Query Caching CACHED property not available

<cfquery datasource="#Application.SessionMgr.getValue("DataSource")#"  name="getCounts" cachedwithin="#createtimespan(0,0,5,0)#">
     Some SQL Query Here
</cfquery>
<CFDump var="#getCounts.cached#">
the dump produces an error
file : ERRORlocal.CFM
Message Content:Element CACHED is undefined in GETCOUNTS
But if I CFDump the query, then I see the CACHED value as part of the dump ... Why can't I access it?
D?

there is one thing you missed... the result attribute
<cfquery datasource="DSN"  name="getCounts" cachedwithin="#CacheTime#" result="BucketQueryResult">
</cfquery>
<cfif BucketQueryResult.cached eq true>
<CFElse>
</cfif>
hth,
D.

Similar Messages

  • Blinking property not available in PDA Module?

    Hi All,
    Is the Blinking property not available for a Boolean indicator in PDA Module?
    hellolv

    Perhaps you can help investigate this problem. I am attaching a zip file that has a sequence file and 4 VIs. The sequence has 2 steps, one that calls blinktop2.vi and the other calls blinktop.vi. Each calls a subVI that blink certain controls. Here is what I have found in experimenting with these.
    After starting the sequence editor if I run only the first step, skipping the second, then blinking will not occur. After doing this not other blinking property in any other VI called by TS will work either.
    I then ***shut down*** LV. Returning to my sequence I run the second step skipping the first. Now the controls blink. Without shutting down LV if I then execute the first step it also blinks.
    In the second step Blinking is first set to False before being set to
    true. What do you find?
    Attachments:
    blink.zip ‏33 KB

  • RFC Debugging and Cache Admin not available in PPM 5.0

    Hi Experts,
    RFC Debugging  and Cache Administration is not available  in PPM 5.0 as compared to RPM 4.5. Is it due to single ABAP stack model of PPM 5.0 ? Complete transformation from java to ABAP webdynpro components  ?  Any idea on this ?
    Best Regards,
    Siva

    Siva,
    Yes, you are right, As enerything is moved to ABAP stack in SAP PPM 5.0. RFC Debugging and Cache Administration is not required in SAP PPM 5.0.
    Cheers
    Amit

  • Cached data not available to clients

    Hi All ,
    I am loading table data to cache using putAll() . Data loads fine and if i try to access the data in same program , i can get it .
    but if i try to access the cached data with some other client it is not available (getting null pointer exceptions ) .
    is there any config change required to keep data in cache for other clients .I am using JPA beans for persistence .
    Code to access the data
    =========================================
    public static void main(String[] args) {   
    CacheFactory.ensureCluster();
    NamedCache currcache = CacheFactory.getCache("ClientMaster");
    //These two commented lines of code is used to load data in cache using putAll .
    // DataLoad data = new DataLoad();
    //data.preload(currcache);
    Set setKeys = currcache.keySet();
    Map mapResult = currcache.getAll(setKeys);
    ClientMaster cm =(ClientMaster)mapResult.get("ABC");
    System.out.println("key"+cm.getClientDesc());
    =========================================
    Here is my cache-config file .
    <?xml version="1.0" encoding="windows-1252" ?>
    <cache-config>
    <caching-scheme-mapping>
    <cache-mapping>
    <!-- Set the name of the cache to be the entity name -->
    <cache-name>ClientMaster</cache-name>
    <!-- Configure this cache to use the scheme defined below -->
    <scheme-name>jpa-distributed</scheme-name>
    </cache-mapping>
    <cache-mapping>
    <!-- Set the name of the cache to be the entity name -->
    <cache-name>CurrencyMaster</cache-name>
    <!-- Configure this cache to use the scheme defined below -->
    <scheme-name>jpa-distributed</scheme-name>
    </cache-mapping>
    </caching-scheme-mapping>
    <caching-schemes>
    <distributed-scheme>
    <scheme-name>jpa-distributed</scheme-name>
    <service-name>JpaDistributedCache</service-name>
    <backing-map-scheme>
    <read-write-backing-map-scheme>
    <!--
    Define the cache scheme
    -->
    <internal-cache-scheme>
    <local-scheme>
    </local-scheme>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <class-name>com.tangosol.coherence.jpa.JpaCacheStore</class-name>
    <init-params>
    <!--
    This param is the entity name
    This param is the fully qualified entity class
    This param should match the value of the
    persistence unit name in persistence.xml
    -->
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>project1.{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>Project1</param-value>
    </init-param>
    </init-params>
    </class-scheme>
    </cachestore-scheme>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
    <autostart>true</autostart>
    </distributed-scheme>
    </caching-schemes>
    </cache-config>

    Now Getting this error when trying load data in cache using putAll , This is really weird as this same piece of code worked fine earlier.
    Any ideas ?
    ============================================================
    2011-01-11 13:45:47.165/14.266 Oracle Coherence GE 3.6.1.0 <Error> (thread=Distr
    ibutedCache:JpaDistributedCache, member=6): BackingMapManager com.tangosol.net.D
    efaultConfigurableCacheFactory$Manager: failed to instantiate a cache: ClientMaster
    2011-01-11 13:45:47.165/14.266 Oracle Coherence GE 3.6.1.0 <Error> (thread=DistributedCache:JpaDistributedCache, member=6):
    (Wrapped: Missing or inaccessible constructor "com.tangosol.coherence.jpa.JpaCacheStore(String,String,String)"
    <class-scheme>
    <class-name>com.tangosol.coherence.jpa.JpaCacheStore</class-name>
    <init-params>
    <!--
    This param is the entity name
    This param is the fully qualified entity class
    This param should match the value of the
    persistence unit name in persistence.xml
    -->
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>ClientMaster</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>project1.ClientMaster</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>Project1</param-value>
    </init-param>
    </init-params>
    </class-scheme>) java.lang.reflect.InvocationTargetException
    at com.tangosol.util.Base.ensureRuntimeException(Base.java:293)
    at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2542)
    at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2426)
    at com.tangosol.coherence.jpa.JpaCacheStore.<init>(JpaCacheStore.java:44
    ... 29 more
    Caused by: java.lang.ClassNotFoundException: project1.ClientMaster
    at java.net.URLClassLoader$1.run(URLClassLoader.java:200)
    at java.security.AccessController.doPrivileged(Native Method)
    at java.net.URLClassLoader.findClass(URLClassLoader.java:188)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:252)
    at com.tangosol.coherence.jpa.JpaCacheLoader.initialize(JpaCacheLoader.j
    ava:143)

  • Named query cache not hit

    Hi,
    I'm using Toplink ORM 10.1.3.
    I have a table called STORE_CONFIG which has a primary key called KEYWORD (a VARCHAR2). The POJO mapped to this table is called StoreConfig.
    In the JDeveloper (10.1.3.1.0) mapping workbench I've defined a named query to query by the PK called "getConfigPropertyByKeyword". The type of the named query is ReadObjectQuery.
    SELECT keyword, key_value
    FROM STORE_CONFIG
    WHERE (keyword = #key)
    Under the options tab I have the following settings:
    Cache Statement: true
    Bind Parameters: true
    Cache Usage: Check Cache by Primary Key
    The application logs show that the same database queries are executed multiple times for the same PK (keyword)! Why is that? Shouldn't it be checking the Object Cache rather than going to the DB?
    I've tried it with "Cache Statement: false" and "Bind Parameters: false" with the same problem.
    If I click the Advanced tab and check "Cache Query Results" then the database is not hit twice for the same record. However it was my understanding that since I am querying by PK that I wouldn't need to set "Cache Query Results".
    Doesn't "Cache Query Results" apply to the Query Cache and not the Object Cache?

    Your issue seems to be that you are using custom SQL for the query, not a TopLink expression. When you use an Expression query TopLink's know if the query is by primary key and can get a cache hit.
    When you use custom SQL, TopLink does not know that the SQL is by primary key, so does not get a cache hit.
    You could either use an Expression for the query,
    or when using custom SQL you should be able to name your query argument the same as your database field defined as the primary key in your descriptor (case sensitive).
    i.e.
    SELECT keyword, key_value
    FROM STORE_CONFIG
    WHERE (keyword = #KEYWORD)

  • Query.setRange() not available on standard javax.jdo.Query interface?

    Hi Kodo people, I'm trying out Kodo 3.2.0, most specifically for the
    spiffy new setRange() functionality. However, I'm somewhat confused
    because while the docs clearly indicate that the setRange() method is
    available on the standard javax.jdo.Query interface, it's not available
    on the javax.jdo.Query class that ships with Kodo 3.2.0. I have to cast
    my query to a KodoQuery in order for this to work. Is this expected?
    Thanks,
    -Mike
    Michael Allen
    Technical Lead
    PGP Corporation

    I think you missed the note at the beginning of the query chapter in the docs:
    Much of the functionality we discuss in this chapter is new to JDO 2. Though
    Kodo supports all of the features defined in the following sections, many JDO
    implementations may not. Additionally, because official JDO 2 jars are not yet
    available, you will have to cast your query objects to kodo.query.KodoQuery to
    access any JDO 2 APIs. The UML diagram above depicts these APIs in bold. For
    simplicity, casts have been left out of the example code throughout the chapter.

  • 0repmode  "Restatement" value is not available in the BEx query

    Dear SEM-BCS experts,
    We need to select 0repmode  "Restatement" value in a BEx query, but it is not available (only # and Standard). The query is based on the SEM-BCS virtual cube. We use SEM-BCS version 6.02.
    I would be grateful if you could help us to resolve this problem.
    Mila

    Dear Mila,
    Report mode. "Restatement" availability purely depends on the settings made for
    the transaction data transfer ; between Transaction cube (BCS cube) to Reporting cube (BW cube). 
    Please check in the info package created for Transaction cube(BCS Cube) where in you can see the available options. If the data transfer is happening with reporting mode "Restatement" then it should be available.
    Most of the people use one Standard (S), Purchase (P).. but Restatement (R) option also available.
    Hope this will give you clarity.
    Thanks
    kamal

  • The Security Token Service is not available error on dedicated Distributed Cache server

    I have an error on a dedicated Distributed Cache server stating that the Security Token Service is not available.  I was under the impression that when Distributed Cache was running on a dedicated server that the only service that should be enabled
    is Distributed Cache. 
    The token service is working as expected on all other servers but this one.  Does this service need to be started or should I just ignore this error message?
    Jennifer Knight (MCITP, MCPD)

    as per my little experience with 2013, if STS is working fine on Web server then I am sure that sharepoint will be fine...Distributed cache stores the ST issued by STS. NO need to worry about this error.
    Login
    Token Cache
    DistributedLogonTokenCache
    This
    cache stores the security token issued by a Secure Token Service for use by any web server in the server farm. Any web server that receives a request for resources can access the security token from the cache, authenticate the user, and provide access to the
    resources requested.
    I would say check the ULS logs and get more details about the error why its not working on that server.
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • Entity-Named Query Caching - Need live results from DB not Cached

    I've developed an application EJB3 web application which uses Entity Beans and NamedQueries to retrieve my results.
    Aswell as my main EJB3 web application I also have an external interface (Oracle Stored Procedure) which is updating data in my database tables. The problem is that the when the NamedQueries are called they only bring back "Cached" resultsets and not the live results that exists in the database at that time. If my Stored Procedure updates a value (or if I update a value via Toad for testing) the NamedQuery does not bring back these changes. From what I have read about NamedQueries it seems that they do actually query the live db but send back the cached values for performance reasons.
    Is there a way to turn off caching of Query results?
    If not is there another way to deal with this issue?
    ps. I'm using Jdev 10.1.3.4 and appserver 10.1.3.1

    Anyone?

  • Cache....schema not available

    Hi guys...
    I'm having Cache problems ...
    I've been checking the logs and I've discovered the error. I thought that the error could be a cache problem but when I was at the Integration Directory -Cache Notifications, I noticed that my Integration Server(ABAP Cache) was having problems.
    I went to the problems tag and saw the same error from the log.
    At the Cache Instance, Integration Server(Central Adapter Engine), I had the same error for my communication channel.
    This error is not enabling the file adapter to be created... can anyone help me...?
    I'm over this error for the last couple of days, I've read all the howtos (maybe not...) and helps...
    Any suggestions?
    ERROR
    com.sap.aii.af.service.cpa.CPAException: Schema not available for File|3b787a8035c111d6bbe0efe50a1145a5|http://sap.com/xi/XI/System.
         at com.sap.aii.af.service.cpa.impl.cache.directory.DirectoryDataSAXHandler.endElement(DirectoryDataSAXHandler.java:262)
         at com.sap.engine.lib.xml.parser.handlers.SAXDocHandler.endElement(SAXDocHandler.java:154)
         at com.sap.engine.lib.xml.parser.XMLParser.scanEndTag(XMLParser.java:1826)
         at com.sap.engine.lib.xml.parser.XMLParser.scanElement(XMLParser.java:1722)
         at com.sap.engine.lib.xml.parser.XMLParser.scanContent(XMLParser.java:2298)
         at com.sap.engine.lib.xml.parser.XMLParser.scanElement(XMLParser.java:1719)
         at com.sap.engine.lib.xml.parser.XMLParser.scanDocument(XMLParser.java:2701)
         at com.sap.engine.lib.xml.parser.XMLParser.parse0(XMLParser.java:162)
         at com.sap.engine.lib.xml.parser.AbstractXMLParser.parseAndCatchException(AbstractXMLParser.java:126)
         at com.sap.engine.lib.xml.parser.AbstractXMLParser.parse(AbstractXMLParser.java:136)
         at com.sap.engine.lib.xml.parser.AbstractXMLParser.parse(AbstractXMLParser.java:209)
         at com.sap.engine.lib.xml.parser.Parser.parseWithoutSchemaValidationProcessing(Parser.java:270)
         at com.sap.engine.lib.xml.parser.Parser.parse(Parser.java:331)
         at com.sap.engine.lib.xml.parser.SAXParser.parse(SAXParser.java:125)
         at javax.xml.parsers.SAXParser.parse(SAXParser.java:345)
         at javax.xml.parsers.SAXParser.parse(SAXParser.java:143)
         at com.sap.aii.af.service.cpa.impl.cache.directory.DirectoryDataParser.updateCentralCache(DirectoryDataParser.java:54)
         at com.sap.aii.af.service.cpa.impl.cache.CacheManager.updateCacheWithDirectoryData(CacheManager.java:713)
         at com.sap.aii.af.service.cpa.impl.cache.CacheManager.performCacheUpdate(CacheManager.java:595)
         at com.sap.aii.af.service.cpa.impl.cache.CacheManager$CacheUpdateRunnable.run(CacheManager.java:440)
         at com.sap.engine.frame.core.thread.Task.run(Task.java:60)
         at com.sap.engine.core.thread.impl5.SingleThread.execute(SingleThread.java:73)
         at com.sap.engine.core.thread.impl5.SingleThread.run(SingleThread.java:145)

    Well, the problem is not knowing where to refresh the cache, the problem is that the cache is not performing the refresh...
    I'm having this error:
    <?xml version="1.0" encoding="UTF-8"?>
    <CacheRefreshError>
      <EngineType>AE</EngineType>
      <EngineName>af.xir.srv04</EngineName>
      <RefreshMode>C</RefreshMode>
      <GlobalError>
        <Message>
          CPA Cache not updated with directory data.
        </Message>
        <Trace>
          com.sap.aii.af.service.cpa.impl.exception.CPADirectoryCacheException: Couldn't open Directory URL (http://srv04:50000/dir/hmi_cache_refresh_service/ext?method=CacheRefresh&mode=C&consumer=af.xir.srv04), due to: HTTP 503: Service Unavailable
         at com.sap.aii.af.service.cpa.impl.cache.directory.DirectoryAccess.getDataFromDirectory(DirectoryAccess.java:382)
         at com.sap.aii.af.service.cpa.impl.cache.directory.DirectoryAccess.getDeltaCacheUpdateXML(DirectoryAccess.java:545)
         at com.sap.aii.af.service.cpa.impl.cache.CacheManager.performCacheUpdate(CacheManager.java:551)
         at com.sap.aii.af.service.cpa.impl.cache.CacheManager.initializeCPACache(CacheManager.java:260)
         at com.sap.aii.af.service.cpa.impl.startup.StartupManager.initializeCPACache(StartupManager.java:220)
         at com.sap.aii.af.service.cpa.impl.j2ee.sapengine630.ServiceFrameImpl.startCPACache(ServiceFrameImpl.java:129)
         at com.sap.aii.af.service.cpa.impl.j2ee.sapengine630.ServiceFrameImpl.start(ServiceFrameImpl.java:49)
         at com.sap.engine.core.service630.container.ServiceRunner.startApplicationServiceFrame(ServiceRunner.java:201)
         at com.sap.engine.core.service630.container.ServiceRunner.run(ServiceRunner.java:142)
         at com.sap.engine.frame.core.thread.Task.run(Task.java:60)
         at com.sap.engine.core.thread.impl5.SingleThread.execute(SingleThread.java:73)
         at com.sap.engine.core.thread.impl5.SingleThread.run(SingleThread.java:145)
        </Trace>
      </GlobalError>
    meaning that there is a service, i suppose that this service will perform the full refresh, is unavailable.
    Can you help me figuring this out?How can I activate the service?
    I've launched the logviewer and i've seen that the service "SAP XI AF CPA Cache" is active, so it's not this service that is responsible... Can anyone help?

  • Live cache lock handler not available

    Hi we are frequently getting this error Live cache lock handler not available and this causes our process chains to fail.
    Can anyone please explain in details what this error refers to and direct to any good links for error resolution.

    Hello,
    1. As SAP customer, you could create the SAP message to the component BC-DB-LCA and get SAP support.
    2. If you set the flag 'activate livecache lock' in tcode /SAPAPO/MSDP_ADMIN for your planning area on SCM system, please run the connection test for the LEA connection in /ndb59 to check the LEA connection settings.
       The liveCache installation guide has additional information about LEA connection.
    Regards, Natalia Khlopina

  • NO_MAPPINGPROGRAM_FOUND : Mapping program is not available in runtime cache

    Hi,
    I have a SOAP to RFC scenario. Where the source is a service interface and target is ZBAPI. When I test this interface I get 
    'NO_MAPPINGPROGRAM_FOUND : Mapping program is not available in runtime cache' error.
    I have tried the following things but none worked.
    1. Checked the Message Mappings and reimported it in the Operation Mapping.
    2. Deleted and created the Operation Mapping again.
    3. Cleared the SLD Cache.
    4. Performed the Complete Cache Refresh from SXI_CACHE.
    5. Reimported the ZBAPI.
    6. Deleted all the IR and ID objects and created it once again after performing the Cache refresh.
    Note: I had one observation that the newly created Operation Mapping object had the same object ID as that of the deleted one.

    Hi experts,
    I have a SOAP to RFC scenario. Where the source is a service interface and target is ZBAPI. When I test this interface I get 
    'NO_MAPPINGPROGRAM_FOUND : Mapping program is not available in runtime cache : Object ID <ID of Operation Mapping>' error.
    I have tried the following things but none worked.
    1. Checked the Message Mappings and reimported it in the Operation Mapping.
    2. Deleted and created the Operation Mapping again.
    3. Cleared the SLD Cache.
    4. Performed the Complete Cache Refresh from SXI_CACHE.
    5. Reimported the ZBAPI.
    6. Deleted all the IR and ID objects and created it once again after performing the Cache refresh.
    Note: I had one observation that the newly created Operation Mapping object had the same object ID as that of the deleted one.
             The ZBAPI is a wrapper for a standard BAPI with Request, Response and Exception as a structure.
    Can anyone please highlight on the above issue also please consider the note above.
    Regards,
    Rohit.

  • Page Zero (0) - Regions Caching / Page Caching not available

    Hi,
    My page zero display information from a calendar table (fiscal year, etc.). This information will be the same all day long (get updated at midnight). It's not possible to enable caching of regions on page zero. It's also impossible to enable caching of the page zero.
    Why?
    Thanks,
    Louis-Guillaume
    [My Blog|http://lgcarrier.blogspot.com]
    [APEX and jQuery demos|http://www.insum.ca/jquery]

    Louis-Guillaume,
    Good question. We have asked ourselves this and are considering removing the restriction in 4.0. I think it initially had to do with preventing unwanted side effects and anomalies although none of us can recall the details right now. We will have to carefully look at ways to make this something developers can use easily while preventing strange results. First of all, caching of page 0 doesn't make sense. You want to be able to cache regions on page 0. Now, say your page 0 has two regions, P0_CACHED and P0_DYN (one cached, one dynamic). And your page 10 has region P10_CACHED. When page 10 is rendered, you'll get:
    P0_CACHED
    P0_DYN
    P10_CACHED
    That looks okay, you get one cached region from page 0, one dynamic region from page 0, and one cached region from page 10. Of course you have controls with which to purge any individual region from the cache so you can cause P0_CACHED to be refreshed whenever you like.
    But say page 20 is a cached page. When it renders you'll get:
    P0_CACHED <-- not from the page 0 region cache but from the page 20 page cache, regardless of the "stale-ness" of the page 0 cache
    P0_DYN <-- not rendered dynamically from page 0 but retrieved as part of the page 20 page cache. But this content may be different from the content for the same page 0 region displayed on page 10 one second ago or one second from now.
    P20_DYN <-- from the page cache as is normal for a dynamic region on a cached page
    Hardly a thorough treatment, I realize. Just wanted you to know some of the aspects we have to consider. We will also have to read this thread again to come back up to speed: V3 Caching - any more info? .
    Scott

  • SAP BW 3.5 Query Cache - no data for queries

    Dear experts,
    we do have a problem with the SAP BW Query Cache (BW 3.5). Sometimes the
    problem arise that after the queries have been pre-calculated using web templates there are no figures available when running a web-report, or when doing a drilldown or filter navigation. A solution to solve that issue is to delete the cache for that query, however this solution is not reasonable or passable. The problem occurs non-reproducible for different queries.
    Is this a "normal" error of the SAP BW we have to live with, or are there any solutions for it? Any hints are greatly appreciated.
    Thanks in advance & kind regards,
    daniel

    HI Daniel
    Try to wirk without cache for those queries.
    Anyway, you should check how the cache option is configured for those queries .
    You can see that, in RSRV tx
    Hope this help

  • Best way to query cache - get() vs filters?

    Hi,
    I am in a dillemma. Whether to use NamedCache.get() or entrySet(filter) methods to query the cache. Please guide me..
    My understanding is that when using
    1. get() or getAll(), Coherence checks whether the entry is in the cache, if it does exist in the cache, Coherence get its from the DataStore
    2. entrySet(Filters), Coherence just checks the cache and returns results based on whats available in the cache.
    In that case, Isnt it better to use get instead of entrySet in a case where one is not sure whether up to date data is available in the Cache?
    1. What is the difference between using a get and using entrySet ?
    2. How does one make sure that up to date data is available in the cache when not using a write-behind scenario?
    I am newbie...Gurus, please guide me..

    sjohn wrote:
    Hi,
    I am in a dillemma. Whether to use NamedCache.get() or entrySet(filter) methods to query the cache. Please guide me..
    My understanding is that when using
    1. get() or getAll(), Coherence checks whether the entry is in the cache, if it does exist in the cache, Coherence get its from the DataStoreThat's not the relevant part.
    In this case because you specified the keys to the entries, Coherence knows exactly where each entry resides, and optimally communicates with the owner nodes (1 network call / owner node) to get the data, and it will return the data without deserialization on the owner node.
    2. entrySet(Filters), Coherence just checks the cache and returns results based on whats available in the cache.
    In this case, depending on the actual filter (hierarchy) Coherence has to contact all nodes (if you did not filter it with KeyAssociatedFilter or PartitionFilter), which is not scalable, then depending on the filter(s) used it may have to deserialize possibly all cached entries (which is expensive) to evaluate the filter. On the other hand this method is usable for non-key-based access, which the get/getAll is not able to do.
    In that case, Isnt it better to use get instead of entrySet in a case where one is not sure whether up to date data is available in the Cache?
    1. What is the difference between using a get and using entrySet ?Heaven and earth...
    2. How does one make sure that up to date data is available in the cache when not using a write-behind scenario?
    You have to preload it, or trigger fetching it from a cache-store.
    Best regards,
    Robert

Maybe you are looking for

  • .jar file suddenly not working

    I have some minimal experience working with .jar files, but I am baffled here. Recently, the jar file for one of my programs stopped working. I had just made a change to my main class, so i repackaged the jar. Now it says "Can't find main class" when

  • Pulseaudio - user settings and microphone auto adjustment

    Hi all, Few questions if I may - I have search the wiki/forum but couldn't find any answers. When recording my webcam with ffmpeg along with my webcam microphone the recording level of the microphone is automatically adjusted by pulseaudio (I am assu

  • Mac Mini with 24" LED ACD

    I've been reading several threads about the new mac mini's not being compatible with the 2008-2009 24" LED ACD (flickering and blank screens), are the bugs worked out, because I want to buy a mac mini but not if I'm going to have problems with my dis

  • JAXB mapping of field to sub nodes

    I have the following XML I consume: <root>      <node1>           <node2 expirationDate="10/10/2012" />      </node1> </root> And the following Java class: @XmlRootElement(name= "root") class Foo private String expirationDate; My question is whether

  • Intermittent JNI Panic exception when accessing Oracle

    I was wondering if anyone had seem this error before and if so, did you find a resolution. The system is a Solaris system running oracle 8.0.6 using the JDBC 9.0.1 thin client drivers. Any help would be greatly appriciated. The error as reported is:-