Distributed cache size limit

Hi,
I want to create a distributed cache with 2 node.
Each node can have maximum 500 entries.
Total entries in cache of both node together should not exceed more than 1000.
If user try to put more than 1000 element in cache then some old entries from cache (LRU, LFU) should be removed from cache and new entries should be added.
Can you please help me with the schema for the above scenario.
Your help will be appreciated.
Thanks & Regards,
Viral Gala

Hi,
I tried below code <high-units> was not working
Cache size is 1010  (greater than 500)
Java code
package com.splwg.ccb.domain.pricing;
import com.tangosol.net.CacheFactory;
import com.tangosol.net.NamedCache;
public class CoherenceSizeTest {
     public static final int NO_OF_MAIN = 4;
     public static void main(String[] args) {
     NamedCache cache = CacheFactory.getCache("CheckSize");
      for (int i = 0; i < 1010; i++) {
            String key = "key" + i;
            cache.put(key, new Long(i));
     System.out.println(" Cache size : " + cache.size());
config file
<?xml version="1.0"?>
<!DOCTYPE cache-config SYSTEM "cache-config.dtd">
<cache-config>
    <caching-scheme-mapping>
        <cache-mapping>
            <cache-name>CheckSize</cache-name>
            <scheme-name>default-distributed</scheme-name>
        </cache-mapping>
    </caching-scheme-mapping>
    <caching-schemes>
        <distributed-scheme>
            <scheme-name>default-distributed</scheme-name>
            <service-name>DistributedCache</service-name>
            <backing-map-scheme>
                <local-scheme />
            </backing-map-scheme>
            <high-units>500</high-units>
            <autostart>true</autostart>
        </distributed-scheme>
    </caching-schemes>
</cache-config>
Console output
MasterMemberSet(
  ThisMember=Member(Id=1, Timestamp=2014-11-07 16:31:21.123, Address=10.180.7.97:8088, MachineId=16932, Location=site:,machine:OFSS310723,process:4036, Role=SplwgCcbDomainCoherenceSizeTest)
  OldestMember=Member(Id=1, Timestamp=2014-11-07 16:31:21.123, Address=10.180.7.97:8088, MachineId=16932, Location=site:,machine:OFSS310723,process:4036, Role=SplwgCcbDomainCoherenceSizeTest)
  ActualMemberSet=MemberSet(Size=1
    Member(Id=1, Timestamp=2014-11-07 16:31:21.123, Address=10.180.7.97:8088, MachineId=16932, Location=site:,machine:OFSS310723,process:4036, Role=SplwgCcbDomainCoherenceSizeTest)
  MemberId|ServiceVersion|ServiceJoined|MemberState
    1|3.7.1|2014-11-07 16:31:24.375|JOINED
  RecycleMillis=1200000
  RecycleSet=MemberSet(Size=0
TcpRing{Connections=[]}
IpMonitor{AddressListSize=0}
Cache size : 1010

Similar Messages

  • Regarding cache size limit

    hi all,
    What is the maximum possible cache size set for Webcache.
    Any help would be much appreciated
    Regards,
    Geeta

    hi all,
    What is the maximum possible cache size set for Webcache.
    Any help would be much appreciated
    Regards,
    Geeta

  • Limitation on number of objects in distributed cache

    Hi,
    Is there a limitation on the number (or total size) of objects in a distributed cache? I am seeing a big increase in response time when the number of objects exceeds 16,000. Normally, the ServiceMBean.RequestAverageDuration value is in the 6-8ms range as long as the number of objects in the cache is less than 16K - I've run our application for weeks at a time without seeing any problems. However, once the number of objects exceeds the magic number of 16K the average request duration almost immediately jumps to over 100ms and continues to climb as more objects are added.
    I'm fairly confident that the cache is indexed properly (as Dimitri helped us with that). Are there any configuration changes that could possibly help out here? We are using Coherence 3.3.
    Any suggestions would be greatly appreciated.
    Thanks,
    Jim

    Hi Jim,
    The results from the load test look quite normal, the system fairly quickly stabilizes at a particular performance level and remains there for the duration of the test. In terms of latency results, we see that the cache.putAll operations are taking ~45ms per bulk operation where each operation is putting 100 1K items, for cache.getAll operations we see about ~15ms per bulk operation. Additionally note that the test runs over 256,000 items, so it is well beyond the 16,000 limit you've encountered.
    So it looks like your application are exhibiting different behavior then this test. You may wish to try to configure this test to behave as similarly to yours as possible. For instance you can set the size of the cache to just over/under 16,000 using the -entries parameter, set the size of the entries to 900 bytes using the -size parameter, and set the total number of threads per worker using the -threads parameter.
    What is quite interesting is that at 256,000 1K objects the latency measured with this test is apparently less then half the latency you are seeing with a much smaller cache size. This would seem to point at the issue being related to or rooted in your test. Would you be able to provide a more detailed description of how you are using the cache, and the types of operations you are performing.
    thanks,
    mark

  • Generate Events based on Cache size

    Hi,
    I have a use case where I need to generate events when the size of the cache has reached a specific value. The cache will be highly transactions (lot of inserts and deletes totaling to 6 million transactions per day). Also, I cannot limit the size of the cache as I don't want any data to be lost (trying to insert when cache size has reached its limit).
    One of the ways I could do this is to use a MapListener to generate an event on every insert to the Cache, and then check for size each time to decide if the further processing needs to be done. With the nature of the traffic in the cache, the items in the cache are to be processed when the size threshold is reached and then deleted. But, I'm not sure how much of an overhead these events and listeners will have (with 6 million items going through the cache each day)
    Is there any other more efficient way to do this? I have looked at Continuous Query cache which uses a filter based on a cache. But, I'm not able to figure out a way to get the size of the cache using a filter.
    Any pointers are helpful
    Thanks in advance
    Regards
    Vikas
    Edited by: vikascv on Nov 17, 2009 8:48 AM

    Hi Vikas,
    Typically a size type calculation is the cumulative result of all storage enabled nodes, i.e. a Count aggregation. In your use case you want to be made aware of changes to an aggregation but as an aggregation is point in time of request opposed to continually being updated it does not fit your use case. One option is to create a MapEventTransformer that you register with the cache which increments a counters cache (on a separate service) using an EntryProcessor, i.e.
    MapEventTransformer...
    public class CounterMapEventTransformer implements MapEventTransformer, PortableObject
        String m_cacheName;
        public CounterMapEventTransformer()
        public CounterMapEventTransformer(final String cacheName)
            m_cacheName = cacheName;
        public MapEvent transform(MapEvent event)
            Long prevCount    = null;
            Long currentCount = null;
            switch( event.getId() )
                case( MapEvent.ENTRY_INSERTED ):
                    // perform counter++
                    currentCount = (Long) CacheFactory.getCache("counters").invoke(m_cacheName, new CounterOp(true));
                    break;
                case( MapEvent.ENTRY_DELETED ):
                    // perform counter--
                    currentCount = (Long) CacheFactory.getCache("counters").invoke(m_cacheName, new CounterOp(false));
                    break;
            MapEvent newEvent = new MapEvent(event.getMap(), event.getId(), event.getKey(), currentCount-1, currentCount);
            return newEvent;
        public void writeExternal(PofWriter out) throws IOException
            out.writeString(0, m_cacheName);
        public void readExternal(PofReader in) throws IOException
            m_cacheName = in.readString(0);
        }Incrementing EP...
    public class CounterOp extends AbstractProcessor implements PortableObject
        boolean m_increment;
        public CounterOp()
        public CounterOp(boolean increment)
            m_increment = increment;
        @Override
        public Object process(Entry entry)
            Long newVal = null;
            if( !entry.isPresent() )
                entry.setValue(newVal = Long.valueOf(1));
            else
                newVal = ((Long)entry.getValue()).longValue() + (isIncrement() ? 1 : -1);
                entry.setValue( newVal );
            return newVal;
        protected boolean isIncrement()
            return m_increment;
        public void writeExternal(PofWriter out) throws IOException
            out.writeBoolean(0, m_increment);
        public void readExternal(PofReader in) throws IOException
            m_increment = in.readBoolean(0);
        }cache-config
         <!--
              | listener-cacheconfig.xml | | Copyright 2001-2009 by Oracle. All
              rights reserved. | | Oracle is a registered trademarks of Oracle
              Corporation and/or its | affiliates. | | This software is the
              confidential and proprietary information of Oracle | Corporation. You
              shall not disclose such confidential and proprietary | information and
              shall use it only in accordance with the terms of the | license
              agreement you entered into with Oracle. | | This notice may not be
              removed or altered.
         -->
    <!DOCTYPE cache-config SYSTEM "cache-config.dtd">
    <cache-config>
         <caching-scheme-mapping>
              <cache-mapping>
                   <cache-name>dist-*</cache-name>
                   <scheme-name>distributed-scheme</scheme-name>
              </cache-mapping>
              <cache-mapping>
                   <cache-name>counters</cache-name>
                   <scheme-name>counter-scheme</scheme-name>
              </cache-mapping>
         </caching-scheme-mapping>
         <caching-schemes>
              <!--
        Distributed caching schemes
        -->
              <distributed-scheme>
                   <scheme-name>distributed-scheme</scheme-name>
                   <service-name>DistributedCache</service-name>
                   <serializer>
                        <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
                        <init-params>
                             <init-param>
                                  <param-type>string</param-type>
                                  <param-value>pof-config.xml</param-value>
                             </init-param>
                        </init-params>
                   </serializer>
                   <backing-map-scheme>
                        <local-scheme />
                        <!--
                             <read-write-backing-map-scheme> <internal-cache-scheme>
                             <local-scheme /> </internal-cache-scheme> <cachestore-scheme>
                             <class-scheme>
                             <class-name>com.oracle.coherence.cachecounter.store.CacheCounterStore</class-name>
                             </class-scheme> </cachestore-scheme>
                             </read-write-backing-map-scheme>
                        -->
                   </backing-map-scheme>
                   <autostart>true</autostart>
              </distributed-scheme>
              <distributed-scheme>
                   <scheme-name>counter-scheme</scheme-name>
                   <service-name>CounterDistributedCache</service-name>
                   <serializer>
                        <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
                        <init-params>
                             <init-param>
                                  <param-type>string</param-type>
                                  <param-value>pof-config.xml</param-value>
                             </init-param>
                        </init-params>
                   </serializer>
                   <backing-map-scheme>
                        <local-scheme />
                   </backing-map-scheme>
                   <autostart>true</autostart>
              </distributed-scheme>
         </caching-schemes>
    </cache-config>pof-config...
    <?xml version="1.0" encoding="UTF-8"?>
    <!DOCTYPE pof-config SYSTEM "pof-config.dtd">
    <pof-config>     
       <user-type-list>          
          <!-- include all of the standard Coherence type -->
          <include>coherence-pof-config.xml</include>          
          <!-- our custom types (user types) 1000+ -->          
          <user-type>                 
             <type-id>1000</type-id>            
             <class-name>com.oracle.coherence.cachecounter.event.CounterMapEventTransformer</class-name>          
          </user-type>
          <user-type>                 
             <type-id>1001</type-id>            
             <class-name>com.oracle.coherence.cachecounter.processors.CounterOp</class-name>          
          </user-type>
       </user-type-list>
       <allow-interfaces>true</allow-interfaces>
       <allow-subclasses>true</allow-subclasses>
    </pof-config>Test class
    public class MapEventTransformerTest
        public static final String m_cacheName = "dist-test1";
        public static void setUp()
            System.setProperty("tangosol.coherence.cacheconfig", "counter-cacheconfig.xml");
            System.setProperty("tangosol.coherence.distributed.localstorage", "false");
        public static void loadVals()
            final NamedCache cache = CacheFactory.getCache(m_cacheName);
            final int MAX_SIZE = 100;
            for( int i=0; i<MAX_SIZE; i++ )
                cache.put(i, "msg "+i);
        public static void removeVals()
            final NamedCache cache = CacheFactory.getCache(m_cacheName);
            final int MAX_SIZE = 100;
            for( int i=0; i<MAX_SIZE; i++ )
                if( i%2==0 )
                    cache.remove(i);
        public static void testTransformer()
            final NamedCache cache = CacheFactory.getCache(m_cacheName);
            cache.addMapListener(new CounterAwareListener(),
                    new MapEventTransformerFilter(AlwaysFilter.INSTANCE, new CounterMapEventTransformer(m_cacheName)),
                    false);
        public static void main(String[] args) throws IOException
            setUp();
            testTransformer();
            loadVals();
            removeVals();
            System.out.println("press a key to end...");
            System.in.read();
        public static class CounterAwareListener extends MultiplexingMapListener
            @Override
            protected void onMapEvent(MapEvent evt)
                System.out.format( "# of items in cache: %d\n", evt.getNewValue() );
        }Note: Events could be received out of order but for your use case where you are interested in a threshold you may not care! -- food for thought :)
    --harvey                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Distributed cache doesn't work. Please help.

    Hi,
    I am trying to use distributed cache by:
    1. use coherence as 2nd level cache for hibernate in the application server (weblogic 9). Configuration as follows:
    <cache-config>
    <caching-scheme-mapping>
    <cache-mapping>
    <cache-name>*</cache-name>
    <scheme-name>DistributedInMemoryCache</scheme-name>
    </cache-mapping>
    </caching-scheme-mapping>
    <caching-schemes>
    <distributed-scheme>
    <scheme-name>DistributedInMemoryCache</scheme-name>
    <service-name>DistributedCache</service-name>
    <backing-map-scheme>
    <local-scheme>
    <high-units>{size-limit 0}</high-units>
    </local-scheme>
    </backing-map-scheme>
    <autostart>true</autostart>
    </distributed-scheme>
    </caching-schemes>
    </cache-config>
    2. start a standalone jvm locally to form a 2 nodes cluster with the coherence node above using the same cache-config.xml as shown above. I use the following command to start this cache:
    %JAVA_HOME%/bin/java -server -showversion -jar %LIB_HOME%\coherence.jar
    I enable jmx on both jvm, the coherence cache on the coherence node in weblogic shows a bunch of caches exist (I use @Cache(usage = CacheConcurrencyStrategy.READ_ONLY) annontation in my code to enable cache for an entity), also there are a few cache services including DistributedCache started. However, from the standalone jvm, I didn't see any cache nor Distributed cache service (the only service is management from jconsole).
    My goal is to use the standalone jvm as a cache server that holds all the caches while the coherence in weblogic as the client which has no local storage but a near cache.
    However, I could not even get the distributed cache to work. Please help.
    Thanks

    Hi,
    To start the cache server you need to use the command like this:
    %JAVA_HOME%/bin/java -server -Dtangosol.coherence.cacheconfig=/path/to/cache_configuration_descriptor -cp %LIB_HOME%\coherence.jar com.tangosol.net.DefaultCacheServer
    Regards,
    Dimitri

  • Error message when using a MessageListener using a distributed cache

    Hi --
    I'm getting the following error message when I attach a message listener to a distributed cache. I get the same message if I attach the listener to the NearCache in front of the DistributedCache, or to the DistributedCache itself.
    My message listener listens for a create() operation and writes the created value out to the database. Both the key and value are java objects that are getting "serialized" when they're pushed in the cache. The listener is never called.
    The error spits out two messages, which look like:
    2003-04-07 21:48:05.281 Tangosol Coherence 2.1/239 <Error> (thread=DistributedCache:EventDispatcher): An exception occurred while dispatching this event:
    CacheEvent: MapEvent{com.tangosol.coherence.component.util.daemon.queueProcessor
    .service.DistributedCache$BinaryMap added: key=Binary(length=269, value=0x0005AC
    ED000573720021636F6D2E6F6C742E646174612E696E7465726E616C2E4461746162617365554944
    6FABB5383C6013B402000078720021636F6D2E6F6C742E646174612E696E7465726E616C2E416273
    7472616374554944D04F591196E4DC1B0200024C000D657874656E73696F6E4461746174000F4C6A
    6176612F7574696C2F4D61703B4C0009756964537472696E677400124C6A6176612F6C616E672F53
    7472696E673B7870737200116A6176612E7574696C2E486173684D61700507DAC1C31660D1030002
    46000A6C6F6164466163746F724900097468726573686F6C6478703F400000000000087708000000
    0B0000000078740011363930395F436F6D706F6E656E74426964), value=Binary(length=1069,
    value=0x0005ACED000573720026636F6D2E6562726576696174652E61756374696F6E2E6269642
    E436F6D706F6E656E744269648EC95C4DE33A88D802000D5A0007626573744269644A000B6269645
    3657175656E6365440004636F73745A0007696E697469616C5A00066E65774269645A00067469654
    2696444000576616C75654C000B61756374696F6E4D6F64657400294C636F6D2F656272657669617
    4652F61756374696F6E2F6576656E742F41756374696F6E4D6F64653B4C000A61756374696F6E554
    9447400124C636F6D2F6F6C742F646174612F5549443B4C000A636F6D70616E7955494471007E000
    24C000C636F6D706F6E656E7455494471007E00024C000A737472696E67436F73747400124C6A617
    6612F6C616E672F537472696E673B4C000B737472696E6756616C756571007E00037872002D636F6
    D2E6562726576696174652E636F6D6D6F6E2E416273747261637450657273697374656E744F626A6
    56374497E2729A24CA5790200034C000A637265617465446174657400104C6A6176612F7574696C2
    F446174653B4C000375696471007E00024C000A7570646174654461746571007E000578707372000
    E6A6176612E7574696C2E44617465686A81014B59741903000078707708000000F46B9A286278737
    20021636F6D2E6F6C742E646174612E696E7465726E616C2E44617461626173655549446FABB5383
    C6013B402000078720021636F6D2E6F6C742E646174612E696E7465726E616C2E416273747261637
    4554944D04F591196E4DC1B0200024C000D657874656E73696F6E4461746174000F4C6A6176612F7
    574696C2F4D61703B4C0009756964537472696E6771007E00037870737200116A6176612E7574696
    C2E486173684D61700507DAC1C31660D103000246000A6C6F6164466163746F72490009746872657
    3686F6C6478703F4000000000000877080000000B0000000078740011363930395F436F6D706F6E6
    56E744269647371007E00077708000000F46B9A286278000000000000000000402E0000000000000
    00000402E00000000000073720027636F6D2E6562726576696174652E61756374696F6E2E6576656
    E742E41756374696F6E4D6F6465BD0C9E245C328B4F02000078720029636F6D2E656272657669617
    4652E636F6D6D6F6E2E41627374726163745479706553616665456E756D506D8C41B0144DB302000
    249000A696E744C69746572616C4C000D737472696E674C69746572616C71007E000378700000000
    374000A50524F44554354494F4E7371007E00097371007E000D3F4000000000000877080000000B0
    00000007874000A38335F41756374696F6E7371007E00097371007E000D3F4000000000000877080
    000000B000000007874000A34325F436F6D70616E797371007E00097371007E000D3F40000000000
    00877080000000B00000000787400103131315F426964436F6D706F6E656E747070)}
    2003-04-07 21:48:05.687 Tangosol Coherence 2.1/239 <Warning> (thread=CoherenceLogger): Asynchronous logging character limit exceeded; discarding 3 log messages (lines=17, chars=1416)

    Kris,
    First of all you should increase the value of logging-config/character-limit element in tangosol-coherence.xml to see the message entirely. The default setting is 4096 which is not enough to see your exception text.
    When you do that I believe you will see that the actual exception is java.lang.ClassNotFoundException indicating that the node that has the listener installed doesn't know about the class that is being put into the cache and could be easily fixed as shown here: http://www.tangosol.com/faq-coherence.jsp#classnotfound
    Please let me know if that doesn't help.
    Gene

  • Messaging server size limit

    Hallo
    As I see, I have common problem among the Sun Messaging Server administrators. I have whole system distributed on several virtual solaris machines and some days ago emerged message size problem. I noticed it in relation with incoming mail. When I create new user, he or she can't get larger mail then 300K. Sender get will known message:
    This message is larger than the current system limit or the recipient's mailbox is full. Create a shorter message body or remove attachments and try sending it again.
    <server.domain.com #5.3.4 smtp;552 5.3.4 a message size of 302 kilobytes exceeds the size limit of 300 kilobytes computed for this transaction>
    Interesting thin is, that this problem arised with no correlation with other actions. I noticed this problem with new users before, but i could successful manage it with different service packs. Now this method with new users, this method doesn't work! Old users normally recieve messages bigger then 300k, as before.
    I tried to set default setting blocklimit 2000 in imta.cnf, but I didn't succeed.
    I know, that size limit can be set on different places, but is there a simple way, to set sending and recieving message size unlimited?
    Messaging server version is:
    Sun Java(tm) System Messaging Server 7u2-7.02 64bit (built Apr 16 2009)*
    libimta.so 7u2-7.02 64bit (built 03:03:02, Apr 16 2009)*
    Using /opt/sun/comms/messaging64/config/imta.cnf (compiled)*
    SunOS mailstore 5.10 Generic_138888-01 sun4v sparc SUNW,SPARC-Enterprise-T5120*
    Regards
    Matej

    For the sake of correctness, the attribute name in LDAP is mailMsgMaxBlocks.
    I also stumbled upon this - the values like 300 blocks or 7000 blocks are set in (sample) service packages but are not advertised in Delegated Admin web-interface. When packages are assigned, these values are copied into each user's LDAP entry as well, and can not be seen or changed in web-interface.
    And then mail users get "weird" errors like:
    550 5.2.3 user limit of 7000 kilobytes on message size exceeded: [email protected]
    or
    550 5.2.3 user limit of 300 kilobytes on message size exceeded: [email protected]
    resulting in
    <[email protected]>... User unknown
    or
    552 5.3.4 a message size of 7003 kilobytes exceeds the size limit of 7000 kilobytes computed for this transaction
    or
    552 5.3.4 a message size of 302 kilobytes exceeds the size limit of 300 kilobytes computed for this transaction
    resulting in
    Service unavailable
    I guess there are other similar error messages, but these two are most common.
    I hope other people googling up the problem would get to this post too ;)
    One solution is to replace the predefined service packages with several of your own, i.e. ldapadd entries like these (fix the dc=domain,dc=com part to suit your deployment, and both cn parts if you rename them), and restart the DA webcontainer:
    dn: cn=Mail-Calendar - Unlimited,o=mailcalendaruser,o=cosTemplates,dc=domain,dc=com
    cn: Mail-Calendar - Unlimited
    daservicetype: calendar user
    daservicetype: mail user
    mailallowedserviceaccess: imaps:ALL$pops:ALL$+smtps:ALL$+https:ALL$+pop:ALL$+imap:ALL$+smtp:ALL$+http:ALL
    mailmsgmaxblocks: 20480
    mailmsgquota: -1
    mailquota: -1
    objectclass: top
    objectclass: LDAPsubentry
    objectclass: costemplate
    objectclass: extensibleobject
    dn: cn=Mail-Calendar - 100M,o=mailcalendaruser,o=cosTemplates,dc=domain,dc=com
    cn: Mail-Calendar - 100M
    daservicetype: calendar user
    daservicetype: mail user
    mailallowedserviceaccess: imaps:ALL$pops:ALL$+smtps:ALL$+https:ALL$+pop:ALL$+imap:ALL$+smtp:ALL$+http:ALL
    mailmsgmaxblocks: 20480
    mailmsgquota: 10000
    mailquota: 104857600
    objectclass: top
    objectclass: LDAPsubentry
    objectclass: costemplate
    objectclass: extensibleobject
    dn: cn=Mail-Calendar - 500M,o=mailcalendaruser,o=cosTemplates,dc=domain,dc=com
    cn: Mail-Calendar - 500M
    daservicetype: calendar user
    daservicetype: mail user
    mailallowedserviceaccess: imaps:ALL$pops:ALL$+smtps:ALL$+https:ALL$+pop:ALL$+imap:ALL$+smtp:ALL$+http:ALL
    mailmsgmaxblocks: 20480
    mailmsgquota: 10000
    mailquota: 524288000
    objectclass: top
    objectclass: LDAPsubentry
    objectclass: costemplate
    objectclass: extensibleobject
    See also limits in config files -
    * msg.conf (in bytes):
    service.http.maxmessagesize = 20480000
    service.http.maxpostsize = 20480000
    and
    * imta.cnf (in 1k blocks): <channel block definition> ... maxblocks 20000 blocklimit 20000 sourceblocklimit 20000
    i.e.:
    tcp_local smtp mx single_sys remotehost inner switchchannel identnonenumeric subdirs 20 maxjobs 2 pool SMTP_POOL maytlsserver maysaslserver saslswitchchannel tcp_auth missingrecipientpolicy 0 loopcheck slave_debug sourcespamfilter2optin virus destinationspamfilter2optin virus maxblocks 20000 blocklimit 20000 sourceblocklimit 20000 daemon outwardrelay.domain.com
    tcp_intranet smtp mx single_sys subdirs 20 dequeue_removeroute maxjobs 7 pool SMTP_POOL maytlsserver allowswitchchannel saslswitchchannel tcp_auth missingrecipientpolicy 4 maxblocks 20000 blocklimit 20000 sourceblocklimit 20000
    tcp_submit submit smtp mx single_sys mustsaslserver maytlsserver missingrecipientpolicy 4 slave_debug maxblocks 20000 blocklimit 20000 sourceblocklimit 20000
    tcp_auth smtp mx single_sys mustsaslserver missingrecipientpolicy 4 maxblocks 20000 blocklimit 20000 sourceblocklimit 20000
    If your deployment uses other SMTP components, like milters to check for viruses and spam, in/out relays separate from Sun Messaging, other mailbox servers, etc. make sure to use a common size limit.
    For sendmail relays sendmail.mc (m4) config source file it could mean lines like these:
    define(`SMTP_MAILER_MAX', `20480000')dnl
    define(`confMAX_MESSAGE_SIZE', `20480000')dnl
    HTH,
    //Jim Klimov
    PS: Great thanks to Shane Hjorth who originally helped me to figure all of this out! ;)

  • SGA Max Size limit?

    Hi,
    I have Fujitsu mid range Server with 16gb RAM and 64 bit Windows Server 2003,10g R2 db installed, current i have SGA size 4gb..
    What is SGA max size limit????
    One of my report runs in 24 seconds...*will this issue b solved increasing the SGA size upto 10,12 gb?*

    Yes,
    You can also go for a 10046 event tracing...
    ACCEPT sid PROMPT 'Enter SID: '
    ACCEPT serial PROMPT 'Enter SERIAL#: '
    ACCEPT action PROMPT 'Enter TRUE or FALSE: '
    EXEC sys.DBMS_SYSTEM.SET_SQL_TRACE_IN_SESSION(&sid,&serial,&action);
    prompt Trace &action for &sid,&serial
    exec DBMS_SYSTEM.SET_EV(10,20,10046,12,”);
    Then you can check your dump file and see whcih events are higher......
    For Eg. content could be like:
    =====================
    PARSING IN CURSOR #6 len=107 dep=1 uid=44 oct=6 lid=44 tim=1621758552415 hv=3988607735 ad='902c07a8'
    UPDATE rn_lu_lastname_loca set entr_loca_id_plz14 = translate(entr_loca_id_plz14,'_','-') where rowid = :b1
    END OF STMT
    PARSE #6:c=0,e=981,p=0,cr=0,cu=0,mis=1,r=0,dep=1,og=0,tim=1621758552403
    BINDS #6:
    bind 0: dty=1 mxl=32(18) mal=00 scl=00 pre=00 oacflg=13 oacfl2=1 size=32 offset=0
    bfp=10331d748 bln=32 avl=18 flg=09
    value="AAAHINAATAAAwTTABV"
    WAIT #6: nam='db file sequential read' ela= 12170 p1=6 p2=197843 p3=1
    WAIT #6: nam='db file sequential read' ela= 8051 p1=14 p2=261084 p3=1
    WAIT #6: nam='db file sequential read' ela= 7165 p1=19 p2=147722 p3=1
    WAIT #6: nam='db file sequential read' ela= 9604 p1=19 p2=133999 p3=1
    WAIT #6: nam='db file sequential read' ela= 6381 p1=19 p2=133801 p3=1
    EXEC #6:c=10000,e=45750,p=5,cr=1,cu=10,mis=0,r=1,dep=1,og=4,tim=1621758598343
    FETCH #5:c=0,e=357,p=0,cr=5,cu=0,mis=0,r=0,dep=1,og=4,tim=1621758598896
    EXEC #1:c=30000,e=116691,p=36,cr=35,cu=10,mis=0,r=1,dep=0,og=4,tim=1621758599043
    WAIT #1: nam='SQL*Net message to client' ela= 5 p1=1413697536 p2=1 p3=0
    WAIT #1: nam='SQL*Net message from client' ela= 2283 p1=1413697536 p2=1 p3=0
    Lines that start with WAIT
    len Length of SQL statement.
    dep Recursive depth of the cursor.
    uid Schema user id of parsing user.
    oct Oracle command type.
    lid Privilege user id.
    ela Elapsed time. 8i: in 1/1000th of a second, 9i: 1/1'000'000th of a second 
    tim Timestamp. Pre-Oracle9i, the times recorded by Oracle only have a resolution of 1/100th of a second (10mS). As of Oracle9i some times are available to microsecond accuracy (1/1,000,000th of a second). The timestamp can be used to determine times between points in the trace file. The value is the value in v$timer when the line was written. If there are TIMESTAMPS in the file you can use the difference between 'tim' values to determine an absolute time. 
    hv Hash id.
    ad SQLTEXT address (see v$sqlarea and v$sqltext).
    Lines that start with PARSE, EXEC or FETCH
    #n  n = number of cursor 
    c  cpu time 
    e  elapsed time 
    p  physical reads 
    cr  consistant reads 
    cu  current mode reads 
    mis miss in cache (?) 
    r  rows processed 
    dep recursive depth 
    og  optimizer goal 
    tim time  Content

  • How to increase cache size in firefox 6.02 mac os x

    I just want to be able to increase the cache size for larger files. As it is now large files are not caching and the limit of 1024 mb is very limiting.

    See if this reference helps:
    http://www.cyberciti.biz/faq/linux-unix-apache-increase-php-upload-limit/

  • Preposition size limit was exceeded

    Hi Everyone,
    I have a preposition job that's trying to copy a folder roughly 108GB in size (comprised of subfolders and files).  I have the job running nightly and it always terminates after several hours copying about 27GB of data saying, "Preposition size limit was exceeded."  I understand that we might be running into the reserved cache limit, and I have already set the preposition to use 100% of the cache.
    My question is, if this job terminates after copying 27GB of the 108GB, what happens the next time this job starts on the next scheduled iteration the following night.  Right now, it is set to "Files changed since last preposition."  Will the preposition resume where it last Status:Terminated or will it start over since it wasn't Status:Complete task (ie. I'll never copy the full 108 GB since it keeps starting over)?
    The versions of the two WAAS devices on each end:
    Cisco Wide Area Application Services Software (WAAS)
    Copyright (c) 1999-2012 by Cisco Systems, Inc.
    Cisco Wide Area Application Services (accelerator-k9) Software Release 4.4.7 (build b4 Apr 25 2012)
    Version: nme-wae-502-4.4.7.4
    Cisco Wide Area Application Services Software (WAAS)
    Copyright (c) 1999-2012 by Cisco Systems, Inc.
    Cisco Wide Area Application Services (accelerator-k9) Software Release 4.4.7 (build b4 Apr 25 2012)
    Version: oe512-4.4.7.4
    Thanks,
    JP Babiera

    Hi JP,
    I was digging into details and i got this in user guide:
    Step 4 Ensure that the Status column shows Completed.
    If this column shows a failure, look in the Reason column for an explanation that can help you troubleshoot why the preposition task failed. After resolving the issue, you can schedule the preposition task to run again now, or wait until the scheduled start time and check the status again later
    So looking at this explanation it does seem that prepositioning should resume where it actually left off.
    For more details please visit the below link and section: Checking the Preposition Status
    http://www.cisco.com/en/US/docs/app_ntwk_services/waas/waas/v413/configuration/guide/filesvr.html#wp1043724
    Regards,
    Kanwal

  • Default cache size reduced to 350 MB (from 1024 MB)

    Why was the default size of the automatic cache management reduced from 1024 MB to 350 MB ?
    How does that improve Firefox ?

    That is a consequence of this bug:
    *[https://bugzilla.mozilla.org/show_bug.cgi?id=709297 bug 709297] - Limit disk cache size until/unless "clear recent data" can be done async
    <i>(please do not comment in bug reports: [https://bugzilla.mozilla.org/page.cgi?id=etiquette.html])</i>

  • ResourceBundle cache&size

    Hello there,
    If there is needed the content of the ResourceBundle, which have been already needed, it is loaded just from the cache. The ResourceBundle is loaded once for application not for each client, is it right? Is the cache cleaned time to time, or is it there all forever? Is it better to have more smaller files, one huge or doesn't it matter?

    Proxy 4.0 also has a maximum of 32GB as the cache capacity.
    The cache capacity is used to derive the sections under the cache directory.
    For a 32GB cache capacity there will be a maximum of 256 sections and under each section there are 64 subsections.
    256 sections are just directories created under the cache directory and each of them will have 64 subdirectories.
    This hierarchy will help to store the number of cache files and not to limit the size of all the cached files put together to 32GB.
    So if you have a disk size of say 72GB its ok.
    But you need to configure it as 32GB to get the sections/subsections hierarchy and under these the cache files will be stored.
    There is no size limit for each section/subsection.
    As the files are being stored it occupies the space in the 72GB hard disk.
    As per the caching algorithm the max cache capacity/size is 32GB to cache say 70 million urls.
    So if the disk size is more its ok. But this is the maximum that can be configured to get 256 sections.
    We are agree its bit confusing as we say that its cache capacity.
    One thing to understand is 32GB is the max to derive 256 sections for the algorithm and
    the sections/subsections are not limited for a particular disk size.
    So if your disk size is more its ok.

  • Setup failover for a distributed cache

    Hello,
    For our production setup we will have 4 app servers one clone per each app server. so there will be 4 clones to a cluster. And we will have 2 jvms for our distributed cache - one being a failover, both of those will be in cluster.
    How would i configure the failover for the distributed cache?
    Thanks

    user644269 wrote:
    Right - so each of the near cache schemes defined would need to have the back map high-units set to where it could take on 100% of data.Specifically the near-scheme/back-scheme/distributed-scheme/backing-map-scheme/local-scheme/high-units value (take a look at the [Cache Configuration Elements|http://coherence.oracle.com/display/COH34UG/Cache+Configuration+Elements] ).
    There are two options:
    1) No Expiry -- In this case you would have to size the storage enabled JVMs to that an individual JVM could store all of the data.
    or
    2) Expiry -- In this case you would set the high-units a value that you determine. If you want it to store all the data then it needs to be set higher than the total number of objects that you will store in the cache at any given time or you can set it lower with the understanding that once that high-units is reached Coherence will evict some data from the cluster (i.e. remove it from the "cluster memory").
    user644269 wrote:
    Other than that - there is not configuration needed to ensure that these JVM's act as a failover in the event one goes down.Correct, data fault tolerance is on by default (set to one level of redundancy).
    :Rob:
    Coherence Team

  • Cache size question...

    howdy,
    just a quick one.
    i've noticed in the new firefox (1.5) there is an option in preferences to limit the cache size. however, there seems to be no option in safari.
    :: is there a method to set the cache size in safari? ::
    any info is greatly appreciated.
    cheers...
    ryan

    ryan,
    I am Terminally challenged, so I use CLIX and it has both Caches Off/Caches On for Safari.
    I do not see a command for setting a cache limit.
    Caches Off (rm -fr ~/Library/Caches/Safari;ln -s /dev/null ~/Library/Caches/Safari)
    Caches On (rm ~/Library/Caches/Safari)
    ;~)

  • Distributed Cache Errors in WFE

    Unexpected error occurred in method 'Put' , usage 'Distributed Logon Token Cache' - Exception 'Microsoft.ApplicationServer.Caching.DataCacheException: ErrorCode<ERRCA0016>:SubStatus<ES0001>:The
    connection was terminated, possibly due to server or network problems or serialized Object size is greater than MaxBufferSize on server. Result of the request is unknown. ---> System.TimeoutException: The socket was aborted because an asynchronous receive
    from the socket did not complete within the allotted timeout of 00:02:00. The time allotted to this operation may have been a portion of a longer timeout. ---> System.IO.IOException: The read operation failed, see inner exception. ---> System.TimeoutException:
    The socket was aborted because an asynchronous receive from the socket did not complete within the allotted timeout of 00:02:00. The time allotted to this operation may have been a portion of a longer timeout. ---> System.Net.Sockets.SocketException: The
    I/O operation has been aborted because of either a thread exit or an application request   
     at System.ServiceModel.Channels.SocketConnection.HandleReceiveAsyncCompleted()   
     at System.ServiceModel.Channels.SocketConnection.OnReceiveAsync(Object sender, SocketAsyncEventArgs eventArgs)     -
    -- End of inner exception stack trace ---   
     at System.Runtime.AsyncResult.End[TAsyncResult](IAsyncResult result)   
     at System.ServiceModel.Channels.ConnectionStream.ReadAsyncResult.End(IAsyncResult result)   
     at System.Net.FixedSizeReader.ReadCallback(IAsyncResult transportResult)     -
    -- End of inner exception stack trace ---   
     at System.Net.Security.NegotiateStream.EndRead(IAsyncResult asyncResult)   
     at System.ServiceModel.Channels.StreamConnection.EndRead()     -
    -- End of inner exception stack trace ---   
     at System.Runtime.AsyncResult.End[TAsyncResult](IAsyncResult result)   
     at System.ServiceModel.Channels.TransportDuplexSessionChannel.EndReceive(IAsyncResult result)   
     at Microsoft.ApplicationServer.Caching.WcfClientChannel.CompleteProcessing(IAsyncResult result)     -
    -- End of inner exception stack trace ---   
     at Microsoft.ApplicationServer.Caching.DataCache.ThrowException(ResponseBody respBody, RequestBody reqBody)   
     at Microsoft.ApplicationServer.Caching.DataCache.InternalPut(String key, Object value, DataCacheItemVersion oldVersion, TimeSpan timeout, DataCacheTag[] tags, String region,
    IMonitoringListener listener)   
     at Microsoft.ApplicationServer.Caching.DataCache.<>c__DisplayClass25.<Put>b__24()   
     at Microsoft.ApplicationServer.Caching.DataCache.Put(String key, Object value, TimeSpan timeout)   
     at Microsoft.SharePoint.DistributedCaching.SPDistributedCache.Put(String key, Object value)'.

    Hi Amol,
    check those links, contains similar issues and resolution
    https://habaneroconsulting.com/insights/sharepoint-2013-distributed-cache-bug#.VJAjsyuUdP0
    http://blogs.msdn.com/b/sambetts/archive/2014/05/28/troubleshooting-appfabric-timeouts-on-sharepoint.aspx
    http://sharepointchips.com/distributed-cache-errors-in-the-uls-log/
    https://www.dmcinfo.com/latest-thinking/blog/id/8657/fix-sharepoint-2013-distributed-cache-timeouts
    https://social.technet.microsoft.com/Forums/sharepoint/en-US/2fad2277-8f5f-4323-8c18-621ae4bfe11a/refresh-the-sp2013-distributed-cache-services-logon-token-cache?forum=sharepointgeneral
    Kind Regards,
    John Naguib
    Technical Consultant/Architect
    MCITP, MCPD, MCTS, MCT, TOGAF 9 Foundation
    Please remember to mark your question as answered if this solves your problem

Maybe you are looking for

  • How do I know if my MBP is 64 bit compatable ?

    Hello all, I purchased my Intel core2duo 2.16 Ghz MBP in March 2007, it is still running Tiger. Now that Snow Leopard is due out in September, I would like to upgrade to that version, either end of year, or January. I saw that it will run as 64bit, i

  • Photos present in I Photo but not appearing in folders on hard disk

    OK so I have been warned about the idiosynchronies of I Photo but this? ...Help! I am able to successfully load from the camera into IPhoto and I am able to open the pics in Photoshop via various roundabout methods yet the new folders that appear in

  • Iweb can't find mobileme gallery

    hello, i have published 2 mobileme gallerys, but iweb wingit can't see it. what can be wrong?

  • Can u please give me two critical issues in app

    < MODERATOR:  Message locked.  Please read the [Rules of Engagement|https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/rulesofEngagement] before posting next time. > can u please tell me two critical issues in app and explain me how i can solve

  • "Importing" table-comments into an existinting server model with a text file

    Hello i have redesigned a database with the "capture design" tool - server model and server model diagramm is ok. Now, i want add many table- and column-comments into the server model. I wrote these comments with the "comment on table..." and "commen