Different distributed caches within the cluster

Hi,
i've three machines n1 , n2 and n3 respectively that host tangosol. 2 of them act as the primary distributed cache and the third one acts as the secondary cache. i also have weblogic running on n1 and based on some requests pumps data on to the distributed cache on n1 and n2. i've a listener configured on n1 and n2 and on the entry deleted event i would like to populate tangosol distributed service running on n3. all the 3 nodes are within the same cluster.
i would like to ensure that the data directly coming from weblogic should only be distributed across n1 and n2 and NOT n3. for e.g. i do not start an instance of tangosol on node n3. and an object gets pruned from either n1 or n2. so ideally i should get a storage not configured exception which does not happen.
The point is the moment is say CacheFactory.getCache("Dist:n3") in the cache listener, tangosol does populate the secondary cache by creating an instance of Dist:n3 on either n1 or n2 depending from where the object has been pruned.
from my understanding i dont think we can have a config file on n1 and n2 that does not have a scheme for n3. i tried doing that and got an illegalstate exception.
my next step was to define the Dist:n3 scheme on n1 and n2 with local storage false and have a similar config file on n3 with local-storage for Dist:n3 as true and local storage for the primary cache as false.
can i configure local-storage specific to a cache rather than to a node.
i also have an EJB deployed on weblogic that also entertains a getData request. i.e. this ejb will also check the primary cache and the secondary cache for data. i would have the statement
NamedCahe n3 = CacheFactory.getCache("n3") in the bean as well.

Hi Jigar,
i've three machines n1 , n2 and n3 respectively that
host tangosol. 2 of them act as the primary
distributed cache and the third one acts as the
secondary cache.First, I am curious as to the requirements that drive this configuration setup.
i would like to ensure that the data directly coming
from weblogic should only be distributed across n1
and n2 and NOT n3. for e.g. i do not start an
instance of tangosol on node n3. and an object gets
pruned from either n1 or n2. so ideally i should get
a storage not configured exception which does not
happen.
The point is the moment is say
CacheFactory.getCache("Dist:n3") in the cache
listener, tangosol does populate the secondary cache
by creating an instance of Dist:n3 on either n1 or n2
depending from where the object has been pruned.
from my understanding i dont think we can have a
config file on n1 and n2 that does not have a scheme
for n3. i tried doing that and got an illegalstate
exception.
my next step was to define the Dist:n3 scheme on n1
and n2 with local storage false and have a similar
config file on n3 with local-storage for Dist:n3 as
true and local storage for the primary cache as
false.
can i configure local-storage specific to a cache
rather than to a node.
i also have an EJB deployed on weblogic that also
entertains a getData request. i.e. this ejb will also
check the primary cache and the secondary cache for
data. i would have the statement
NamedCahe n3 = CacheFactory.getCache("n3") in the
bean as well.In this scenario, I would recommend having the "primary" and "secondary" caches on different cache services (i.e. distributed-scheme/service-name). Then you can configure local storage on a service by service basis (i.e. distributed-scheme/local-storage).
Later,
Rob Misek
Tangosol, Inc.

Similar Messages

  • How is index data managed within the cluster?

    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    Peter

    user11279467 wrote:
    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    PeterHi Peter,
    up to 3.4.x indexes were supported only on partitioned topology, and yes, indexes on each node contain only data corresponding to entries which have their primary copy on that node (referred to as local entries from now on).
    The indexes are not stored in a cache but they are managed as part of the information Coherence maintains about the backing map (the map which contains local entries).
    The indexes are made of 2 parts:
    - forward index: this is a mapping from the cache key in an internal representation (backing map keys from now on) to the value extracted with the extractor used to create the index from the entry belonging to the cache key (extracted value from now on)
    - reverse index (aka. reverse map): this is a mapping from the extracted value to a set of backing map keys of entries from which the reverse index key was extracted with the extractor used to create the index
    The index can be sorted which means that the reverse index will be a SortedMap (so sorting order is the ordering of extracted values)
    Let's see an example. Assuming you have the following cache content within a particular storage-enabled node :
    A --> Object (getX=5 ,...)
    B --> Object (getA=3 ,...)
    C --> Object (getA=3 ,...)
    D --> Object (getA=4 ,...)
    The forward index in that node will contain:
    Binary(A) --> 5
    Binary(B) --> 3
    Binary(C) --> 3
    Binary(D) --> 4
    The reverse index will contain:
    3 --> { Binary(B), Binary(C) }
    4 --> { Binary(D) }
    5 --> { Binary(A) }
    Where Binary(...) means the internal (binary) representation of the ... object.
    If the index is not ordered, then the order of iteration on entries in the reverse index are not deterministic.
    If the index is ordered, then the iteration of entries in the reverse index will be as shown above. Also, you can cast the reverse index to SortedMap so you have the very useful headMap and tailMap and firstKey and lastKey methods.
    Hope this helps.
    Best regards,
    Robert

  • I subscribed to iTunes Match and would like to sync my music on multiples devices that have different Apple IDs within the same iCloud family.

    I subscribed to iTunes Match and would like to sync my music on multiples devices that have different Apple IDs within the same iCloud family.
    Can someone indicate the way to proceed?
    Thanks.

    Hi,
    You can't use family sharing or multiple Apple IDs with iTunes match.
    ITunes match is only attached to Apple id that you used to subscribe.
    Jim

  • RetrieveJXM metrics from within the cluster?

    I would like to know if it is possible to retrieve the metrics the JMX adapter exposes from within the cluster e.g. by running a thread on the invocation service that would query the underlying data structures to get the metrics out.

    You could still snapshot the stats if you wanted to but you would have to do it using queries into the local JMS server running in the JVM rather than using Coherence API calls.
    Or, depending on what you want to do with the stats, you could look at the JMX reporter that will log stats to a file every so often or you could look at one of the third-party monitoring tolls that will do this.
    JK

  • Different  disaggregation rules within the same key figure

    Hello All,
    We need to apply different disaggregation rules within the different levels of the same key figure.
    As an Example:
    Key Figure: SALES VOLUME
    Level1: Market
    Level2: Segment
    Level3: Model
    We want to define such rules:
    1- Disaggregate from Market Level to Segment Level using Historical Data
    2- Disaggregate from Segment Level to Model Level using APODPNDANT (a key figure in which the input is supplied from business)
    Does anybody know if this (defining two different rules within the different levels of the same key figure) is posssible?
    Thanks in advance...

    Hi Ergul
    Ideally its better to have APODPDANT to be the disaggregation KF at all levels
    you can choose to manipulate the APODPDANT instead by making it an Editable Kf in the Planning book
    To start with all levels can be based on the proportional KF calculation
    By Loading the selections at the Segment to model level you can manually enter the values based on business input
    you might need some macors to fix the totals and maybe to make sure the Sales volume is redisaggregated
    i think its something that can be solved by a god business process and simple functionality in the way you see the data

  • RMAN to connect to all nodes within the cluster to parallelize backups

    Hi,
    I wish to configure my RMAN catalog to backup my Production RAC backups. I wish to parallelize my backups to backup database files of different nodes.
    Can anyone help me how can it be done?
    Regards,
    Srini

    Hi,
    May be you are right. I see my RAC cluster databases grow very fast and some nodes are not really performing well resources wise. So, i may need to parallalize the backups across the nodes.
    However, i have another idea of finding the best node and take backup on it. checking with my system admins at this point.
    Is there any way RMAN can know if the node went down so that it needs to backup the database on a different node instead of the first node it attempted?
    May be this is a different question.
    Regards,
    srini

  • Can the node's UID be different when re-joining the cluster?

    Let's say that we have a split brain scenario and when both islands meet again one of them is "marked as invalid" and is restarted. Upon cluster restart will the restarted nodes get new UID values?
    Thanks,
    -- Gato

    Gato,
    Absolutely! The UID is the member's identity and if the cluster service restarts (within a given Java process), it will be always assigned a new and unique UID.
    Regards,
    Gene

  • SBO - How to use two different Stock Account within the same Item Group

    Hi,
    I'm currently stuck within one of my implementation.
    I'm deploying the same solution worldwide which worked pretty well until now. All the items are connected to an item group, with a unique GL account, supposed to book at the same time the good receipt and the good issue (Balance-Sheet Account = 14xxxxx).
    My new unit is willing to use two different GL accounts when performing its stock entry (let say 14xxx10) and another GL account for the good issue, let say 14xxx20.
    Do you have any clue so i can please them and keep using at the same time the item group ?
    Thanking you in advance
    Stephane

    Hi,
    In fact, we use the GL account as the HQ Account (Reporting Account) while the "Export Code" will be used to store the local GL Account (greek account).
    As such, we can, somehow handle two different chart of accounts.
    In my reporting, i'm only using a single GL account for stock movements, while my greek counterparts want to handle two different accounts.
    Threfore, i have only two solution since there is normally a one to one relationship between a HQ account and a local account that are :
    1 - Duplicate the HQ accounts
    In that case, i do respect anytime the one to one relationship. I've got one HQ account for one local account. But how can i put this since i can only handle one single stock account.
    2 - I keep my single stock account like this but i won't be managing the possibility to get two different Export Codes.
    Regards
    Stephane

  • How to create a different browser session within the same profile

    Internet Explorer 8 lets you create a new session within a new browser window but how can I do the same with Firefox? I know if you switch user profiles then you can have two browsers in different sessions but I would like to have two sessions using the same user profile.

    Sorry, that's not possible with Firefox. Firefox uses a single executable instance to run all windows and tabs, where IE starts another process for each window. Chrome uses a separate process for each tab (or window), which seems real nice until it sucks the life out of an older processor with the "overhead" of all those separate processes.
    Depending upon what you want a "different session" for, there are 3 extensions that allow for multiple cookie sessions which allow for multiple, simultaneous logins to the same server. That works to allow the user to log into multiple webmail accounts at the same service provider. <br />
    Multifox: <br />
    http://br.mozdev.org/multifox/ <br />
    Cookie Swap extension: <br />
    https://addons.mozilla.org/firefox/3255/ <br />
    Cookie Pie extension: <br />
    http://www.nektra.com/oss/firefox/extensions/cookiepie/

  • Different Page Sizes within the same document - Printing

    I am having trouble printing a document which contains pages of different sizes (using Adobe Professional 7.0):
    i.e. The first 10 pages are 8.5" x 11", the next few pages are 11" x 17" and the last few pages are 8.5" x 11"
    The printer only spits everything out as 8.5" x 11" and the larger sheets are just cut off.
    In the "Print" menu that comes up when you hit FILE->PRINT it suggests that everything is alright... I can scroll through the small "page preview" and it shows some pages 8.5x11 and some pages 11x17... but when it actually prints, it doesn't do that. Is it the printer that is the problem perhaps?
    (The printer is a Lexmark x945e)
    Any insight would be greatly appreciated.
    ~Paul

    If the goal is to make a flap, then you can do this in InDesign. A spread can be any contiguous number of pages, so if your document is generally speaking double-page spreads, you can have a single spread that is composed of three contiguous pages.
    To do this, in the Pages palette, select a double-page spread you want to expand. Go to the flyout menu and deselect "Allow selected spread to shuffle." Then drag new page to the left of the left page or to the right of the right page (depending on where your flap is).
    Then you're all set.
    Here's a screenshot of the Pages panel:
    In this case, I've created a flap to the right of page 3 (numbered page a) and its reverse to the left of page 4 (numbered page b).

  • Is it possible to create different font styles within the same text field?

    The form I am creating is getting really complex because when different fields require different font treatments (bold, italic, different font size, different font) I have to create a seperate text box and get it to align with the others. This wouldn't be an issue in most cases but this form is really detailed and has many text boxes. Before I stuggle too much longer on it, I thought I would check and see if there is a way to make them all in the same text field. If not, perhaps I am trying to use Acrobat for something it is not meant for. Any recommendations on other programs I could use? Thanks

    Thank you try76. I'm not sure if I did it correctly. As soon as I tick the Rich Text Formatting box it automatically changes the appearance of the text in the field (Helvetica Light to something like Helvetica Heavy) I can't seem to get the font back to Helvetica light or change the font size while the Rich Text Formatting box is ticked.

  • How to read data with different XML schemas within the single connection?

    I have Oracle 11g database
    I access it through jdbc:oracle:thin, version 11.2.0.3, same as xdb.
    I have several tables, each has one XMLType column, all schema-based.
    There are three different XML schemata registered in the DB
    I may need to read the XML data from several tables.
    If all the XMLTypes have the same XML schema ,there is no problem,
    If the schemata are different, the second read throws BindXMLException.
    If I reset the connection between the reads of the XMLType column with different schemata, it works.
    The question is: how can I configure the driver, or the connection to be able to read the data with different XML schemata without resetting the connection (which is expensive).
    The code to get the XMLType data is textbook implementation:
    1   ResultSet resultSet = statement.executeQuery( sql ) ;
    2   String result = null ;
    3    while(resultSet.next()) {
    4   SQLXML sqlxml = resultSet.getSQLXML(1) ;
    5   result = sqlxml.getString() ;
    6   sqlxml.free();
    7   }
    8   resultSet.close();
    9    return result ;

    It turns out, that I needed to serialize the XML on the server and read it as Blob. Like this:
    1    final Statement statement = connection.createStatement() ;
    2    final String sql = String.format("select xmlserialize(content xml_content_column as blob encoding 'UTF-8') from %s where key='%s'", table, key ) ;
    3   ResultSet resultSet = statement.executeQuery( sql ) ;
    4   String result = null ;
    5    while(resultSet.next()) {
    6   Blob blob = resultSet.getBlob( 1 );
    7   InputStream inputStream = blob.getBinaryStream();
    8   result = new Scanner( inputStream ).useDelimiter( "\\A" ).next();
    9   inputStream.close();
    10   blob.free();
    11   }
    12   resultSet.close();
    13   statement.close();
    14
    15   System.out.println( result );
    16    return result ;
    17
    Then it works. Still, can't get it work with XMLType in resultset.On the client unwrapping XML blows up when trying to switch to different XML schema. JDBC/XDB problem?

  • HT4352 Will homesharing share the content of different iTunes accounts within the home? ie wife has account and purchased songs. I have different account and want to listen to them on my Apple TV.

    Have 3 family members with three accounts. Homesharing is supposed to be set up all computers with the same (one) account. Then, will the content ONLY of the set account be shared? or will ALL content of those who sync with those computers be shared?
    Products: (2) iPhones 4s, (2)iPad2, Apple TV, iPod, Dell PC with Windows 7.

    There's an issue currently that a lot of us are experiencing in that we all get the error message "iTunes is currently unavailable; please try later"
    It's an issue with Apple and we're awaiting an update re this
    I've posted a topic in regards to this after having a long conversation with Apple support this morning and will be receiving a call from their tech team tomorrow to run through it further
    As it stands iTunes cannot be accessed via ATV's and nor can it be streamed via AirPlay from iTunes to the ATV
    However I can access all films/iCloud content via my Mac Mini with no issues!

  • Can I create different Coherence nodes in the same cluster with defferent?

    Can I create different Coherence nodes in the same cluster with defferent cache-config.xml file ?
    Can a cache be distributed in these deffirent nodes?

    Yes. You can create different Coherence nodes in the same cluster with defferent cache-config.xml files as long as you use the same tangosol-coherence.xml file and the same tangosol-coherence-override.xml file. But you cannot store the cache data in the different nodes (started with different cache-config file). In other word, a node only create caches in their own's modes which are started with the same cache-config.xml file.
    See the following demo:
    I start a cache server using the cache config file examples-cache-server.xml. Then I start a storage-disabled cache console (cache client) using the cache config file coherence-cache-config.xml. Both of them using the same tangosol-coherence.xml file and the same tangosol-coherence-override.xml file.
    The cache server uses a cache service PartitionedPofCache. But the client side is using the Distributedcache service. The cluster address is same 224.3.5.2.
    The cluster name is also samme. They know each other.
    D:\coherence\lib>D:\examples\java\bin\run-cache-server.cmd
    D:\coherence\lib>D:\examples\java\bin\run-cache-server.cmd
    The system cannot find the file D:\coherence.
    The system cannot find the file C:\Oracle\Middleware\jdk160_11.
    2009-12-22 12:09:31.400/4.987 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational configurat
    ion from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence.xml"
    2009-12-22 12:09:31.450/5.037 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational overrides
    from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence-override-dev.xml"
    2009-12-22 12:09:31.470/5.057 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/tangosol-coherence-override.xml" is not specified
    2009-12-22 12:09:31.540/5.127 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/custom-mbeans.xml" is not specified
    Oracle Coherence Version 3.5.2/463
    Grid Edition: Development mode
    Copyright (c) 2000, 2009, Oracle and/or its affiliates. All rights reserved.
    2009-12-22 12:09:33.864/7.451 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=n/a): Loaded cache configuration
    from "file:/D:/examples/java/resource/config/examples-cache-config.xml"
    2009-12-22 12:09:39.983/13.570 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Service Cluster joined t
    he cluster with senior service member n/a
    2009-12-22 12:09:43.187/16.774 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Created a new cluster
    "cluster:0xD3FB" with Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Locatio
    n=process:144, Role=CoherenceServer, Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1) UID=0xC0A8085000
    000125B75D888C60501F98
    2009-12-22 12:09:43.508/17.095 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:Management, member=1): Service Mana
    gement joined the cluster with senior service member 1
    2009-12-22 12:09:46.582/20.169 Oracle Coherence GE 3.5.2/463 <D5> (thread=DistributedCache:PartitionedPofCache, member=1
    ): Service PartitionedPofCache joined the cluster with senior service member 1
    2009-12-22 12:09:46.672/20.259 Oracle Coherence GE 3.5.2/463 <Info> (thread=DistributedCache:PartitionedPofCache, member
    =1): Loading POF configuration from resource "file:/D:/examples/java/resource/config/examples-pof-config.xml"
    2009-12-22 12:09:46.702/20.289 Oracle Coherence GE 3.5.2/463 <Info> (thread=DistributedCache:PartitionedPofCache, member
    =1): Loading POF configuration from resource "jar:file:/D:/coherence/lib/coherence.jar!/coherence-pof-config.xml"
    2009-12-22 12:09:47.734/21.321 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=1): Started DefaultCacheServer.
    SafeCluster: Name=cluster:0xD3FB
    Group{Address=224.3.5.2, Port=35463, TTL=4}
    MasterMemberSet
      ThisMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process
    :144, Role=CoherenceServer)
      OldestMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=proce
    ss:144, Role=CoherenceServer)
      ActualMemberSet=MemberSet(Size=1, BitSetCount=2
        Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Rol
    e=CoherenceServer)
      RecycleMillis=120000
      RecycleSet=MemberSet(Size=0, BitSetCount=0
    Services
      TcpRing{TcpSocketAccepter{State=STATE_OPEN, ServerSocket=192.168.8.80:8088}, Connections=[]}
      ClusterService{Name=Cluster, State=(SERVICE_STARTED, STATE_JOINED), Id=0, Version=3.5, OldestMemberId=1}
      InvocationService{Name=Management, State=(SERVICE_STARTED), Id=1, Version=3.1, OldestMemberId=1}
      DistributedCache{Name=PartitionedPofCache, State=(SERVICE_STARTED), LocalStorage=enabled, PartitionCount=257, BackupCo
    unt=1, AssignedPartitions=257, BackupPartitions=0}
    2009-12-22 12:12:29.737/183.324 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member(Id=2, Timestamp=20
    09-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, Role=CoherenceConsole) joined
    Cluster with senior member 1
    2009-12-22 12:12:30.498/184.085 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member 2 joined Service M
    anagement with senior member 1
    2009-12-22 12:12:31.860/185.447 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): TcpRing: connecting to me
    mber 2 using TcpSocket{State=STATE_OPEN, Socket=Socket[addr=/192.168.8.80,port=8089,localport=2463]}
    2009-12-22 12:12:51.338/204.925 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member 2 joined Service D
    istributedCache with senior member 2The following command starts a cache client.
    D:\coherence\bin>coherence.cmd
    D:\coherence\bin>coherence.cmd
    ** Starting storage disabled console **
    java version "1.6.0_11"
    Java(TM) SE Runtime Environment (build 1.6.0_11-b03)
    Java HotSpot(TM) Server VM (build 11.0-b16, mixed mode)
    2009-12-22 12:12:21.054/3.425 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational configurat
    ion from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence.xml"
    2009-12-22 12:12:21.355/3.726 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational overrides
    from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence-override-dev.xml"
    2009-12-22 12:12:21.365/3.736 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/tangosol-coherence-override.xml" is not specified
    2009-12-22 12:12:21.415/3.786 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/custom-mbeans.xml" is not specified
    Oracle Coherence Version 3.5.2/463
    Grid Edition: Development mode
    Copyright (c) 2000, 2009, Oracle and/or its affiliates. All rights reserved.
    2009-12-22 12:12:29.316/11.687 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Service Cluster joined t
    he cluster with senior service member n/a
    2009-12-22 12:12:29.356/11.727 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Failed to satisfy the
    variance: allowed=16, actual=20
    2009-12-22 12:12:29.356/11.727 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Increasing allowable v
    ariance to 17
    2009-12-22 12:12:29.807/12.178 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): This Member(Id=2, Time
    stamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, Role=CoherenceConsole,
    Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1) joined cluster "cluster:0xD3FB" with senior Member(I
    d=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Role=CoherenceS
    erver, Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1)
    2009-12-22 12:12:29.977/12.348 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Member 1 joined Service
    Management with senior member 1
    2009-12-22 12:12:29.977/12.348 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Member 1 joined Service
    PartitionedPofCache with senior member 1
    2009-12-22 12:12:30.578/12.949 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:Management, member=2): Service Mana
    gement joined the cluster with senior service member 1
    SafeCluster: Name=cluster:0xD3FB
    Group{Address=224.3.5.2, Port=35463, TTL=4}
    MasterMemberSet
      ThisMember=Member(Id=2, Timestamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=proces
    s:1188, Role=CoherenceConsole)
      OldestMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=proce
    ss:144, Role=CoherenceServer)
      ActualMemberSet=MemberSet(Size=2, BitSetCount=2
        Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Rol
    e=CoherenceServer)
        Member(Id=2, Timestamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, R
    ole=CoherenceConsole)
      RecycleMillis=120000
      RecycleSet=MemberSet(Size=0, BitSetCount=0
    Services
      TcpRing{TcpSocketAccepter{State=STATE_OPEN, ServerSocket=192.168.8.80:8089}, Connections=[]}
      ClusterService{Name=Cluster, State=(SERVICE_STARTED, STATE_JOINED), Id=0, Version=3.5, OldestMemberId=1}
      InvocationService{Name=Management, State=(SERVICE_STARTED), Id=1, Version=3.1, OldestMemberId=1}
    Map (?):
    2009-12-22 12:12:49.505/31.906 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=2): Loaded cache configuration
    from "jar:file:/D:/coherence/lib/coherence.jar!/coherence-cache-config.xml"
    2009-12-22 12:12:51.358/33.729 Oracle Coherence GE 3.5.2/463 <D5> (thread=DistributedCache, member=2): Service Distribut
    edCache joined the cluster with senior service member 2
    <distributed-scheme>
      <!--
      To use POF serialization for this partitioned service,
      uncomment the following section
      <serializer>
      <class-
      name>com.tangosol.io.pof.ConfigurablePofContext</class-
      name>
      </serializer>
      -->
      <scheme-name>example-distributed</scheme-name>
      <service-name>DistributedCache</service-name>
      <backing-map-scheme>
        <local-scheme>
          <scheme-ref>example-binary-backing-map</scheme-ref>
        </local-scheme>
      </backing-map-scheme>
      <autostart>true</autostart>
    </distributed-scheme>But when I try to store data into cache from the client side, it report error message: it's staorage-disabled. It shows that this cache console cannot store the data in the existing cache server because then using different cache config files.
    Map (ca3): cache ca2
    <distributed-scheme>
      <!--
      To use POF serialization for this partitioned service,
      uncomment the following section
      <serializer>
      <class-
      name>com.tangosol.io.pof.ConfigurablePofContext</class-
      name>
      </serializer>
      -->
      <scheme-name>example-distributed</scheme-name>
      <service-name>DistributedCache</service-name>
      <backing-map-scheme>
        <local-scheme>
          <scheme-ref>example-binary-backing-map</scheme-ref>
        </local-scheme>
      </backing-map-scheme>
      <autostart>true</autostart>
    </distributed-scheme>
    Map (ca2): put 1 one
    2009-12-22 14:00:04.999/6467.370 Oracle Coherence GE 3.5.2/463 <Error> (thread=main, member=2):
    java.lang.RuntimeException: Storage is not configured
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.onMissing
    Storage(DistributedCache.CDB:9)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.ensureReq
    uestTarget(DistributedCache.CDB:34)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.put(Distr
    ibutedCache.CDB:22)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.put(Distr
    ibutedCache.CDB:1)
            at com.tangosol.util.ConverterCollections$ConverterMap.put(ConverterCollections.java:1541)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$ViewMap.put(Distrib
    utedCache.CDB:1)
            at com.tangosol.coherence.component.util.SafeNamedCache.put(SafeNamedCache.CDB:1)
            at com.tangosol.coherence.component.application.console.Coherence.processCommand(Coherence.CDB:581)
            at com.tangosol.coherence.component.application.console.Coherence.run(Coherence.CDB:39)
            at com.tangosol.coherence.component.application.console.Coherence.main(Coherence.CDB:3)
            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
            at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
            at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
            at java.lang.reflect.Method.invoke(Method.java:597)
            at com.tangosol.net.CacheFactory.main(CacheFactory.java:1400)

  • The Security Token Service is not available error on dedicated Distributed Cache server

    I have an error on a dedicated Distributed Cache server stating that the Security Token Service is not available.  I was under the impression that when Distributed Cache was running on a dedicated server that the only service that should be enabled
    is Distributed Cache. 
    The token service is working as expected on all other servers but this one.  Does this service need to be started or should I just ignore this error message?
    Jennifer Knight (MCITP, MCPD)

    as per my little experience with 2013, if STS is working fine on Web server then I am sure that sharepoint will be fine...Distributed cache stores the ST issued by STS. NO need to worry about this error.
    Login
    Token Cache
    DistributedLogonTokenCache
    This
    cache stores the security token issued by a Secure Token Service for use by any web server in the server farm. Any web server that receives a request for resources can access the security token from the cache, authenticate the user, and provide access to the
    resources requested.
    I would say check the ULS logs and get more details about the error why its not working on that server.
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

Maybe you are looking for

  • HOW TO FIND UNUSED initialization blocks  IN OBIEE RPD

    Hi Gurus, I have a task to remove unused initialization blocks and subject areas from RPD. how to find them ?how to check whether any reports are using these initialization blocks? Please reply me. I need to complete this task.Pls help me Regards, si

  • Goods issue availability check

    Dear All, we are facing one problem in SAP. My scenerio is i am procurring one Raw Material say Eg - RM1 - 100 kg Batch No - ABC01 on 12.08.2010 so my stock will be availble only 12.08.2010 only, but when i am doing goods issue to order / cost centre

  • Perf.Tuning

    Hi everybody, Iam new to this oracle sql tuning. call count cpu elapsed disk query current rows Parse 1 0.00 0.00 0 0 2 0 Execute 46 0.42 4.61 717 47066 4332 6049 Fetch 0 0.00 0.00 0 0 0 0 total 47 0.42 4.61 717 47066 4334 6049 Misses in library cach

  • No UI redraw while Mousedown in Illustrator with flash panel (MAC only)

    Hello all, I've noticed in Illustrator whenever the mouse is held down over a FLASH UI object on a mac no updates are drawn to the UI until the mouse button is released. This makes things like highlighting text , using scrollbars etc very awkward. Ha

  • Change the Size of  "Code"  field in Master Table of UDO

    Hi there, I am using UDO in my form, where "Code" is the Primary key field of the Master Table. Now the Size of the same is 8 characters. Can any one of you help me, to change the "Code" size to 15 characters? Thanking you, Pradeep