Data Caching in a Clustered Environment

          I want to cache read-only reference/code table data that will run in a clustered
          WLS6 environment. It's a JSP application and I am storing a complete HTML Select
          Control per reference/code table data in the cache. The question is where to
          cache it? I was going to put it in the ServletContext (JSP "application" implicit
          object), but the ServletContext is not replicated. I considered using JNDI, but
          there are problems with duplicate name errors when another server who doesn't
          originally bind the object tries to lookup, change and rebind the object. I guess
          JMS Multicasting is an option, but I don't want to implement JMS just for an application
          data cache.
          Any suggestions for a simple reference/code table read-only caching strategy that
          will work in a clustered WLS6 environment?
          

If the data is strictly read-only, and you do not have to worry about cache
          integrity, then look at WebLogic JSP cachetag:
          http://www.weblogic.com/docs51/classdocs/API_jsp.html#cachetag
          You can use it to cache both the output and the calculations results
          (variables calculated inside the cache tag).
          The scenario will be exactly the same for non-clustered and clustered
          cases - using multicast to broadcast small invalidation messages (so the
          data can be refreshed from the database) is ok, but replicating application
          data is not (and you definitely do not want to use JNDI for this purpose).
          BTW, the initial CacheTag implementation in 5.1 (supposedly) had a 'cluster'
          scope and I assume it was multicasting fresh data after cache miss - there
          is no such scope in 6.0 implementation.
          If you still want replication you can look at javagroups:
          http://sourceforge.net/projects/javagroups/
          (distributedhashtable example).
          Olsen <[email protected]> wrote:
          > Cameron,
          > Thanks for the reply. However, as I stated below, I am not interested in
          > JMS, nor an EJB solution to the problem. It really is not that complicated of
          > a concept and I know a solution or two (ServletContext, JNDI), but none that works
          > in a WLS6 clustered environment.
          > Any other ideas???
          > Thanks...
          > "Cameron Purdy" <[email protected]> wrote:
          >>Dimitri had a clever (as ever) solution using JMS to maintain cache
          >>integrity:
          >>
          >>explanation at
          >>http://dima.dhs.org/misc/readOnlyUpdates.html
          >>
          >>d/l from
          >>http://dima.dhs.org/misc/readOnlyUpdates.jar
          >>
          >>--
          >>Cameron Purdy
          >>Tangosol, Inc.
          >>http://www.tangosol.com
          >>+1.617.623.5782
          >>WebLogic Consulting Available
          >>
          >>
          >>"Olsen" <[email protected]> wrote in message
          >>news:[email protected]...
          >>>
          >>> I want to cache read-only reference/code table data that will run in
          >>a
          >>clustered
          >>> WLS6 environment. It's a JSP application and I am storing a complete
          >>HTML
          >>Select
          >>> Control per reference/code table data in the cache. The question is
          >>where
          >>to
          >>> cache it? I was going to put it in the ServletContext (JSP "application"
          >>implicit
          >>> object), but the ServletContext is not replicated. I considered using
          >>JNDI, but
          >>> there are problems with duplicate name errors when another server who
          >>doesn't
          >>> originally bind the object tries to lookup, change and rebind the object.
          >>I guess
          >>> JMS Multicasting is an option, but I don't want to implement JMS just
          >>for
          >>an application
          >>> data cache.
          >>> Any suggestions for a simple reference/code table read-only caching
          >>strategy that
          >>> will work in a clustered WLS6 environment?
          >>
          >>
          Dimitri
          

Similar Messages

  • Error while Enabling Active Data Cache Clustering Services: error 1753

    I have installed the Oracle BAM with Enterprise link.All the installation was successful.But while Enabling Active Data Cache Clustering Services using the following command:
    cluster.exe restype "Oracle Business Activity Monitoring Active Data Cache" /create /dll:"C:\OracleBAM\ADCClusterResourceType.dll"
    i am getting the folllowing error:
    System error 1753 has occurred.
    There are no more endpoints available from the endpoint mapper.
    Although i stopped the Active Data Cache Service..?
    I am unable to resolve it
    Kindly help me out
    Ramesh Nambala

    Hi.
    I encountered the same problem. In my case, the reason was that I created a new user for BAM (to be provided when installing BAM) but ran the installation under another user account. It seems like these users must be the same to make it work.
    Greetings,
    cor

  • Global cache in clustered environment

    We have a clustered external facing portal application with four servers and each server has two nodes. We want use to IcacheService to store some objects in a cache. But this cache is specific to each JVM and cannot be shared across the servers.
    Is there any global cache mechanism in SAP through which we can share the objects across all the servers in a clustered environment?
    Thanks in adavace.
    Ram

    Hi Rambhupal,
    Would you share any solution/information for the same as we have a similar requirement.
    Regards,
    Melwyn

  • Data Caching in a WLS6 Cluster - Storage Location

    I want to cache read-only reference/code table data that will run in a clustered
    WLS6 environment. It's a JSP application and I am storing a complete HTML Select
    Control per reference/code table data in the cache. The question is where to cache
    it? I was going to put it in the ServletContext (JSP "application" implicit object),
    but the ServletContext is not replicated. I considered using JNDI, but there are
    problems with duplicate name errors when another server who doesn't originally
    bind the object tries to lookup, change and rebind the object. I guess JMS Multicasting
    is an option, but I don't want to implement JMS just for an application data cache.
    Any suggestions for a simple reference/code table read-only caching strategy that
    will work in a clustered WLS6 environment? I know how to do everything except
    how to have it available and in sync on all servers in a clustered environment?
    Any suggestions?

    Read-Only entity ejb?
    .raja
    P.S. Also read Rob Wollen's post in the JDBC news group where he explains
    the new features with example.
    <Chris Olsen chris.olsen> wrote in message
    news:3b09493f$[email protected]..
    >
    I want to cache read-only reference/code table data that will run in aclustered
    WLS6 environment. It's a JSP application and I am storing a complete HTMLSelect
    Control per reference/code table data in the cache. The question is whereto cache
    it? I was going to put it in the ServletContext (JSP "application"implicit object),
    but the ServletContext is not replicated. I considered using JNDI, butthere are
    problems with duplicate name errors when another server who doesn'toriginally
    bind the object tries to lookup, change and rebind the object. I guess JMSMulticasting
    is an option, but I don't want to implement JMS just for an applicationdata cache.
    Any suggestions for a simple reference/code table read-only cachingstrategy that
    will work in a clustered WLS6 environment? I know how to do everythingexcept
    how to have it available and in sync on all servers in a clusteredenvironment?
    >
    >
    Any suggestions?

  • Woking in clustered environment

    Are there any special concerns I should know about KODO JDO when running
    in a
    clustered environment???

    Can you post some details about the cluster? Do you mean you will be
    using EJBs, or that you will be running multiple separate intances of
    Kodo on different machines? If the latter, then you will either want to
    use pessimistic locking, or use some other mechanism to make sure that
    your cache dosen't have stale data.
    karim qazi <[email protected]> wrote:
    Are there any special concerns I should know about KODO JDO when running
    in a
    clustered environment???--
    Marc Prud'hommeaux [email protected]
    SolarMetric Inc. http://www.solarmetric.com
    Kodo Java Data Objects Full featured JDO: eliminate the SQL from your code

  • Singleton behaviour in clustered environment

    Hi,
              I have a very basic question. In a WebLogic clustered
              environment, is a Singleton replicated? In other words, if the
              mySingleton on node A in a cluster is updated with a piece of data,
              will this show in the mySingleton in node B?
              Thanks
              Ciao
              Ferruccio
              

    No. If you need this type of functionality, use Coherence.
              http://www.tangosol.com/coherence.jsp
              Peace,
              Cameron Purdy
              Tangosol, Inc.
              http://www.tangosol.com/coherence.jsp
              Tangosol Coherence: Clustered Replicated Cache for Weblogic
              "Ferruccio" <[email protected]> wrote in message
              news:[email protected]..
              > Hi,
              > I have a very basic question. In a WebLogic clustered
              > environment, is a Singleton replicated? In other words, if the
              > mySingleton on node A in a cluster is updated with a piece of data,
              > will this show in the mySingleton in node B?
              >
              > Thanks
              >
              > Ciao
              > Ferruccio
              

  • Design Considerations for a Repository Manager in a clustered environment

    Hi,
    I am currently building a repository manager for a backend document management system.  Our portal is installed in a load balanced clusterd environment and we are creating the following functionality in the repostiory manager -
      -  Browse content
      -  Read/Write document Metadata 
      -  Upload/download documentation
      -  Reserve/Unreserve documentation
      -  Search
    How will a clustered environment affect the implementation of the repository manager?  Would I need to check for a clustered installation in any kind of way in my code?  Currently I do not see why this would be necessary, but i'm not entirely sure and I need this manager to work in a clustered environment.
    I could see this as being an issue if we are caching information about the documents in the repository.  However, we are not.  We will only be caching user information which can be stored separately in a single cache on each server. 
    Thanks for your help,
    Scott

    Hi Scott!
    There might be two reasons for knowing whether other cluster nodes are running or not and for communicating with them:
    - synchronizing access to the backend
    - synchronizing caches
    If your repository manager must care for serialized access to the backend you will have to synchronize cluster-wide. You don't have to, if your backend can handle parallel access on its own (like a web-server).
    If you cache data in your repository and you want to update the caches of the repositories on the other cluster nodes (maybe because you don't get events for changed data from your backend), you must think of sending messages between cluster nodes. You don't have to, if you think that the expiration of your cache entries is shorter than the required timeliness of the data.
    Kind Regards, Dirk

  • Issue in Synchronous File Read in clustered environment

    Hi,
    We are using clustered environment (4 managed servers) in Unix. In a OSB 11gR3 proxy service we are using Synchronous File Read. Randomly we are getting the below error. Let us know what could cause the issue. But the same code is working fine in a single stand-alone server configuration.
    Error Code : BEA-380002 , Error Reason : Invoke JCA outbound service failed with connection error, exception: com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/wsdlPathAndName [ SynchRead_ptt::SynchRead(Empty,body) ] - WSIF JCA Execute of operation 'SynchRead' failed due to: No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content ; nested exception is:
    BINDING.JCA-11007
    No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderNamer/Filename.txt to be processed was not found or not available or has no content Please make sure that the file exists in the specified directory.
    com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/wsdlPathAndName [ SynchRead_ptt::SynchRead(Empty,body) ] - WSIF JCA Execute of operation 'SynchRead' failed due to: No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content ; nested exception is:
    BINDING.JCA-11007
    No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content Please make sure that the file exists in the specified directory.
    at com.bea.wli.sb.transports.jca.binding.JCATransportOutboundOperationBindingServiceImpl.invoke(JCATransportOutboundOperationBindingServiceImpl.java:153)
    at com.bea.wli.sb.transports.jca.JCATransportEndpoint.sendRequestResponse(JCATransportEndpoint.java:209)
    at com.bea.wli.sb.transports.jca.JCATransportEndpoint.send(JCATransportEndpoint.java:170)
    at com.bea.wli.sb.transports.jca.JCATransportProvider.sendMessageAsync(JCATransportProvider.java:598)
    at sun.reflect.GeneratedMethodAccessor1115.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at com.bea.wli.sb.transports.Util$1.invoke(Util.java:83)
    at $Proxy142.sendMessageAsync(Unknown Source)
    at com.bea.wli.sb.transports.LoadBalanceFailoverListener.sendMessageAsync(LoadBalanceFailoverListener.java:148)
    at com.bea.wli.sb.transports.LoadBalanceFailoverListener.sendMessageToServiceAsync(LoadBalanceFailoverListener.java:603)
    at com.bea.wli.sb.transports.LoadBalanceFailoverListener.sendMessageToService(LoadBalanceFailoverListener.java:538)
    at com.bea.wli.sb.transports.TransportManagerImpl.sendMessageToService(TransportManagerImpl.java:558)
    at com.bea.wli.sb.transports.TransportManagerImpl.sendMessageAsync(TransportManagerImpl.java:426)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.doDispatch(PipelineContextImpl.java:670)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.dispatchSync(PipelineContextImpl.java:551)
    at stages.transform.runtime.WsCalloutRuntimeStep$WsCalloutDispatcher.dispatch(WsCalloutRuntimeStep.java:1391)
    at stages.transform.runtime.WsCalloutRuntimeStep.processMessage(WsCalloutRuntimeStep.java:236)
    at com.bea.wli.sb.stages.StageMetadataImpl$WrapperRuntimeStep.processMessage(StageMetadataImpl.java:346)
    at com.bea.wli.sb.stages.impl.SequenceRuntimeStep.processMessage(SequenceRuntimeStep.java:33)
    at com.bea.wli.sb.pipeline.PipelineStage.processMessage(PipelineStage.java:84)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.execute(PipelineContextImpl.java:1055)
    at com.bea.wli.sb.pipeline.Pipeline.processMessage(Pipeline.java:141)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.execute(PipelineContextImpl.java:1055)
    at com.bea.wli.sb.pipeline.PipelineNode.doRequest(PipelineNode.java:55)
    at com.bea.wli.sb.pipeline.Node.processMessage(Node.java:67)
    at com.bea.wli.sb.pipeline.PipelineContextImpl.execute(PipelineContextImpl.java:1055)
    at com.bea.wli.sb.pipeline.Router.processMessage(Router.java:214)
    at com.bea.wli.sb.pipeline.MessageProcessor.processRequest(MessageProcessor.java:96)
    at com.bea.wli.sb.pipeline.RouterManager$1.run(RouterManager.java:593)
    at com.bea.wli.sb.pipeline.RouterManager$1.run(RouterManager.java:591)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:146)
    at com.bea.wli.sb.security.WLSSecurityContextService.runAs(WLSSecurityContextService.java:55)
    at com.bea.wli.sb.pipeline.RouterManager.processMessage(RouterManager.java:590)
    at com.bea.wli.sb.transports.TransportManagerImpl.receiveMessage(TransportManagerImpl.java:375)
    at com.bea.wli.sb.transports.jca.binding.JCATransportInboundOperationBindingServiceImpl$4.run(JCATransportInboundOperationBindingServiceImpl.java:415)
    at com.bea.wli.sb.transports.jca.binding.JCATransportInboundOperationBindingServiceImpl$4.run(JCATransportInboundOperationBindingServiceImpl.java:413)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:146)
    at weblogic.security.Security.runAs(Security.java:61)
    at com.bea.wli.sb.transports.jca.binding.JCATransportInboundOperationBindingServiceImpl.sendMessage(JCATransportInboundOperationBindingServiceImpl.java:413)
    at com.bea.wli.sb.transports.jca.binding.JCATransportInboundOperationBindingServiceImpl.invokeOneWay(JCATransportInboundOperationBindingServiceImpl.java:126)
    at com.bea.wli.sb.transports.jca.binding.JCAInboundRequestListener.post(JCAInboundRequestListener.java:39)
    at oracle.tip.adapter.sa.impl.inbound.JCAInboundListenerImpl.onMessage(JCAInboundListenerImpl.java:170)
    at oracle.tip.adapter.fw.jca.messageinflow.MessageEndpointImpl.onMessage(MessageEndpointImpl.java:502)
    at oracle.tip.adapter.file.inbound.Publisher.onMessageDelegate(Publisher.java:493)
    at oracle.tip.adapter.file.inbound.Publisher.publishMessage(Publisher.java:419)
    at oracle.tip.adapter.file.inbound.InboundTranslatorDelegate.xlate(InboundTranslatorDelegate.java:484)
    at oracle.tip.adapter.file.inbound.InboundTranslatorDelegate.doXlate(InboundTranslatorDelegate.java:121)
    at oracle.tip.adapter.file.inbound.ProcessorDelegate.doXlate(ProcessorDelegate.java:388)
    at oracle.tip.adapter.file.inbound.ProcessorDelegate.process(ProcessorDelegate.java:174)
    at oracle.tip.adapter.file.inbound.ProcessWork.run(ProcessWork.java:349)
    at weblogic.work.ContextWrap.run(ContextWrap.java:41)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:528)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:207)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:176)
    Caused by: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/wsdlPathAndName [ SynchRead_ptt::SynchRead(Empty,body) ] - WSIF JCA Execute of operation 'SynchRead' failed due to: No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content ; nested exception is:
    BINDING.JCA-11007
    No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content Please make sure that the file exists in the specified directory.
    at oracle.tip.adapter.sa.impl.JCABindingReferenceImpl.request(JCABindingReferenceImpl.java:259)
    at com.bea.wli.sb.transports.jca.binding.JCATransportOutboundOperationBindingServiceImpl.invoke(JCATransportOutboundOperationBindingServiceImpl.java:150)
    ... 56 more
    Caused by: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/wsdlPathAndName [ SynchRead_ptt::SynchRead(Empty,body) ] - WSIF JCA Execute of operation 'SynchRead' failed due to: No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content ; nested exception is:
    BINDING.JCA-11007
    No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content Please make sure that the file exists in the specified directory.
    at oracle.tip.adapter.sa.impl.fw.wsif.jca.WSIFOperation_JCA.performOperation(WSIFOperation_JCA.java:662)
    at oracle.tip.adapter.sa.impl.fw.wsif.jca.WSIFOperation_JCA.executeOperation(WSIFOperation_JCA.java:353)
    at oracle.tip.adapter.sa.impl.fw.wsif.jca.WSIFOperation_JCA.executeRequestResponseOperation(WSIFOperation_JCA.java:312)
    at oracle.tip.adapter.sa.impl.JCABindingReferenceImpl.invokeWsifProvider(JCABindingReferenceImpl.java:350)
    at oracle.tip.adapter.sa.impl.JCABindingReferenceImpl.request(JCABindingReferenceImpl.java:253)
    ... 57 more
    Caused by: BINDING.JCA-11007
    No Data to process.
    No Data to process.
    File /root/oracle/domains/osb/11.1.1.4/cluster/data/osb2/FolderName/Filename.txt to be processed was not found or not available or has no content Please make sure that the file exists in the specified directory.
    at oracle.tip.adapter.file.outbound.FileReader.readFile(FileReader.java:277)
    at oracle.tip.adapter.file.outbound.FileReader.executeFileRead(FileReader.java:181)
    at oracle.tip.adapter.file.outbound.FileInteraction.executeFileRead(FileInteraction.java:331)
    at oracle.tip.adapter.file.outbound.FileInteraction.execute(FileInteraction.java:395)
    at oracle.tip.adapter.sa.impl.fw.wsif.jca.WSIFOperation_JCA.performOperation(WSIFOperation_JCA.java:529)
    ... 61 more
    Edited by: 842347 on Jul 6, 2011 3:11 AM

    I face the same issue and I have given all permissions to the folder for OS user.
    Because of this error my server is not starting up . Is there any way I can undeploy this composite to get my server running.
    I cant do this from EM because SOA server is failing to start up.
    I have tried removing it from $DOMAIN_HOME/deployed-composites but still when i try restarting the soa server the composite comes up there. Do we need to delete the entry some where else too. Kindly help.
    Thanks,
    Sri.

  • Web Application - Data caching of enterprise data

    Sorry in advance if this is off-topic but I can't find anywhere else to post this type of question.
    I am looking for information/suggestions such as books, technology or design methodology for my enterprise web applications. These sites are currently up and functional using only JSP, servlets and regular Java classes stored in a web application session to provide data caching and access. We are using Weblogic Server 6.1 running on an AIX Unix system at this time. I'm not sure that this is the best design architecture as our web sessions are getting too large but I can't think of any other Java technology to use and I need some help. Here's an overview of our environment and our needs.
    Our core data is held in a mainframe based IMS system. Some DB2 is also used. Access to this data is through IMS COBOL transactions which we can execute with IMS Connect. We also use some JDBC to get to the DB2 tables directly where available.
    Some overall application data is cached when the web application is deployed. We use singleton classes which are created and refreshed at deployment and they then refresh themselves from the sources every 24 hours.
    Each time a user logs in we execute several IMS transactions and JDBC calls to cache user specific data in regular Java classes which are then simply placed in the users web session where we access them from JSP, servlets and other Java classes. The fields in these Java classes range from any type of primitive field to TreeMaps of other Java classes. As the data is cached it is sorted and other fields are calculated and stored in these classes. As the user progresses through the system we then may have to do several other IMS transaction and JDBC calls to collect other types of data. All of this is then also added to the users session. Most of this in inquiry. We do allow transactions but those are built from user input and data already cached and are then we just execute the IMS transactions with the input.
    As our application has grown these Java classes have gotten larger and larger. And since these are simply stored in server memory in the web sessions then these are also getting huge. I'm concerned that this is not the best way for this application to be architected. Is there something else we should be doign? I simply don't understand how Entity Java Beans could be used but then again I don't know much about them. I wouldn't think that caching the data to a local database and accessing it from there would be any more efficient and would probably just slow down the system from all the I/O.
    Any help or direction would be greatly appreciated.

    The best book you can buy is 'Professional Java Server Programming, j2ee edition' by Wrox. It is by far the best reference I've used. Another quick reference consideration might be the j2ee book provided by codenotes... its quick and to the point.

  • Issue while polling in clustered environment.

    Hi All,
    I am using a polling adapter which polls records from two database tables. I have set Database rows per transaction as 10 and polling interval as 30 secs. Our servers are running in clustered mode.
    The issue is that, when the rows are picked by node 1 , without waiting for the polling interval node 2 picks up the next set of rows from database.
    As per our requirement, rows should be polled according to the value of Database rows per transaction size. Then adapter should wait for the polling interval and then pick up the next set of rows (in the clustered environment only). Can it be achieved using custom SQL or some java code. If anyone has tried it please help me.
    Thanks in advance.

    See this is misconception that both nodes picks the data in turn one by one.
    It totally depends on the availability of the nodes. This is how a typical active active cluster should behave.
    Both nodes will show similar set of instances as they are pointing to same soa_infra schema.
    We can not differentiate by checking EM that which node actually processed a particular instances.
    I would like to know how are you sure that both nodes are picking file one by one without waiting other one to complete its job.
    I have experience with Active active cluster with uses of DB Adapter in many scenarios. But I never observed this behavior.
    Probably by understanding your applications architecture, answer to your problem can be found.
    Thanks,
    Ashu

  • File Protocol in OSB Clustered Environment

    Hi ,
    I have a OSB Server in a Clustered Environment ....
    Here Admin Server is in Machine1
    Managed Server1 is in Machine 2
    Managed Server2 is in Machine 3
    so both the Managed Servers are put in a Cluster
    I have Access to Machine1 Unix Box where the Admin Server is Installed ...I can get the File only in this Location and have to Read the Same from here , So i was planning to Use File Protocol for Polling the Message from Admin Server .Since OSB has option for Selecting only the Managed Server , I would request suggestions how i can Read the Data from this Admin Server by using File Protocol .

    have you used "service type" as "Messaging Service" ? that should give you all the options you need

  • Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).

    Hi,
    Our Environment is Essbase 11.1.2.2 and working on Essbase EAS and Shared Services components.One of our user tried to run the Cal Script of one Application and faced this error.
    Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).
    I have done some Google and found that we need to add something in Essbase.cfg file like below.
    1012704 Dynamic Calc processor cannot lock more than number ESM blocks during the calculation, please increase CalcLockBlock setting and then retry (a small data cache setting could also cause this problem, please check the data cache size setting).
    Possible Problems
    Analytic Services could not lock enough blocks to perform the calculation.
    Possible Solutions
    Increase the number of blocks that Analytic Services can allocate for a calculation:
    Set the maximum number of blocks that Analytic Services can allocate to at least 500. 
    If you do not have an $ARBORPATH/bin/essbase.cfg file on the server computer, create one using a text editor.
    In the essbase.cfg file on the server computer, set CALCLOCKBLOCKHIGH to 500.
    Stop and restart Analytic Server.
    Add the SET LOCKBLOCK HIGH command to the beginning of the calculation script.
    Set the data cache large enough to hold all the blocks specified in the CALCLOCKBLOCKHIGH setting. 
    Determine the block size.
    Set the data catche size.
    Actually in our Server Config file(essbase.cfg) we dont have below data  added.
    CalcLockBlockHigh 2000
    CalcLockBlockDefault 200
    CalcLockBlocklow 50
    So my doubt is if we edit the Essbase.cfg file and add the above settings and restart the services will it work?  and if so why should we change the Server config file if the problem is with one application Cal Script. Please guide me how to proceed.
    Regards,
    Naveen

    Your calculation needs to hold more blocks in memory than your current set up allows.
    From the docs (quoting so I don't have to write it, not to be a smarta***:
    CALCLOCKBLOCK specifies the number of blocks that can be fixed at each level of the SET LOCKBLOCK HIGH | DEFAULT | LOW calculation script command.
    When a block is calculated, Essbase fixes (gets addressability to) the block along with the blocks containing its children. Essbase calculates the block and then releases it along with the blocks containing its children. By default, Essbase allows up to 100 blocks to be fixed concurrently when calculating a block. This is sufficient for most database calculations. However, you may want to set a number higher than 100 if you are consolidating very large numbers of children in a formula calculation. This ensures that Essbase can fix all the required blocks when calculating a data block and that performance will not be impaired.
    Example
    If the essbase.cfg file contains the following settings:
    CALCLOCKBLOCKHIGH 500  CALCLOCKBLOCKDEFAULT 200  CALCLOCKBLOCKLOW 50 
    then you can use the following SET LOCKBLOCK setting commands in a calculation script:
    SET LOCKBLOCK HIGH; 
    means that Essbase can fix up to 500 data blocks when calculating one block.
    Support doc is saying to change your config file so those settings can be made available for any calc script to use.
    On a side note, if this was working previously and now isn't then it is worth investigating if this is simply due to standard growth or a recent change that has made an unexpected significant impact.

  • Error starting Oracle BAM active data cache service

    Hi
    after installing BAM every thing working fine ,but if restart my system Oracle BAM active data cache service throwing following error
    "The Oracle BAM Active Data Cache service on Local computer started and then stopped.Some services stop automatically if they have no work to do,for example the performance logs and alerts service"
    Database is running fine
    Following is the ADC log file error
    2007-12-07 17:19:29,640 [2928] ERROR - ActiveDataCache The Oracle BAM Active Data Cache service failed to start. Oracle.BAM.ActiveDataCache.Common.Exceptions.CacheException: ADC Server exception in Startup(). ---> System.DllNotFoundException: Unable to load DLL (OraOps10.dll).
    at Oracle.DataAccess.Client.OpsTrace.GetRegTraceInfo(UInt32& TrcLevel, UInt32& StmtCacheSize)
    at Oracle.DataAccess.Client.OraTrace.GetRegistryTraceInfo()
    at Oracle.DataAccess.Client.OracleConnection..ctor(String connectionString)
    at Oracle.DataAccess.Client.OracleConnection..ctor(String connectionString)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleDataFactory.GetConnection()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetServerVersion()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.Startup(IDictionary oParameters)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    --- End of inner exception stack trace ---
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    at Oracle.BAM.ActiveDataCache.Kernel.Server.Server.Startup()
    at Oracle.BAM.ActiveDataCache.Service.DataServer.Run()
    2007-12-07 17:24:45,250 [1524] ERROR - ActiveDataCache Unable to load DLL (OraOps10.dll).
    2007-12-07 17:24:45,265 [1524] WARN - ActiveDataCache Exception occurred in method Startup
    Please help me in resolving this issue .Am getting this issue every time
    Thanks
    BS

    Make sure the path to the ODAC used by BAM (C:\OracleBAM\ClientForBAM\bin) is the first item in the system PATH
    environment variable. Restart your computer after fixing this.
    If that doesn't fix it, please check the Troubleshooting section in the BAM Install Guide.
    Regards, Stephen

  • Pessimistic lock exception in a clustered environment

    Hello,
    i am using distributed Cache(DataCache) in a Cluster environment
    (Weblogic 7.0 cluster).
    I got one AdminServer and two ManagedServer running on the same machine.
    The problem occurs when i am using optimistic locking. It seems to be that
    the cache is not beeing synchronized, when i am successivly persisting
    objects and therfore get optimistic lock exceptions.
    When i am using pessimistic locking there is no problem.
    Waht could be the reason for this, maybe the case the all instances
    running on the same machine?
    My properties for the cache are the following:
    com.solarmetric.kodo.DataCacheClass=de.d2vodafone.webe.common.security.
    authorization.jdo.Cache (Cache is a subclass of UDPCache)
    com.solarmetric.kodo.DataCacheProperties=UseMulticast=true
    MulticastGroup=224.0.0.123 (is unique in my environment)
    In the log of my server i got the following messages:
    ignoring update from localhost
    ####<12.02.2003 13:25:11 CET> <Error> <EJB> <pc120514> <ManagedServer2>
    <ExecuteThread: '4' for queue: '__weblogic_admin_rmi_queue'> <kernel
    identity> <> <010026> <Exception during commit of transaction
    4:b57355e6636065c2: javax.jdo.JDOUserException: Errors were detected when
    flushing to the data store. If using optimistic transactions, this could
    indicate that lock violations were detected. The "getNestedExceptions"
    method of this Exception will return the specific errors.
    NestedThrowables:
    javax.jdo.JDOUserException:
    An instance was concurrently modified in another transaction (optimistic
    lock #: 3).
    FailedObject:de.webe.common.security.authorization.DomainImpl-5
         at
    com.solarmetric.kodo.runtime.PersistenceManagerImpl.flush(PersistenceManagerImpl.java:551)
         at
    com.solarmetric.kodo.ee.EEPersistenceManager.beforeCompletion(EEPersistenceManager.java:192)
         at
    weblogic.transaction.internal.ServerSCInfo.callBeforeCompletions(ServerSCInfo.java:540)
         at
    weblogic.transaction.internal.ServerSCInfo.startPrePrepareAndChain(ServerSCInfo.java:92)
         at
    weblogic.transaction.internal.ServerTransactionImpl.localPrePrepareAndChain(ServerTransactionImpl.java:1044)
         at
    weblogic.transaction.internal.ServerTransactionImpl.globalPrePrepare(ServerTransactionImpl.java:1581)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:237)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    NestedThrowablesStackTrace:
    javax.jdo.JDOUserException: An instance was concurrently modified in
    another transaction (optimistic lock #: 3).
    FailedObject:de.d2vodafone.webe.common.security.authorization.DomainImpl-5
         at
    com.solarmetric.kodo.impl.jdbc.ormapping.ClassMapping$1.afterUpdate(ClassMapping.java:611)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executePreparedStatementNonBatch(SQLExecutionManagerImpl.java:563)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executePreparedStatement(SQLExecutionManagerImpl.java:512)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executeInternal(SQLExecutionManagerImpl.java:406)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.flush(SQLExecutionManagerImpl.java:273)
         at
    com.solarmetric.kodo.impl.jdbc.runtime.JDBCStoreManager.flush(JDBCStoreManager.java:421)
         at
    com.solarmetric.kodo.runtime.datacache.DataCacheStoreManager.flush(DataCacheStoreManager.java:296)
         at
    com.solarmetric.kodo.runtime.PersistenceManagerImpl.flush(PersistenceManagerImpl.java:549)
         at
    com.solarmetric.kodo.ee.EEPersistenceManager.beforeCompletion(EEPersistenceManager.java:192)
         at
    weblogic.transaction.internal.ServerSCInfo.callBeforeCompletions(ServerSCInfo.java:540)
         at
    weblogic.transaction.internal.ServerSCInfo.startPrePrepareAndChain(ServerSCInfo.java:92)
         at
    weblogic.transaction.internal.ServerTransactionImpl.localPrePrepareAndChain(ServerTransactionImpl.java:1044)
         at
    weblogic.transaction.internal.ServerTransactionImpl.globalPrePrepare(ServerTransactionImpl.java:1581)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:237)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.d2vodafone.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.d2vodafone.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    --------------- nested within: ------------------
    weblogic.transaction.RollbackException: Unexpected exception in
    beforeCompletion: sync=com.solarmetric.kodo.ee.EEPersistenceManager@472ddf
    Errors were detected when flushing to the data store. If using optimistic
    transactions, this could indicate that lock violations were detected. The
    "getNestedExceptions" method of this Exception will return the specific
    errors. - with nested exception:
    [javax.jdo.JDOUserException: Errors were detected when flushing to the
    data store.  If using optimistic transactions, this could indicate that
    lock violations were detected.  The "getNestedExceptions" method of this
    Exception will return the specific errors.
    NestedThrowables:
    javax.jdo.JDOUserException: An instance was concurrently modified in
    another transaction (optimistic lock #: 3).
    FailedObject:de.d2vodafone.webe.common.security.authorization.DomainImpl-5]
         at
    weblogic.transaction.internal.TransactionImpl.throwRollbackException(TransactionImpl.java:1561)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:284)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    Thanks for your help
    Best Regards
    Orhan Keskin

    Orhan-
    Is this consistently reproducable, or a rare or one-time occurance? If
    the former, it may be a configuration issue, and we will investigatge
    along that path.
    If it is a rare or one-time occurance, keep in mind that data cache
    synchronization does not guarantee that you will be able to avoid
    optimistic locking exceptions. Whenever using optimistic locking, you
    always need to assume that someday, someone is going to violate that
    lock, unless you place some sort of synchronization atop all the
    database access (e.g., an EJB). It makes any application much more error
    prone (especially since an OL exception is a RuntimeException, and
    frequently people forget that it always needs to be checked for).
    Anyway, if you provide us with some more details, we will be able to
    investigate further.
    In article <[email protected]>, Orhan Keskin wrote:
    Hello,
    i am using distributed Cache(DataCache) in a Cluster environment
    (Weblogic 7.0 cluster).
    I got one AdminServer and two ManagedServer running on the same machine.
    The problem occurs when i am using optimistic locking. It seems to be that
    the cache is not beeing synchronized, when i am successivly persisting
    objects and therfore get optimistic lock exceptions.
    When i am using pessimistic locking there is no problem.
    Waht could be the reason for this, maybe the case the all instances
    running on the same machine?
    My properties for the cache are the following:
    com.solarmetric.kodo.DataCacheClass=de.d2vodafone.webe.common.security.
    authorization.jdo.Cache (Cache is a subclass of UDPCache)
    com.solarmetric.kodo.DataCacheProperties=UseMulticast=true
    MulticastGroup=224.0.0.123 (is unique in my environment)
    In the log of my server i got the following messages:
    ignoring update from localhost
    ####<12.02.2003 13:25:11 CET> <Error> <EJB> <pc120514> <ManagedServer2>
    <ExecuteThread: '4' for queue: '__weblogic_admin_rmi_queue'> <kernel
    identity> <> <010026> <Exception during commit of transaction
    4:b57355e6636065c2: javax.jdo.JDOUserException: Errors were detected when
    flushing to the data store. If using optimistic transactions, this could
    indicate that lock violations were detected. The "getNestedExceptions"
    method of this Exception will return the specific errors.
    NestedThrowables:
    javax.jdo.JDOUserException:
    An instance was concurrently modified in another transaction (optimistic
    lock #: 3).
    FailedObject:de.webe.common.security.authorization.DomainImpl-5
         at
    com.solarmetric.kodo.runtime.PersistenceManagerImpl.flush(PersistenceManagerImpl.java:551)
         at
    com.solarmetric.kodo.ee.EEPersistenceManager.beforeCompletion(EEPersistenceManager.java:192)
         at
    weblogic.transaction.internal.ServerSCInfo.callBeforeCompletions(ServerSCInfo.java:540)
         at
    weblogic.transaction.internal.ServerSCInfo.startPrePrepareAndChain(ServerSCInfo.java:92)
         at
    weblogic.transaction.internal.ServerTransactionImpl.localPrePrepareAndChain(ServerTransactionImpl.java:1044)
         at
    weblogic.transaction.internal.ServerTransactionImpl.globalPrePrepare(ServerTransactionImpl.java:1581)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:237)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    NestedThrowablesStackTrace:
    javax.jdo.JDOUserException: An instance was concurrently modified in
    another transaction (optimistic lock #: 3).
    FailedObject:de.d2vodafone.webe.common.security.authorization.DomainImpl-5
         at
    com.solarmetric.kodo.impl.jdbc.ormapping.ClassMapping$1.afterUpdate(ClassMapping.java:611)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executePreparedStatementNonBatch(SQLExecutionManagerImpl.java:563)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executePreparedStatement(SQLExecutionManagerImpl.java:512)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.executeInternal(SQLExecutionManagerImpl.java:406)
         at
    com.solarmetric.kodo.impl.jdbc.SQLExecutionManagerImpl.flush(SQLExecutionManagerImpl.java:273)
         at
    com.solarmetric.kodo.impl.jdbc.runtime.JDBCStoreManager.flush(JDBCStoreManager.java:421)
         at
    com.solarmetric.kodo.runtime.datacache.DataCacheStoreManager.flush(DataCacheStoreManager.java:296)
         at
    com.solarmetric.kodo.runtime.PersistenceManagerImpl.flush(PersistenceManagerImpl.java:549)
         at
    com.solarmetric.kodo.ee.EEPersistenceManager.beforeCompletion(EEPersistenceManager.java:192)
         at
    weblogic.transaction.internal.ServerSCInfo.callBeforeCompletions(ServerSCInfo.java:540)
         at
    weblogic.transaction.internal.ServerSCInfo.startPrePrepareAndChain(ServerSCInfo.java:92)
         at
    weblogic.transaction.internal.ServerTransactionImpl.localPrePrepareAndChain(ServerTransactionImpl.java:1044)
         at
    weblogic.transaction.internal.ServerTransactionImpl.globalPrePrepare(ServerTransactionImpl.java:1581)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:237)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.d2vodafone.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.d2vodafone.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    --------------- nested within: ------------------
    weblogic.transaction.RollbackException: Unexpected exception in
    beforeCompletion: sync=com.solarmetric.kodo.ee.EEPersistenceManager@472ddf
    Errors were detected when flushing to the data store. If using optimistic
    transactions, this could indicate that lock violations were detected. The
    "getNestedExceptions" method of this Exception will return the specific
    errors. - with nested exception:
    [javax.jdo.JDOUserException: Errors were detected when flushing to the
    data store.  If using optimistic transactions, this could indicate that
    lock violations were detected.  The "getNestedExceptions" method of this
    Exception will return the specific errors.
    NestedThrowables:
    javax.jdo.JDOUserException: An instance was concurrently modified in
    another transaction (optimistic lock #: 3).
    FailedObject:de.d2vodafone.webe.common.security.authorization.DomainImpl-5]
         at
    weblogic.transaction.internal.TransactionImpl.throwRollbackException(TransactionImpl.java:1561)
         at
    weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:284)
         at
    weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:208)
         at
    weblogic.ejb20.internal.BaseEJBObject.postInvoke(BaseEJBObject.java:278)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl.createUser(AuthorizationServiceBean_8zrrq4_EOImpl.java:170)
         at
    de.webe.common.security.authorization.jdo.AuthorizationServiceBean_8zrrq4_EOImpl_WLSkel.invoke(Unknown
    Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:362)
         at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java:114)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:313)
         at
    weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:785)
         at
    weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:308)
         at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:30)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    Thanks for your help
    Best Regards
    Orhan Keskin
    Marc Prud'hommeaux [email protected]
    SolarMetric Inc. http://www.solarmetric.com

  • How to configure params for buffer pool for named data cache?

    when create a named data cache on ase 12.5, it will setup 2K I/O buffer pool by default with Configured size=0, wash size = 60M
    1. if 2K can be changed to 8K for this buffer pool?
    2. If add another 16k buffer pool, should Affected Pool be changed to the right pool?
    3. How to decide then pagesize, configured size and wash size for a buffer pool? Are they part of total memory size allocated for this cache?

    > 1. if 2K can be changed to 8K for this buffer pool?
    You should be able to create an 8K i/o pool, then drop the 2k pool by setting its size to 0.
    > 2. If add another 16k buffer pool, should Affected Pool be changed to the right pool?
    If you don't specify the Affected Pool (when calling sp_poolconfig), the procedure uses the pool with the smallest i/o.  So if you had an 8k but not had dropped the 2k, the space for the new 16k pool would come from the 8k pool.
    > 3. How to decide then pagesize, configured size and wash size for a buffer pool? Are they part of total memory size allocated for this cache?
    The wash is included in the pool. I don't think it usually needs to be adjusted.
    Which page size pools to have will depend on how the cache is used.  Tables with a clustered index that have a lot of range queries will benefit from larger page size pools, as will text/image/java.   Syslogs is said to do well on a 4k pool.
    -bret

Maybe you are looking for