Pof Serialization Error leads to partial cache updates in XA Tran

I am using coherence jca adapter to enlist in XA transactions with the database operations. The data is being stored in distributed caches with the cache member running on weblogic server with "Storage disabled". POF is being used for serialization. As a part of a single transaction, multiple caches, which are obtained from the CacheAdapter, are being updated. The application code does explicit updates to the cache and the database within the same transaction with the write to the cache happening after the write the database has been executed.
It is being observed that when an exception happens during the serialization of an object all the cache updates prior to this error are not rolled back. Namely,
I have CacheA for Object A, Cache B for Object B and Cache C for Object C.
I am updating A, B, C with the same transaction and in the same order as the objects are listed. So database for A is updated followed by cache update for A, DataBase B is updated followed by Cache update for B and similarly for C.
If there is an error while serializing C, all the database updates are rolled back, however updates to A and B are committed to the cache.
Why aren't all the cache updates being rolled back ? Has this been fixed already in coherence 3.6.
Thanks,
Shamsur
Application Server: Weblogic 10.3
jdbc driver : XA thin driver
coherence : 3.5.0
Caused by: java.lang.IllegalStateException: decimal value exceeds IEEE754r 128-bit range: 7777777788888888888899999999999900000000000000044444444444447777777777777.00
at com.tangosol.io.pof.PofHelper.calcDecimalSize(PofHelper.java:1517)
at com.tangosol.io.pof.PofBufferWriter.writeBigDecimal(PofBufferWriter.java:562)
at com.tangosol.io.pof.PofBufferWriter.writeObject(PofBufferWriter.java:1325)
at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.writeObject(PofBufferWriter.java:2092)
at com.apx.core.datalayer.data.basictypes.BigDecimalMoneyImpl.writeExternal(BigDecimalMoneyImpl.java:127)
at com.tangosol.io.pof.PortableObjectSerializer.serialize(PortableObjectSerializer.java:88)
at com.tangosol.io.pof.PofBufferWriter.writeObject(PofBufferWriter.java:1439)
at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.writeObject(PofBufferWriter.java:2092)
at com.apx.instrument.datalayer.data.domain.impl.DOtcInstrumentImpl.writeExternal(DOtcInstrumentImpl.java:200)
at com.apx.core.datalayer.data.impl.AbstractDomainObject.writeExternal(AbstractDomainObject.java:109)
... 185 more
at com.tangosol.coherence.ra.component.connector.resourceAdapter.cciAdapter.CacheAdapter$ManagedConnection$LocalTransaction.commit(CacheAdapter.CDB:37)
at weblogic.connector.security.layer.AdapterLayer.commit(AdapterLayer.java:570)
at weblogic.connector.transaction.outbound.NonXAWrapper.commit(NonXAWrapper.java:84)
at weblogic.transaction.internal.NonXAServerResourceInfo.commit(NonXAServerResourceInfo.java:330)
at weblogic.transaction.internal.ServerTransactionImpl.globalPrepare(ServerTransactionImpl.java:2251)
at weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:270)
at weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:230)
at weblogic.transaction.internal.TransactionManagerImpl.commit(TransactionManagerImpl.java:283)
at $Proxy150.create(Unknown Source)
javax.resource.spi.LocalTransactionException: CoherenceRA: Commit failed:
java.lang.RuntimeException: error with the class: com.apx.instrument.datalayer.data.domain.impl.DOtcInstrumentImpl
at com.apx.core.datalayer.data.impl.AbstractDomainObject.writeExternal(AbstractDomainObject.java:111)
at com.tangosol.io.pof.PortableObjectSerializer.serialize(PortableObjectSerializer.java:88)
at com.tangosol.io.pof.PofBufferWriter.writeObject(PofBufferWriter.java:1439)
at com.tangosol.io.pof.ConfigurablePofContext.serialize(ConfigurablePofContext.java:338)
at com.tangosol.util.ExternalizableHelper.serializeInternal(ExternalizableHelper.java:2508)
at com.tangosol.util.ExternalizableHelper.toBinary(ExternalizableHelper.java:205)
at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$ConverterValueToBinary.convert(DistributedCache.CDB:3)
at com.tangosol.util.ConverterCollections$AbstractConverterEntry.getValue(ConverterCollections.java:3333)
at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.putAll(DistributedCache.CDB:19)
at com.tangosol.util.ConverterCollections$ConverterMap.putAll(ConverterCollections.java:1570)
at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$ViewMap.putAll(DistributedCache.CDB:1)
at com.tangosol.coherence.component.util.SafeNamedCache.putAll(SafeNamedCache.CDB:1)
at com.tangosol.coherence.component.util.collections.WrapperMap.putAll(WrapperMap.CDB:1)
at com.tangosol.coherence.component.util.DeltaMap.resolve(DeltaMap.CDB:9)
at com.tangosol.coherence.component.util.deltaMap.TransactionMap.commit(TransactionMap.CDB:1)
at com.tangosol.coherence.component.util.TransactionCache.commit(TransactionCache.CDB:14)
at com.tangosol.coherence.component.util.transactionCache.Local.commit(Local.CDB:1)
at com.tangosol.coherence.ra.component.connector.resourceAdapter.cciAdapter.CacheAdapter$ManagedConnection$LocalTransaction.commit(CacheAdapter.CDB:25)
at weblogic.connector.security.layer.AdapterLayer.commit(AdapterLayer.java:570)
at weblogic.connector.transaction.outbound.NonXAWrapper.commit(NonXAWrapper.java:84)
at weblogic.transaction.internal.NonXAServerResourceInfo.commit(NonXAServerResourceInfo.java:330)
at weblogic.transaction.internal.ServerTransactionImpl.globalPrepare(ServerTransactionImpl.java:2251)
at weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:270)
at weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:230)
at weblogic.transaction.internal.TransactionManagerImpl.commit(TransactionManagerImpl.java:283)
at org.springframework.transaction.jta.JtaTransactionManager.doCommit(JtaTransactionManager.java:1009)
at org.springframework.transaction.support.AbstractPlatformTransactionManager.processCommit(AbstractPlatformTransactionManager.java:754)
at org.springframework.transaction.support.AbstractPlatformTransactionManager.commit(AbstractPlatformTransactionManager.java:723)
at org.springframework.transaction.interceptor.TransactionAspectSupport.commitTransactionAfterReturning(TransactionAspectSupport.java:374)
at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:120)
at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:172)

Hi SR-APX
The problem is that even though you are using the JCA adaptor Coherence (pre-3.6) is not really transactional. Once you commit all that data is still being pushed out to the distributed cluster members who all work independently. An error on one member will not stop data being written successfully to others.
In Coherence 3.6 there are real transactions but you would need to see if the limitations on them fit your use-cases.
JK

Similar Messages

  • Error in Directory Cache Update

    Hi,
    because we changed from two SLDs (PROD & DEV) to one (DEV) we did all changes like given in note 720717.
    Everything seems to run fine except the Adapter Engine - ok parts of it.
    When checking the Cache-Infos in Integration Directory we get following error in Integration Server (Central Adapter Engine). What do we need to do?!
    br
    com.sap.aii.ib.server.abapcache.CacheRefreshException: Unable to find an associated SLD element (source element: SAP_XIIntegrationDirectory, [CreationClassName, SAP_XIIntegrationDirectory, string, Name, directory.px1.sapru03, string], target element type: SAP_XIIntegrationServer)
         at com.sap.aii.ibdir.server.abapcache.content.CacheCPA.addContent(CacheCPA.java:483)
         at com.sap.aii.ibdir.server.abapcache.content.CacheCPA.addContent(CacheCPA.java:154)
         at com.sap.aii.ibdir.server.abapcache.CacheRefreshRequest.addContent(CacheRefreshRequest.java:388)
         at com.sap.aii.ibdir.server.abapcache.CacheRefreshRequest.addContent(CacheRefreshRequest.java:326)
         at com.sap.aii.ibdir.server.abapcache.CacheRefreshRequest.processHTTPRequest(CacheRefreshRequest.java:145)
         at com.sap.aii.ibdir.server.abapcache.CacheRefreshRequest.handleHTTPRequest(CacheRefreshRequest.java:103)
         at com.sap.aii.ibdir.web.abapcache.HmiMethod_CacheRefresh.process(HmiMethod_CacheRefresh.java:67)
         at com.sap.aii.utilxi.hmis.server.HmisServiceImpl.invokeMethod(HmisServiceImpl.java:169)
         at com.sap.aii.utilxi.hmis.server.HmisServer.process(HmisServer.java:178)
         at com.sap.aii.utilxi.hmis.sbeans.HmisBeanImpl.process(HmisBeanImpl.java:86)
         at com.sap.aii.utilxi.hmis.sbeans.HmisLocalLocalObjectImpl10.process(HmisLocalLocalObjectImpl10.java:259)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.processRequestByHmiServer(HmisServletImpl.java:290)
         at com.sap.aii.utilxi.hmis.web.workers.HmisExternalClient.doWork(HmisExternalClient.java:75)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.doWork(HmisServletImpl.java:496)
         at com.sap.aii.utilxi.hmis.web.HmisServletImpl.doPost(HmisServletImpl.java:634)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:390)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:264)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:347)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:325)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:887)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:241)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:92)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:148)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(AccessController.java:207)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)

    Hi,
    Hereu2019s a list of common errors/problems in SAP XI and their possible resolutions. This Guide will help you troubleshoot your integration scenarios in SAP XI/PI. This is in no way an exhaustive list. You can add your points/ideas to this list. Please feel free to post your inputs using the comments form at the end of this article.
    Cache Update Problems
    Use transaction SXI_CACHE to update the Integration Directory cache. Alternatively, you can use the following URLs to update the CPA cache. Use XIDIRUSER to refresh the cache.
    For complete cache refresh - http://<hostname>:<port>/CPACache/refresh?mode=full
    For delta cache refresh - http://<hostname>:<port>/CPACache/refresh?mode=delta
    If this does not solve the issue, check transaction SLDCHECK to ensure that connection to SLD is available. If the connection fails, check the configuration in the transaction SLDAPICUST. Make sure that the password maintained is correct and the maintained service user is not locked.
    Now in the Integration Repository go to Environment u2192 Clear SLD Data Cache. Also go to Integration Directoy and clear the cache using menu Environment u2192 Clear SLD Data Cache.
    Open the XI Start Page and click on Administration. On the Repository tab, choose Cache Overview. Refresh the cache using the buttons/icons on the right. Use XIDIRUSER to refresh the cache. Carry out cache refresh in the same way on the Directory and Runtime tabs.
    If you are facing cache update problems in your BPM (say you have modified the BPM, but when executed old version of the BPM is picked up instead of the new one), run the transaction SWF_XI_CUSTOMIZING and press F9 carry out automatic BPM/Workflow Customizing.
    Routing Errors
    NO_RECEIVER_CASE_BE or NO_RECEIVER_CASE_ASYNC
    This means no receiver could be found. Check your Receiver Determination. Activate and update cache. Asysnchronous messages can be manually restarted.
    TOO_MANY_RECEIVERS_CASE_BE
    More than one receiver found. Check your ID configuration to ensure that there is exactly one receiver for the synchronous message. Multiple receivers for synchronous interfaces are not permitted.
    Mapping Errors
    JCO_COMMUNICATION_FAILURE
    Check whether RFC destination AI_RUNTIME_JCOSERVER is correctly configured
    NO_MAPPINGPROGRAM_FOUND
    Ensure that mapping program exists and is activated. If it exists then update the cache.
    EXCEPTION_DURING_EXECUTE
    This error occurs due to erroneous XML formatting. Check your mapping program and ensure that you supply valid input data.
    Messages stuck in queues
    Check the queues using transactions SMQ1 (outbound)/SMQ2 (inbound). Resolve the displayed errors. You can cancel the messages from SXMB_MONI. Execute LUW if necessary and avoid deleting entries manually.
    Conversion Errors
    Unable to convert the sender service XXXX to an ALE logical system
    This error occurs in case of scenarios with IDoc adapters. Whenever you use business systems, make sure that the corresponding logical system name is maintained in the SLD.
    Open your business system in the Integration Directory. Switch to Change mode. Access the menu path Service u2192 Adapter Specific Identifiers. Click the button that says u2018Compare with System Landscape Directoryu2019 and chose Apply. Save and activate your change list.
    In case of business services, you can manually type a logical system name in the Adapter Specific Identifiers if required. This name should match the corresponding logical system name defined in the partner SAP systemu2019s partner profiles.
    Errors on the outbound side
    Sometimes the link between SAP XI and the target system (say ERP) goes down and messages fail on the outbound side. It may not be possible to restart them from using RWB or the transactions like SXI_MONITOR/SXMB_MONI. In such cases, you can follow the procedure outlined in the following article - Dealing with errors on the outbound side.
    Refer this article:
    http://help.sap.com/saphelp_nwpi71/helpdata/en/0d/28e1c20a9d374cbb71875c5f89093b/content.htm
    Refer this portal
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/ep/pointers%2bfor%2btroubleshooting%2bportal%2bruntime%2berrors
    Regards,
    Suryanarayana

  • Error while POF serialization

    Hi i am getting the following error while doing the POF serialization. please help
    java.lang.StackOverflowError
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)

    saiyansharwan wrote:
    Hi i am getting the following error while doing the POF serialization. please help
    java.lang.StackOverflowError
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    at com.tangosol.io.pof.PofBufferWriter$UserTypeWriter.isEvolvable(PofBufferWriter.java:2753)
    Hi saiyansharwan,
    which Coherence version is this?
    Best regards,
    Robert

  • Can I enable pof serialization for one cache and other JAVA serialization

    I had coherence cluster with few cache , Is there any way i can enable pof serialization for one cache and other to use normal JAVA serialization

    839051 wrote:
    I had coherence cluster with few cache , Is there any way i can enable pof serialization for one cache and other to use normal JAVA serializationHi,
    you can control serialization on a service-by-service basis. You can specify which serializer to use for the service with the <serializer> element in the service-scheme element corresponding to the service in the scache configuration file.
    Be aware, though, that if you use Coherence*Extend, and the service serializer configuration for the proxy service does not match the serializer configuration of the service which you are proxying to the extend client then the proxy node has to de- and reserialize the data which it moves between the service and the client.
    Best regards,
    Robert

  • POF serialization with replicated cache?

    Sorry again for the newbie question.
    Can you use POF serialized objects in a replicated cache?
    All of the examples show POF serialized objects being used with a partitioned cache.
    If you can do this, are there any caveats involved with the "replication" cache? I assume it would have to be started using the same configuration as the "master" cache.

    Thanks Rob.
    So, you just start up the Coherence instance on the (or at the) replication site, using the same configuration as the "master"? (Of course with appropriate classpaths and such set correctly)

  • Issue with POF serialization - Failure to deserialize an Invocable object: java.io.StreamCorruptedException

    Am running into following exception even after following all guidelines to implement POF. The main objective is to perform Distributed Bulk cache loading.
    Oracle Coherence GE 3.7.1.10 <Error> (thread=Invocation:InvocationService, member=1): Failure to deserialize an Invocable object: java.io.StreamCorruptedException: unknown user type: 1001
    java.io.StreamCorruptedException: unknown user type: 1001
      at com.tangosol.io.pof.PofBufferReader.readAsObject(PofBufferReader.java:3312)
      at com.tangosol.io.pof.PofBufferReader.readObject(PofBufferReader.java:2604)
      at com.tangosol.io.pof.ConfigurablePofContext.deserialize(ConfigurablePofContext.java:371)
      at com.tangosol.coherence.component.util.daemon.queueProcessor.Service.readObject(Service.CDB:1)
      at com.tangosol.coherence.component.net.Message.readObject(Message.CDB:1)
      at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.InvocationService$InvocationRequest.read(InvocationService.CDB:8)
      at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.deserializeMessage(Grid.CDB:19)
      at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.onNotify(Grid.CDB:31)
      at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
      at java.lang.Thread.run(Thread.java:662)
    Following is the pof-config.xml
    <?xml version="1.0" encoding="UTF-8" ?>
    <pof-config xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
                xmlns="http://xmlns.oracle.com/coherence/coherence-pof-config"
                xsi:schemaLocation="http://xmlns.oracle.com/coherence/coherence-pof-config/1.1/coherence-pof-config.xsd">
      <user-type-list>
        <include>coherence-pof-config.xml</include>
        <user-type>
          <type-id>1001</type-id>
          <class-name>com.westgroup.coherence.bermuda.loader.DistributedLoaderAgent</class-name>
          <serializer>
            <class-name>com.tangosol.io.pof.PofAnnotationSerializer</class-name>     
            <init-params>
              <init-param>
                <param-type>int</param-type>
                <param-value>{type-id}</param-value>
              </init-param>
              <init-param>
                <param-type>java.lang.Class</param-type>
                <param-value>{class}</param-value>
              </init-param>
              <init-param>
                <param-type>boolean</param-type>
                <param-value>true</param-value>
              </init-param>
            </init-params>
          </serializer>
        </user-type>
         <user-type>
          <type-id>1002</type-id>
          <class-name>com.westgroup.coherence.bermuda.profile.lpa.LPACacheProfile</class-name>
          <serializer>
            <class-name>com.tangosol.io.pof.PofAnnotationSerializer</class-name>     
            <init-params>
              <init-param>
                <param-type>int</param-type>
                <param-value>{type-id}</param-value>
              </init-param>
              <init-param>
                <param-type>java.lang.Class</param-type>
                <param-value>{class}</param-value>
              </init-param>
              <init-param>
                <param-type>boolean</param-type>
                <param-value>true</param-value>
              </init-param>
            </init-params>
          </serializer>
        </user-type>
         <user-type>
          <type-id>1003</type-id>
          <class-name>com.westgroup.coherence.bermuda.profile.lpa.Address</class-name>
          <serializer>
            <class-name>com.tangosol.io.pof.PofAnnotationSerializer</class-name>     
            <init-params>
              <init-param>
                <param-type>int</param-type>
                <param-value>{type-id}</param-value>
              </init-param>
              <init-param>
                <param-type>java.lang.Class</param-type>
                <param-value>{class}</param-value>
              </init-param>
              <init-param>
                <param-type>boolean</param-type>
                <param-value>true</param-value>
              </init-param>
            </init-params>
          </serializer>
        </user-type>
         <user-type>
          <type-id>1004</type-id>
          <class-name>com.westgroup.coherence.bermuda.profile.lpa.Discipline</class-name>
          <serializer>
            <class-name>com.tangosol.io.pof.PofAnnotationSerializer</class-name>     
            <init-params>
              <init-param>
                <param-type>int</param-type>
                <param-value>{type-id}</param-value>
              </init-param>
              <init-param>
                <param-type>java.lang.Class</param-type>
                <param-value>{class}</param-value>
              </init-param>
              <init-param>
                <param-type>boolean</param-type>
                <param-value>true</param-value>
              </init-param>
            </init-params>
          </serializer>
        </user-type>
         <user-type>
          <type-id>1005</type-id>
          <class-name>com.westgroup.coherence.bermuda.profile.lpa.Employment</class-name>
          <serializer>
            <class-name>com.tangosol.io.pof.PofAnnotationSerializer</class-name>     
            <init-params>
              <init-param>
                <param-type>int</param-type>
                <param-value>{type-id}</param-value>
              </init-param>
              <init-param>
                <param-type>java.lang.Class</param-type>
                <param-value>{class}</param-value>
              </init-param>
              <init-param>
                <param-type>boolean</param-type>
                <param-value>true</param-value>
              </init-param>
            </init-params>
          </serializer>
        </user-type>
      </user-type-list>
      <allow-interfaces>true</allow-interfaces>
      <allow-subclasses>true</allow-subclasses>
    </pof-config>
    cache-config.xml
    <cache-config xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
      xmlns="http://xmlns.oracle.com/coherence/coherence-cache-config"
      xsi:schemaLocation="http://xmlns.oracle.com/coherence/coherence-cache-config http://xmlns.oracle.com/coherence/coherence-cache-config/1.1/coherence-cache-config.xsd">
       <defaults>
        <serializer>pof</serializer>
        </defaults>
      <caching-scheme-mapping>
      <cache-mapping>
      <cache-name>DistributedLPACache</cache-name>
      <scheme-name>LPANewCache</scheme-name>
      <init-params>
      <init-param>
      <param-name>back-size-limit</param-name>
      <param-value>250MB</param-value>
      </init-param>
      </init-params>
      </cache-mapping>
      </caching-scheme-mapping>
      <caching-schemes>
      <!-- Distributed caching scheme. -->
      <distributed-scheme>
      <scheme-name>LPANewCache</scheme-name>
      <service-name>HBaseLPACache</service-name>
      <serializer>
      <instance>
                 <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
                 <init-params>
                   <init-param>
                     <param-type>java.lang.String</param-type>
                     <param-value>pof-config.xml</param-value>
                   </init-param>
                 </init-params>
            </instance>
      </serializer>
      <backing-map-scheme>
      <read-write-backing-map-scheme>
      <internal-cache-scheme>
      <class-scheme>
      <class-name>com.tangosol.util.ObservableHashMap</class-name>
      </class-scheme>
      </internal-cache-scheme>
      <cachestore-scheme>
      <class-scheme>
      <class-name>com.westgroup.coherence.bermuda.profile.lpa.LPACacheProfile</class-name>
      </class-scheme>
      </cachestore-scheme>
      <read-only>false</read-only>
      <write-delay-seconds>0</write-delay-seconds>
      </read-write-backing-map-scheme>
      </backing-map-scheme>
      <autostart>true</autostart>
      </distributed-scheme>
      <invocation-scheme>
           <scheme-name>InvocationService</scheme-name>
           <service-name>InvocationService</service-name>
           <thread-count>5</thread-count>
           <autostart>true</autostart>
        </invocation-scheme>
      </caching-schemes>
    </cache-config>
    DistributedLoaderAgent (user type 1001)
    import java.io.IOException;
    import java.io.Serializable;
    import java.lang.annotation.Annotation;
    import org.apache.log4j.Logger;
    import com.tangosol.io.pof.PofReader;
    import com.tangosol.io.pof.PofWriter;
    import com.tangosol.io.pof.PortableObject;
    import com.tangosol.io.pof.annotation.Portable;
    import com.tangosol.io.pof.annotation.PortableProperty;
    import com.tangosol.net.AbstractInvocable;
    import com.tangosol.net.InvocationService;
    @Portable
    public class DistributedLoaderAgent extends AbstractInvocable implements PortableObject{
      private static final long serialVersionUID = 10L;
      private static Logger m_logger = Logger.getLogger(DistributedLoaderAgent.class);
      @PortableProperty(0)
      public String partDumpFileName = null;
      public String getPartDumpFileName() {
      return partDumpFileName;
      public void setPartDumpFileName(String partDumpFileName) {
      this.partDumpFileName = partDumpFileName;
      public DistributedLoaderAgent(){
      super();
      m_logger.debug("Configuring this loader ");
      public DistributedLoaderAgent(String partDumpFile){
      super();
      m_logger.debug("Configuring this loader to load dump file "+ partDumpFile);
      partDumpFileName = partDumpFile;
      @Override
      public void init(InvocationService service) {
      // TODO Auto-generated method stub
      super.init(service);
      @Override
      public void run() {
      // TODO Auto-generated method stub
      try{
      m_logger.debug("Invoked DistributedLoaderAgent");
      MetadataTranslatorService service = new MetadataTranslatorService(false, "LPA");
      m_logger.debug("Invoking service.loadLPACache");
      service.loadLPACache(partDumpFileName);
      }catch(Exception e){
      m_logger.debug("Exception in DistributedLoaderAgent " + e.getMessage());
      @Override
      public void readExternal(PofReader arg0) throws IOException {
      // TODO Auto-generated method stub
      setPartDumpFileName(arg0.readString(0));
      @Override
      public void writeExternal(PofWriter arg0) throws IOException {
      // TODO Auto-generated method stub
      arg0.writeString(0, getPartDumpFileName());
    Please assist.

    OK I have two suggestions.
    1. Always create and flush the ObjectOutputStream before creating the ObjectInputStream.
    2. Always close the output before you close the input. Actually once you close the output stream both the input stream and the socket are closed anyway so you can economize on this code. In the above you have out..writeObject() followed by input.close() followed by out.close(). Change this to out.writeObject() followed by out.close(). It may be that something needed flushing and the input.close() prevented the flush from happening.

  • Int Server (ABAP Cache) - Cache Updated column shows "red"

    ...long read, appreciate your patience...
    In the IB:Config, under Cache Notifications, I'm getting the red square in the Cache Updated column for all "ABAP Cache" entries (green for notification).   The "Java Cache" and "Central Adapter Engine" entries are green all the way through.
    I've read a bunch of other threads here, and I've read and worked through the entire "How to Handle Caches" document.  I ran in to a couple issues, but I don't know how to solve them.....
    1) When I run SXI_CACHE, I receive the following messages:
    <green> "Cache contents are up to date"
    <red> "Error during last attempt to refresh cache"
    --->double-clicking shows: Error ID = BUSINESS_SYSTEM, Message = LCR_GET_OWN_BUSINESS_SYSTEM - NO_BUSINESS_SYSTEM
    --->Running the LCR... fxn via SE37 is successful, so I don't understand why SXI_CACHE has a problem running it.
    2) When I run the Cache Connectivity Test where the Yellow triangle is displayed for the IS-ABAP with message "Attempt to fetch cache data from Integration Directory not yet started or still in progress".
    3) When I try to load http://<host>:<port>/CPACache/refresh?mode=delta or full, I get a "403 Forbidden - You are not authorized to view the requested resource."  But I'm never prompted for a user/pw?  Since I'm the only user in this XI box, I do have some of my logons saved in Internet Explorer - could it be using one of those without prompting me?
    Note:
    I have been experiencing on and off network issues, and the machine we have XI installed on is slow.  I don't know if that could affect anything or not.
    Thanks for taking the time to read this...
    Brian
    Message was edited by: Brian Vanderwiel

    Thanks for the replies...
    Prashanth,
    I think SXI_CACHE is for items in the Java stack only (correct me if I'm wrong).  The main problem I have is with items in the ABAP stack not making it in to the cache.
    Moorthy,
    I walked through the Readiness Check - everything passed except for the items I already mentioned (Cache test shows yellow).  The document is a nice gathering of tests, but it doesn't offer any ideas has to how to correct problems encountered.
    Integration_directory_hmi tested successfully, and I read through the other OSS Notes, but none solved the issue.

  • Error occurred saving the cache

    I'm creating an AIR application using the Model Driven Development tools and hoping to make use of the offline cache. The generated value objects and services are working well when connected but I have been unable to get the cache working.
    Currently I am generating custom offline adaptors for each of my Model entities and assigning them to the offlineAdapter property of each data service as follows:
      _terminalService.serviceControl.autoCommit=false;
      _terminalService.serviceControl.autoConnect=true;
      _terminalService.serviceControl.fallBackToLocalFill=true;
      _terminalService.serviceControl.autoSaveCache=true;
      _terminalService.serviceControl.autoMerge=true;
      _terminalService.serviceControl.autoSyncEnabled=true;
      _terminalService.serviceControl.encryptLocalCache=false;
      _terminalService.serviceControl.offlineAdapter=new TerminalOfflineAdapter();
      _terminalService.serviceControl.cacheID="ss-sclient-39";
    When I run my application it works as anticipated but I am seeing the following error in the console log:
    7/27/2010 12:21:25.677 [DEBUG] mx.data.DataStore Saving cached query: store id: ["getByMacAddress","001CC4313675"] lastWritten: Tue Jul 27 12:21:25 GMT+0100 2010 lastAccessed: Tue Jul 27 12:21:25 GMT+0100 2010 created: Tue Jul 27 12:21:25 GMT+0100 2010 metadata: (null) type: 1 referenced ids: ["common.Terminal:#:13"]
    7/27/2010 12:21:25.677 [DEBUG] mx.data.LSODataStore dbStore lock released: 965366152
    7/27/2010 12:21:25.677 [ERROR] mx.data.DataStore Error occurred saving the cache: TypeError: Error #1009: Cannot access a property or method of a null object reference. stack: TypeError: Error #1009: Cannot access a property or method of a null object reference.
         at mx.data.offline::EntityManager/save()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\offline\EntityManager.as:104]
         at mx.data.offline::EntityManager/update()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\offline\EntityManager.as:345]
         at mx.data::SQLiteOfflineAdapter/updateOfflineItems()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\SQLiteOfflineAdapter.as:120]
         at mx.data::DataStore/persistCacheItems()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4543]
         at mx.data::DataStore/persistDataService()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4679]
         at mx.data::DataStore/doSaveCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:3755]
         at Function/()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:3711]
         at mx.data::DataStore/http://www.adobe.com/2006/flex/mx/internal::saveCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:3727]
         at mx.data::DataList/http://www.adobe.com/2006/flex/mx/internal::processSequenceResult()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataList.as:3063]
         at mx.data::DataListRequestResponder/result()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataListRequestResponder.as:103]
         at mx.rpc::AsyncRequest/acknowledge()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\AsyncRequest.as:84]
         at NetConnectionMessageResponder/resultHandler()[E:\dev\4.x\frameworks\projects\rpc\src\mx\messaging\channels\NetConnectionChannel.as:547]
    If I then disconnect from the server and attempt to run my client again I get the following error:
    7/27/2010 12:27:46.548 [DEBUG] mx.data.DataService.common.Terminal DataService.fill() called for destination: common.Terminal with args: ["getByMacAddress","001CC4313675"] includesProperties: (include default - excludes: [])
    7/27/2010 12:27:46.548 [INFO] mx.messaging.Producer 'ds-producer-common.Terminal' producer sending message 'EF832901-F3A4-7AE3-EF12-13A8897408CC'
    TypeError: Error #1009: Cannot access a property or method of a null object reference.
         at mx.data.offline::EntityManager/executeSelectQuery()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\offline\EntityManager.as:446]
         at mx.data.offline::EntityManager/executeQuery()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\offline\EntityManager.as:432]
         at mx.data::SQLiteOfflineAdapter/internalExecuteOfflineQuery()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\SQLiteOfflineAdapter.as:318]
         at mx.data::SQLiteOfflineAdapter/executeOfflineQuery()[C:\depot\DataServices\trunk\frameworks\projects\airfds\src\mx\data\SQLiteOfflineAdapter.as:202]
         at Function/()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:1651]
         at mx.data::DataStore/http://www.adobe.com/2006/flex/mx/internal::fill()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:1765]
         at mx.data::ConcreteDataService/internalFill()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\ConcreteDataService.as:7235]
         at Function/()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\ConcreteDataService.as:1317]
         at mx.data::ConcreteDataService/fill()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\ConcreteDataService.as:1336]
         at mx.data::DataManager/fill()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataManager.as:1560]
         at _Super_TerminalService/getByMacAddress()[C:\Documents and Settings\benp.NB\My Documents\src\qms\workspace\CommonClient\src\common\model\_Super_TerminalService.as:226]
         at airclient.application::AirClientImpl/initialiseTerminalDetails()[C:\Documents and Settings\benp.NB\My Documents\src\qms\workspace\AirClient\src\airclient\application\AirClientImpl.as:146]
         at Function/()[C:\Documents and Settings\benp.NB\My Documents\src\qms\workspace\AirClient\src\airclient\application\AirClientImpl.as:268]
         at mx.collections::ItemResponder/result()[E:\dev\4.x\frameworks\projects\framework\src\mx\collections\ItemResponder.as:129]
         at mx.rpc::AsyncToken/http://www.adobe.com/2006/flex/mx/internal::applyResult()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\AsyncToken.as:239]
         at mx.rpc.events::ResultEvent/http://www.adobe.com/2006/flex/mx/internal::callTokenResponders()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\events\ResultEvent.as:207]
         at mx.data::ConcreteDataService/http://www.adobe.com/2006/flex/mx/internal::dispatchResultEvent()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\ConcreteDataService.as:3427]
         at Function/http://adobe.com/AS3/2006/builtin::apply()
         at mx.rpc::AsyncDispatcher/timerEventHandler()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\AsyncDispatcher.as:50]
         at flash.utils::Timer/_timerDispatch()
         at flash.utils::Timer/tick()
    I'm slightly confused by the fact that the offline adaptors are derived from SQLiteOfflineAdapter and are being run within an AIR application but I am still seeing references to the LSODataStore in the logs (which I thought was only used for Flash applications). Also if I attempt to set encryptLocalCache to true I get the following error:
    7/27/2010 12:32:10.409 [DEBUG] mx.data.DataStore Adding data service: common.Terminal to the data store: my-rtmp:true initialized: false
    Error: Encryption is not supported for local shared objects.
         at mx.data::LSODatabase/set encryptLocalCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\LSODatabase.as:149]
         at mx.data::DataStore/set encryptLocalCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:391]
         at mx.data::ConcreteDataService/set encryptLocalCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\ConcreteDataService.as:227]
         at mx.data::DataManager/set encryptLocalCache()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataManager.as:392]
         at airclient.application::AirServiceFactory/configureService()[C:\Documents and Settings\benp.NB\My Documents\src\qms\workspace\AirClient\src\airclient\application\AirServiceFactory.as:236]
    Does anyone know what's going on here?
    Thanks!

    Ah, thanks! That certainly solves the LSO issue - I've changed to linking with airfds.swc and I'm now seeing lots of healthy looking SQLDBCache output in the logs when I run the application with a connection. Unfortunately if I then stop the server and run the application again I'm still seeing a problem. The app does load some data from the cache (I have several services) but then stops with the following error before it can load everything:
    7/27/2010 15:43:46.929 [DEBUG] mx.data.SQLDBCache SQLDBCache - getCollection(_common.Customer_ic) creating new collection
    7/27/2010 15:43:46.929 [DEBUG] mx.data.SQLDBCache SQLDBCache - before load data - create table if not exists: CREATE TABLE IF NOT EXISTS [_common_Customer_ic] (id VARCHAR PRIMARY KEY, data BLOB);
    7/27/2010 15:43:46.960 [DEBUG] mx.data.SQLDBCache SQLDBCache - after load data - create table if not exists
    7/27/2010 15:43:46.960 [DEBUG] mx.data.SQLDBCache SQLDBCache - before select: SELECT id,data FROM [_common_Customer_ic]
    TypeError: Error #1009: Cannot access a property or method of a null object reference.
         at mx.data::DataStore/restoreReferencedIds()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4766]
         at mx.data::DataStore/restoreReferencedItems()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4736]
         at mx.data::DataStore/restoreAssociations()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4880]
         at mx.data::DataStore/restoreReferencedItems()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:4759]
         at Function/()[C:\depot\DataServices\trunk\frameworks\projects\data\src\mx\data\DataStore.as:1742]
         at mx.rpc::Responder/result()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\Responder.as:56]
         at mx.rpc::AsyncToken/http://www.adobe.com/2006/flex/mx/internal::applyResult()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\AsyncToken.as:239]
         at mx.rpc.events::ResultEvent/http://www.adobe.com/2006/flex/mx/internal::callTokenResponders()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\events\ResultEvent.as:207]
         at Function/http://adobe.com/AS3/2006/builtin::apply()
         at mx.rpc::AsyncDispatcher/timerEventHandler()[E:\dev\4.x\frameworks\projects\rpc\src\mx\rpc\AsyncDispatcher.as:50]
         at flash.utils::Timer/_timerDispatch()
         at flash.utils::Timer/tick()
    In both connected and disconnected operation there is no user input, so the sequence of queries should be identical. Each service I use is configured identically except that they each use their own automatically-generated OfflineAdaptor. Any suggestions?
    Many thanks

  • I'm getting error messages when I download updates for lightroom and flash for the mac. can't figure out how to trouble shoot.

    I'm getting error messages when I download updates for lightroom and flash for the mac. can't figure out how to trouble shoot.

    It sounds as if your browser and some of your apps are not color-managed.  This is pretty typical.  Even IE9 is only partially color-managed.
    You can expect color-managed and non-color-managed applications to show you different things with the same images.  How different will depend upon how different your monitor color profile is from the image's color profile.
    For web publication and most general use, experts usually advise saving images with the sRGB profile.  If such images, saved through the Save for Web & Devices function, look different to you than you expect, it may be that your input images have previously been saved in another color space.
    You should really try to get your head around color-management by reading more on it.  It can seem baffling and it's difficult to understand without some background.  A quick web search turns up many overviews.  Beware, though, even people writing articles sometimes don't fully understand it.
    -Noel

  • Supplier collab BP : Cache update

    Hi,
    In Supplier collaboration Business Package Cache update for master data
    is not working.
    Jco is created and configured.
    Had a look in to the Backend System ECC and found from the trace that
    Jco RFC user is executing the functional module but it is not updating
    in the Portal.
    Error in FI customization says" Could not read invoice document type
    settings from database; contact the system administrator"
    Please provide me some solution on this.

    Would anyone have any suggestion for this problem.

  • Cache update / Missing objects

    Dear all!
    Since two days we have following problem:
    After activating changes in integration directory sometimes the changes are not transmitted to the Runtime Cache, sometimes they are. The Status is everytime "green". A manually (full and delta) refresh is not succesfull (means status is still green, no error message is displayed but the objects are still missing). I.e. i can see only five of eight configured communication channels.
    I read the thread "Automtic Cache Update is not done !!
    Any ideas?
    Kind regards,
    Klaus Leng

    Hi Klaus,
    Which parts are showing up yellow.
    For me it was the Integration Engine - adapter component.
    In my case I had to reset the password for the SAP XI AF CPA cache component in the visual administrator.
    Also make sure you haven't changed any of the standard users in the exchangeProfile. Because they didn't have enough authorisations I once changed them over there, and that also caused similar problems.
    Best regards
    Alwin

  • Issues with cache updates

    Hi ,
    I have some issues with runtime xi cache update.
    i tested the RFC dest for connection RFC Dest - INTEGRATION_DIRECTORY_HMI.
    that failed with HTTP 500 internal server error .
    I even tried to update the cache from TCode - sxi_cache
    the error i got there was
    Prefix number: entry missing for system XI1 client 200
    If anyone has any idea as to why is such an error occuring
    Thanks
    Nikhil

    Hi
    <i>that failed with HTTP 500 internal server error .</i>
    is not an error
    Please can you check on this thread:
    XI Runtime cache error: Cache contents are obsolete
    Regards
    Vijaya

  • I now get an 500 Internal Server error when loading Firefox, after updating RealPlayer

    I now get an 500 Internal Server error when loading Firefox, after updating RealPlayer.
    When I try to open Firefox, it opens. But there is the error message.
    What is this? How do I did rid of it?
    '''bold text'''

    Clear the cache and the cookies from sites that cause problems.
    * "Clear the Cache": Tools > Options > Advanced > Network > Offline Storage (Cache): "Clear Now"
    * "Remove the Cookies" from sites causing problems: Tools > Options > Privacy > Cookies: "Show Cookies"
    Start Firefox in [[Safe Mode]] to check if one of the add-ons is causing the problem (switch to the DEFAULT theme: Tools > Add-ons > Appearance/Themes).
    * Don't make any changes on the Safe mode start window.
    See:
    * [[Troubleshooting extensions and themes]]
    You can also check the connection settings in Tools > Options > Advanced : Network : Connection<br />
    If you do not need to use a proxy to connect to internet then select "No Proxy"
    See "Firefox connection settings":
    * [[Firefox cannot load websites but other programs can]]

  • Serialization error in inbound idoc

    Hi Friends
    Am getting inbound idocs with this error : Serialization errror for object 01,S, 34343434,Expected counter 000001<
    and error in details says,
    The expected serialization counter has the value 000001. However, the serialization counter in the IDoc has the value 000002 and is therefore too big. There are therefore older IDocs with this HR object with the serialization counter values in between. These IDocs have either not yet been posted, or have been posted incorrectly.
    Intrested Part is   i put check ( Selected ) on  as per client suggestion in ,
    IMG -> Sap Netweaver -> Application server ->Idoc Interface / Application Link Enabling (ALE ) -> Modelling and imple  Business process -> Configure  Predefine  ALE Business Process -> Human Resource -> Master data Distribution ->Serialize HR Master Data  -> ( Serialize HR Master data in ALE inbound processing ) ...   settings  and i was getting this error but when i take back this check ( unselect ) flag   idoc posted successfully    Now am confuse i shd keep this or not because am getting data from diff countries also.
    what shd i do for this few expert said i need to do some thing for Serialization so automatically system shd take care for this,
    Need your support on this .
    Regards
    Meeta

    I have been working with IDocs in a serilization-active HR-ALE system and have learned a few things along the way ....
    - Serialization must be active in both sending/originating and receiving/destination systems.  (Turn on flag in customizing, as you have already done).
    - Both systems maintain a registry of HR objects that includes in each record the object type, object id, and serial counter.  The serial counter is increased by one in the sending system every time the object is sent via ALE and increased by one when the object is correctly processed in the receiving system.  The serial counter is passed in the E1PLOGI record of each object in the IDoc.
    - When a similar serialization error is received (where the counter received is lower than the expected counter), the error message gives a little more detail:   "Decide what to do based on your own particular situation. Responses could range from ignoring or editing the IDoc, to manually adjusting the serialization counter in the registry of the receiving system to the expected value."
    Options are therefore:
      - Edit IDoc - can use IDoc test tool (WE19) to create and edit a copy for processing.
      - Use transaction RE_RHALE_HRMDRGIN to maintain the serial counters
      - Use transaction RE_RHALE_RGIN2IDOC to adjust the serial counters for the offending IDoc
    I strongly recommend, however, that you take some time to further understand the serialization process before updating the counters manually.  If you've turned on, then off, then on the serialization flag(s) through customizing, you might consider initializing the serial counters in both sending and receiving systems and starting from scratch.

  • POF serialization testing

    hi
    I would like to write an integration test for POF serialization.
    Is it possible to make Coherence to serialize all objects I put into named cache ?
    Or I need to start two nodes (in two classloaders) in my test ?

    If you want to just test the POF serialization of your classes you can just serialize, deserialize and check the results
    e.g.
    MyPofClass original = new MyPofClass();
    // populate fields of original
    // Serialize to Binary
    ConfigurablePofContext pofContext = new ConfigurablePofContext("name of pof config file .xml");
    Binary binary = ExternalizableHelper.toBinary(original, pofCOntext);
    // Deserialize back to original class
    MyPofClass deserialized = ExternalizableHelper.fromBinary(binary, pofContext);
    // now do some asserts to check that the fields in the original match the fields in the deserialized instanceI usually do that test for all but the most simple POF classes I write.
    JK

Maybe you are looking for

  • Issue in creating Invoice when route is not relevant for shipment cost.

    Dear All, We are facing a problem while creating an Invoice. The route used in shipment in not relevant for shipment costing. Shipment type is relevant for costing. So in normal scenario if I don't remove the tick mark from "Shipment cost relanace" a

  • Problem in Pricing Condition Type : In Service Confirmation

    Hi, I am trying to create Service Confirmation from Sevice Order. I am tring to edit service Item details. I assigned valuation type as "Saturday". As per the actual result, one condition type(ZS13) for surcharge should get added in Item Pricing deta

  • Custom Error handling in DataSheet View with pop-up message or label controls or HTML

    I have a requirement in which DataSheet view is used to bulk edit or input list items and handle validations while in the process. We have validations done in event recievers before creating the list item. But could not provide a pop up or message wi

  • Removing the preview pane in iCloud

    I didn't know where to send the suggestion for this or if I'm just not looking in the right place for it. I would like to be able to remove the preview pane in my iCloud mail. It's really the biggest thing keeping me from using iCloud for email; I do

  • Delete account taking a long time

    Hi, yesterday evening I try to delete my account, but the process is still running, I read on the internet that this process takes much time (30 hours, 2 days) Is it normal? If it is not normal behavior how can I solve it ? I look forward your reply