Data Caching in PI

Hello all,
I have an interesting client requirement regarding PI caching datasets which it can query in order to minimize amount of hits to the SAP database. These datasets would be refreshes in a specified batch.
For example, there is a requirement to query SAP BOM tables from an external system in order to find out a value of a specific attribute. Instead of querying SAP database each time, the client would like PI to cache datasets (which would be refreshed every night), then use those datasets to provide values to queries coming from the external system.
Is anyone aware whether PI supports such requirement?
Thank you all in advance.
Alex

Hi Alex,
there is no functionality caching such data for a specific time in PI.
From the performance perspective it is the best to read the BOM data in the message receiver system, not via database look-ups in PI.
If the delivered message needs further processing (enrichment with data from the receiver system), the receiver system is responsible for that specific processing/enrichment (e.g. using IDOC -> build own IDOC inbound processing function module).
There should no business logic excluded BPM in PI.
Best regards,
André

Similar Messages

  • Error in Starting Oracle BAM Active Data Cache

    I am not able to start "Oracle BAM Active Data Cache" on my machine.
    The other two components "Oracle BAM Event Engine" and "Oracle BAM Report Cache" are starting properly.
    When I see the event log file of my Computer I could see the details as below:
    Event Type: Error
    Event Source: Oracle BAM Active Data Cache
    Event Category: None
    Event ID: 0
    Date: 2/7/2007
    Time: 3:51:25 PM
    User: N/A
    Computer: CHNANDA-WXP
    Description:
    ActiveDataCache: The Oracle BAM Active Data Cache service failed to start. Oracle.BAM.ActiveDataCache.Common.Exceptions.CacheException: ADC Server exception in Startup(). ---> Oracle.DataAccess.Client.OracleException ORA-12541: TNS:no listener at Oracle.DataAccess.Client.OracleException.HandleErrorHelper(Int32 errCode, OracleConnection conn, IntPtr opsErrCtx, OpoSqlValCtx* pOpoSqlValCtx, Object src, String procedure)
    at Oracle.DataAccess.Client.OracleConnection.Open()
    at Oracle.DataAccess.Client.OracleConnection.Open()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetServerVersion()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.Startup(IDictionary oParameters)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    --- End of inner exception stack trace ---
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    at Oracle.BAM.ActiveDataCache.Kernel.Server.Server.Startup()
    at Oracle.BAM.ActiveDataCache.Service.DataServer.Run()
    For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp.
    Could anyone pls help me?
    Thanks and Regards,
    Chinmaya Nanda

    hi Chinmaya -can yoy tell us your companyname,project ? Your problem is very simple.BAM ADC is notable to reachoracle db.fromyour dos prompt- try tnsping <yrDB> [default  is oraclebam  or orcl  ]/ Also see FAQ pages. there is a requirement on dos prompt setting, with <clientforBAM>as 1st parameter

  • Web Application - Data caching of enterprise data

    Sorry in advance if this is off-topic but I can't find anywhere else to post this type of question.
    I am looking for information/suggestions such as books, technology or design methodology for my enterprise web applications. These sites are currently up and functional using only JSP, servlets and regular Java classes stored in a web application session to provide data caching and access. We are using Weblogic Server 6.1 running on an AIX Unix system at this time. I'm not sure that this is the best design architecture as our web sessions are getting too large but I can't think of any other Java technology to use and I need some help. Here's an overview of our environment and our needs.
    Our core data is held in a mainframe based IMS system. Some DB2 is also used. Access to this data is through IMS COBOL transactions which we can execute with IMS Connect. We also use some JDBC to get to the DB2 tables directly where available.
    Some overall application data is cached when the web application is deployed. We use singleton classes which are created and refreshed at deployment and they then refresh themselves from the sources every 24 hours.
    Each time a user logs in we execute several IMS transactions and JDBC calls to cache user specific data in regular Java classes which are then simply placed in the users web session where we access them from JSP, servlets and other Java classes. The fields in these Java classes range from any type of primitive field to TreeMaps of other Java classes. As the data is cached it is sorted and other fields are calculated and stored in these classes. As the user progresses through the system we then may have to do several other IMS transaction and JDBC calls to collect other types of data. All of this is then also added to the users session. Most of this in inquiry. We do allow transactions but those are built from user input and data already cached and are then we just execute the IMS transactions with the input.
    As our application has grown these Java classes have gotten larger and larger. And since these are simply stored in server memory in the web sessions then these are also getting huge. I'm concerned that this is not the best way for this application to be architected. Is there something else we should be doign? I simply don't understand how Entity Java Beans could be used but then again I don't know much about them. I wouldn't think that caching the data to a local database and accessing it from there would be any more efficient and would probably just slow down the system from all the I/O.
    Any help or direction would be greatly appreciated.

    The best book you can buy is 'Professional Java Server Programming, j2ee edition' by Wrox. It is by far the best reference I've used. Another quick reference consideration might be the j2ee book provided by codenotes... its quick and to the point.

  • Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).

    Hi,
    Our Environment is Essbase 11.1.2.2 and working on Essbase EAS and Shared Services components.One of our user tried to run the Cal Script of one Application and faced this error.
    Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).
    I have done some Google and found that we need to add something in Essbase.cfg file like below.
    1012704 Dynamic Calc processor cannot lock more than number ESM blocks during the calculation, please increase CalcLockBlock setting and then retry (a small data cache setting could also cause this problem, please check the data cache size setting).
    Possible Problems
    Analytic Services could not lock enough blocks to perform the calculation.
    Possible Solutions
    Increase the number of blocks that Analytic Services can allocate for a calculation:
    Set the maximum number of blocks that Analytic Services can allocate to at least 500. 
    If you do not have an $ARBORPATH/bin/essbase.cfg file on the server computer, create one using a text editor.
    In the essbase.cfg file on the server computer, set CALCLOCKBLOCKHIGH to 500.
    Stop and restart Analytic Server.
    Add the SET LOCKBLOCK HIGH command to the beginning of the calculation script.
    Set the data cache large enough to hold all the blocks specified in the CALCLOCKBLOCKHIGH setting. 
    Determine the block size.
    Set the data catche size.
    Actually in our Server Config file(essbase.cfg) we dont have below data  added.
    CalcLockBlockHigh 2000
    CalcLockBlockDefault 200
    CalcLockBlocklow 50
    So my doubt is if we edit the Essbase.cfg file and add the above settings and restart the services will it work?  and if so why should we change the Server config file if the problem is with one application Cal Script. Please guide me how to proceed.
    Regards,
    Naveen

    Your calculation needs to hold more blocks in memory than your current set up allows.
    From the docs (quoting so I don't have to write it, not to be a smarta***:
    CALCLOCKBLOCK specifies the number of blocks that can be fixed at each level of the SET LOCKBLOCK HIGH | DEFAULT | LOW calculation script command.
    When a block is calculated, Essbase fixes (gets addressability to) the block along with the blocks containing its children. Essbase calculates the block and then releases it along with the blocks containing its children. By default, Essbase allows up to 100 blocks to be fixed concurrently when calculating a block. This is sufficient for most database calculations. However, you may want to set a number higher than 100 if you are consolidating very large numbers of children in a formula calculation. This ensures that Essbase can fix all the required blocks when calculating a data block and that performance will not be impaired.
    Example
    If the essbase.cfg file contains the following settings:
    CALCLOCKBLOCKHIGH 500  CALCLOCKBLOCKDEFAULT 200  CALCLOCKBLOCKLOW 50 
    then you can use the following SET LOCKBLOCK setting commands in a calculation script:
    SET LOCKBLOCK HIGH; 
    means that Essbase can fix up to 500 data blocks when calculating one block.
    Support doc is saying to change your config file so those settings can be made available for any calc script to use.
    On a side note, if this was working previously and now isn't then it is worth investigating if this is simply due to standard growth or a recent change that has made an unexpected significant impact.

  • Xi Data Cache Refresh Problem

    Hello,
    We have following Problem:
    We send an XML-Request with HTTP-Adapter to XI and get Data from R/3, works fine.
    But when we send a Second Request we become an Error.
    The error comes because in the Java Runtime the XI writes wrong values into Message.
    When we do a XI Data Cache Refresh for INTEGRATION_SERVER
    and INTEGRATION_RUNTIME we can send a message and all works fine again.´
    Any Idea why the XI Caches Data from Request and set this values in the message of request that was send later?
    Thanks,
    Robin

    Hi,
    I think both are same. Because in SXi_ADM we also define our IS. so i think it may be same ...
    If I am wrong please share your knowledge . in this thread for more clarification ..
    Regards,
    Amit

  • Error in Starting Active Data Cache

    Hi All,
    I am trying to start the BAM services, all other services are getting started except 'Active Data Cache'.
    Its erroring out with the following msg in the log....
    Could any one look in to this issue......Thanks in Advance.
    here is the log msg
    2009-01-16 15:56:56,538 [10688] ERROR - ActiveDataCache DPAPI was unable to decrypt data. CryptUnprotectData failed. Error -2146893813: Key not valid for use in specified state.
    2009-01-16 15:56:56,538 [10688] WARN - ActiveDataCache Exception occurred in method Startup
    Stack trace:
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(Byte[] cipherTextBytes, Byte[] entropyBytes, String& description)
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(String cipherText, String entropy, String& description)
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(String cipherText)
    at Oracle.BAM.Common.Security.DataProtector.DataProtector.Decrypt(String strData)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.ConnectionStringDecrypter.Decrypt(String strEncrypted)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleDataFactory.GetInstance()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetDataFactory()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetServerVersion()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.Startup(IDictionary oParameters)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    2009-01-16 15:56:56,553 [10688] ERROR - ActiveDataCache The Oracle BAM Active Data Cache service failed to start. Oracle.BAM.ActiveDataCache.Common.Exceptions.CacheException: ADC Server exception in Startup(). ---> System.Exception: DPAPI was unable to decrypt data. CryptUnprotectData failed. Error -2146893813: Key not valid for use in specified state.
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(Byte[] cipherTextBytes, Byte[] entropyBytes, String& description)
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(String cipherText, String entropy, String& description)
    at Oracle.BAM.Common.Security.DataProtector.DPAPI.Decrypt(String cipherText)
    at Oracle.BAM.Common.Security.DataProtector.DataProtector.Decrypt(String strData)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.ConnectionStringDecrypter.Decrypt(String strEncrypted)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleDataFactory.GetInstance()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetDataFactory()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetServerVersion()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.Startup(IDictionary oParameters)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    --- End of inner exception stack trace ---
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    at Oracle.BAM.ActiveDataCache.Kernel.Server.Server.Startup()
    at Oracle.BAM.ActiveDataCache.Service.DataServer.Run()

    Hi.
    I encountered the same problem. In my case, the reason was that I created a new user for BAM (to be provided when installing BAM) but ran the installation under another user account. It seems like these users must be the same to make it work.
    Greetings,
    cor

  • Data Cache Error

    Hi Experts,
    Im new in this forum and please tell me if Im in the wrong Forum.
    I think the problem Im facing is a data cache one.
    The Scenario is a proxy -> Web Sevice. Synchronous.
    The Symptoms are: problems in the ESR. when im testing my scenario I get diferent errors depending on the refresh of the cache.
    If I refresh the cache in the ABAP stack transaction sxi_cache. the error showed is: Operation Mapping XXXXXXXXX Runtime error Mapping program is not available in runtime cache
    If I refresh the cache in the JAVA stack in the administration tool the error showed is:  Receiver Determination  Runtime error Error while refreshing the XI runtime cache
    I never faced this before and cant find it in any other posts.
    Thanks in advance and I hope you can help me.
    Agus.

    Hi Edgarzino,
    Please take a look at this guides, they hace useful topics:
    ´´´´ How to handle caches in XI
    (NW 2004, XI 3.0)
        https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/1a69ea11-0d01-0010-fa80-b47a79301290
    (NW 2004s, PI 7.0)
        https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/c0332b2a-eb97-2910-b6ba-dbe52a01be34

  • ADC exception=BAM-01262: Active Data Cache server exception in openViewset

    Hi All,
    I am getting the below mentioned error when i try to open a report from Active Studio in BAM.
    If i give parameter value ' All ' in the prompt, it shows all the value. But if i give specific values like one parameter value 1007 and other All etc, it gives the below error. I tried giving parameter value ' All ' in the prompt after getting the error, its not showing the values.
    The below error is from front end:-
    SQL EXCEPTION NULL : Invalid column index
    The below error from back end:-
    [2012-01-10T06:39:07.916+05:30] [bam_server1] [ERROR] [] [oracle.bam.reportcache] [tid: [ACTIVE].ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: weblogic] [ecid: 83cdf926093045b8:4fde7fa7:134bf505061:-8000-00000000000044f4,0] [APP: oracle-bam#11.1.1] *ReportCache: ReportCacheServer.OpenViewSet: ADC exception=BAM-01262: Active Data Cache server exception in openViewset(). [[*
    at oracle.bam.adc.kernel.util.Util.getCacheException(Util.java:101)
    at oracle.bam.adc.kernel.util.Util.getCacheException(Util.java:154)
    at oracle.bam.adc.kernel.util.Util.getCacheException(Util.java:172)
    at oracle.bam.adc.kernel.server.DataStoreServer.openViewset(DataStoreServer.java:1110)
    at oracle.bam.adc.ejb.BamAdcServerBean.openViewset(BamAdcServerBean.java:841)
    at sun.reflect.GeneratedMethodAccessor702.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at com.bea.core.repackaged.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:310)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)
    at com.bea.core.repackaged.springframework.jee.intercept.MethodInvocationInvocationContext.proceed(MethodInvocationInvocationContext.java:104)
    at oracle.bam.adc.ejb.BamAdcServerBean.interceptor(BamAdcServerBean.java:266)
    at sun.reflect.GeneratedMethodAccessor374.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    [2012-01-10T06:39:07.913+05:30] [bam_server1] [ERROR] [] [oracle.bam.adc] [tid: [ACTIVE].ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: weblogic] [ecid: 83cdf926093045b8:4fde7fa7:134bf505061:-8000-00000000000044f4,0] [APP: oracle-bam#11.1.1] ActiveDataCache: Exception occurred in method openViewset(_TPT_PO_HEADER_STG_TBL,0)[[
    Exception: java.sql.SQLException: SQLError(17003) SQLState(99999) Invalid column index
    at oracle.jdbc.driver.OraclePreparedStatement.setStringInternal(OraclePreparedStatement.java:6336)
    at oracle.jdbc.driver.OraclePreparedStatement.setObjectCritical(OraclePreparedStatement.java:10605)
    at oracle.jdbc.driver.OraclePreparedStatement.setObjectInternal(OraclePreparedStatement.java:10518)
    at oracle.jdbc.driver.OraclePreparedStatement.setObjectInternal(OraclePreparedStatement.java:11574)
    at oracle.jdbc.driver.OraclePreparedStatement.setObject(OraclePreparedStatement.java:11544)
    at oracle.jdbc.driver.OraclePreparedStatementWrapper.setObject(OraclePreparedStatementWrapper.java:249)
    at oracle.bam.adc.common.externaldatasources.JDBC.getDataReader(JDBC.java:322)
    at oracle.bam.adc.kernel.datasets.ExternalStorageEngine.getDataReader(ExternalStorageEngine.java:76)
    at oracle.bam.adc.kernel.viewsets.utilities.externaldata.DataImporter.executeQuery(DataImporter.java:95)
    at oracle.bam.adc.kernel.viewsets.utilities.externaldata.ExternalDataManager.importExternalData(ExternalDataManager.java:228)
    at oracle.bam.adc.kernel.viewsets.utilities.externaldata.ExternalDataManager.importExternalFactData(ExternalDataManager.java:151)
    at oracle.bam.adc.kernel.viewsets.utilities.externaldata.ExternalDataManager.getExternalData(ExternalDataManager.java:103)
    at oracle.bam.adc.kernel.viewsets.Viewset.loadData(Viewset.java:259)
    at oracle.bam.adc.kernel.viewsets.ViewsetBase.initialize(ViewsetBase.java:171)
    at oracle.bam.adc.kernel.viewsets.Viewset.initialize(Viewset.java:220)
    at oracle.bam.adc.kernel.viewsets.ViewsetBase.open(ViewsetBase.java:154)
    Please give me any suggestion on this.
    Thanks,
    Manikandan

    This problem was solved when I installed a new product version.

  • Error while Enabling Active Data Cache Clustering Services: error 1753

    I have installed the Oracle BAM with Enterprise link.All the installation was successful.But while Enabling Active Data Cache Clustering Services using the following command:
    cluster.exe restype "Oracle Business Activity Monitoring Active Data Cache" /create /dll:"C:\OracleBAM\ADCClusterResourceType.dll"
    i am getting the folllowing error:
    System error 1753 has occurred.
    There are no more endpoints available from the endpoint mapper.
    Although i stopped the Active Data Cache Service..?
    I am unable to resolve it
    Kindly help me out
    Ramesh Nambala

    Hi.
    I encountered the same problem. In my case, the reason was that I created a new user for BAM (to be provided when installing BAM) but ran the installation under another user account. It seems like these users must be the same to make it work.
    Greetings,
    cor

  • BAM Report and Active data cache

    I am having problems with reports that have automatic active data retrieval.
    It does not work.
    If I load the report it extracts the data correctly from an external data source. If I Reprompt or refresh it brings back changed data from the data source.
    It does not automatically poll for changes even though it is set in the report properties.
    If I stop and start the Active data cache service it starts polling and reconnecting, but not resynching the data.
    In the Active data cache log I have the following entries :
    2005-12-01 09:36:54,062 [3576] ERROR - ActiveDataCache Viewset not found:
    2005-12-01 09:36:54,078 [3576] WARN - ActiveDataCache Exception occurred in method GetChangeList
    Stack trace:
    at Oracle.BAM.ActiveDataCache.Kernel.Viewsets.ViewsetManager.GetViewset(String strViewsetID)
    at Oracle.BAM.ActiveDataCache.Kernel.Viewsets.ViewsetManager.GetChangeList(String strViewsetID, Int32 iTimeout)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.GetChangeList(String strViewsetID, Int32 iTimeout)
    Has anyone any ideas on what the problem is ?
    Thanks

    Sriram
    Can you identify your company and project.?
    a) Maximum capacity of ADC == maximum capacity of your underlying Oracle DB used by BAM as repository.
    b) Retention period - infinite - data is there in BAM ADC (also in BAM repository) till you manually delete it (there is no automatic delete of old data)
    c) Yes ADC stores its data internally in the Oracle database (aka BAM repository)
    d) No - we generally donot publish ADC internals. This is a known decision since end users will only work either to put data into ADC or plot reports/dashboards from BAM GUI. You have to be very specific on what your are trying to acheive and we can suggest possible alternates.

  • BAM Active data Cache Capacity

    what is the maximum capacity of the active data cache?
    what is the retention period of active data cache?
    does ADC store the data in the internal oracle data base?
    Please provide the documents for the internal implementation of ADC?
    Sriram.S

    Sriram
    Can you identify your company and project.?
    a) Maximum capacity of ADC == maximum capacity of your underlying Oracle DB used by BAM as repository.
    b) Retention period - infinite - data is there in BAM ADC (also in BAM repository) till you manually delete it (there is no automatic delete of old data)
    c) Yes ADC stores its data internally in the Oracle database (aka BAM repository)
    d) No - we generally donot publish ADC internals. This is a known decision since end users will only work either to put data into ADC or plot reports/dashboards from BAM GUI. You have to be very specific on what your are trying to acheive and we can suggest possible alternates.

  • Error starting Oracle BAM active data cache service

    Hi
    after installing BAM every thing working fine ,but if restart my system Oracle BAM active data cache service throwing following error
    "The Oracle BAM Active Data Cache service on Local computer started and then stopped.Some services stop automatically if they have no work to do,for example the performance logs and alerts service"
    Database is running fine
    Following is the ADC log file error
    2007-12-07 17:19:29,640 [2928] ERROR - ActiveDataCache The Oracle BAM Active Data Cache service failed to start. Oracle.BAM.ActiveDataCache.Common.Exceptions.CacheException: ADC Server exception in Startup(). ---> System.DllNotFoundException: Unable to load DLL (OraOps10.dll).
    at Oracle.DataAccess.Client.OpsTrace.GetRegTraceInfo(UInt32& TrcLevel, UInt32& StmtCacheSize)
    at Oracle.DataAccess.Client.OraTrace.GetRegistryTraceInfo()
    at Oracle.DataAccess.Client.OracleConnection..ctor(String connectionString)
    at Oracle.DataAccess.Client.OracleConnection..ctor(String connectionString)
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleDataFactory.GetConnection()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.GetServerVersion()
    at Oracle.BAM.ActiveDataCache.Kernel.StorageEngine.Oracle.OracleStorageEngine.Startup(IDictionary oParameters)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    --- End of inner exception stack trace ---
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.Startup()
    at Oracle.BAM.ActiveDataCache.Kernel.Server.Server.Startup()
    at Oracle.BAM.ActiveDataCache.Service.DataServer.Run()
    2007-12-07 17:24:45,250 [1524] ERROR - ActiveDataCache Unable to load DLL (OraOps10.dll).
    2007-12-07 17:24:45,265 [1524] WARN - ActiveDataCache Exception occurred in method Startup
    Please help me in resolving this issue .Am getting this issue every time
    Thanks
    BS

    Make sure the path to the ODAC used by BAM (C:\OracleBAM\ClientForBAM\bin) is the first item in the system PATH
    environment variable. Restart your computer after fixing this.
    If that doesn't fix it, please check the Troubleshooting section in the BAM Install Guide.
    Regards, Stephen

  • How to check if table or index bound to data cache loaded in memory?

    In my system, thare are many named data caches created and certain objects bind to those data cache.
    For example, I have a large table mytab which is bound to data cache mycache.
    When I issue a sql like select * from mytab where x=y, run it at first time is is very slow.  run it again and again. then it very fast. it means data loaded in cahce and ready for same query.
    Question:
    1. How can I know if the table mytab has been loaded in data cache mycache?
    2. How to load mytab into data cahce mycache one time for all query in the feature?

    one way to monitor is:
    select CacheName,DBName,OwnerName,ObjectName,IndexID,sum(CachedKB) as "CachedKb"
    from master..monCachedObject
    group by CacheName,DBName,OwnerName,ObjectName,IndexID
    order by CacheName,DBName,OwnerName,ObjectName,IndexID
    But, you should understand that caches have a finite size, and caches contain "pages" of objects including data pages, index pages, and LOB pages.  Also, caches may have different pool sizes, so a page can be in only one cache pool.  So, if you want  a table and all of it's indexes, text/image pages  to be loaded into a dedicated cache, you need a large enough cache to fit all of those pages, and decide which buffer pool you want them in (typically either the 1 page pool, or the 8 page pool).
    Then, simply execute SQL (or dbcc) commands that access all of those pages in the manner you wish to find them in the cache.  For example, two statements, one that scans the table using 2k reads, and another that scans the index (mytab_ind1) using 2k reads.
    select count(*) from mytab plan '( i_scan mytab_cl mytab) ( prop mytab ( prefetch 2 ) ( lru ) )'
    select count(*) from mytab plan '( i_scan mytab_ind1 mytab) ( prop mytab ( prefetch 2 ) ( lru ) )'
    etc etc.
    used count(*) to limit result sets of examples

  • Question about MAXCPU and clearing data cache

    Hi all,
    We're running MaxDB 7.6.03.15 and have following questions. I apologize for not grouping them together with my earlier set of questions along the similar lines.
    1. What would be an appropriate value for our application for MAXCPU when running on a quad processor machine. The application consists of the MaxDB along with 4 windows services and an instance of tomcat. Can we use 4 for MAXCPU or does it always need to be something less than the total number of CPUs on the machine (less than 4 here?). Please advise.
    2. When we're doing query execution time analysis we need to be able to clear data cache in between query runs. We achieve this so far by simply bringing the database offline and then online again. The problem with this approach is its time consuming and if we do 100 query runs a lot of time is spent in waiting for the database go from online to offline to online again. Please let us know if there's an alternative faster way to achieve the same.
    Thanks for your time.
    Sincerely,
    Sameer

    > 1. What would be an appropriate value for our application for MAXCPU when running on a quad processor machine. The application consists of the MaxDB along with 4 windows services and an instance of tomcat. Can we use 4 for MAXCPU or does it always need to be something less than the total number of CPUs on the machine (less than 4 here?). Please advise.
    Of course you may set MAXCPU to 4 - but you risk that the four Usertask UKTs will use up all CPU resources on your system then.
    So it is usually better to be rather defensive with this. If you're unsure about the CPU needs of your other services better try out and check starting with MAXCPU 2.
    > 2. When we're doing query execution time analysis we need to be able to clear data cache in between query runs. We achieve this so far by simply bringing the database offline and then online again. The problem with this approach is its time consuming and if we do 100 query runs a lot of time is spent in waiting for the database go from online to offline to online again. Please let us know if there's an alternative faster way to achieve the same.
    Well, x_cons pagecache_release should do that for you.
    But I never was able to really see the effect of it (e.g. reading the pages in again).
    If I'm not mistaken in this point - which is pretty easy because the command is rarely used and sparsely documented - it just moves the pages to the lower end of the lru list and thereby making them candidates for getting moved out of cache but not actually doing it.
    However why don't you just use the resource monitor to figure out how many pages are touched by your statements?
    Then you don't have to bounce your instance.
    regards,
    Lars

  • _WLS_ADMIN000000.DAT Caching Issue

    Hi all,
    I have a strange situation here. My intention is to point my application to a new database. My application is deployed using an .ear file in weblogic 11g and reads database connection properties from a properties file placed in the classpath of the weblogic server.
    Now, I kept the old database up and changed the properties file in Weblogic with the new database server details and restarted the weblogic server and it worked. Then I reverted back my changes to point my application to the old database using the same procedure.
    Then, I turned off my old database and then tried doing the same and while I was trying to start my application from weblogic console,it failed saying "Could not establish a Network Connection", though the new database which I was trying to point was up and running.
    Then I found a file _WLS_<domain_name>_ADMIN000000.DAT in the path ".../weblogic/<domain_name>/servers/payments_admin/data/store/default that was caching the old database details and then I removed the file and restarted the server.
    Now here are the inferences I can draw from this...
    1. While the old database is up and running, the application starts with those old cached details in the .DAT file and still points to the new server.
    2. When the old database is down, the application tries to connect to the old database and hence fails.
    3. So the application uses .DAT file only while starting and then points to the new database.
    I just want to verify if my findings are correct or am I missing something here.
    Please suggest.
    Thanks in advance.
    Regards,
    Pritam

    Pl do not post duplicates - _WLS_ADMIN000000.DAT Caching Issue

  • Weblogic _WLS_ADMIN000000.DAT caching issue

    Hi all,
    I have a strange situation here. My intention is to point my application to a new database. My application is deployed using an .ear file in weblogic 11g and reads database connection properties from a properties file placed in the classpath of the weblogic server.
    Now, I kept the old database up and changed the properties file in Weblogic with the new database server details and restarted the weblogic server and it worked. Then I reverted back my changes to point my application to the old database using the same procedure.
    Then, I turned off my old database and then tried doing the same and while I was trying to start my application from weblogic console,it failed saying "Could not establish a Network Connection", though the new database which I was trying to point was up and running.
    Then I found a file _WLS_<domain_name>_ADMIN000000.DAT in the path ".../weblogic/<domain_name>/servers/payments_admin/data/store/default that was caching the old database details and then I removed the file and restarted the server.
    Now here are the inferences I can draw from this...
    1. While the old database is up and running, the application starts with those old cached details in the .DAT file and still points to the new server.
    2. When the old database is down, the application tries to connect to the old database and hence fails.
    3. So the application uses .DAT file only while starting and then points to the new database.
    I just want to verify if my findings are correct or am I missing something here.
    Please suggest.
    Thanks in advance.
    Regards,
    Pritam

    Pl do not post duplicates - _WLS_ADMIN000000.DAT Caching Issue

Maybe you are looking for

  • Impact of Merging multiple company codes into ONE

    Hi, We are currently doing a study for a client,  they have the following scenario where the client has got multiple company codes in SAP (due to some tax benefits they had initially created the company codes) now the tax benefits cease to exist and

  • No pro-rating with cancellation

    I have been a long time verizon customer about 10 years sticking to their wireless voice plan, and about last 5 years also month to month. just before year 2014 starts, I finally disconnected services, and made a call to make sure they are aware of i

  • Consumption from Unrestricted Stock for MTO

    Dear Experts, Kindly help me in meeting the below requirement: When I create a Production Order with reference to sales order (T Code Co08), the GI (MB1A - 261 Mvt) can be done only from the Sales Order Stock. The system doesn;t allow the unrestricte

  • Archiving financial document using Data aging

    Hi, reading documentation about Data archiving, ILM and SAP Hana I came across archiving of finacial accounting documents in SMART FI 1.0. It is stated, that the archiving with the archiving object FI_DOCUMENT is replaced by Data Aging for Financial

  • How to properly cite images and media used as to copyright

    All of the images and media used in my iBook were either created by me or gotten from an Open Educational Resource repository such as Wikimedia Commons and are CC licensed such that I am permitted to use them.  Wanting to avoid any delay in getting o