Types of XI cache

Hi, can anybody explain different types of XI cache and where they are used?
Edited by: sree nivas on May 28, 2008 9:03 PM

Hi,
There are 2 types of caches.....
CPA cache - is your Collaboration profile agreement cache, ie to mean it will contains the cache copy of all objects created during configuration time (ID).CPA (Collaboration profile agreement )Cache-mainly for J2EE engine/Adapter engine.
About Cache Monitoring-
http://help.sap.com/saphelp_nw2004s/helpdata/en/92/2fd93f130f9115e10000000a155106/content.htm
Runtime cache - it will contain cache copies of all the objects in XI (IR/ID)
and active version of all your objects during runtime.
Check this.......
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1a69ea11-0d01-0010-fa80-b47a79301290
sxi_cache - this tcode is used to manually refresh the cache and also to view the status of the objeccts in the cache. (/people/sravya.talanki2/blog/2005/12/02/sxicache--ripped-off)
Thanks
Swarup

Similar Messages

  • "Global Type declaration duplicated", cached xsd conflicting in jdeveloper?

    We're having problems recompiling a project after changing a schema in a second project on which it depends:
    Our HR composite app imports a schema from our FaultHandler app. I added 2 new elements to the FaultHandler schema and redeployed to the server - now our HR composite app won't recompile. It errors saying:
    [scac] error: in EmpTransformSynch.bpel(87): query "/ns9:HandleFaultRequest/ns9:TimestampeInit" is invalid, because Global Type declaration/definition of name '{http://mydomain.com/SOA/FaultHandler}HandleFaultRequestType' are duplicated at the following locations:
    [scac] http://mydomain.com:8001/soa-infra/services/default/FaultHandlerProject/xsd/FaultHandler.xsd [line#: 4]
    [scac] http://mydomaincom:8001/soa-infra/services/default/FaultHandlerProject/FaultHandler?XSD=xsd/FaultHandler.xsd [line#: 7]
    [scac] There are at least two of them looking different:
    [scac] http://mydomain.com:8001/soa-infra/services/default/FaultHandlerProject/xsd/FaultHandler.xsd [difference starting at line#:16]
    [scac] http://mydomain.com:8001/soa-infra/services/default/FaultHandlerProject/FaultHandler?XSD=xsd/FaultHandler.xsd [difference starting at line#:19]
    I can view the 2 URIs in my browser, and while they differ slightly in how they reference another schema, the schemas (and imported schema) are semantically equivalent.
    We've taken the following steps to try to resolve it:
    Undeploy the previous revision of Fault handler
    Restart soa server
    Redepoy Fault Handler
    Clean HR app
    Make HR app
    We have the issue in both jdeveloper 11.1.1.3.0 and 11.1.1.4.0
    This is surely something very simple, but it's blockiing us at the moment.
    Any help much appreciated.
    ..Garret
    Edited by: user10714498 on 09-Mar-2011 02:27

    If your schema contains any characters which will be escaped when parsed by a browser, then version present in the actual .xsd and the version referenced via HTTP server will be different causing the issue.
    If you can post the content of your actual schema (as you see it in jDev and not the one opened in the browser) here it will help identifying the issue if any.

  • Fill Olap cache distribution type

    Hi,
    I'm trying to fill the olap cache with information broadcasting, but I cannot see option "Fill olap cache" in the distribution type.
    Only distribution type "broadcast E-Mail" and "Broadcast to portal" are present.
    I reached the broadcasting screen from Tcode RSRT -> ABAP Web.
    Is a Precalc. server required to get access to "fill olap cache" distribution type ?
    How can I get access to distribution type "fill olap cache"  ?
    System: BW 7.01 SP06

    Hi Brian,
    Please check below:
    Broadcasting in BI 7.x - Do I need Portal, Missing Distribution Types
    Also check the expert for
    In SAP NetWeaver BI 7.0 (2004s), which BEx Web functions require SAP NetWeaver 7.0 Portal?
    As soon as you use any of the new BI capabilities for SAP NetWeaver 7.0 (2004s) BEx tools (such as BEx Analyzer & Broadcasting), you require usage types BI Java and Portal. The only exception is if you do not use SAP NetWeaver 7.0 BEx tools and ONLY use the old 3.x BEx tools.
    https://www.sdn.sap.com/irj/sdn/nw-datamanagement?rid=/library/uuid/a5067965-0901-0010-6f8a-bbf0b7424283#q-6
    -Vikram

  • Controlling Content Cache by Mime Type?

    I'm trying to control the application of the "Cache-control" response header based on the mime type of the resource requested. For instance I want to only set it for only static, not dynamic (i.e. JSPs/servlets) content.
    Have figured out how to set it for all content by adding the following:
    PathCheck fn="set-cache-control" control="private,max-age=3600"
    But how can a restrict it to certain mime types such as "image/*" or "application/x-*"?
    I suspect an new Object is needed, but can't figure out how to invoke it by mime type:
    <Object name="cache-static">
    PathCheck fn="set-cache-control" control="private,max-age=3600"
    Service fn="send-file"
    </Object>
    Any help would be apprciated.

    Unfortunately, you won't be able to use a PathCheck directive. The MIME type hasn't yet been determined when the PathCheck directives execute.
    If you're using Sun ONE Web Server 6.1 or later, you could add the following to the default Object:Output type="(image/*|application/x-*)"
           fn="set-cache-control"
           control="private,max-age=3600"

  • Unable to run project on upgrading to azure sdk 2.3 while using cache

    We upgraded azure sdk from 1.8 to 2.3 and we are unable to run our project.
    following is the error.
    Attempt by method 'Microsoft.Web.DistributedCache.DataCacheFactoryWrapper.CreateDataCacheFactoryConfiguration(System.String)' to access method 'Microsoft.ApplicationServer.Caching.DataCacheFactoryConfiguration..ctor(System.String)' failed.
    However if we comment out the following lines of code in web.config we are able to run the project.
    <sessionState mode="Custom" customProvider="AFCacheSessionStateProvider">
    <providers>
    <add name="AFCacheSessionStateProvider" type="Microsoft.Web.DistributedCache.DistributedCacheSessionStateStoreProvider, Microsoft.Web.DistributedCache" cacheName="AppName" dataCacheClientName="default" applicationName="AFCacheSessionState"/>
    </providers>
    </sessionState>
    Below is the screen shot of error :

    Hi Jambor,
    Thanks for the reply.
    We have no other sessionState caching configured other than this one in WebRole.
    So, in webrole we.config, we have:
    <section name="dataCacheClients" type="Microsoft.ApplicationServer.Caching.DataCacheClientsSection, Microsoft.ApplicationServer.Caching.Core" allowLocation="true" allowDefinition="Everywhere" />
    <section name="cacheDiagnostics" type="Microsoft.ApplicationServer.Caching.AzureCommon.DiagnosticsConfigurationSection, Microsoft.ApplicationServer.Caching.AzureCommon" allowLocation="true" allowDefinition="Everywhere" />
    <sessionState mode="Custom" customProvider="AFCacheSessionStateProvider">
    <providers>
    <add name="AFCacheSessionStateProvider" type="Microsoft.Web.DistributedCache.DistributedCacheSessionStateStoreProvider, Microsoft.Web.DistributedCache" cacheName="AppName" dataCacheClientName="default" applicationName="AFCacheSessionState"/>
    </providers>
    </sessionState>
    <dataCacheClients>
    <dataCacheClient name="default">
    <!--To use the in-role flavor of Windows Azure Cache, set identifier to be the cache cluster role name -->
    <!--To use the Windows Azure Cache Service, set identifier to be the endpoint of the cache cluster -->
    <autoDiscover isEnabled="true" identifier="appname.cache.windows.net" />
    <!--<localCache isEnabled="true" sync="TimeoutBased" objectCount="100000" ttlValue="300" />-->
    <!--Use this section to specify security settings for connecting to your cache. This section is not required if your cache is hosted on a role that is a part of your cloud service. -->
    <securityProperties mode="Message" sslEnabled="true">
    <messageSecurity authorizationInfo="{authInfo}" />
    </securityProperties>
    </dataCacheClient>
    </dataCacheClients>
    <cacheDiagnostics>
    <crashDump dumpLevel="Off" dumpStorageQuotaInMB="100" />
    </cacheDiagnostics>
    And here are the installers that are installed in our win7 system:
    Windows Azure Authoring Tools - v2.3
    Windows Azure Compute Emulator - v2.3
    Windows Azure Libraries for .NET - v2.3
    Windows Azure Storage Emulator - v2.3
    Windows Azure Storage Tools - v2.2.2
    Windows Azure Tools for Microsoft LightSwitch for Visual Studio 2012 - October 2012
    Windows Azure Tools for Microsoft Visual Studio 2012 - v2.3
    EDIT: Tried with
    Windows Azure Tools for Microsoft LightSwitch for Visual Studio 2013 - v2.3
    But no luck.
    Please let us know if we are missing something.
    Thanks,
    - Sovan
    sovan kumar das

  • Maximum PDF file size for browsers to cache?

    My site distributes daily newsletters via PDF that are typically about 2MB in size. A link is provided to the PDF, and when clicked, it opens the Adobe Reader plugin within the browser (Firefox, IE) and opens the file therein.
    Recently it was brought to my attention by a dialup user (they do still exist) that when he navigates back or forward in his browser's history, the PDF has to be redownloaded rather than reloaded from his browser's cache. I, with a broadband connection and some wonderful Firefox extensions (Firefox Throttle and Live HTTP Headers) confirmed this behavior.
    However, a similarly sized file (2MB) of a different type (text) is cached by the browser and I can go back and forth in the history without the file having to be redownloaded. I also had success with a smaller sized PDF (70kb).
    What appears to happen is this:
    1) Click the link to the 2MB PDF, and it loads in the browser.
    2) Click browser's back button to back to previous page in browser history.
    3) Click forward button to return to PDF.
    4) A request with a "Range: <byte range>" header is sent which I think requests part of the PDF file (the byte ranges).
    5) The server responds with a "206 Partial Content" response code and resends some/most/all of the PDF to the browser.
    What should happen (and does with other file types and smaller sized PDFs) is that the browser should make a simple request for the PDF file, and assuming the PDF hasn't been modified since the latest request for it, the server should send a "304 Not Modified" response code which instructs the browser to use what it has in cache.
    I've increased my browser cache sizes dramatically to no avail. Cleared them before starting the process to no avail. Should be no problem with any browser's cache limit.
    I'm using Firefox 2 and IE7. My co-worker has IE6 and Acrobat Reader 6 installed on another machine and it works fine there (PDF is cached and does not need to be redownloaded).
    Help! Why aren't these larger PDF files being cached correctly, and what can we do to get them saved in the cache so our poor dialup users don't have to redownload them everytime they navigate their browser history?
    Thanks!

    Just wanted to reply with my discovery and fix. I'm not sure if it is Adobe Reader or the web browsers (I tested Firefox 2 and IE7), but they appear to request parts of large PDF files "on demand." That is, they send requests to the web server with a "Range: <byte range(s)>" header which instructs the web server to only send segments of the file. I think they use the feature of HTTP 1.1 that keeps the connection open or alive to request/send the data incrementally.<br /><br />While this might seem preferrable, in the case of dialup users that navigate through their browser(s) history, they should instead download the PDF in its entirety and allow the browser to cache the file. This way, when they click the link to open the PDF again (or go back/forward in their browser's history), they use what they have in their cache instead of redownloading the PDF (which in the case of dialup users is quite painful).<br /><br />We fixed the problem by instructing our web server to ignore/disregard byte range requests. In a nutshell, you need to have the web server issue a header:<br /><br />Accept-Ranges: none<br /><br />In Apache, you can do this on a per-file-type basis using a FilesMatch directive. In IIS (we're using 5.0), you cannot specify a file type for this behavior; you have to set it in Internet Services Manager by adding a custom header under the "HTTP Headers" tab.<br /><br />Personally, I prefer folks use their browser cache than repeatedly request data (PDF file in this case) from our server. To each his own I guess. Hope this helps someone.

  • How to query data from grid cache group after created global AWT group

    It is me again.
    as I mentioned in my previous posts, I am in progress of setup IMDB grid environment, and now I am at stage of creating cache group. and I created global AWT cache group on one node(cachealone2), but I can not query this global cache group from another node(cachealone1)
    thanks Chirs and J, I have done successfully setup IMDB grid env, and have two node in this grid as below
    Command> call ttGridNodeStatus;
    < MYGRID, 1, 1, T, igs_imdb02, MYGRID_cachealone1_1, 10.214.10.176, 5001, <NULL>, <NULL>, <NULL>, <NULL>, <NULL> >
    < MYGRID, 2, 1, T, igsimdb01, MYGRID_cachealone2_2, 10.214.10.119, 5002, <NULL>, <NULL>, <NULL>, <NULL>, <NULL> >
    2 rows found.
    and I create group ATW cache group on cachealone2
    Command> cachegroups;
    Cache Group CACHEUSER.SUBSCRIBER_ACCOUNTS:
    Cache Group Type: Asynchronous Writethrough global (Dynamic)
    Autorefresh: No
    Aging: LRU on
    Root Table: ORATT.SUBSCRIBER
    Table Type: Propagate
    1 cache group found.
    Command> SELECT * FROM oratt.subscriber;
    0 rows found.
    however I can not query this from another node cachealone1
    Command> SELECT * FROM oratt.subscriber WHERE subscriberid = 1004;
    2206: Table ORATT.SUBSCRIBER not found
    The command failed.
    Command> SELECT * FROM oratt.subscriber WHERE subscriberid = 1004;
    2206: Table ORATT.SUBSCRIBER not found
    The command failed.
    Command> SELECT * FROM oratt.subscriber;
    2206: Table ORATT.SUBSCRIBER not found
    this is example from Oracle docs, I an not sure where I missed for this. thanks for your help.

    Sounds like you have not created the Global AWT cache groupo in the second datastore? There is a multi-step process needed to roll out a cache grid and various things must be done on each node in the correct order. have you done that?
    Try checking out the QuickStart example here:
    http://download.oracle.com/otn_hosted_doc/timesten/1121/quickstart/index.html
    Chris

  • Query from Distributed Cache

    Hi
    I am newbie to Oracle Coherence and trying to get a hands on experience by running a example (coherence-example-distributedload.zip) (Coherence GE 3.6.1). I am running two instances of server . After this I ran "load.cmd" to distribute data across two server nodes - I can see that data is partitioned across server instances.
    Now I run another instance(on another JVM) of program which will try to join the distributed cache and try to query on the loaded on server instances. I see that the new JVM is joining the cluster and querying for data returns no records. Can you please tell me if I am missing something?
         NamedCache nNamedCache = CacheFactory.getCache("example-distributed");
         Filter eEqualsFilter = new GreaterFilter("getLocId", "1000");
         Set keySet = nNamedCache.keySet(eEqualsFilter);
    I see here that keySet has no records. Can you please help?
    Thanks
    sunder

    I got this problem sorted out - the was problem cache-config.xml.. The correct one looks as below.
    <distributed-scheme>
    <scheme-name>example-distributed</scheme-name>
    <service-name>DistributedCache1</service-name>
    <backing-map-scheme>
         <read-write-backing-map-scheme>
         <scheme-name>DBCacheLoaderScheme</scheme-name>
         <internal-cache-scheme>
         <local-scheme>
         <scheme-ref>DBCache-eviction</scheme-ref>
         </local-scheme>
         </internal-cache-scheme>
              <cachestore-scheme>
              <class-scheme>
                   <class-name>com.test.DBCacheStore</class-name>
                   <init-params>
                                  <init-param>
                                       <param-type>java.lang.String</param-type>
                                       <param-value>locations</param-value>
                                  </init-param>
                                  <init-param>
                                       <param-type>java.lang.String</param-type>
                                       <param-value>{cache-name}</param-value>
                                  </init-param>
                   </init-params>                     
                   </class-scheme>
              </cachestore-scheme>
              <cachestore-timeout>6000</cachestore-timeout>
              <refresh-ahead-factor>0.5</refresh-ahead-factor>     
         </read-write-backing-map-scheme>
         </backing-map-scheme>
         <thread-count>10</thread-count>
    <autostart>true</autostart>
    </distributed-scheme>
    <invocation-scheme>
    <scheme-name>example-invocation</scheme-name>
    <service-name>InvocationService1</service-name>
    <autostart system-property="tangosol.coherence.invocation.autostart">true</autostart>
    </invocation-scheme>
    Missed <class-scheme> element inside <cachestore-scheme> of <read-write-backing-map-scheme>.
    Thanks
    sunder

  • Error with Microsoft.ApplicationServer.Caching.VelocityPacketTooBigException

    Hello All
    I am getting the below error while I am searching for a huge record (more than 8.5 MB).
    Microsoft.ApplicationServer.Caching.VelocityPacketTooBigException: Size of message to be sent is 8837220 bytes which is larger than the maximum message size 8388608 specified in configuration.
    I have tried to change the settings in web.config cloud as below:-
    <dataCacheClient name="default" isCompressionEnabled="true">
    <autoDiscover isEnabled="true" identifier="foo" />
    <!--<localCache isEnabled="true" sync="TimeoutBased" objectCount="100000" ttlValue="300" />-->
    <transportProperties maxBufferSize="67108864"/>
    </dataCacheClient>
    But no luck...
    Please help me as am new to Azure.
    Thanks in Advance
    Deepak Gupta
    [email protected]
    919910164863

    Hi Deepak,
    Looks like, you only changed the Cache client config setting in app.config/web.config.
    You have to set the maxbuffersize on both the client and server.
    You also need to set the maxBufferSize for the server. To do this
    Right click your role
    Add -> New Item -> XML File
    Name it "CacheSettings.xml"
    Paste the following XML:
    <?xml version="1.0" encoding="utf-8" ?>
    <configuration>
    <configSections>
    <section name="dataCache" type="Microsoft.ApplicationServer.Caching.DataCacheSection, Microsoft.ApplicationServer.Caching.Core" allowLocation="true" allowDefinition="Everywhere" />
    </configSections>
    <dataCache size="Small">
    <advancedProperties>
    <transportProperties maxBufferSize="XXXX" maxBufferPoolSize="XXXX" />
    </advancedProperties>
    </dataCache>
    </configuration>
    Regards,
    Manu Rekhar

  • Using the cache in Data Integrator

    Hi,
    I'm trying to understand if is possible restrict the use of cache in Data Integrator to load initial datasource to limit the use of server resources.
    I understand from the manual to set the option to use the type of dataflow cache: Pageable or In-Memory. This option does not solve my problem. I would like to avoid to load all the cached datasource.
    Is possible to perform query objects directly without first loading all data sources in the tables?

    base /open SQL Statement etc) the first time i defined the system everything was fine but when i changed the Database (using M.S.Access) the " open SQL Statement" it would show the tables but not thier columns ,I"m using win98 sec edition / Teststand 1.0.1Hello Kitty -
    Certainly it is unusual that you can still see the tables available in your MS Access database but cannot see the columns? I am assuming you are configuring an Open Statement step and are trying to use the ring-control to select columns from your table?
    Can you tell me more about the changes you made to your file when you 'changed' it with MS Access? What version of Access are you using? What happens if you try and manually type in an 'Open Statement Dialog's SQL string such as...
    "SELECT UUT_RESULT.TEST_SOCKET_INDEX, UUT_RESULT.UUT_STATUS, UUT_RESULT.START_DATE_TIME FROM UUT_RESULT"
    Is it able to find the columns even if it can't display them? I am worried that maybe you are using a version of MS Access that is too new for the version of TestSt
    and you are running. Has anything else changed aside from the file you are editing?
    Regards,
    -Elaine R.
    National Instruments
    http://www.ni.com/ask

  • Azure Cache ErrorCode ERRCA0017 :SubStatus ES0006

    Hello everyone.
    Im trying to access MS Azure Cache service and pretty often I get this error:
    Microsoft.ApplicationServer.Caching.DataCacheException: ErrorCode<ERRCA0017>:SubStatus<ES0006>:There is a temporary failure. Please retry later. (One or more specified cache servers are unavailable, which could be caused by busy network or servers. For on-premises cache clusters, also verify the following conditions. Ensure that security permission has been granted for this client account, and check that the AppFabric Caching Service is allowed through the firewall on all cache hosts. Also the MaxBufferSize on the server must be greater than or equal to the serialized object size sent from the client.). Additional Information : The client was trying to communicate with the server: net.tcp://remoto.cache.windows.net: line 23233.       at Microsoft.ApplicationServer.Caching.DataCache.ThrowException(ErrStatus errStatus, Guid trackingId, Exception responseException, Byte[][] payload, EndpointID destination)       at Microsoft.ApplicationServer.Caching.SocketClientProtocol.Get(String key, ref DataCacheItemVersion version, ref TimeSpan timeout, ref ErrStatus err, String region, IMonitoringListener listener)       at Microsoft.ApplicationServer.Caching.DataCache.InternalGet(String key, ref DataCacheItemVersion version, String region, IMonitoringListener listener)       at Microsoft.ApplicationServer.Caching.DataCache.<>c__DisplayClass53.<Get>b__52()       at Infrastructure.Azure.Cache.AzureCacheServiceClient.<>c__DisplayClass6`1.<Get>b__5() in AzureCacheServiceClient.cs: line 88       at Microsoft.Practices.EnterpriseLibrary.TransientFaultHandling.RetryPolicy.ExecuteAction(Func`1 func)
    It happens when Im trying to access some object in cache like this:
    public T Get<T>(string key)
    return retryPolicy.ExecuteAction(() =>(T) (_cache.Get(key)));
    Here is my initialization code:
    var cacheFactory = new DataCacheFactory();_cache = cacheFactory.GetDefaultCache();
    var retryStrategy = new FixedInterval(15, TimeSpan.FromSeconds(2));
    _retryPolicy = new RetryPolicy<CustomCacheTransientErrorDetectionStrategy>(retryStrategy);
    And app.config:
    <configSections>
    <section name="dataCacheClients" type="Microsoft.ApplicationServer.Caching.DataCacheClientsSection, Microsoft.ApplicationServer.Caching.Core" allowLocation="true" allowDefinition="Everywhere" />     
    <section name="cacheDiagnostics" type="Microsoft.ApplicationServer.Caching.AzureCommon.DiagnosticsConfigurationSection, Microsoft.ApplicationServer.Caching.AzureCommon" allowLocation="true" allowDefinition="Everywhere" />
    </configSections>
    <dataCacheClients>
    <dataCacheClient channelOpenTimeout="1000" requestTimeout="45000" name="default">
    <autoDiscover isEnabled="true" identifier="[some.host.name]" />
    <securityProperties mode="Message" sslEnabled="true">
    <messageSecurity authorizationInfo="***" />
    </securityProperties>
    </dataCacheClient>
    </dataCacheClients>
    It happend at least three times already while the Azure Health Status (https://azure.microsoft.com/en-us/status/) said everything was fine at that time.
    As the exception message says - there are some 'temporary' failures on MS side still maybe I'm doing smth wrong in my code?

    Hi,
    This issue might be happed due to hitting the Windows Azure Cache limitation, where you can not add more than 8MB data at a time. The simple workaround for this issue is to split the object into smaller chunks, i.e  you can  serialize the object
    into smaller byte arrays and pass the arrays to cache as multiple objects.  I recommend you could refer to this blog:
    http://blogs.msdn.com/b/pkc/archive/2011/12/21/datacacheexception-errorcode-lt-errca0017-gt-substatus-lt-es0006-gt.aspx
    Any questions, please let me know.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • SharePoint 2010 Cache Not working for Publishing Images Document Library

    Hi
    I am using SharePoint 2010 Publishing Site and have enabled anonymous access.
    I have enabled blob cache and object cache in the site.
    While i was checking with fiddler those objects in style library and other document libraries are cached.
    Whereas the images from the document library Publishing images is not cached; its returning 304 Not Modified status code.
    Is there a way to include the Publishing images to be cacheable as well ?
    Please advise
    Ready for Action

    Did you perform IISreset on all sharepoint server after enabling cache
    http://technet.microsoft.com/en-us/library/cc261797%28v=office.15%29.aspx
    To use the page output cache and the associated cache profile settings, you must be using the Publishing feature on your site.
    http://sharepointinterface.com/tag/caching/
    Wiley went out to SharePoint, checked the article that he wrote, and saw that he did indeed use a series of dynamite sticks for an image. The page should have actually appeared as it does in the screenshot that is above and to the left. After a quick facepalm,
    Wiley realized that he needed to make a change – and fast.
    Wiley went out to the Publishing Images library for the site collection and uploaded a new version of the
    newsarticleimage.jpg image file – one that contained a barrel of TNT instead of a bundle of dynamite. He then browsed to the article page and did a refresh.
    Nothing changed.
    Wiley hit F5 in his browser. Still nothing changed.
    Over the course of the hour that followed, Wiley grew increasingly more bewildered and panicked as he tried in vain to get the new TNT barrel to show up on the article page. He uploaded the image several more times, closed and re-opened his browser, deleted
    and then reloaded the image, re-published and re-approved the actual article page, and even got the administrators to flush the SharePoint BLOB Cache. None of the actions made a difference.
    The Coyote Never Wins
    Why didn’t any of Wiley’s efforts make a difference? Because what Wiley didn’t understand was that there was nothing he could do short of flushing his cache that would prompt the browser to re-request the updated image. The browser started using the cached
    copy of the image after the first request Wiley made in the morning; i.e., the request to verify that the image on the page was incorrect as Fulfillment indicated. For another 24 hours (86400 seconds), the browser would continue to use the cached image.
    Wiley’s image problem was just one of the potential issues that might surface as a result of the BLOB Cache change. It was also one of the more visible problems. In looking at the
    path attribute of the BlobCache element, you might have noticed some of the other file types that got cached by default – file types with
    js (JavaScript) and css (Cascading Style Sheets) extensions, for example. Any of those file types which were served from site collection lists and libraries would also be impacted by the “fetch once and use for 24 hours” behavior.
    If this helped you resolve your issue, please mark it Answered

  • Fill OLAP cache via broadcaster

    Hi,
    I'm trying to use the broadcaster with distribution type 'Fill OLAP cache' and schedule it using " Including Event Data Changes in Process Chains ". The process chain has been adapted with the 'Event data change' process and scheduled. Process chain runs fine.  All was set up according to the link below.
    http://help.sap.com/saphelp_nw70/helpdata/EN/a5/359840dfa5a160e10000000a1550b0/frameset.htm
    However, when I try to schedule the broadcaster setting, I only get the options of "Execution at Predefined Time"  or direct scheduling.
    Should I not be able to see "Execution with Data Change in the InfoProvider " as an option? Is there some other setting to be made?
    Very grateful for any ideas.
    Thanx

    Check if you have S_RS_BCS role authorization.
    You have to use process type DATACHANGE in the process chain for data load.
    Check this http://help.sap.com/saphelp_nw04/Helpdata/EN/ec/0d0e405c538f5ce10000000a155106/content.htm

  • Windows Azure Caching Previewing Service No Longer Works

    This is the error I get back:
    "ErrorCode<ERRCA0030>:SubStatus<ES0001>:Unable to complete authorization request for connecting to the caching endpoint. Please try again after sometime. If the issue persists, contact Microsoft support."
    Any ideas?

    Hi,
    I encounter the same issue but am not sure regenerating keys will fixed it. 
    once I got this error, I redeploy the application and it worked but after few hrs I got the same exception.
    Inner Exception
    Type : Microsoft.ApplicationServer.Caching.DataCacheException, Microsoft.ApplicationServer.Caching.Core, Version=1.0.0.0, Culture=neutral, PublicKeyToken=31bf3856ad364e35
    Message : ErrorCode&lt;ERRCA0030&gt;:SubStatus&lt;ES0001&gt;:Unable to complete authorization request for connecting to the caching endpoint. Please try again after sometime. If the issue persists, contact
    Microsoft support. 
    Source : Microsoft.ApplicationServer.Caching.Core
    Need to investigate how keys regenerate will fix this issue. Any Thoughts?

  • Does Cover Flow leave a cache?

    Does viewing documents, pictures, or video through Coverflow in Finder leave any type of thumbnail cache file in Leopard?

    Coverflow is just a different manner of display. It can be used with either, say, an iMac desktop (with system 10.5x). The same cache happens as before, but now there are some new ways of displaying that information visually. Coverflow is less like creating new cache than finding three ways to see the same cache you always had.

Maybe you are looking for