Client side cacheing!

Hi,
We have a Java Web Start application, which caches around 500MB of heavy weight objects in memory. These corresponds to different reports and it kind of grows over the time. I am looking for some kind of architecture framework or some work around to maintain this cache at the client side. It is bit late to change the architecture and it is important to cache all the objects.
Could somebody suggest your views on this?
Thank You,
Vikram

Cross post:
http://forum.java.sun.com/thread.jspa?threadID=649133
http://forum.java.sun.com/thread.jspa?threadID=649134
http://forum.java.sun.com/thread.jspa?threadID=649135
http://forum.java.sun.com/thread.jspa?threadID=649136
http://forum.java.sun.com/thread.jspa?threadID=649137
http://forum.java.sun.com/thread.jspa?threadID=649138

Similar Messages

  • Possible to delete Offline Files content for a specific user from the Client Side Cache (CSC) ?.

    Hello Everyone,
    We would like to implement a script to delete the offline files in the Client Side Cache (CSC) for a nominated user (on Windows 7 x64 enterprise).
    I am aware that;
    1. We can use a registry value to flush the entire CNC cache (for all users) next time the machine reboots.
    2. If we delete the user's local profile it appears that Windows 7 also removes their content from the local CSC.
    However, we would like to just delete the CSC content for a particular nominated user without having to delete their local user profile.
    In our environment we have many users that share workstations but only use them occasionally. We don't use roaming profile so we would like to retain all the users' local profiles but still delete the CSC content for any users that haven't
    logged on in a week.
    Any ideas or info would be appreciated !
    Thanks, Makes

    Hi,
    I don't think this is possible.
    If you want to achieve it via script, I suggest you post it in official script forum for more professional help:
    http://social.technet.microsoft.com/Forums/scriptcenter/en-US/home?category=scripting
    The reason why we recommend posting appropriately is you will get the most qualified pool of respondents, and other partners who read the forums regularly can either share their knowledge or learn from your interaction with us. Thank you for your understanding.
    Karen Hu
    TechNet Community Support

  • Client-Side Caching of Static Objects

    Hi,
    We just installed SPS12 for NWs.  I learned of this new client-side caching of static objects while reading through the release notes, and I went in to our portal to check it out.  I went to the location where the <a href="http://help.sap.com/saphelp_nw04s/helpdata/en/45/7c6336b6e5694ee10000000a155369/content.htm">release notes</a> (and an <a href="https://www.sdn.sap.com/irj/sdn/weblogs?blog=/pub/wlg/6622">article in the blog</a>) specified, System Administration > System Configuration > Knowledge Management > Content Management > Global Services.
    Problem is that I do not see Client Cache Service and Client Cache Patterns.  I enabled "Show Advanced Options," but I do not see those two configuration items.
    Does anyone know why I am not seeing those two configuration items and how I can get to those?
    Thank you.

    Hi,
    We are using SPS 12.
    KMC-CM  7.00 SP12 (1000.7.00.12.0.20070509052145) 
    KMC-BC  7.00 SP12 (1000.7.00.12.0.20070509052008) 
    CAF-KM  7.00 SP12 (1000.7.00.12.0.20070510091043) 
    KMC-COLL  7.00 SP12 (1000.7.00.12.0.20070509052223) 
    KM-KW_JIKS  7.00 SP12 (1000.7.00.12.0.20070507080500)

  • Client side caching

    Hi all,
    I wonder how client side caching (in memory) would work with BlazeDS.
    There was a project called FAST that would support client side caching for Remoting:
    http://www.adobe.com/devnet/flex/articles/fast_userguide.html, but this doesn't seem to work for Flex 3.
    Do I have to invent my own caching mechanism?
    Without client side caching support I don't see how an application build on BlazeDS remoting would ever scale to many users.
    Regards,
    Markus

    Hi Markus. I took a look at that FAST project. There isn't anything like that in BlazeDS currently. It probably wouldn't be too hard to port that code to make it work with Flex 3 but it's usefulness looks somewhat limited. There doesn't look to be any way to timeout a request that has been cached for example which I think most people would want. If you think this is something that would be useful you can log an enhancement request in Jira.
    http://bugs.adobe.com/jira/browse/BLZ
    As to your point about not seeing how an application built on BlazeDS remoting would scale, I think it's a fairly common practice to not expose functions to the end user that directly call backend services. It's pretty straightforward, for example, to request some data when the application loads and then refresh the data at some interval. You don't need a caching API to do this.
    -Alex

  • Client-side caching of Javascript (for Google Web Toolkit *.cache.* files

    Hi all,
    I'm trying out the use of Google Web Toolkit (GWT) for AJAX code development, leveraging RPC calling back-end Web Services for a document browser prototype.  When the JavaScript code is generated by GWT, it has the ability to automatically distinguish between cacheable and non-cacheable content via file extensions (.cache. and .nocache.).
    Now when running in a Tomcat environment with appropriate caching settings; the application runs extremely fast even on really poor latency sites (>500ms round trip); however on a NetWeaver stack, I can't find any information on how to set an attribute on .cache. files to set an expiry of 1 year.
    Anyone know how to do this on a NetWeaver J2EE stack?
    Thanks,
    Matt
    PS. For reference, GWT is a very cool open source project worth watching in the Enterprise space for targeted high-usability, high performance apps.  Just the image bundles concept themselves are an awesome approach to minimizing impact of small images on performance.

    Hi again,
    I thought I should post the solution I came up with in case people search on GWT in the future.  In terms of caching, the Portal does a good job of making nearly everything that GWT produces to be cached at the client; and for the life of me, I couldn't get nocache files not cached at the client side. 
    So thanks to my friendly local SAP experts in the portal space; I discovered there's not much you can do to get this working except try use standard browser tags like the following to prevent caching on the browser:
    Note - Can't seem to save this post in SDN with Meta tags that I'm trying to display so check out http://www.willmaster.com/library/tutorials/preventing_content_cache.php for more info...
    Unfortunately, the way that GWT creates the nocache files, it is not possible to modify the .js files to do this so I needed an alternative approach.
    Then I thought, well, the html file that includes the JS file is cached, but it's not very big, so how about I just write a script to modify the html file when I deploy to production to point to the nocache file with a date suffix.
    Let me explain in more detail:
    1. The original html file that includes the nocache.js file gets modified to point to nocache.<date>.js.
    2. The original nocache.js file gets changed to nocache.<date>.js
    3. Now produce an external bat file to do this automatically for you before deployment which DevStudio can call as part of an Ant script or similar.
    That's it.  So when you are developing, you can manually delete your cache, but just before you go to production, you do this conversion, and you will never have an issue with clients having the wrong version of javascript files. 
    Note - GWT takes care of caching of most of the files by using a GUID equivalenet each time you compile, so it's just a couple of files that really need to not be cached.
    Anyway, I'm not explaining this solution very well, but drop us a line if you need to understand this approach further.
    And lastly, my feedback about GWT - It Rocks!  Everyone loves the result, it's easy to build and maintain, it offers a great user experience, and it's fast.  Not a replacement for transactional apps; but for the occasional user - go for it.
    Matt
    Edited by: Matt Harding on May 22, 2008 7:48 AM

  • Controlling the client-side cache

    At the current stage on my project I'm finding that I'm changing my mapping definitions fairly frequently (because the country side is just the wrong shade of egg-shell). Clearing the server side cache is fairly easy to do, provided you remember to do it, but my problem is on the client side.
    Right now, the tiles are returned with a one week time to live. So if I change the mapping definitions it might not propagate through to all the clients until up to a week later. What I'd like to do, at least for now, is turn off storing of map tiles on the client-side. The easiest way to do that would be to specify the cache-control header but I can't seem to find a configuration option for that.
    Has anyone setup MapViewer to use a different cache-control?

    Hi Mark,
    You can specify the cache control statements for the page itself.
    If this does not help, try to set expires header for the images. ex for Apache see mod_expires for a directory/location setting
    regards, michael

  • Client-Side Cache Question

    Let's say I have a FLV that "lives" on a server, and I serve
    it up through, say, Ruby. The Ruby script takes care of obtaining
    the FLV from the filesystem and renders it to the browser.
    Inside my client-side SWF, my code to connect to the Ruby
    application and get the FLV may look like this:
    var nc:NetConnection = new NetConnection();
    nc.connect(null);
    var ns:NetStream = new NetStream(nc);
    tvid.attachVideo(ns);
    ns.setBufferTime(2);
    statusID = setInterval(videoStatus, 200);
    ns.onStatus = function(info) {
    trace(info.code);
    if(info.code == "NetStream.Buffer.Full") {
    bufferClip._visible = false;
    ending = false;
    clearInterval( statusID );
    statusID = setInterval(videoStatus, 200);
    if(info.code == "NetStream.Buffer.Empty") {
    if ( !ending ) {
    bufferClip._visible = true;
    if(info.code == "NetStream.Play.Stop") {
    bufferClip._visible = false;
    //ending = true;
    if(info.code == "NetStream.Play.Start") {
    ending = false;
    if(info.code == "NetStream.Buffer.Flush") {
    ending = true;
    //Play it
    ns.play("
    http://localhost:3000/stream");
    ==============
    It seems to me that the "decision" of whether or not to cache
    is entirely dependent on the access method within the client-side
    SWF. So, if in this case, I'm using NetStream to stream the video,
    will it still be cached on the client end? Or do I -have- to use
    FMS to prevent client caching - and if so, why? How does FMS
    prevent the client from caching the data (isn't it up to the client
    to delete the data bits after they're viewed?)
    Thanks a bunch for the help.

    Let's say I have a FLV that "lives" on a server, and I serve
    it up through, say, Ruby. The Ruby script takes care of obtaining
    the FLV from the filesystem and renders it to the browser.
    Inside my client-side SWF, my code to connect to the Ruby
    application and get the FLV may look like this:
    var nc:NetConnection = new NetConnection();
    nc.connect(null);
    var ns:NetStream = new NetStream(nc);
    tvid.attachVideo(ns);
    ns.setBufferTime(2);
    statusID = setInterval(videoStatus, 200);
    ns.onStatus = function(info) {
    trace(info.code);
    if(info.code == "NetStream.Buffer.Full") {
    bufferClip._visible = false;
    ending = false;
    clearInterval( statusID );
    statusID = setInterval(videoStatus, 200);
    if(info.code == "NetStream.Buffer.Empty") {
    if ( !ending ) {
    bufferClip._visible = true;
    if(info.code == "NetStream.Play.Stop") {
    bufferClip._visible = false;
    //ending = true;
    if(info.code == "NetStream.Play.Start") {
    ending = false;
    if(info.code == "NetStream.Buffer.Flush") {
    ending = true;
    //Play it
    ns.play("
    http://localhost:3000/stream");
    ==============
    It seems to me that the "decision" of whether or not to cache
    is entirely dependent on the access method within the client-side
    SWF. So, if in this case, I'm using NetStream to stream the video,
    will it still be cached on the client end? Or do I -have- to use
    FMS to prevent client caching - and if so, why? How does FMS
    prevent the client from caching the data (isn't it up to the client
    to delete the data bits after they're viewed?)
    Thanks a bunch for the help.

  • Windows 8.1 Client Side Caching - Is it possible to specify the location of the offline file cache?

    Hi,
    Is there any approved way to change the location of the offline files cache in Windows 8.1?  I'm asking before moving client systems from Windows 7 where a registry change was required - is this still the way to go?
    I believe that if the default user-profile directory is moved, this causes problems for updates to Windows 8/8.1 now, is the same true if you move CSC?
    Just a bit of background.  The systems I want to do this for have smallish SSD boot drives with a higher capacity spinning disk.  The machines are members of a domain with Folder Redirection setup.  I could setup group policies to disable
    caching of some of the larger directories or even disable CSC entirely on the machines which have an SSD boot drive, but would prefer not to do so.
    Thanks in advance.

    Hi,
    According to your descryption, In windows 7, the tool is migwiz.exe, namely the Windows Easy Transfer, you can modify the registry in the way listed in the post below:
    http://social.technet.microsoft.com/Forums/en-US/bbf5890c-b3d7-4b38-83d8-d9a5e025fb2b/how-to-move-clientside-caching-to-a-new-location?forum=w7itproinstall
    In Windows 8, Windows easy transfer is also a built in tool to migrate user profiles and user settings, but as for the offline transfer, I suggest to use the USMT for your situation.
    Because with USMT you can specify the folder that you want to migrate the files to, by using an offline.xml you can set the path just as the sample listed below:
    <offline>
         <winDir>
              <path>C:\Windows</path>
              <path>D:\Windows</path>
              <path>E:\</path>
         </winDir>
         <failOnMultipleWinDir>1</failOnMultipleWinDir>
    </offline>
    You can also refer t othe details at:
    http://technet.microsoft.com/en-us/library/hh824880.aspx
    Regards
    Wade Liu
    TechNet Community Support

  • Client Side Cache Manipulation

    This thread is to clarify a previous thread.
    In JNLP is there a directory that gets created when a program gets installed, it is never cleared by JNLP while the program is installed, and, upon uninstallation of that program, that directory gets completely cleared of all files, not just jars, or, the directory is removed.
    Thanks =)

    Your comments make me suspect you have not discovered the
    JNLP cache yet.. It is mentioned in the Java Console settings
    and I have discussed how to find it, on messages to this
    forum over the last 24-48 hours (cannot quite recall to whom).
    Do you know where to find the cache?
    (I am guessing that being able to explore the cache
    would answer many of your questions)

  • Client side result set cache

    Hello,
    I try to get the client side result set cache working, but i have no luck :-(
    I'm using Oracle Enterprise Edition 11.2.0.1.0 and as client diver 11.2.0.2.0.
    Executing the query select /*+ result_cache*/ * from p_item via sql plus or toad will generate an nice execution plan with an RESULT CACHE node and the v$result_cache_objects contains some rows.
    After I've check the server side cache works. I want to cache the client side
    My simple Java Application looks like
    private static final String ID = UUID.randomUUID().toString();
    private static final String JDBC_URL = "jdbc:oracle:oci:@server:1521:ORCL";
    private static final String USER = "user";
    private static final String PASSWORD = "password";
    public static void main(String[] args) throws SQLException {
    OracleDataSource ds = new OracleDataSource();
    ds.setImplicitCachingEnabled(true);
    ds.setURL( JDBC_URL );
    ds.setUser( USER );
    ds.setPassword( PASSWORD );
    String sql = "select /*+ result_cache */ /* " + ID + " */ * from p_item d " +
    "where d.i_size = :1";
    for( int i=0; i<100; i++ ) {
    OracleConnection connection = (OracleConnection) ds.getConnection();
    connection.setImplicitCachingEnabled(true);
    connection.setStatementCacheSize(10);
    OraclePreparedStatement stmt = (OraclePreparedStatement) connection.prepareStatement( sql );
    stmt.setLong( 1, 176 );
    ResultSet rs = stmt.executeQuery();
    int count = 0;
    for(; rs.next(); count++ );
    rs.close();
    stmt.close();
    System.out.println( "Execution: " + getExecutions(connection) + " Fetched: " + count );
    connection.close();
    private static int getExecutions( Connection connection ) throws SQLException {
    String sql = "select executions from v$sqlarea where sql_text like ?";
    PreparedStatement stmt = connection.prepareStatement(sql);
    stmt.setString(1, "%" + ID + "%" );
    ResultSet rs = stmt.executeQuery();
    if( rs.next() == false )
    return 0;
    int result = rs.getInt(1);
    if( rs.next() )
    throw new IllegalArgumentException("not unique");
    rs.close();
    stmt.close();
    return result;
    100 times the same query is executed and the statement exection count is incemented every time. I expect just 1 statement execution ( client database roundtrip ) and 99 hits in client result set cache. The view CLIENT_RESULT_CACHE_STATS$ is empty :-(
    I'm using the oracle documentation at http://download.oracle.com/docs/cd/E14072_01/java.112/e10589/instclnt.htm#BABEDHFF and I don't kown why it does't work :-(
    I'm thankful for every tip,
    André Kullmann

    I wanted to post a follow-up to (hopefully) clear up a point of potential confusion. That is, with the OCI Client Result Cache, the results are indeed cached on the client in memory managed by OCI.
    As I mentioned in my previous reply, I am not a JDBC (or Java) expert so there is likely a great deal of improvement that can be made to my little test program. However, it is not intended to be exemplary, didactic code - rather, it's hopefully just enough to illustrate that the caching happens on the client (when things are configured correctly, etc).
    My environment for this exercise is Windows 7 64-bit, Java SE 1.6.0_27 32-bit, Oracle Instant Client 11.2.0.2 32-bit, and Oracle Database 11.2.0.2 64-bit.
    Apologies if this is a messy post, but I wanted to make it as close to copy/paste/verify as possible.
    Here's the test code I used:
    import java.sql.ResultSet;
    import java.sql.PreparedStatement;
    import java.sql.SQLException;
    import oracle.jdbc.pool.OracleDataSource;
    import oracle.jdbc.OracleConnection;
    class OCIResultCache
      public static void main(String args []) throws SQLException
        OracleDataSource ods = null;
        OracleConnection conn = null;
        PreparedStatement stmt = null;
        ResultSet rset = null;
        String sql1 = "select /*+ no_result_cache */ first_name, last_name " +
                      "from hr.employees";
        String sql2 = "select /*+ result_cache */ first_name, last_name " +
                      "from hr.employees";
        int fetchSize = 128;
        long start, end;
        try
          ods = new OracleDataSource();
          ods.setURL("jdbc:oracle:oci:@liverpool:1521:V112");
          ods.setUser("orademo");
          ods.setPassword("orademo");
          conn = (OracleConnection) ods.getConnection();
          conn.setImplicitCachingEnabled(true);
          conn.setStatementCacheSize(20);
          stmt = conn.prepareStatement(sql1);
          stmt.setFetchSize(fetchSize);
          start = System.currentTimeMillis();
          for (int i=0; i < 10000; i++)
            rset = stmt.executeQuery();
            while (rset.next())
            if (rset != null) rset.close();
          end = System.currentTimeMillis();
          if (stmt != null) stmt.close();
          System.out.println();
          System.out.println("Execution time [sql1] = " + (end-start) + " ms.");
          stmt = conn.prepareStatement(sql2);
          stmt.setFetchSize(fetchSize);
          start = System.currentTimeMillis();
          for (int i=0; i < 10000; i++)
            rset = stmt.executeQuery();
            while (rset.next())
            if (rset != null) rset.close();
          end = System.currentTimeMillis();
          if (stmt != null) stmt.close();
          System.out.println();
          System.out.println("Execution time [sql2] = " + (end-start) + " ms.");
          System.out.println();
          System.out.print("Enter to continue...");
          System.console().readLine();
        finally
          if (rset != null) rset.close();
          if (stmt != null) stmt.close();
          if (conn != null) conn.close();
    }In order to show that the results are cached on the client and thus server round-trips are avoided, I generated a 10046 level 12 trace from the database for this session. This was done using the following database logon trigger:
    create or replace trigger logon_trigger
    after logon on database
    begin
      if (user = 'ORADEMO') then
        execute immediate
        'alter session set events ''10046 trace name context forever, level 12''';
      end if;
    end;
    /With that in place I then did some environmental setup and executed the test:
    C:\Projects\Test\Java\OCIResultCache>set ORACLE_HOME=C:\Oracle\instantclient_11_2
    C:\Projects\Test\Java\OCIResultCache>set CLASSPATH=.;%ORACLE_HOME%\ojdbc6.jar
    C:\Projects\Test\Java\OCIResultCache>set PATH=%ORACLE_HOME%\;%PATH%
    C:\Projects\Test\Java\OCIResultCache>java OCIResultCache
    Execution time [sql1] = 1654 ms.
    Execution time [sql2] = 686 ms.
    Enter to continue...This is all on my laptop, so results are not stellar in terms of performance; however, you can see that the portion of the test that uses the OCI client result cache did execute in approximately half of the time as the non-cached portion.
    But, the more compelling data is in the resulting trace file which I ran through the tkprof utility to make it nicely formatted and summarized:
    SQL ID: cqx6mdvs7mqud Plan Hash: 2228653197
    select /*+ no_result_cache */ first_name, last_name
    from
    hr.employees
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute  10000      0.10       0.10          0          0          0           0
    Fetch    10001      0.49       0.54          0      10001          0     1070000
    total    20002      0.60       0.65          0      10001          0     1070000
    Misses in library cache during parse: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 94 
    Number of plan statistics captured: 1
    Rows (1st) Rows (avg) Rows (max)  Row Source Operation
           107        107        107  INDEX FULL SCAN EMP_NAME_IX (cr=2 pr=0 pw=0 time=21 us cost=1 size=1605 card=107)(object id 75241)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                   10001        0.00          0.00
      SQL*Net message from client                 10001        0.00          1.10
    SQL ID: frzmxy93n71ss Plan Hash: 2228653197
    select /*+ result_cache */ first_name, last_name
    from
    hr.employees
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.01          0         11         22           0
    Fetch        2      0.00       0.00          0          0          0         107
    total        4      0.00       0.01          0         11         22         107
    Misses in library cache during parse: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 94 
    Number of plan statistics captured: 1
    Rows (1st) Rows (avg) Rows (max)  Row Source Operation
           107        107        107  RESULT CACHE  0rdkpjr5p74cf0n0cs95ntguh7 (cr=0 pr=0 pw=0 time=12 us)
             0          0          0   INDEX FULL SCAN EMP_NAME_IX (cr=0 pr=0 pw=0 time=0 us cost=1 size=1605 card=107)(object id 75241)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       2        0.00          0.00
      log file sync                                   1        0.00          0.00
      SQL*Net message from client                     2        1.13          1.13The key differences here are the execute, fetch, and SQL*Net message values. Using the client-side cache, the values drop dramatically due to getting the results from client memory rather than round-trips to the server.
    Of course, corrections, clarifications, etc. welcome and so on...
    Regards,
    Mark

  • Server Side Caching of pdf files

    Hi all,
    I've run into a rather odd problem that I can't find a way around. I've created some URL iViews to display some html files and some pdf files. I've adjusted the Load parameters for these iviews to the following:
    Allow Client-Side caching: Yes
    Cache Level: Shared
    Cache Validity Period: 12 hrs
    Fetch Mode: Server-Side
    I've found that the html URL iViews work fine like this, but the pdf URL iViews do not. The pdf ones seem to just hang.
    I've noticed that if I change the Fetch Mode to client-side for the pdf files, the iview then begins to work. However, I don't believe that this is the way that I want to configure my URL iviews. I believe that I want them all to be server-side cached.
    I've followed the instructions on 'setting up system properties for proxy server' (System Administration -> System Configuration -> Service Configuration -> Applications -> com.sap.portal.ivs.httpservice -> Services -> Proxy). After making these configuration changes, I've stopped/restarted portal services. The problem still occurs.
    Does anyone have any idea as to what may be causing this problem and how do I fix it?
    Thanks!
    -Stephen Spalding
    Web Developer
    Graybar

    Hello.
    I have the same problem.  When I tried open a pdf file with  fecth = server-side,  I waitting for 500 seconds... But if I tried open with Fetcf = clien-side I only wait 2 seconds.
    Did you found the solution?
    Thanks.
    Paco.

  • Is NFS client data cacheing possible?

    Yesterday, I was viewing an HD 1080 video with VLC, and noticed that Activity Monitor was showing about 34MB/sec from my NAS box. My NAS box runs OpenSolaris (I was at Sun for over 20 years, and some habits die hard), and the 6GB video file was mounted on my iMac 27" (10.7.2) using NFSv3 (yes, I have a gigabit network).
    Being a long term UNIX performance expert and regular DTrace user, I was able to confirm that VLC on Lion was reading the file at about 1.8MB/sec, and that the NFS server was being hit at 34MB/sec. Further investigation showed that the NFS client (Lion) was requesting each 32KB block 20 times!
    Note: the default read size for NFSv3 over TCP is 32KB).
    Digging deeper, I found that VLC was reading the file in 1786 byte blocks. I have concluded that Lion's NFSv3 client implement at least one 32KB read for each application call to read(2), and that no data is cached betweem reads (this fully accounts for the 20x overhead in this case).
    A workaround is to use say rsize=1024, which will increase the number of NFS ops but dramatically reduce the bandwidth consumption (which means I might yet be able to watch HD video over wifi).
    That VLC should start issuing such small reads is a bug, so I have also written some notes in the vlc.org forums. But client side cacheing would hide the issue from the network.
    So, the big question: is it possible to enable NFS client data cacheing in Lion?

    The problem solved itself mysteriously overnight, without any interference from myself.
    The systems are again perfectly happily mounting the file space (650 clients of them all at the same time
    mounting up to 6 filesystems from the same server) and the server is happily serving again as it has been for the past 2 years.
    My idea is that there has been a network configuration clash, but considering that the last modification of NIS hosts file was about 4 weeks ago and the latest server was installed then and has been serving since then, I have no
    idea how such clash could happen without interference in the config files. It is a mystery and I will have to make
    every effort to unravel it. One does not really like to sweep incidents like that un-investigated under the carpet.
    If anybody has any suggestions and thoughts on this matter please post them here.
    Lydia

  • Error writing data to client-side row cache.  URGENT!!!

    Hi everyone
    I had had this error for 3 weeks and I don't know what am I doing wrong?
    I have an Oracle Database. and when I Run my project from JSC I get this error:.
    Description: An unhandled exception occurred during the execution of the web application. Please review the following stack trace for more information regarding the error.
    Exception Details: javax.faces.FacesException
    java.sql.SQLException: [sunm][Oracle JDBC Driver]Error writing data to client-side row cache.
    Possible Source of Error:
    Class Name: com.sun.jsfcl.data.RowSetDataModel
    File Name: RowSetDataModel.java
    Source not available. Information regarding the location of the exception can be identified using the exception stack trace below.
    I uninstalled and installed the JSC and the first time that I run my project worked fine, but just that time, because then the error appeared again.
    What can I do????
    Thanks!!!
    Wendy,
    Costa Rica

    Duplicate thread of
    http://swforum.sun.com/jive/thread.jspa?threadID=53424&tstart=0

  • Why does IE8 running on XP and Win7 Virtual Machine deletes the history while I'm still browsing the same site with client-side hashbang routing?

    Hello, 
    I have a asp.net mvc 5 web application running on .net 4.5 hosted on my local windows 8 machine on macbook pro using parallels. I'm running Internet Explorer 8 Version: 8.0.6001.18702 running on XP and Version: 8.0.7601.17514 running on Windows 7 Enterprise.
     I've got both of the virtual machine source from www.modern.ie 
    My web application is attempting to provide single page user experience by having some client-side routing by using '#!/xxx' hashbangs in the url so that I can get and post via ajax and change the hash in  the url without causing the page to reload. My
    code works fine in IE9+ and other modern browsers.
    However, I'm facing this strange issue in both of the IE8 versions. When running my demo app in IE8, I can login, view home page and can navigate from this home page to many details page. From the very first login page, all the pages are the result
    of form's post action via ajax which then expects html from server and loads in the DOM. The page structure of my app is like summary form => Details page (can also contain summary forms) => Details...so on.
    The above process works fine for few of the navigation steps. If I keep navigating from one page to another & go back, I have observed that the browser back button is removing the history items slowly. And a time comes when it totally forgets about
    all the history pages that I visited to reach this current page I'm viewing!! It treats like I have just started my browsing session and this current page is the first page I've landed on with no back option. Now if I again try to navigate, I have encountered
    problems like both the back and forward buttons gets disabled.
    Hope, my words above is clear enough to explain this problem. The same application works fine in IE9+ and other browsers. 
    Till now, I have tried following steps on both VMs: 
    Resetting IE8.
    Increasing disk space to 1024mb for temporary internet files storage.
    Setting 'Never' for Check for newer versions of stored pages.
    Disabling the automatic crash recovery feature didn't helped in this case. Found this on support website.
    Deleted temporary files, history, cache, etc many a times.
    Search many forums, posted question on StackOverflow, ASP.NET
    Forums - but didn't helped.
    Tried 'beforeunload' event of browser to see if any of the page is reloading because of submit or any other reason. But the page doesn't reload at all.
    The issue is browser forgets about the browsing history while I'm still browsing the same site. 
    Is there any possible fix for the issue above? Does IE8 have any issues as many ajax form post is happening on every page? 

    I have really tried many things to identify any possible reason of the issue above. Even rewrote all my javascript navigation code and checked server-side code.
    But the only place where I got the solution is at site: http://www.enhanceie.com/ie/bugs.asp which states that there is a bug in IE8 which reads like:
     IE0012: IE Travellog broken when navigating back/away from a page with
     a large POST form If there is a form input field with a value longer
     than 523,659 characters, when you navigate away from the page, IE may
     clear the current session's travellog (similar to history), disabling
     the back and forward buttons. Repros in IE6 & IE7.
    There are following solutions that you can go for:
    - Check for the input field that has lots of characters as mentioned above and solve your problem.
    - If you have control over the system where IE8 will be used, you can add the following registry key on that machine. There is no existing key, so, you need to add a new one:
    [HKEY_CURRENT_USER\Software\Microsoft\Windows\CurrentVersion\Explorer\TravelLog] "MaxSize"=dword:ffffffff
    I am really relaxed to find this information as I spent really long time to solve this issue. Hope this will help someone.

  • How to 'client-side include for secondary menu on page'

    Hi there.
    The following page refers:
    test page for early
    draft
    The linked page is in fact a DW CS3 tmp preview file, for
    that reason the CSS has been pulled into the html file by DW, it
    won't look that way in the eventual page file.
    After lengthy battles with the Spry Horizontal menu, I
    eventually resorted to Pop Menu Magic 2 (PPM2) to generate menus.
    I'm very happy with the product (I'm not looking for PPM2 support,
    but rather support pertaining to CSS-driven menu's in general
    relating to caching). I started out with a single main navigation
    menu, which was very comprehensive - but also very big (ito file
    size). I was hoping to cache a part of the menu on the client-side
    to reduce download times once the site visitor has downloaded the
    first page in the site (where he enters the site is not
    applicable). But that proved impossible (I think...).
    I then followed some advice to break up my main navigation
    menu, keep only the more important stuff in there and rather
    distribute the lower levels of the main menu into their own
    dedicated menus. From there the page linked above.
    The main menu at the top is fine, no problem there. On the
    left below the masthead I've created two menus, one above the other
    (it displays as a single menu but the first menu ends above the
    "Other tour lists" item). The first of these two menus contains the
    bulk of the heavy navigation elements (±160kb of the total
    page file size of 220kb). The div info that applies to this menu =
    <div id="p7PMM_2" class="p7PMMv06">. I'd love to cache this
    menu, either in full or the bulk thereof, on the client-side. That
    is implement a client-side include. When a visitor first enters our
    site the include file with the 160kb (secondary) menu gets
    downloaded together with CSS stylesheets, etc. When the visitor
    links through to a following page in our site, his machine
    (client-side) dishes up the menu to that next page without having
    to download it from our hosting server a second time.
    Is this possible and how? Any and all suggestions will be
    appreciated!
    Our clients are mostly wealthy and thus probably (virtually)
    all have high-speed broadband connections. I seriously doubt that
    any of them are running dial-up connections. Thus this issue is not
    going to sink our site. But I'd love to cut page file size if
    possible.
    Furthermore, I'm not too worried about the small portion of
    visitors that may have their javascript disabled in their browsers.
    The main navigation menu at the top of the page will still allow
    them to navigate the site. Much the same goes for SEO.
    I've tried one or two CSI javascripts and succeeded in
    actually getting the relevant tags and contents inserted in the
    html document, but the javascript would then 'break'. That is all
    the coding and menu contents would be there, but the javascript
    won't function (menu's won't pop / fly-out). I'm a little clueless
    when it comes to Javascript, and many other things :-), but I
    suspect that the Javascript in the include file does not get
    called. If the Javascript is kept in the html file, the menu
    content (elements) in the include file arrives in the html file
    after the relevant script has been called and doesn't get
    rendered??
    Once again any help / suggestions will be appreciated!

    > But that proved impossible (I
    > think...).
    Even without doing anything on your part, the images, the
    CSS, and the
    javascript is already cached client-side. You cannot cache
    any of the
    structural code client-side.
    Murray --- ICQ 71997575
    Adobe Community Expert
    (If you *MUST* email me, don't LAUGH when you do so!)
    ==================
    http://www.projectseven.com/go
    - DW FAQs, Tutorials & Resources
    http://www.dwfaq.com - DW FAQs,
    Tutorials & Resources
    ==================
    "afrilux" <[email protected]> wrote in
    message
    news:[email protected]...
    > Hi there.
    >
    > The following page refers:
    http://www.afrilux.co.za/test/1/dev2.htm
    >
    > The linked page is in fact a DW CS3 tmp preview file,
    for that reason the
    > CSS
    > has been pulled into the html file by DW, it won't look
    that way in the
    > eventual page file.
    >
    > After lengthy battles with the Spry Horizontal menu, I
    eventually resorted
    > to
    > Pop Menu Magic 2 (PPM2) to generate menus. I'm very
    happy with the product
    > (I'm
    > not looking for PPM2 support, but rather support
    pertaining to CSS-driven
    > menu's in general relating to caching). I started out
    with a single main
    > navigation menu, which was very comprehensive - but also
    very big (ito
    > file
    > size). I was hoping to cache a part of the menu on the
    client-side to
    > reduce
    > download times once the site visitor has downloaded the
    first page in the
    > site
    > (where he enters the site is not applicable). But that
    proved impossible
    > (I
    > think...).
    >
    > I then followed some advice to break up my main
    navigation menu, keep only
    > the
    > more important stuff in there and rather distribute the
    lower levels of
    > the
    > main menu into their own dedicated menus. From there the
    page linked
    > above.
    >
    > The main menu at the top is fine, no problem there. On
    the left below the
    > masthead I've created two menus, one above the other (it
    displays as a
    > single
    > menu but the first menu ends above the "Other tour
    lists" item). The first
    > of
    > these two menus contains the bulk of the heavy
    navigation elements (?160kb
    > of
    > the total page file size of 220kb). The div info that
    applies to this menu
    > =
    > <div id="p7PMM_2" class="p7PMMv06">. I'd love to
    cache this menu, either
    > in
    > full or the bulk thereof, on the client-side. That is
    implement a
    > client-side
    > include. When a visitor first enters our site the
    include file with the
    > 160kb
    > (secondary) menu gets downloaded together with CSS
    stylesheets, etc. When
    > the
    > visitor links through to a following page in our site,
    his machine
    > (client-side) dishes up the menu to that next page
    without having to
    > download
    > it from our hosting server a second time.
    >
    > Is this possible and how? Any and all suggestions will
    be appreciated!
    >
    > Our clients are mostly wealthy and thus probably
    (virtually) all have
    > high-speed broadband connections. I seriously doubt that
    any of them are
    > running dial-up connections. Thus this issue is not
    going to sink our
    > site. But
    > I'd love to cut page file size if possible.
    >
    > Furthermore, I'm not too worried about the small portion
    of visitors that
    > may
    > have their javascript disabled in their browsers. The
    main navigation menu
    > at
    > the top of the page will still allow them to navigate
    the site. Much the
    > same
    > goes for SEO.
    >
    > I've tried one or two CSI javascripts and succeeded in
    actually getting
    > the
    > relevant tags and contents inserted in the html
    document, but the
    > javascript
    > would then 'break'. That is all the coding and menu
    contents would be
    > there,
    > but the javascript won't function (menu's won't pop /
    fly-out). I'm a
    > little
    > clueless when it comes to Javascript, and many other
    things :-), but I
    > suspect
    > that the Javascript in the include file does not get
    called. If the
    > Javascript
    > is kept in the html file, the menu content (elements) in
    the include file
    > arrives in the html file after the relevant script has
    been called and
    > doesn't
    > get rendered??
    >
    > Once again any help / suggestions will be appreciated!
    >
    >

Maybe you are looking for

  • Trying to restoring a late 2013 macbook pro from a 2010 macbook pro

    Hi I have problems with my 13"retina macbook pro from late 2013.There is something on it that makes it 4 times slower in boot up than the other two  one from 2009 and 20010 all are running 10.9.5, so I thought I would clone one of them that has all t

  • Disk Utility on Time Capsule

    I have and older (non-dual band) Time Capsule and I want to reformat to extended journaled so I can use it for transferring Aperture 3.0 projects. I dimly recall using FAT32 format when I first got it (don't ask). However, the time capsule is not app

  • New GL activated without data migration

    Hi all, we have some three projects lined up, one of it is new gl, due to some reasons we have postponed the go live of new gl but other two projects wants to go ahead with their go live, so we provided them an environment where new gl is active but

  • Is this good? (RAM test)

    sandra screen: i can't compare it as you can see, i get N/A :( don't know why, but maybe someone else can tell me how im doing for an Nforce2 system i would like to know how this is compared with a Intel Pentium 4 system.

  • Callback script - Just last Callback number works!

    Hi, I have created a Callback script, the issue is if there is one CallBack number, it works fine, if there are multiple Callback Number, as soon as Agent is ready, his Phone rings but when he picks it up, there is fast busy, however, the last Call b