DB Cache issue

Hello, I am troubleshooting performance issues on a customer database. I was looking at issues with disk i/o and was wondering if this output may indicate an issue. There are a few tablespaces that use a 16k block size and several that use an 8k block size. Below indicates that db_8k_cache_size is 0. Does that mean that there is no db cache for the 8k tables and indexes. db_cache_size is also 0.
NAME                                 TYPE        VALUE
=======================================================
db_16k_cache_size                    big integer 687865856
db_2k_cache_size                     big integer 0
db_32k_cache_size                    big integer 0
db_4k_cache_size                     big integer 0
db_8k_cache_size                     big integer 0
db_cache_size                        big integer 0
db_keep_cache_size                   big integer 0
db_recycle_cache_size                big integer 0

Have you identified which datafile or tempfile this I/O is related to? You can check DB Cache Advice
SET PAGES 44
SET LINES 140
COLUMN size_for_estimate FORMAT 9,999,999,999,999 heading 'Cache Size (MB)'
COLUMN buffers_for_estimate FORMAT 9,999,999,999 heading 'Buffers'
COLUMN estd_physical_read_factor FORMAT 999.90 heading 'Estd Phys|Read Factor'
COLUMN estd_physical_reads FORMAT 9,999,999,999 heading 'Estd Phys| Reads'
SELECT size_for_estimate, buffers_for_estimate, estd_physical_read_factor, estd_physical_reads
FROM V$DB_CACHE_ADVICE
WHERE name = 'DEFAULT'
AND block_size = (SELECT value FROM V$PARAMETER WHERE name = 'db_block_size')
AND advice_status = 'ON';
But do not focus on only DB Cache. Also check PGA usage if it is Tempfile usage that is causing it.
column name  format a60
column value format 9,999,999,999,999,999
select * from gv$pgastat order by inst_id,name
column BYTES_PROCESSED format 9,999,999,999,999,999
column EST_RW_EXTRA_BYTES format 9,999,999,999,999,999
select inst_id,round(pga_target_for_estimate/1024/1024) as target_size_MB,
              bytes_processed,estd_extra_bytes_rw as est_rw_extra_bytes,
              estd_pga_cache_hit_percentage as est_hit_pct,
              estd_overalloc_count as est_overalloc
from gv$pga_target_advice  order by inst_id,target_size_mb
/

Similar Messages

  • Directory Caching issue with Cisco Jabber client for Windows

    Hi ,
    I am facing cache issue with Cisco Jabber client for Windows. If I do any change related to modification or deletion of contacts in Active Directory/ Callmanager, it does not reflect in the Jabber. Because jabber takes the contacts from the locally stored cache file in the Windows system.
    Every time I have to remove the cache file to overcome this issue, practically it's not possible to do the same with all the Widows users. As, if any employee leaves the company and still I can see his contact appears in the "Cisco Jabber client". I have not seen this issue with Android/Apple iOS.
    Is there any automated way to remove the cache file? 
    Here is the detail of CUCM,Presence and Jabber.
    CUCM version: 9.1.x
    Presence          : 9.1.X
    Jabber              : 10.5 and 10.6

    Hello
    On our environment we had to install a dedicated Microsoft Certificate Authority "just for Cisco Jabber usage" to house the
    Network Device Enrollment Service.
    Our certificate for the CUPS were generated on this Certification Authority too.
    I discussed this certificate matter with my colleagues this afternoon and nobody seems to remember how these certificates were deployed into the
    Enterprise Trust store for the users.
    But I think they asked all 400 users to accept the 3 certificates by answering "yes" to the popup instead of using a script deployed by GPO...
    I wish you success with that deployment and really hope you have a technical partner that *Knows* this subject.
    Our partner left us alone with that unfortunately.
    Florent
    EDIT: If the "Certutil script method" works, please let me know. This could be useful in our own deployment.

  • Sql query cache issue

    I am trying to see the log file in Manage sessions for the sql query in Answers. I see that if we run the same report multiple times, the sql query is showing up only the first time. Second time if I run it is not showing up. If I do a brand new report with diff columns picked it is giving me the sql then. Where do I set this option to show the sql query everytime I run a report even if it is the same report run multiple times. Is this caching issue?

    It shouldn't.... Have you unchecked the "Cache" on the physical layer for this table? If you go onto the Advanced tab, is the option "Bypass the Oracle BI cache" checked?

  • WEB u2013 I cache issue.

    We have BO XI 3.1 PF 1.7 and SAP BW 7.0 SP 18
    We are facing strong caching issue  with WEB- I XI 3.1 , what we observed is it cache previously fetched data and does not recognize data been updated in BW Infoprovider ( in our case we have BEx query on Direct update DSO )
    I have tried u2013
    1.     I made caching inactive for that BEx query in BW side.
    2.     I have check caching setting in BO , they are not helping me much to in active cache in BO
    On executing BEx query in BW, I have checked RSRT cash monitor; the query result is not getting cached, and get up-to-date data. Query works perfectly file.
    Also observed Events in RSDDSTAT_OLAP table, I could find very well Event 9000. This shows me that BEx query hits database and bring data back.
    When I execute BO report on the top of same BEx query, I do not get up to date data , it shows me previously fetched data and does not recognize data been updated in BW.
    Now question is where this data being getting cashed , and how to make that inactive to make sure I do get up to date data in WEB-I .
    Many thanks for your kind response.
    Regards
    Ashutosh D

    HI Ashutosh,
    This would be stored in the C:\Program Files\Business Objects\BusinessObjects Enterprise 12.0\Data\servername_6400\storage folder.  Try renaming this folder and then testing again.  If this allows for updated records, then that proves that the Webi Processing Server is getting the cached data from this directory.
    There are also a few settings you can try to disable in the Webi Processing Server settings within the CMC:
    Disable cache Sharing (Checked)
    Enable Real-Time Caching (Unchecked)
    Enable Document Cache (Unchecked)
    Cache Timeout (1 min) - sets the cache to expire after 1 min.
    Try playing with these settings and see if that resolves the issue. 
    Thanks
    Jb

  • SharedObject and Caching issues

    Background:
         I am working on a project where I am using sharedObject to store data for a favorites list. The project contains 2 swfs, one in AS 3.0 (using Flex 4.5) and the other an AS 2.0 flash 8 project.  Basically the AS 2.0 project read and writes data to the shared object and the AS 3.0 project just reads the data.
    Setup:
         OS: Windows 7
         Flash Player: 10.3     Browsers: IE 9, IE 8, Chrome 14, and some misc Firefox testing -- Chrome seems to be working the best.  
    Problems:
         First off it has been extremely hard to get consistent results across different computers so if anyone has a suggestion as to why that would be (other than different browsers, different flash player versions, different flash player settings/global settings and different OS versions) because I have been making sure all that stuff is consistent across devices for testing purposes.
         1. It appears that some computers/browers won't store the shared object at all even though everything to allow it is turned on.
         2. Some computers/browsers will store it, but won't update until I refresh the page
    General Questions:
         1. Is the storage of sharedObject object data in a .sol file configured the same in both AS 2.0 and AS 3.0 and if so, would this be a means of communicating between two swfs running off the same domain?
         2. Is it possible to have a browser cache sharedObject data(the .sol file)?
          3. I have noticed as new flash player releases are coming out(especially recently i.e. 10.3) I have had to make a lot of repairs to the actionscipt 2 stuff.  Is actionscript 2 support slowly being phased out? Is anyone else experiencing the same problems? 
    Possible Solutions/Findings:
         I am starting to believe that it has something to do with browser caching since Chrome seems to perform the best and Chrome has performed best when it comes to browser caching issues.
    ** code is available if needed but since it was working before I don't think it is relevant.

    Hello,
    #1
    are you using explicit call to flush method immediately afer you're writing data to reference to SharedObject?
    http://help.adobe.com/en_US/FlashPlatform/reference/actionscript/3/flash/net/SharedObject. html#flush()
    If not you could have different data read by different browsers sessions run at the same time (e.g. IE/FireFox both hosting your test Flash movies) - as data are flushed only when given Flash runtime session is to be terminated (e.g. movie is to be unloaded) and in other cases as outlined in documentation.
    #2
    the SharedObject is reference to binary file stored on local machine by Flash runtime (Air runtime too). It has nothing similar to web-fetched content to be cached:
    http://en.wikipedia.org/wiki/Local_Shared_Object
    regards,
    Peter

  • SecurityDomain and Caching Issues

    I am running into some caching issues when setting the securityDomain of an imported SWF to match the calling SWF file and I was curious if anyone had any ideas on how to get around this issue.
    Here is the scenario:
    A.swf is launched on Domain A, it sends a Loader request for B.swf on Domain B.
    B.swf will be updated frequently so caching is disabled, A.swf however can be cached.
    B.swf has some ExternalInterface.calls so it requires being in the same securityDomain as A.swf otherwise it will receive a Security Error #2006.
    The code I am using for this is fairly straightforward:
    ActionScript Code:
    var request:URLRequest = new URLRequest("http://domainB/B.swf");
    var loader:Loader = new Loader();
    var context:LoaderContext = new LoaderContext();
    context.securityDomain = SecurityDomain.currentDomain;
    loader.load(request, context);
    I believe B.swf is inheriting the caching setting of A.swf because it is residing in the same securityDomain. If I make a small update to B.swf and refresh A.swf in a browser it will not load the B.swf updates until I clear cache.
    If I get rid of the securityDomain context on the load it will always update B.swf with the most current version, but I run into security issues with ExternalInterface.
    ActionScript Code:
    var request:URLRequest = new URLRequest("http://domainB/B.swf");
    var loader:Loader = new Loader();
    loader.load(request);
    I have tried appending random strings to the end of the URLRequest while using the securityDomain context but it will always used a cached version of B.swf. I have also tried loading B.swf into a ByteArray then using loadBytes on a Loader but that didn't work either.
    Any ideas are appreciated!

    Hello,
    #1
    are you using explicit call to flush method immediately afer you're writing data to reference to SharedObject?
    http://help.adobe.com/en_US/FlashPlatform/reference/actionscript/3/flash/net/SharedObject. html#flush()
    If not you could have different data read by different browsers sessions run at the same time (e.g. IE/FireFox both hosting your test Flash movies) - as data are flushed only when given Flash runtime session is to be terminated (e.g. movie is to be unloaded) and in other cases as outlined in documentation.
    #2
    the SharedObject is reference to binary file stored on local machine by Flash runtime (Air runtime too). It has nothing similar to web-fetched content to be cached:
    http://en.wikipedia.org/wiki/Local_Shared_Object
    regards,
    Peter

  • Caching issue in jsp or servlet

    I am not sure where I should post the issue that I have. I have a j2ee web app (struts) running on Apache Web Server + Glassfish + MySQL. My server has multi core.
    The problem is that whenever I insert a new data or delete it from a table, I don't see the change right away. Sometimes, I see the change. Sometimes I don't. It's very inconsistent.
    If the old data is cached in JSP, I shouldn't see the change in log file, but I do see it even in the log file.
    For example, I have a page managing user's folders. when I delete a certain folder name from jsp page, this folder is deleted from a db table. but when I refresh the page, I still see the folder name that is not supposed to show up. or when I go to a different page and come back to this page, I don't see it. The behavior is very inconsistent.
    If it's a browser caching issue, I don't think that I should see it in the log file, but I still see the folder name(which is supposed to be deleted) in the log file.
    I am including these lines in all included jsp pages.
    response.setHeader("Cache-Control","no-cache");
    response.setHeader("Pragma","no-cache");
    response.setDateHeader ("Expires", -1);
    does anybody have any opinion about this?
    It's hard to debug it and describe the behavior.
    But it would be very helpful if someone who had a same experience about this explains about this and tells me how to fix it.
    Thanks.

    caesarkim1 wrote:
    I am including these lines in all included jsp pages.
    response.setHeader("Cache-Control","no-cache");
    response.setHeader("Pragma","no-cache");
    response.setDateHeader ("Expires", -1); Instead of including these lines in all jsp's, make a filter and add these lines to it.

  • Acrobat Connect Pro LMS 7.5 server cache issue - displaying old content

    Our Adobe Acrobat Connect Pro server is showing old Captivate-created content from about 4-6 weeks ago.
    I loaded 35+ sets of Captivate 3 (SCORM 1.2, HTML, zipped) content onto our Acrobat Connect Pro LMS about 6 weeks ago.
    I converted all training content from Captivate 3.0.1 to 4.0.1, 4 weeks ago by opening each file in Captivate 4 and following prompts to "Save As" new files with different file names.
    I reloaded all content 4 weeks ago, and again 2 weeks ago.
    I reloaded about half of all content again last week.
    End User Acceptance testing performed this week showed that most of the courses are showing old content, ranging from 2-6 weeks old
    Attempted fixes and workarounds:
    Deleting content entirely and reloaded from scratch - this will not work long term, as we lose usage data each time we reload completely new files.
    Contacted Adobe, provided times to track incidents of the issue.  We reached Tier 2 - who told us it was our problem and that everything appeared to be working fine from their side.
    New workaround - load new content and reattach course to new content.  This presents the same long term issue as the first workaround, but enables us to retain older versions of content in the system, should we need to revert or report on it.
    Gaining server side access is a bit challenging due to the hosting situation we have, so I am looking (ideally) for a solution that can be performed from the Administrator/Author Frontend.  However, I want to learn the real cause of the problem, wherever it might reside, so that it can be properly corrected and avoided in the future.  I am calling this a server cache issue, as it seems the server has somewhere retained unwanted old versions of content, preventing current content from being displayed to end users.  Viewing content as an end user = see old content.  Viewing content from the Content area (Author view) shows the current files, so I know they are on the server and are loading correctly, up to a point.
    I am preparing all content for another round of loading/reloading due to other issues and updates, so republishing and reloading all 35+ files into the LMS is unavoidable at this point.
    This issue is keeping our LMS from launching to several thousand users across the country, so any suggestions or helpful tips are much appreciated.

    I think I have isolated the source of this problem. It's the Pitstop Professional 9 plug in. I un-installed this, and everything opens quicker than greased lightning. I re-installed it and it's back to slowsville.
    Unfortunately Pitstop is essential to my workflow.
    Until recently I did my pre-press on a Mac G5 with Acrobat Pro 7 and Pitstop 6.5. I never had this problem with slow file opening. But it seems that the delays would occur when I used the plug-in with large complex files.. So it would open files as fast as you'd expect from an elderly machine. But starting to use Pitstop would result in a prolonged period of staring at a spinning beachball.
    I wonder is there any way to stop the Pitstop plug-in from initializing until it is used? So the plug-in stays inert until you select the tool from from the menus.

  • IE Caching Issue while closing and opening a UI thrice or more

    Hi Everybody..
    Greetings of the day..
    I am working on a web application (Can not disclose the name) where at certain level a word document generated contains a customized toolbar.
    It contains one print option, when clicked opens a UI which shows default printer available (Its different from our normal printing).
    When I repeat opening and then closing of this UI thrice or more times, the entire application gets hanged and I have
    to terminate it by killing the associated process from task manager.
    When this issue was analysed our by onsite team they arrived at a conclusion saying that its an IE caching issue.
    But upto my knowledge I know that any application which is .SSL secured, by default IE never does cache control.
    What could be the solution if application is not .SSL secured or if it is.
    Kindly help asap.
    Thanks.

    Hi all
    Did anyone ever manage to fix this? I'm having this issue across multiple iPads. Even if i completely reset the iPad (erasing all content and settings) it does it again when restored.
    I've just got an iPad Air and that does it too. I even created a new Apple ID to see if that would work (trying to avoid any kind of caching aross iCloud, etc) and it does it straight away in those books. It's massivley frsustrating.
    I have noticed it only seems to do it for iBooks formatted books, not ePubs.
    Anyone have a resolution - it's quite embarrassing when you're doing a presentation about book features and the audience assumes youve only the wrong book each time. More annoying when clsoing the book and the wrong cover appears again, but also takes you to that books collection page, rather than the one you were in.

  • Possible session caching issue in SSRS2014

    Using custom FormsAuth, User A can sign into our own main asp.net mvc app (WIF cookie), then SSRS (FormsAuth cookie) and all is well.  Here is where things go bad.  User A signs out of our main application (WIF cookie deleted) then back in into
    our main application as User B then back into SSRS.  SSRS report that displays User!UserID show UserA instead of current User B.  Its like there is either a session or cookie caching issue going on but not for sure.  
    1. What is the proper way to sign out of SSRS and prevent session caching?
    2. Do I need to worry about making my SSRS logon page non-cacheable?  If so, what is the recommended way of doing this?
    thanks
    scott

    Hi scott_m,
    According to your description, you used custom FormsAuthentication in Reporting Services, after user A sign out the application an sign in as user B, SSRS built-in user is shows user A instead of user B.
    Based on my search, once we configured SSRS to use Custom (Forms) authentication by deploying a custom security extension, we can logon to MS Report Manager (MSRM) using credentials of our custom security framework via a logon web page. But there is no way
    to logout or to expire the authentication cookie, so we need to close the browser manually. As a workaround, we can add a logout button to the Report Manager which is using Forms Authentication, then use code to clear the cookie and redirect to home page.
    In addition, if you extend Reporting Services to use Forms Authentication, it’s better to use Secure Sockets Layer (SSL) for all communications with the report server to prevent malicious users from gaining access to another user's cookie. SSL enables clients
    and a report server to authenticate each other and to ensure that no other computers can read the contents of communications between the two computers. All data sent from a client through an SSL connection is encrypted so that malicious users cannot intercept
    passwords or data sent to a report server.
    Here is a relevant thread you can reference:
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/5e33949d-7757-45d1-9c43-6dc3911c3ced/how-do-you-logout-of-report-manager
    For more information about Forms Authentication, please refer to the following document:
    https://technet.microsoft.com/en-us/library/aa902691%28v=sql.80%29.aspx?f=255&MSPPError=-2147217396
    If you have any more questions, please feel free to ask.
    Thanks,
    Wendy Fu
    If you have any feedback on our support, please click
    here.
    Wendy Fu
    TechNet Community Support

  • Parallel Caching Issue

    We have found issues with the parallel loading of the OLAP cache using reporting agent jobs where entries are not populated correctly.We rely on the OLAP cache for delivering the very high levels of concurrency during the peak times.
    Once the main batch data loading has been completed we run some background reporting agent jobs to pre-populate the OLAP cach.Each job contains a web application template that holds 1 or more queries and will process 1500+ stores for each run.
    We have different reporting agent jobs for the different web application templates, but we have discovered that if we run these jobs in parallel we do not get the full benefit of the OLAP cache compared to if we run them sequentially.
    If we run the jobs in parallel, when we look in RSRCACHE for these queries it would appear to have populated correctly, but when we check RSDDSTATS for the query performance the following day, we can see that a large number of the
    stores still hit the database and did not benefit from the cache entries. Sometime this can be as much as 60% failure to hit the cache.
    If we run the same job sequentially, then check RSDDSTATS the following
    day, we can see 100% success rate with every store hitting the OLAP cache.
    Is anyone able to advise how we can resolve this parallel caching issue?

    Hi Ganesh,
    I am currently having similar trouble with TBB1 where I receive error message below:
    TRL intialization for MM, FX, Derivatives, co. code WTRD, valn area 003 is not yet complete
    Message no. TPM_TRL052
    Are you familiar with this issue? any help would be greatly appreciated! i hope that you can help..

  • EJB CMP remove create cache issue? DuplicateKeyException

    EJB CMP remove create cache issue? DuplicateKeyException
    Hi I have an EJB 2.1 application using CMP. Most things work fine. But if I in a transaction tries to remove a bean and create a new one with the same primary key I get a DuplicateKeyException:
    2007-11-16 09:25:31,963 ERROR [RMICallHandler-6] AdminGroupData_ConcreteSubClass147 - Error adding AccessRules:
    javax.ejb.DuplicateKeyException: Exception [EJB - 10007]: Exception creating bean of type [AccessRulesData]. Bean already exists.
    at oracle.toplink.internal.ejb.cmp.EJBExceptionFactory.duplicateKeyException(EJBExceptionFactory.java:195)
    I suspect that the remove call only removes from the cache (until commit), but that the create call checks the database or something?
    My code is simple like the following:
    AdminPreferencesDataLocal apdata1 = adminpreferenceshome.findByPrimaryKey(certificatefingerprint);
    adminpreferenceshome.remove(certificatefingerprint);
    adminpreferenceshome.create(certificatefingerprint,newadminpreference);
    Is there some configuration I can set in toplink-ejb-jar.xml to fix this?
    I use OC4j 10.1.3.3
    Cheers,
    Tomas

    The bean.remove() was executed but the sql DELETE was executed to the database as the result, oc4j manages all sql statements and they are optimized to commit to the database in batch at once when the transaction commit. Oc4j ejb container is smart and preventing you from creating the same entity cached in the transaction because the sql delete was not really committed to the database when the create is called.
    My guess is that the reason it works with IBM WAS was because it issued sql for each remove/create call right away instead of doing them in batch as oc4j, or your WAS remove then create were called in separate transactions.

  • Is there a caching issue with DPS

    Facing a unique issue. When I am creating a folio it references to redundent files while build which are not used in the folio, and then the build does not upload to folio producer. Could there be a caching issue and how can I clear the DPS cache (Not the indesign cache.) After some time it shows the dreaded network error message.

    Hopefully you are using Windows:
    Close InDesign and follow the steps below:
        1.       Go to C: drive.
        2.       Click on Organize button on top left hand side.
        3.       Click on Folder search option.
        4.       Click on view tab and select show all hidden items.
        5.       Apply and ok.
        6.       Then  c:\users\username\AppData\Roaming\StageManager.BD092818F67280F4B42B04 877600987F0111B594.1\Local Store.
        7.       Delete dmp and shared object folder.
        8.       Empty Recycle Bin.
    Then re-launch InDesign

  • TOC caching issue on IE

    Hello all
    RH10 to webhelp
    I just received a call from an ongoing client. He's reporting that they are experiencing TOC caching issues in IE only. A topic and a book in the TOC that we removed in the latest updates are still appearing in the TOC, but obviously are dead links to topics that no longer exist. He said they cleared the folder and replaced it entirely with the new output files. They clear the cache but they keep showing up. He says they don't have this issue in Chrome.
    Any ideas?
    Thanks,
    Laurie

    He would ideally be looking for a way to not cache or to automatically clear the cache so his clients don't have to manually do it each time (they likely wouldn't know that they even would need to, and would just be confused by orphan links in the TOC).
    I found info re using meta tags to disable the cache (Disable Browser Caching with Meta HTML Tags [Cristian Sulea]). I inserted the tags into the master page header according to the instructions, but curiously they are not included in the output topics. I manually inserted it into the index.htm, and have sent the output to the client to test. I can't test it because I'm not experiencing the issue.

  • BitmapImage cache issue

    On WIndows 8 RP, I create a BitmapImage with the Uri of a file on disk. The user has the ability to change the image file (using a picker which will replace the file on disk that is currently being used). The issue is that even when the BitmapImage is recreated,
    because the Uri is pointing to the same file, the new updated image is never loaded. The only way to get the new image to load is to close and reopen the app. Clearly there is a cache issue.
    I have two questions:
    1. Is the issue with the cache of the BitmapImage that I construct using the Uri? Or is the issue with the cache of the Image UIElement that is displayed using the BitmapImage I mentioned previously?
    2. It looks like I have far fewer options when dealing with BitmapImages on WinRT. I cannot do bi.BeginInit() or bi.CacheOption. All I can do is bi.CreateOptions = BitmapCreateOptions.IgnoreImageCache; ... but this does not help.
    Here is how I'm constructing my BitmapImage:
    BitmapImage bi = new BitmapImage();bi.CreateOptions = BitmapCreateOptions.IgnoreImageCache;bi.UriSource = new Uri(currentFile.Path, UriKind.Absolute);
    Here is how I'm setting the grid cell BitmapImage (which is binded to a UIElement Image):
    gridViewCellObject.tileImage = bi;
    Here is the code for the gridViewCellObject tileImage get/set
    public BitmapImage muteImage
    get { return _muteImage; }
    set
    _muteImage = value;
    // Call NotifyPropertyChanged when the source property
    // is updated.
    NotifyPropertyChanged("muteImage");
    Here is the XAML which binds the Image to tileImage:
    <Image Source="{Binding tileImage, Mode=OneWay}" Width="140" Height="140" HorizontalAlignment="Center" VerticalAlignment="Top" Margin="0,10,0,0" Stretch="Fill"/>
    Any advice on how get around this caching?
    Bryan L. Keller

    Here is the solution. As the comments in my code suggest, this is ridiculous and shouldn't even be necessary.
    Here is the revised getPhotos method. I copy each bitmap from its original source to the temp folder. I give it a random name that will never be repeated while the application is open. This is done by getting some information about the current system time.
    I calculated it out to be a 1 in 10^-8 chance of getting a duplicate file name using my method of naming.
    public async Task<List<BitmapImage>> getPhotos()
    photos.Clear();
    IReadOnlyList<IStorageFile> files = (IReadOnlyList<IStorageFile>) await folderHierarchy.Last().GetFilesAsync(); //gets storagefiles of original png files
    DateTime currentTime;
    foreach (StorageFile currentFile in files)
    if (currentFile.Name.EndsWith(".png")) //get only png files
    //copy to and read from temp to prevent bitmap caching. (It's ridiculous that I even need to do this)
    currentTime = DateTime.Now;
    String tempFileName = currentTime.Hour + " " + currentTime.Minute + " " + currentTime.Second + " " + currentTime.Millisecond; //generate unique name
    await currentFile.CopyAsync(Windows.Storage.ApplicationData.Current.TemporaryFolder, tempFileName); //copy to temp with unique name
    //because the name is new each time, the bitmap will never be cached. photos.Add(new BitmapImage(new Uri((await Windows.Storage.ApplicationData.Current.TemporaryFolder.GetFileAsync(tempFileName)).Path))); //add bitmap to array of photos. This is used by the rest of the app to populate the gridview
    return photos; //return array of photos
    Microsoft, I shouldn't need to waste time solving your API issues. This stuff should have been fixed in one of the Developer Preview builds, NOT the Release preview.
    Get your act together so developers can do their job.
    Bryan L. Keller

  • CC2014 Cache Issue/Bug?

    Has anyone had any Cache issues withe After Effects CC2014? I'm using the latest version and myself and a co-worker are having cache-related(we think) issues whenever we re-time something.
    Specifically I was trying to create a looping comp using a 1second comp that uses the loopOut expression. It's basically a rangefinder set of lines that continuously move to the right. Inside the 1s comp it loops perfectly, but when I use loopOut, no matter what I do, or check, it creates a huge amount of motion blur on the last frame of the comp(even though it doesn't happen at all in the precomp)
    My coworker was re-timing a scene with a person using twixtor pro, the footage is touched up to remove some wrinkles under the subjects eyes, but when twixtor is applied it appears to be using the wrinkled-eyed footage. It's 100% opacity and doesn't have a blend mode or any expressions. The original un-retouched image with wrinkles is a hidden(not shy) layer under the new footage for timing/matching purposes.
    When the original footage is removed from the comp and the cache is cleared it seems to clear up.
    Any thoughts/suggestions?

    hey sandeep--
    if i'm understanding you correctly, this does indeed sound like a cache issue. based on what i'm thinking is your expected behavior, it sounds as if you're forgetting to clear the cache for your other PARAM items when a value is selected from one of your PARAM items that, from a logical perspective, comes before it. so to use your example...
    After I get Report based on (1). And then go back & select (2). It works fine. But when I go back & select (1) again, it still shows the Regions displayed based on the (2).
    ...i'd say you'd want to add a process to your page that fires after submit and clears the cache for PARAM2 when the value of :REQUEST = 'PARAM1'.
    hope this helps,
    raj
    ps-if i've misunderstood your issue, please explain it in more detail.

Maybe you are looking for

  • FI_GL Line items-Opening Balance

    Hi Frieds, We have installed 0FI_GL_4 datasource and loaded data 0FIGL_O02 DSO. We have bulit a cube based the 0FIGL_O02 and passing the data from 0FIGL_O02 to Newly created custom cube.  We are now building report called Trail Balance. In the report

  • Using the Apple TV can I stream face time though to the TV the apple TV is hooked up to?

    Usung the apple TV can I stream face time though to the TV the apple TV is hooked up to?

  • Problem in starting sdm

    HI Help me . I have problem in sdm server. initially it was started.but second time onwards it doent start. what is the problem for that. pls help me... regards BHI

  • Photo album support needed

    Apple - i can take pics, edit red eye, crop - which is cool.  But for the love of God, when will you improve photo album managent via the device?  Imagine a world where I can select a photo(s) and hit "MOVE TO ALBUM..." and it really moves it!  I mea

  • Where I can download OCA 10g DBA dumps

    Hello... I want dumps for OCA certification in Oracle 10g in DBA track. So I need dumps for that exam. Can anyone suggest me any site where I can download it. Prashant T.