Performance issue when accessing a CRL file

Hi,
In an application we receive sign and verify files(ordinary information files) as seperate processes. When signing files we do a check on the CRL of the signing certificate. Below are the steps in which the check is done:
1- Receive a file. Get its signing certificate from a repository.
2-save/cache the CRL file in a directory(we do check if the file already exists when creating). We have configured a time out period with which the CRL will be updated in case of modifications.
File file = new File(cacheDirSystem.getProperty("file.separator")+crlUrl(url));+
fs = new FileOutputStream(file);
X509CRL crl.writeTo(fs);
3-Get the CRL Distribution Point from the signing certificate and load the file that was saved to check with the Certificate Revocation Lists.
String as[] = getDistributionPointNameURIs((GeneralNames)dpName);
File file = new File(cacheDirSystem.getProperty("file.separator")+modifyUrl(as));+
lastModified=file.lastModified();
fstream = new FileInputStream(file);
crlObject.crl=new X509CRL(fstream);
When running signing processes with more than 150 files(the file size can go upto 2MB) over a period of time, the CRL is sometimes not found and InvalidCRLException is thrown. I think its some problem with accessing the CRL file concurrently since it could also be updated at some point.
Im trying to figure out if a file lock or some other way would help to reslove this. I need help/suggestions on this.
Thanks in advance

I think its some problem with accessing the CRL file concurrentlyUndoubtedly. You should write the new data to a new file and do a delete/rename operation. All clients should cope with not finding the CRL file by retrying a couple of times in case they tried between the delete and the rename.
This is a file handling question, or concurrency question, not really a security question.

Similar Messages

  • Performance issues while accessing the Confirm/Goods Services' transaction

    Hello
    We are using SRM 4.0 , through Enterprise Portal 7.0.
    Many of our users are crippled by Performance issues when accessing the Confirm/Goods Services tab( Transaction bbpcf02).
    The system simply clocks and would never show the screen.
    This problem occurs for some users all the time, and some users for some time.
    It's not related to the User's machine as others are able to access it fast using the same machine.
    It is also not dependent on the data size (i.e.no . of confirmations created by the user).
    We would like to know why only some users are suffering more pronouncedly, and why is this transaction generally slower than all others.
    Any directions for finding the Probable cause will be highly rewarded.
    Thanks
    Kedar

    Hi Kedar,
    Please go through the following OSS Notes:
    Note 610805 - Performance problems in goods receipt
    Note 885409 - BBPCF02: The search for confirmation and roles is slow
    Note 1258830 - BBPCF02: Display/Process confirmation response time is slow
    Thanks,
    Pradeep

  • Performance issues when creating a Report / Query in Discoverer

    Hi forum,
    Hope you are can help, it involves a performance issues when creating a Report / Query.
    I have a Discoverer Report that currently takes less than 5 seconds to run. After I add a condition to bring back Batch Status that = ‘Posted’ we cancelled the query after reaching 20 minutes as this is way too long. If I remove the condition the query time goes back to less than 5 seconds.
    Please see attached the SQL Inspector Plan:
    Before Condition
    SELECT STATEMENT
    SORT GROUP BY
    VIEW SYS
    SORT GROUP BY
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    TABLE ACCESS BY INDEX ROWID GL.GL_CODE_COMBINATIONS
    AND-EQUAL
    INDEX RANGE SCAN GL.GL_CODE_COMBINATIONS_N2
    INDEX RANGE SCAN GL.GL_CODE_COMBINATIONS_N1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUES_N1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUE_SETS
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUE_SETS_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES_TL
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUES_TL_U1
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUE_NORM_HIER_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_LINES
    INDEX RANGE SCAN GL.GL_JE_LINES_N1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_HEADERS
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_DAILY_CONVERSION_TYPES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_SOURCES_TL
    INDEX UNIQUE SCAN GL.GL_JE_SOURCES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_CATEGORIES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    INDEX UNIQUE SCAN GL.GL_BUDGET_VERSIONS_U1
    INDEX UNIQUE SCAN GL.GL_ENCUMBRANCE_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_BATCHES
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_PERIODS
    INDEX RANGE SCAN GL.GL_PERIODS_U1
    After Condition
    SELECT STATEMENT
    SORT GROUP BY
    VIEW SYS
    SORT GROUP BY
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    NESTED LOOPS
    NESTED LOOPS OUTER
    NESTED LOOPS
    TABLE ACCESS FULL GL.GL_JE_BATCHES
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_HEADERS
    INDEX RANGE SCAN GL.GL_JE_HEADERS_N1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    INDEX UNIQUE SCAN GL.GL_ENCUMBRANCE_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_DAILY_CONVERSION_TYPES_U1
    INDEX UNIQUE SCAN GL.GL_BUDGET_VERSIONS_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_SOURCES_TL
    INDEX UNIQUE SCAN GL.GL_JE_SOURCES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_CATEGORIES_TL_U1
    INDEX UNIQUE SCAN GL.GL_JE_BATCHES_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_JE_LINES
    INDEX RANGE SCAN GL.GL_JE_LINES_U1
    INDEX UNIQUE SCAN GL.GL_SETS_OF_BOOKS_U2
    TABLE ACCESS BY INDEX ROWID GL.GL_CODE_COMBINATIONS
    INDEX UNIQUE SCAN GL.GL_CODE_COMBINATIONS_U1
    TABLE ACCESS BY INDEX ROWID GL.GL_PERIODS
    INDEX RANGE SCAN GL.GL_PERIODS_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUES_N1
    INDEX RANGE SCAN APPLSYS.FND_FLEX_VALUE_NORM_HIER_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUES_TL
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUES_TL_U1
    TABLE ACCESS BY INDEX ROWID APPLSYS.FND_FLEX_VALUE_SETS
    INDEX UNIQUE SCAN APPLSYS.FND_FLEX_VALUE_SETS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    INDEX UNIQUE SCAN GL.GL_JE_HEADERS_U1
    Is there anything i can do in Discoverer Desktop / Administration to avoid this problem.
    Many thanks,
    Lance

    Hi Rod,
    I've tried the condition (Batch Status||'' = 'Posted') as you suggested, but the qeury time is still over 20 mins. To test i changed it to (Batch Status||'' = 'Unposted') and the query was returned within seconds again.
    I’ve been doing some more digging and have found the database view that is linked to the Journal Batches folder. See below.
    I think the problem is with the column using DECODE. When querying the column in TOAD the value of ‘P’ is returned. But in discoverer the condition is done on the value ‘Posted’. I’m not too sure how DECODE works, but think this could be the causing some sort of issue with Full Table Scans. How do we get around this?
    Lance
    DECODE( JOURNAL_BATCH1.STATUS,
    '+', 'Unable to validate or create CTA',
    '+*', 'Was unable to validate or create CTA',
    '-','Invalid or inactive rounding differences account in journal entry',
    '-*', 'Modified invalid or inactive rounding differences account in journal entry',
    '<', 'Showing sequence assignment failure',
    '<*', 'Was showing sequence assignment failure',
    '>', 'Showing cutoff rule violation',
    '>*', 'Was showing cutoff rule violation',
    'A', 'Journal batch failed funds reservation',
    'A*', 'Journal batch previously failed funds reservation',
    'AU', 'Showing batch with unopened period',
    'B', 'Showing batch control total violation',
    'B*', 'Was showing batch control total violation',
    'BF', 'Showing batch with frozen or inactive budget',
    'BU', 'Showing batch with unopened budget year',
    'C', 'Showing unopened reporting period',
    'C*', 'Was showing unopened reporting period',
    'D', 'Selected for posting to an unopened period',
    'D*', 'Was selected for posting to an unopened period',
    'E', 'Showing no journal entries for this batch',
    'E*', 'Was showing no journal entries for this batch',
    'EU', 'Showing batch with unopened encumbrance year',
    'F', 'Showing unopened reporting encumbrance year',
    'F*', 'Was showing unopened reporting encumbrance year',
    'G', 'Showing journal entry with invalid or inactive suspense account',
    'G*', 'Was showing journal entry with invalid or inactive suspense account',
    'H', 'Showing encumbrance journal entry with invalid or inactive reserve account',
    'H*', 'Was showing encumbrance journal entry with invalid or inactive reserve account',
    'I', 'In the process of being posted',
    'J', 'Showing journal control total violation',
    'J*', 'Was showing journal control total violation',
    'K', 'Showing unbalanced intercompany journal entry',
    'K*', 'Was showing unbalanced intercompany journal entry',
    'L', 'Showing unbalanced journal entry by account category',
    'L*', 'Was showing unbalanced journal entry by account category',
    'M', 'Showing multiple problems preventing posting of batch',
    'M*', 'Was showing multiple problems preventing posting of batch',
    'N', 'Journal produced error during intercompany balance processing',
    'N*', 'Journal produced error during intercompany balance processing',
    'O', 'Unable to convert amounts into reporting currency',
    'O*', 'Was unable to convert amounts into reporting currency',
    'P', 'Posted',
    'Q', 'Showing untaxed journal entry',
    'Q*', 'Was showing untaxed journal entry',
    'R', 'Showing unbalanced encumbrance entry without reserve account',
    'R*', 'Was showing unbalanced encumbrance entry without reserve account',
    'S', 'Already selected for posting',
    'T', 'Showing invalid period and conversion information for this batch',
    'T*', 'Was showing invalid period and conversion information for this batch',
    'U', 'Unposted',
    'V', 'Journal batch is unapproved',
    'V*', 'Journal batch was unapproved',
    'W', 'Showing an encumbrance journal entry with no encumbrance type',
    'W*', 'Was showing an encumbrance journal entry with no encumbrance type',
    'X', 'Showing an unbalanced journal entry but suspense not allowed',
    'X*', 'Was showing an unbalanced journal entry but suspense not allowed',
    'Z', 'Showing invalid journal entry lines or no journal entry lines',
    'Z*', 'Was showing invalid journal entry lines or no journal entry lines', NULL ),

  • Performance Issues when editing large PDFs

    We are using Adobe 9 and X Professional and are experiencing performance issues when attempting to edit large PDF files.  (Windows 7 OS). When editing PDFs that are 200+ pages, we are seeing pregnated pauses (that feel like lockups), slow open times and slow to print issues. 
    Are there any tips or tricks with regard to working with these large documents that would improve performance?

    You said "edit." If you are talking about actual editing, that should be done in the original and a new PDF created. Acrobat is not a very good editing tool and should only be used for minor, critical edits.
    If you are talking about simply using the PDF, a lot depends on the structure of the PDF. If it is full of graphics, it will be slow. You can improve this performance by using the PDF Optimize to reduce graphic resolution and such. You may very likely have a bloated PDF that is causing the problem and optimizing the structure should help.
    Be sure to work on a copy.

  • Performance issue when a Direct I/O option is selected

    Hello Experts,
    One of my customers has a performance issue when a Direct I/O option is selected. Reports that there was increase in memory usage when Direct I/O storage option is selected when compared to Buffered I/O option.
    There are two applications on the server of type BSO. When using Buffered I/O, experienced a high level of Read and Write I/O's. Using Direct I/O reduces the Read and Write I/O's, but dramatically increases memory usage.
    Other Information -
    a) Environment Details
    HSS - 9.3.1.0.45, AAS - 9.3.1.0.0.135, Essbase - 9.3.1.2.00 (64-bit)
    OS: Microsoft Windows x64 (64-bit) 2003 R2
    b) What is the memory usage when Buffered I/O and Direct I/O is used? How about running calculations, database restructures, and database queries? Do these processes take much time for execution?
    Application 1: Buffered 700MB, Direct 5GB
    Application 2: Buffered 600MB to 1.5GB, Direct 2GB
    Calculation times may increase from 15 minutes to 4 hours. Same with restructure.
    c) What is the current Database Data cache; Data file cache and Index cache values?
    Application 1: Buffered (Index 80MB, Data 400MB), Direct (Index 120MB; Data File 4GB, Data 480MB).
    Application 2: Buffered (Index 100MB, Data 300MB), Direct (Index 700MB, Data File 1.5GB, Data 300MB)
    d) What is the total size of the ess0000x.pag files and ess0000x.ind files?
    Application 1: Page File 20GB, Index 1.7GB.
    Application 2: Page 3GB, index 700MB.
    Any suggestions on how to improve the performance when Direct I/O is selected? Any performance documents relating to above scenario would be of great help.
    Thanks in advance.
    Regards,
    Sudhir

    Sudhir,
    Do you work at a help desk or are you a consultant? you ask such a varied range of questions I think the former. If you do work at a help desk, don't you have a next level support that could help you? If you are a consultant, I suggest getting together with another consultant that actually knows more. You might also want to close some of your questions,. You have 24 open and perhaps give points to those that helped you.

  • I'm facing performance issue while accessing the PLAF Table

    Dar all,
    I'm facing performance issue while accessing the PLAF Table.
    The START-OF-SELECTION of the report starts with the following select query.
        SELECT plnum  pwwrk matnr gsmng psttr FROM plaf
        INTO CORRESPONDING FIELDS OF TABLE it_tab
        WHERE matnr IN s_matnr
          AND pwwrk = p_pwwrk
          AND psttr IN s_psttr
          AND auffx = 'X'
          AND paart = 'LA' .
    While executing the report in the Quality system it does not face any performance issue...
    When it comes to Production System the above said select query itself it is taking 15 - 20 minutes time to move further.
    Kindly help me to over come this problem...
    Regards,
    Jessi

    Hi,
    "Just implement its primary Key
    WHERE PLNUM BETWEEN '0000000001' AND '9999999999' " By this you are implementing the Primary Key
    This statement has nothing to do with performance, because system is not able to use primary key or uses every row.
    Jessica, your query uses secondary index created by SAP:
    1     (Material, plant) which uses fields MANDT MATNR and PLWRK.
    but it is not suitable in your case.
    You can consider adding new one, which would containt all fields: MANDT, MATNR, PWWRK, PSTTR AUFFX PAART
    or - but it depends on number of rows meeting and not meeting (auffx = 'X' AND paart = 'LA' ) condition.
    It could speed the performance, if you would create secondary index based on fields MANDT, MATNR, PWWRK, PSTTR
    and do like Ramchander suggested: remove AUFFX and PAART from index and where section, and remove these unwanted rows
    after the query using DELETE statement.
    Regards,
    Przemysław
    Please check how many rows in production system

  • Issue when accessing standard LM transactions via ITS

    Hi,
    We are implementing WM using ITS Mobile.
    We have developed certain custom screens and also used standard transactions such as LM02 and LM04.We have created  a internet service for transaction LM01 and published all our custom and standard screens within the service.
    When testing the transactions in R/3 using LM01 transaction..there were no issue...
    But when testing the same in ITS....once we enter into a standard transaction like LM02(By clicking the corresponding menu entry) and click on Back button...it must bring me back to my menu( transaction LM01)...instead...my ITS session gets logged off...displaying message...'You have been logged off from SAP NetWeaver Application Server'.
    We are facing the same issue when we try to access standard LM transactions via ITS...
    I tried to analyze by debugging from ITS....but the flow of R/3 and ITS remains the same...
    These are my service parameters..
    ~THEME     99
    ~SINGLETRANSACTION     1
    ~ITSMOBILE     1
    ~TRANSACTION          LM01
    ~SOURCES     ZWM_MENU,itsmobile
    ~GENERATEDYNPRO     1
    ~POPUPS     1
    Did any of you face a similar kind of issue when accessing standard LM transactions in ITS ?
    Kindly Suggest.
    Regards,
    Thyagu.
    Edited by: Thyagu Natarajan on Jan 25, 2011 1:00 PM

    Hello,
    I am also trying to implement WM using ISTMobile, but i have a little problem... when i am in my local network the service works and calls the page in the browser, but when i am in other network it gives me "Could not connect to remote server" (off course).
    I undertand that i can´t access my server through other networks...  but can anyone tell me if i it is possible to access SAP ITSMobile transactions through the internet?? Is the solution create a VPN network to my local network or something like that?
    Thanks a lot.

  • Oracle Retail 13 - Performance issues when open, save, approving worksheets

    Hi Guys,
    Recently we started facing performance issues when we started working with Oracle Retail 13 worksheets from within the java GUI at clients desktops.
    We run Oracle Retail 13.1 powered by Oracle Database 11g R1 and AS 10g in latest release.
    Issues:
    - Opening, saving, approving worksheets with approx 9 thousands of items takes up to 15 minutes.
    - Time for smaller worksheets is also around 10 minutes just to open a worksheet
    - Also just to open multiple worksheets takes "ages" up to 10-15 minuts
    Questions:
    - Is it expected performance for such worksheets?
    - What is your experience with Oracle Retail 13 in terms of performance while working with worksheets - how much time does it normally take to open edit save a worksheet?
    - What are the average expected times for such operations?
    Any feedback and hints would be much appreciated.
    Cheers!!

    Hi,
    I guess you mean Order/Buyer worksheets?
    This is not normal, should be quicker, matter of seconds to at most a minute.
    Database side tuning is where I would look for clues.
    And the obvious question: remember any changes to anything that may have caused the issue? Are the table and index statistics freshly gathered?
    Best regards, Erik Ykema

  • How to get around a performance issue when dealing with a lot of data

    Hello All,
    This is an academic question really, I'm not sure what I'm going to do with my issue, but I have some options.  I was wondering if anyone would like to throw in their two cents on what they would do.
    I have a report, the users want to see all agreements and all conditions related to the updating of rebates and the affected invoices. From a technical perspective ENT6038-KONV-KONP-KONA-KNA1.  THese are the tables I have to hit.  The problem is that when they retroactively update rebate conditions they can hit thousands of invoices, which blossoms out to thousands of conditions...you see the problem. I simply have too much data to grab, it times out.
    I've tried everything around the code.  If you have a better way to get price conditions and agreement numbers off of thousands of invoices, please let me know what that is.
    I have a couple of options.
    1) Use shared memory to preload the data for the report.  This would work, but I'm not going to know what data is needed to be loaded until report run time. They put in a date. I simply can't preload everything. I don't like this option much. 
    2) Write a function module to do this work. When the user clicks on the button to get this particular data, it will launch the FM in background and e-mail them the results. As you know, the background job won't time out. So far this is my favored option.
    Any other ideas?
    Oh...nope, BI is not an option, we don't have it. I know, I'm not happy about it. We do have a data warehouse, but the prospect of working with that group makes me whince.

    My two cents - firstly totally agree with Derick that its probably a good idea to go back to the business and justify their requirement in regards to reporting and "whether any user can meaningfully process all those results in an aggregate". But having dealt with customers across industries over a long period of time, it would probably be bit fanciful to expect them to change their requirements too much as in my experience neither do they understand (too much) technology nor they want to hear about technical limitations for a system etc. They want what they want if possible yesterday!
    So, about dealing with performance issues within ABAP, I'm sure you must be already using efficient programming techniques like using Hash internal tables with Unique Keys, accessing rows of the table using Field-Symbols and all that but what I was going to suggest to you is probably look at using [Extracts|http://help.sap.com/saphelp_nw04/helpdata/en/9f/db9ed135c111d1829f0000e829fbfe/content.htm]. I've had to deal with this couple of times in the past when dealing with massive amount of data and I found it to be very efficient in regards to performance. A good point to remember when using Extracts that, I quote from SAP Help, "The size of an extract dataset is, in principle, unlimited. Extracts larger than 500KB are stored in operating system files. The practical size of an extract is up to 2GB, as long as there is enough space in the filesystem."
    Hope this helps,
    Cheers,
    Sougata.

  • Complete Outlook 2010 Lock-Up when accessing archive pst files

    Hi,
    I am having a real issue with Outlook 2010 (32 bit) and accessing archive pst files that are stored on a network share on our SBS server.
    Like many we operate almost soley using email and this creates ever larger mailboxes, so most items are archived after six-months, but still need to be accessed occassionaly.
    Each user has a series of archive pst files that are stored on our server in a common location for ease, but since switching some users to Office 2010 we now have a major issue that if a user opens one of the archive pst files from its network location they
    can view everything in all folders of the archive, but after accessing a couple of messages Outlook "stops responding" and totally locks.
    It seems that Outlook is trying to access the message in the archive pst, but cannot complete this and even restarting Outlook doesn't solve this as on restart the splash screen sticks at loading profile and never restarts! Worse yet this lock-up in Outlook
    actually prevents Windows 7 from closing down, so when you try to restart the PC to resolve the problem Windows 7 just sits at "shutting down" and never closes (left for 20 mins), so a reset or hard reboot is then needed.
    Has anyone seen similar issues with pst files created from Outlook 2007?
    Is this issue due to network location of pst's?
    Anyone got any ideas, as this issue drives me nuts when it happens, which thankfully is rare as we dont access archives often and can do so fine if needed from a PC running Outlook 2007.
    Thanks
    Ridesy

    Here is my 2 cents, a few years down the road from the original posting.
    I had the same problem as the original poster... extremely slow .pst performance for a current archive folder.
    I have a SBS 2008 system in place at this client. Ironically they acquired another small business and everyone moved in--along with all their data which lead to the following discovery.
    Company A's SBS 2008 server was a bit tight on disk space so I talked them into setting up a second server to be primarily a file server.
    I moved all of Company B's data --including lots of good old .PST archive files to sever #2.
    I setup all Company B staff with a new mapped drive and had their Outlook 2010 clients open their old archive .pst mail from this mapped drive.
    Current archive files, like the rest of Company A's staff, point to the SBS server via a UNC share \\server\Usershares\%username%\....I just like to keep consistency and this was how it was setup all along.
    With the latest archive .pst folder on Server #1 (the sbs server) for some of my recent Company B setups--performance is horrible. Exactly what everyone is complaining about. I have 10gig archive.pst files for these same users on the second server using
    a mapped network drive and they open fine. Even some archive files that are 15+ gigs in size (archives are named for each year--2006.pst etc).
    So, either there is a problem hosting your .pst via UNC shares, or on a SBS server in general.
    Short story... try setting it up via a Mapped network drive on a different server that isn't doing everything like the SBS server is...I tend to think it is UNC mapping related but not sure.
    Not guaranteeing this is the 'fix' but it definitely was a night and day difference when I just tried this and I am not going to do anything different from here out.

  • Performance degradation when using proxy.pac file with FF ESR 31

    With Bug 923458 many people complained about a performance issue compared to other browsers when a proxy.pac file is used.
    The issue initially reported with the bug was resolved for ESR25 according to the statistics, but the general performance issue remained.
    I had the same issue with ESR24 and ESR31.3 .
    I was testing with www.bild.de.
    It took about 40 seconds to load the content completely. Without the proxy.pac file it took about 10 seconds.
    I added a few alerts to the pac-File in order to get logs within the console for some analyses.
    I found the following:
    1. the pac.file is executed for every request, no matter if the host changed or not.
    With us the pac-File checks for IP-Adresses and host-names only.
    It is not necessary to execute the pac file for each and every request to the same remote host.
    So the question is, if we are able to disable this behaviour via about:config?
    2. the content referenced by www.bild.de seems to be loaded sequentially and with a delay
    The overall time consumed by the proxy.pac file executions was about 4 Seconds compared to the 40 seconds of overall load time.
    So I checked the delay between executions of the pac-file and found an overall delay of 40 seconds. I expect that the delay between the calls to the pac-file is caused by the retrieval of contents from the remote host.
    So why are the requests executed sequentially?
    Hint: Due to the times necessary for executing the pac-file and downloading the contents from the remote host, I would expect the logs generated by my alerts to be mixed (especially if myIpAddress took 1 Second). But the log is cleanly ordered by URL. (see attachment)

    Hi guigs2,
    thanks for your response. As we only use myIpAddress once within our pac-File and only rely on dnsDomainIs(), ==-Comparisons and shExpMatch() and the sum of all pac-Executions was about 4 seconds compared to 40 seconds overall load time, I do not think that dns resolving is our issue.
    I checked the seetings of the configuration you mentioned above. It is set to "false", so the client would try the resolve the dns names. Our admin told, that we do not use socks-Proxies, only http-Proxies.
    Regarding sequential load of the contents included on www.bild.de from other web sites, I attached a screenthot.
    Please note the red highlights. These show the start time in milliseconds of the pac-execution. I added this as a kind of id which represents a unique identifier together with the URL if the log items are mixed. But they are not, instead they are cleanly ordered by URL (for all 360 pac-file calls).
    Moreover in the picture you can see the delay between the end of the last pac-file execution and the next one (blue timestamp in millisonds compared to the red timestamp of the next row saying "entered proxy.pac"). The delay sum up exactly to the 40 seconds the FF took to load the page completely.
    Alone the fragment shown represents a delay of 630ms between the pac-file executions. If the contents would be loaded in parallel, there should be no such delay.

  • How to Improve performance issue when we are using BRM LDB

    HI All,
    I am facing a performanc eissue when i am retriving the data from BKPF and respective BSEG table....I see that for fiscal period there are around 60lakhs records. and to populate the data value from the table to final internal table its taking so much of time.
    when i tried to make use of the BRM LDB with the SAP Query/Quickviewer, its the same issue.
    Please suggest me how to improve the performance issue.
    Thanks in advance
    Chakradhar

    Moderator message - Please see Please Read before Posting in the Performance and Tuning Forum before posting - post locked
    Rob

  • Issue when I upload CSV file for insert contact list on Cisco Attendant Console Advanced

    Hi Friends,
    I have error "Provided file is not valid CSV File" when I upload CSV file to Cisco Attendant Console Advanced system. I'm not sure what happen about my CSV. My purpose is insert contact list to system and create CSV base on documentation.
    Anyone have idea to solve this issue please kindly help me.
    Thanks

    You are right, whenever Cisco device boots, the IOS files gets loaded on the DRAM.
    But in this process, some temporary files are also generated which gets saved in the flash/Disk, that’s the only reason you got these error messages. It’s not recommended at all to have less space in the Flash than what is recommended on Cisco.com. I would say please remove some files from Disk and have minimum 256 MB flash otherwise your router may drop into rommon mode at the time of next reload.
    Well, it’s good to upgrade the bootstarp image too. Currently you are running 15.X IOS code, I would say run 15.X bootstarp image on the box.
    You may download bootstarp image for 7206VRX NPEG2 from the link below:-
    http://software.cisco.com/download/release.html?mdfid=282188585&flowid=1380&softwareid=280805685&release=15.2.4S4&relind=AVAILABLE&rellifecycle=ED&reltype=latest
    If you want to know the procedure of the upgrade, click the link mentioned below:-
    http://www.cisco.com/en/US/docs/ios/12_2/configfun/command/reference/frf010.html#wp1017654
    -Amant

  • Display issues when opening InDesign CC files

    I have had CC since it was released and I am having an issue. when i open a INDD file, indesign will open correctly, but the document will not display. I can see the scroll bars, and can move the mouse auround and the cursor will change like it's hovering over a frame. but not visual document. It also collapses all my panels.
    It lookes like this
    I am running Indesign CC, Win 7 pro 64bit
    Core i7-3770 3.4ghz
    16gb memory
    NVIDIA gforse GT 640
    500GB WD Velociraptor
    Any thoughts?

    Second report like this today...
    Try trashing your prefs. See Replace Your Preferences

  • IOS - Performance issues when touching screen

    Hello,
    I am having perfomance issues when the use keep moving one finger on the screen.
    I am testing with :
    - flash builder 4.7 beta 2
    - air 3.5 (beta too I guess)
    - iPhone 3GS running on iOS 6.0.1 (also tested on an iPad3, and even though it's less horrible, the impact is still huge compared to what you could expect).
    - gpu mode
    - always published as an addhoc release build (best)
    Test 1-A) :
    - Have a completely empty project, add a fps counter (with mouseEnabled and children false).
    - run a while each frame so you force the fps under 60 (if not you won't see it going down from let's say theoric 100 to theoric 70 since fps max is 60).
    like this for example :
                                            protected function framingMouseTest(e : Event) : void {
                           var t : int = getTimer();
                           while (getTimer()-t<30) {}//this will keep flash busy
    Notice that when you keep your finger on the screen and move it you loose approximately 7ms per frame. (for example from 26 to 22fps or from 48 to 36 fps, depending on the time waited in your while).
    Test 1-B) do the same with even stage mouseEnabled false (and you could even set it + mouseChildren off recursively from stage) : same result (-7ms). Why ? Since it is all off it shouldn't take more than 1 ms no ?
    Test 1-C) add Multitouch.inputMode = MultitouchInputMode.TOUCH_POINT; Same result (found some threads saying it would help but not that much...)
    Test 1-D) Multitouch.mapTouchToMouse = false;          It is a bit better but still around 5ms (didnt write down this one)
    Test 1-D) Multitouch.mapTouchToMouse = false; Multitouch.inputMode = MultitouchInputMode.NONE;          It is a bit better but still around 4ms (didnt write down this one) And you have now nothing left to remove I think.
    Now the horrible part :
    Test 2-A)
    Add some clips and subclips to your scene, containing bitmaps, results may vary depending on how much (let's say 30, I don't have the exact count on subcontainers), etc. but what's important is the difference between no finger and 1 finger moving on screen.
    On normal conditions (keep some mouseEnabled for the clips you need interaction add those a listener, lets say 10 total, default values for the rest) you can get about 40ms per frame down ! (for example from 36 to 14 fps) just by touching the screen.
    Test 2-B) Recursively set everything to mouseEnabled and children false after everything is added, you still get about 30ms down because of the finger, whereas it shouldn't test above the main container which already has mouseChildren false.
    Test 2-C) Multitouch.mapTouchToMouse = false; Multitouch.inputMode = MultitouchInputMode.NONE; you still get about 20ms down !!! whereas you're in a mode when touch and mouseEvents shouldn't even exist.
    I am really confused. My conclusions are :
    Performance are really impacted whatever you do when a user touch the screen + it will get worse the more clips/depth you have. Even if everything is made to disable this.
    I think anyone could test this with pretty much any app done with AIR on iOS, as long as the framerate is already <max (<60 for example), touch and move your finger and tell me what happens regarding framerate.
    What I would love :
    1) tell me what I am missing if you think this is due to some mistake on my part, if not :
    2) better perfomance when containers have mouseChildren false. It seems obvious it still takes longer to process.
    3) a mode where the touch/mouse is REALLY disabled including stage etc. It is quite obvious air does takes longer to handle touch as the scene gets more complex, even if you disable it all (seems to work only 20%...)
    4) a mode where the touch/mouse is REALLY disabled, on anything but the stage. Because you might want to handle everythin just from stage coordinates, if your project allows you to, like mine. Unfortunately right now it wouldn't give you enough performance boost (well, a bit but still maybe +20ms).
    PS : you don't even need to add mouselistener or touchlisteners (tested both), but if you do, it is worse.
    Compared to this, I am finding rendering performance pretty decent - even though it's tough for a game on a 3GS- as long as you do it properly.

    Could you open a bug report on this at bugbase.adobe.com?  Please include any sample code or applications so we can quickly reproduce the problem.  If you'd like to keep this private, feel free to send them to me directly ([email protected]).  Also, please note in the bug if this is new behavior with the latest AIR sdk or if you've seen this in past versions too.  Once added, let me know the bug number and I'll follow up internally.  I've alerted the iOS team to expect this bug.
    Thanks,
    Chris

Maybe you are looking for