Strange Performance Problem

Hello. I'm very new to Flash so if I haven't provided enough
information or the proper terminology I apologize. If you can help
let me know and I'll get any additional informaiton you need.
Here's what's up
I have a slideshow with about 40 slides. Some slides have a
small video loop, a few MC symbols used throughout that have a
glowing effect or flashing and a lot of FLV's that play once the
play button is pressed on the associated slide. Here's what's
strange.
Since I'm new this has taken me quite a while to build. When
I started this thing and would get all excited to publish it and
take a look it would take hours to publish. Once published I was
loving it and didn't care how long it was taking to publish. This
carried on for a long while during my building of this project.
When I would play the published file, everything worked just right.
Now though, it's a much different story. I'm not sure what I did or
how I did it but now when I go to publish my file it publishes in
an astonishing 3 minutes. This is great and all but now when the
file publishes so quickly and I execute it, my slides run slowly
when changing pages, my video clips are all choppy and everything
within the show is severely delayed. (delayed as in placing my
cursor over a button and it won't light up right away, I press play
on the video and it doesn't play right away like it once did).
What's going on? Did I change something somewhere that now makes my
publish run so quickly but the output file runs so poorly that it's
not even worth using any more? Please. I beg for any help. I put so
much time, study, learning and effort into making this, aquiring
content and having fun with it and now this stupid thing is
worthless. Thank you so much to any of you that can help me. I
sincerely appreciate it more than you will ever know. If you know
how to fix this but will take your time with me I would be willing
to pay if you want to work something out with me. Thanks so much
everyone!

Hi Thomas,
the creation of the index will invalidate the statement of the cache. Therefore for any statement executed after
the new index was created the optimizer will see it and take it into account and will not use an old cached plan.
As far as i know this is true on all databases that are supported by SAP. (i never came across problems here).
Data distribution changes... now we talk about statistics. Same databases do not have a problem with this.
E.g. MaxDB since they do a so called evaluation all the time. The new version of DB6 (9.5 or 9.7) uses so
called real time statistics and at least theoretically should see the new data distribution as well (of course
there is some threshold where this feature kicks in, below the threshold the old behaviour is valid. The old
behaviour is what we have on all other databases as far as i know. The old cached plan is used the new
data distribution is only taken into account when new statistics are created. The creation of new statistics
leads to invalidation of the old plans.
For your problem it is most likely (this is an educated guess), as you said, the parameter issue. MSSQL
evaluates the parameters for the FIRST execution of a statement. These parameters led to a full table
scan in your case. All other statements with different parameters will use that plan until the sql statement
is invalidated in the cache. On MSSQL this happens if:
- a new index is created
- new statistics are created
- the statement is recompiled / reparsed  (then the parameters for the execution with the recompile / reparse hint are used).
- the statement cache is cleared / invalidated
A new plan will be build as well if the statement is displaced from the statement cache.
Kind regards,
Hermann

Similar Messages

  • Out.print performance problems?

    Hi
    I'm facing strange performance problems I would not expect generating long html pages on the fly.
    I have a page that is approximately 30 to 40Kb long. Profiling code generation I obtain a 30 second time to process. I have excluded every server-side processing activity since I have also used a StringBuffer to accumulate all the html code to be flushed: generating the StringBuffer content require less than 1ms... flushing out the content with:
    out.print(stb);
    out.flush()
    is taking 5 to 6 seconds. The same happens both on a development environment using CQ servlet engine, both in production where we are using CQ with JBoss as app server.
    I tried the same operation on a simple tomcat installation with a JSP... less than 10 ms to generate the full page.
    This problem is slowing down a lot performances.
    The contents are generated inside a nested component included with a <cq:include> tag.
    Anyone can give me suggestions on how to cut down a so big delay in html content generation?
    thanks

    I made further analysis and now I can give a clearer scenario.
    The *very low* performances has been registered on CQ 5.2.1 when generating html code full of relative links. Every link *cost* big processing time. With 60-70 links the component ends up paying 3 to 4 seconds (too much!).
    The same component on a CQ 5.4 instance takes approximately 500 milliseconds, still too much considering that without link the same html code is generated in less than 10 milliseconds.
    The reason is clearly in the filters applied in the output pipeline --> link checker? remapping on links based on etc/map configurations?
    How can I change configurations in order to avoid these performance problems?
    thanks
    Ignazio

  • Very strange performance issue

    Hi!
    I've got a very very strange performance problem and I have no more ideas what it might be:
    The following code fragment contains the useless line int count=0;
    However, deleting the line reduces(!) the execution speed of the program by ~30%. This is not only weird because the line is absolutely useless, but also because the whole function only takes <1% total execution time in the program.
        public void simuliere_bis_soll_erreicht()
            int count=0;  //????
            while(welt.soll_simulationszeit>welt.ist_simulationszeit && simu_active)
                simuliere_einen_schritt();
        }The problem occurs both under Java 1.5 and 1.4.2, using the HotSpot-Client.
    Cleaning and rebuilding the project does not help.
    It occurs on different computers in the same way, too.
    Thank you very much in advance! :-)
    Mirko Seithe

    well, this is what you get:
    1.) run totally interpreted since compilation threshhold is reached (1500 (client jvm) - 10000 (server-jvm)
    2.) start background compilation, run further in interpreted mode till compilation finishes (depends on code, but about 1000-10000)
    3.) run compiled code
    so this is what you get:
    ~20.000 invokations interpreted (about 70-90%n of total time)
    ~80.000 invokations compiled (30-10% of total time)
    maybe your int takes much longer to optimize, so the loop executes much longer in interpreted code.
    I would not bet on such microbenchmarks, since they do not tell you real stories under real circumstances. Believe me, I am the programmer working on performance optimizations in the company I am employed.
    lg Clemens

  • Strange performance issue with 3510/3511 SAM-FS disk cache

    Hi there!
    I'm running a small SAM-QFS environment and have some strange performance issue on the disk storage part, which somebody here might be able to explain.
    Configuration: one 3510, dual controller, RAID-5 9+1, one hot spare and one disk not configured for whatever reason. The R5 logical drive hosts a 150GB LUN for SAM-QFS metadata (mm in SAM-FS speak) and a 1TB LUN for data (mr in SAM-FS speak). Further, there are two small LUNs (2GB, 100GB) for some other purpose. Those two LUNs have nearly no I/O. All disks are SUN146G. Host connection is 2GBit, multipathing enabled and working.
    Then the disk cache became too small, and the customer added a 3511 expansion unit with SUN300G disks. One logical drive is a RAID-1, 1+1, used for NetBackup catalog. The other is a RAID-5, 8+1, providing two LUNs: 260GB SAM-FS metadata (mm) and 1.999TB SAM-FS data (mr).
    For SAM-FS, the LUNs form two file systems: one "residing" in the 3510, the other "residing" in the 3511 expansion. Cabling is according to the manual and checked several times by several independant people. Operating system is Solaris 10, hardware is a V880.
    The problem we observe: SAM-FS I/O on LUNs on disks inside the 3510 is fine. With iostat, I see 100MB/s read and 50MB/s write at the same time. On the SAM-FS file system which is running on the two LUNs in the 3511, the limit seems to be at 40MB/s read/write. Both SAM-FS file systems are configured the same in regards of block size.
    In case I have activity on both SAM-FS file systems, I see 100MB/s+ on the LUN running inside the controller shelf and another 40MB/s on the disk runnin in the 3511 expansion chassis. So, the controller is easily capable of handling 150MB/s.
    Cache settings in the 3510 controller are default I think (wasn't installed by me), batteries are fine.
    Is this 40MB/s we experience a limitation by the expansion shelf? Don't think so. Anybody has any ideas on this? What parameters to check or to change? Any hint appreciated. I can also provide further details if needed. Thank you.
    wolfgang

    SUN300G disks sound like 300GB FC disks.
    Depending on how many files are in the SAMFS file system, sharing the mm and mr devices on the same RAID array can be a pretty horrible idea. In my opinion and experience, it's almost always better to NEVER put more than one LUN on a RAID array. Period. Putting more than one LUN on an array results in IO contention on that array. And large, unnaturally configured (9+1? Why?) RAID arrays will have problems from the start.
    What are the block sizes used on the RAID arrays? It wouldn't surprise me to see that the RAID array on the expansion tray has a very large block size. Larger block sizes are, in general, not better. Especially for SAMFS metadata - which IIRC is something like 8k or 16k blocks.
    I suspect what is happening is most of the metadata updates are going to the mm device on the new array, contending with the IO operations on the file data.
    How much space is left on each mm device? What does "iostat -sndxz 2" show when you're having the IO problems?

  • CRM IC Webclient - Massive Performance Problems Searching in Agent Inbox

    Hi Forum,
    can somebody help me. We have massive Search Performance problems in Interaction Center Webclient in Agent Inbox. When agents searching, for example for open emails, the time to get a result takes approx. 2-3 minutes.
    Thats absolute inadmissible and endangered the running business.
    The queury reading the eMails from the workitem tables is very very slow.
    Could somebody help me with ideas to solve this big performance problem. Have somebody the same problems?
    Thank you very much in advance for any information.
    We use Interaction Center Webclient on CRM Release 5.0
    Thorsten

    Another aspect you could check is the following:
    Analysis performed didn't show anything strange nor any high consumers
    of the response time. The processing time occurs, since IC-Web on CRM 5
    is a bit demanding for CPU power.
    The only thing that could improve the performance a bit is the
    <b>buffering of org structure</b>, which is currently switched off for
    SALES and SERVICE scenarios as per table T77OMATTSC.
    The report <b>HRBCI_ATTRIBUTES_BUFFER_UPDATE</b> is running regulary
    however no scenario is being placed in the buffer to speed up
    org structure read. Please check this and provide feedback.
    How to use this report, can be found back in the Reports own documentation and also in SAP HELP.
    cheers
    Davy

  • Performance Problem - MS SQL 2K and PreparedStatement

    Hi all
    I am using MS SQL 2k and used PreparedStatement to retrieve data. There is strange and serious performance problem when the PreparedStatement contains "?" and using PreparedStatement.setX() functions to set its value. I have performed the test with the following code.
    for (int i = 0; i < 10; i ++) {
    try {
    con = DBConnection.getInstance();
    statement = con.prepareStatement("SELECT * FROM cardno WHERE car_no = '" + cardNo + "'");
    // statement = con.prepareStatement("SELECT * FROM cardno WHERE car_no = ?");
    // statement.setString(1, cardNo);
    rs = statement.executeQuery();
    if (rs.next()) {
    catch(SQLException e) {
    e.printStackTrace();
    finally {
    try {
    rs.close();
    statement.close();
    catch(SQLException e) {
    e.printStackTrace();
    Iteration Time (ms)
    1 961
    10 1061
    200 1803
    for (int i = 0; i < 10; i ++) {
    try {
    con = DBConnection.getInstance();
    // statement = con.prepareStatement("SELECT * FROM cardno WHERE car_no = '" + cardNo + "'");
    statement = con.prepareStatement("SELECT * FROM cardno WHERE car_no = ?");
    statement.setString(1, cardNo);
    rs = statement.executeQuery();
    if (rs.next()) {
    catch(SQLException e) {
    e.printStackTrace();
    finally {
    try {
    rs.close();
    statement.close();
    catch(SQLException e) {
    e.printStackTrace();
    Iteration Time (ms)
    1 1171
    10 2754
    100 18817
    200 36443
    The above test is performed with DataDirect JDBC 3.0 driver. The one uses ? and setString functions take much longer to execute, which supposed to be faster because of precompilation of the statement.
    I have tried different drivers - the one provided by MS, data direct and Sprinta JDBC drivers but all suffer the same problem in different extent. So, I am wondering if MS SQL doesn't support for precompiled statement and no matter what JDBC driver I used I am still having the performance problem. If so, many O/R mappings cannot be used because I believe most of them if not all use the precompiled statement.
    Best regards
    Edmond

    Edmond,
    Most JDBC drivers for MS SQL (and I think this includes all the drivers you tested) use sp_executesql to execute PreparedStatements. This is a pretty good solution as the driver doesn't have to keep any information about the PreparedStatement locally, the server takes care of all the precompiling and caching. And if the statement isn't already precompiled, this is also taken care of transparently by SQL Server.
    The problem with this approach is that all names in the query must be fully qualified. This means that the driver has to parse the query you are submitting and make all names fully qualified (by prepending a db name and schema). This is why creating a PreparedStatement takes so much using these drivers (and why it does so every time you create it, even though it's the same PreparedStatement).
    However, the speed advantage of PreparedStatements only becomes visible if you reuse the statement a lot of times.
    As about why the PreparedStatement with no placeholder is much faster, I think is because of internal optimisations (maybe the statement is run as a plain statement (?) ).
    As a conclusion, if you can reuse the same PreparedStatement, then the performance hit is not so high. Just ignore it. However, if the PreparedStatement is created each time and only used a few times, then you might have a performance issue. In this case I would recommend you try out the jTDS driver ( http://jtds.sourceforge.net ), which uses a completely different approach: temporary stored procedures are created for PreparedStatements. This means that no parsing is done by the driver and PreparedStatement caching is possible (i.e. the next time you are preparing the same statement it will take much less as the previously submitted procedure will be reused).
    Alin.

  • Performance problem while CPU is 80% Idel ?

    Hi,
    My end users are claim for performance problem during execution of batch process.
    As you can see there are 1,745 statement executing each second.
    Awr report shows 98.1% of the time , waits on CPU .
    Also Awr report shows that Host CPU is :79.9% Idel.
    The second wait event shows only 212 seconds waits on db file sequential read.
    Yet , 4 minute in 1 hour period is seems not an issue.
    Please advise
    DB Name         DB Id    Instance     Inst Num Startup Time    Release     RAC
    QERP          xxx        erp                 1 21-Jan-13 15:40 11.2.0.2.0 ; NO
    Host Name        Platform                         CPUs Cores Sockets Memory(GB)
    erptst           HP-UX IA (64-bit)                  16    16       4     127.83
                  Snap Id      Snap Time      Sessions Curs/Sess
    Begin Snap:     40066 22-Jan-13 20:00:52       207       9.6
      End Snap:     40067 22-Jan-13 21:00:05       210       9.6
       Elapsed:               59.21 (mins)
       DB Time:              189.24 (mins)
    Cache Sizes                       Begin        End
    ~~~~~~~~~~~                  ---------- ----------
                   Buffer Cache:     8,800M     8,800M  Std Block Size:         8K
               Shared Pool Size:     1,056M     1,056M      Log Buffer:    49,344K
    Load Profile              Per Second    Per Transaction   Per Exec   Per Call
    ~~~~~~~~~~~~         ---------------    --------------- ---------- ----------
          DB Time(s):                3.2 ;               0.1 ;      0.00 ;      0.05
           DB CPU(s):                3.1 ;               0.1 ;      0.00 ;      0.05
           Redo size:          604,285.1 ;          27,271.3
       Logical reads:          364,792.3 ;          16,463.0
       Block changes:            3,629.5 ;             163.8
      Physical reads:               21.5 ;               1.0
    Physical writes:               95.3 ;               4.3
          User calls:               68.7 ;               3.1
              Parses:              212.9 ;               9.6
         Hard parses:                0.3 ;               0.0
    W/A MB processed:                1.2 ;               0.1
              Logons:                0.3 ;               0.0
            Executes:            1,745.2 ;              78.8
           Rollbacks:                1.2 ;               0.1
        Transactions:               22.2
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
                Buffer Nowait %:  100.00 ;      Redo NoWait %:  100.00
                Buffer  Hit   %:   99.99 ;   In-memory Sort %:  100.00
                Library Hit   %:   99.95 ;       Soft Parse %:   99.85
             Execute to Parse %:   87.80 ;        Latch Hit %:   99.99
    Parse CPU to Parse Elapsd %:   74.76 ;    % Non-Parse CPU:   99.89
    Shared Pool Statistics        Begin    End
                 Memory Usage %:   75.37 ;  76.85
        % SQL with executions>1:   95.31 ;  85.98
      % Memory for SQL w/exec>1:   90.33 ;  82.84
    Top 5 Timed Foreground Events
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
                                                               Avg
                                                              wait   % DB
    Event                                 Waits     Time(s)   (ms)   time Wait Class
    DB CPU                                           11,144          98.1
    db file sequential read              52,714         214      4    1.9 User I/O
    SQL*Net break/reset to client        29,050           6      0     .1 Applicatio
    log file sync                         2,536           6      2     .0 Commit
    buffer busy waits                     4,338           2      1     .0 Concurrenc
    Host CPU (CPUs:   16 Cores:   16 Sockets:    4)
    ~~~~~~~~         Load Average
                   Begin       End     %User   %System      %WIO     %Idle
                    0.34 ;     0.33 ;     19.7 ;      0.4 ;      1.8 ;     79.9

    Nikolay Savvinov wrote:
    if the users are complaining about performance of the batch process, then that's what you should be looking at, not the entire system.I find it strange to see "end users" and "the batch process" in the same sentence (as it was in the first post). "End users" gives me the feeling of a significant number of concurrent sessions with people waiting for results in real time at the far end, while "batch process" carries the image a small number of large scale processes running overnight to prepare the data for the following morning.
    I mention this because my first view of the AWR output was: you've got 16 CPUs, only three in use, virtually no users, and doing very little work, how can the users complain. (One answer, of course, is that the 13 CPUs could be locked out of use as far as Oracle is concerned). On the second read I decided that the "users" had gone home, and the complaint was simply that the batch process wasn't completing in time.
    In this case I think "the entire system" IS "the batch process"
    Determine which stored procedures and/or SQL statements took longer than usual and then find out why. Most likely you'll be able to find
    everything you need in AWR views (DBA_HIST_SQL%) and ASH archive (DBA_HIST_ACTIVE_SESS_HISTORY).
    If the batch process has changed dramatically and recently, then a simple first step might be to look at the current AWR report, find the few most time-consuming SQL statements, and use the awrsqrpt.sql script to find their history of execution plans.
    But I'd also just look at the expensive SQL - bearing in mind, particularly, that there are very few user calls per second, yet many hundred executions per second: it strikes me that there could be quite a lot of PL/SQL going on doing something a little bit expensive many times or some PL/SQL function that calls some SQL that used to be called rarely from an SQL statement but is now (due, perhaps to a change in plan) being called much more frequently - so check SQL Ordered by Executions.
    Regards
    Jonathan Lewis

  • Performance problem - XML and Hashtable

    Hello!
    I have a strange (or not strange, we don't know) performance problem. The situation is simple: our application runs on Oracle IAS, and gets XML messages from an MQ, and we have to process them.
    Sample message:
    <RequestDeal RequestId="RequestDeal-20090209200010-998493885" CurrentFragment="0" EndFragment="51" Expires="2009-03-7T08:00:10">
         <Deal>
              <Id>385011</Id>
              <ModifyDate>2009-02-09</ModifyDate>
              <PastDuesOnly>false</PastDuesOnly>
         </Deal>
         <Deal>
              <Id>385015</Id>
              <ModifyDate>2009-02-09</ModifyDate>
              <PastDuesOnly>false</PastDuesOnly>
         </Deal>
    </RequestDeal>(there's an average of 50000-80000 deals in the whole message)
    After the application receives the whole message, we'd like to parse it, and put the values into a hashtable for further processing:
        Hashtable ht = new Hashtable();
        Node eDeals = getElementNode(docReq, "/RequestDeal");
        Node eDeal = (Element)(eDeals.getFirstChild());
        long start = System.currentTimeMillis();
        while (eDeal != null) {
          String id = getElementText((Element)eDeal, "Id");
          Date modifyDate = getElementDateValue((Element)eDeal, "ModifyDate");
          boolean szukitett = getElementBool((Element)eDeal, "PastDuesOnly", false);
          ht.put(id, new UgyletStatusz(id, modifyDate, szukitett));
          eDeal = eDeal.getNextSibling();
        logger.info("Request adatok betöltve.");
        long end = System.currentTimeMillis();
        logger.info("Hosszu muvelet ideje: " + (end - start) + "ms");The problem is the while statement. On my PC it runs for 15-25 seconds, depends on the number of the deals; but on our customer's server, it runs for 2-5 hours with very high CPU load. The application uses oracle xml parser.
    On my PC there are Websphere MQ 5.3 and Oracle 10ias with 1.5 JVM.
    On our customers's server there are Websphere MQ 5.3 and Oracle 9ias with 1.4 HotSpot JVM.
    Do you have any idea, what can be the cause of the problem?

    gyulaur wrote:
    Hello!
    The problem is the while statement. On my PC it runs for 15-25 seconds, depends on the number of the deals; but on our customer's server, it runs for 2-5 hours with very high CPU load. The application uses oracle xml parser.
    On my PC there are Websphere MQ 5.3 and Oracle 10ias with 1.5 JVM.
    On our customers's server there are Websphere MQ 5.3 and Oracle 9ias with 1.4 HotSpot JVM.
    Do you have any idea, what can be the cause of the problem?All sorts of things are possible.
    - MQ is configured differently. For instance transactional versus non-transactional.
    - The customer uses a network (multiple computers) and you use a box (singlur) and something is using a lot of bandwidth on the network. Could be your process, one of the dependent processes or something else that has nothing to do with your system
    - The processing computer is loaded with something else or something is artificially limiting the processing time of your application (there is a app that allows one to limit another app to one CPU.)
    - At least one version of 1.4 had a xml bug that consumed massive amounts of memory when processing large xml. Sound familar? If the physical memory is not up to it then it will be thrashing the hard drive as it swaps virtual in and out.
    - Of course virtual memory causing swaps would impact it regardless of the cause.
    - The database is loaded.
    You can of course at least get the same version of java that they are using. Actually that would seem like a rather good idea to me regardless.

  • Performance Problem with several Database accesses

    Hello
    We have SAP ERP 2004 running on a Oracle DB 9.x and we face a strange behavior of our system. Sometimes it is very slow. MB51 with 1 Material and 1 Month as selection runs into timeout. But i faced the most significant (or strange) problem  using SE16 with table bseg. If i use Company code,  one item number and year as a selection, it is very fast. If i put in 2 items numbers instead of one, it takes 15 minutes to show the result. This happens everytime i try it.
    For me it looks like the system is doing a sequential read instead of using the proper index. This "wrong index" behavior could be the source of many of our performance problems.
    Does anybody have a suggestion for me, how to fix this problem?
    Thanks in advance
    Hans-Peter

    > We have SAP ERP 2004 running on a Oracle DB 9.x and we face a strange behavior of our system. Sometimes it is very slow. MB51 with 1 Material and 1 Month as selection runs into timeout. But i faced the most significant (or strange) problem  using SE16 with table bseg. If i use Company code,  one item number and year as a selection, it is very fast. If i put in 2 items numbers instead of one, it takes 15 minutes to show the result. This happens everytime i try it.
    >
    > For me it looks like the system is doing a sequential read instead of using the proper index. This "wrong index" behavior could be the source of many of our performance problems.
    As mentioned above you can active trace with ST05 to monitor the problematic SQL request. With the explain function, you will see the estimated cost of the request and the used index.
    An other way is during the execution of the request --> ST04 -> performance -> Wait event analysis -> Session monitor
    If you use the IN option, your request could be very expensive. And in many cases, the reads can only performed with sequential access.
    For an overlook of possible indexes for a table, you can use the DB05. To interpret the results : an index is representative when you have numerous distinct values for a few 'rows per generic keys'
    Hope this will help you

  • Strage Performance Problems with SQL Statements....

    Hi,
    I have realized a strange performance collapse while changing simple things in SQL Statements...
    I have build a simple Statement:
    SELECT
    a.AAA,
    a.BBB,
    a.CCC,
    b.AAA
    from
    TABLE_A a,
    TABLE_B b,
    TABLE_C c
    where
    a.XXX=b.XXX
    and c.XXX = a.XXX
    and c.yyy = 'SOMETHING'
    Its very fast, even with complex XSL Transformations...
    After Putting a GROUP BY or DISTINCT into the statement (to surpress Dataset "Clones") it tooks around 100 times more Time then before....
    I have tested the statement in a SQLPlus, it was as fast as before, but in the XMLPublisher it tooks much longer.....
    Have you realizesd this Problem before?
    Greetings...

    PROBLEM SOLVED !
    It has to be 8.1.6 across the whole environment.
    null

  • Can't access root share sometimes and some strange performance issues

    Hi :)
    I'm sometimes getting error 0x80070043 "The network name cannot be found" when accessing \\dc01 (the root), but can access shares via \\dc01\share.
    When I get that error I also didn't get the network drive hosted on that server set via Group Policy, it fails with this error:
    The user 'W:' preference item in the 'GPO Name' Group Policy Object did not apply because it failed with error code '0x80070008 Not enough storage is available to process this command.' This error was suppressed.
    The client is Windows Server 2012 Remote Desktop and file server is 2012 too. On a VMware host.
    Then I log off and back on, and no issues.
    Maybe related and maybe where the problem is: When I have the issue above and sometimes when I don't (the network drive is added fine) I have some strange performance issues on share/network drive: Word, Excel and PDF files opens very slowly. Offices says
    "Contacting \\dc01\share..." for 20-30 sec and then opens. Text files don't have that problem.
    I have a DC02 server also 2012 with no issues like like this.
    Any tips how to troubleshoot?

    Hi,
    Based on your description, you could access shares on DC via
    \\dc01\share. But you couldn’t access shares via \\dc01.
    Please check the
    Network Path in the Properties of the shared folders at first. If the network path is
    \\dc01\share, you should access the shared folder by using
    \\dc01\share.
    And when you configure
    Drive Maps via domain group policy, you should also type the Network Path of the shared folders in the
    Location edit.
    About opening Office files very slow. There are some possible reasons.
     File validation can slow down the opening of files.
     This problem caused by the issue mentioned above.
    Here are a similar thread about slow opening office files from network share
    http://answers.microsoft.com/en-us/office/forum/office_2010-word/office-2010-slow-opening-files-from-network-share/d69e8942-b773-4aea-a6fc-8577def6b06a
    For File Validation, please refer to the article below,
    Office 2010 File Validation
    http://blogs.technet.com/b/office2010/archive/2009/12/16/office-2010-file-validation.aspx
    Best Regards,
    Tina

  • A very strange synchronization problem

    Hi Everybody,
    I would like to get some help regarding a strange synchronization problem that I have found. I have created a java server application and after a lot of testing I have found the following problem:
    A certain thread (Thread-A) is stuck forever on a synchornized (objX) line, whereas another thread (Thread-B) every 30 seconds performs a section which is protected by the same synchornized (objX) line.
    It seems that the objX is not held by any thread, but Thread-A is stuck forever.
    The BUG is reproducible after the server runs about 1 day.
    The server is running on Linux machine (Red Hat 7.3) with JVM 1.4 (Java(TM) 2 Runtime Environment, Standard Edition (build 1.4.1_01-b01)).
    I would appreciate any help.
    Thanks In Advance,
    Idan Zach

    This is not a starvation problem since Thread-B performs its job every 30 seconds. The job is very short (I have added log messages before Thread-B starts its job and after it completes the job).
    However, the priority of both threads are the same.
    Thanks In Advance,
    Idan Zach

  • Strange wifi problem on a brand-new MacBook Pro

    Hey there, I'm new here. Desperation has got the better of me, and I hope I can get some advice.
    So, last sunday I bought a brand new 13" MacBookPro with 2.8g i7 processor and 10.7.2 OS (I can't update to 10.7.4 because System Update tells me the download has been tampered with or corrupted after I waited 2 hours for it to finish, but I'm trying a direct download from the website). Here's the exact product: http://www.amazon.com/Apple-MacBook-MD314LL-13-3-Inch-VERSION/dp/B005CWIZ4O/ref= sr_1_5?ie=UTF8&qid=1340301247&sr=8-5&keywords=macbook+pro+13
    Everything is all perfect except for a strange wifi problem. Note that the wifi does not turn-on or off, it doesn't disconnect after 'sleep', it doesn't get less speed, it doesn't constantly leave a preferred connection (once I chose my home wifi it always connects to it no problem). The wifi bar is always full. What is wrong is that the internet is kinda 'glitchy'. The best metaphor I can think of is that it's like a car that has to be pushed down the road before the engine can start. Please, bear with me because this is kinda hard to explain and I'm not good at tech lingos.
    One of the symptoms is browsing can get really slow and inconsistent, almost as if the wifi gets turned off in the middle of browsing. For example: if I look something up on Google and then open 5 new tabs, probably only 2 or 3 of those tabs will eventually load while the other two will not get past the long 'http:.......google.....etc etc' address and will eventually show one of those 'error loading page because there's no Internet connection' notice. This happens a lot. Sometimes I have to refresh/retry loading a page a couple of times before it finally loads, then usually everything will get back to normal and I will be able to open multiple tabs with no problem, but in a couple of minutes I would try to open another site and the problem will start again.
    So I tried to see how it does with downloading using iGetter. Fortunately I also have a Macbook (white-plastic casing one) and my brother has an old 2006 MacbookPro so I can compare. With those two older Apples iGetter starts receiving data in less than 8 seconds after I click Start, while my brand-new Macbook Pro can take up to 15-20 seconds, sometimes it doesn't even start at all and I have to re-click the start button. But once the download actually starts (on my new MacBook Pro), it downloads like normal, it doesn't stop and the speed is also as good as I can get. Strange huh? Hence, like a car that has to be pushed before the engine starts.
    I also tried 'speed testing' from a website with all 3 laptops. On my old MacBook and my brother's MacBook pro, once I click 'begin test' it takes only a couple of seconds of 'ping' before the download/upload speed measurements start, but with my new MacBook Pro (just like with iGetter) it takes up to almost a minute before the measurements start, sometimes it won't start at all and I have to re-start the test.
    Well, thanks for reading. And hope you can help me shed some light into this. Because it is really preventing me from really enjoying my new MacBook Pro. And if this is most likely a hardware problem, I'd better get back to the store and try to get it replace or something while it's only been 5 days. *cheers*

    Hey guys.. Thanks for the replies. Unfortunately, I bought the MacBook in Indonesia where there are no such thing as an AppleCare. The only option is to leave the Apple product to one of the 'Authorized Apple care centre' and have whoever do whatever to it without guarantee of success. As for returning it and getting a brand new one, I also don't know if there is such a deal here, but that's actually a good idea, will try to go to the store tomorrow. Although, I'm worried that the problem might be with OSX Lion or something and getting a new one won't help.
    mark75090, I've tried resetting the router when I first noticed te problem, didn't work though. I've checked my new MacBook's wi-fi (option+click wifi bar) and it says 'PHY Mode: 802.11g', so does my MacBook and the old MacBook Pro. I don't know what those mean though, and I don't know whether or not my router is an N., can you tell me more please.. I have tried with the other two turned off, and only the new MacBook using the wi-fi but the problem persists. I have also tried connecting directly to the router, but as I described, the problem is too subtle for me to notice the difference.
    Also, if I create a brand new Wi-Fi connection in network preferences and connect to my router through the new Wi-Fi profile, the problem seems to be gone for an hour or so, then it always comes back though. I have also just finally installed 10.7.4, but it doesn't help either.

  • PL/SQL Performance problem

    I am facing a performance problem with my current application (PL/SQL packaged procedure)
    My application takes data from 4 temporary tables, does a lot of validation and
    puts them into permanent tables.(updates if present else inserts)
    One of the temporary tables is parent table and can have 0 or more rows in
    the other tables.
    I have analyzed all my tables and indexes and checked all my SQLs
    They all seem to be using the indexes correctly.
    There are 1.6 million records combined in all 4 tables.
    I am using Oracle 8i.
    How do I determine what is causing the problem and which part is taking time.
    Please help.
    The skeleton of the code which we have written looks like this
    MAIN LOOP ( 255308 records)-- Parent temporary table
    -----lots of validation-----
    update permanent_table1
    if sql%rowcount = 0 then
    insert into permanent_table1
    Loop2 (0-5 records)-- child temporary table1
    -----lots of validation-----
    update permanent_table2
    if sql%rowcount = 0 then
    insert into permanent_table2
    end loop2
    Loop3 (0-5 records)-- child temporary table2
    -----lots of validation-----
    update permanent_table3
    if sql%rowcount = 0 then
    insert into permanent_table3
    end loop3
    Loop4 (0-5 records)-- child temporary table3
    -----lots of validation-----
    update permanent_table4
    if sql%rowcount = 0 then
    insert into permanent_table4
    end loop4
    -- COMMIT after every 3000 records
    END MAIN LOOP
    Thanks
    Ashwin N.

    Do this intead of ditching the PL/SQL.
    DECLARE
    TYPE NumTab IS TABLE OF NUMBER(4) INDEX BY BINARY_INTEGER;
    TYPE NameTab IS TABLE OF CHAR(15) INDEX BY BINARY_INTEGER;
    pnums NumTab;
    pnames NameTab;
    t1 NUMBER(5);
    t2 NUMBER(5);
    t3 NUMBER(5);
    BEGIN
    FOR j IN 1..5000 LOOP -- load index-by tables
    pnums(j) := j;
    pnames(j) := 'Part No. ' || TO_CHAR(j);
    END LOOP;
    t1 := dbms_utility.get_time;
    FOR i IN 1..5000 LOOP -- use FOR loop
    INSERT INTO parts VALUES (pnums(i), pnames(i));
    END LOOP;
    t2 := dbms_utility.get_time;
    FORALL i IN 1..5000 -- use FORALL statement
    INSERT INTO parts VALUES (pnums(i), pnames(i));
    get_time(t3);
    dbms_output.put_line('Execution Time (secs)');
    dbms_output.put_line('---------------------');
    dbms_output.put_line('FOR loop: ' || TO_CHAR(t2 - t1));
    dbms_output.put_line('FORALL: ' || TO_CHAR(t3 - t2));
    END;
    Try this link, http://download-west.oracle.com/docs/cd/B10501_01/appdev.920/a96624/05_colls.htm#23723

  • Performance problem in Zstick report...

    Hi Experts,
    I am facing performance problem in Custoom Stock report of Material Management.
    In this report i am fetching all the materials with its batches to get the desired output, at a time this report executes 36,000 plus unique combination of material and batch.
    This report takes around 30 mins to execute. And this report is to be viewed regularly in every 2 hours.
    To read the batch characteristics value I am using FM -> '/SAPMP/CE1_BATCH_GET_DETAIL'
    Is there any way out to increase the performance of this report, the output of the report is in ALV.
    May i have any refresh button in the report so that data may get refreshed automatically without executing it again. or is there any cache memory concept.
    Note: I have declared all the itabs with type sorted, all the select queries are fetched with key and index.
    Thanks
    Rohit Gharwar

    Hello,
    SE30 is old. Switch on trace on ST12 while running this progarm and identify where exactly most of the time is being spent. If you see high CPU time this problem with the ABAP code. You can exactly figure out the program/function module from ST12 trace where exactly time is being spent. If you see high database time in ST12, problem is with database related issue. So basically you have to analyze sql statement from performance traces in ST12. These could resolve your issue.
    Yours Sincerely
    Dileep

Maybe you are looking for

  • 2013 Mac Mini slow...

    I bought this Mac Mini less than 5 months ago, and for whatever reason it's running slow as ****.  Nothing extra is open, very minimal programs, totaling less than 10% cpu and 100MB ram.  So why is it running slow?  Startup took like 5 minutes, and e

  • URGENT inventory reports

    Hi friends, I would like to see a report by Product line/Profit center, I checked these reports, MCBE,MMBE,MCBC,ME80AN. Where I don’t see any field that can show Profit center, And also I would like to see a inventory report  by material, qty, batch,

  • Google app run on mac outlook 2011

    is there a way to setup google mail to run on mac outlook 2011?

  • Do i need to re-set bootblk after replaced the second SCSI hdd in sun box?

    I have a sun box with 2 SCSI hdd inside. The second hdd faulty.. then i follow the instruction to detach and metaclear..and metainit again... but i am not sure that do i need to set bootblk again in the new hdd (2nd)? after metainit, i have attach to

  • After Effects CS5 Installation Problem

    I'm running Windows 7 64 bit and can't seem to install a trial verion of CS5 After Effects.  I have a licensed version of CS5 Design Basic installed.  I've followed all of Adobe's troubleshooting tips and Adobe Support Advisor is not detecting any pr