Huge Memory Consumption for Transparent Stage

When executing the test below the memory consumption in the Task Manager grows up to 2GB within seconds on my test system - the issue does not occur when the stage style is changed from TRANSPARENT to UNDECORATED. Tested on Win7, 8GB with FX8b116.
public class TransparentStageMemoryIssueTest extends Application
  public static void main(String[] args)
    Application.launch(args);
  @Override
  public void start(Stage stage)
    System.err.println(System.getProperty("javafx.runtime.version"));
    //stage.initStyle(StageStyle.UNDECORATED);
    stage.initStyle(StageStyle.TRANSPARENT);
    final BorderPane root = new BorderPane();
    root.setStyle("-fx-background-color:#808080;-fx-background-radius:30;-fx-border-radius:30;-fx-border-width:14;-fx-border-color:blue;");
    Scene scene = new Scene(root, 800, 600);
    stage.setScene(scene);
    stage.setX(10);
    stage.setY(10);
    stage.show();
    new Thread()
      public void run()
        for (int i = 0; i < 1000; i++)
          Platform.runLater(new Runnable()
            @Override
            public void run()
              int minSize = 500;
              int maxSize = 1000;
              int w = minSize + new Random().nextInt(maxSize - minSize);
              int h = minSize + new Random().nextInt(maxSize - minSize);
              Window win = root.getScene().getWindow();
              win.setWidth(w);
              win.setHeight(h);
          try
            sleep(50);
          catch (InterruptedException e)
            e.printStackTrace();
    }.start();

Could you post this issue to the JavaFX Jira as well as an email to openjfx-dev?
Thanks,
-- Jonathan

Similar Messages

  • Huge Memory Consumption

    Hello Everyone,
    I have been checking a maxdb system and can see huge memory consumption occurring:
    top - 12:29:10 up 842 days,  2:59,  5 users,  load average: 18.59, 17.80, 11.77
    Tasks: 728 total,   2 running, 723 sleeping,   0 stopped,   3 zombie
    Cpu(s):  8.5%us,  0.7%sy,  0.0%ni, 88.8%id,  0.3%wa,  0.0%hi,  0.3%si,  1.4%st
    Mem:     47509M total,    47129M used,      379M free,      210M buffers
    Swap:    20479M total,     1475M used,    19004M free,     5386M cached
      PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
    26644 sdb       20   0 39.8g  38g 306m S  126 82.3  52654:06 kernel
    7136 sdb       20   0 77304  16m 4692 S    8  0.0   1:12.69 dbmsrv
    2860 root      20   0     0    0    0 S    3  0.0   2837:24 nfsiod
      208 root      20   0     0    0    0 S    2  0.0   2448:01 rpciod/2
    I could not see any long running db transactions, background jobs, deadlocks
    Please provide any valuable inputs on how to check the issue further
    Regards,
    Vinay

    Hello Vinay,
    what are you wondering about?
    That the database uses 82.3% of the memory?
    In MAXDB there exists a parameter called:
    CACHE_SIZE (MaxDB < 7.7.03) respectively CacheMemorySize (MaxDB >= 7.7.03)
    "The value specifies the i/o capable memory used by different
    MaxDB components in particular the data cache, converter and
    the shared catalog cache."
    When the database starts it allocates the memory given by the parameter. This memory is reserved for the db. It will not become free if there is no activity on the db.
    See also:
    MaxDB Memory Sizing - MaxDB - SCN Wiki
    Regards,
    Michael

  • How to find out memory consumption for table in HANA without load it into memory

    Hi,
    To determine the memory consumption for a table in HANA, you can query table M_CS_TABLES, however, it requires load table into memory first, I just wonder if there has another table store memory consumption information for all HANA tables regardless it load into memory or not. Below is screenshot for one of table in my system, since that table is partially loaded into memory, "Total Memory Consumption (KB):" tell me the memory consumption of the portion load into memory, what I am looking for is something like "Estimated Maximun Memory Consumption (KB)" which provides me total memory consumption for that table including portion doesn't load into memory, of course I can use this Esitmated information, but consider I have close to thousand tables in my HANA system already, it's not pratical to check tables one by one.
    Thanks,
    Xiaogang.

    Hi Xiaogang,
    Estimated Memory Size that you see in the Table Run time Information - same is available in M_CS_TABLES also
    If you don't get the size of any Table in M_CS_TABLES View, then the same will also not be available in Runtime information of the Table
    Even if tables are not loaded into memory, you can get the Estimated Size, just try running the query with filter LOADED = 'NO'
    Regards,
    Vivek

  • Memory consumption for an ASO app with a very large outline

    Solaris 10 running Essbase 9.3.1.3.08 64-bit
    We have an app with a very large outline (3Gb). When we do a dimension restructure, it takes about 30 minutes which I suppose is reasonable. However, it takes up about 13Gb RAM to do this and doesn't release the memory when finished unless we restart the app.
    Is this normal? It seems an incredible amount of RAM to use, even for such a large outline and it doesn't release it when done.
    The box has 32Gb RAM.

    I think it was version 9.3.1.3 that outline compaction started working. The first thing I would try is to compact the outline. The MaxL statement for it is Alter database xxxx.yyyy compact outline;
    ASO Outlines tend to grow with every change, do you really have that many members and attributes that would make that big an outline. If it keeps growing at some point it will corrupt

  • Portal Session Memory Consumption

    Dear All,
                          I want to see the user sessions memory consumption for portal 7.0. i.e. if a Portal user opens a session, how much memory is consumed by him/her. How can i check this. Any default value that is associated with this?
    Backend System memory load will get added to portal consumption or to that specific Backend System memory consumption.
    Thanks in Advance......
    Vinayak

    I'm seeing the exact same thing with our setup (it essentially the same
    as yours). The WLS5.1 documentation indicates that java objects that
    aren't serializeable aren't supported with in-memory replication. My
    testing has indicated that the <web_context>._SERVLET_AUTHENTICATION_
    session value (which is of class type
    weblogic.servlet.security.ServletAuthentication) is not being
    replicated. From what I can tell in the WLS5.1 API Javadocs, this class
    is a subclass of java.lang.object (doesn't mention serializeable) as of
    SP9.
    When <web_context>._SERVLET_AUTHENTICATION_ doesn't come up in the
    SECONDARY cluster instance, the <web_context>.SERVICEMANAGER.LOGGED.IN
    gets set to false.
    I'm wondering if WLCS3.2 can only use file or JDBC for failover.
    Either way, if you learn anything more about this, will you keep me
    informed? I'd really appreciate it.
    >
    Hi,
    We have clustered two instances of WLCS in our development environment with
    properties file configured for "in memory replication" of session data. Both the
    instances come up properly and join the cluster properly. But, the problem is
    with the in memory replication. It looks like the session data of the portal is
    getting replicated.
    We tried with the simplesession.jsp in this cluster and its session data is properly
    replicated.
    So, the problem seems to be with the session data put by Portal
    (and that is the reason why I am posting it here). Everytime the "logged in "
    check fails with the removal of one of the instances, serving the request. Is
    there known bug/patch for the session data serialization of WLCS? We are using
    3.2 with Apache as the proxy.
    Your help is very much appreciated.--
    Greg
    GREGORY K. CRIDER, Emerging Digital Concepts
    Systems Integration/Enterprise Solutions/Web & Telephony Integration
    (e-mail) gcrider@[NO_SPAM]EmergingDigital.com
    (web) http://www.EmergingDigital.com

  • Measuring Swing GUI memory consumption

    Hi community,
    First , I beg you for excuse for posting this here and not in the Swing forum. I think the question is not clear GUI question, but a more generic one.
    I have a GUI that seems to consume a lot of memory(i.e. opening several screens causes OutOfMemoryError and data received on network measures to a very small amount (<5%) compared to the total memory shown by Task Manager).
    I would like to take these screens apart to discover which of their parts are heavy so that I can concentrate my efforts on optimizing them. I cannot use ObjectOutputStream serialization since it seems there are lots of classes that do not implement Serializable, so I need some other way of determining component memory sizes. I believe that XMLEncoder serialzation will leed to nowhere because it contains a lot of text describibg properties that cannot be accurately measured.
    What can I use to measure the memory consumption for a given component tree?
    Thank you for your time
    Mike

    Swing related questions should be...
    Umm have you considered the use of a Profiler? I kind of doubt it is the GUI per se but a model or something that is leaking memory. I mean there are lot's of fairly complex GUI's out there that don't have these issues so I suspect there is a very deep or complex structure somewhere that is eating all of the memory.
    But in short I think running in some sort of profiler would be good becuase I also suspect that this may be one of this issues where looking at each component may not help because it's as the whole thing is used at runtime where some complexity comes into play and it dies.
    Try http://www.manageability.org/blog/stuff/open-source-profilers-for-java for a start.

  • Continuously refreshing a tab after an interval leads to high memory consumption (400MB to 800MB in 30 seconds for 3 refreshes at 10 secs interval), why?

    Environment:
    MAC OSX 10.9.5
    Firefox 32.0.3
    Firefox keeps consuming lot of memory when you keep refreshing a tab after an interval.
    I opened a single tab in my firefox and logged into my gmail account on that. At this stage the memory consumption was about 400MB. I refreshed the page after 10 seconds and it went to 580MB. Again i refreshed after 10 seconds and this time it was 690MB. Finally, when i refreshed 3rd time after 10 seconds, it was showing as 800MB.
    Nothing was changed on the page (no new email, chat conversation, etc. nothing). Some how i feel that the firefox is not doing a good job at garbage collection. I tested this use case with lot of other applications and websites and got the similar result. Other browsers like Google chrome, safari, etc. they just work fine.
    For one on of my application with three tabs open, the firefox literally crashed after the high memory consumption (around 2GB).
    Can someone tell me if this is a known issue in firefox? and is firefox planning to fix it? Right now, is there any workaround or fix for this?

    Hi FredMcD,
    Thanks for the reply. Unfortunately, i don't see any crash reports in about:crashes. I am trying to reproduce the issue which will make browser to crash but somehow its not happening anymore but the browser gets stuck at a point. Here is what i am doing:
    - 3 tabs are open with same page of my application. The page has several panels which has charts and the javascript libraries used for this page are backbone.js, underscore.js, require.js, highcharts.js
    - The page automatically reloads after every 30 seconds
    - After the first loading of there three tabs, the memory consumption is 600MB. But after 5 minutes, the memory consumption goes to 1.6GB and stays at this rate.
    - After sometime, the page wont load completely for any of the tabs. At this stage the browser becomes very slow and i have to either hard refresh the tabs or restart the browser.

  • Is there any way to avoid a huge memory spike when using CGContextDrawImage for a high resolution image?

    Hi,
    I'm trying to fix the orientation of an UIImage in my app (iPad 2, iOS 7). I'm using the method described here: https://gist.github.com/alex-cellcity/1531596, but a huge memory spike appears when CGContextDrawImage is called. The image size I'm trying to fix is 2848x4288.
    Anybody knows any algorithm to avoid this huge spike?
    Thanks,
    Roberto.

    While I never done iOS code it appears that the code you link to basically does the following:
    takes a large image A
    generates a transformation matrix by doing rotation in 90 degree increments and by mirroring
    creates a large image B
    draws image A into B using the transformation matrix
    releases memory for image A
    goes on using image B
    This results in a memory peak as you have two large images in RAM, until image A gets deallocated, which can't be done until B has been drawn.
    I can see two possible solutions to this, both which probably require you to explicitly modify individual pixels:
    1) Try to simplify the solution - if 180 degree rotation and mirroring is enough to solve your image issues, then a simple in-place (same bitmap) pixel swap could be done.
    2) Find more memory -  if you need 90, 270 or arbitrary degrees of rotation as well, you could instead keep either image A or B on disc.
    Keep image A on disc: read pixels from disc and place appropriately in image B. This may require you to first load image A and write it back to disc as ARGB data (one byte for each color) for easy access.
    Keep image B on disc: get pixel from image A and place it at the right location in image (file) B - this will require pre-creating a file of the correct size for Bs ARGB data and then moving the insertion point for writing around. When done dealloc image A and load image B.
    Keep image A and B on disc: combine using disc for image A and B - this will minimize RAM usage but will probably be the slowest solution and will increases disc (SSD/flash) load and wear.

  • High memory consumption in XSL transformations (XSLT)

    Hello colleagues!
    We have the problem of a very high memory consumption when transforming XML
    files with CALL TRANSFORMATION.
    Code example:
    CALL TRANSFORMATION /ipro/wml_translate_cls_ilfo
                SOURCE XML lx_clause_text
                RESULT XML lx_temp.
    lx_clause_text is a WordML xstring (i.e. it is a Microsoft Word file in XML
    format) and can therefore not be easily splitted into several parts.
    Unfortunately this string can get very huge (e.g. 50MB). The problem is that
    it seems that CALL TRANSFORMATION allocates memory for the source and result
    xstrings but doesn't free them after the transformation.
    So in this example this would mean that the transformation allocates ~100MB
    memory (50MB for source, ~50MB for result) and doesn't free it. Multiply
    this with a couple of transformations and a good amount of users and you see
    we get in trouble.
    I found this note regarding the problem: 1081257
    But we couldn't figure out how this problem could be solved in our case. The
    note proposes to "use several short-running programs". What is meant with
    this? By the way, our application is done with Web Dynpro for ABAP.
    Thank you very much!
    With best regards,
    Mario Düssel

    Hi,
    q1. how come the Ram consumption is increased to 99% on all the three boxes?If we continue with the theory that network connectivity was lost between the hosts, the Coherence servers on the local hosts would form their own clusters. Prior to the "split", each cache server would hold 1/12 of the primary and 1/12 of the backup (assuming you have one backup). Since Coherence avoids selecting a backup on the same host as the primary when possible, the 4 servers on each host would hold 2/3 of the cache. After the spit, each server would hold 1/6 of the primary and 1/6 of the backup, i.e., twice the memory it previously consumed for the cache. It is also possible that a substantial portion of the missing 1/3 of the cache may be restored from the near caches, in which case, each server would then hold 1/4 of the primary and 1/4 of the backup, i.e., thrice the memory it previously consumed for the cache.
    q2: where is the cache data stored in the coherence servers?on which memory?The cache data is typically stored in the jvm's heap memory area.
    Have you reviewed the logs?
    Regards,
    Harv

  • Query memory consumption

    Hi,
    Need some expert in SQL here. May i know how much memory (RAM) consumption for a simple query like 'SELECT SUM(Balance) FROM OCRD' cost.
    What about query like
    select (select sum(doctotal) from ordr) + (select sum(doctotal) from odln) + (select sum(doctotal) from oinv)
    How much memory would it normally takes? The reason is that i have a query that is quite similar to this and it would be run quite often. So i wonder if it is feasible to use this type of queries withought making the server to a crawl.
    Please note that the real query would include JOINS and such. Thanks
    Any information is appreciated

    Hi Melvin,
    Not sure I'd call myself an expert but I'll have a go at an answer
    I think you are going to need to set up a test environment and then stress test your solution to see what happens. There are so many different variables that affect the memory consumption that no-one is likely to be able to say just what the impact will be on your server. SQL Server, by default will allocate 1024Kb to each query but, of course, quite a number of factors will affect whether SQL needs more memory than this to execute a particular query (e.g. the number of joins, the locks created, whether the data is grouped or sorted, the size of the data etc etc). Also, SQL will release memory as soon as it can (based on its own algorithms) so a query that is run periodically has much less impact on the server than a query that will be run concurrently by multiple users. For these reasons, the impact can only really be assessed if you test it in a real-world scenario.
    If you've ever seen SQL Server memory usage when XL Reporter is running a very large report then you'll know that this is a very memory hungry operation. XL Reporter bombards SQL with a huge number of separate little queries and SQL Server starts grabbing significant amounts of memory to fulfill these queries. As the queries are coming so fast, SQL hasn't yet got around to releasing the memory used by previous queries so SQL instead grabs available memory from the server.
    You'll get better performance and scaleability by using stored procedures but SDK certification does not allow the use of SPs in the SBO databases.
    Hope this helps,
    Owen

  • Memory handling for big footage, wrong settings, bug?

    Hi
    I am getting huge time differences in rendering, a full res render takes ca 100 times longer than a half res, where I would normally expect maybe 4 times longer.
    Here's the details:
    I'm making an 30 sec animation in NTSC, the material is high resolution illustrations in .psd format with lots of layers. I parent and animate the layers in after effects, ie no cell animation or sequences.
    A couple of characters with loose limbs and some backgrounds, and the comp is probably a few hundred layers nested within comps etc.
    All this is fine when previewing in half resolution, it takes a little while for the program to load all the layers for every new scene, and then it ticks along good, since it's static illustrations with position and rotation keys.
    Previewing the 30 secs in half res takes a couple of minutes tops. Totally fine.
    But when switching to full res everything goes unbelievably slow. It takes more than 1 hour 30 minutes. The exact same animation that previewed in 1-2 minutes in half res.
    I am using
    AE CS3, 8.0.2.27
    8-core Mac Pro, Leopard 10.5.3
    10 GB Ram
    Multiprocessor rendering is switched on.
    I see all the AE instances in the Activity Monitor, most of the time they're working on less than 10%.
    So I'm guessing the cache is sufficient for the half res stuff, and for full res it has to load everything for every single frame? Still though, it seems that it's even slower than that. It seems like there's some sort of bottle neck.
    Some good advice would be greatly appreciated.
    Thanks, L

    Turn off multiprocessing. Loading and unloading your hunky files may take way longer than the actual processing. For the half res previews, they would be loaded from the cache in many situations, thus not affecting I/O speed at all. Also your math is not really right. Depending on how an effect works, render times can go up indefinitely, e.g. if it works strictly pixel-per-pixel insrtead of a coherent single one-pass buffer. In any case, the res is 2^2 times as high, so a multiplier of 8 is to be expected on memory consumption and possibly processing as well... Not sure where your project goes wrong, though, would require soem more info about footage types, blendmodes, effects and so on.
    Mylenium

  • Measure thread's memory consumption

    Hello.
    Nice to see you here.
    Please tell me is there any possibility to measure thread's memory consumption?
    I'm trying to tune application server.
    Totally physical server with Power AIX 5.3 on board has 8GB of memory.
    For example I allocate 1408m for Application Server Java heap (-Xms1408m -Xmx1408m).
    Then I tune Application server thread pools (web-threads, EJB-threads, EJB alarm threads, etc...).
    As I understood Java treads live in native memory, not in Java heap.
    I would like to know how to measure size of thread in native memory.
    After that I can set size of thread pools (to avoid OutOfMemory native or heap).

    holod wrote:
    As I understood Java treads live in native memory, not in Java heap.The data the JVM uses to manage threads may live in the JVM's own memory outside of the Java heap. However, that data will be a very tiny fraction of what the JVM is consuming (unless you have huge number of threads, which are all using very, very little memory).
    I would like to know how to measure size of thread in native memory. It will almost certainly be so small as to not matter.
    After that I can set size of thread pools (to avoid OutOfMemory native or heap).No, that will almost certain not help at all.

  • SetTransform() memory consumption

    Hi,
    I'm currently working on a application which needs to move a sphere very quickly. The position is calculated every 40 ms and set via the TransformGroup.setTransfom() method. This raises a problem as this statement rapidly consumes huge amounts of memory, especially when called in short time intervals.
    I also tested it with the java3d example program "AWTInteraction" by simply putting the statement in a for loop and watched the memory climb:
    for (int i=0; i<1e+6; i++)
         objTrans.setTransform(trans);The result is a java.lang.OutOfMemoryError
    Is there a soultion or workaround for this kind of problem?
    Any hints appreciated. (Project has to be finished on Monday. It's really urgent.)
    TIA
    Erich

    Erich,
    i've never had any memory problems when dealing with Transforms. Is it perhaps possible that the leakage results from another instruction? For high memory consumption I saw responsible only working with textures so far. In order to find the problem, I would check out three things in your code:
    1. Are you working with textures and if yes, how does the program behave, if the textures are omitted?
    2. Are there any "new" instructions inside your loop? If yes, try to reuse objects and eliminate all "new" commands inside the loop.
    3. Did you consider the Mantra to do all changes on a live scene graph within a behavior (and from the behavior scheduler)? It seems unusual to me to change transforms inside a loop.
    Good luck,
    Oliver

  • How to measure memory consumption during unit tests?

    Hello,
    I'm looking for simple tools to automate measurement of overall memory consumption during some memory-sensitive unit tests.
    I would like to apply this when running a batch of some test suite targetting tests that exercise memory-sensitive operations.
    The intent is, to verify that a modification of code in this area does not introduce regression (raise) of memory consumption.
    I would include it in the nightly build, and monitor evolution of summary figure (a-ah, the "userAccount" test suite consumed 615Mb last night, compared to 500Mb the night before... What did we check-in yesterday?)
    Running on Win32, the system-level info of memory consumed is known not to be accurate.
    Using perfmon is more accurate but it seems an overkill - plus it's difficult to automate, you have to attach it to an existing process...
    I've looked in the hprof included in Sun's JDK, but it seems to be targetted at investigating problems rather than discovering them. In particular there isn't a "summary line" of the total memory consumed...
    What tools do you use/suggest?

    However this requires manual code in my unit test
    classes themselves, e.g. in my setUp/tearDown
    methods.
    I was expecting something more orthogonal to the
    tests, that I could activate or not depending on the
    purpose of the test.Some IDEs display mmeory usage and execution time for each test/group of tests.
    If I don't have another option, OK I'll wire my own
    pre/post memory counting, maybe using AOP, and will
    activate memory measurement only when needed.If you need to check the memory used, I would do this.
    You can do the same thing with AOP. Unless you are using an AOP library, I doubt it is worth additional effort.
    Have you actually used your suggestion to automate
    memory consumption measurement as part of daily builds?Yes, but I have less than a dozen tests which fail if the memory consumption is significantly different.
    I have more test which fail if the execution time is siginificantly different.
    Rather than use the setUp()/tearDown() approach, I use the testMethod() as a wrapper for the real test and add the check inside it. This is useful as different test will use different amounts of memory.
    Plus, I did not understand your suggestion, can you elaborate?
    - I first assumed you meant freeMemory(), which, as
    you suggest, is not accurate, since it returns "an
    approximation of [available memory]"freeMemory gives the free memory from the total. The total can change so you need to take the total - free as the memory used.
    - I re-read it and now assume you do mean
    totalMemory(), which unfortunately will grow only
    when more memory than the initial heap setting is
    needed.more memory is needed when more memory is used. Unless your test uses a significant amount of memory there is no way to measure it reliably. i.e. if a GC is perform during a test, you can have the test appear to use less memory than it consumes.
    - Eventually, I may need to inlcude calls to
    System.gc() but I seem to remember it is best-effort
    only (endless discussion) and may not help accuracy.if you do a System.gc(); followed by a Thread.yield() at the start it can improve things marginally.

  • Huge memory leaks after upgrading from kernel 2.6.38(?)

    My system began leaking memory some months ago and I've little to no idea what is causing it. The closest hint I can give is that I think these problems started around upgrading from Linux 2.6.38 to 2.6.39 and have continued with 3.0 aswell. Sometimes my whole system freezes due to extensive swapping (HDD usage led stays lit), sometimes it stops after consuming almost all of my 4 GB RAM.
    I think the leak has to come from inside the kernel because all sysmon apps claim no application is eating too much RAM yet the system memory consumption is at 100 % or even using swap. Even after I kill Xorg the memory usage will stay at least at 2.0 GB, when it should be less than half a gig.
    I'd like to guess the blame is on Nouveau, because it was giving me a hell of a time back with Linux 2.6.39. It had a habit of freezing my system when re-enabling suspended compositing (both with Kwin and Mutter). Also video playback was corrupted sometimes. These probs have since been fixed with Linux 3.0.
    I cannot name any spesific app that might trigger this memory leaking. I don't play any games and mainly use Firefox, Chromium and VLC day in, day out.
    There seems to be absolutely no way of reclaiming the reserved memory than rebooting the whole system. I have even tried unloading Nouveau.
    I am currently running the [testing] repo with ~daily upgrades.
    Last edited by Verge (2011-08-16 14:04:09)

    OK, this kind of surprised me. No wonder I couldn't find any relevant stuff while googling for kernel issues. I have had Strigi disabled for some time now because it is miserably broken, but I now disabled the whole Nepomuk too. Let's see if my problem now goes away or what...
    The release of KDE 4.7 Beta 1 actually lines up with the launch of Linux 2.6.39 so this really might be the case. I moved to KDE 4.7 already with the first beta.
    Last edited by Verge (2011-08-16 14:30:37)

Maybe you are looking for

  • I just switched to Charter. When going to Mozilla Firefox a box appears saying, Authentication Required and asks for a User name and Password. What do I do?

    I switched from CenturyTel to Charter. I can't use Firefox because a box appears asking me for Authentication Required, User Name and Password. When I hit cancel, this message appears: HTTP 401 - Unauthorized Authorization is required to access the c

  • Pop 3 mail on iPad

    I'm having trouble setting up my POP3 e-mail account on my iPad 3...I know all the settings I've entered are correct (verified with provider) but I receive the message the "secureserver.net" is not responding.  I enabled SSL and have shut down, reboo

  • Changing Pixel Colours

    Hi,      Is there a way of changing pixel colours in (say) a Graphics (or image) object? I want to be able to create a Graphics object, and then change pixel colours, as determined by the results of a numerical simulation, as results are generated. A

  • Handling process in a remote system

    how to list processes, start processes, and kill processes in a remote system from a server. and how to get configuration information of a remote system. Return information regarding: logical memory configuration, physical and logical disk and partit

  • Diff between wifi and wifi +3G?

    Thinking of getting my kids Ipads for Christmas. Which version do I need to get so they can grab hotspots when available? I do not want to have to buy data plans through AT&T. I assume it's the wifi +3G but not sure. Assume that because it's the more