Out of Memory ODBC - 100

Dear Experts,
         I have created a Query Print Layout to generate a picture report. When i run the report through manager id the report gets generated, on the other hand when i run the Query Print Layout from user/ id it throws the following error Out of Memory ODBC - 100. I tried checking the user giving super user rights but it did not generate report. The report screen is blank. Could you be kind enough to let me know any possible solution for the same.
Saravanan
Edited by: Sanbrahma39 on Jan 18, 2010 5:35 AM
Edited by: Sanbrahma39 on Jan 18, 2010 5:38 AM

Dear Rahul,
         I worked out your suggestion like dropping the SP from the SQL/FMS and try running the report. Again i face the same issue, when i am generating the report with manager user id the report is perfectly getting generated, same time when i try to generate the report with the user the issue arises.
Dear Gordon,
I agree that the pictures that is there are around 1000+ items in the report. But as stated earlier if the issue occurs across the users i could accept, but here for manager id is generating only at the user level it is not happening.

Similar Messages

  • Out of Memory (ODBC)

    Hi all,
    My client is currently experiencing an issue when trying to print a Purchase Order where he gets an error( out of memory(ODBC)) when previewing the Order.
    Please help,
    Aubrey

    Hi,
    If you have changed the layout in terms of adding any field without a link to the master then query will not give proper results and it will crash giving you a memory error.
    Try running the system template and compare that with your default PLD and track the changes.
    thanking you
    Malhaar

  • IMac CPU3.06 GHz Intel Core 2 Duo, 8GB, Extremely slow and repeatedly "out of Memory" since migration from Mountain Lion to Mavericks

    Dear Apple Support Community.
    As mentioned by a number of users reporting the same problem, my 2009 iMac CPU3.06 GHz Intel Core 2 Duo, 8GB RAM, has become extremely slow and repeatedly "out of Memory" since it was migrated from Mountain Lion to Mavericks (OS 10.9 then OS 10.9.1).
    I have already run CleanMyMac, deleted 5GB of useless stuffs, which did not translate into any improvement. Following the advices collected in other Apple Support Community threads, I have applied EtreCheck the outcome of which is provided below.
    For information, I have also upgraded my 2011 MacBook CPU2.3 GHz Core i7, 8GB which became only a bit slower than before, sometimes freezing which it almost never did before. However, I am used to the fact that Mac get usually slower and more prompt fo freeze after upgrading to a more recent OS.
    Both my 2009 iMac and my 2011 MacBook have F-Secure installed. So the problem with the 2009 IMac is unlikly to be due to the presence of F-Secure. Interestingly, the few abnormalities found by Etrecheck were exacly identical before I ran CleanMyMac. There are two more findings (mentioned below) since I ran CleanMyMac on the 2009 iMac, which again didn't improve the problem.
    So please find below the details of the EtreCheck experience and some additional measurements made with iStat while running EtreCheck:
    Hardware Information:
    iMac (24-inch, Early 2009)
    iMac - model: iMac9,1
    1 3.06 GHz Intel Core 2 Duo CPU: 2 cores
    8 GB RAM
    Video Information:
    ATI Radeon HD 4850 - VRAM: 512 MB
    Audio Plug-ins:
    BluetoothAudioPlugIn: Version: 1.0 - SDK 10.9
    AirPlay: Version: 1.9 - SDK 10.9
    AppleAVBAudio: Version: 2.0.0 - SDK 10.9
    iSightAudio: Version: 7.7.3 - SDK 10.9
    System Software:
    OS X 10.9.1 (13B42) - Uptime: 0 days 2:40:57
    Disk Information:
    WDC WD1001FALS-40K1B0 disk0 : (1 TB)
    EFI (disk0s1) <not mounted>: 209.7 MB
    XooS5 HD (disk0s2) /: 999.35 GB (189.85 GB free)
    Recovery HD (disk0s3) <not mounted>: 650 MB
    OPTIARC DVD RW AD-5670S 
    USB Information:
    Apple Inc. Built-in iSight
    Canon CanoScan
    Apple Computer, Inc. IR Receiver
    Apple Inc. BRCM2046 Hub
    Apple Inc. Bluetooth USB Host Controller
    FireWire Information:
    WD My Passport 071D 800mbit - 800mbit max
    EFI (disk1s1) <not mounted>: 209.7 MB
    disk1s2 (disk1s2) <not mounted>: 2 TB
    Boot OS X (disk1s3) <not mounted>: 134.2 MB
    Thunderbolt Information:
    Kernel Extensions:
    com.f-secure.kext.fsauth          (1.0.0d1)
    Problem System Launch Daemons:
    [failed] com.apple.wdhelper.plist [REMARK: this abnormality appeared only after running CleanMyMac. It does not appear on my 2011 MacBook]
    Problem System Launch Agents:
    Launch Daemons:
    [loaded] com.adobe.fpsaud.plist
    [invalid] com.bjango.istatlocaldaemon.plist [REMARK: this abnormality pre-existed to running CleanMyMac. It appears also on my 2011 MacBook]
    [loaded] com.f-secure.fsavd.dbhelper.plist
    [loaded] com.f-secure.fsavd.plist
    [loaded] com.f-secure.fsmac.firewall.plist
    [loaded] com.f-secure.fsmac.fsupdated_guts2.plist
    [loaded] com.f-secure.fsmac.licensetool.plist
    [loaded] com.macpaw.CleanMyMac2.Agent.plist
    Launch Agents:
    [loaded] com.babylon.activation.plist
    [loaded] com.bjango.istatlocal.plist
    [loaded] com.f-secure.relauncher.plist
    [loaded] com.f-secure.trasher.plist
    User Launch Agents:
    [loaded] com.adobe.ARM.[...].plist
    [loaded] com.google.keystone.agent.plist
    [loaded] com.macpaw.CleanMyMac2Helper.diskSpaceWatcher.plist
    [loaded] com.macpaw.CleanMyMac2Helper.scheduledScan.plist
    [loaded] com.macpaw.CleanMyMac2Helper.trashWatcher.plist
    User Login Items:
    iTunesHelper
    Dropbox
    F-Secure Anti-Virus for MAC
    3rd Party Preference Panes:
    Flash Player
    Internet Plug-ins::
    FlashPlayer-10.6: Version: 11.9.900.170 - SDK 10.6
    Default Browser: Version: 537 - SDK 10.9
    AdobePDFViewerNPAPI: Version: 11.0.04 - SDK 10.6
    AdobePDFViewer: Version: 11.0.04 - SDK 10.6
    DivXBrowserPlugin: Version: 2.1
    Flash Player: Version: 11.9.900.170 - SDK 10.6
    OfficeLiveBrowserPlugin: Version: 12.2.8
    Google Earth Web Plug-in: Version: 6.1
    QuickTime Plugin: Version: 7.7.3
    iPhotoPhotocast: Version: 7.0
    User Internet Plug-ins::
    Picasa: Version: 1.0
    Bad Fonts:
    None
    Old applications:
    None
    Time Machine:
    Skip System Files: NO
    Mobile backups: OFF
    Auto backup: YES
    Volumes being backed up:
    XooS5 HD: Disk size: 930.71 GB Disk used: 753.90 GB
    Destinations:
    XooS5 TM [Local] (Last used)
    Total size: 2 
    Total number of backups: 32
    Oldest backup: 2013-09-15 01:11:39 +0000
    Last backup: 2013-12-07 12:48:45 +0000
    Size of backup disk: Too small. Backup size 2  < (Disk used 753.90 GB X 3)
    Time Machine details may not be accurate. All volumes being backed up may not be listed.
    REMARK: The above finding is unlikely to be accurate. In fact, my Time Machine is supported by a 2TB WD My Passport Studio external HD connected with a Firewire 800. As mentionned, only 754GB of the 2000GB are used.
    Top Processes by CPU:
    5%          backupd
    5%          Mail
    5%          WindowServer
    3%          Finder
    3%          Dropbox
    Top Processes by Memory:
    1.60 GB          mds_stores REMARK: this item was marked in red in the outcome of EtreCheck, so I assume there is something wrong there
    82 MB          Finder
    74 MB          Mail
    49 MB          softwareupdated
    49 MB          Dropbox
    Virtual Memory Statistics:
    97 MB          Free RAM
    1.21 GB          Active RAM
    1.19 GB          Inactive RAM
    1.21 GB          Wired RAM
    1.83 GB          Page-ins
    36 MB          Page-outs
    Please find now some additional measurements collected with iStat while running EtreCheck:
    Practically:
    The iMac can work slowly with only one software e.g. either Mail or Safari or Firefox, but still the below "out of Memory" message occurs about every 3 minutes.
    Aperture has becomre to slow to be useable
    If trying to write a mail (with Mail) while Safari or Firefox is open, then the "out of Memory" message occurs about every 20 seconds so I need to check on "Resume" several times per minute.
    Conclusions
    From the above, unless anyone in the Apple Support Community proves to be able to identify a specific problem leading to a way to find a cure, I tend to conclude that Mavericks is not suitable for computers operating with a Core 2 Duo CPU even with a 8GB of RAM.
    Unless I missed something, I did not see any warning from Apple Computer Inc. informing their customers they should refrain from upgrading Core 2 Duo Computers.
    The only solution I see so far is to downgrade from Mavericks to Mountain Lion. And the Apple Support Community would confirm that it is the only solution, I would appreciate to get your guidance on how to proceed.
    Intended action plan
    Migrate the above 2009 iMac to a 2014 Mac Mini Server operating on Mavericks to ensure that no data may be lost while downgrading to Montain Lion.
    Provided the above downgrading operation is successful, assign this 2009 iMac to partial retirement e.g. Mail, Skype, iTunes  and DVD player.
    In advance, I would like to thank the Apple Community for helping me either i) finding a cure to the problem, ii) or confirming that Mavericks cannot operate properly on Core 2 Duo computers, iii) and in such a case, guiding me on how to downgrade to Mountain Lion.
    Yours sincerely
    Core

    ... what anti-virus would you recommend using?
    Use what is already included with OS X. It already includes everything it needs to protect itself from viruses and malware. Keep it that way with software updates from Apple.
    A much better question is "how should I protect my Mac":
    Never install any product that claims to "speed up", "clean up", "optimize", or "accelerate" your Mac. Without exception, they will do the opposite.
    Never install pirated or "cracked" software, software obtained from dubious websites, or other questionable sources. Illegally obtained software is almost certain to contain malware.
    Don’t supply your password in response to a popup window requesting it, unless you know what it is and the reason your credentials are required.
    Don’t open email attachments from email addresses that you do not recognize, or click links contained in an email:
    Most of these are scams that direct you to fraudulent sites that attempt to convince you to disclose personal information.
    Such "phishing" attempts are the 21st century equivalent of a social exploit that has existed since the dawn of civilization. Don’t fall for it.
    Apple will never ask you to reveal personal information in an email. If you receive an unexpected email from Apple saying your account will be closed unless you take immediate action, just ignore it. If your iTunes or App Store account becomes disabled for valid reasons, you will know when you try to buy something or log in to this support site, and are unable to.
    Don’t install browser extensions unless you understand their purpose. Go to the Safari menu > Preferences > Extensions. If you see any extensions that you do not recognize or understand, simply click the Uninstall button and they will be gone.
    Don’t install Java unless you are certain that you need it:
    Java, a non-Apple product, is a potential vector for malware. If you are required to use Java, be mindful of that possibility.
    Disable Java in Safari > Preferences > Security.
    Despite its name JavaScript is unrelated to Java. No malware can infect your Mac through JavaScript. It’s OK to leave it enabled.
    Block browser popups: Safari menu > Preferences > Security > and check "Block popup windows":
    Popup windows are useful and required for some websites, but popups have devolved to become a common means to deliver targeted advertising that you probably do not want.
    Popups themselves cannot infect your Mac, but many contain resource-hungry code that will slow down Internet browsing.
    If you ever see a popup indicating it detected registry errors, that your Mac is infected with some ick, or that you won some prize, it is 100% fraudulent. Ignore it.
    Ignore hyperventilating popular media outlets that thrive by promoting fear and discord with entertainment products arrogantly presented as "news". Learn what real threats actually exist and how to arm yourself against them:
    The most serious threat to your data security is phishing. To date, most of these attempts have been pathetic and are easily recognized, but that is likely to change in the future as criminals become more clever.
    OS X viruses do not exist, but intentionally malicious or poorly written code, created by either nefarious or inept individuals, is nothing new.
    Never install something without first knowing what it is, what it does, how it works, and how to get rid of it when you don’t want it any more.
    If you elect to use "anti-virus" software, familiarize yourself with its limitations and potential to cause adverse effects, and apply the principle immediately preceding this one.
    Most such utilities will only slow down and destabilize your Mac while they look for viruses that do not exist, conveying no benefit whatsoever - other than to make you "feel good" about security, when you should actually be exercising sound judgment, derived from accurate knowledge, based on verifiable facts.
    Do install updates from Apple as they become available. No one knows more about Macs and how to protect them than the company that builds them.
    Summary: Use common sense and caution when you use your Mac, just like you would in any social context. There is no product, utility, or magic talisman that can protect you from all the evils of mankind.

  • Cannot re-open project after Out of Memory message

    Hi,
    Problem with Encore CS3.
    Here is my problem: I have spent about 2 weeks working on a DVD project made of 2h movie (about 100 files and 2h of subtitles) and about 20 slideshows with a few hundred pictures and music for each slideshow. There are about 15 pages of menus and submenus. All in all it was getting longer and longer to save and open but it was okay.
    I eventually tried to produce the DVD in a hard drive folder. Then, after about 14h of computation I got the message "Out of memory". I had to quit at that point. Since then I cannot open the project anymore!
    When I open the project I get a window telling me how long it will take to open the project (with an incredibly wrong estimate by the way), once reached the 100%, it never gives me control back anymore. I have to kill the process with the task manager. I have tried to let it work for hours but that does not change much.
    The hardware configuration is the following:
    At the crash time: Windows XP pro SP2, Duo core processor, 2 GB RAM, a couple of 500GB hard drives connected in USB.
    I have since added 2 more GB RAM and now have 4GB with 3.4GB detected (as we all know Windows 32 cannot handle the full 4GB). I thought the add of the memory would make a difference. It does not! Encore uses up to 1.5GB of Ram while trying to open the project (the task manager tells me).
    I have tried many many times to re-open my project as I spent too much time on it to give-up easily.
    I actually managed to reopen the project a couple of times in a kind of degraded mode by removing some files from the "Transcodes" directory. I then get an error message and sometimes manage to get in. I managed twice amd twice immediately saved the project under a new name (over 7GB of files).
    The problem is: I have got a couple of error message while trying to re-setup the project as I left it (remember it was a degraded mode with missing files). I never managed to go to the build stage again and I cannot reopen my project consistently. I now have 3 versions of it. None is working. I am back to starting point, unable to open the project.
    If anyone has a solution to recover my project, it is much more than welcome! I hate the idea to start from scratch again. If I have to do so, I am seriously thinking to switch to Vegas and DVD architect because Premiere is unable to handle my new HDV files and I start to be fed-up with all the CS3 bugs. Saying that learning these 2 tools was long enough to not wish to learn 2 new ones.
    Thanks for taking time to read all my message.
    Thanks in advance for your answers.

    Well, here is the solution to my own problem. I give it to you in case you get the same problem one day.
    The solution is as simple as impossible to find on your own: launch Bridge before launching Encore! Yes, that's it. If Bridge is running when you start Encore and open your project, it all works just fine.
    In my case, the project was a bit damaged as many files where not located properly. I just hit the "Skip all" button when prompted and got into my project. Then I re-assigned each unfound file (in italic) with the original one. That was quite easy in the end.
    I find it incredible that Encore will work only when Bridge is launched. In that case, I can only recommend to have plenty of RAM. In all cases it is a nasty bug.
    I hope it will help someone some day.
    Cheers.

  • While creating DB using DBCA getting ORA-27102: out of memory in Linux

    Hi All,
    I am working on 11.2.0.3 oracle Redhat linux. I am getting error "ORA-27102: out of memory" while creating a new database using dbca
    Below are the DB ans OS details. Please check it and let me know what i need to do to overcome this issue.
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    PL/SQL Release 11.2.0.3.0 - Production
    CORE 11.2.0.3.0 Production
    TNS for Linux: Version 11.2.0.3.0 - Production
    NLSRTL Version 11.2.0.3.0 - Production
    $uname -a
    Linux greenlantern1a 2.6.18-92.1.17.0.1.el5 #1 SMP Tue Nov 4 17:10:53 EST 2008 x86_64 x86_64 x86_64 GNU/Linux
    $cat /etc/sysctl.conf
    # Controls the maximum shared segment size, in bytes
    kernel.shmmax = 68719476736
    # Controls the maximum number of shared memory segments, in pages
    kernel.shmall = 4294967296
    kernel.shmall = 2097152
    kernel.shmmax = 4294967295
    kernel.shmmni = 4096
    kernel.sem = 250 32000 100 128
    net.core.rmem_default = 4194304
    net.core.wmem_default = 262144
    net.core.rmem_max = 4194304
    net.core.wmem_max = 1048576
    fs.file-max = 6815744
    fs.aio-max-nr = 1048576
    net.ipv4.ip_local_port_range = 9000 65500
    $free -g
    total used free shared buffers cached
    Mem: 94 44 49 0 0 31
    -/+ buffers/cache: 12 81
    Swap: 140 6 133
    $ulimit -l
    32
    $ipcs -lm
    Shared Memory Limits
    max number of segments = 4096
    max seg size (kbytes) = 4194303
    max total shared memory (kbytes) = 8388608
    min seg size (bytes) = 1
    Also created a trace file under trace loction and it suggesting to changes shm parameter value. but i am not sure which parameter (shmmax or shmall) and value i need to modify.
    below are trace file info
    Trace file /u02/app/oracle/diag/rdbms/beaconpt/beaconpt/trace/beaconpt_ora_9324.trc
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORACLE_HOME = /u02/app/oracle/product/11.2.0.3
    System name: Linux
    Node name: greenlantern1a
    Release: 2.6.18-92.1.17.0.1.el5
    Version: #1 SMP Tue Nov 4 17:10:53 EST 2008
    Machine: x86_64
    Instance name: beaconpt
    Redo thread mounted by this instance: 0 <none>
    Oracle process number: 0
    Unix process pid: 9324, image: oracle@greenlantern1a
    *** 2012-02-02 11:09:53.539
    Switching to regular size pages for segment size 33554432
    Switching to regular size pages for segment size 4261412864
    skgm warning: ENOSPC creating segment of size 00000000fe000000
    fix shm parameters in /etc/system or equivalent
    Please let me what are the kernel parameter values i need to chage to work this.
    Thanks in advance.

    Yes it is same question, but i didn't have any solution there and still looking for some help. the solution it was provided in the last post is not working and getting the same error even with less thn 20% of memory. Please let me know how to overcome this issue.
    Thanks

  • Getting ORA-27102: out of memory while creating DB using DBCA

    Hi All,
    I am working on 11.2.0.3 oracle version and linux OS. I am trying to create a new database using dbca and getting error "ORA-27102: out of memory".
    Please find the DB version and OS level parameters info below and let me know what i need to do to overcome this issue.
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    PL/SQL Release 11.2.0.3.0 - Production
    CORE 11.2.0.3.0 Production
    TNS for Linux: Version 11.2.0.3.0 - Production
    NLSRTL Version 11.2.0.3.0 - Production
    $uname -a
    Linux greenlantern1a 2.6.18-92.1.17.0.1.el5 #1 SMP Tue Nov 4 17:10:53 EST 2008 x86_64 x86_64 x86_64 GNU/Linux
    $cat /etc/sysctl.conf
    kernel.shmall = 2097152
    kernel.shmmax = 4294967295
    kernel.shmmni = 4096
    kernel.sem = 250 32000 100 128
    net.core.rmem_default = 4194304
    net.core.wmem_default = 262144
    net.core.rmem_max = 4194304
    net.core.wmem_max = 1048576
    fs.file-max = 6815744
    fs.aio-max-nr = 1048576
    net.ipv4.ip_local_port_range = 9000 65500
    $free -g
    total used free shared buffers cached
    Mem: 94 44 49 0 0 31
    -/+ buffers/cache: 12 81
    Swap: 140 6 133
    $ulimit -l
    32
    $ipcs -lm
    ------ Shared Memory Limits --------
    max number of segments = 4096
    max seg size (kbytes) = 4194303
    max total shared memory (kbytes) = 8388608
    min seg size (bytes) = 1
    Please let me know for any other details.
    Thanks in advance.

    Ok, first, let's set aside the issue of hugepages for a moment. (Personally, IMHO, if you're doing manual memory mangement, and you're not using hugepages, you're doing it wrong.)
    Anyhow, looking at your SHM parameters:
    kernel.shmall = 2097152
    kernel.shmmax = 4294967295
    kernel.shmmni = 4096
    Let's take those in reverse order:
    1.) shmmni - This is the max number of shared memory segments you can have on your system, regardless of the size of each segment.
    2.) shmmax - Contrary to popular belief, this is NOT the max amount of shared memory you can allocate system wide! This is the max size, in bytes of a single shared memory segment. You currently have it set to 4GB-1. This is probably fine. Even if you wanted an SGA larger than 4GB, having shmmax set to this wouldn't hurt you. Oracle would simply allocate multiple shared memory segments, until it had allocated enough memory for the SGA. There's really no harm there, unless this parameter is set really low, causing a huge number of tiny shared memory segments to be allocated.
    3.) shmall - This is the real shared memory segment limit. This number is the total amount of shared memory you're permitted to allocate, system wide, expressed in pages. Pagesize here is the native OS pagesize, which is 4096 bytes, so, this is 2097152 * 4096 = 8589934592, or, 8GB. So, 8GB is the maximum amount of memory that can currnetly be allocated to shared memory, on your machine.
    So, having said all that, you haven't mentioned how many, if any, other Oracle databases are running on the server or their sizes. Secondly, we have no idea what memory sizing parameters you have set on the database that you're trying to create, that's getting the error.
    So, if you can provide more details, in terms of how many other databases are already on this server, and their SGA sizes, and the parameters you've chosen for the database that's failing to create, perhaps we can help more.
    Finally, if you're not using SGA_TARGET or MEMORY_TARGET, you really need to take the time to configure hugepages. Particularly if you've got a server that has as much memory as you do, and you're planning to have non-trivially sized SGA (10s of GB), then you really want to configure hugepages.
    Hope that helps,
    -Mark

  • Unable to create titles, getting "Out of memory" error message

    I am editing a project on FCE 2.0.3.
    I recently re-opened the project after about a year. It is stored on a 500GB external firewire hard-drive. I am working on a Mac G5 Powermac 7.2 with OS 10.4.11 with 1GB of RAM. I have my system settings on FCE set at 100% for memory usage and still cache. All my system settings on FCE are set to save to the external drive. I have 16GB available on my G5 memory.
    Every time I try to create a title using Title crawl, i get this error message:
    AE Effects Error: Out of memory using "title crawl."
    Also, every time I close the project and re-open it, I have to reconnect all the media and re-render everything, even though I save the project each time. I have also Saved As a different project name multiple times and this doesn't help.
    One other issue: The titles that I have on the project already look clear when the video is paused on a frame, but when they are moving they look very low-res and pixilated. The pixilated look also happens when I output my video to a QT movie. How can I make the titles look sharp?
    Please help!
    THANKS
    -Pierre
    Message was edited by: Pierre Degaillande1

    Before you do anything, trash your preferences exactly as detailed in this link:
    http://fcpbook.com/Misc1.html
    +"AE Effects Error: Out of memory using "title crawl.""+
    Try pushing down the memory usage and still cache a lot, and see if that helps. You only have 1GB of RAM, so keeping them at 100% is not a good idea. A computer can easily use about half a gigabyte just maintaining regular processes. To make sure you don't overload your RAM, you need to set the amount of memory FCE can use to a much smaller amount.
    +"Also, every time I close the project and re-open it, I have to reconnect all the media and re-render everything, even though I save the project each time. I have also Saved As a different project name multiple times and this doesn't help."+
    Are you moving files around on your scratch disk or renaming them? If so, you will have to reconnect them every time you change things.
    +"The titles that I have on the project already look clear when the video is paused on a frame, but when they are moving they look very low-res and pixilated. The pixilated look also happens when I output my video to a QT movie. How can I make the titles look sharp?"+
    Are you looking at the fully rendered movie? Click Option-R to render everything out.

  • ORA-27102: out of memory (while creation of drsite problem)

    Hi all,
    I am trying to create DRSITE at remote location, but whilw using the pfile of primary server i am getting the error ORA-27102: out of memory we are using Oracle 9.2 and RHEL and another is that in the primary server we are haing oracle 9.2.0.8 and at the drsite we are using oracle 9.2.0.6,actually aour patch got corrupted that's why we are using oracle 9.2.0.6, because of the differences os the patch creating a problem.....but i dno't think so..pls correct me if i am wrong
    SQL> conn sys/pwd as sysdba
    Connected to an idle instance.
    SQL> startup nomount pfile='/u01/initicai.ora';
    ORA-27102: out of memory
    SQL>we are haing total 8gb memory out of which we using 6gb for oracle i.e
    [oracle@icdb u01]$ cat /proc/meminfo
    MemTotal:      8175080 kB
    MemFree:         39912 kB
    Buffers:         33116 kB
    Cached:        7780188 kB
    SwapCached:         32 kB
    Active:          78716 kB
    Inactive:      7761396 kB
    HighTotal:           0 kB
    HighFree:            0 kB
    LowTotal:      8175080 kB
    LowFree:         39912 kB
    SwapTotal:    16779884 kB
    SwapFree:     16779660 kB
    Dirty:              28 kB
    Writeback:           0 kB
    Mapped:          48356 kB
    Slab:           265028 kB
    CommitLimit:  20867424 kB
    Committed_AS:    61372 kB
    PageTables:       2300 kB
    VmallocTotal: 536870911 kB
    VmallocUsed:    271252 kB
    VmallocChunk: 536599163 kB
    HugePages_Total:     0
    HugePages_Free:      0
    Hugepagesize:     2048 kB
    and
    [oracle@icdb u01]$ cat /etc/sysctl.conf
    # Kernel sysctl configuration file for Red Hat Linux
    # For binary values, 0 is disabled, 1 is enabled.  See sysctl(8) and
    # sysctl.conf(5) for more details.
    # Controls IP packet forwarding
    net.ipv4.ip_forward = 0
    # Controls source route verification
    net.ipv4.conf.default.rp_filter = 1
    # Do not accept source routing
    net.ipv4.conf.default.accept_source_route = 0
    # Controls the System Request debugging functionality of the kernel
    kernel.sysrq = 0
    # Controls whether core dumps will append the PID to the core filename.
    # Useful for debugging multi-threaded applications.
    kernel.core_uses_pid = 1
    kernel.shmall=2097152
    kernel.shmmax=6187593113
    kernel.shmmni = 4096
    kernel.sem = 250 32000 100 128
    fs.file-max =65536
    net.ipv4.ip_local_port_range = 1024 65000
    [oracle@icdb u01]$and bash profile is
    PATH=$PATH:$HOME/bin
    ORACLE_BASE=/u01/app/oracle
    ORACLE_HOME=$ORACLE_BASE/product/9.2.0
    ORACLE_SID=ic
    LD_LIBRARY_PATH=$ORACLE_HOME/lib:$ORACLE_HOME/lib32
    PATH=$PATH:$ORACLE_HOME/bin
    export  ORACLE_BASE ORACLE_HOME ORACLE_SID LD_LIBRARY_PATH PATH
    export PATH
    unset USERNAME
    ~
    ~please suggest me...

    init file
    [oracle@icdb u01]$ cat initicai.ora
    *.aq_tm_processes=1
    *.background_dump_dest='/u01/app/oracle/admin/ic/bdump'
    *.compatible='9.2.0.0.0'
    *.control_files='/bkp/data/ctl/control03.ctl'
    *.core_dump_dest='/u01/app/oracle/admin/ic/cdump'
    *.db_block_size=8192
    *.db_cache_size=4294967296
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='icai'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=icaiXDB)'
    *.fast_start_mttr_target=300
    *.hash_join_enabled=TRUE
    *.instance_name='icai'
    *.java_pool_size=157286400
    *.job_queue_processes=20
    *.large_pool_size=104857600
    *.open_cursors=300
    *.pga_aggregate_target=938860800
    *.processes=1000
    *.query_rewrite_enabled='FALSE'
    *.remote_login_passwordfile='EXCLUSIVE'
    *.shared_pool_size=818103808
    *.sort_area_size=524288
    *.star_transformation_enabled='FALSE'
    *.timed_statistics=TRUE
    *.undo_management='AUTO'
    *.undo_retention=10800
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='/u01/app/oracle/admin/ic/udump'
    #log_archive_dest='/bkp/arch/ic_'
    Log_archive_start=True
    sga_max_size=6444442450944
    log_archive_dest_1='location=/bkp/arch/ mandatory'
    log_archive_dest_2='service=prim optional reopen=15'
    log_archive_dest_state_1=enable
    remote_archive_enable=true
    standby_archive_dest='/arch/ic/ic_'
    standby_file_management=auto
    [oracle@icdb u01]$Edited by: user00726 on Nov 11, 2009 10:27 PM

  • Uploading large files from applet to servlet throws out of memory error

    I have a java applet that needs to upload files from a client machine
    to a web server using a servlet. the problem i am having is that in
    the current scheme, files larger than 17-20MB throw an out of memory
    error. is there any way we can get around this problem? i will post
    the client and server side code for reference.
    Client Side Code:
    import java.io.*;
    import java.net.*;
    // this class is a client that enables transfer of files from client
    // to server. This client connects to a servlet running on the server
    // and transmits the file.
    public class fileTransferClient
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    // this method transfers the prescribed file to the server.
    // if the destination directory is "", it transfers the file to
    "d:\\".
    //11-21-02 Changes : This method now has a new parameter that
    references the item
    //that is being transferred in the import list.
    public static String transferFile(String srcFileName, String
    destFileName,
    String destDir, int itemID)
    if (destDir.equals(""))
    destDir = "E:\\FTP\\incoming\\";
    // get the fully qualified filename and the mere filename.
    String fqfn = srcFileName;
    String fname =
    fqfn.substring(fqfn.lastIndexOf(File.separator)+1);
    try
    //importTable importer = jbInit.getImportTable();
    // create the file to be uploaded and a connection to
    servlet.
    File fileToUpload = new File(fqfn);
    long fileSize = fileToUpload.length();
    // get last mod of this file.
    // The last mod is sent to the servlet as a header.
    long lastMod = fileToUpload.lastModified();
    String strLastMod = String.valueOf(lastMod);
    URL serverURL = new URL(webadminApplet.strServletURL);
    URLConnection serverCon = serverURL.openConnection();
    // a bunch of connection setup related things.
    serverCon.setDoInput(true);
    serverCon.setDoOutput(true);
    // Don't use a cached version of URL connection.
    serverCon.setUseCaches (false);
    serverCon.setDefaultUseCaches (false);
    // set headers and their values.
    serverCon.setRequestProperty("Content-Type",
    "application/octet-stream");
    serverCon.setRequestProperty("Content-Length",
    Long.toString(fileToUpload.length()));
    serverCon.setRequestProperty(FILENAME_HEADER, destDir +
    destFileName);
    serverCon.setRequestProperty(FILELASTMOD_HEADER, strLastMod);
    if (webadminApplet.DEBUG) System.out.println("Connection with
    FTP server established");
    // create file stream and write stream to write file data.
    FileInputStream fis = new FileInputStream(fileToUpload);
    OutputStream os = serverCon.getOutputStream();
    try
    // transfer the file in 4K chunks.
    byte[] buffer = new byte[4096];
    long byteCnt = 0;
    //long percent = 0;
    int newPercent = 0;
    int oldPercent = 0;
    while (true)
    int bytes = fis.read(buffer);
    byteCnt += bytes;
    //11-21-02 :
    //If itemID is greater than -1 this is an import file
    transfer
    //otherwise this is a header graphic file transfer.
    if (itemID > -1)
    newPercent = (int) ((double) byteCnt/ (double)
    fileSize * 100.0);
    int diff = newPercent - oldPercent;
    if (newPercent == 0 || diff >= 20)
    oldPercent = newPercent;
    jbInit.getImportTable().displayFileTransferStatus
    (itemID,
    newPercent);
    if (bytes < 0) break;
    os.write(buffer, 0, bytes);
    os.flush();
    if (webadminApplet.DEBUG) System.out.println("No of bytes
    sent: " + byteCnt);
    finally
    // close related streams.
    os.close();
    fis.close();
    if (webadminApplet.DEBUG) System.out.println("File
    Transmission complete");
    // find out what the servlet has got to say in response.
    BufferedReader reader = new BufferedReader(
    new
    InputStreamReader(serverCon.getInputStream()));
    try
    String line;
    while ((line = reader.readLine()) != null)
    if (webadminApplet.DEBUG) System.out.println(line);
    finally
    // close the reader stream from servlet.
    reader.close();
    } // end of the big try block.
    catch (Exception e)
    System.out.println("Exception during file transfer:\n" + e);
    e.printStackTrace();
    return("FTP failed. See Java Console for Errors.");
    } // end of catch block.
    return("File: " + fname + " successfully transferred.");
    } // end of method transferFile().
    } // end of class fileTransferClient
    Server side code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.util.*;
    import java.net.*;
    // This servlet class acts as an FTP server to enable transfer of
    files
    // from client side.
    public class FtpServerServlet extends HttpServlet
    String ftpDir = "D:\\pub\\FTP\\";
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    public void doGet(HttpServletRequest req, HttpServletResponse resp)
    throws ServletException,
    IOException
    doPost(req, resp);
    public void doPost(HttpServletRequest req, HttpServletResponse
    resp)
    throws ServletException,
    IOException
    // ### for now enable overwrite by default.
    boolean overwrite = true;
    // get the fileName for this transmission.
    String fileName = req.getHeader(FILENAME_HEADER);
    // also get the last mod of this file.
    String strLastMod = req.getHeader(FILELASTMOD_HEADER);
    String message = "Filename: " + fileName + " saved
    successfully.";
    int status = HttpServletResponse.SC_OK;
    System.out.println("fileName from client: " + fileName);
    // if filename is not specified, complain.
    if (fileName == null)
    message = "Filename not specified";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // open the file stream for the file about to be transferred.
    File uploadedFile = new File(fileName);
    // check if file already exists - and overwrite if necessary.
    if (uploadedFile.exists())
    if (overwrite)
    // delete the file.
    uploadedFile.delete();
    // ensure the directory is writable - and a new file may be
    created.
    if (!uploadedFile.createNewFile())
    message = "Unable to create file on server. FTP failed.";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // get the necessary streams for file creation.
    FileOutputStream fos = new FileOutputStream(uploadedFile);
    InputStream is = req.getInputStream();
    try
    // create a buffer. 4K!
    byte[] buffer = new byte[4096];
    // read from input stream and write to file stream.
    int byteCnt = 0;
    while (true)
    int bytes = is.read(buffer);
    if (bytes < 0) break;
    byteCnt += bytes;
    // System.out.println(buffer);
    fos.write(buffer, 0, bytes);
    // flush the stream.
    fos.flush();
    } // end of try block.
    finally
    is.close();
    fos.close();
    // set last mod date for this file.
    uploadedFile.setLastModified((new
    Long(strLastMod)).longValue());
    } // end of finally block.
    } // end - the new file may be created on server.
    } // end - we have a valid filename.
    // set response headers.
    resp.setContentType("text/plain");
    resp.setStatus(status);
    if (status != HttpServletResponse.SC_OK)
    getServletContext().log("ERROR: " + message);
    // get output stream.
    PrintWriter out = resp.getWriter();
    out.println(message);
    } // end of doPost().
    } // end of class FtpServerServlet

    OK - the problem you describe is definitely what's giving you grief.
    The workaround is to use a socket connection and send your own request headers, with the content length filled in. You may have to multi-part mime encode the stream on its way out as well (I'm not about that...).
    You can use the following:
    http://porsche.cis.udel.edu:8080/cis479/lectures/slides-04/slide-02.html
    on your server to get a feel for the format that the request headers need to take.
    - Kevin
    I get the out of Memory Error on the client side. I
    was told that this might be a bug in the URLConnection
    class implementation that basically it wont know the
    content length until all the data has been written to
    the output stream, so it uses an in memory buffer to
    store the data which basically causes memory issues..
    do you think there might be a workaround of any kind..
    or maybe a way that the buffer might be flushed after
    a certain size of file has been uploaded.. ?? do you
    have any ideas?

  • Cannot see images occassionally: "Out of memory"

    Hi,
    I love Lightroom, but have the problem that is really interfering with my work. When I look at a larger collection of images (100+), every now and then the image area will be greyed out, and it will say 'Out Of Memory' in red (upside down and back to front) in the lower left corner.
    This never happens when I am looking at thumbnails in Library-mode. It does happen when I am looking one image in Library-mode, in deverloper mode, and frequently in the Slide-show mode. Interestingly enough: when an image is not visible in Library-mode, chances are it is visible in Developer-mode and Slide-show (and vice-versa).
    I use Lightroom for a lot of subtle retouching, so I can imagine the files are large. I would understand in Slide-show that the CPU simply doesn't have the time to render the image with all the alterations I apply to it -- but in Developer and Libarary-mode I can wait and wait, and the image nevr shows up. Then I go to Developer, and there it is!
    I store my images on a seperate dedicated external harddisk connected through USB 2.0. I have browsed through the forum already, and found a solution in changing the windows pageview setting. I did this, but it made no difference. The problem still occurs with the same frequency. I have not installed LR 1.1 yet; the negative reports on this forum have scared me: I really don't need my database messed up right now.
    Any suggestions you might have will be greatly appreciated!
    Rogier Bos
    Rotterdam, The Netherlands.

    Guys,
    Before I start this rant... I LOVE LIGHTROOM... but no more ADOBE, I cannot stand it anymore... sort out the OOM errors now, before I put my Lightroom CD into the shredder and then post this SHAM on every ADOBE forum I can find.
    I have experienced the out of memory error 11 times today. I am working with 30,000+ images in Lightroom 1.1 on XP, but only looking through 58 files while doing an import of 100 files I get the "OUT OF MEMORY A2" error. This looks suspiciously like another of the, now famous, debug errors used by the developers to track the 'pain' issues. I would challenge Adobe that they know EXACTLY why this happens, a limitation in the programming techniques for their chosen development platform is highlighting issues in the Windows Memory sub-system and they cannot do diddly squat about it... they are just 'attempting' to clean up memory leaks where they can, by putting markers (A2) into the code. It would seem that they are down a creek without the proverbial paddle.
    Tom Hogarty & ADOBE listen up... stabilize the memory issues stop listening to your marketing bods who are probably leaning over your shoulder as I write these words. You are going to suffer at the hands of Aperture and Capture One if you dont get your act together. I have used both of the above with little or NO errors at all.
    Tell us ADOBE why do none of your engineers or developers talk openly about OOM errors, stop silencing them... lets get a discussion going there are some very capable developers out there willing to help for FREE! Are you mad... admit the problems and get people to help... Open Standards could teach you guys a thing or two!
    Remember for everyone of of us that actually has the balls to complain, there are a 1000 who just sit and suffer not knowing what to do!
    Fix the OOM issues before you do anything else this week!! Provide an interim 1.1.x update and make people happy they bought ADOBE and above all else remember at the end of the day these people PAY your wages.
    ...Now where is my Capture One license key...
    Damon Knight
    Photographer
    London

  • Photoshop CS6 has out of memory errors under Mac OS X 10.7.4

    Has anyone who has a 2008 MacBoo Pro with the NVIDIA GeForce 8600M GT 512 MB graphics card tried running Photoshop CS6 under Mac OS X 10.7.4?
    On my MacBook Pro, PS CS6 launches; however, any action results in an out of memory error.  The CS6 versions of After Effects, Premiere Pro and Illustrator all launch and run as expected under 10.7.4.
    If I switch to my Mac OS X 10.6.8 startup volume, Photoshop CS6 runs as expected.
    I have support cases open with Adobe, NVidia and Apple.  Adobe thinks I need to update the display drivers.  NVidia says display drivers must be provided by Apple.  Apple says that the display drivers are current.  The Apple support agent said that he can escalate the case if the issue can be reproduced on more than one computer.
    I'd love to keep my MacBook Pro running another year or so without having to reboot in 10.6.8 every time I need to use Photoshop.  I'd stay with 10.6.8, but of course iCloud requires 10.7.4.
    I have not tried 10.8 yet, but it's on my troubleshooting list.
    Also, PS CS6 runs fine under Mac OS X 10.7.4 on my 2010 iMac.
    Thanks in advance for any feedback.
    - Warren

    It looks like it was indeed the display driver for the NVIDIA GeForce 8600 GT 512 MB graphics card inside my 2008 MacBook Pro.
    It seems that this driver only gets installed if you upgrade from Mac OS X 10.6.7 to 10.7.4 while the startup drive is connected to the MacBook Pro itself.  I had updated while the startup drive was connected to my 2010 iMac. Accordingly, I could have reinstalled Photoshop 100 times without it ever launching as expected.  Having used the external startup drive with three other Apple computers (all 2010 or newer machines) with Photoshop CS6 launching as expected, I had just assumed (mistakenly) that it would work as expected with my 2008 computer.
    So, this was simple enough to resolve by upgrading the OS on the external startup drive while having it connected to the MacBook Pro.   In hindsight, this makes perfect sense.
    -Warren

  • Adobe X Pro 10.1.10 Out of Memory error message

    Hello One, Hello All
    Since updating our Adobe X Pro machines to version update 10.1.10, we occasionally receive an "Out of Memory" error message, which depending on what we are doing, may force us to shutdown all Adobe windows and re-open Adobe, or simply click OK and continue working with the PDF document like nothing is wrong. It seems to have no rhyme or reason to when or why it occurs. I do not see any warnings or errors in the Windows Event Log. It happens with files we have created, and also happens with files received from internal and external sources through email. It is affecting our high end machines: Lenovo X Series and W Series laptops with Windows 8.1 3rd/4th gen i5 CPUs, 8+GB RAM, 128+GB SSD's. We have all system and Windows updates applied. We have Trend Micro and Malwarebytes real-time protection and system scans do not find any malware.
    I have found a few other recent threads on Adobe forum related to this error message but the responses are weak at best with no definitive fix. The system and user temp folders size are less than 100MB each so this cannot be the issue. When the error occurs I check task manager and system utilization, including memory, is well below 100%.
    We did not have this issue prior to version update 10.1.10.
    Really hoping Adobe can step in to help here and hoping boilerplate responses are not used.

    Hi wayne103,
    We released a new security update yesterday that is v10.1.11
    Please update to this version and check the performance.
    I have seen this error message occur while opening a 3rd party created pdf that has corrupted file structure.
    Please let me know if the issue occurs for some specific files or random
    Also let me know the PDF producer for these pdfs.
    Regards,
    Rave

  • Error 500 Out of Memory on x64 bit machine

    Using windows vista 64bit business
    x64 bit machine 32 bit application
    using visual fox pro odbc driver
    Happens on Crystal reports 8 and 10
    After 3+ million records with our query we receive an Error 500 out of memory.
    Under windows 32bit machine, we haven't encountered this issue regardless the size of data.
    Any thoughts?

    Hi Alex
    As mentioned earlier Crystal Reports 9 and 10 are not tested platfrom with Windows Vista.
    Try using Crystal Reports 2008 and check if you are facing any issues.
    Hope this helps.
    Please let us know if you have any further queries.
    Regards
    Ashwini Yadav

  • Out of memory error - JS Runtime: How many users can one connect?

    Not talking video here.  Talking interactive apps, like chat.  Ours crashes at about 500 connected users.  When I report this I'm told "make sure you're not creating too many objects serverside" or "increase the JSRuntimeSize setting in your application.xml file to the max".
    Have now done both of those things but still get this out of memory error.  Let's say I optomized my app and got 100% more connection capacity.  That would be 1,000 connected users - still nowhere near enough.
    Are my dreams of 6,000 or 10,000 connected users enjoying all of the fruits of the FMS interactivity pipe dreams?  Is it not meant for sessions of that size?  Where does one find documentation or advice or application assistance on this issue?
    How do large social media applications connect so many people concurrently.
    Thoughts appreciated.
    Thanks

    Yes.  I'm using the max.
    <RuntimeSize>51200</RuntimeSize>
    See:
    http://help.adobe.com/en_US/FlashMediaServer/3.5_AdminGuide/WS5b3ccc516d4fbf351e63e3d119f2 926bcf-7ff0.html#WS5b3ccc516d4fbf351e63e3d119f2926bcf-7ed2
    Don't think 100MB or 200MB would be valid settings.

  • Out of Memory Runtime Error

    Hello,
    I'm new to Java, this is only my third time using the language, and first time writing an applet. What I'm trying to do is create an applet that will plot 2D a set of coordinates based on an input string. Inexplicably, the VM gives an "<<Out of Memory>>" error while running. I urgently need a solution to this problem (as in, in the next two days... by August 9th, 2001). Any help or suggestions would be greatly appreciated.
    -Mark Radtke, radrik2001<REMOVE>@yahoo.com
    import java.awt.*;
    import java.applet.*;
    import java.io.*;
    import Orbit;
    public class TwoDApplet extends Applet {
         private Orbit stringHolder = new Orbit();
         public void init() {
              System.out.println("execution");
              stringToDraw();
              System.out.println("Done executing.");
         public boolean stringToDraw() {
              Graphics g = getGraphics();
              g.setColor(Color.black);
               StringReader sReader = new StringReader(stringHolder.case1);
               int XYline = 0, pArrayPlace = 0;
               int X = 0, Y = 0, last_x = 0, last_y = 0;
               float temp_x = 0, temp_y = 0;
               final char lineBreaker = '%', separator = ' ';     //These char's set the char that denotes line breaks and spaces between words, '\n' and ' ' by default
               StringBuffer buffer = new StringBuffer();
               final StringBuffer blankBuffer = new StringBuffer(" ");
              char temp;
              boolean flag = false;
              try {
                    temp = (char)sReader.read();
               } catch(IOException e) {
                    System.out.println("\n\n\tERROR: " + e);
                    return false;
         try {
              do {
                   switch(temp) {
                        case lineBreaker:
                             switch(XYline) {
                                  case 0:
                                       last_x = X;
                                       temp_x = Float.parseFloat(buffer.toString());
                                       X = (int)(temp_x * Math.pow(10.0,11.0));
                                       g.drawLine(last_x, 0, X, 0);
                                       break;
                                  case 1:
                                       last_y = Y;
                                       temp_y = Float.parseFloat(buffer.toString());
                                       Y = (int)(temp_y * Math.pow(10.0,11.0));
                                       g.drawLine(last_x, last_y, X, Y);
                                       break;
                                  default: //anything beyond 2 numbers in a row is ignored
                                       g.drawLine(last_x, last_y, X, Y);
                                       break;
                             } //end of nested switch
                             buffer = blankBuffer;
                             XYline = 0;
                             break;
                        case separator: //case separator:
                             switch(XYline) {
                                  case 0:
                                       System.out.print(buffer.toString());
                                       last_x = X;
                                       temp_x = Float.parseFloat(buffer.toString());
                                       X = (int)(temp_x * Math.pow(10.0,11.0));
                                       System.out.print("X = " + X + " ");
                                       break;
                                  case 1:
                                       last_y = Y;
                                       temp_y = Float.parseFloat(buffer.toString());
                                       Y = (int)(temp_y * Math.pow(10.0,11.0));
                                       System.out.print("Y = " + Y + " ");
                                       break;
                                  default:
                                       break;
                             } //end of nested switch
                             buffer = blankBuffer;
                             XYline++;
                             System.out.println("OK\n");
                             break;
                        default:
                             if(buffer.toString().equals(" ")) //i used to test is buffer == new StringBuffer(" "), but I figured that may have been part of the problem... it didn't help
                                  buffer.setCharAt(0, temp);
                             else
                                  buffer.append(temp);
                             break;
                   } //end of switch
                   try {
                         temp = (char)sReader.read();
                    } catch(IOException e) {
                         System.out.println("IOException caught... throwing...");
                         flag = true;
                         throw e;
                    } catch(NullPointerException e) {
                         System.out.println("Caught NullPointerException. You suck."); }
              } while(flag == false); //end of while
         } //end of try
         catch(NullPointerException e) {
              System.out.println("Caught NullPointerException.");
         catch(IOException e) {
              System.out.println("Caught IOException: " + e + "\n");
              return true;
    public class Orbit { //when totally completed, this class will just hold a few strings, most of which are much, much larger than this one
         public String case1 = "3.0745036727705142e-009 3.9417146050976244e-009%4.9852836681565192e-009 3.5573952047714837e-010%3.6200601148208685e-009 3.4445682318680393e-009%8.1006295549636224e-011 4.9953846385005008e-009%3.7105805578461100e-009 3.3440347969092934e-009%4.9772361223912569e-009 4.1037439683639425e-010%3.1322774963314919e-009 3.8884603960830088e-009%";

    Do you need to hold those values in a String? If not then simply store those values in a doubly-dimensioned array and avoid parsing the string altogether (see sample applet below). If you do need to store the values in a String consider using a StringTokenizer for parsing.
    import java.awt.*;
    import java.applet.*;
    import java.io.*;
    public class TwoDApplet extends Applet implements Runnable
    Thread th;
    int x;
    int y;
    int lastX;
    int lastY;
    double coordinates[][] = {{3.0745036727705142e-009,3.9417146050976244e-009},{4.9852836681565192e-009,3.5573952047714837e-010},{3.6200601148208685e-009,3.4445682318680393e-009},{8.1006295549636224e-011,4.9953846385005008e-009},{3.7105805578461100e-009,3.3440347969092934e-009},{4.9772361223912569e-009,4.1037439683639425e-010},{3.1322774963314919e-009,3.8884603960830088e-009}};
    final double operand = Math.pow(10.0,11.0);
    public void start()
      if (th == null)
       th = new Thread(this);
       th.start();
    public void stop()
      if (th != null)
       th = null;
    public synchronized void run()
      for (int z = 0; z < coordinates.length; z++)
       lastX = x;
       lastY = y;
       x = (int)(coordinates[z][0] * operand);
       y = (int)(coordinates[z][1] * operand);
       repaint();
       try
        wait();
        Thread.sleep(100);
       catch(InterruptedException ie)
    public void update(Graphics g)
      paint(g);
    public void paint(Graphics g)
      g.drawLine(lastX, lastY, x, y);
      synchronized(this)
       notifyAll();
    }

Maybe you are looking for

  • Date in numeric format - how to map ?

    Hi, I have a scenario, where i get all the characteristics of an Equipment from R3 to XI. One of the characteristics (installation date) is maintained as YYYYMMDD format in R3 so when it reaches XI it appears as a numeric value(exponential). How can

  • Horribe image quality in PDF

    Hello, my problem is very simple. The quality of the image of my PDF is really horrible. I need to have a PDF file because in order to print a magazine, that's the file that I need. So I converted my picture: http://image.noelshack.com/fichiers/2013/

  • Form fields won't display input information

    We are having a very odd issue that's happened on two occassions. We have a basic PDF form where someone has input information into various fields, but the information doesn't display, unless you click within the field. At that point you can see what

  • Problems synchronising new iPhone 4

    bought an iphone 4, synchronisation doesn't work because it cannot make a backupcopy it starts the synchronisation and starts making the backupcopy, but stucks after 30sec Cannot stop it, ITunes does not respond anymore

  • Error Moving Sidecar JPEGS

    Hi, I am new to lightroom but have watched this: http://tv.adobe.com/watch/getting-started-with-adobe-photoshop-lightroom-3/be-organized/ http://tv.adobe.com/watch/adobe-evangelists-julieanne-kost/moving-files-in-lightroom-and-t he-finder/ I have jus