File not being written!! really weird! help!!

Hello everyone.
My problem is very strange.
    private void jButton2ActionPerformed(java.awt.event.ActionEvent evt) {
        fc.setFileFilter(new MLPfilter());         // my own custom file filter
        int returnVal = fc.showSaveDialog(this);
        if (returnVal == JFileChooser.APPROVE_OPTION)
            file = fc.getSelectedFile();
        else return;
        try{
            outFile=new PrintWriter(new FileWriter(file.getName()+".mlp"));    // DOESNT SAVE!
            outFile.println(jTextArea1.getText());
        catch (Exception edfgfd){}
        outFile.close();
    }Everything compiles fine. But, when I try to write it wont write!
The problem lies in this line:
outFile=new PrintWriter(new FileWriter(file.getName()+".mlp"));
Whats funny is that if I replace "file.getName()+".mlp" with "file" it works! >.<
But, I want my files to be saved with a ".mlp" extension.
HALP!

For the future why don't you simplify your code by using:
String fileName = file.getName() + ".mlp";
System.out.println(fileName);
outFile=new PrintWriter(new FileWriter(fileName));
This way you can display the actual name of the file and make sure it is what you are expecting. Basic Debugging 101.

Similar Messages

  • Ocrfile is not being written to.  open file issues.  Help please.

    I've been troubleshooting an open file issue on our Test environment for quite a while now. Oracle has had me update to latest CRS bundle for 10.2.0.3, then upgrade to 10.2.0.4, then two more patches via OPatch to bring 10.2.0.4 RAC to it's most recent patch. None of these patches resolved our problem. We have ~8700 datafiles in the database and once the database is started, we're at ~11k on Production but on Test we're at ~37K or higher. It takes 1-2 days to hit the 65536 limit before it crashes. I have to 'bounce' the database to keep it from crashing. Yes, I could raise the ulimit but that isn't solving the problem.
    Over the weekend I noticed that on Production and DEV, the ocrfile is being written to constantly and has a current timestamp but on Test, the ocrfile has not been written to since the last OPatch install. I've checked the crs status via 'cluvfy stage -post crsinst -n all -verbose' and everything comes back as 'passed'. The database is up and running, but the ocrfile is still timestamped at April 14th and open files jump to 37k upon opening the database and continue to grow to the ulimit. Before hitting the limit, I'll have over 5,000 open files for 'hc_<instance>.dat, which is where I've been led down the path of patching Oracle CRS and RDBMS to resolve the 'hc_<instance>.dat bug which was supposed to be resolved in all of the patches I've applied.
    From imon_<instance>.log:
    Health check failed to connect to instance.
    GIM-00090: OS-dependent operation:mmap failed with status: 22
    GIM-00091: OS failure message: Invalid argument
    GIM-00092: OS failure occurred at: sskgmsmr_13
    That info started the patching process but it seems like there's more to it and this is just a result of some other issue. The fact that my ocrfile on Test is not being written to when it updates frequently on Prod and Dev, seems odd.
    We're using OCFS2 as our CFS, updated to most recent version for our kernel (RHEL AS 4 u7 -- 2.6.9-67.0.15.ELsmp for x86_64)
    Any help greatly appreciated.

    Check Bug... on metalink
    if Bug 6931689
    Solve:
    To fix this issue please apply following patch:
    Patch 7298531 CRS MLR#2 ON TOP OF 10.2.0.4 FOR BUGS 6931689 7174111 6912026 7116314
    or
    Patch 7493592 CRS 10.2.0.4 Bundle Patch #2
    Be aware that the fix has to be applied to the 10.2.0.4 database home to fix the problem
    Good Luck

  • PDF files not being displayed correctly, instead I get a blank screen with some sort of small pinned icon in the centre.  It was working fine until today HELP!

    PDF files not being displayed correctly, instead I get a blank screen with some sort of small pinned icon in the centre.  It was working fine until today HELP!

    What is your operating system?  Reader version?  Are these local or online PDFs?  If online, in what browser?
    Can you post a screenshot: https://forums.adobe.com/thread/1070933

  • File corruption on SDCard when multiple files are being written from WinCE 6.0R3

    We currently have file corruption problems which we have been able to reproduce on our system which uses WinCE 6.0R3. We have an SDCard in our system which is mounted as the root FS.  When multiple files are being written to the file system we occasionally
    see file corruption with data destined from one file, ending up in another file, or in another location in the same file.  We have already written test SW that we have been able to use to reproduce the problem, and have worked with the SDCard vendor to
    check that the memory controller on the card is not the source of the problems.
    We know that the data we send to WriteFile() is correct, and that the data which eventually gets sent through the SDCard driver to the SD card is already corrupted.
    We believe that the problem is somewhere in the microsoft private sources between the high level filesystem API calls and the low level device calls that get the data onto the HW.
    We have confirmed that the cards that get corrupted are all good and this is not a case ofpoor quality flash memory in the cards. The same cards that fail under WinCE 6.0R3 never fail under the same types of testing on Windows, Mac OX, or linux.  We
    can hammer the cards with single files writes over and over, but as soon as multiple threads are writing multiple files it is only a matter of time before a corruption occurs.
    One of the big problems is that we are using the sqlcompact DB for storing some data and this DB uses a cache which get's flushed on it's own schedule. Often the DB gets corrupted because other files are being written when the DB decides to flush.
    So we can reproduce the error (with enough time), and we know that data into the windows CE stack of code is good, but it comes out to the SDcard driver corrupted.  We have tried to minimize writes to the file system, but so far we have not found a
    way to make sure only one file can be written at once. Is there a setting or an API call that we can make to force the OS into only allowing one file write at a time, or a way of seeing how the multiple files are managed in the private sources?
    Thanks
    Peter

    All QFE's have been applied we are building the image so we have some control.
    I have build an image which used the debug DLL's of the FATFS and I have enabled all of the DebugZones.  The problem is still happening. From the timings in the debug logs and the timestamps in the data which corrupts the test file I have been able
    to see that the file is corrupted AFTER the write is complete. Or at least that's how it seems.
    We finished writing the file and closed the handle. Then more data is written to other files. When we get around to verifying the file it now contains data from the files that were subsequently written.
    What I think I need to do is figure out in detail how the two files were "laid down" onto the SDCard.  If the system used the same cluster to write the 2 files then that would explain the issue.

  • IO Labels are NOT being written to prefs!

    I've followed all the advice on existing threads about this and it's definitely a serious bug for me. IO Labels are not being written to prefs for me.
    Any ideas? Already tried deleting prefs, creating new blank project and not saving, nothing has worked.

    I found a workaround for anyone having this issue - and this is the ONLY thing that has worked after a week of trying everything on several forums.
    Open Logic, set your labels how you want.
    While Logic is open go to ~/Library/Preferences and delete com.apple.logic.pro.plist
    Quit Logic. I don't think it matters whether you save your project or not. When Logic quits a new plist will be written, and this one WILL have your labels!
    Seems on my machine Logic would not update the IO labels bit of the prefs unless it was writing a complete new prefs file.

  • I ran into the issue of the 32-bit file not being recognized in LR 5.6

    I ran into the issue of the 32-bit file not being recognized in LR 5.6 after it has been saved in PS. I tried every type of file I could save a file as in CC but none of them would be recognized or said it was corrupt.  I had exported 3 files to Merge in HDR Pro Photoshop as 32 bit files.  I then saved in PS and returned to LR 5.6, but it said the bit depth was not supported .  Is this a known issue with LR 5.6 / PS CC?

    Hi,
    this might help: Video Tutorial – 32-bit HDR TIFF files in LR 4.1 « Julieanne Kost's Blog (Only 32 bit TIFF is supported).

  • 3?'s: Message today warning lack of memory when using Word (files in Documents) something about "idisc not working" 2. Message week ago "Files not being backed up to Time Capsule"; 3. When using Mac Mail I'm prompted for password but none work TKS - J

    3 ?'s:
    1  Message today warning lack of memory when using Word (files in Documents) something about "idisc not working"
    2. Message week ago "Files not being backed up to Time Capsule";                                                                                                                                             
    3. When using Mac Mail I'm prompted for password but none work
    Thanks - J

    Thanks Allan for your quick response to my amateur questions.
    Allan:     I'm running version Mac OS X Version 10.6.8     PS Processor is 2.4 GHz Intel core 15 
    Memory  4 gb  1067   MHz  DDr3  TN And @ 1983-2011 Apple Inc.
    I just "Updated Software" as prompted.
    Thanks for helping me!    - John Garrett
    PS.
    Hardware Overview:
      Model Name:          MacBook Pro
      Model Identifier:          MacBookPro6,2
      Processor Name:          Intel Core i5
      Processor Speed:          2.4 GHz
      Number Of Processors:          1
      Total Number Of Cores:          2
      L2 Cache (per core):          256 KB
      L3 Cache:          3 MB
      Memory:          4 GB
      Processor Interconnect Speed:          4.8 GT/s
      Boot ROM Version:          MBP61.0057.B0C
      SMC Version (system):          1.58f17
      Serial Number (system):          W8*****AGU
      Hardware UUID:          *****
      Sudden Motion Sensor:
      State:          Enabled
    <Edited By Host>

  • Log4j log file not being created

    Using websphere for a web app. At first I was getting the error log4j:WARN No appenders could be found for logger....
    So I created the property file and I assume correctly referenced it. The error went away and my logging messages are showing up in the websphere console, but the .log file specified in my log4j.properties file is not being written to... it is only writing to my systemOut.log.
    If I remove the ROOT.File line it still does not create the file (I've done a search on the IBM directory
    #Default log level to ERROR. Other levels are INFO and DEBUG.
    log4j.rootLogger=INFO,ROOT
    log4j.appender.ROOT=org.apache.log4j.RollingFileAppender
    log4j.appender.ROOT.File=c:\myapplication.log
    log4j.appender.ROOT.MaxFileSize=1000KB
    #Keep 5 old files around.
    log4j.appender.ROOT.MaxBackupIndex=5
    log4j.appender.ROOT.layout=org.apache.log4j.PatternLayout
    #Format almost same as WebSphere's common log format.
    log4j.appender.ROOT.layout.ConversionPattern=[%d] %t %c %-5p - %m%n
    #Optionally override log level of individual packages or classes
    log4j.logger.com.webage.ejbs=INFO       
    private static final Logger logger = Logger.getLogger(LoginAction.class);
        public ActionForward execute(ActionMapping mapping, ActionForm form,
                HttpServletRequest request, HttpServletResponse response)
                throws IOException, ServletException {
            initializeLogger();
    private void initializeLogger() {
            org.apache.log4j.BasicConfigurator.configure();
    //trying the above just to get it to work.. because by default this
    //should look in WEB-INF/classes/log4j.properties... I thought
            /*try {
                String log4jUrl = servlet.getServletContext().getInitParameter(
                        "LOG4J_XML");
                if (!(log4jUrl == null || log4jUrl.equals("")))
                    DOMConfigurator.configure(servlet.getServletContext()
                            .getResource(log4jUrl));
            } catch (MalformedURLException e) {
                e.printStackTrace();
            } catch (FactoryConfigurationError e) {
                e.printStackTrace();
        }    Edited by: gmachamer on Nov 30, 2007 6:37 AM

    ok changed to xml file and found a few things out.
    now when I debug though the logger that was created has an empty level... but if I look at the parent logger it is correctly pulling the root logger from my xml (if I change the priority attribute then it changes when debugging the code)
    <log4j:configuration xmlns:log4j="http://jakarta.apache.org/log4j/" debug="true">
         <!-- this appender would be the same as having a System.out -->
         <appender name="console" class="org.apache.log4j.ConsoleAppender">
              <param name="Target" value="System.out"/>
              <layout class="org.apache.log4j.PatternLayout">
                         <param name="ConversionPattern" value="%-5p %c{1} - %m%n"/>
                  </layout>
           </appender>
           <appender name="rollingFileAppender" class="org.apache.log4j.RollingFileAppender">
              <!-- name and location of the file to log to -->
                 <param name="File" value="c:/appLog.log"/>
              <!-- the maximum size the file will be before it rolls the file -->
                 <param name="MaxFileSize" value="1000kb"/>
              <!-- the number of backups you want to maintain -->
              <param name="MaxBackupIndex" value="5"/>
              <!--
                   This is the layout of your messages, you can do alot with this.
                   See the java docs for the class PatternLayout for an explanation of
                   the different values you can have.
              -->
                 <layout class="org.apache.log4j.PatternLayout">
                          <param name="ConversionPattern" value="%t %-5p %c{2} - %m%n"/>
                      </layout>          
              </appender>
           <root>
                  <priority value ="error" />
                  <appender-ref ref="rollingFileAppender" />
                  <appender-ref ref="console" />
           </root> 
    </log4j:configuration>

  • Anybody know what's up with the menu.xml.files not being available for DW after installing on a new comp?

    Anybody know what's up with the menu.xml.files not being available for DW after installing on a new comp?

    Usually that error can be cleared up by renaming the personal config folder. Turn on your OS's hidden files and then go to...
    C: > Users > your username > AppData > Roaming > Dreamweaver (version) > your language > configuration
    Rename the configuration folder configuration-old and start DW up. That should create an entirely new configuration folder and correct the menu.xml file.

  • I bought my iPhone5 4 months ago and now the camera is not working, it only displays black in the screen. I've tried reset, restore for how many times but still the problem is not being resolve. Any help from you Apple team?

    I bought my iPhone5 4 months ago and now the camera is not working, it only displays black in the screen. I've tried reset, restore for how many times but still the problem is not being resolve. Any help from you Apple team?

    There is no Apple here, this is a user to user forum.  When you said you restored, then you restore as NEW?
    If you restored as New and still the problem persiist, then it time to take a trip to the Apple store.

  • Hprof heap dump not being written to specified file

    I am running with the following
    -Xrunhprof:heap=all,format=b,file=/tmp/englog.txt (java 1.2.2_10)
    When I start the appserver, the file /tmp/englog1.txt gets created, but
    when I do a kill -3 pid on the .kjs process, nothing else is being written to
    /tmp/englog1.txt. In the kjs log I do see the "Dumping java heap..." message
    and a core file is generated.
    Any ideas on why I'm not getting anything else written to /tmp/englog1.txt?
    Thanks.

    Hi
    It seems that the option you are using is correct. I may modify it to something like
    java -Xrunhprof:heap=all,format=a,cpu=samples,file=/tmp/englog.txt,doe=n ClassFile
    This seems to work on 1.3.1_02, so may be something specific to the JDK version that
    you are using. Try a later version just to make sure.
    -Manish

  • HELP: Assembly Initialization Error: File not found. Something weird about the used reference

    I moved one project from solution A to solution B using TFS move (with uncommitted changes). I got compile time errors since the references were not found. There were 2 references used. I removed and added them again and compile time errors were fixed. I
    ran a test case and I get file not found exception.
    As per the reference properties, the Path of used reference is C:_src\HIP\Out\Debug\CIP\Bin\amd64\DataPipe\MaterializedEntityManagerTest\Caradigm.Platform.Pipeline.MaterializedEntityManager.dll
    As per the error below, I think it is looking for this, file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager
    Note that it is looking for it in the folder structure -  CIP\Test instead of
    CIP\Bin. This file does not exist in the Test folder. ANyways I decided to copy to the test folder and give it a shot, the folder being... 
    C:_src\HIP\Out\Debug\CIP\test\amd64\DataPipe\MaterializedEntityManagerTest\Caradigm.Platform.Pipeline.MaterializedEntityManager.dll. Unexpectedly, I still get the same error.
    I'm now going crazy with this error. What am i missing???
    Test Name: CdcEnableWithNoTableTest
    Test FullName: Caradigm.Platform.Pipeline.MaterializedEntityManager.Test.DefiningQueryParserTests.CdcEnableWithNoTableTest
    Test Source: c:\_src\HIP\prod\test\DataPipeline\MaterializationTest\EntityViewManagerTest\DefiningQueryParserTests.cs : line 15
    Test Outcome: Failed
    Test Duration: 0:00:00
    Result Message:
    Assembly Initialization method Caradigm.Platform.Pipeline.MaterializedEntityManager.Test.UnitTestInitializer.AssemblyInitialize threw exception. System.IO.FileNotFoundException: System.IO.FileNotFoundException: Could not load file or assembly 'Caradigm.Platform.Pipeline.MaterializedEntityManager, Version=3.0.0.0, Culture=neutral, PublicKeyToken=a03c60abe3a7c396' or one of its dependencies. The system cannot find the file specified.=== Pre-bind state information ===
    LOG: DisplayName = Caradigm.Platform.Pipeline.MaterializedEntityManager, Version=3.0.0.0, Culture=neutral, PublicKeyToken=a03c60abe3a7c396
    (Fully-specified)
    LOG: Appbase = file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager
    LOG: Initial PrivatePath = NULL
    Calling assembly : Caradigm.Platform.Pipeline.MaterializedEntityManager.Test, Version=3.0.0.0, Culture=neutral, PublicKeyToken=a03c60abe3a7c396.
    ===
    LOG: This bind starts in default load context.
    LOG: Using application configuration file: C:\PROGRAM FILES (X86)\MICROSOFT VISUAL STUDIO 12.0\COMMON7\IDE\COMMONEXTENSIONS\MICROSOFT\TESTWINDOW\vstest.executionengine.x86.exe.Config
    LOG: Using host configuration file:
    LOG: Using machine configuration file from C:\Windows\Microsoft.NET\Framework\v4.0.30319\config\machine.config.
    LOG: Post-policy reference: Caradigm.Platform.Pipeline.MaterializedEntityManager, Version=3.0.0.0, Culture=neutral, PublicKeyToken=a03c60abe3a7c396
    LOG: Attempting download of new URL file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager.DLL.
    LOG: Attempting download of new URL file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager.DLL.
    LOG: Attempting download of new URL file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager.EXE.
    LOG: Attempting download of new URL file:///C:/_src/HIP/Out/Debug/CIP/Test/amd64/DataPipe/MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager/Caradigm.Platform.Pipeline.MaterializedEntityManager.EXE.
    . Aborting test execution.
    Result StackTrace: at Caradigm.Platform.Pipeline.MaterializedEntityManager.Test.UnitTestInitializer.AssemblyInitialize(TestContext context)

    Hi,
    Maybe TFS forum is the correct forum for your question, but I'm not sure. I move your question to [where is the forum for...] where the moderators may direct you to the right place.
    Thanks for your understanding.
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Header files  not being included in JNI

    I've written a java program which interfaces with some C code.In the c code i use certain functions which are defined in a header file that i include in the same c file.The problem i have ,is that the compiler shows an error when it comes across one of those functions.It seems to me that the header file is not being included.Here's what i've done at the start of the jni c code,any help will be greatly appreciated.
    #include<stdio.h>
    #include " mnt/cprog/ca.h"
    JNIEXPORT void JNICALL......
    Can other header files be included over here ,or do they have to be included from the java file?..but the problem there is that the header file is of .h extension.....any suggestions??
    ps..The os is Linux
    ..Derick

    If the header files are not found, the compiler will tell you. At least gcc does so. If only some functions you are using are missing, they are probably not in the header file you are including.
    By the way, there is no way to include headers for your native code through Java. What kind of mechanism should that be?

  • GIF files not being cached for toolbar (WEB Forms)

    I realized that the GIF files we are using
    in our toolbars are not being cached by the
    browser. I analyzed the file XLF.LOG and
    the files always have access code "200"
    (not cached). The file REGISTRY.DAT has
    the same problem. Only the JAR file has
    access code "304" (cached). Although these
    GIF files are small (~900 bytes), it takes
    many seconds to open each form when someone
    executes our application using a low speed
    line (many of our forms have toolbars
    with more than 30 icons). Is there any
    special configuration I have to do or this
    will be corrected in future versions?
    XLF.LOG (first and second execution):
    #Version: 1.0
    #Software: Oracle WRB Log Server
    #Fields: clf
    1.0.5.57 - - [24/Jan/2000:07:45:48 -0300] "GET /htm/sales HTTP/1.0" 200 1440
    1.0.5.57 - - [24/Jan/2000:07:45:51 -0300] "GET /htm/ HTTP/1.0" 200 760
    1.0.5.57 - - [24/Jan/2000:07:45:51 -0300] "GET /forms_code/f60all.jar HTTP/1.0" 304 0
    1.0.5.57 - - [24/Jan/2000:07:45:53 -0300] "GET /forms_code/javax/swing/JInternalFrame.class HTTP/1.0" 404 99
    1.0.5.57 - - [24/Jan/2000:07:45:55 -0300] "GET /forms_code/oracle/forms/registry/Registry.dat HTTP/1.0" 200 4122
    1.0.5.57 - - [24/Jan/2000:07:46:06 -0300] "GET /img/PASSWORD.gif HTTP/1.0" 200 853
    1.0.5.57 - - [24/Jan/2000:07:46:06 -0300] "GET /img/HELP.gif HTTP/1.0" 200 898
    1.0.5.57 - - [24/Jan/2000:07:46:06 -0300] "GET /img/KEYS.gif HTTP/1.0" 200 864
    1.0.5.57 - - [24/Jan/2000:07:46:06 -0300] "GET /img/EXIT.gif HTTP/1.0" 200 888
    1.0.5.57 - - [24/Jan/2000:07:47:28 -0300] "GET /htm/sales HTTP/1.0" 304 0
    1.0.5.57 - - [24/Jan/2000:07:47:30 -0300] "GET /htm/ HTTP/1.0" 304 0
    1.0.5.57 - - [24/Jan/2000:07:47:30 -0300] "GET /forms_code/f60all.jar HTTP/1.0" 304 0
    1.0.5.57 - - [24/Jan/2000:07:47:32 -0300] "GET /forms_code/javax/swing/JInternalFrame.class HTTP/1.0" 404 99
    1.0.5.57 - - [24/Jan/2000:07:47:32 -0300] "GET /forms_code/oracle/forms/registry/Registry.dat HTTP/1.0" 200 4122
    1.0.5.57 - - [24/Jan/2000:07:47:45 -0300] "GET /img/PASSWORD.gif HTTP/1.0" 200 853
    1.0.5.57 - - [24/Jan/2000:07:47:45 -0300] "GET /img/HELP.gif HTTP/1.0" 200 898
    1.0.5.57 - - [24/Jan/2000:07:47:45 -0300] "GET /img/KEYS.gif HTTP/1.0" 200 864
    1.0.5.57 - - [24/Jan/2000:07:47:45 -0300] "GET /img/EXIT.gif HTTP/1.0" 200 888
    null

    Have you tried turning on the "File Caching" setting under the "Server" branch for you Listener under OAS?

  • XMP files not being created with DNGs...

    I'm used to using Bridge with ACR to process Nikon NEFs. But a recent camera update has me experimenting with converting the D3's NEFs to DNG files and editing them with CS2 and Bridge 1.04 which works better on my system. I've also experimented with CS3 and Bridge 2.1.1.9 with the same DNG files.
    In both cases XMP files are not being created when working with DNG files. I have enabled "Save image settings in : Sidecar ".xmp" files" turned ON. But they do not seem to be generated and I assume the ACR adjustments are being embedded in the DNG file. Meanwhile doing edits on NEF files continues to create .xmp files as expected.
    I find .xmp files to be very useful and I'm wondering why they're missing with DNG files? Are they hiding somewhere other than in the same folder that the images reside in?
    Thanks.
    Russell

    > The information contained in the XMP file is stored in the DNG file so there is no need for a separate file.
    Aww, yuck!! That's a real drag...
    When editing lots of files (yesterday's shoot produced 8GB) I back everything up to off-line storage and then, after doing all my edits and crops in ACR, I only have to copy over the small XMP files which takes no time at all.
    So if I re-edit the DNG with ACR I have to re-copy ALL the files again to the off-line storage.
    I used to sometimes even save 2 versions of the xmp files when I needed 2 different crops of the same images, such as doing a wide screen 'cinema' crop for a corporate client's Intranet presentation of their event, as well as more standard crops to be used for their newsletters.
    Guess I won't be using DNG any more..
    Thanks for the help Kees :-)
    Russell

Maybe you are looking for

  • IPhoto wont open - have been advised by computer to reinstall

    ...app store says the app is installed, which it is - do I uninstall by moving to trash and trashing? Is there any danger of losing data doing this? once I've done that will I be able to install from the app store? Ed

  • Having problems passing more than one parameter with html:link tag

    Hi guys, for my web application I�m using Struts. I�ve got a database with user details. I would like to get users list and link to the details of each user. I wrote the code and everything is working fine only the users list is repeating as many tim

  • Control Safari browser tabs on laptop

    Hello! Anyone have any ideas how I could quickly change between tabs on my laptop using my iPhone?  I like to walk around the classroom while I teach, but would also like to switch between webpages real quick so I can use AirSketch on my iPad, but al

  • Bessel function best fit

    Hi all, Quite new to LabVIEW. I am capturing diffraction images that are described by Bessel J functions. The maxima of the diffraction correspond to the maxima of the Bessel J functions ( or zeros of 1st derivative).  I wish to fit the diffraction i

  • Proxy error when downloading report (IR) in csv format

    I'm getting the error below when downloading IR report in csv format. I'm getting that whenever the records is more than 108,000. Is there a limit in records to download? Is there a resolution. I am using apex 4.1, EXCEL 2010. Thanks, Pinky