Exit labview (executables) after using large text files

Hello,
I am using LabView 6.0 and his aplication builder / runtime engine. I wrote some VI`s to convert large Tab delimited textfiles (up to 50 mb). When I am finished with the file it is staying in the memory somehow and is staggered with other (text)files in such a way the computer is slowing down.
When I want to exit the VI (program) it will take a very long time to get  lost of the program (resetting LabView) and get my speed back.
How kan I solve this problem for these large files?
Martin.

OK, this may be a bit of a problem to track down, but let's start.
First, while your front panel looks great, your code is very hard to read. Overlapping elements, multiple nested structures and a liberal use of locals make this a problem. My first suggestion would be to with a massive cleanup operation. Make more room, make wires straight, make sure things are going left-to-right, make subVIs, place some documentation and so on. You won't believe the difference this makes.
After you did that, we can turn to find the problems. Some likely suspects are the local variables and the array functions. You use many local variables and perform resizing operations which are certain to generate copies. If you do this on arrays with dozens of MBs of data, this looks like the most likely source of the problem. Some suggestions to deal with this - if you have repeating code, make subVIs or move the code outside of the structures, so that it only has to be run once. Also, you seem to have some redundant code. For instance, you open  the file only to see if you get an error. You should be able to do this with the VIs in the advanced palette without opening it (and you won't need to close it, either). Another example - you check the exit conditions in many places in your code. If your loop runs fast enough, there is no need for it. Some more suggestions - use shift registers instead of locals and avoid setting the same properties over and over again in the loop.
After you do these, it will probably be much easier to find the problem.
To learn more about LabVIEW, I suggest you try searching this site and google for LabVIEW tutorials. Here and here are a couple you can start with. You can also contact your local NI office and join one of their courses.
In addition, I suggest you read the LabVIEW style guide and the LabVIEW user manual (Help>>Search the LabVIEW Bookshelf).
And one last thing - having the VI run automatically and then use the Quit VI at the end is not very nice. Since you are building it, it will run automatically on its own and you can use the Application>>Kind property to quit only if it's an executable.
Try to take over the world!

Similar Messages

  • Arbitrary waveform generation from large text file

    Hello,
    I'm trying to use a PXI 6733 card hooked up to a BNC 2110 in a PXI 1031-DC chassis to output arbitrary waveforms at a sample rate of 100kS/s.  The types of waveforms I want to generate are generally going to be sine waves of frequencies less than 10 kHz, but they need to be very high quality signals, hence the high sample rate.  Eventually, we would like to go up to as high as 200 kS/s, but for right now we just want to get it to work at the lower rate. 
    Someone in the department has already created for me large text files > 1GB  with (9) columns of numbers representing the output voltages for the channels(there will be 6 channels outputting sine waves, 3 other channels with a periodic DC voltage.   The reason for the large file is that we want a continuous signal for around 30 minutes to allow for equipment testing and configuration while the signals are being generated. 
    I'm supposed to use this file to generate the output voltages on the 6733 card, but I keep getting numerous errors and I've been unable to get something that works. The code, as written, currently generates an error code 200290 immediately after the buffered data is output from the card.  Nothing ever seems to get enqued or dequed, and although I've read the Labview help on buffers, I'm still very confused about their operation so I'm not even sure if the buffer is working properly.  I was hoping some of you could look at my code, and give me some suggestions(or sample code too!) for the best way to achieve this goal.
    Thanks a lot,
    Chris(new Labview user)

    Chris:
    For context, I've pasted in the "explain error" output from LabVIEW to refer to while we work on this. More after the code...
    Error -200290 occurred at an unidentified location
    Possible reason(s):
    The generation has stopped to prevent the regeneration of old samples. Your application was unable to write samples to the background buffer fast enough to prevent old samples from being regenerated.
    To avoid this error, you can do any of the following:
    1. Increase the size of the background buffer by configuring the buffer.
    2. Increase the number of samples you write each time you invoke a write operation.
    3. Write samples more often.
    4. Reduce the sample rate.
    5. Change the data transfer mechanism from interrupts to DMA if your device supports DMA.
    6. Reduce the number of applications your computer is executing concurrently.
    In addition, if you do not need to write every sample that is generated, you can configure the regeneration mode to allow regeneration, and then use the Position and Offset attributes to write the desired samples.
    By default, the analog output on the device does what is called regeneration. Basically, if we're outputting a repeating waveform, we can simply fill the buffer once and the DAQ device will reuse the samples, reducing load on the system. What appears to be happening is that the VI can't read samples out from the file fast enough to keep up with the DAQ card. The DAQ card is set to NOT allow regeneration, so once it empties the buffer, it stops the task since there aren't any new samples available yet.
    If we go through the options, we have a few things we can try:
    1. Increase background buffer size.
    I don't think this is the best option. Our issue is with filling the buffer, and this requires more advanced configuration.
    2. Increase the number of samples written.
    This may be a better option. If we increase how many samples we commit to the buffer, we can increase the minimum time between writes in the consumer loop.
    3. Write samples more often.
    This probably isn't as feasible. If anything, you should probably have a short "Wait" function in the consumer loop where the DAQmx write is occurring, just to regulate loop timing and give the CPU some breathing space.
    4. Reduce the sample rate.
    Definitely not a feasible option for your application, so we'll just skip that one.
    5. Use DMA instead of interrupts.
    I'm 99.99999999% sure you're already using DMA, so we'll skip this one also.
    6. Reduce the number of concurrent apps on the PC.
    This is to make sure that the CPU time required to maintain good loop rates isn't being taken by, say, an antivirus scanner or something. Generally, if you don't have anything major running other than LabVIEW, you should be fine.
    I think our best bet is to increase the "Samples to Write" quantity (to increase the minimum loop period), and possibly to delay the DAQmx Start Task and consumer loop until the producer loop has had a chance to build the queue up a little. That should reduce the chance that the DAQmx task will empty the system buffer and ensure that we can prime the queue with a large quantity of samples. The consumer loop will wait for elements to become available in the queue, so I have a feeling that the file read may be what is slowing the program down. Once the queue empties, we'll see the DAQmx error surface again. The only real solution is to load the file to memory farther ahead of time.
    Hope that helps!
    Caleb Harris
    National Instruments | Mechanical Engineer | http://www.ni.com/support

  • Sorting of a large text file

    I am rewriting a UNIX script to Java. The UNIX script has syncsort code contained in it. The sort is used to sort large text files. The files consist of very large records, up to 3000 characters. I have the file layouts for the fields within the records. Below are a couple syncsorts showing what I need to re-write.
    The first sort, sorts the file by transaction type and only keeps certain transaction types.
    $SYNCSORT << ____endsort1
    /WARNING 1 /SKIPBLANK
    /INFILE $XJOB/tran.ACHP
    /OUTFILE $XJOB/tran.ACHP.careabout
    /FIELDS tran 1 CHAR 3, seq 4 CHAR 2, check_field 9 CHAR 1
    /CONDITION CARE_ABOUT ((tran="104" OR tran="154" OR tran="165" OR tran="301" OR tran="385") AND (seq="00" OR seq="01")) OR (tran="104" AND check_field !="$")
    /COPY
    /INCLUDE CARE_ABOUT
    /END
    endsort1
    The second sort. sorts by plan number, participant number in ascending order and summarizes the amount. I can summarize the amount after the sort in Java so that does not need to be done in the same sort step.
    $SYNCSORT << ____endsort2
    /WARNING 1 /SKIPBLANK
    /INFILE $XJOB/extract.ACHP
    /FIELDS plan 1 CHAR 6, amount 8 TS 15, partid 23 CHAR 9
    /KEYS plan, partid ASCENDING
    /SUMMARIZE TOTAL amount
    /OUTFILE $XJOB/extract.ACHP.summary
    /END
    endsort2
    What is the best way to read in the file, store the data, i.e. vector or array, and how would I go about sorting a large file with very long records?
    Much thanks in advance!!!

    try looking at the comparable interface and the collections.sort() methods. maybe they will save u a little work. as for using arrays or vectors i would use arraylists/vectors so i could use the collections methods. and for the rest of it, it will depend more on your exact needs and i guess you'll just have to go about and make your own configuration systems etc.. shouldnt be too tuff.

  • Importing schema+ data using a text file containing sql commands

    I have the 2012 SQL SVR Mgmt Studio installed. I received a file--snippet below... from an export of our current DB provider, Advantage. I know I have to edit a lot of commands, but my question is quite simple--how do I use a text file with commands in it
    like below to create tables, columns and load data into SQL SVR.
    ---data.txt file contents----
    -- Table Type of Beneficiary is ADT
    Create Table Beneficiary(
       PlanId Char( 9 ),
       Hash Integer,
       Line1 Char( 128 ),
       Line2 Char( 128 ),
       Batch Char( 16 ) );
    EXECUTE PROCEDURE sp_CreateIndex90( 'Beneficiary', 'LAC_LACV1_Beneficiary.adi', 'KEY1', 'Hash;PlanId', '', 2051, 4096, NULL );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 983, 'Judith Pursifull~Spouse~100~~~', 'Michael Pursifull~Son~50~Thomas Pursifull~Son~50', '761011042' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 996, 'Anna Brownlow~Spouse~100~~~', 'Kaitlin Brownlow~Daughter~85~Lawanda Brownlow~Mother~15', '49036615' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 1005, 'Weldon Shelton~Other~50~Star Shelton~Other~50', 'Danica Shelton~Sister~50~Ryan Shelton~brother~50', '109075816' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 1031, 'Donald D Duffy~Spouse~100~~~', 'Brandon M Duffy~Son~100~~~', '219979254' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 1063, 'Lynne Roffino~Other~50~John Roffino~Other~50', 'Katy Roffino~Sister~50~Margaret Roffino~Sister~50', '604232358' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 1062, 'John Teaven Redstone~Spouse~100~~~', '~~~~~', '482024691' );
    INSERT INTO "Beneficiary" VALUES( 'LACV1', 1032, 'Judith Anne Brown (for cat care)~Other~50~Judith Ann (Kappler) Barklage~Other~50', 'LaVerne Cocke (for cat care)~Friend-PantegoBibl~50~~~', '358324107' );
    -- Table Type of Date is ADT
    Create Table Date(
       Hash Integer,
       PlanId Char( 9 ),
       Type Char( 24 ),
       Date Date,
       Override Logical,
       Batch Char( 32 ) );
    EXECUTE PROCEDURE sp_CreateIndex90( 'Date', 'LAC_LACV1_Date.adi', 'KEY1', 'Hash;Date;PlanId;Type', '', 2051, 4096, NULL );
    INSERT INTO "Date" VALUES( 1018, 'LACV1', 'HARDSHIPEND', '2010-02-20', False, '20090820-9414719' );
    INSERT INTO "Date" VALUES( 1018, 'LACV1', 'HARDSHIPSTART', '2009-08-20', False, '20090820-9414719' );
    INSERT INTO "Date" VALUES( 1001, 'LACV1', 'HARDSHIPEND', '2010-02-06', False, '20090806-9371968' );
    INSERT INTO "Date" VALUES( 1001, 'LACV1', 'HARDSHIPSTART', '2009-08-06', False, '20090806-9371968' );
    INSERT INTO "Date" VALUES( 1022, 'LACV1', 'LUMPSUMDISTRIBUTION', '2009-07-21', False, '20090721-9337640' );
    charles.leggette

    from an export of our current DB provider, Advantage.....
    -- Table Type of Date is ADT
    Create Table Date(
       Hash Integer,
       PlanId Char( 9 ),
       Type Char( 24 ),
       Date Date,
       Override Logical,
       Batch Char( 32 ) );
    Hello Charles,
    The SQL Syntax especially for DDL commands are different between MS SQL Server and Advantage Database, so you have to modify them to get them work.
    As David already wrote we don't have a stored procedure "sp_CreateIndex90" in SQL Server, you may remove these commands, and we also don't have a data type "Logical", next could be "bit"
    Olaf Helper
    [ Blog] [ Xing] [ MVP]

  • Problems with a waking up a labview executable after computer hibernation?

    I am just wondering if anyone has had any problems with waking up a labview executable after hibernating a computer.  The executable was running before the computer was hibernated and it seems to have frozen not crashed after the computer was woken up.  I am using Windows 7 and Labview 2009.
    Thanks for your time.

    What is the application doing?  (if that even matters)
    If it were performing a data-acquisition hardward process then that might matter..
    I haven't experienced this problem.

  • Editing and changing large text file

    hi,
    new to this, so bare with me.
    got a large text file 44meg and i need to change some values in it.
    example:
    TSX ;20030102;40302216;40300579;1980;1900;3762000
    i need to change the lines so that they read:
    TSX ;20030102;302216;300579;1980;1900;3762000
    thus removing the leading 40 in the middle cols.
    Thanks in advance
    john

    crap, small mistake
    1) use BufferedReader to read in the file line by line (BufferedReader.readLine())
    2a) for each line, split it on the semicolons (String.split())
    2b) change the middle value using String.substring()
    2c) construct a new line by appending all strings in the array returned by 2a) to eachother
    2d) write this new line to a file using PrintStream (PrintSteam.println())
    3) when done, close both the reader and the printstream.

  • When using large text my lines overpal, how can I increase the line spacing?

    When using large text my lines overpal, how can I increase the line spacing?

    Press 'Ctrl + Enter' after the numbered item to insert blank space below it. If you press 'Enter', the next line will start with the next number. As long as you want to type lines without a numbering, press 'Ctrl + Enter' and type the line.

  • Problems using larger text

    I have a few machines using Appleworks and trying to use larger text. When they are typing text with a 48 or 72 pt font the cursor is behind 3 letters. Is there any way to fix this?

    AppleWorks, being an OS 9 application carbonized to run in OS X, doesn't take advantage of a lot of OS X features, one being the display of text. This is not something that got fixed with the 6.2.9 update. Usually turning Fractional Character Widths on in Preferences > General… > Topic: Text will fix the problem. It is not at all unusual for the problem to appear on one Mac & not another, even though they are running the exact same system & AppleWorks version. It even can vary document to document on the same Mac.

  • "Damaged" fonts in pdf file after using TouchUp Text Tool

    Hello,
    I have a question regarding an issue with lost fonts in a pdf file.
    The document was created in Adobe InDesign and printed to pdf.
    However, after using the "TouchUp Text Tool" in order to check the font for a certain text, several attributes of the font have been damaged in the entire pdf: all characters in "small caps" have been transformed to regular.
    We've had 3 cases so far.
    Does anyone know how we can prevent this from happening again?
    When exporting the document to pdf, all fonts were embedded, I am not sure if this can be the cause.
    Thank you in advance for any suggestions!
    Paula

    We are using Adobe 8 Professional.
    Bill@VT, I am not sure what you are asking me
    I have checked the Preferences for the pdf file and the "Use system fonts" case is checked.
    When opening the pdf file, the fonts were ok, but somewhere along the process, the small caps were lost.
    Maybe a print screen will be of use: the presentation above is the correct presentation (the one immediately after exporting to pdf); the presentation below is the one after all fonts were damaged.

  • Execute code from a text file ...

    Hello ,  Would like to know whether it is possible to have my ABAP code in a text file in desktop and then execute the code by reading it into an ITAB in another ABAP program .
    Note : Pl. dont suggest how to upload and create a new program . I dont want to create a new program after uploading into ITAB using INSERT or function modules . I want to execute the code as part of the uploading program only . So that once the execution is complete the uploaded code will not be available in the SAP system.
    Thanks in advance ,
    Jee.R

    >
    jeeva R wrote:
    > Note : Pl. dont suggest how to upload and create a new program . I dont want to create a new program after uploading into ITAB using INSERT or function modules . I want to execute the code as part of the uploading program only . So that once the execution is complete the uploaded code will not be available in the SAP system.
    Oh dear.  The only solution to your requirement is, sadly, to upload and create a new program, and then one further step.  There is no other way.  So it looks like you'll have to do without.
    matt

  • BADI or User Exit that executes after BADI ORDER_SAVE

    Hi.  I am trying to find a BADI or user-exit that occurs after the BADI ORDER_SAVE has executed completely and saved the values to database.  This is because ORDER_SAVE only gives the item GUID not the header GUID and it appears the header data is only created when the save has taken place.  I am trying to use the function CRM_ORDER_MAINTAIN but I need the header data to manipulate it.
    Thanx in advance
    Martin

    I will tell you four methods to find a BADI,
    Method 1. If you think that it is a simple requirement then got SE18 and click on F4 and Iformation Systems. In the short description enter the short text and search.
    Method2. Another simple method is to search with Package.
    Method3. Put a break point in the FM SXV_GET_CLIF_BY_NAME and then run your transaction---> your requirement may satisfy here after when you press the SAVE Button. There are two more BADIs that will be triggered in general.
    a) Before Update
    b) In Update.
    Method4. Put a break point in the class CL_EXITHANDLER for METHOD get_instance. and then run your transaction.
    Hope this material helps.
    Bhargav.

  • Loading large text files into java vectors and outof memmory

    Hi there,
    need your help for the following:
    i'm trying to load large ammoubnts of data into a Vector in order to concatenate several text files and treat them, but i'm getting outofmemory error. I even tried using xml structure and saving to database but the error is still the same. Can you help?
    thanks
    here's the code:
    public void Concatenate() {
    try {
    //for(int i=0;i<1;i++) {
    vEntries =  new Vector();
    for(int i=0;i<BopFiles.length;i++) {
    MainPanel.WriteLog("reading file " + BopFiles[i] + "...");
    FileInputStream fis = new FileInputStream(BopFiles);
    BufferedInputStream bis = new BufferedInputStream(fis);
    DataInputStream in = new DataInputStream(bis);
    String line = in.readLine();
    Database db = new Database();
    Connection conn = db.open();
    while(line != null) {
    DivideLine(BopFiles[i], line);
    line = in.readLine();
    FreeMemory(db, conn);
    MainPanel.WriteLog("Num of elements: " + root.getChildNodes().getLength());
    MainPanel.WriteLog("Done!");
    } catch (Exception e) {
    e.printStackTrace();
    public void DivideLine(String file, String line) {
         if (line.toLowerCase().startsWith("00694")) {
              Header hd = new Header();
              hd.headerFile = file;
              hd.headerLine = line;
              vHeaders.add(hd);
         } else if (line.toLowerCase().startsWith("10694")) {
              Line entry = new Line();
              Vector vString = new Vector();
              Vector vType = new Vector();
              Vector vValue = new Vector();
              entry.name = line.substring(45, 150).trim();
              entry.number = line.substring(30, 45).trim();
              entry.nif = line.substring(213, 222).trim();
              entry.index=BopIndex;
              entry.message=line;
              entry.file=file;
              String series = line.substring(252);
              StringTokenizer st = new StringTokenizer(series, "A");
              while (st.hasMoreTokens()) {
                   String token=st.nextToken();
                   if(!token.startsWith(" ")) {
                        vString.add(token);
                        vType.add(token.substring(2,4));
                        vValue.add(token.substring(4));
                   token=null;
              entry.strings= new String[vString.size()];
              vString.copyInto(entry.strings);
              entry.types= new String[vType.size()];
              vType.copyInto(entry.types);
              entry.values= new String[vType.size()];
              vValue.copyInto(entry.values);
              vEntries.add(entry);
              entry=null;
              vString=null;
              vType=null;
              vValue=null;
              st=null;
              series=null;
              line=null;
              file=null;
              MainPanel.SetCount(BopIndex);
              BopIndex ++;
    public void FreeMemory(Database db, Connection conn) {
    try {
    //db.update("CREATE TABLE entries (message VARCHAR(1000))");
                   db.update("DELETE FROM entries;");
                   PreparedStatement ps = null;
                   for( int i=0; i<vEntries.size(); i++ ) {
                        Line entry = (Line) vEntries.get(i);
                        String value = "" + entry.message;
                        if(!value.equals("")) {
                             try {
                                  ps = conn.prepareStatement("INSERT INTO entries (message) VALUES('" + Tools.RemoveSingleQuote(value) + "');");
                                  ps.execute();
                             } catch(Exception e) {
                                  e.printStackTrace();
                                  System.out.println("error in number->" + i);
                   MainPanel.WriteLog("Releasing memory...");
         vEntries = null;
         vEntries = new Vector();
         System.gc();
              } catch (Exception e1) {
                   e1.printStackTrace();

    Well, i need to treat those contents, and calculate values withing those files, so wrinting files using FileInputstream wont do. for instance i need to get line 5 from file 1, split it, grab a value according to its class (value also taken) and compare it with another line of another file, adding those values to asingle file.
    that's why i need vector capabilities, but since these files have more than 5 Mb each, an out of memory error is returned by loading those values into vector.
    A better explanaition:
    Each file has a line like
    CLIENTNUM CLASS VALUE
    so if the client is the same withing 2 files, i need to sum the lines into a single file.
    If class is the same, then sum values, if not add it to the front.
    we could have a final line like
    CLIENTNUM CLASS1 VALUE1 CLASS2 VALUE2

  • Creating large text files

    What is the fastest as well as less memory intensive way to create large files.
    Current State
    I have an application where I am reading the database and processing the information (formatting it) and using some third party API to add the information and then finally save as text file. The problem is since its third party API and we do not have any control on that and it takes very long time to generate the files
    Future state
    I want to build the file generator which will read from the database and then process/format the information one by one. Now I have following options.
    1) Add the processed information in a StringBuffer line by line and at the end of it create a file from the StringBuffer and save it.
    2) Create a custom object with different ArrayLists and keep on adding the processed lines into appropriate lists and at the end of it while saving it to a file read the custom object and save it as a file.
    3) Create a file at the start of it and then keep on adding and flushing the lines one by one and at the end of it close the file.
    For handling files I was thinking of using PrintWriter. I am talking about the text files anyware from 50 KB to 20 MB.
    I have performance concerns as well as memory issues. So I want a balanced solution so that I am able to handle both.

    Use a BufferedWriter to write each line/entry/record as you process it.
    Don't do any special flushing() (unless you need special transactional properties in which case you need a lot more than simple flush() calls).

  • Labview crashes when creating large image files

    I have a problem with Labview 6.0.2( I've tested evaluation version 7.0 too).
    I'm constructing a very large image, for example: 4500x4500 pixels. Labview crashes when converting the pixture to a pixmap. The image is fully constructed on my screen (in a picture control), but when converting it to a pixmap (for saving the image in a known format (bmp, jpg, tiff)), Labview crashes.
    I did some testing and when the number of pixels exceeded the limit of 2^24(16777216), the file 'image.cpp' crashes on line 1570. The vi to convert it to a pixmap is: P'icture to pixmap.vi'
    Does someone know a workaround for this problem? Or is there a fix for it?
    Thank you!

    I've tested the 6i version of my VI in Labview 7.0 evalutation version. It raised an error but not the same error:
    d:\lvworm\src\lvsource\compatexport.cpp(37) : DAbort: Called a routine not in the compatibility LVRT table
    $Id: //labview/branches/Wormhole/dev/lvsource/compatexport.cpp#11 $
    0x004BD4CB - LabVIEW_Eval + 0
    0x0EB710D9 - lvs248 + 0
    0x094C87A0 - + 0
    So i replaced the picture VI's with the 7.0 evalutation version VI's, and it worked. It is now possible for me to construct very large image files!
    I see no attached VI to test. But i guess it is also solved in Labview 7.0
    I used this file to convert the picture to image data:
    C:\Program Files\National Instruments\LabVIEW 7.0 Evaluation\vi.lib
    \picture\pictutil.llb\Picture to Pixmap.vi
    And this file to convert image data to bmp:
    C:\Program Files\National Instruments\LabVIEW 7.0 Evaluation\vi.lib\picture\bmp.llb\Write BMP File.vi
    I guess i have to write a workaround for this problem:
    divide the picture in blocks of 4096 x 4096 and then merge the image data arrays of the bloks together.

  • Executing commands read from text file

    I'm trying to execute full command lines that I am reading in from a text file. For example, each line in the file contains a full call to a method that I have already written, including arguments. I have opened the text file, can read in line by line and store in an array, but how do I go about executing the commands?
    For example, if the first slot in my array contains the string "myMethod(50, 100);" how can I effectively get that to execute, followed by the next comand, etc.?
    I think I'm overshooting this, because it really shouldn't be difficult.
    Help would be reciprocated with undying gratitude and perhaps the sacrifice of a pot nellied pig in your name.
    D.

    this article seems like it might be helpful to you.
    http://www.javaworld.com/javaworld/jw-03-2005/jw-0314-
    scripting.htmlSlappy would be proud!
    But before i am scolded, i will promise not to bring it up
    again.
    w

Maybe you are looking for

  • "How do I defrag my mac?" "You don't need to, it does it itself" Fragmented Free Space: 99%. READ THIS NOW

    Just a heads up for my fellow Mac users. "How do I defrag my mac?" "You don't need to, ever. It does it automatically." "But it's running really slowly..." "Repair permissions or something. Defragging is not necessary on a mac." Techtool pro is now d

  • Sponsor Portal after upgrade ISE 1.2 - 1.3

    Hi, After upgrade ISE to version 1.3 I can't access to Sponsor Portal via ://ISE_IP:8443/sponsorportal/ as it was done in version 1.2 (error: [ 404 ] Sponsor Portal Resource Not Found. The resource requested cannot be found). I have to open it throug

  • 2008 Service Pack to Fix Enterprise Login

    Post Author: CalvinMiranda CA Forum: Crystal Reports Any idea when this Service Pack is due out in 2008?

  • Today option in adding attachments

    Hi all, I'm very disappointed that the Today is no longer an option in the pull-down list of folders when you are trying to attach a file (e.g. in an email message) in Mountain Lion.  I have tried using All My Files, but this invariably hangs, so it'

  • I forgotten password of client 000/ddic

    Hello All,             I am working in sap ecc5.0 and I am working in client :810, user:sapuser and now I tried to open client 000 with user: sap* , p/w:06071992. and user:ddic, p/w:19920706, and I tried with the user:sap* , p/w:pass…but it is not op