Reading data files saved using FTP Append

Hi All,
The code posted will write me a data file using FTP Append, which represents the data coming from a cRIO chassis.
However I do not seem to be able to write the code to pull the data back out of the saved files.
I have success in retrieving an isolated value (ie setting the count to 1 on the read binary file.vi), but as soon as I try to retrieve anything any bigger I have some serious issues.  I have checked through the various different tutorials and I don't seem to be doing anything particularly wrong.
Any help would be good ta.
Attachments:
FTP streaming.vi ‏29 KB

OK thanks Dom.
I have a few things which you may consider trying in order to troubleshoot the issue.
In your FTP Strem File Read code, try setting the count input of the read binary file vi to -1 (read entire file) and see what the result is. It may be worth creating a new string indicator rather than an array for this.
Also I noticed you are performing a get file size function in parallel with the read binary vi. It is good practice to keep the flow of file reference and error cluster through these vi's sequentially in order to keep data flow and be sure of the execution order. In this case issues may occur if both vi's try to perform their functions concurrently.
Any chance you could send the file created by the FTP Stream vi so I could take a look?
Thanks.
Paul
http://www.paulharris.engineering

Similar Messages

  • Problem on loading DAT file when using 3G network modem

    Hello,
    I'm having some strange problem when I'm trying to load my game on the part where DAT file with Map Object is read, using 3G Network Modem. This issue started when I migrated my applet from one host to another. Everything loads well, until the user authentication. On user authenticatiom I'm recieving map name which I should load and show the user on it. I'm recieving the message where map name is mentioned, but after, when the process of loading map begins(when I need to read DAT file which is only 15KB) application blocks(browser and JVM also block).
    I thought it could be because of slow network communication, but it's not the reason, since I have tested(using Bandwidth limiter software) with 5 KBs/Second and it loads well and application is running well.
    Any clue why this can happen? If code is needed I'll post some pieces related to the process of map creation.
    URL or application: [ http://mimosa.dei.uc.pt/serhiy/demo/hoonline.html|http://mimosa.dei.uc.pt/serhiy/demo/hoonline.html]
    Accounts: test01/test01 ... test0n/test0n ... test05/test05 (n is number from 1 to 5)
    Thanks in advance!

    You have to upload it with FileReference.upload() to a PHP
    (or other server-side) script which saves it to a folder on the
    server. When the DataEvent.UPLOAD_COMPLETE_DATA event has been
    dispatched you can then use the FileReference.name to load from the
    file on the server just like any other image.

  • Error in Rcvr File Adapter using FTP ?

    Hi @,
    I am getting the follwoing error while trying to put file in the folder thru Rcvr File Adapter using FTP and it is giving me following error :
    java.io.FileNotFoundException :(Permission Denied )
    I have verified the permissions on the folder and it is having write permissions.
    Please suggest Its urgent.
    REgards

    hi amit,
    can you
    1. check whether the access to FTP is Authenticated, if so provide the username and password
    2. try giving different file name at the receiver end
    3. try normal ftp commands to access the folder
    regards
    Ramesh P

  • How to read .dat files that are hidden?

    Is it possible to read .dat files that are set to hidden? 'cause what i did was i created a .dat file and set its attributes to hidden and my Read file code cannot read hidden files.
    here's my Read file code:
    Scanner read=new Scanner(FileReader("C:\password.dat"));
    and another thing, how do you Write a .dat file that is hidden too? 'cause my password.dat file needs to be updated.
    here's my Write file code:
    PrintWriter write=new PrintWriter("C:\password.dat");
    thank for the help.

    There should not be any problems opening up hidden files. Your problem is probably elsewhere (for example, you didn't escape the backslash, so what you posted is not your code).

  • Unable to read PDF files created by FTP Adapter

    Hi,
    I am unable to read pdf files generated by FTP Adapter with its default settings.
    Can anybody guide me how to proceed with it. I guess I need to change something in encoding mechanism in FTP Adapter Outbound Connection Pool, but not sure what has to be.
    Thanks,
    Richa

    Hi all,
    I have got its solution. I was unable to read .pdf files because FTP adapter  was sending the files just by changing the extension from .txt /.xml to .pdf that is the reason that it was not readable.
    We had to write an java code to convert .xml file to .pdf file and then paas it to FTP ADapter.
    Thanks

  • Error when reading data from XML using JAXB?

    Hi all,
    I get the error message when reading data from XML using JAXB.
    [org.xml.sax.SAXParseException: The entity name must immediately follow the '&' in the entity reference.]
         at javax.xml.bind.helpers.AbstractUnmarshallerImpl.createUnmarshalException
    (AbstractUnmarshallerImpl.java:315)
         at com.sun.xml.internal.bind.v2.runtime.unmarshaller.UnmarshallerImpl.createUnmarshalException(UnmarshallerImpl.java:481)
         at com.sun.xml.internal.bind.v2.runtime.unmarshaller.UnmarshallerImpl.unmarshal0(UnmarshallerImpl.java:199)
         at com.sun.xml.internal.bind.v2.runtime.unmarshaller.UnmarshallerImpl.unmarshal(UnmarshallerImpl.java:168)
         at javax.xml.bind.helpers.AbstractUnmarshallerImpl.unmarshal(AbstractUnmarshallerImpl.java:137)
         at javax.xml.bind.helpers.AbstractUnmarshallerImpl.unmarshal(AbstractUnmarshallerImpl.java:142)
    I don't find the reason, please help!
    Thanks in advance!

    The problem is sovled.

  • How to read the whole text file lines using FTP adapter

    Hi all,
    How to read the whole text file lines when error occured middle of the text file reading.
    after it is not reading the remaining lines . how to read the whole text file using FTP adapter
    pls can you help me

    Yes there is you need to use the uniqueMessageSeparator property. Have a look at the following link for its implementation.
    http://download-west.oracle.com/docs/cd/B31017_01/integrate.1013/b28994/adptr_file.htm#CIACDAAC
    cheers
    James

  • Internet toolkit get file date without year using ftp!!!

    Hi!
    I'm using "FTP Directory Listing.vi" to get the modification date of files on a ftp server (PXI-RT): I only get the time, day and month of the last modification but I do not get the year (For example: Aug 3 04:52)!?!
    Have anybody an idea how to get the complete date information (For example: 2005 Aug 3 04:52)?
    Thanks in advance for your help!

    I agree with Corey that you may have issues with the host computer blocking your VI, but the error 15425 seems to fix itself if you change the ftp mode from active to passive. The default is true, but wiring-in a false constant should take care of the error and enable successful file transferring.
    The reasons for this are explained here.
    National Instruments
    RIO Embedded Hardware PSE
    CompactRIO Developers Guide

  • Need to read data files originally created in LabVIEW in C++

    I have originally created binary files in LabVIEW. These files are made by flattening labview data to a string, then converting the string to a U8 byte array which is then saved to a binary data file. A separate file is used to keep track of the variables written to the data file. (FAT) a 2D-String array (first element = variable name, 2nd element = staring byte location in .dat file, 3rd element = byte length of variable) is flattened to U32 and then written to a binary file.
    Opening the files in LabVIEW follows the inverse proceedure. The FAT file is read as a U32 binary data, converted to a byte array, converted to a string, and then unflattened from the string into a LabVIEW 2D-string data type. From this the starting location and length of the variable is extracted. These are used to read the data file, in U8, which is then converted bu a string and then unflattened to LabVIEW data according to the LabVIEW data type.
    I now need to now open these files in a Linux environment using C++. I am at a loss at how to emulate all of this data-type conversion, flattening, and unflattening in C++. Does anyone have any ideas?

    Faraclas wrote:
    > Thanks rolfk, I will do the reading. But just to make sure you don't misunderstand
    > me, everything works beautifully in LabVIEW.
    I gathered that. But reading through the description gave me the
    impression that a few conversions/typecasts may actually be superfluous
    without any change to the actual data.
    > The problem I am having is trying to open up these files in C++. This system has
    > been around for a couple of years now, so I cannot change the data structure in
    > the LabVIEW programs (well I could but I don't want to because it is really a very
    > flexible system that works well).
    Well, it really isn't that difficult, it just won't be a one hour job
    for sure. A few guidelines:
    1) LabVIEW flattened data contains the actual data as is for all fixed
    size elements (scalars and clusters of scalars). Any variable sized data
    (strings, arrays) is prepended with an int32 for each dimension
    containing the number of elements for that dimension.
    2) LabVIEW flattened data is always in Big Endian (Most significant Byte
    first) format. This means for normal C(++) programs developed on an x86
    system you will have to byteswap every single 16 bit and 32 bit integer.
    3) LabVIEW Typecast really is similar to Flattened data but no
    prepending of the number of elements is done for variable sized data.
    And Typecast can't convert any type containing variable sized data, but
    any arbitrary complex datatype only containing fixed size data, or a
    simple array or string can be typecasted.
    Rolf Kalbermatter
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • Jar file downloaded using FTP gets damaged

    Hi people, Am working on a solution that implements an auto update of an application developed using Java. I am using Apache commons package for downloading the updated JAR file from an FTP site. Once I have the new JAR file on my local machine, I double click it and I get the following error. Invalid or corrupt jarfile C:\FIDS\FIDS.jar When I use a different FTP client to download my JAR file, the file works well when double clicked. What could be happening on my JAR file. Thanks and regards.

    tdulce_aboo wrote:
    I know that FTP is not the best solution, however, sometimes we are bound by what the client has in place.I know, it wasn't trying to attack you, it was just a general rant. FTP is so common and when it's the only thing available you obviously have to use it. I just don't understand those who actively enable it on their server and suggest (or even require) its use.
    What is your suggestion,that I use http? or ?That depends on what it's used for.
    My normal one-stop replacement would be SFTP: based on SSH, secure, authenticated, tunnelable, uses a single connection, doesn't do stupid "translations" of the transfered data, ...

  • Problem reading data file (JDK1.3x)

    Here's what I'm trying to do:
    I want to read in data from a text file, parse it into records and fields and insert it into an oracle database.
    Info about the data file:
    I have an excel file that contains rows of data. Some cells can contain data (text) that has multiple lines and line breaks. This text data can contain commas, single quotes and double quotes.
    Using excel, I save this excel file as either a tab delimited or csv file.
    The problem:
    I use the following java code to open the file, and reads it's contents:
    String fileData = "";
    BufferedReader br = new BufferedReader(new FileReader( "filePathAndName" ) );
    StringBuffer contents = new StringBuffer();
    String line = "";
    while( ( line = br.readLine() ) != null )
    {   contents.append( line ); }
    br.close();
    fileData = contents.toString();
    //code to parse data goes here
    Remember, the file is delimited by tabs or commas. A record (an excel row) is delimited by a carriage return. A single value/field can contain multi-line text.
    The problem I'm experiencing is that using the code specified above to read in the data from the file, I lose all tabs, new lines and carriage returns, so my stringTokenizer/regular expression code to parse the data doesn't work correctly. In the case of tab-delimited files, the fields cannot be distinguished. In both file types, the last field of the previous record and the first field of the next record run together.
    It doesn't seem logical that java won't read in tabs/newlines/carriage-returns. Can someone tell me what's wrong? <I confirmed that the tabs/newlines/carriage returns exist in the text file, so it's definately something going wrong during the file reading process>
    Thanks

    Using the FileInputStream, is there anyway to read in multiple characters at a time, like a readline() function? I've looked at various buffered input streams and none of them seemed to have this function.

  • Help needed for file processing using FTP

    Hi All,
    I am new to ODI tool and currently we are implementing one project in ODI. Could you please, any one help me on the following requirement... how to implement it?
    Scenario:
    i) I need to pick a file from remote host 'outbox' dir using FTP process, and place into local 'temp' dir. (I can do it with odiFTPGet tool)
    ii) Change the file as per target business requirement.??? ( this step challenging me)
    for ex:
    if source file name : <project name>_<source ID>_<transaction name>_<unique id>.txt
    target name should be : 1234_<transaction name>_<sysdate in yyyyMMddhhmmss>_<unique id>_2345.txt ( here 1234 and 2345 are hard coded values)
    iii) Later move to local 'temp' to local 'outbox' dir. ( I can do it with odiFileMove tool)
    iv) After successful process move the file from remote 'outbox' to remote 'archive' dir ??? (again this is challenging me).
    For this my questions are:
    1) How to get file name into a variable ( In one of the post , it says we need to create data model and data stores, and interface to get the file names into a file. apart from is there any easy way i can implement it in package it self?)
    2) How to tokenize a long string ( i.e., need to tokenize the source file to get <unique id> and <transaction name> and map to target file name)
    3) How to define implementation for Java in user function?? do we need to use class declarations and import statements?? I need a sample user function in Java.
    4) How to move/rename a file in remote server?? ( need to move file from remote 'outbox' server to remote archive' dir)
    5) If any step in package fails, How to get the error msg into a variable which i can use for sending mail and raise a ticket in remedy?? ( I will pass that variable to as content to mail and remedy ticket)
    6) How to handle list type data (string list) by variables in ODI?? (As we can define single variable not as list.)
    I am sorry to ask all my queries I a single post. But I need to solutions for all to implement this in my project.
    Thanks in advance. Appreciate early response.
    Regards,
    Kiran.N

    Can any one share your thoughts for my request.
    Thanks in advance.
    Regards,
    Kiran

  • Error reading data from Infocube using shell script.

    Dear all ,
    I am facing a problem while reading data from an infocube using a shell script.
    The details are as follows.
    One of the shell script reads the data from the infocube to extract files with the values.
    The tables used for extraction by the shell script are :
    from   SAPR3."/BIC/F&PAR_CUBE.COPA"     FCOPA,
           SAPR3."/BIC/D&PAR_CUBE.COPAU"    COPAU,
           SAPR3."/BIC/D&PAR_CUBE.COPAP"    COPAP,
           SAPR3."/BIC/D&PAR_CUBE.COPA1"    CCPROD,
           SAPR3."/BIC/D&PAR_CUBE.COPA2"    CCCUST,
           SAPR3."/BIC/D&PAR_CUBE.COPA3"    COPA3,
           SAPR3."/BIC/D&PAR_CUBE.COPA4"    COPA4,
           SAPR3."/BIC/D&PAR_CUBE.COPA5"    COPA5,
           SAPR3."/BIC/MCCPROD"      MCCPROD,
           SAPR3."/BIC/SCCPROD"      SCCPROD,
           SAPR3."/BIC/MCCCUSTOM"    MCCCUSTOM,
           SAPR3."/BIC/SCCCUSTOM"    SCCCUSTOM,
           SAPR3."/BIC/SORGUNIT"     SORGUNIT,
           SAPR3."/BIC/SUNIMOYEAR"   SUNIMOYEAR,
    /*     SAPR3."/BI0/SFISCPER"     SFISCPER, */
           SAPR3."/BI0/SREQUID"      SREQUID,
           SAPR3."/BI0/SCURRENCY"    SCURRENCY,
           SAPR3."/BIC/SSCENARIO"    SSCENARIO,
           SAPR3."/BIC/SSOURCE"      SSOURCE
    The problem is that the file generation by this script (after reading the data from teh infocube) is taking an unexpected time of 2 hours which needs to be maximum 10 mins only.
    I used RSRV to get the info about these tables for the infocube:
    Entry '00046174', SID = 37 in SID table is missing in master data table /BIC/MCUSLEVEL2
    Entry '00081450', SID = 38 in SID table is missing in master data table /BIC/MCUSLEVEL2
    and so on for SID = 39  and SID = 35 .
    Checking of SID table /BIC/SCUSLEVEL2 produced errors
    Checking of SID table /BIC/SCUSLEVEL3 produced errors
    Can you please let me know if this can be a reason of delay in file generation (or reading of data from the infocube).
    Also , Please let me know how to proceed with this issue.
    Kindly let me know for more information, if required.
    Thanks in advance for your help.
    -Shalabh

    Hi ,
    In continuation with searching the solution to the problem , I could manage to note a difference in the partition of the Fact table of the infocube.
    Using SE14 -> Storage Parameters, I could find the partition done for the fact table as :
    PARTITION BY: RANGE
    COLUMN_LIST: KEY_ABACOPA
    and subsequently there are partitions with data in it.
    I need to understand the details of these partitions .
    Do they correspond to each requests in the infocube(which may not be possible as there are 13 requests in infocube and much more partitions).
    Most importantly, since this partition is observed for this onfocube only and not for other infocubes, it is possible that it can be a reason for SLOW RETRIEVAL of data from this ionfocube( not sure since the partition is used to help in fast retreival of data from the infocubes).
    Kindly help.
    Thanks for your co-operation in advance.
    -Shalabh

  • Having difficulty reading data file and doing analysis, maybe DIAdem can help

    Hello,
    I am new to DIAdem but I think it may be helpful to my application.
    I used Write Labview Measurement File VI with n-signals input to generate a data file containing 2*n columns(time column and data column iteratively). Now I need to do data analysis from the file. I'd tried to use Read Labview Measurement File VI to load the data but didn't work. I found that it can only read file with one signal as input and without file header. But why?
    What I want to do is to extract specific columns from the file. Some column has special flag values which divide the column into sections. I also need to extract these sections from the column and plot each of them. I know this can be easily done on other data analysis software but is there any way I can do it on Labview? Can DIAdem help with this?
    Thank you

    Hello,
    To answer your last question first, DIAdem can certainly help. You can open the TDM file you created with the Write LabVIEW Measurement File VI with DIAdem. If you navigate to the TDM file in the External Data window of DIAdem NAVIGATOR, you can drag and drop the file into the Data Portal and use the channels of data at will.
    You can also right-click on the file in the External Data window and choose Selective Opening…. There is a Help button in the Selective Opening window which will tell you more about using this feature.
    Either way, once your data is in the Data Portal, you can use the DIAdem ANALYSIS to perform a wide range of analysis on your data. For more on the capabilities of DIAdem, I recommend starting at NI DIAdem.
    Now, in regards to using the LabVIEW to access the data in your TDM file, this is also possible. If you are using LabVIEW 7.1, you will want to look at the Read Channel Groups example. This example uses a sequence of VIs from the Functions>>All Functions>>FileI/O>>Storage palette. The example demonstrates reading multiple channels from a TDM file.
    Regards,
    Eric M

  • Problem in Receiver side File Adapter using FTPS mode

    Hello,
    Here I am facing some problem in receiver side while using FTPS mode in the channel configuration.
    The error message which i could see in the audit log is,
    Message processing failed. Cause: com.sap.aii.af.ra.ms.api.RecoverableException: Error when getting an FTP connection from connection pool: com.sap.aii.af.service.util.concurrent.ResourcePoolException: Unable to create new pooled resource: java.lang.NullPointerException
    Can anyone help me out in solving this issue ASAP?
    I am using Per file transfer mode ni the reciever channel.
    Thanks in advance,
    Yours
    Soorya

    HI Surya,
    First check wether server started or not and then check you are connecting to FTP server by
    go to run -> cmd and write ping and ipaddress which is used and see whether u r getting reponse from teh FTP server.
    Try to login to the ftp server which you have mentioned in CC using the user name and pwd, to chk whether you have the permissions to login to the server.
    Also the check whether the folder you are trying to access is having permission for delete/read/write.
    Restart the FTP server and try it again.
    Regards
    Sridhar Goli

Maybe you are looking for

  • How do I print in Black and White only (greyscale) with my HP OfficeJet 8600 Pro Plus

    Apple OSX Lion, MacBook Pro Product Name HP Officejet Pro 8600 N911g Product Model Number CN579A Product Serial Number {Removed for privacy} Service ID 22112 Firmware Version CLP1CN1150CR Total Page Count 236 Country/Region USA

  • Acrobat 9 Portfolio can't be printed by Reader 9

    I have addressed this issue with Adobe Support and have either received references to KB documents that are totally unrelated, or have been told to just uninstall Reader 9 and install Reader 8, since Reader 8 works ok. In my opinion that answer is un

  • Some files not showing up in iPhoto

    I used the Grand Perspective app on my iPhoto library package to find large video files I could reduce to save space. In the process, I've found some files that are in the package but aren't displayed in iPhoto. If I do a search for the file in iPhot

  • Adobe Reader 8.1.2 wont finish install and wont uninstall

    I downloaded adobe 8.1.2 onto my Toshiba laptop and it fails to open pdf files on my Windows Vista 64 bit version. i removed all i could and Adobe wont uninstall i get an error message 2330. i tried to install Adobe 9 on top but it would only go as f

  • How can i get fallout 2 to work

    How can I get Fallout 2 to work on my mac book pro.  I just got my computer and the graphics are all messed up.  I tryed installing it on a friends computer and it worked.  He has a mac book pro that is only about 8 months old.  Why wont it work on m