Process and Streams

I'm doing a program that will use psftp-program to transfer files to server. Part of the code is attached below.
Process yhteys = aja.exec("c:\\temp\\psftp [email protected] -bc -b tuo.src");
InputStream input = yhteys.getInputStream();
OutputStreamWriter out = new OutputStreamWriter( new BufferedOutputStream(yhteys.getOutputStream()) );
char merkki;
while ((merkki = (char)input.read()) != ':') {
System.out.print(merkki);
System.out.print(merkki);
out.write("salasana\n");
out.flush();
while ((merkki = (char)input.read()) != ':') {
System.out.print(merkki);
System.out.print(merkki);
input.close();
out.close();
yhteys.getErrorStream().close();
yhteys.destroy();This prints the following:
[email protected]'s password:
[email protected]'s password:So the input seems to work but out doesn't. Psftp doesn't seem to receive the password, because it keeps asking it after it should have been received.
How this data transfer in and out of process should be done right?

try using the ftp url in the following form:
nn:[email protected] ftp clients can correctly parse the format user:password@address url, since it is a standard. Anyway, if you want to dig deeper into Java, try to find a FTP library that can suit your needs. In that way, you won't need a 3rd party ftp client.

Similar Messages

  • Spawned Processes and Memory

    Hi Guys,
    Does the JVM limit the amount of memory that a process uses when spawning using Runtime.exec()
    Basically we are trying to run the microsoft linker from Java (via Maven) and it keeps erroring with 'out of memory' this seems to happen when the process uses around 275MB of RAM.
    Are processes spawned by Java limited to the amount of memory allocated to the JVM via -Xmx ?
    Any insight on this matter would be appreciated.

    Check the source code of the function
    Java_java_lang_Win32Process_create
    in j2se\src\windows\native\java\lang\Win32Process_md.c (you can get such
    source code downloading the SCSL Source code.
    The relevant information is posted below.
        si.dwFlags = STARTF_USESTDHANDLES;
        processFlag = CREATE_NO_WINDOW;
        ret = CreateProcess(0,                /* executable name */
                            cmd,              /* command line */
                            0,                /* process security attribute */
                            0,                /* thread security attribute */
                            TRUE,             /* inherits system handles */
                            processFlag,      /* selected based on exe type */
                            envcmd,           /* environment block */
                            cwd,              /* change to the new current directory */
                            &si,              /* (in)  startup information */
                            &pi);             /* (out) process information */It simply does not adjust memory size, or do something mysterious. It simply creates the process, and redirects the input , output and error streams.
    You really must get more memory (or adjust the size of your Windows swap file)...
    You can try lowering (not raising) the -Xmx limit of your Java program to make more room for the "link" process.

  • SAP MII 14.0 SP5 Patch 11 - Error has occurred while processing data stream Dynamic Query role is not assigned to the Data Server

    Hello All,
    We are using a two tier architecture.
    Our Corp server calls the refinery server.
    Our CORP MII server uses user id abc_user to connect to the refinery data server.
    The user id abc_user has the SAP_xMII_Dynamic_Query role.
    The data server also has the checkbox for allow dynamic query enabled.
    But we are still getting the following error
    Error has occurred while processing data stream
    Dynamic Query role is not assigned to the Data Server; Use query template
    Once we add the SAP_xMII_Dynamic_Query role to the data server everything works fine. Is this feature by design ?
    Thanks,
    Kiran

    Thanks Anushree !!
    I thought that just adding the role to the user and enabling the dynamic query checkbox on the data server should work.
    But we even needed to add the role to the data server.
    Thanks,
    Kiran

  • Problem with FMIS 4 and streaming of live events

    We have a problem on our platform and its driving us nuts... no seriously... NUTS.
    We have triple checked every possible component from a hardware level up to a software configuration level.
    The problem :  Our platform consists of 2 origin servers with 6 edges talking to them (really beefy hardware).  Once we inject a live stream into our two origins... we can successfully get the stream out via the edges and stream it successfully via our player.  Once we hit around 2200 concurrent connections, the FMIS servers drops all the connections busy with streams.   From the logs the only thing we can see is the following - Tons of disconnects with the Status code 103's which according to the online documentation means Client disconnected due to server shutdown (or application unloaded).
    We simulated the scenario with the FMS load simulator utility... and we start seeing errors + all connections dropped around the 2200 mark.
    The machines are Dell blades with dual CPU Xeons (quad cores) with around 50 gigs of ram per server... The edges are all on 10 Gb/s ethernet interfaces as well. 
    We managed to generate a nice big fat coredump on the one origin and the only thing visible from inspecting the core dumps + logs is the following :
    2011-10-05 
    15:44:10   
    22353   (e)2641112 
    JavaScript runtime is out of memory; server shutting down instance (Adaptor:
    _defaultRoot_, VHost: _defaultVHost_, App: livestreamcast_origin/_definst_). Check the JavaScript runtime size for this application
    in the configuration file.
    And from the core dump :
    warning: no loadable sections found in added symbol-file system-supplied DSO at 0x7fff9ddfc000
    Core was generated by `/opt/adobe/fms/fmscore -adaptor _defaultRoot_ -vhost _defaultVHost_ -app -inst'.
    Program terminated with signal 11, Segmentation fault.
    #0  0x00002aaaab19ab22 in js_MarkGCThing () from /opt/adobe/fms/modules/scriptengines/libasc.so
    (gdb) bt
    #0  0x00002aaaab19ab22 in js_MarkGCThing () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #1  0x00002aaaab196b63 in ?? () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #2  0x00002aaaab1b316f in js_Mark () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #3  0x00002aaaab19a673 in ?? () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #4  0x00002aaaab19a6f7 in ?? () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #5  0x00002aaaab19ab3d in js_MarkGCThing () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #6  0x00002aaaab19abbe in ?? () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #7  0x00002aaaab185bbe in JS_DHashTableEnumerate () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #8  0x00002aaaab19b39d in js_GC () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #9  0x00002aaaab17e6d7 in js_DestroyContext () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #10 0x00002aaaab176bf4 in JS_DestroyContext () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #11 0x00002aaaab14f5e3 in ?? () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #12 0x00002aaaab14fabd in JScriptVMImpl::resetContext() () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #13 0x00002aaaab1527b4 in JScriptVMImpl::postProcessCbk(unsigned int, bool, int) ()
       from /opt/adobe/fms/modules/scriptengines/libasc.so
    #14 0x00002aaaab1035c7 in boost::detail::function::void_function_obj_invoker3<boost::_bi::bind_t<void, boost::_mfi::mf3<void, IJScriptVM, unsigned int, bool, int>, boost::_bi::list4<boost::_bi::value<IJScriptVM*>, boost::arg<1>, boost::arg<2>, boost::arg<3> > >, void, unsigned int, bool, int>::invoke(boost::detail::function::function_buffer&, unsigned int, bool, int) ()
       from /opt/adobe/fms/modules/scriptengines/libasc.so
    #15 0x00002aaaab0fddf6 in boost::function3<void, unsigned int, bool, int>::operator()(unsigned int, bool, int) const ()
       from /opt/adobe/fms/modules/scriptengines/libasc.so
    #16 0x00002aaaab0fbd9d in fms::script::AscRequestQ::run() () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #17 0x00002aaaab0fd0eb in boost::detail::function::function_obj_invoker0<boost::_bi::bind_t<bool, boost::_mfi::mf0<bool, fms::script::AscRequestQ>, boost::_bi::list1<boost::_bi::value<fms::script::IntrusivePtr<fms::script::AscRequestQ> > > >, bool>::invoke(boost::detail::function::function_buffer&) () from /opt/adobe/fms/modules/scriptengines/libasc.so
    #18 0x00000000009c7327 in boost::function0<bool>::operator()() const ()
    #19 0x00000000009c7529 in fms::script::QueueRequest::run() ()
    #20 0x00000000008b868a in TCThreadPool::launchThreadRun(void*) ()
    #21 0x00000000008b8bd6 in TCThreadPool::__ThreadStaticPoolEntry(void*) ()
    #22 0x00000000008ba496 in launchThreadRun(void*) ()
    #23 0x00000000008bb44f in __TCThreadEntry(void*) ()
    #24 0x000000390ca0673d in start_thread () from /lib64/libpthread.so.0
    #25 0x000000390bed44bd in clone () from /lib64/libc.so.6
    From what it looks like above, FMS is hard crashing when trying to use clone(2) (basically it means when its trying to spawn a new/another process).
    I am really hoping there is someone out there who can guide us in the right direction with regards to how we can pinpoint why our platform cannot cope with a pathetic 2200 connections before the FMIS daemon drops all connected streams.
    There has to be someone out there that has run into this or a similiar problem like this...  HELP !!!!
    Any feedback / ideas would be greatly appreciated.

    Thank you very much for the reply :-)
    We have been fiddling with the platform on many levels yesterday, and one thing we did do was bump that value up from 1024 to 8192... This made a HUGE improvement in ensuring the platform now holds the live streaming connections. (up to 8000 per edge)
    I think for other future reference and to aid other people that might run into this problem in the future, its a good idea to increase this value.  From what we have seen, read and heard, that default value is fairly conservative, its suppose to grow when the load demands it, however, if you have a large scale of connections coming in at once from multiple locations, it can happen that it grows too quickly which can result in the application to be reloaded (which disconnects all users, i.e. all edge servers connected to this origin).
    Another option we were recommended to modify was the following :
    In adaptor.xml you currently have this line:
    <HostPort name="edge1" ctl_channel="localhost:19350" rtmfp="${ADAPTOR.HOSTPORT}">${ADAPTOR.HOSTPORT}</HostPort>
    You can set this to
                    <HostPort name="edge1" ctl_channel="localhost:19350" rtmfp=":80">:80</HostPort>
    <HostPort name="edge2" ctl_channel="localhost:19351" rtmfp=":1935">:1935</HostPort>
    This will create two edge processes for both ports 80 and 1935. Currently both ports are served from the same fmsedge process.
    Of course this is not a huge performance improvement, but should further distribute load over more processes which is a good thing. Especially when there are that many connections.
    This setting can be made on all machines (origin + edge)
    Hopefully this could help other people also running into the same problems we have seen ...

  • How do you improve the speed of processing and printing of PDF from CAD in indesign?

    I work for an engineering consultancy company and for a long while I have suffered with using PDFs files generated from AutoCAD when placed within an InDesign report. There are three issues that arise:
    The files take a long time to process when preparing the InDesign report. (This is mainly in high quality preview mode)
    The files take a long time to export to PDF
    (The most annoying one) The files take a long time to process and print when sent from the PDF to the colour laser printer. So much so, a single page from InDesign that contains a PDF generated from CAD can take up to 15 mins to print.
    I have tried converting the PDFs to jpegs prior to insertion into the InDesign document, but the quality and colour is dramatically reduced thus them useless. Is there any pre processing on the PDFs generated from CAD that I can do to reduce the impact that it has within InDesign and in the post processing?
    We also get missing font warnings that arise from the inserted PDFs from CAD that may or may not be an issue. Your advice here would also be appreciated.
    I would be the first to admit that I am not the most advanced user of InDesign though I am sure that my workflow could be improved to alleviate the situation.
    Your assistance would be appreciated!

    Thanks for your quick response.
    I have tried both options you have recommended, but unfortunately the file size doesn't seem to be the issue. Most of our CAD files (except for when they contain raster backgrounds) are around 0.5-2Mb which isn't massive compared to some jpegs. The PDF optimization process does make a small difference to the file size but only in the region of perhaps 5% reduction.
    If it helps the most awkward file I am using at the moment has this audit result, pre-optimization. This is typical of the CAD files produced by AutoCAD. As you can see the content streams take up the highest percentage that doesn't reduce that much even when the optimizer has all options engaged.
    What do you think?

  • Java.lang.Process input stream waiting until process is complete to print

    I have tried to simplify this problem as much as possible. Basically, I have a java.lang.Process which executes a simple C program.
    test.c
    #include <stdio.h>
    #include <unistd.h>
    int main()
      printf("foo\n");
      sleep(2);
      printf("bar\n");
    ...The process has an input stream and error stream handler each on a separate thread. I have tried both buffered and unbuffered (BufferedReader, BufferedInputStream, InputStreamReader...) stream handlers. Both produce the same problem of waiting until the process has exited to receive anything from the process's streams.
    The only time this does not happen is when I call fflush(stdout); after each printf(). This can't be a solution because the real application calls a massive C application which would require thousands of fflush()'s to be added. What is causing this to happen? This doesn't happen when the C program is executed from the shell. Is there a way the InputStream can be forced to extract from the stream?

    hi.....
    I have closed the output stream of the process as you told me to do...
    The hitch is that, if my program contains only printf() statements,it works fine
    as soon as scanf() statement is encountered within the C code,it is totally neglected,and the output comes as if no scanf() statement existed in the C code.
    Consequently the thread doesnt wait for input which was bound for scanf() from the thread
    the code...
        public void run()
         try
             PrintWriter out = new PrintWriter(socket.getOutputStream(),true);
             BufferedReader in = new BufferedReader(
                            new InputStreamReader(
                            socket.getInputStream()));
             try
                     Process p;
              p=new ProcessBuilder("./a.out").start();
                     PrintWriter exOut=null;
                     BufferedReader exIn=null;
              exOut = new PrintWriter(p.getOutputStream(),true);
              exIn = new BufferedReader(
                           new InputStreamReader(
                           p.getInputStream()));
                  //String inputLine="", outputLine="";        
                  String str="";
                     int c;          
                  while(true)                   
                        //System.out.println("In While");
                  str="";exOut.close();                  
                        while((c=exIn.read())!=-1)
                                 str=str+(char)(c);
                                    System.out.print(str);
                        str=str+(char)(0);
                        System.out.print(str+"outside");
                        out.print(str);
                        sleep(100);
                        try
                            int x=p.exitValue();
                              out.print(str);
                   System.out.print("Bye 1");
                            String str1="Bye"+(char)(0);
                   out.println(str1);              
                   break;
                        catch(IllegalThreadStateException e)
                            //System.out.println("The Process has not ended yet");
                        //str=str+((char)-1);
                        //System.out.print(str+"Control reaches here too");
                        str="";
                        exOut = new PrintWriter(p.getOutputStream(),true);//I have tried to run the program without this also but the effect is the same
                        while((c=in.read())!=-1)
                            str=str+(char)(c);                                    
                        if(str.contentEquals(""))
                                System.out.print("Bye 2");
                                String str1="Bye"+(char)(0);
                                out.println(str1); 
                                p.destroy();
                                exOut.close();
                                exIn.close();
                                out.close();
                                in.close();        
                                socket.close();
                                break;
                        //str=str+(char)(0);
                  exOut.print(str);
                        try
                            int x=p.exitValue();
                            System.out.print("Bye 3");
                            String str1="Bye"+(char)(0);
                   out.println(str1);
                            break;
                        catch(IllegalThreadStateException e)
                            //System.out.println("The Process has not ended yet");
                  /*while ((inputLine = in.readLine()) != null)
                        exOut.println(inputLine);
                        outputLine=exIn.readLine();
                        //outputLine=inputLine;
                        //out.println(outputLine);}*/                   
             exOut.close();
             exIn.close();
             catch(IOException e)
                  System.err.println("Accept failed."+e);
             out.close();
             in.close();        
             socket.close();
         catch (Exception e)
             e.printStackTrace();
    }

  • Airport issues and streaming music

    Hi all, after quite a bit of frustration with my AX, which I might add, worked well for years, I installed a NetGear wireless modem and am very happy with it.
    One issue is that previously, when the AX connected with ethernet cable to my router, I was streaming music around the house, something I really enjoyed. This AX has now been removed.
    I assumed that it was possible to connect the second AX to my new wireless network. (created with the new NetGear)
    Not so, Airport utility still recognised the second unit and I proceeded to complete the process. But the AX would not re-start. Now, admittedly I only tried it once, but is it possible to still connect the second AX to my wireless network, and stream music, via AirTunes?
    Appreciate any comments....am I trying to do the impossible?

    Ronald,
    Have you reset the Airport Express?
    Then try and join your new network.... which is now the NetGear...
    I am running 3 Airport Express units to 3 different stereos and they all work fine. Even though I have a different Main Base station than you have, you should be able to "Join" your network.

  • Golden Gate and Streams

    Oracle 11.1.0.7:
    Before I ask my simple question. Is there a place I can get detailed documentation of Golden Gate? I got some documentation but that is very high level. I am looking for some document or tutorial like the ones Oracle has on tahiti.oracle.com.
    We are currently using streams and personally worked on it. I see following issues:
    1. When some errors occur in dba_apply_error it's really slow to re-execute them. Recovery is very slow
    2. Streaming the LOBs over of just 512 bytes makes the CCA very slow
    3. Manageability of scripts, configuration etc. is a headache
    4. Come accross streams bugs all the time
    From what I have read about golden gate is that is super fast and better than streams. I haven't read about it so I don't have complete comparision. Could someone with similar experience provide the insight.
    Also, if someone can please help me get right detailed documents would be really great.

    Hi,
    Think the goldengate documentation provide adequate help to install and configure oracle goldengate process. I can put down a sample configuration for you to get started.
    Installing GoldenGate and configuring a extract replicat process.
    First FTP the installation zip in binary mode from the PC to the installation directory on machine where Oracle GoldenGate suppose to be installed.
    Use gzip to decompress the file.
    gzip –d <file_name>.gz
    Extract the files
    Use tar to extract the files. All GoldenGate files will be placed in the current subdirectory.
    tar –xvf <file_name>.tar
    Create folders
    Create the required sub-directories in the installation directory by using
    Shell> cd <install location>
    Shell> ggsci
    GGSCI> create subdirs
    GGSCI> exit
    Oracle-specific installation steps
    The below steps are only required when GoldenGate will be capturing data from the server. Therefore it is only needed when you are installing GoldenGatef or Oracle as your <source> environment.
    You need to turn on supplemental logging at the database level. This command requires the ALTER DATABASE permissions.
    Shell> sqlplus <login>/<password>
    SQL> alter database add supplemental log data;
    Now Switch to the next redo log file.
    SQL> alter system switch logfile;
    SQL> exit
    Then now configure manager parameter file.
    GGSCI> EDIT PARAMS MGR
    •     Use the editor to assign a port
    --GoldenGate Manager parameter file
    PORT <port>
    example: PORT 7840
    •     To Start the Manager, issue following command
    GGSCI> START MGR
    •     To Verify that the Manager has started, issue the following command
    GGSCI> INFO MGR
    Now Configure Supplemental logging.
    By default, if not supplimentory logging enabled , only the change vectors will be logged in redo logs So inorder to get the columns required for constructing the image by replicat in target we require atleast minimum supplimentory logging enabled in database level and table level logging required for proper functioning of Oracle Goldengate Software.
    Using GGSCI, log in to the database on the <source> and turn on supplemental logging for the the tables configured for replication
    Shell> ggsci
    GGSCI> DBLOGIN USERID <login>
    GGSCI> ADD TRANDATA <owner/schema>.<table-name>
    Do this for all the tables.
    Configure change capture
    1. Add the Extract group
    Execute the following command on the <source> system to add an Extract group
    GGSCI> ADD EXTRACT <extract-name>, TRANLOG, BEGIN NOW
    Verify the results:
    GGSCI> INFO EXTRACT <extract-name>
    Create the Extract parameter file
    Execute the following commands on the <source> system.
    GGSCI> EDIT PARAM <extract-name>
    EXTRACT <extract-name>
    USERID <login>, PASSWORD <password>
    RMTHOST <target>, MGRPORT <port>
    RMTTRAIL ./dirdat/<trail id>
    TABLE <owner/schema>.<table-name>
    TABLE <owner/schema>.<table-name>
    for example
    extract testext
    userid ggs password ggs
    rmthost 172.25.25.177
    rmttrail ./dirdat/tr
    table scott.mytable;
    table scott.yourtable;
    Define the GoldenGate trail
    Execute the following command on the <source> to add the trail that will store the changes on the <target>.
    GGSCI> ADD RMTTRAIL ./dirdat/<trail id>, EXTRACT <extract-name>
    Verify the results:
    GGSCI> INFO RMTTRAIL *
    Start the capture process
    GGSCI> START EXTRACT <extract-name>
    Configure Change Delivery or replicat
    we have to do the following
    •     Set up the checkpoint table on the target system.(this is optional)
    •     Create a named group that includes the Replicat process and the checkpoint tables.
    •     Configure the Replicat group by adding parameters.
    •     Start the Replicat group.
    Now we will check how to set up the checkpoint table (this is a optional settings)
    Create a GLOBALS file on the target system
    Execute the following commands on the <target> system.
    Create and edit the GLOBALS parameter file to add the checkpoint table.
    Shell> cd <install location>
    Shell> ggsci
    GGSCI> EDIT PARAMS ./GLOBALS
    In the text editor, type:
    CHECKPOINTTABLE <owner/schema>.ggschkpt
    •     Record the checkpoint table owner and name, then save and close the file.
    Please Note: You could name the table anything you want
    •     Verify that the GLOBALS file was created in the root GoldenGate directory, and remove any file extension that was added.
    Activate the GLOBALS parameters
    For the GLOBALS configuration to take effect, you must exit the session in which
    the changes were made. Execute the following command to exit GGSCI.
    GGSCI> EXIT
    Add a Replicat checkpoint table     
    On the <target> system, execute the following commands in GGSCI:
    Shell> cd <install location>
    Shell> ggsci
    GGSCI> DBLOGIN USERID <login>, PASSWORD <password>
    GGSCI> ADD CHECKPOINTTABLE
    Configure Change Delivery
    Add the Replicat group
    Execute the following command on the <target> system to add a delivery group
    named <replicat-name>
    GGSCI> ADD REPLICAT <replicat-name>, EXTTRAIL ./dirdat/<trail id>
    Note: Refer to your Extract set up for the correct two-character <trail id>.
    Create Replicat parameter file
    Execute the following commands on the <target> system to bring up the parameter
    file in the editor.
    GGSCI> EDIT PARAM <replicat-name>
    -Type in the following parameters
    REPLICAT <replicat-name>
    USERID <login>, PASSWORD <password>
    HANDLECOLLISIONS
    ASSUMETARGETDEFS
    DISCARDFILE ./dirrpt/RORA<unique id>.DSC, PURGE
    MAP <owner/schema>.<source table-name>, TARGET <owner/schema>.<target table-name>;
    Start the Replicat process
    GGSCI> START REPLICAT <replicat-name>
    Verify the results:
    GGSCI> INFO REPLICAT <replicat-name>
    The above i scrapped is a basic extract-replication process which you can scale up once you get accustomed with this beautiful product.
    Thanks & Regards,
    Antony Joseph Jethu.

  • Java.lang.Process output stream problem

    Hi, I have a program that starts a process (java.lang.Process) using the java.lang.Runtime.exec() and it attemtps to interface with it using the provieded io streams. I have both the output and error streams being handled on their own threads and I have a hashmap of output lines/command pairs that are checked so that when the process outputs certain lines to the console it feed the proper input into the process. My problem is that when I feed the input into the process it dosen't respond to it almost like the user hasn't pressed enter, The process hangs. I have tried using /n /r and permutations thereof but nothing works. The thread does read the lines from the process and does output to the process from what i can gather. Can you help me!
    here is some of the code..
    public void run() {
    try {
         //the process's output
         InputStreamReader isrOutput = new InputStreamReader(inOutput);
         //the process's input(our output)
         PrintWriter pw = new PrintWriter(outInput);
         String line = null;
         while(true){
              if(brOutput.ready()){
                   line = "";
                   while(brOutput.ready())
         line+=(char)brOutput.read();
                   System.out.print(line);
                   if(commands.containsKey(line)){
         pw.println((String)commands.get(line));
         System.out.println((String)commands.get(line));;
    } catch (IOException ioe) {
              ioe.printStackTrace();
    }Thanks

    Oops.. i forgot to flush my PrintWriter /blushing......... Thanks

  • 6534 DAQ card- i want to process continous stream of data....through DAQ card...

    I have NI -6534 DAQ card.i have continous stream of 16-bit parallel data. I use two port as input port .(16 bits parallel data), such that i recieve  12000 bytes per second in my card buffer.  i  use  DIO read.vi. then on this bunch of 12000 bytes, i do some processing and display a XY chart/waveform. 
    code structure is as follows:
    1) DIO config. vi
    2) DIO start. vi
    3) while loop
        {          DIO read.vi (reads 12000/or fixed number of bytes, inserts into an array )
                   Insert array.vi
                   while loop(this processes data till end of 12000 bytes,then stops)
                                      stacked sequence(this processes the 12000 bytes data, )
                                        (      seq 0: intialize counters,
                                                seq 1: extract some pre-defined bytes,convert and display on frontpanel
                                                seq 2: extract specified bunch of bytes, create  xy chart/waveform
                                 }(12000 bytes data ends)
         }(16-bit parallel data is unavaible/or user control stop)
    4)dio clear.vi
    the problem persceptable here is , that processing data takes time.. few seconds for processing bucnh of data..then goes back to DIO read to fetch new bunch of data. till then the front panel waveforms and digital displays are on hold/static.
    I want to process data simulatanouesly , and also being recieve in buffer through DAQ.real time processing.
    simulatanoeus aqcuuistion..and process on continous data rather than in bunch of data.(i.e. TWO vi. should run in paralell  DIO READ.vi acqusition...and process.. on that data(vi) )
    Can queue.vi slove this problem? any other method
    your help is greatly appreciated..Thanks.

    Hi,
    This is definitely something that can be solved using a producer consumer loop.  You would have two loops running in parallel, have all of your data acquisition/generation done in the producer loop.  Use queues to transfer all that data from the producer loop to the consumer loop and then do any kind of processing in your consumer loop.
    The producer and consumer loop will run in parallel, and the consumer loop cannot run faster that the producer loop.
    The good thing is that you can get a producer consumer template easily.  In the LabVIEW startup window, select New (not New VI) >> From Template >> Frameworks >>Desgin Patterns >> Producer/Consumer Design Pattern (Data).
    Hope that helps.
    Regards,
    Raajit L
    National Instruments

  • When I plug in my iPhone to sync, it starts but does not complete the process and several pieces of data are not being transferred to my iPhone from my MacBook Air.

    Problem:
    When I plug in my iPhone to sync, it starts but does not complete the process and several pieces of data are not being transferred to my iPhone from my MacBook Air.
    Any help that you can provide in helping my iPhone accurately sync with iPhoto and iTunes will be greatly appreciated.
    Symptoms:
    1)   Sync:  It’s not completing the sync.  Below, I’ve provided all of my settings from the iTunes Summary tab so that you might tell me if I’ve selected any incorrect options.  I prefer to sync the “old school” way – by connecting to the computer – as opposed to syncing over the cloud. Perhaps this is what’s causing the problem? Here is a list of the steps displayed in the iTunes window as the sync progresses:
    waiting for sync to start (step 1 of 7)
    backing up (step 2 of 7)
    preparing to sync (step 3 of 7)
    determining apps to sync (step 4 of 7)
    preparing apps to sync (step 5 of 7)
    importing photos (step 6 of 7)
    waiting for changes to be applied (step 7 of 7)
    syncing apps / copying 1 of 4 (App name) (step 7 of 7)
    canceling sync
    apple icon
    2)   Photos: I've selected only certain of my iPhoto albums to sync to my iPhone.  All of the albums are correct/complete in iPhoto.  All of the albums are listed on my iPhone, both before and after the sync, but the albums are empty (no actual photos) before and after the sync. Perhaps this is tied to the fact that the sync isn’t completing, but because “importing photos” is one of the steps that the incomplete sync displays, I don’t know.
    3)   Apps: When I launch iTunes and click on the Apps folder under the Library listing, then click on the Updates tab, iTunes searches for any Apps needing to be updated and provides a list.  If I click on Update All, the Apps are successfully updated in iTunes.  But, when I plug in my iPhone so that the updates will transfer to the apps on my iPhone, the updates don’t transfer to the apps on my iPhone and those apps still reflect that they need updating on the iPhone.
    Other Potential Pertinent Info:
    The flash memory hard drive on my MacBook Air recently died (perhaps a month or two ago).  Apple had emailed me about a known issue and literally the next day, my MacBook Air crashed.  I installed a new flash memory drive and re-installed everything from a backup off of an external hard drive.  Everything seems to be working fine; it recreated accurately all of my software and data, including iPhoto and iTunes, the pictures and songs (respectively) for which are stored on that hard drive, as opposed to being on the flash memory in the MacBook Air itself.  However, I don’t recall if the start of the sync problem described herein started happening at the same time that I replaced the flash memory drive.  All I know is that the computer is working perfectly in all respects and that even as the sync is failing, it at least says that it’s doing the right things and looking in the right places (e.g., the list of albums on my iPhone matches the list of albums in iTunes, etc.).
    Settings/Status:
    MacBook Air
    OSX v. 10.9
    iPhoto ’11 v. 9.5 (902.7)
    iPhone iOS 7.0.4
    iTunes v. 11.1.3 (8)
    Summary Tab
    Backups (This Computer)
    Options
    Automatically sync when this iPhone is connected
    Sync only checked songs and videos
    Photos Tab
    Sync Photos from iPhoto (429 Photos)
    Selected albums, Events, and Faces, and automatically include (no Events)
    Albums – 9 are selected

    You need to download iTunes on your computer. iOS 6 requires the latest version of iTunes which is 10.7.

  • Third Party Process and Inventory

    We are using the SAP function "Third-Party Processing". We are using the function in it's standard (SAP R/3 4.7).
    The auditor now has the following issue: Shipping the goods starts after paying the goods. Out of this, while shipping the goods, we are the owner of them. In this circumstance the ship is our warehouse.
    SAP does not allow an Inventory Management within Third-Party Processing:
    "Since a third-party order involves a movement from the vendor direct to the customer, Inventory Management is not in any way affected. If you wish to document the movement in the system, however, you can create a goods receipt for the third-party order."
    With the SAP functions in this scenario it is correct that the warehouse stock is not being impacted. This is a conflict with the opinion of the auditor.
    Let me explain in detail what is wrong out of the view of our auditor:
    For example we initiated the order of a customer in June and we have paid the goods in June. The shipment starts in June and the goods will arive in July. Due to the fact, that we are the owner of the goods on the ship, we have to increase the stock. This is the opinion of our auditor.
    Does any body know, how to solve the issue? How do you manage Third Party Processing & Inventory?

    Some auditors just dont understand the SAP process. Instead of talking to weak users, they should talk to SAP directly.
    What about shipments from standard sales orders that are shipped end of June and do arrive at the customer in July?
    When do you bill your customer? when the goods arrives there, or earlier? Depending on the Incoterms you may still be the owner until the truck is unloaded.
    If you change your process and post goods receipt to your warehouse, then you show inventory where no physical inventory is, you posted a goods receipt without a physical receipt. I am sure the next auditor will not allow such therotical process.
    Get together, IT consultants of FI and MM and SD, strong business people for the same modules and the auditor, and discuss and find an optimal solution.And let the auidor sign the decision paper!!! Very important, I recently had to convice German auditors because they had different opinion than the Australian and US auditors of the same auditing company (and the issue had nothing to do with local legal requirements).

  • What is the optimum core configuration for a new Mac Pro to process and manipulate very large (80 megapixel) images using PhotoShop and Camera Raw?

    Hello:
    I will be using creative techniques to process and manipulate a large number (hundreds) of very large (80 megapixel) images captured using a medium format digital back (Phase One IQ180).
    Final output will be digital fine art imagery printed using an Epson 11880 at large sizes (up to 60 inches x ?), retaining the highest possible quality and resolution. I will be using Adobe CC PhotoShop and Camera RAW as well as Capture One software. PhotoShop filters will be used extensively.
    The Mac Pro needs to be optimized for the above purpose and be useful for at least five years. I plan to max out all the other options (RAM, graphics cards, storage). Performance is more important than cost.
    The few discussions I have found that mention optimum core configurations seem to lean toward 6 or 8 (but likely are not taking into consideration my need for manipulating a large number of very large files), so I am looking to this foum for opinions.
    Thank you,
    Kent

    See if this helps
    http://macperformanceguide.com/index_topics.html#MacPro2013

  • Image Processing and machine vision don't appear in Function Palette

    Hi,
    I have problem and I can't solve it
    Image Processing and machine vision don't appear in Function Palette, Although I set up NI imaq and NI Vision Acquisition Why??
    Anyone can help me?
    Solved!
    Go to Solution.

    If the image is already a Grayscale U8 image you would just have to save it to a bitmap file using the IMAQ WriteFile VI. If not, use the IMAQ Cast Image VI to convert it to a Grayscale U8 and then save it.
    Also, it would probably be better to create a new thread for questions like these.
    Cameron T
    Applications Engineer
    National Instruments

  • I tried to install yosemite yesterday and my computer goes through the whole installation process and then to a blank white screen.  I've rebooted, I've reinstalled several times and get the same result.  Two days wasted!

    I tried to install yosemite yesterday and my computer goes through the whole installation process and then to a blank white screen.  I've rebooted, I've reinstalled several times and get the same result.  Two days wasted and I can't use my computer!!

    I don't know if you've already resolved your problem, but I had the same thing, it took me 6 hours to fix it. I had the exact same as you, installed the update and then it went to a white screen. After trying start up holding down cmmd r or holding down the alt key to try to install it again, nothing worked, same result every time. Then an angel came to me here somewhere but I can't find it now to thank him. Start in safe mode, press shift once you hear the start up chime and hold it down until you hear it again. Installation completion box came up and it was working. turn off computer and start up again normally and all is well. Except my final cut pro x doesn't work with it, have to update that now too. same as when I updated to mavericks.
    Hope this helps.

Maybe you are looking for

  • Problem setting up Sky Sports

    Hi, Received a BT Vision Box yesterday with the Viewing Card for Sky Sports 1+2. Bt Vision TV is all working fine, and put the viewing card into the slot within half an hour of setting the tv up. Followed the instructions and waited for the hour on t

  • Can Compressor 4 remove duplicate frames?

    I have some videos which were converted to 30fps from a lower frame count. It was obviously a simple conversion that duplicated frames as needed. The problem is, when I use these videos for slow motion, the duplicated frames cause the slow motion to

  • Print command doesn't work.  A_8_pro

    New issue, worked previous use.  Know of no system changes.  Wireless printer works with all other programs.  Tried "fix" and "image" without success.  Any suggestions other than remove and re-install?

  • Ipod windows to new imac?

    Any information on how to transfer songs from an ipod that was previously used with windows to a new imac? There is no hard back up of the windows itunes cause the computer is fried. The Manual manage doesnt seem to be working correctly with the ipod

  • How to copy from in linux??

    i've just installed linux. trying to install packages ftp://ftp.archlinux.org/core/os/i686/core.db.tar.gz ftp://ftp.archlinux.org/extra/os/i686/extra.db.tar.gz ftp://ftp.archlinux.org/community/os/i6 - .db.tar.gz so that i can install xkfe offline as