VI Reference Output LV 2009

I have 2 applications. Client made in LV 2009 call the server made in LV 6.0. The previous methode was calling: Open Application Reference (from Client on Server), Open VI Reference (VI on Server) and send input data to Call By Reference Node and get the result from the output if Call By Reference Node. But, the frist time I call this, I have no output (data = 0), but on server, everyting is ok. When I did a second time (without exit the Client Software), it's working proprely.
I attach a png when you can see the flow I use.
Thank
Attachments:
VI_Reference.png ‏29 KB

CFiset,
Hello! Would it be possible to convert your LV 6.0 Server to LV 2009 to see if it exhisbits the same behavior? Also, are you running the server first before running the client application?
Ben Sisney
FlexRIO V&V Engineer
National Instruments

Similar Messages

  • Open reference / output in the same window

    I want a vi (opened and started at runtime per reference) to display
    its data on the frontpanel of the calling VI. How can this be done?
    Thanks,
    Andre

    THis can be done using control references.
    Pop-up on the control you want the sub-VI to update. Create a reference to the control. Pass this reference to the called sub-VI when it is called (via its connector pane).
    Inside the sub-VI, use a property node>>>value to update the indicator when required.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • NI5112 PFI Reference Clock Input Problem

    I'm writing C code to use an NI5112 O-Scope card as a digitizer and I would like to synchronize it with an HP 8648C signal generator (which has the high-stability timebase option) through PFI1. I want to use the signal generator for the reference clock as the NI5112 samples about 40 Hz high at 5 MSamples/s and because I need the two to drift together. When I use the signal generator as reference clock the result is a very "noisy", i.e. with an input sine wave the FFT is noisy, and I the timing is too bad for processing input GPS signals (which have stringent timing requirements). If I use an Agilent 33120A function generator as the reference clock, I get better results, but GPS processing is still compromised. When not using a
    n external reference everything looks clean and works fine except for 40 Hz error in the sample rate. The reference outputs from both the 8648C and 33120A work fine with each other and look good and clean with the o-scope, so I can't figure what the problem is. Maybe the card is just extremely picky about external reference clocks. The function call I'm using to initialize this is:
    handleErr(niScope_ConfigureClock(vi, NISCOPE_VAL_PFI_1,
    NISCOPE_VAL_NO_SOURCE,
    NISCOPE_VAL_NO_SOURCE,
    NISCOPE_VAL_FALSE));
    The program runs normally with no errors, it's just that the results are bad. Any help would be appreciated.

    Hello,
    If the data that you are acquiring is very noisy, I would start getting suspicious about the input impedance (1MOhm or 50Ohm). Does the impedance of the signal generator match the scope's? How does your signal look in the Scope Soft fron Panel? Does it make a difference if you use the internal clock?
    Another thing that I would consider is the following parameter from the niScope_ConfigureClock function: NISCOPE_ATTR_CLOCK_SYNC_PULSE_SOURCE. In order to synchronize your arb to the scope, not only they need to share the same clock (10MHz), but also the slave receives this sync pulse from the master. In you application, if the signal generator has the ability to send a sync pulse, you can route that signal through PFI2 to ensure synchronization.
    Hope this helps. Good luck with your application!
    Annette Perez

  • Close reference from event data node

    When using an event structure, I often times use
    the Control Reference output in the Event Data Node.  Do I need to
    close these references each time the corresponding event fires?  In
    other words, are the reference created from this node unique each time
    that particular case is called?

    The Event Structure does not create duplicate references to the same object if the event case is called multiple times, so you don't have to worry about having dozens of open references. You can actually verify this by type-casting your control reference to an I32 to see the exact 32 bit value (or just probe the refnum). Firing the event case multiple times produces the same reference.
    You really don't need to worry about closing the reference in this case. In general, LabVIEW will figure out that the control references are no longer valid as soon as the VI goes idle, and it will clean them up. Closing references is good practice if you're going to be opening a large number of references (for instance, getting references to all controls on a large Front Panel), and you don't want them all to stay in memory the whole time.
    Jarrod S.
    National Instruments

  • Online Backup Failed (Status Its shows Scheduling failed)

    Dear all ,
    OS AIX , DB Oracle , SAP ECC6.0 , Tivoli Storage Manager
    Scheduled  Online Backups are  failed  (Backups are fired through DB13)
    JOB Status : Scheduling failed
    Job Log:
    02.12.2009     17:00:01     Job started
    02.12.2009     17:00:01     Step 001 started (program RSDBAJOB, variant &0000000000657, user ID BASIS)
    02.12.2009     17:00:01     No application server found on database host - rsh/gateway will be used
    02.12.2009     17:00:01     Execute logical command BRBACKUP On host PRODORADB
    02.12.2009     17:00:01     Parameters:-u / -jid ALLOG20091201170000 -c force -t online -m all -p initIRPoffsite.sap -a -c force -p initIRP
    02.12.2009     17:00:01     offsite.sap -cds
    02.12.2009     17:01:01     SXPG_STEP_XPG_START: is_local_host: rc = 403
    02.12.2009     17:01:01     SXPG_STEP_XPG_START: host = PRODORADB
    02.12.2009     17:01:01     SXPG_STEP_XPG_START: is_local_r3_host: rc = 802
    02.12.2009     17:01:01     SXPG_STEP_XPG_START: RFC_TCPIP_CONNECTION_OPEN: rc = 1003
    02.12.2009     17:01:01     SXPG_STEP_COMMAND_START: SXPG_STEP_XPG_START returned: 1.003
    02.12.2009     17:01:01     SXPG_COMMAND_EXECUTE(LONG)
    02.12.2009     17:01:01     <timestamp> = 20091202170101
    02.12.2009     17:01:01     COMMANDNAME = BRBACKUP
    02.12.2009     17:01:01     ADDITIONAL_PARAMETERS =
    02.12.2009     17:01:01     -u / -jid ALLOG20091201170000 -c force -t online -
    02.12.2009     17:01:01     m all -p initIRPoffsite.sap -a -c force -p initIRP
    02.12.2009     17:01:01     offsite.sap -cds
    02.12.2009     17:01:01     LONG_PARAMS
    02.12.2009     17:01:01     OPERATINGSYSTEM = ANYOS
    02.12.2009     17:01:01     TARGETSYSTEM = PRODORADB
    02.12.2009     17:01:01     DESTINATION
    02.12.2009     17:01:01     SY-SUBRC = 1003
    02.12.2009     17:01:01     SXPG_COMMAND_EXECUTE failed for BRBACKUP - Reason: program_start_error: For More Information, See SYS
    02.12.2009     17:01:01     Job cancelled after system exception ERROR_MESSAGE
    Note :
    when i execute immediately or through Command line  backup  is going fine
    but when i scheduling  backup it was Cancelled  (Status shows Schudling Failed)
    pls give me suggestion
    sathessh

    Dear Yoganand.V
    thanks for your immediate response
    and still i am facing this issue
    the Backups are triggered from DB13 and Scheduled from Central Instance not Application Server and Auto Script is running at OS level and pls look FINISHED BACKUP LOG for your Reference
    08.12.2009     17:50:18     Job started
    08.12.2009     17:50:18     Step 001 started (program RSDBAJOB, variant &0000000000717, user ID S2K_BASIS)
    08.12.2009     17:50:18     No application server found on database host - rsh/gateway will be used
    08.12.2009     17:50:18     Execute logical command BRBACKUP On host PRODORADB
    08.12.2009     17:50:18     Parameters:-u / -jid ALLOG20091208175017 -c force -t online -m all -p initIRPdaily.sap -a -c force -p initIRPda
    08.12.2009     17:50:18     ily.sap -cds
    08.12.2009     18:29:23     BR0051I BRBACKUP 7.00 (34)
    08.12.2009     18:29:23     BR0055I Start of database backup: becbwhwk.anf 2009-12-08 17.50.18
    08.12.2009     18:29:23     BR0484I BRBACKUP log file: /oracle/IRP/sapbackup/becbwhwk.anf
    08.12.2009     18:29:23     BR0477I Oracle pfile /oracle/IRP/102_64/dbs/initIRP.ora created from spfile /oracle/IRP/102_64/dbs/spfileIRP.ora
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 17.50.19
    08.12.2009     18:29:23     BR0319I Control file copy created: /oracle/IRP/sapbackup/cntrlIRP.dbf 15122432
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 17.50.20
    08.12.2009     18:29:23     BR0057I Backup of database: IRP
    08.12.2009     18:29:23     BR0058I BRBACKUP action ID: becbwhwk
    08.12.2009     18:29:23     BR0059I BRBACKUP function ID: anf
    08.12.2009     18:29:23     BR0110I Backup mode: ALL
    08.12.2009     18:29:23     BR0077I Database file for backup: /oracle/IRP/sapbackup/cntrlIRP.dbf
    08.12.2009     18:29:23     BR0061I 42 files found for backup, total size 141794.742 MB
    08.12.2009     18:29:23     BR0143I Backup type: online
    08.12.2009     18:29:23     BR0130I Backup device type: util_file_online
    08.12.2009     18:29:23     BR0109I Files will be saved by backup utility
    08.12.2009     18:29:23     BR0142I Files will be switched to backup status during the backup
    08.12.2009     18:29:23     BR0289I BRARCHIVE will be started at the end of processing
    08.12.2009     18:29:23     BR0134I Unattended mode with 'force' active - no operator confirmation allowed
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 17.50.20
    08.12.2009     18:29:23     BR0229I Calling backup utility with function 'backup'...
    08.12.2009     18:29:23     BR0278I Command output of '/usr/sap/IRP/SYS/exe/run/backint -u IRP -f backup -i /oracle/IRP/sapbackup/.becbwhwk.lst -t file_onli
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     Data Protection for SAP(R)
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     Interface between BR*Tools and Tivoli Storage Manager
    08.12.2009     18:29:23     - Version 5, Release 5, Modification0.0  for AIX LF 64-bit -
    08.12.2009     18:29:23     Build: 316B  compiled on Oct 23 2007
    08.12.2009     18:29:23     (c) Copyright IBM Corporation, 1996, 2007, All Rights Reserved.
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BKI2027I: Using TSM-API version 5.5.1.0 (compiledwith 5.3.0.0).
    08.12.2009     18:29:23     BKI2000I: Successfully connected to ProLE on porttdpr3ora64.
    08.12.2009     18:29:23     BKI0005I: Start of program at: Tue Dec  8 17:50:20 IST 2009 .
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     -- Parameters --
    08.12.2009     18:29:23     Input File            : /oracle/IRP/sapbackup/.becbwhwk.lst
    08.12.2009     18:29:23     Profile               : /oracle/IRP/102_64/dbs/initIRPdaily.utl
    08.12.2009     18:29:23     Configfile            : /oracle/IRP/102_64/dbs/initIRP.bki
    08.12.2009     18:29:23     Manual sorting file   : disabled
    08.12.2009     18:29:23     Tracefile             : disabled
    08.12.2009     18:29:23     Traceflags            : disabled
    08.12.2009     18:29:23     Parallel sessions     : 1
    08.12.2009     18:29:23     Multiplexed files     : 1
    08.12.2009     18:29:23     RL compression        : 0
    08.12.2009     18:29:23     Exit on error         : disabled
    08.12.2009     18:29:23     BATCH                 : enabled
    08.12.2009     18:29:23     Buffer size           : 131072
    08.12.2009     18:29:23     Buffer Copy Mode      : SIMPLE
    08.12.2009     18:29:23     Redologcopies         : disabled
    08.12.2009     18:29:23     Versioning            : enabled
    08.12.2009     18:29:23     Current Version      : 581
    08.12.2009     18:29:23     Versions to keep     : 6
    08.12.2009     18:29:23     Delete Versions      : <= 575
    08.12.2009     18:29:23     Backup Type           : file_online
    08.12.2009     18:29:23     TSM log server        : disabled
    08.12.2009     18:29:23     TSM server            : [email protected] with 2 sessions configured, using 1 session
    08.12.2009     18:29:23     TSM client node      : PRODORADB_TDP
    08.12.2009     18:29:23     Days for backup      : Sun Mon Tue Wed Thu Fri Sat
    08.12.2009     18:29:23     Backup mgmt class    : DAILYMGMTCLASS
    08.12.2009     18:29:23     Archiv mgmt class    : DAILYMGMTCLASS
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 18.24.44
    08.12.2009     18:29:23     BR0232I 8 of 8 files saved by backup utility
    08.12.2009     18:29:23     BR0230I Backup utility called successfully
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0056I End of database backup: becbwhwk.anf 2009-12-08 18.24.44
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 18.24.44
    08.12.2009     18:29:23     BR0052I BRBACKUP completed successfully
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 18.24.44
    08.12.2009     18:29:23     BR0291I BRARCHIVE will be started with options '-U -jid ALLOG20091208175017 -d util_file -c force -p initIRPdaily.sap -cds'
    08.12.2009     18:29:23     
    08.12.2009     18:29:23     BR0280I BRBACKUP time stamp: 2009-12-08 18.29.23
    08.12.2009     18:29:23     BR0292I Execution of BRARCHIVE finished with return code 0
    08.12.2009     18:29:23     Job finished
    regards

  • FSL-06002  Error 14001 This application has failed to start

    Hello Folks
    I have problem installing SAP R/3 Enterprise (4.72 Enterprise Ext2) IDES Version
    The installation is under Windows 2003 server SP2 and Oracle 920
    I'm in the Database Instance Installation (non-unicode) and phase 29 (Updating database statistics) 
    The following error is:
    TRACE      [synxccuren.cpp:452]
               grantAccessTo
    Granted access rights 0xf01ff for object default to user chavez-server\mauadm with inheritance flags 0.
    TRACE<i>   [synxccuren.cpp:1045]
               CSyCurrentProcessEnvironmentImpl::setUser()
    Switched to user: mauadm.
    ERROR 2009-08-23 13:27:58
    FSL-06002  Error 14001 (This application has failed to start because the application configuration is incorrect. Reinstalling the application may fix this problem.
    ) in execution of a 'CreateProcess' function, line (284), with parameter (E:\usr\sap\MAU\SYS\exe\run/brconnect ...).
    When i ran the brconnect manually i received the same error  (This application has failed to start because the application configuration is incorrect. Reinstalling the application may fix this problem)
    When i ran brconnect using cmd.exe i received the following error
    E:\usr\sap\MAU\SYS\exe\run>brconnect.exe
    The system cannot execute the specified program.
    I think is important mention this, i had problem 1 setp before about the brtools 7.00
    Execution of the command "E:\usr\sap\MAU\SYS\exe\run/brconnect -u / -c -f crsyn -o SAPMAU" finished with return code 3. Output:
    BR0801I BRCONNECT 7.00 (16)
    BR0999E Loading of SQL client library oci.dll failed - please check Oracle configuration
    BR0310E Connect to database instance MAU failed
    BR0280I BRCONNECT time stamp: 2009-08-23 13.06.43
    BR0804I BRCONNECT terminated with errors
    I downloaded the correct version (640) DBATOOLS Package for Oracle 9.2 and i replaced the files
    Please folks someone have any idea about this, i tried search on the web and nothing, on the marketplace and nothing
    Thanks all for your help

    >
    Markus Doehr wrote:
    > > I have problem installing SAP R/3 Enterprise (4.72 Enterprise Ext2) IDES Version
    > > The installation is under Windows 2003 server SP2 and Oracle 920
    >
    > and something on top: 640_EX2 is only for Oracle 10g - it  won't work with Oracle 9.2. So you should change your kernel to 640 back - then the R3DLLINS works.
    >
    >
    >
    >
    > Markus
    Ok Markus, i did the following
    First i downloaded vcredist_<platform>.exe from Microsoft web page and retry the installation
    And the following error i got
    Operative System open's a Window that say following
    brconnect.exe Unable to locate component "this application has failed to start because orasql10.dll was not found. Re-installing the application may fix this problem"
    And this is the log
    WARNING 2009-08-23 17:52:07
    Execution of the command "E:\usr\sap\MAU\SYS\exe\run/brconnect -u / -c -f crsyn -o SAPMAU" finished with return code 128. Output:
    ERROR 2009-08-23 17:52:07
    CJS-00288  Could not update database statistics.<br>DIAGNOSIS: Command brconnect -u / -c -f crsyn -o SAPMAU returned 128.<br>SOLUTION: See brconnect.log for details.
    Second i downloaded the kernel 640 whitout ext2 and i get the following error
    INFO 2009-08-23 17:41:32
    Output of E:\usr\sap\MAU\SYS\exe\run/brconnect -u / -c -f crsyn -o SAPMAU is written to the logfile brconnect.log.
    WARNING 2009-08-23 17:41:32
    Execution of the command "E:\usr\sap\MAU\SYS\exe\run/brconnect -u / -c -f crsyn -o SAPMAU" finished with return code 3. Output: BR0801I BRCONNECT 6.40 (51)BR0252E Function _popen() failed for '( F:\ORACLE\MAU\920\bin\SQLPLUS /nolog ) < \nul 2>&1' at location BrPipeOpen-4BR0253E errno 2: No such file or directoryBR0272E Execution of program '( F:\ORACLE\MAU\920\bin\SQLPLUS /nolog ) < \nul 2>&1' through pipe failedBR0303E Determination of Oracle version failed
    ERROR 2009-08-23 17:41:32
    CJS-00288  Could not update database statistics.<br>DIAGNOSIS: Command brconnect -u / -c -f crsyn -o SAPMAU returned 3.<br>SOLUTION: See brconnect.log for details.

  • FTP Get File List Action Block, It's double listing files!  ver 11.5

    Hi guys.. I have a good one!   I have an FTP Get File List action block in my BLS transaction.  Occasionally, it double lists the files in its output.   For testing I put a repeater with a logevent output where I log the filename, date, and size.  Heres what I saw for my action block output.
    2009-02-13 00:38:00,963  [UserEvent] : File Name: DMM_Export_0010056.txt, File Date 2009-02-13T00:36:00, File Size 339
    2009-02-13 00:38:00,963  [UserEvent] : File Name: DMM_Export_0010056.txt, File Date 2009-02-13T00:36:00, File Size 339
    This is xMII  version 11.5.6 b73  with java 1.4.2_07
    I have a workaround by putting in a distinct action block, after the filelist, but anybody have an idea why this might happen?   My theory is that something might be occuring if the file is being written to while we try to process it, but not sure. 
    I've been building BLS parsers since 2003, (Remember those fun times with Jeremy?)   I've never seen this happen.

    My example is a sample log file before the distinct action.  The general log shows nothing other than the subsequent transaction errors I get as a result of running the same error twice (Tcode return from BAPI calls etc)
    Here is something else interesting..  my userlog file is acting funny, like its trying to write on top of itself.  could it be the transaction is actually running twice or parts of it? 
    For example look at the following log entries
    This is how my log file entry for a production confirmation should look
    2009-02-13 00:38:06,854 [LHScheduler-Int10_NestingWOProdConf] INFO   UserLog - [UserEvent] :
    However sometimes... its looking like this...
    2009-02-13 2009-02-13 00:38:11,854 [LHScheduler-Int10_NestingWOProdConf] INFO   UserLog - [UserEvent] :
    Like it started writing to the log, then started again.
    The problem we are having is that we have JCO calls to SAP in this transaction that does goods movement, we get locking / block errors back from our  saying that we (our sap account) is already updating the information.   Sometimes the information would be posted twice!  You can see how this has become a HUGE issue posting data to a LIVE system twice.
    This is happening on 2 xMII servers.

  • Web photo gallery won't open on internet... why?

    I hope someone can help me with this.
    I always made slideshows with Photoshop until I got Elements. The newest version of Photoshop does not have that option anymore.
    I made a web photo gallery with elements, put it up in Dreamweaver, made a link, tested it on three browsers and it all worked perfectly.
    I then uploaded it all to a server.
    When I go to the page on internet, the link takes me to the index of the photo gallery but the slideshow does not open. It seems like the index file does not open the resource folder.
    I have compared the site on my computer with the site on the server and everything is exactly the same. The photo gallery is on th server with the index file and resource folder.
    I can't understand why the photo gallery doesn't behave on internet like it does on my computer.
    Can anyone explain this to me?
    I really like the way the new photo gallery looks compared with the slideshow made with my older version of Photoshop and would like to use it. Plus the fact that I can control the speed of the slideshow on Elements is a real boon.
    Thanks.

    Thanks for your reply.
    I will try to be as explicate as possible.
    This is a site that I maintain for a marathon. This is the 10th year. The participants contribute a series of photos that I have always made into slide shows.
    I have a page that references slideshows from 2009 for example. I made these photos into slideshows using Photoshop CS3. The different folders are called "Slideshow xxx"  with the corresponding index file. When I link to the index file in a certain folder the slideshow opens in a blank page and wonderful.
    Now, this year, I have done the same thing using Element with the results I mention in my origiinal message: the photo gallery does not open in any browser  in spite of being linked to the index file in the web photo gallery I created with Element.
    What happens when I go to the index file is simply nothing. I do not get an error message.
    You can see for yourself by going to
    http://www.el-ocejon.com/fotos%202010.html
    and clicking on either of the two last options (the first links to a picasa page that is not a problem).
    The solution is simple: create a slideshow using Photoshop CS3 and forget the whole thing.  But I would like to know what the problem is here and why I cannot get the photo gallery created with Elements to work on the internet.
    Or is this a "two steps forward and one step back" issue with Photoshop?

  • Write trace to Citadel

    I'm trying to take spreadsheet data and write it to individual traces inside DSC2012 to a Citadel 5 database. I keep getting an error -1967386570 Data has Back in time timestamp.
    Searching the NI website, back in 2006 there was a way to do this with a vi server.
    http://www.ni.com/white-paper/3485/en
    Is this still possible with the current DSC version?
    From the 2012 DSC help file.
    Writing a Value to a Citadel Trace (DSC Module)
    You can use the Write Trace VI to append a data point to a Citadel trace. Complete the following steps to write a value:
    Add the Write Trace VI on the block diagram.
     Add  Find
    Wire the trace reference output of the Open Trace VI to the trace referenceinput of the Write Trace VI.
    Wire the value and timestamp inputs of the Write Trace VI. Leave the timestamp input unwired to use the current time. The Write Trace VI fails if the timestamp input is earlier than the timestamp of the last point written to the trace. You can determine the timestamp of the last point in the trace using the Get Trace Info VI.
    So, is it no longer possible to write old data into Citadel traces?
    I also saw some posts about a registry key for Citadel 5 about server timestamps, but I don't see a registry key where that note says it should be located.
    Logging Back-in-Time
    Most data logging systems generate ever-increasing time stamps. However, if you manually set the system clock back-in-time, or if an automatic time synchronization service resets the system clock during logging, a back-in-time data point might be logged. Citadel handles this case in two ways.
    When a point is logged back-in-time, Citadel checks to see if the difference between the point time stamp and the last time stamp in the trace is less than the larger of the global back-in-time tolerance and the time precision of the subtrace. If the time is within the tolerance, Citadel ignores the difference and logs the point using the last time stamp in the trace. For example, the Shared Variable Engine in LabVIEW 8.0 and later uses a tolerance level of 10 seconds. Thus, if the system clock is set backwards up to ten seconds from the previous time stamp, a value is logged in the database on a data change, but the time stamp is set equal to the previous logged point. If the time is set backwards farther than 10 seconds, Citadel creates a new subtrace and begins logging from that time stamp.
    Beginning with LabVIEW DSC 8.0, you can define a global back-in-time tolerance in the system registry. Earlier versions of DSC or Lookout always log back-in-time points. Use the backInTimeToleranceMS key located in the HKLM\SOFTWARE\National Instruments\Citadel\5.0 directory. Specify this value in milliseconds. The default value is 0, which indicates no global tolerance.
    This key doesn't exist on my system.
    This link from July 2012 seems to mention that it is still possible to use custom timestamps.
    http://www.ni.com/white-paper/6579/en
    Citadel Writing API
    The DSC Module 8.0 and later include an API for writing data directly to a Citadel trace. This API is useful to perform the following operations:
    · Implement a data redundancy system for LabVIEW Real-Time targets.
    · Record data in a Citadel trace faster than can be achieved with a shared variable.
    · Write trace data using custom time stamps.
    The Citadel writing API inserts trace data point-by-point with either user-specified or server-generated time stamps.
    Is there some more documentation out there that explains this process a bit better?

    Hi unclebump,
    I have been trying to determine what the best course of action would be and I think you need to move the data to a new trace. What I am thinking is for you to open a reference to the trace as it currently exists. Then you will need to read in all the data of that trace. While you read that trace you should also be reading in the data from your file. Once you have both sets of data you will need to iterate over all the data and merge the two sets of data based off their timestamps. The VIs to accomplish this should all exist in the DSC Palette >> Historical or DSC >> Historical >> Database Writing. There is a writing example in the example finder that is called Database Direct Write Demo that would probably be worth looking at. The write trace help says, "
    This VI returns an error if you try to write a point with a timestamp that is earlier than the timestamp of the last point written to the trace." which means that if your data is merged and written in order you should not get this error.
    Hope this helps and let me know if you have any questions.
    Patrick H | National Instruments | Software Engineer

  • Use of Sequence in Match-Merge

    Dear All,
    I have a source table containing name and address data. I want to do householding, which means I need to split the data using a match-merge operator. The address data is the merged output, and the name data is the cross-reference output. All well and good so far.
    I want eventually to load the outputs into two levels of a Customer dimension (although for testing right now, I'm just using two ordinary target tables). I know that I will need a 'business key' to link the merged and the cross-reference outputs (as the two dimensional levels). I decide that I shall create a column in both outputs called HH_SEQ, and load this Household Sequence number from a Sequence (which I prepared earlier).
    In Merge Rules, for the Attribute HH_SEQ, I use a Rule Type of Sequence, and select the appropriate Sequence from the drop-down list box. I map this attribute in the MERGE outgroup to the appropriate column in the target merged table, and MM_HH_SEQ in the XREF outgroup to the appropriate column in the target cross-reference table.
    Here is the issue: to get HH_SEQ to be available as a potential Output Attribute in either Merge Output or Cross-Reference Output, it must first be present as a Source Attribute. Which means I had to add it to Input Attributes first. But of course it is not on my input source table! And thus I get a "VLD-1000 Input parameter INGRP1.HH_SEQ does not have input" warning.
    Either I'm just not understanding how to use the Sequence rule type properly (the documentation is pretty atrocious, mind), or there is some technique required here that is just not apparent from several days of experimentation.
    All help gratefully received!
    Regards,
    Donna
    p.s. it does actually compile and run just fine!

    Hi Donna
    I think you are using it in the correct way! I've seen other examples like this from the MM creators.
    The warning can be ignored, I think it is there to highlight a potential issue if it was being used for a scenario other than what you are doing.
    Cheers
    David

  • Java system copy problems with the Java Migration Toolkit

    During the JAVA system copy, we have have got the error in second last
    step i.e;Java Migration Toolkit. We tried to do the system copy twice
    but every time we have got the below error.
    Is this a know bug? I have got the OSS note 966752 but the value
    of 'src.ci.host' is already correct. has anybody come across the similar issue. please help.
    " CJSlibModule::writeInfo_impl()
    Output of /usr/java14_64/bin/java -
    classpath /usr/sap/CMT/SYS/global/sltools/sharedlib/launcher.jar:. -Xj9
    com.sap.engine.offline.OfflineToolStart
    com.sap.sdt.jmt.migrationtool.MigrationToolImport /usr/sap/CMT/SYS/global/sltools/jmt:/usr/sap/CMT/SYS/global/sltools/sharedlib import.switch
    resume downtime mti.properties is written to the
    logfile /home/cmtadm/CI_Import_08312009/runJmt.log.
    WARNING 2009-09-01 20:49:43.473
    CJSlibModule::writeWarning_impl()
    Execution of the command "/usr/java14_64/bin/java -
    classpath /usr/sap/CMT/SYS/global/sltools/sharedlib/launcher.jar:. -Xj9
    com.sap.engine.offline.OfflineToolStart
    com.sap.sdt.jmt.migrationtool.MigrationToolImport /usr/sap/CMT/SYS/global/sltools/jmt:/usr/sap/CMT/SYS/global/sltools/sharedlib import.switch
    resume downtime mti.properties" finished with return code 27. Output:
    ERROR 2009-09-01 20:49:43.477
    CJSlibModule::writeError_impl()
    CJS-20068 Error when running the J2EE Migration Toolkit.<br>SOLUTION:
    See output of logfile /home/cmtadm/CI_Import_08312009/runJmt.log: ''.
    ERROR 2009-09-01 20:49:43.479
    CJSlibModule::writeError_impl()
    CJS-20068 Error when running the J2EE Migration Toolkit.<br>SOLUTION:
    See output of logfile /home/cmtadm/CI_Import_08312009/runJmt.log: ''.
    ERROR 2009-09-01 20:49:43.480 [iaxxgenimp.cpp:731]
    showDialog()
    FCO-00011 The step RunMigrationController with step key
    |NW_Doublestack_OneHost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|1|0|NW_CI_Instance|ind|ind|ind|ind|11|0|NW_CI_Instance_Configure_Java|ind|ind|ind|ind|4|0|NW_RUN_MIGRATION_CONTROLLER|ind|ind|ind|ind|2|0|RunMigrationController was executed with status ERROR ."
    There is no log file by the name runJmt.log.
    Regards,
    Sagar

    This issue was resolved by OSS note 966752 itself. Below is the important line in the OSS note.
    Note: The default value is crucial! If a custom value other than the default value was applied, we highly recommend that you restore the default value.
    Thanks,
    Sagar

  • Customer Creation - Want to pass custom field data to KNA1(Customer Master)

    Hi,
    I have created BAPI for customer creation which uses standard program RFBIDE00 for customer creation.
    I have appended new custom field(ZZATTR1) in the table KNA1 and its currosponding stucture used in the standard program BKNA1.
    Import parameters of my BAPI is one structure which is having this custom field, on executing this BAPI it creates the customer but data which I give through this field is not getting transfered in the table KNA1.
    This sstandard progaram uses file to containing customer information, In the BAPI I on the data from the structure to the file and further it gets submitted to the program.
    My requirement is to pass the data of teh custom field which I have created, How to achieve this ?
    Pls Suggest.
    Regards
    Vishal

    Hi Vishal,
    When implementing a customer exit (i.e. field, user etc) have you ever wanted to be able to read additional
    field values available in the main SAP program but not available in your exit. Now with the use of
    field-symbols you can have access to any field available in the main SAP program. The code below
    demonstrates how this is done within a field exit on BSTNR which can be tested using transaction
    ME21. Information on creating field exits can be found here. 
    FUNCTION field_exit_bstnr.
    ""Local interface:
    *"  IMPORTING
    *"     REFERENCE(INPUT)
    *"  EXPORTING
    *"     REFERENCE(OUTPUT)
    create field symbol
      FIELD-SYMBOLS: <status>.
    Assign value of variable from calling prog to field symbol
      ASSIGN ('(SAPMM06E)RM06E-BSART') TO <status>.
    Display value retrieved in message
    Note: Messages of type i and w are not allowed
      IF sy-subrc = 0.
        MESSAGE e003(zr) WITH <status> 'kkk'.
      ENDIF.
    ENDFUNCTION.
    Rewords some point if it is helpful.
    Rgds,
    P.Naganjana Reddy

  • Reading Tables in MS WORD using LabView & ActiveX

    I would like to be able to read tables contained in an MS WORD document using LabView and ActiveX. (I've been successful in doing this with Excel files.  ( I am not using the report generation toolkit.)  After searching numerous posts with similar questions about LV & ActiveX  I've not found any solution to my specific app.  In LabView I'm successful doing an "Automation Open", followed by a Property Node, to an Invoke Node ("Documents" Open) and then a Property Node with "TABLES" selected. That's where I get stuck. Now that I have a reference (?) to "Tables" I can't seem to figure out what I need to connect and follow that with in order to now read the Tables in my document.  I guess that's due to my lack of understanding as to what exactly this "Table" property node output actually is.  Can someone tell me what this reference output is referring to and how I can get to the Tables within the document from it?  I've tried connecting both a property node and invoke node to no avail.  Can someone point me in the right direction?
    Thanks. Ed

    Dear Ed,
    Figuring out "what I need to connect and follow that with" relates to the organization of the MS Word Object Model.  Please refer to the Word Object Model Reference MSDN >> MSDN Library >> Office Development >> 2007 Microsoft Office System >> Word 2007 >> Word 2007 Developer Reference >> Word Object Model Reference.  You can think of the Object Model as a software hierarchy.  You must start at the top of the hierarchy and work your way down to the item of interest.  You must descend the hierarchy one level at a time; you cannot skip levels.
    At each level of the hierarchy, you have access to the methods and properties of that "level".
    Click on the Tables Collection link from the Word Object Model Reference page.  Notice how Tables is at the top of the hierarchy.  When you have a LabVIEW reference to a Tables object (technically speaking a collection), you then have access to the methods and properties supported by this "level".  Click on the Tables Object Members link.  Here you will see that the Tables collection supports two methods, Add and Item, and also supports five properties, Application, Count, Creator, NestingLevel, and Parent.
    Notice how Mike gets the Count property of the Tables object.  He then uses the Item method to return a reference to an individual table.  The key thing to take away from this is how Mike uses the Item output node to gain access to the methods and properties of the Table object (not the Tables collection).  Mike has descended the hierarchy from Tables to Table.  He descends the hierarchy again with the Cell method and Range property of the Table and Cell object respectively.  Lastly, the Text property of the Range object, "Returns or sets the text in the specified range or selection."
    Effective LabVIEW programming with this type of ActiveX control requires a fair understanding of how Microsoft organized its software libraries.  The Object Model Reference is a good place to start visualizing how these libraries are organized.  Thank you for using the forums Ed, and thanks for your post, Mike.

  • Error in Select-Option

    Hi Experts,
    In my report i need to declare a select-option . This select-option is a combination of Period and fiscal year i..e the values in the select-option should be like
    <b>PPP/YYYY</b> period and year need to be separated by <b>"/".</b>
    I had declared in that format. <b>Here the period starts from 001-012</b>.So when ever the User enters <b>010/2005</b> in low and <b>007/2007</b> in high it is giving an error that <b>Lower limit is greater than Higher limit</b> .

    Technically, I think you could actually achive this with the use of a custom data domain and on that domain refer to a custom conversion exit...this would mean that the value typed by the user in "PPP/YYYY" format could be converted immediately into "YYYYPPP" or whatever internally before screen validation kicks in... e.g. (logic below needs more work!!):
    function conversion_exit_zppyy_input.
    *"*"Local Interface:
    *"  IMPORTING
    *"     REFERENCE(INPUT)
    *"  EXPORTING
    *"     REFERENCE(OUTPUT)
    * Data should be in form PPP/YYYY
      data:
        l_ppp(3)            type c,
        l_yyyy(4)           type c.
      output = input.
    *" add some validation here...
      split output at '/'
        into l_ppp l_yyyy.
    *" add more validation here
      if sy-subrc is initial.
    *" make it into YYYYPPP format
        concatenate l_yyyy l_ppp into output.
      endif.
    endfunction.
    You'd probably need an equivalent 'conversion_exit_zppyy_output" function too.
    The question remaining in my mind is what you are going to do with the values once you have them in the range table...?  And what if user inputs wildcards etc?
    Jonathan

  • Read/Write file in Unicode (UTF-16)

    Hi, I have some issue to write a file in Unicode (UTF-16)
    I have to read a file with LabView, change some parameters and write the new data into the same file. The file uses Unicode UTF-16.
    I downloaded some library here: https://decibel.ni.com/content/docs/DOC-10153
    I can read the file, convert the data to UNI/ASCII/UNI, and then write the file. But when I open the new file with an editor like Notepad++ there are some unexpected characters at the end of the line.
    Even reading the file and writing exactly the same data doesn'not work.
    I attached an example.
    Thanks for you kind support.
    Solved!
    Go to Solution.
    Attachments:
    TEST.zip ‏7 KB

    Right-click on your Read and Write Text File functions.  There is an option to "Convert End Of Line".  Turn that off on both functions.
    As a side not, you don't need the Close File functions.  The Read and Write Text File functions will close the file if the file reference output is not wired.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

Maybe you are looking for