Wanna compare data in excel with real time signal

Hi,
 my application is to check the PCB i/o pins... first i have to corresponding signal to every pin in GOOD PCB board and the output of each pin should saved in excel with the help of write to spreadsheet.. then the database be act as master for other PCB boards..
If i wanna check the other pcb board means the same corresponding signals should be send to the each pin of the PCB board.. the Output received from the board should check with the data already saved in the database if it is match's means Status LED should Glow otherwise its not to glow...
pls help me to finish this application.. i have attach model vi with post and excel too....
Regards,
N. Srinivasan
Solved!
Go to Solution.
Attachments:
PCB_COMPARING.vi ‏184 KB
DATABASE.xls ‏2 KB

The "method" you have in your last post doesn't come anywhere near to what you say you want to do.
It creates a simulated sinewave and writes it to a text file.  Waits 1/10th of a second, reads it back from that text file, does a lot of manipulation to that data to convert it to a dynamic waveform type.  Then compares it to another simulated sinewave to say if they are equal or not.
"My doubt is if all the received signal are saved in the excel sheet first..".  I don't know why you would doubt that this would happen.  If you receive a signal and write it to the text file, then why don't you believe it would be written to a text file?  By the way, the Read from Spreadsheet File and Write to Spreadsheet File functions have nothing to do with Excel.  They read or write data to a text file that happens to be formatted in a way (delimiters between data in a row, end of line characters between rows) that is a text file version of a spreadsheet.  Excel has the ability to import from or export to a text file, but that file is NOT an Excel file.
You have major portions of your code missing and questions to be answered.
Where are you actually outputing a signal to the printed circuit board?
Where are you actually receiving a signal from the printed circuit board?
Are you receiving one signal or 40 signals?
What type of DAQ board are you using to do this?
Okay, let's say once you have the first portion of your code set up to do that, you can write the data to a text file.
Now in another part of your code, you basically repeat the procedure but instead of writing to a text file, you read the data from your "good board" text file and compare the two.
How do you compare the signals and determine whether the newly read signal is good or bad?  Right now, you are doing a comparison for equality.  Do you think the "good" signal and the newly test signal will ever be equal?  Even if you just repeat the test on the good board, do you think those signals will be equal?  NO.  It is impossible for the signals to ever be equal, even the slightest variation in the signal is going to make them unequal.  You are never going to measure identical real world signals identically.  You need to come up with some comparison algorithm with criteria that determines whether another good signal actually matches your saved good signal, and a bad signal does not match the good signal taking into account the variations you inevitably get when measuring real world signals.
Those are big questions you need to answer.  Assuming you work all that out, on paper in a design specification document, only then can you start worrying about the LabVIEW program and its architecture.  Ultimately, a state machine architecture will be your likely choice as you have multiple states to work with.  Initialize, test board, save data to file, read data from file, compare data, possibly more states such as wait, close program, multiple states within the test board process as it is probably a multiple step process.

Similar Messages

  • RMAN in active data guard with real time apply ON without a catalog DB

    I would like to know whether I can take RMAN backups (both level 0 & incremental) in active data guard with real time apply ON without a catalog database. version is 11.2.0.3.2. Will it make difference in the control file of standby syncing with primary? Can I use these backups on primary to restore and recover the database(including control file backup)?

    Thank you very much msberg. Could you please answer the below queries as well?
    Should I cancel Real Time Apply before starting RMAN?
    Will it make difference in the control file of standby syncing with primary? Is it possible to use these backups on primary to restore and recover the database(including control file backup)? If possible, how will I do that?

  • B2MML with real time data

    Hi All
             I want some B2mml files with real time data.
    So exact( maintenance infromation and Production) B2MML that is coming from MIS.
    So if anyone have some information regarding this please tell me.
    Thanks in Advance
    Ramshanker

    Hi, Ramshanker.
    In the cases where the plant system is directly delivering B2MML documents (such as with certain products from Wonderware, Siemens, and some others), using the data in xMII is quite straightforward.
    1) If the B2MML message is not "standard" B2MML (be aware that very few plant vendors use the baseline standard - most have a number of specific extensions), you can add the vendors' schemas or B2MML example documents to xMII by adding them to a ReferenceDocument folder on the server.
    2) In xMII BLS, create a transaction (let's assume you called it B2MMLHandler) with one input parameter, of type "XML" (let's assume you called it "B2MMLDocument").  Add a "ReferenceSchema" or "ReferenceDocument" action, and configure it to load the vendor's B2MML format.  Use the "assign reference document" menu option to assign this format to the transaction input.
    3) You can now use the B2MML document to perform analytics or to map from the B2MML document into any RFC or BAPI calls or ESA service calls from within BLS.
    4) The plant system system should "post" the XML data to an xMII transaction using the "Runner" URL as in:
    http://<yourserver>/Lighthammer/Runner?Transaction=B2MMLHandler&InputParameter=B2MMLDocument
    Of course, if the MES system isn't capable of posting the documents and they are available on the file system, you can use the xMII file actions (or remote file access via FTP) to access and load the B2MML documents, perhaps at a scheduled interval.
    Hope this helps get you started in the right direction.
    Best regards,
    Rick Bullotta

  • Upload data from excel with vba by calling a function module

    Hello all,
    i have a problem with the function module "ALSM_EXCEL_TO_INTERNAL_TABLE". I will call this function module with vba to load data from excel to sap with a Buttonclick. I have copied this function module and set it remotable. But i can´t call it from excel.
    Can you give me some tips how can i
    upload data from excel with vba by click a button.
    The problem seems the function: call method cl_gui_frontend_services=>clipboard_import in the function module, because when i comment this function call the vba-call is true but no results. 
    How can I call the function module correct with vba?
    Thanks a lot for your tips!!!!
    Chris
    Message was edited by:
            Christoph Kirschner

    HI
    Uploading data directly from Excel file format
    * Upload data direct from excel.xls file to SAP
    REPORT ZEXCELUPLOAD.
    PARAMETERS: filename LIKE rlgrap-filename MEMORY ID M01,
                begcol TYPE i DEFAULT 1 NO-DISPLAY,
                begrow TYPE i DEFAULT 1 NO-DISPLAY,
                endcol TYPE i DEFAULT 100 NO-DISPLAY,
                endrow TYPE i DEFAULT 32000 NO-DISPLAY.
    * Tick don't append header
    PARAMETERS: kzheader AS CHECKBOX.
    DATA: BEGIN OF intern OCCURS 0.
            INCLUDE STRUCTURE  alsmex_tabline.
    DATA: END OF intern.
    DATA: BEGIN OF intern1 OCCURS 0.
            INCLUDE STRUCTURE  alsmex_tabline.
    DATA: END OF intern1.
    DATA: BEGIN OF t_col OCCURS 0,
           col LIKE alsmex_tabline-col,
           size TYPE i.
    DATA: END OF t_col.
    DATA: zwlen TYPE i,
          zwlines TYPE i.
    DATA: BEGIN OF fieldnames OCCURS 3,
            title(60),
            table(6),
            field(10),
            kz(1),
          END OF fieldnames.
    * No of columns
    DATA: BEGIN OF data_tab OCCURS 0,
           value_0001(50),
           value_0002(50),
           value_0003(50),
           value_0004(50),
           value_0005(50),
           value_0006(50),
           value_0007(50),
           value_0008(50),
           value_0009(50),
           value_0010(50),
           value_0011(50),
           value_0012(50),
           value_0013(50),
           value_0014(50),
           value_0015(50),
           value_0016(50),
           value_0017(50),
           value_0018(50),
           value_0019(50),
           value_0020(50),
           value_0021(50),
           value_0022(50),
           value_0023(50),
           value_0024(50),
           value_0025(50),
           value_0026(50),
           value_0027(50),
           value_0028(50),
           value_0029(50),
           value_0030(50),
           value_0031(50),
           value_0032(50),
           value_0033(50),
           value_0034(50),
           value_0035(50),
           value_0036(50),
           value_0037(50),
           value_0038(50),
           value_0039(50),
           value_0040(50),
           value_0041(50),
           value_0042(50),
           value_0043(50),
           value_0044(50),
           value_0045(50),
           value_0046(50),
           value_0047(50),
           value_0048(50),
           value_0049(50),
           value_0050(50),
           value_0051(50),
           value_0052(50),
           value_0053(50),
           value_0054(50),
           value_0055(50),
           value_0056(50),
           value_0057(50),
           value_0058(50),
           value_0059(50),
           value_0060(50),
           value_0061(50),
           value_0062(50),
           value_0063(50),
           value_0064(50),
           value_0065(50),
           value_0066(50),
           value_0067(50),
           value_0068(50),
           value_0069(50),
           value_0070(50),
           value_0071(50),
           value_0072(50),
           value_0073(50),
           value_0074(50),
           value_0075(50),
           value_0076(50),
           value_0077(50),
           value_0078(50),
           value_0079(50),
           value_0080(50),
           value_0081(50),
           value_0082(50),
           value_0083(50),
           value_0084(50),
           value_0085(50),
           value_0086(50),
           value_0087(50),
           value_0088(50),
           value_0089(50),
           value_0090(50),
           value_0091(50),
           value_0092(50),
           value_0093(50),
           value_0094(50),
           value_0095(50),
           value_0096(50),
           value_0097(50),
           value_0098(50),
           value_0099(50),
           value_0100(50).
    DATA: END OF data_tab.
    DATA: tind(4) TYPE n.
    DATA: zwfeld(19).
    FIELD-SYMBOLS: <fs1>.
    AT SELECTION-SCREEN ON VALUE-REQUEST FOR filename.
      CALL FUNCTION 'KD_GET_FILENAME_ON_F4'
           EXPORTING
                mask      = '*.xls'
                static    = 'X'
           CHANGING
                file_name = filename.
    START-OF-SELECTION.
      CALL FUNCTION 'ALSM_EXCEL_TO_INTERNAL_TABLE'
           EXPORTING
                filename                = filename
                i_begin_col             = begcol
                i_begin_row             = begrow
                i_end_col               = endcol
                i_end_row               = endrow
           TABLES
                intern                  = intern
           EXCEPTIONS
                inconsistent_parameters = 1
                upload_ole              = 2
                OTHERS                  = 3.
      IF sy-subrc <> 0.
        WRITE:/ 'Upload Error ', SY-SUBRC.
      ENDIF.
    END-OF-SELECTION.
      LOOP AT intern.
        intern1 = intern.
        CLEAR intern1-row.
        APPEND intern1.
      ENDLOOP.
      SORT intern1 BY col.
      LOOP AT intern1.
        AT NEW col.
          t_col-col = intern1-col.
          APPEND t_col.
        ENDAT.
        zwlen = strlen( intern1-value ).
        READ TABLE t_col WITH KEY col = intern1-col.
        IF sy-subrc EQ 0.
          IF zwlen > t_col-size.
            t_col-size = zwlen.
    *                          Internal Table, Current Row Index
            MODIFY t_col INDEX sy-tabix.
          ENDIF.
        ENDIF.
      ENDLOOP.
      DESCRIBE TABLE t_col LINES zwlines.
      SORT intern BY row col.
      IF kzheader = 'X'.
        LOOP AT intern.
          fieldnames-title = intern-value.
          APPEND fieldnames.
          AT END OF row.
            EXIT.
          ENDAT.
        ENDLOOP.
      ELSE.
        DO zwlines TIMES.
          WRITE sy-index TO fieldnames-title.
          APPEND fieldnames.
        ENDDO.
      ENDIF.
      SORT intern BY row col.
      LOOP AT intern.
        IF kzheader = 'X'
        AND intern-row = 1.
          CONTINUE.
        ENDIF.
        tind = intern-col.
        CONCATENATE 'DATA_TAB-VALUE_' tind INTO zwfeld.
        ASSIGN (zwfeld) TO <fs1>.
        <fs1> = intern-value.
        AT END OF row.
          APPEND data_tab.
          CLEAR data_tab.
        ENDAT.
      ENDLOOP.
      CALL FUNCTION 'DISPLAY_BASIC_LIST'
           EXPORTING
                file_name     = filename
           TABLES
                data_tab      = data_tab
                fieldname_tab = fieldnames.
    *-- End of Program
    <b>Excel Upload Alternative - KCD_EXCEL_OLE_TO_INT_CONVERT</b>
    *Title : Excel Uploading
    TYPES:   BEGIN OF t_datatab,
             col1(25)  TYPE c,
             col2(30)  TYPE c,
             col3(30)  TYPE c,
             col4(30)  TYPE c,
             col5(30)  TYPE c,
             col6(30)  TYPE c,
             col7(30) TYPE c,
             col8(30)  TYPE c,
             col9(30)  TYPE c,
             col10(30)  TYPE c,
             col11(30)    TYPE c,
           END OF t_datatab.
    DATA: it_datatab TYPE STANDARD TABLE OF t_datatab INITIAL SIZE 0,
          wa_datatab TYPE t_datatab.
    Data : p_table type t_datatab occurs 0 with header line.
    DATA : gd_scol   TYPE i VALUE '1',
           gd_srow   TYPE i VALUE '1',
           gd_ecol   TYPE i VALUE '256',
           gd_erow   TYPE i VALUE '65536'.
    DATA: it_tab TYPE filetable,
          gd_subrc TYPE i.
    field-symbols : <fs>.
    *Selection screen definition
    SELECTION-SCREEN BEGIN OF BLOCK b1 WITH FRAME TITLE text-001.
    PARAMETERS:  p_file LIKE rlgrap-filename
                   DEFAULT 'c:test.xls' OBLIGATORY.   " File Name
    SELECTION-SCREEN END OF BLOCK b1.
    AT SELECTION-SCREEN ON VALUE-REQUEST FOR p_file.
      REFRESH: it_tab.
      CALL METHOD cl_gui_frontend_services=>file_open_dialog
        EXPORTING
          window_title     = 'Select File'
          default_filename = '*.xls'
          multiselection   = ' '
        CHANGING
          file_table       = it_tab
          rc               = gd_subrc.
      LOOP AT it_tab INTO p_file.
    *    so_fpath-sign = 'I'.
    *    so_fpath-option = 'EQ'.
    *    append so_fpath.
      ENDLOOP.
    START-OF-SELECTION.
      PERFORM upload_excel_file TABLES   it_datatab
                                 USING   p_file
                                         gd_scol
                                         gd_srow
                                         gd_ecol
                                         gd_erow.
    * END-OF-SELECTION.
    END-OF-SELECTION.
      LOOP AT it_datatab INTO wa_datatab.
        WRITE:/ wa_datatab-col1,
                wa_datatab-col2,
                wa_datatab-col3,
                wa_datatab-col4,
                wa_datatab-col5,
                wa_datatab-col6,
                wa_datatab-col7,
                wa_datatab-col8,
                wa_datatab-col9,
                wa_datatab-col10,
                wa_datatab-col11.
      ENDLOOP.
    *&      Form  UPLOAD_EXCEL_FILE
    *       upload excel spreadsheet into internal table
    *      -->P_TABLE    Table to return excel data into
    *      -->P_FILE     file name and path
    *      -->P_SCOL     start column
    *      -->P_SROW     start row
    *      -->P_ECOL     end column
    *      -->P_EROW     end row
    FORM upload_excel_file TABLES   p_table
                           USING    p_file
                                    p_scol
                                    p_srow
                                    p_ecol
                                    p_erow.
      DATA : lt_intern TYPE  kcde_cells OCCURS 0 WITH HEADER LINE.
    * Has the following format:
    *             Row number   | Colum Number   |   Value
    *      i.e.     1                 1             Name1
    *               2                 1             Joe
      DATA : ld_index TYPE i.
    * Note: Alternative function module - 'ALSM_EXCEL_TO_INTERNAL_TABLE'
      CALL FUNCTION 'KCD_EXCEL_OLE_TO_INT_CONVERT'
        EXPORTING
          filename                = p_file
          i_begin_col             = p_scol
          i_begin_row             = p_srow
          i_end_col               = p_ecol
          i_end_row               = p_erow
        TABLES
          intern                  = LT_INTERN
        EXCEPTIONS
          inconsistent_parameters = 1
          upload_ole              = 2
          OTHERS                  = 3.
      IF sy-subrc <> 0.
        FORMAT COLOR COL_BACKGROUND INTENSIFIED.
        WRITE:/ 'Error Uploading file'.
        EXIT.
      ENDIF.
      IF lt_intern[] IS INITIAL.
        FORMAT COLOR COL_BACKGROUND INTENSIFIED.
        WRITE:/ 'No Data Uploaded'.
        EXIT.
      ELSE.
        SORT lt_intern BY row col.
        LOOP AT lt_intern.
         MOVE lt_intern-col TO ld_index.
         assign component ld_index of structure
         p_table to <fs>.
    move : lt_intern-value to <fs>.
    *     MOVE lt_intern-value TO p_table.
          AT END OF row.
            APPEND p_table.
            CLEAR p_table.
          ENDAT.
        ENDLOOP.
      ENDIF.
    ENDFORM.                    "UPLOAD_EXCEL_FILE
    Regards
    Pavan

  • RAC with real time knowledge

    Hi, I'm interested in learning RAC and dataguard with real time knowledge......... please help me finding some training institution or peoples with real time knowledge
    Thanks

    The term "+real-time+" has a very specific meaning in Information Technology. See http://en.wikipedia.org/wiki/Real-time_computing for the basic details.
    Oracle RAC is not a real-time database.. and nor does the "real-time knowledge" make much sense IMO...
    Always difficult to get to grips with the nuances of the English language where it is your 2nd or even 3rd language. But it is important that we speak the same language and understand the same terms and concepts, in Information Technology.

  • [Solved] QJackCtl won't start in with real time.

    Hey all,
    I'm having a problem with QJackCtl, I got my version from Extra. The problem is that for some reason Jack won't start in real time mode using QJackCtl. Real Time is checked and marked under the set-up tab, but that little piece of info doesn't seem to make it to jackd. If I start Jack through CLI it works with real time, but not from the GUI.
    I figured this out by starting JACK from the command line and then starting QJack. When I would do that it would be running in real time just without the rest of my settings (buffer,periods,sampling rate, etc).  So in my head there's an issue with QJackCtl. I've tried to find a solution, but I can't seem to find one. This isn't the first time I've set JACK up on a system, but this is the first time I've had this problem.
    The rest of the system is set up and ready to go. Rt Kernel, limits.conf and all other tuning is done.
    Is there something I missed when setting this piece of software up? Any ideas?
    Solved Edit:
    Okay, so right after I posted this I got a crazy idea. When you enter setup in qjackctl on the first page there's a box for server path. The path defaults to "jackd". On that line I added "--realtime", so it looked like a basic CLI start-up of JACK and now it works!
    So, if anyone else has this problem go to the Setup tab, in the setup window on the line that says Server Path it should look like this:
    jackd --realtime
    It worked for me and I hope it helps somebody else out.
    Last edited by Zen:Core (2010-04-03 21:41:52)

    Which volume? Assuming you're on the most common usecase (single hw audio device) then whatever command line you use (amixer for example) it'll work fine, it'll control the volume of the device rather than the pulseaudio plugin (so other playing apps get affected as well, which is probably what you want). The 'global' volume so to speak.
    For multiple hw audio devices it gets tricky though.

  • MDX to get last date MEMBER associated with a time LEVEL

    Hi there,
    I need to come up with a MDX calculation in my cube that takes an average 90 days forward from the last date member associated with whatever time level is being viewed in a time hierarchy.  Is this something anyone could help me with?  I can come
    up with code that is specific to a level (Year, Month, Week) but not one that is dynamic.
    Best,
    D

    I don't know exactly what you want from me.  I have a Time dimension with Year, Semester, Quarter, Month and Date.  I want to be able to get to the last date associated with a time member. So, for example, I would want to derive like this...
    October 31, 2014 would get to October 31, 2014
    October 2014 would get me to October 31, 2014
    Quarter 3 would get to to September 30, 2014
    Semester 2, 2014 would get me to December 31, 2014
    2014 would get me to  December 31, 2014

  • I have a PXI-8109 with Real Time installed. That displays USER1 indicator?

    HI all,
    I have a PXI-8109 with Real Time installed.
    That displays USER1 indicator LED on the panel?
                                                      Thanks
    Solved!
    Go to Solution.

    Thank you.
    The NI PXI-8109 front panel has 4 LEDs.
    (Figure 3-1. NI PXI-8109 Front Panel Layout and Dimensions)
    - PWR OK/FAULT LED
    - DRIVE LED
    - USER1 LED
    - USER2 LED
    The NI PXI-8109 User Manual describes only 2 LEDS:
     - PWR OK/FAULT LED
    - DRIVE LED
    What do USER LEDs on the NI PXI-8109 front panel mean?

  • Uploading files to SAP using HTML5 /AJAX/Gateway media links with real-time progress bar

    Hi All,
    Check out the blog post 'Uploading files to SAP using HTML5 /AJAX/Gateway media links with real-time progress bar'
    Uploading files to SAP using HTML5 /AJAX/Gateway media links with real-time progress bar
    Regards
    Prabaharan

    aplay -l
    **** List of PLAYBACK Hardware Devices ****
    card 0: NVidia [HDA NVidia], device 3: NVIDIA HDMI [NVIDIA HDMI]
    Subdevices: 1/1
    Subdevice #0: subdevice #0
    Which correspondents fine with the
    device "hw:0,3"
    in my mpd.conf alsa section.
    Also sound worked fine withh this config before I started installing ffmpeg-svn and alikes
    hokasch:
    I tried changing the mixer to "Master" but that did not change much.

  • The problem with Real-Time Data Acqusutuin

    Hi , all
    I created ds with Generic Delta : Delta-Specific Fields pointed time stamp and in settings checked Real-time Enabl . After that created modeling in BW , made initial update like discribed in SAP Documentation , created one more infopackage with adapter :Real-time extraction from sap , assigned infopackage and dtp to daemon and started job of daemon .
    Request stay yellow all of time and dosn't bring no one records of changes in R3 . Anybody know what is the couse for that? Maby I forgot to check something anywhere or star trigger anyone!?

    The following setup working fine for me
    In R/3 for Generic Delta --- 
            Filed Name - AEDTM
            0Calday -- Radio Button - 'ON'
            Upper Limit - 0
            Lower Limit - 2
            Real_Time enabl - Check Box - 'ON'
            New Status for changed records - Radio Button - 'ON'
    In BI
        Create infopackage (init) for replicated datasource
        Execute the DTP to load this init data to DSO.
        Do NOT activate DSO
        Now create another Infopack (real_time Enable & ) and create real-time DTP.
    In RSRDA
        Create a Daemon
        Assign InfoPack & DTP to Daemon.
        Now start daemon...Here the message says something like 'one of the request is not activated'.
        Now, select DTP under Daemon --> use menu Edit --> Repair Chain --> Execute
        (If repair chain is not yet there then Daemon --> use menu Edit --> Repair Chain --> Generate and  
         once this is done execute the using the above path). Once the chain finishs
        Now start Daemon and refresh.
        Now u can see, an open request (yellow) under infopackage.
        As soon as the new records gets load into PSA, then an open request starts under DTP.
        (Note: If no new data available in PSA, no open request starts under DTP)
    Also note that, RDA requests (in PSA & DSO Manage) should always be in yellow, that means Real-Time Aquisition is working.

  • Problem with Real Time Data Acquisition

    Hello,
    I tried to built a RDA data flow according to the SAP-Help and this blog (/people/kamaljeet.kharbanda/blog/2006/11/13/real-time-data-acquisition-bi2004s), but I ran into a problem. Everything seems to be fine, but even though the daemon is running no new data is added to the DSO. I'll explain in detail what steps I took and hope someone can pinpoint my error(s).
    System is BW4 on SPS 17. My datasource as well as my DSObject are in the same system
    1.) Created a transparent table with a timestamp-field for generic delta as a local object.
    2.) Created a generic datasource from view/table (used the above table), marked it as realtime-enabled and used timestamp as delta-specific field.
    3.) Replicated the metadata for my datasource and activated it.
    4.) Created a Data Storage Object with infoobjects corresponding to the table fields
    5.) Created a transformation between DataSource and DSO
    6.) Created a DTP from Data Source to DSO. The type is realtime-DTP.
    7.) Created an InfoPackage for Delta-Initial Load. Update mode: Initialize Delta Process -> Initialization with Data Transfer.
    8.) Then I changed my DTP (step 6) to "normal DTP" and executed it. Then activated the data in the DSO. The datasets from the table (which i filled with an ABAP-program) were successfully loaded and activated. So now my DSO contains 5 datasets.
    9.) Changed the DTP back to "realtime DTP"
    10.) Created an InfoPackage for RDA. Adapter is set to "Realtime extraction from SAP System"
    11.) Created a daemon, assigned the IP for RDA and the DTP to the daemon and started it up.
    12.) Executed my ABAP program to added 3 datasets to the table which is used as a datasource.
    So now in my source table there are 8 datasets, but the daemon won't load them. In my RDA-Monitor (trx rsrda) it shows my daemon, the assigned IP and the assigned DTP (status green). Under the IP there's one yellow request (that should be right according to the help), but unlike the picture in the blog (link see above) there's no request under the DTP.
    Job overview (sm37) shows an active batch job and some jobs that are ready (every 10minutes a new batch starts and the old one is closed). the job log shows, that the daemon runs, but loads no data...
    i think it might be a problem with the delta extraction somewhere, because the intial load works fine and there are no errors anywhere.
    Edit: perhaps my settings in the data source unter tab: extraction are wrong?
    Delta process: AIE After Images By Extractor (can't change that)
    Direct Access: Allowd
    real Time: Real-Time DA Supported
    Adapter: realtime data extraction (also tried Access to SAP Data through Service API which seems senseless)
    Data Format: Fixed Lenght
    Anyone can explain what those fileds mean and which one i should use?
    Edited by: Christoph Jender on Apr 10, 2008 2:48 PM
    Edited by: Christoph Jender on Apr 10, 2008 3:33 PM

    Hi again, another thing i'd like to try is to just active a business content datasource which is realtime-enabled and has the ability to dynamically add some values so that i can test if it works. but i have no idea what business content would be applicable here, any ideas on that perhaps?

  • Producer/consumer architecture over network variables with Real-Time target.

    Hi all,
    I am maintaining a producer/consumer data acquisition program to be deployed on a real-time target. The main code is deployed to and run on the real-time target during experiments, but was having trouble because the program was originally designed to write all experimental data to disk on the real time during acquisition, which puts the whole experiment at the mercy of the hard drive. I am now trying to rework the code so that the host takes care of logging, so that my time-critical loops don't have to wait.
    I am currently using LabVIEW 8.5
    I have two questions:
    First, how can I programmatically call the data-logging subvi on the host so that it runs in parallel with the main vi which runs the experiment and collects data on the real-time? I have attached the test code that I have been working with to figure this out, but it does not run the logging vi continuously in the background. I am aware that there is better functionality for this in newer versions of LabVIEW, but I would prefer not to upgrade unless there is no other option. I would like to be able to run my data-generating vi and have it start the data logging remotely.
    Second, is there a way in the host VI to read values off the network variable using an event structure rather than polling it for updates?
    Any help would be sincerely appreciated!
    Attachments:
    testRemoteLogging.zip ‏124 KB

    VI server
    Mark the target VI as served on the machine on which it will execute and use VI server Call by reference to invoke the served VI.
    This used to be taught as THE way to communicate syncronously with an RT app.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • About DTP with real-time access.

    Hello Gurus,
           setting for error handling only has an impact while repairing a DTP request (repair) and during the conversion of the DTP to standard DTP (for example, to correct an error during extraction).
          will you please give a simpel scenario to explain above words?
    thank you very much.
    Fran

    Hi,
    Use :
    With SAP NetWeaver 7.0, SPS 14, the following changes and enhancements are available for real-time data acquisition:
    ●      Changes to the menu and the context menu of the monitor for real-time data acquisition
    The menu and the context menus for the individual objects in the monitor for real-time data acquisition have been standardized and enhanced. In particular, it is now possible to assign daemons and data transfer processes on various levels using context menu entries. For a complete overview of the functions in the menu and in the context menus, see Monitor for Real-Time Data Acquisition.
    ●      Assignment of daemons for InfoPackages and data transfer processes (DTP)
    To assign InfoPackages and data transfer processes to a daemon, you can call the monitor for real-time data acquisition in the following ways:
    ○       In the Data Warehousing Workbench using the respective context menu entry Assign RDA Daemon.
    ○       In InfoPackage maintenance using the Schedule tab page, and in data transfer process maintenance using the Execute tab page.
    The button names for jumping to the monitor for real-time data acquisition have changed here. To jump to the monitor, choose Assign Daemon.
    ●      Repair process chains for repairing a broken data transfer or a missing delta in the DataStore object
    In certain situations, a gap in the delta of the DataStore object can occur if there is a closed request in the PSA but there is no corresponding request in the DataStore object. For example, this is the case if a DTP request has terminated due to an error in the transformation. Here, you can create a repair process chain to repair a missing or broken update from the PSA. The repair process chain contains a standard DTP as well as any further processes required for subsequent processing (such as activating the data in the DataStore object or subsequent process chains). When the repair process chain is executed, the complete delta is loaded from the source into the DataStore object and processes for activation and further processing are executed if required.
    ●      Process types for starting and stopping real-time data acquisition
    You can use process chains to control real-time data acquisition using process types Start Real-Time Data Acquisition (RDA) Load Process and Stop Real-Time Data Acquisition (RDA) Load Process.
    More Info :
    http://help.sap.com/saphelp_nw70/helpdata/en/47/2731751c2a2dede10000000a1553f7/frameset.htm
    Regards
    Ram.

  • Problems with real-time extraction from generic datasource.

    Hello,
    I have a real-time extraction from generic datasoruce with numeric field as a delta pointer. I have created an initilization infopackage, delta IP and real-time IP, real-time data transfer process, real-time deamon. The problem is that real-time extraction does not see delta updates. Deamon creates requests in PSA every minute, but there are all empty. When i delete the deamon and shedule not real-time delta infopackage, it loads delta records correctly. Version of BI - 7.0 SP 18.
    The second question: a cannot stop real-time deamon. I stop it in RSRDA, but when i try to shedule standart IP, i get an error, that deamon is still running. So i have to delete it and recreate.
    And the 3rd question, theoretical. What real-time extraction is necessary for? As i understand, i can create a process chain that will be sheduled every 1 (5, 10, ...) minute. What are advantages of real-time extraction rather such variant?
    Thanks.

    Hi,
    1.Note 1113610 - P17:SDL:PC:RDA:Daemon runs: Do not start delta/init IPackage
    Summary
    Symptom
    When a daemon runs or is scheduled, you may not start a 'normal' delta/init InfoPackage.
    Other terms
    RDA, real-time data acquisition, daemon, InfoPackage, scheduler
    Reason and Prerequisites
    This problem is caused by a program error.
    Solution
    SAP NetWeaver 7.0 BI
               Import Support Package 17 for SAP NetWeaver 7.0 BI (BI Patch 17 or SAPKW70017) into your BI system. The Support Package is available when Note 1106569 "SAPBINews BI7.0 Support Package 17", which describes this Support Package in more detail, is released for customers.
    **********you can apply the correction instructions provided in the note
    2.i refered closing of the request in the target i.e DSO not psa.
    3. Note 1000157 - RDA daemon does not get data from delta queue
    Summary
    Symptom
    You use Real-time Data Acquisition (RDA) to load data from an SAP source system. After several hours or days, the system no longer reads the data from the delta queue. In the RDA monitor, you can see that the InfoPackage has been executed, however the data transfer process (DTP) has not been executed for a long time. In transaction RSA7 in the source system, you can see that there is already data in the delta queue.
    Cheers,
    swapna.G

  • Compatibility of Pattern I/O with a PXI-6533 card with real time application

    Hi,
    I use a PXI-6533 card on a chassis with a PXI-8145 RT controller. I want to acquire 32 bits of data with this card at 1,25 kHz. These data are generated by an external FPGA card together with a clock signal. Referring to the 6533 documentation, a simple way to synchronize the acquisition is to use pattern I/O, entering the clock signal on the REQ pin. I tried the LabVIEW example Cont Pattern Input.vi. It seems to go well, but it does not work when I change this .vi priority to Time-critical or any higher priority than normal. Is there an incompatibility between this type of acquisition and real-time ?

    Hi,
    Sorry for this late answer, I effectively added a �wait until next ms function� in the time-critical loop and it seems to work. In fact, I don�t see exactly how, but it works with 1ms up to 50 ms.
    Nevertheless, the next step is to add the content of this example in a more complex system with a communication loop to transfer the acquired data to the Host PC. I already have such an architecture, with a �time-critical priority� .VI to read the data at 1.25 kHz and a RTFIFO to transfer them to a second �normal priority� .VI loaded on the PXI, that ensures a TCP communication with the HOST PC. This architecture works correctly on another application with an internal clock, BUT doesn�t with this example, even with your solution.
    In
    fact, it seems to work when I put the two .VIs loaded on the PXI in �normal� priority, but with a tremendous jitter (the frequency varies from 1 to 2 kHz) and a scan backlog that gets full very fast. Then, the PXI crashes as soon as I change any priority.
    You�ll find attached these three .VIs and I hope this could help to find a solution.
    Cordially
    PS: Since we started this thread, I�m in relation with NI Support France on that subject. Nevertheless, we haven�t found any solution yet and any new idea is welcome !
    Attachments:
    HOST_Comm.vi ‏532 KB
    RT_Comm.vi ‏265 KB
    TCLoop.vi ‏256 KB

Maybe you are looking for