Multiple Comparison​s on Single Data Source: NaN vs. Large Number

Hello All,
  Working in TestStand 2013, and came across a situation with DMM measurements, in which several values would be acceptable for the limit comparison.  For example, if I am making a resistance measurement, and I expect it to be a very high value, say greater than 1M, I would insert that as the lower limit.  However, there are also certain situations when I may see the result of "NaN" at the meter.  For the purposes of the test, any number greater that 1M would be acceptable, but a result of NaN would also be acceptable.  In an effort to keep the code modular and reusable for many instances, I am curious if there is a way in TestStand to compare the test result against two values at the same time, (i.e. >1M  OR  =NaN).  Please advise.  Thanks.
GSinMN  
Solved!
Go to Solution.

Probably because you are trying to set the Low limit property.  You need to set the Low Limit Expression property and set the Use Expression flag to true.
Consider the screen shot attached.  If UseLowExpr is False then the Low numeric property is used.  If it is True then the LowExpr property is used.
Hope this helps,
jigg
CTA, CLA
teststandhelp.com
~Will work for kudos and/or BBQ~
Attachments:
LowLimitExpression.png ‏19 KB

Similar Messages

  • DPM 2012 SP1 - Single Data Source Written to Multiple Tape Drives

    Just a question, really...
    I know that DPM can write to multiple drives for a single protection group with more than one data source.
    However, is it possible for DPM 2012 SP1 (or R2) to write to multiple drives from a single data source from a single protection group?
    So if I have "DataSource A" in "MyProtectionGroup" and I allocate 2 tape drives to the PG, could DPM use both drives to backup the PG?
    I assume the answer is "no," but I just wanted to double check.  Thanks.

    Hi,
    What you are wanting is a feature called striping and DPM does not have that advanced feature.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread. Regards, Mike J. [MSFT]
    This posting is provided "AS IS" with no warranties, and confers no rights.

  • Replicating single data source

    Hi Experts,
    I have a fresh copy of BW system to this I need to replicate single data source .How do I do that .If I replicate on source system all the ds will be replicated but I need to get only on DS .Pls suggest me.
    Swathi

    Hi Swathi,
    Use Function Module   RSAOS_METADATA_UPLOAD
    Datasource name -> I_OSOURCE
    Logical System  -> I_LOGSYS
    refer [Re: Repliction of datasources from BW- FM RSAOS_METADATA_UPLOAD;
    Reagrds
    Pratap Sone

  • How to configure optimal data connections to allow excel to retrieve multiple data sets form a single data source

    Hi all,
    I would like to have a data connection file used by excel to extract data form a SQL Server DB (so that change in location of the DB needs only a change to the  single data connection file). The excel file will retrieve data from only a single database,
    but there are multiple queries (30+) stored on SQL server that are used - each SQL server views' data is returned to a pivot table in an excel worksheet, which has an associated chart with various slicers on the main dashboard worksheet.
    Do I need a seperate data conenction file for each SQL server query being retrieved, or can a single connection file be created as all data comes from a single database?
    (all my attempts and research on the net have led me to believe that a seperate file is needed for each view, but this seems unnecessary)

    Hi, im learning my way with this, so apologies if I am providing too much or too little info.
    There are 8 source files which are very loosely related in that they capture infromation regarding what has happened within a metro railway over a day. However there are only a few relationships between the contents of tehse files.
    These tables are imported into SQL Server using SSIS where I have developed a number of views that query these 8 source tables to generate a number of metrics to provide insight into the service provided to the passengers. Some examples of metrics are: Number
    of passengers in transit at any given time, time taken to travel between adjacent stations, how much power was used during the day, what distance did the trains travel during the day, etc. Some views provide only a handful of rows, some provide 1M plus. There
    are now approx 40 seperate views created.
    I have then used a spreadsheet with a worksheet associated with each SQL server view. Each worksheet is set up as a pivot table, which allows the related chart on the main dashboard worksheet to use standard excel capability to slice and present the data
    in different ways.
    At the moment if the server on which teh database is stored moves, I have had to recreate the spreadsheet from scratch as I dont know how to change the connection information in any other way. Ideally I would like to have connection info in a single place
    to reduce ongoing maintenance, particularly as I would like to place the spreadsheet on a SharePoint server to distribute it to other users.
    Thanks

  • Objects are not referenced in cube data source view. Large model.

    Looking at large model. 100+ measure groups, 200+ dimensions, one cube. Deployed to multiple customers in production as a part of custom developed software.
    in data tools 2010 I'm looking at the cube on sql 2008, sql 2008 R2 (EE, DE).
    In the cube data source diagram we observe 50+ objects that do not have any links to them or from them.
    1)
    Some of these I was unable to find in DMV's of the cube at all, it turns out they are present in the Data source that is in Data Source Views section.
    How did these end/could end up in the cube data source model? they are not available in dimension usage window at all.
    2) Others do not have any links from them or to them in cube’s data source diagram, however the dimension usage clearly shows these objects are part of relationships 
    both regular and many to many depending on the object in question.
    We are trying to identify what objects we can ‘trim’ from the model. I’m not sure what am I looking at in this case: data tools have a trouble interpreting diagram for 200+ dimensions and
    100+ measure groups (it is a large existing model) and this artifact should not be used to decide what in dimensions is obsolete and can be decommissioned.
    We are looking at cube that has over 40 NxM just to one fact not counting regular relationships.
    Am I looking at a feature or maybe an issue with the model itself that somehow has developed in the particular cube.
    It is not something I came across before.
    thanks

    Hi,
    According to your description, you are designing a large SQL Server Analysis Services cube, the problem is that you are unable view the relationship in the data source
    view diagram, right?
    In your description, you said that you can see the relationships in Dimension Usage window. In your scenario, can you see the relationship under the Tables pane when expanding
    a table's Relationships folder?
    General, the relationships will be defined automatic if the relationships were defined in your data source of the cube. And the dimension relationships between cube dimensions
    and measure groups in the cube are defined manually in Dimension Usage. In your scenario, you can define relationship manually in the data source view, please refer to the link below to see the details.
    http://technet.microsoft.com/en-us/library/ms178448.aspx
    Regards,
    Charlie Liao
    TechNet Community Support

  • Running multiple VI's from one data source

    Hello,
    I am newer to Labview, and currently stuck on a problem.
    I have created a program to serial read from an RS232 port and extracted multiple points of the string and display it to the web. I would like to create a second VI from the same RS232 port and display one or two points of information. I have tried to do this with a global VI, but I can't web publish a global. Neither VI will have a user control, they simply display the data coming in. Any suggestions?

    You really should have a separate loop just to read the COM port.  You can then parse the read data and send to the appropriate sections of code using queues.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Multiple Deliverys in to Single Billing document for same part number

    Hi all,
    We want to combine multiple delivery items into one invoice line item for same material code,
    Scenario is like this,
    Sales order getting created with BOM ( Header items and more then 50 Component items with 30 to 40 quantity for each line items  )
    Based on availability we will create a delivery document for header and component items  -fox example
    sales order - Header - material - A with 1 quantity and component is Material is B with  40 quantity, based on availability we will create delivery document, let say 20 quantity and rest 20 quantity we will create another delivery docuemnt,
    while creating billing system creating with reference to two delivery docuemnt  its creating two line items,
    My Client requirement is if the material code is same, in billing document they need only one line item with all quantity, let say .. above example is for 40 quantity in to single line items in billing document.
    This can be possible in standard SAP,
    Can any one guide us or Which billing user exit would be suitable?
    Regards
    Nishad

    Hi,
    For the first part i think you can explore the "Delivery Group" option available in the item category setting.For the second requirement try to explore  USEREXIT_FILL_VBRK_VBRP ( Program RV60AFZC)
    hope it helps
    Regards,
    Shantanu

  • Query on data dictionary results in large number of physical reads

    I don't understand why I am getting 80,000 physicals for this query. I am not looking for help re-writing this. I just don't understand why I would hit the disk at all.
    My understanding had been that v$views where sql structures that pointed to x$tables. These x$tables are sql structures.
    underneath, the x$tables were linked lists stored in memory. This is why when you bounce the database, all the data, gets reset. Since it is not saved to disk.
    I am doing a simple insert/select off of v$open_cursor that is resulting in 80,000+ physical reads. I am posting the tkprof. It is all from v$open_cursor.
    mysid_table has 6 records. It is 1 mb in size
    if I index mysid_table.sid the query reduces to 20,000 physical reads. (but all the physical reads are on v$session_event)
    the sequence number I am passing returns 2 SIDs
    insert into my_save_table
    select *
    from v$session_event
    where sid in (select sid
      from my_sid_table
    where id = vseq);
    vrowcount := sql%rowcount;
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.01       0.01          0          0          0           0
    Execute      1     31.70      47.57      88570         22          0           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2     31.71      47.58      88570         22          0           1
    Misses in library cache during parse: 1
    Optimizer mode: CHOOSE
    Parsing user id: 22 
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      latch: row cache objects                        1        0.00          0.00
      log file sync                                   1        0.00          0.00
      SQL*Net message to client                       1        0.00          0.00
      SQL*Net message from client                     1        0.00          0.00
    ********************************************************************************

    It seems like there is some missing information.
    You have a wait for a log file sync, but no commit.
    Your table my_sid_table is 1 MB for only 6 records?
    Does the target table you are inserting into (my_save_table) have indexes on it?

  • Multiple rows with single date

    I am using a CDC control task to extract and load data from our OLTP source to our Data Warehouse.
    For our Date dimension, we have a DateKey which is a
    DATETIME stamp (created using SSAS dimension wizard). Time is always 00:00:00 as a DATETIME
    value is required for populating the dates.
    Our fact table destination uses this DateKey as a FK.
    Our fact table OLTP data source contains multiple records with a single DATE and TIME field. This fine as I can convert the DATE field to a DATETIME using a derived column.
    However, as our Date dimension has granularity of a day, all we require for our fact table is the most recent daily record from our OLTP data source.
    At present, the data flow task is attempting to write all records and failing on PK constraint.
    How would I go about traversing all records for each date and loading only the most recent for that day? Is there a better way to go about this?

    If you wish to get only one record for today, you can simply use in your source query:
    SELECT TOP 1 ... FROM OLTPdata ORDER BY DEATETIMEcolumn DESC
    if you need one record for each date, you may use ranking functions:
    WITH Q AS (
    SELECT
    ROW_NUMBER() OVER (PARTITION BY CAST(DateTimeColumn AS DATE) ORDER BY DateTimeColumn DESC, some tie breaker column) AS ROWNUM,
    FROM  OLTPData
    SELECT * FROM Q WHERE ROWNUM=1

  • Add multiple simultaneo​us data sources to Continuous Measuremen​t and Logging Template

    I am looking at using the  Continuous Measurement and Logging sample project as a starting point for my project.  The template shows the use of a single data source at a single rate.  In my case I have multiple data sources that will arrive at different rates and need to come together and be logged to the same file with the fastest arrival dictating the log entry and the slower data would be their last values.  It is a slow data collecton process ranging from 1 second to around 10 second sample rate.  More of a monitoring application with logging than a data aquisition application. Since the waveform data type is base-t  plus delta-t and my data arrives at varying rates, my first task would be to remove the use of it and define a typedef for my data. I would use a single logging message loop, but what approach can anyone recommend for the aquisition?  I imagine separate aquisition message loop vis for each data source but what about placement?  Placing them all at the top level vs placing them within a the existing aquisition message loop vi? Any other thoughts or recommendation are appreciated.  Thanks in advance!

    if you want to sample one channel at 10/s and an other one at 1/second....  How is your file going to look like?  Does this mean that for the slower channel you just want to keep the same value for the other 9 values?
    The sampling speeds that you are talking about are really slow.  It would certainly not hurt your hardware to sample eacht channel with the same speed.  And it you really just need one sample every 10 second, you can always decimate you array of values.
    For instance :
    channel 1 : 1 point per second (sample ratio) => 10 points in 10 seconds
    Channel 2 : 1 point per 10 seconds => 10 points per 10 seconds (same sample speed) => decimate array (keep 1+/10 values) => 1 point per 10 seconds
    This is going to be the easy way, or you want to start with parallel loops running, but that's "not that simple" anymore
    Kind regards,
    - Bjorn -
    Have fun using LabVIEW... and if you like my answer, please pay me back in Kudo's
    LabVIEW 5.1 - LabVIEW 2012

  • Multiple Data Sources unavailability in CUP 5.3

    Greetings All,
    Our Organization has acquired another organization and wants to bring them on board in GRC CUP for user provisioning process.
    The complexity here is they have a separate active directory and the merger of their active directory into the corporate AD is not slated for the next 6-8 months.
    The GRC CUP version 5.3 comes with a single data source and multiple details data sources. I now have my CUP pointing out to LDAP1 (MS Active directory) and I am not able to see any users from LDAP2 (Acquired MS Active Directory) which is expected.
    The AD team suggested LDAP Referral Chasing functionality that works well within MS windows or any other tools in the LDAP world.
    I did not see any other facility or option to use LDAP chasing in CUP.
    Any suggestions as how to connect to both LDAP's simultaneously without asking for another instance of CUP will be appreciated.
    Best regards,
    Angara

    Greetings Ashish & Alley,
    Thanks for the information and help. The situation here is, we do not have NW IDM and I wish we had it implemented; it would have solved many of our issues.
    We are looking into AD as the primary source collecting the information and creating the user in 2 places (Backend & Portal). In backend we have almost all applications (HCM/ECC/CRM/BI/XI/SRM/MDM/BPCu2026etc).
    There are about 10,000 users in one Ldap and about 4000 in another LDAP.  When the request is for an existing user, I can map the datasoure to the most available ABAP Stack (Best would be HCM as most users (excluding contractors/consultants) would have ESS) and pull the user info and get the remaining information from Details data source in sequential order.
    The Issue we have is with the New User creation, as the new user might be in LDAP1/LDAP2 based on the hiring organization and none of their information exists anywhere in the backend or in the portal - since we are using CUP to create even ESS accounts.
    We are currently evaluating on one among the below three things:
    1. Purchase NWIDM (not sure about the cost of the tool -Possibly 100 g's)
    2. Upgrade GRC from 5.3 to 10 (To determine the timeline & cost)
    3. Purchase a third party tool like an open data source or Sonatype (This tool can virtually connect multiple MS ADs into one AD tree and point CUP to this tool u2013may costs about 10g or so)
    If anyone of you has any other information, please let us know. Thanks for all your advice and help.
    Best Regards,
    Angara Rao

  • Multiple Data Source (for FAILOVER) in Easy Connect Naming method

    Hi All,
    I was wondering if there is a way to specify multiple addresses as part of data source property in the connection string when trying to connect to Oracle with Easy Connect Naming method? this is basically for failover scenario.
    I know its straight forward, if i use Connect Descriptor. But not sure how to do it with EZConnect
    For ex. this is what I want to achieve with EZConnect
    sales.us.example.com=
    (DESCRIPTION=
    (ADDRESS_LIST=
    (LOAD_BALANCE=off)
    (FAILOVER=ON)
    *(ADDRESS=(PROTOCOL=tcp)(HOST=sales1-server)(PORT=1521))*
    *(ADDRESS=(PROTOCOL=tcp)(HOST=sales2-server)(PORT=1521)))*
    (CONNECT_DATA=(SERVICE_NAME=sales.us.example.com)))
    I appreciate your suggestions...
    Thanks

    You might get better answers in one of the DB forums from a generic fail-over perspective as EZconnect isn't specific to ODP.NET.
    I don't know if there's a way to do it or not with EZConnect version of the connect string, but you can certainly provide the fully qualified connect string without using tnsnames.ora..
    constring = "data source=(DESCRIPTION=(ADDRESS_LIST=(LOAD_BALANCE=off)(FAILOVER=ON)(ADDRESS=(PROTOCOL=tcp)(HOST=sales1-server)(PORT=1521))(ADDRESS=(PROTOCOL=tcp)(HOST=sales2-server)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=sales.us.example.com))); user id=scott;password=tiger"
    Hope it helps,
    Gres

  • Multiple databases in a single query

    I need to query multiple databases in a single query.
    e.g.
    Database d1 contains table t1 with column id
    Database d2 contains table t2 with column id
    I need a query like
    SELECT * FROM t1, t2 WHERE t1.id = t2.id;
    is this possible in jdbc? if yes can anyone help with some sample code?
    thanks,
    Ashish Saraf

    It is not possible using the JDBC API to query tables in two or more databases using a single query. The reason is that a JDBC Connection is made to a single data source.
    However, some databases support the idea of federations or linked tables. What you do is inside the database manager link or federate a table that is present in another DBMS. Then, pose a single JDBC query to the one DBMS which will have its own table and links to the external tables.
    Another approach is to use a special JDBC driver with a built-in integration engine that will perform a query across multiple databases. In effect, the system poses separate queries to the different databases and integrates and joins the data together for you automatically on the client-side. That is much easier than doing it yourself in code, but that is also possible for simple queries.
    As part of my research program, we have released a shareware version of the UnityJDBC driver that can query multiple databases using a single SQL query. For more information, see:
    http://www.unityjdbc.com
    http://www.cs.uiowa.edu/~rlawrenc/research/projects.html
    Sincerely,
    Dr. Ramon Lawrence
    Assistant Professor
    Department of Computer Science
    University of Iowa
    [email protected]

  • The '' secondary data source is not a relational data source, or does not use an OLE DB provider.

    Hi,
    We are using Teradata as data source in SSAS 2008R2, it’s using ‘.Net Providers\.Net Data Provider for Teradata’.
    In DSV, we have two data sources both using Teradata. While processing dimensions, all dimensions coming from secondary data source has error message”
    The '' secondary data source is not a relational data source, or does not use an OLE DB provider.”
    But we are not using OLE DB provider for now, we’ve searched some threads that named queries can be replaced for secondary ds, but since we have a lot of tables, it’s hard to implement using named queries.
    Any inputs would be appreciated. 

    Hi memostone,
    When defining a data source view that contains tables, views, or columns from multiple data sources, the first data source from which you add objects to the data source view is designated as the primary data source (you cannot change the primary data
    source after it is defined). After defining a data source view based on objects from a single data source, you can then add objects from other data sources.
    If an OLAP processing or a data mining query requires data from multiple data sources in a single query, the primary data source must support remote queries using
    OpenRowset. Typically, this will be a SQL Server data source.
    Here are the restrictions:
    The primary data source should be SQL Server and support OpenRowset to the secondary data source.
    Or you design the cube in such a way that rocessing for each object only needs to access data from a single data source
    I recommand you refer to the following article:
    Defining a Data Source View (Analysis Services):
    http://technet.microsoft.com/en-us/library/ms174600.aspx
    Regards,
    Bin Long
    TechNet Community Support

  • Adhoc Reporting using Data Sources instead of Report Model

    Hi Gurus,
    I'm working on SSRS 2008. I'm trying to create Adhoc Report using Report Builder 3.0. To select Data Source, both *.rsds and *.smdl are listed. I selected a Data Source which points to a SQL database. When I add New datasets, I'm able to add the fields and
    it automatically detect relationships (Joins) using Primary Key and Foreign Key defined in the tables. I'm also able to edit the query. I ran the report and it ran successfully.
    My question is I couldn't find any documentation regarding using Data Source for Adhoc Reporting. MS SQL documentation and forums insists that Report Model is required for Adhoc Reporting.
    Please let me know if someone tried to create Adhoc Report using Data Source instead of Report Model. While I understand there are more benefits using Report Model (define entities/attributes, show/hide entity/attributes, organize entities/attributes etc.),
    I just want to know if this is possible.
    Are there any known limitation using this method?
    Thanks In Advance!

    Hi Ereswaran,
    Yes, we can create ad hoc report using report using data source instead of Report Model. Ad hoc reports typically are used to answer a specific question at appoint in time, so consequently people don’t spend a lot of time arranging and formatting the data
    in a precise way. Other report development tools are available in Reporting Services to support more sophisticated reporting requirements.
    Using Report Builder, we can build a simple report that display data form a single data source as defined by a report model. This tool is most useful for business users who need access to information but lack the skills to write a SQL query. If we have full
    experience of SQL query, of course we can use data source instead of Report Model.
    If you have any questions, please feel free to ask.
    Regards,
    Charlie Liao
    Charlie Liao
    TechNet Community Support

Maybe you are looking for