Generating 2 test reports for the same test - XML and HTML
Hi All,
I have a special test requirement.
Need to generate 2 test reports for the same test. One in HTML and the other in XML format.
Is there any direct method for that?
How it can be implemented?
Thanks in advance.
SajK
Hi SajK,
To accomplish this task, you will need to modify the process model. For example, if you are using the sequential process model, you can concentrate on the TestReport sequence call and the various report related steps in the Single Pass and Test UUTs sequences.
The TestReport sequence (also a callback) relies on Parameters.ReportOptions.Format to decide which of the following sequences to use when generating the report:
<TestStand>\Components\NI\Models\TestStandModels\reportgen_txt.seq
<TestStand>\Components\NI\Models\TestStandModels\reportgen_html.seq
<TestStand>\Components\NI\Models\TestStandModels\reportgen_xml.seq
In your case, instead of choosing one of the sequences, you could hard code it to use both the html and xml sequences. Be aware that for this to work, you must also modify and/or duplicate most of the report related steps in the Single Pass and Test UUTs sequences. For instance, the Set Report Format step becomes unnecessary since you are hard-coding the report format. On the other hand, the Write UUT Report step needs to be duplicated, one step to write to an html file location, and another to write to the xml file location.
Please note that modifying the process mode in this way will deprecate some of the settings in the report options dialog box (such as the report format). For more information on TestStand report generation, please refer to the following DevZone article:
Report Generation Explained
Regards,
Message Edited by James M on 05-29-2007 03:08 PM
Similar Messages
-
PT_QTA00 is generating repetative records for the same quota type
PT_QTA00 is generating repetative records for the same quota type and one more thing i have already maintained 10 days of quota in 2006 and when i run for the same employee its creating one more record with 10 days again how can we restrict it in configuration
As said earlier, you want to setup Time Constraints for It2006 which is in table: V_T582A
You should set the Time Constraint to 2.
HR master data uses the following three time constraints:
Time Constraint 1
For the entire time that the employee works at the enterprise, exactly one valid infotype record must exist. The validity periods of the individual records must not overlap. If a new record is created, the system automatically uses the start date of the new record as the delimitation date of the old record. Gaps are only allowed between the employeeu2019s entry date and the start date of the first record.
Time constraint 1 must be used for all of the infotypes containing information that must be available at all times. This is particularly true of personal and organizational assignment data.
If a record is delimited because of time constraint 1, the system displays an appropriate message.
Time Constraint 2
No more than one valid record can exist at any one time. Records with constraint 2 must not overlap. Their existence is not obligatory. If a new record is created, the system automatically delimits the previous record, if one exists.
If a record is delimited because of time constraint 2, the system displays an appropriate message.
Time Constraint 3
Any number of valid records can exist at any one time. The individual records do not conflict with each other. -
upgraded my macbook pro to mountain lion from snow leopard does this mean i can upgrade my imac as well for the same price? And will it mess it up?
You can upgrade the iMac for no extra charge - sign into the App Store there, download and install 10.8.
I recommend that you make a backup (Time Machine is probably the easiest method) before you install, and make sure that you are not using any applications that are incomptible with 10.8 (eg Office 2004) - other than that, it should go smoothly.
Matt -
CBO generating different plans for the same data in similar Environments
Hi All
I have been trying to compare an SQL from 2 different but similar environments build of the same hardware specs .The issue I am facing is environment A, the query executes in less than 2 minutes with plan mostly showing full table scans and hash join whereas in environment B(problematic), it times out after 2 hours with an error of unable to extend table space . The statistics are up to date in both environments for both tables and indexes . System parameters are exactly similar(default oracle for except for multiblock_read_count ).
Both Environment have same db parameter for db_file_multiblock_read_count(16), optimizer(refer below),hash_area_size (131072),pga_aggregate_target(1G),db_block_size(8192) etc . SREADTIM, MREADTIM, CPUSPEED, MBRC are all null in aux_stats in both environment because workload was never collected i believe.
Attached is details about the SQL with table stats, SQL and index stats my main concern is CBO generating different plans for the similar data and statistics and same hardware and software specs. Is there any thing else I should consider .I generally see environment B being very slow and always plans tend to nested loops and index scan whereas what we really need is a sensible FTS in many cases. One of the very surprising thing is METER_CONFIG_HEADER below which has just 80 blocks of data is being asked for index scan.
show parameter optimizer
optimizer_dynamic_sampling integer 2
optimizer_features_enable string 10.2.0.4
optimizer_index_caching integer 0
optimizer_index_cost_adj integer 100
optimizer_mode string ALL_ROWS
optimizer_secure_view_merging boolean TRUE
**Environment**
Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
PL/SQL Release 10.2.0.4.0 - Production
CORE 10.2.0.4.0 Production
TNS for Solaris: Version 10.2.0.4.0 - Production
NLSRTL Version 10.2.0.4.0 - Production
Note: : There are slight difference in the no of records in the attached sheet.However, I wanted to tell that i have tested with exact same data and was getting similar results but I couldn't retain the data untill collecting the details in the attachment
TEST1 COMPARE TABLE LEVE STATS used by CBO
ENVIRONMENT A
TABLE_NAME NUM_ROWS BLOCKS LAST_ANALYZED
ASSET 3607425 167760 5/02/2013 22:11
METER_CONFIG_HEADER 3658 80 5/01/2013 0:07
METER_CONFIG_ITEM 32310 496 5/01/2013 0:07
NMI 1899024 33557 18/02/2013 10:55
REGISTER 4830153 101504 18/02/2013 9:57
SDP_LOGICAL_ASSET 1607456 19137 18/02/2013 15:48
SDP_LOGICAL_REGISTER 5110781 78691 18/02/2013 9:56
SERVICE_DELIVERY_POINT 1425890 42468 18/02/2013 13:54
ENVIRONMENT B
TABLE_NAME NUM_ROWS BLOCKS LAST_ANALYZED
ASSET 4133939 198570 16/02/2013 10:02
METER_CONFIG_HEADER 3779 80 16/02/2013 10:55
METER_CONFIG_ITEM 33720 510 16/02/2013 10:55
NMI 1969000 33113 16/02/2013 10:58
REGISTER 5837874 120104 16/02/2013 11:05
SDP_LOGICAL_ASSET 1788152 22325 16/02/2013 11:06
SDP_LOGICAL_REGISTER 6101934 91088 16/02/2013 11:07
SERVICE_DELIVERY_POINT 1447589 43804 16/02/2013 11:11
TEST ITEM 2 COMPARE INDEX STATS used by CBO
ENVIRONMENT A
TABLE_NAME INDEX_NAME UNIQUENESS BLEVEL LEAF_BLOCKS DISTINCT_KEYS AVG_LEAF_BLOCKS_PER_KEY AVG_DATA_BLOCKS_PER_KEY CLUSTERING_FACTOR NUM_ROWS
ASSET IDX_AST_DEVICE_CATEGORY_SK NONUNIQUE 2 9878 67 147 12982 869801 3553095
ASSET IDX_A_SAPINTLOGDEV_SK NONUNIQUE 2 7291 2747 2 639 1755977 3597916
ASSET SYS_C00102592 UNIQUE 2 12488 3733831 1 1 3726639 3733831
METER_CONFIG_HEADER SYS_C0092052 UNIQUE 1 12 3670 1 1 3590 3670
METER_CONFIG_ITEM SYS_C0092074 UNIQUE 1 104 32310 1 1 32132 32310
NMI IDX_NMI_ID NONUNIQUE 2 6298 844853 1 2 1964769 1965029
NMI IDX_NMI_ID_NK NONUNIQUE 2 6701 1923072 1 1 1922831 1923084
NMI IDX_NMI_STATS NONUNIQUE 1 106 4 26 52 211 211
REGISTER REG_EFFECTIVE_DTM NONUNIQUE 2 12498 795 15 2899 2304831 4711808
REGISTER SYS_C00102653 UNIQUE 2 16942 5065660 1 1 5056855 5065660
SDP_LOGICAL_ASSET IDX_SLA_SAPINTLOGDEV_SK NONUNIQUE 2 3667 1607968 1 1 1607689 1607982
SDP_LOGICAL_ASSET IDX_SLA_SDP_SK NONUNIQUE 2 3811 668727 1 2 1606204 1607982
SDP_LOGICAL_ASSET SYS_C00102665 UNIQUE 2 5116 1529606 1 1 1528136 1529606
SDP_LOGICAL_REGISTER SYS_C00102677 UNIQUE 2 17370 5193638 1 1 5193623 5193638
SERVICE_DELIVERY_POINT IDX_SDP_NMI_SK NONUNIQUE 2 4406 676523 1 2 1423247 1425890
SERVICE_DELIVERY_POINT IDX_SDP_SAP_INT_NMI_SK NONUNIQUE 2 7374 676523 1 2 1458238 1461108
SERVICE_DELIVERY_POINT SYS_C00102687 UNIQUE 2 4737 1416207 1 1 1415022 1416207
ENVIRONMENT B
TABLE_NAME INDEX_NAME UNIQUENESS BLEVEL LEAF_BLOCKS DISTINCT_KEYS AVG_LEAF_BLOCKS_PER_KEY AVG_DATA_BLOCKS_PER_KEY CLUSTERING_FACTOR NUM_ROWS
ASSET IDX_AST_DEVICE_CATEGORY_SK NONUNIQUE 2 8606 121 71 16428 1987833 4162257
ASSET IDX_A_SAPINTLOGDEV_SK NONUNIQUE 2 8432 1780146 1 1 2048170 4162257
ASSET SYS_C00116157 UNIQUE 2 13597 4162263 1 1 4158759 4162263
METER_CONFIG_HEADER SYS_C00116570 UNIQUE 1 12 3779 1 1 3734 3779
METER_CONFIG_ITEM SYS_C00116592 UNIQUE 1 107 33720 1 1 33459 33720
NMI IDX_NMI_ID NONUNIQUE 2 6319 683370 1 2 1970460 1971313
NMI IDX_NMI_ID_NK NONUNIQUE 2 6597 1971293 1 1 1970771 1971313
NMI IDX_NMI_STATS NONUNIQUE 1 98 48 2 4 196 196
REGISTER REG_EFFECTIVE_DTM NONUNIQUE 2 15615 1273 12 2109 2685924 5886582
REGISTER SYS_C00116748 UNIQUE 2 19533 5886582 1 1 5845565 5886582
SDP_LOGICAL_ASSET IDX_SLA_SAPINTLOGDEV_SK NONUNIQUE 2 4111 1795084 1 1 1758441 1795130
SDP_LOGICAL_ASSET IDX_SLA_SDP_SK NONUNIQUE 2 4003 674249 1 2 1787987 1795130
SDP_LOGICAL_ASSET SYS_C004520 UNIQUE 2 5864 1795130 1 1 1782147 1795130
SDP_LOGICAL_REGISTER SYS_C004539 UNIQUE 2 20413 6152850 1 1 6073059 6152850
SERVICE_DELIVERY_POINT IDX_SDP_NMI_SK NONUNIQUE 2 3227 660649 1 2 1422572 1447803
SERVICE_DELIVERY_POINT IDX_SDP_SAP_INT_NMI_SK NONUNIQUE 2 6399 646257 1 2 1346948 1349993
SERVICE_DELIVERY_POINT SYS_C00128706 UNIQUE 2 4643 1447946 1 1 1442796 1447946
TEST ITEM 3 COMPARE PLANS
ENVIRONMENT A
Plan hash value: 4109575732
| Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)| Time |
| 0 | SELECT STATEMENT | | 13 | 2067 | | 135K (2)| 00:27:05 |
| 1 | HASH UNIQUE | | 13 | 2067 | | 135K (2)| 00:27:05 |
|* 2 | HASH JOIN | | 13 | 2067 | | 135K (2)| 00:27:05 |
|* 3 | HASH JOIN | | 6 | 900 | | 135K (2)| 00:27:04 |
|* 4 | HASH JOIN ANTI | | 1 | 137 | | 135K (2)| 00:27:03 |
|* 5 | TABLE ACCESS BY INDEX ROWID| NMI | 1 | 22 | | 5 (0)| 00:00:01 |
| 6 | NESTED LOOPS | | 1 | 131 | | 95137 (2)| 00:19:02 |
|* 7 | HASH JOIN | | 1 | 109 | | 95132 (2)| 00:19:02 |
|* 8 | TABLE ACCESS FULL | ASSET | 36074 | 1021K| | 38553 (2)| 00:07:43 |
|* 9 | HASH JOIN | | 90361 | 7059K| 4040K| 56578 (2)| 00:11:19 |
|* 10 | HASH JOIN | | 52977 | 3414K| 2248K| 50654 (2)| 00:10:08 |
|* 11 | HASH JOIN | | 39674 | 1782K| | 40101 (2)| 00:08:02 |
|* 12 | TABLE ACCESS FULL | REGISTER | 39439 | 1232K| | 22584 (2)| 00:04:32 |
|* 13 | TABLE ACCESS FULL | SDP_LOGICAL_REGISTER | 4206K| 56M| | 17490 (2)| 00:03:30 |
|* 14 | TABLE ACCESS FULL | SERVICE_DELIVERY_POINT | 675K| 12M| | 9412 (2)| 00:01:53 |
|* 15 | TABLE ACCESS FULL | SDP_LOGICAL_ASSET | 1178K| 15M| | 4262 (2)| 00:00:52 |
|* 16 | INDEX RANGE SCAN | IDX_NMI_ID_NK | 2 | | | 2 (0)| 00:00:01 |
| 17 | VIEW | | 39674 | 232K| | 40101 (2)| 00:08:02 |
|* 18 | HASH JOIN | | 39674 | 1046K| | 40101 (2)| 00:08:02 |
|* 19 | TABLE ACCESS FULL | REGISTER | 39439 | 500K| | 22584 (2)| 00:04:32 |
|* 20 | TABLE ACCESS FULL | SDP_LOGICAL_REGISTER | 4206K| 56M| | 17490 (2)| 00:03:30 |
|* 21 | TABLE ACCESS FULL | METER_CONFIG_HEADER | 3658 | 47554 | | 19 (0)| 00:00:01 |
|* 22 | TABLE ACCESS FULL | METER_CONFIG_ITEM | 7590 | 68310 | | 112 (2)| 00:00:02 |
Predicate Information (identified by operation id):
2 - access("METER_CONFIG_HEADER_SK"="METER_CONFIG_HEADER_SK")
3 - access("NETWORK_TARIFF_CD"="NETWORK_TARIFF_CD")
4 - access("SERVICE_DELIVERY_POINT_SK"="TMP"."SERVICE_DELIVERY_POINT_SK")
5 - filter("ROW_CURRENT_IND"='Y' AND ("NMI_STATUS_CD"='A' OR "NMI_STATUS_CD"='D'))
7 - access("ASSET_CD"="EQUIP_CD" AND "SAP_INT_LOG_DEVICE_SK"="SAP_INT_LOG_DEVICE_SK")
8 - filter("ROW_CURRENT_IND"='Y')
9 - access("SERVICE_DELIVERY_POINT_SK"="SERVICE_DELIVERY_POINT_SK")
10 - access("SERVICE_DELIVERY_POINT_SK"="SERVICE_DELIVERY_POINT_SK")
11 - access("SAP_INT_LOGICAL_REGISTER_SK"="SAP_INT_LOGICAL_REGISTER_SK")
12 - filter("REGISTER_TYPE_CD"='C' AND (SUBSTR("REGISTER_ID_CD",1,1)='4' OR
SUBSTR("REGISTER_ID_CD",1,1)='5' OR SUBSTR("REGISTER_ID_CD",1,1)='6') AND "ROW_CURRENT_IND"='Y')
13 - filter("ROW_CURRENT_IND"='Y')
14 - filter("ROW_CURRENT_IND"='Y')
15 - filter("ROW_CURRENT_IND"='Y')
16 - access("NMI_SK"="NMI_SK")
18 - access("SAP_INT_LOGICAL_REGISTER_SK"="SAP_INT_LOGICAL_REGISTER_SK")
19 - filter("REGISTER_TYPE_CD"='C' AND (SUBSTR("REGISTER_ID_CD",1,1)='1' OR
SUBSTR("REGISTER_ID_CD",1,1)='2' OR SUBSTR("REGISTER_ID_CD",1,1)='3') AND "ROW_CURRENT_IND"='Y')
20 - filter("ROW_CURRENT_IND"='Y')
21 - filter("ROW_CURRENT_IND"='Y')
22 - filter("ROW_CURRENT_IND"='Y' AND "CONROL_REGISTER"='X')
ENVIRONMENT B
Plan hash value: 2826260434
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
| 0 | SELECT STATEMENT | | 1 | 181 | 103K (2)| 00:20:47 |
| 1 | HASH UNIQUE | | 1 | 181 | 103K (2)| 00:20:47 |
|* 2 | HASH JOIN ANTI | | 1 | 181 | 103K (2)| 00:20:47 |
|* 3 | HASH JOIN | | 1 | 176 | 56855 (2)| 00:11:23 |
|* 4 | HASH JOIN | | 1 | 163 | 36577 (2)| 00:07:19 |
|* 5 | TABLE ACCESS BY INDEX ROWID | ASSET | 1 | 44 | 4 (0)| 00:00:01 |
| 6 | NESTED LOOPS | | 1 | 131 | 9834 (2)| 00:01:59 |
| 7 | NESTED LOOPS | | 1 | 87 | 9830 (2)| 00:01:58 |
| 8 | NESTED LOOPS | | 1 | 74 | 9825 (2)| 00:01:58 |
|* 9 | HASH JOIN | | 1 | 52 | 9820 (2)| 00:01:58 |
|* 10 | TABLE ACCESS BY INDEX ROWID| METER_CONFIG_HEADER | 1 | 14 | 1 (0)| 00:00:01 |
| 11 | NESTED LOOPS | | 1 | 33 | 116 (2)| 00:00:02 |
|* 12 | TABLE ACCESS FULL | METER_CONFIG_ITEM | 1 | 19 | 115 (2)| 00:00:02 |
|* 13 | INDEX RANGE SCAN | SYS_C00116570 | 1 | | 1 (0)| 00:00:01 |
|* 14 | TABLE ACCESS FULL | SERVICE_DELIVERY_POINT | 723K| 13M| 9699 (2)| 00:01:57 |
|* 15 | TABLE ACCESS BY INDEX ROWID | NMI | 1 | 22 | 5 (0)| 00:00:01 |
|* 16 | INDEX RANGE SCAN | IDX_NMI_ID_NK | 2 | | 2 (0)| 00:00:01 |
|* 17 | TABLE ACCESS BY INDEX ROWID | SDP_LOGICAL_ASSET | 1 | 13 | 5 (0)| 00:00:01 |
|* 18 | INDEX RANGE SCAN | IDX_SLA_SDP_SK | 2 | | 2 (0)| 00:00:01 |
|* 19 | INDEX RANGE SCAN | IDX_A_SAPINTLOGDEV_SK | 2 | | 2 (0)| 00:00:01 |
|* 20 | TABLE ACCESS FULL | REGISTER | 76113 | 2378K| 26743 (2)| 00:05:21 |
|* 21 | TABLE ACCESS FULL | SDP_LOGICAL_REGISTER | 5095K| 63M| 20245 (2)| 00:04:03 |
| 22 | VIEW | | 90889 | 443K| 47021 (2)| 00:09:25 |
|* 23 | HASH JOIN | | 90889 | 2307K| 47021 (2)| 00:09:25 |
|* 24 | TABLE ACCESS FULL | REGISTER | 76113 | 966K| 26743 (2)| 00:05:21 |
|* 25 | TABLE ACCESS FULL | SDP_LOGICAL_REGISTER | 5095K| 63M| 20245 (2)| 00:04:03 |
Predicate Information (identified by operation id):
2 - access("SERVICE_DELIVERY_POINT_SK"="TMP"."SERVICE_DELIVERY_POINT_SK")
3 - access("SERVICE_DELIVERY_POINT_SK"="SERVICE_DELIVERY_POINT_SK" AND
"SAP_INT_LOGICAL_REGISTER_SK"="SAP_INT_LOGICAL_REGISTER_SK")
4 - access("ASSET_CD"="EQUIP_CD")
5 - filter("ROW_CURRENT_IND"='Y')
9 - access("NETWORK_TARIFF_CD"="NETWORK_TARIFF_CD")
10 - filter("ROW_CURRENT_IND"='Y')
12 - filter("ROW_CURRENT_IND"='Y' AND "CONROL_REGISTER"='X')
13 - access("METER_CONFIG_HEADER_SK"="METER_CONFIG_HEADER_SK")
14 - filter("ROW_CURRENT_IND"='Y')
15 - filter("ROW_CURRENT_IND"='Y' AND ("NMI_STATUS_CD"='A' OR "NMI_STATUS_CD"='D'))
16 - access("NMI_SK"="NMI_SK")
17 - filter("ROW_CURRENT_IND"='Y')
18 - access("SERVICE_DELIVERY_POINT_SK"="SERVICE_DELIVERY_POINT_SK")
19 - access("SAP_INT_LOG_DEVICE_SK"="SAP_INT_LOG_DEVICE_SK")
20 - filter((SUBSTR("REGISTER_ID_CD",1,1)='4' OR SUBSTR("REGISTER_ID_CD",1,1)='5' OR
SUBSTR("REGISTER_ID_CD",1,1)='6') AND "REGISTER_TYPE_CD"='C' AND "ROW_CURRENT_IND"='Y')
21 - filter("ROW_CURRENT_IND"='Y')
23 - access("SAP_INT_LOGICAL_REGISTER_SK"="SAP_INT_LOGICAL_REGISTER_SK")
24 - filter((SUBSTR("REGISTER_ID_CD",1,1)='1' OR SUBSTR("REGISTER_ID_CD",1,1)='2' OR
SUBSTR("REGISTER_ID_CD",1,1)='3') AND "REGISTER_TYPE_CD"='C' AND "ROW_CURRENT_IND"='Y')
25 - filter("ROW_CURRENT_IND"='Y')Edited by: abhilash173 on Feb 24, 2013 9:16 PM
Edited by: abhilash173 on Feb 24, 2013 9:18 PMHi Paul,
I misread your question initially .The system stats are outdated in both ( same result as seen from aux_stats) .I am not a DBA and do not have access to gather system stats fresh.
select * from sys.aux_stats$
SNAME PNAME PVAL1 PVAL2
SYSSTATS_INFO STATUS NULL COMPLETED
SYSSTATS_INFO DSTART NULL 02-16-2011 15:24
SYSSTATS_INFO DSTOP NULL 02-16-2011 15:24
SYSSTATS_INFO FLAGS 1 NULL
SYSSTATS_MAIN CPUSPEEDNW 1321.20523 NULL
SYSSTATS_MAIN IOSEEKTIM 10 NULL
SYSSTATS_MAIN IOTFRSPEED 4096 NULL
SYSSTATS_MAIN SREADTIM NULL NULL
SYSSTATS_MAIN MREADTIM NULL NULL
SYSSTATS_MAIN CPUSPEED NULL NULL
SYSSTATS_MAIN MBRC NULL NULL
SYSSTATS_MAIN MAXTHR NULL NULL
SYSSTATS_MAIN SLAVETHR NULL NULL -
Merging two records in a single record for the same matnr,werks and bwart.
Hi I have a requirement to merge two records into one single record using the quantity field from MSEG table.
I am selecting two records from the table MSEG. Now for these two records i have to add the quantity values into one variable.
Then there should be only one record for the same matnr,werks,bwart.
Any suggestion would be appreciated.
Regards,
Amithi,
You can COLLECT statement only if all non key fields are numeric.
This statement Adds all such fields and give us a single record.
Regards
Sumit Agarwal -
A software application was developed to collect and process readings from capacitance sensors and a tachometer in a running spin rig. The sensors were connected to an Aerogate Model HP-04 H1 Band Preamp connected to an NI PXI-6115. The sensors were read using AI Config and AI Start VIs. The data was saved to a file using hsdlConfig and hsdlFileWriter VIs. In order to add the capability of collecting synchronized data from two Eddy Current Position sensors in addition to the existing sensors, which will be connected to a BNC-2144 connected to an NI PXI-4495, the AI and HSDL VIs were replaced with DAQmx VIs logging to TDMS. When running identical tests, the new file format (TDMS) produces reads that are higher and inconsistent with the readings from the older file format (HSDL).
The main VIs are SpinLab 2.4 and SpinLab 3.8 in folders "SpinLab old format" and "Spinlab 3.8" respectfully. SpinLab 3.8 requires the Sound and Vibration suite to run correctly, but it is used after the part that is causing the problem. The problem is occuring during data collection in the Logger segment of code or during processing in the Reader/Converter segment of code. I could send the readings from the identical tests if they would be helpful, but the data takes up approximately 500 MB.
Attachments:
SpinLab 3.8.zip 1509 KB
SpinLab 2.4.zip 3753 KB
SpinLab Screenshots.doc 795 KBFirst of all, how different is the data? You say that the reads are higher and inconsistent. How much higher? Is every point inconsistent, or is it just parts of your file? If it's just in parts of the file, does there seem to be a consistent pattern as to when the data is different?
Secondly, here are a couple things to try:
Currently, you are not calling DAQmx Stop Task outside of the loop; you're just calling DAQmx Clear Task. This means that if there were any errors that occured in the logging thread, you might not be getting them (as DAQmx Clear Task clears outstanding errors within the task). Add a DAQmx Stop Task before DAQmx Clear Task to make sure that you're not missing an error.
Try "Log and Read" mode. "Log and Read" is probably going to be fast enough for your application (as it's pretty fast), so you might just try it and see if you get any different result. All that you would need to do is change the enum to "Log and Read", then add a DAQmx Read in the loop (you can just use Raw format since you don't care about the output). I'd recommend that you read in even multiples of the sector size (normally 512) for optimal performance. For example, your rate is 1MHz, perhaps read in sizes of 122880 samples per channel (something like 1/8 of the buffer size rounded down to the nearest multiple of 4096). Note: This is a troubleshooting step to try and narrow down the problem.
Finally, how confident are you in the results from the previous HSDL test? Which readings make more sense? I look forward to hearing more detail about how the data is inconsistent (all data, how different, any patterns). As well, I'll be looking forward to hearing the result of test #2 above.
Thanks,
Andy McRorie
NI R&D -
Can I print 2 different reports for the same execution?
Hello all,
I currently have a test sequence that branches off into several subsequences, which branch off, etc. Currently, I have the report options filter out to only report 'pass' or 'fail' results. Even with this, my report is 90-100 pages. This is a good, very detailed report.
However(you knew there would be a 'however'), some people want a condensed version of this report, where it only lists whether the sequence itself failed, and not the individual steps within.
My question is, can I run my test and have both the detailed and compact reports run for that execution? Either that, or simply list if each sequence passes, and if it fails, then display the individual steps?
Either way would work for
me.
Thanks in advance for your help!
Dave Neumann
[email protected]Hi,
As far as I know, TestStand does not include a built in feature to do what your are looking for. However this doesn't mean you can not do what your are looking for.
Like Zvezdana mentioned, you can either use the dll or sequence files to generate reports and you have to modify either of them to achieve what you want.
The dll generates de report faster that the sequences, however, the sequences are easier to modify. You can select the option you want from the Report Options.
If you select to use the sequences, there are two sequence files that would be called depending on the format you choose for your report. reportgen_txt.seq or reportgen_html.seq
They are located under ...\testStand\components\ni\models\testStand models\ Coy this to the
user directory (same path as above but user instead of ni) and make the changes that you need.
For your second option (only printing the details when a step fails) you can start with AddSequenceResultToReport sequence. This sequence has the code that inserts the begin and end sequence markers. You can getrid of them and this will make your report smaller. You can replace that with code that includes if your sequence passed or failed. Look under parameters for the result property. In this sequence there is a call to a subsequence that adds all the results of all the steps within the sequence. You can add a precondition to this step to execute only if the sequence failed.
There are many more things that you could do, but I hope this helps you get started. If you have specific questions once you start doing this, let me know.
I hope this helps,
Marcela. -
Acknowlegment of PO sent through E mail & report for the same
Hi SAP Experts,
We have started sending PO through E Mails.Now our client have requirement that they should receive acknowledgement from vendor about the receipt of PO & also should able to generate report vendor wise about the receipts of PO.Currently we are on SAP 4.7.
Is it possible to map this requirement?Hi,
Upon receiving the acknowledgement the buyer should update the same in the confirmations Tab using the appropriate confirmation control.
Then monitoring can be done using ME92F.
If this needs to be automated, then you need to consider the mode of receiving the acknowledgement electronically and accordingly do a development to update the same in PO.
Hope this helps. -
Open multiples reports for the same document
Hi,
I created 3 PLD reports for PO, I am wondering in SDK, how can I open all 3 PLD reports when I click on the preview button for a PO document?
Thanks
ChrisHi Chris,
You cannot. I use Crystal Reports when I need to open more then one document layout at the same time.
Regards,
Vítor Vieira -
Require two different Financial Reports for the Same Financial Year - 2011
Hi Experts,
The financial Year of the client is from jan-Dec.The client needs different financial reports (P&L , B/S), one from Jan-Aug and another from Sept-Dec for the FY 2011 since the management of the company has changed from Sept-2011 onwards. So when tried to generate the financial reports from Sept-Dec , system shows the carry forward opening balances of income and expenses from Aug 2011 which they do not want. Since we have not done any closing activities at the end of aug-2011 this figure is bound to come. So in this scenario is there any possibilities to get a fresh financial reports from Sept to Dec 2011 where there should not be any opening balances of expenses and incomes for Sept 2011.
Since transferring of balances from income and expenses to retained earnings accounts has done yearly basis , how can we do this for the periods also or any other method to sort out this issue.
Hope my question is clear. if not pls revert with the queries.
Thanks in Advance,
ArabindaHi RP,
I have checked F.54 . Everything is ok in that report expect the Profit and Loss which are two differnt line item in that report and also calculated fields do not display in this report. But anyway this report meets most of the requirements.
Thank you very much.
regards,
Arabinda -
Inventory report for the selection of requisitioner and MRP controller
Hello Experts,
Can someone please help me with a standard Inventory report in which we can give input thr requistioner and MRP controller
regards,
YKHi,
You will get the requisition no. from EKKO (provided PO has been created otherwise you won't get any record). Find out PO & item no. for the requisition item. Then you can check EKBE table whether GR has been posted or not. If GR has not been posted then there won't be entry in EKBE for that PO & item.
Regards,
Rakesh -
I purchased my wife an Iphone and we currently use the same icloud account, but we just ran out of storage. I'm not sure if we need so much more space that I'm ready to upgrade to 15 total GB of storage. If my wife sets up her own icloud account for her iphone will she and I still be able to share our photos, downloaded apps and music with different icloud accounts?
Apps and music aren't anything to do with your iCloud account, indeed you can use different ID's for each.
If you create and use another ID for iCloud, you will not be able to share both photostreams to the same user account on the computer.
This could be worked around with a second user account on the computer for the second user to do all their computing in. However you may find that photostreams shared albums will work around the problem for you. For example if you create a shared album and share it with your partner, any photos you add to that album will be available in your partners photostream on the computer. -
Report for the list of materials with component & assembly scrap
Dear Experts ,
I have been told to generate a report for the materials which are subjected to component and assembly scrap . Please guide me for the same . Many thanks in advance.
Yours
Aman KhanDear Aman,
It you are maintaining both assembly scrap & component scrap both in MM you can read MARC table by creating simple query.
Read Field MARC-AUSSS & MARC-KAUSF.
If you are maintaining component scrap in BOM item read STPO-AUSCH.
[Query Creation|How do create query in SAP?] -
I made a purchase of about 3 songs for the price of 5.16. Well when I was trying to download the songs a message kept appearing saying that "the account inforamtion does not match my bank records". Looking at my bank account 3 days later it shows up under my accout as pre-authorized charges 9 times. The post dates of these charges are not all today and these charges are pending transactions.
Why did they charge me nine times for the same purchase? And why is it charged for tomorrow?This happend to me to---kept getting msg that info did not match bank records and I tried a few times then gave up THEN checked bank account and they charged me SIX TIMES
of course, you cannot even call them or email...its frustrating..I reported a problem under the songs bought but I have spent over half an hour on this trying to call etc....this is nuts -
SAP BW reports for the users?
Hi,
i have a scenario in my new project. i have given all standard reports to the users and apart from these reports they are asking for new reports which uses for their activities. we have SD, FI,MM,HR and PM modules are there. they are not giving me any requirement but they are asking me to give some ideas and create reports.
How can i proceed in these case and i dont have any functional knowledge. Kindly suggest.
ThanksHi,
My Suggestion will be first check with SD, FI,MM,HR and PM modules related ECC reports and then check standard reports in BI either possible to match with ECC reports .
IF not then try to analyze existing reports of each module in ECC & check is there any possibility of using standard D.S & standard Cubes for those reports then create reports for the same in BI.
IF possible try to interact with functional team they can help in giving exact Bushiness Requirement.
Regards,
Seshu.P
Maybe you are looking for
-
I need help I updated my phone and when It was done updating if centime was missing and when I ask Siri to upend it she says I have to make an account how do I do this if i don't have the app plz help me apple
-
RMAN-05001: auxiliary filename
I am trying to duplicate a database using rman nocatalog on local host. I have set the convert file names to db_file_name_convert and log_file_name_convert in PFILE to AUXILIARY instance. when I execute the duplicate command in Rman I do get this err
-
Binary XML lightest-weight transform?
Database 11.2.0.2. I have a set of small XML documents stored in a schema-based binary XML column (pubdate, author, language, title, etc.). I'm using the XML, combined with a bit of longer (blob-based) text to generate a text index using a user_datas
-
Installing iWeb Update 2.0.2
Hi I am having trouble installing the update 2.0.2 for iWeb. I double click the install icon and the installation page appears then an alert boc appears stating 'An eligible iWeb application was not found in the location/Applications.' This is not th
-
Showing the music properties of .M4A files in Windows Explorer
Dear All, You probably noticed that Windows Vista does not provide much information or integration for Apple's .m4a music files. Would you like to see the music file properties for .m4a files in the columns and details pane of Windows Explorer? Would