Statistics in Diadem

Hello everyone,
I have a problem with statistics and respectively histogram generation from a set of data
Below you can see  how my data channels look like. 
In DIAdem VIEW I have ->
<All Channels>
Name  |Ch1 Dev1|Ch2 Dev1|Ch3 Dev1|Ch1 Dev2|Ch2 Dev2|Ch3 Dev2|Ch1 Dev3|Ch2 Dev3|Ch3 Dev3|
Number |1             | 2            |3            |4             |5             |6             |7            |8             |9             |
Length  |3             | 3            |3            |3             |3             |3             |3            |3             |3             |
Unit     | V            | V            | V           | V            | V            | V            | V           | V            | V            |
Channel contents
1         | 1.52      | 1.51       | 1.53       | 1.52       | 1.55       | 1.51       | 1.50     | 1.53        | 1.55       |
2         | 1.62      | 1.61       | 1.63       | 1.62       | 1.65       | 1.60       | 1.62     | 1.67        | 1.65       |
3         | 1.52      | 1.51       | 1.53       | 1.52       | 1.55       | 1.51       | 1.50     | 1.53        | 1.55       |
What I need is to create a histogram for a given row ->in this case number 2
It this straight forward or should I make some adjustments before that?
I hope my question is clear and is easy to follow.
Thanks in advance to anyone who can propose a working solution. 
Solved!
Go to Solution.

Hi,
in the histogram generation function in Diadem Analysis you can choose how many classes you need.
The other settings can be left to automatic. Then you get 2 result channels SampleCount and ClassMean. Class Mean
are the classes and SampleCount is how many values are in every class.
In my case I choosed 5 classes and the result diagram is attached. You can see that most values are in class 1.62
Regards
DianaS
Attachments:
NONAME.pdf ‏21 KB

Similar Messages

  • Using DIAdem, how to get the slope of a signal (Linear or non-linear​)

    Rate of rise or rate of fall on a temperature signal, I'm trying to include the slope in my report using DIAdem,
    I couldn't find , Is there a slope function?
    or a way to do that
    Thanks
    Solved!
    Go to Solution.

    Hi amat,
    You can get the slope fit calculation from the built-in function "ANALYSIS >> Curve Fitting >> Calculate Regression".  If you want to fit only a subset of the data, you will need to extract that subset of data to a new data channel and then run the approximation calculation on the new channel.
    Here is a script I have created to do the range extraction and statistics/slope calculation automatically.
    Brad Turpin
    DIAdem Product Support Engineer
    National Instruments
    Attachments:
    Selected Region Stats.zip ‏5 KB

  • DIAdem Standard Deviation Weighting

    I have been doing some data analysis in DIAdem; I have been taking some statistics on some data - working out variance and standard deviation.  I asked NI Support team whether 'Population' or 'Sample' weighting is used, as it was not mentioned in the help file. I understand that by default Variance and Standard Deviation are calculated using the 'Sample' rather than 'Population' formula. My company uses Population to work out some of the key performance data of the products we make.  At the moment I don't see a way of selecting which formula to use so a little drop down next to the checkboxes so I can pick the type would be great.  For now I'm going to have to develop some VBScript or start using excel to do my data analysis.  As a footnote to this idea - may I also illustrate that calculating standard deviation directly feeds into process capability (Cp and Cpk).  Sample weighting is a little less friendly than Population weighting, so unwitting process control engineers may be finding their processes a little more (I think!) capable than what they truly are.
    Regards,
    Sam
     

    Is this an idea about DIAdem?  There is a separate idea exchange for that. 
    DIAdem Idea Exchange
    I'll ask a moderator to move the idea to there.

  • Statistics Currency Error while posting Sales Order (No: V1453)

    Hi Sappers,
    I am in process of making a company in SAP ERP SD - IDES, and while posting Sales Order, after I input Sold to Party, PO no, Del Plant and Incoterms and I press ENTER: The error comes up as:
    Statistics: The currency from in INR for date 25.11.2011 could not be determined.
    Kindly suggest a possible solution.
    Thanks
    Rahul Tikku

    Hi,
    Just go to "OBD2", select the Accounts group of your customer and "Double Click" on it, then "Double  cliick" on "Sales Data" > Further goto,  "Sales" and check if "Currency" field is suppressed. if so then change it to required or optional Entry. save and update you customer master in XD02. Then try your process.
    Reagrds
    DSR

  • How oracle decide whetehr to use index or full scan (statistics)

    Hi Guys,
    Let say i have a index on a column.
    The table and index statistics has been gathered. (without histograms).
    Let say i perform a select * from table where a=5;
    Oracle will perform a full scan.
    But from which statistics it will be able to know indeed most of the column = 5? (histograms not used)
    After analyzing, we get the below:
    Table Statistics :
    (NUM_ROWS)
    (BLOCKS)
    (EMPTY_BLOCKS)
    (AVG_SPACE)
    (CHAIN_COUNT)
    (AVG_ROW_LEN)
    Index Statistics :
    (BLEVEL)
    (LEAF_BLOCKS)
    (DISTINCT_KEYS)
    (AVG_LEAF_BLOCKS_PER_KEY)
    (AVG_DATA_BLOCKS_PER_KEY)
    (CLUSTERING_FACTOR)
    thanks
    Index Column (A)
    ======
    1
    1
    2
    2
    5
    5
    5
    5
    5
    5

    I have prepared some explanation and have not noticed that the topic has been marked as answered.
    This my sentence is not completely true.
    A column "without histograms" means that the column has only one bucket. More correct: even without histograms there are data in dba_tab_histograms which we can consider as one bucket for whole column. In fact these data are retrieved from hist_head$, not from histgrm$ as usual buckets.
    Technically there is no any buckets without gathered histograms.
    Let's create a table with skewed data distribution.
    SQL> create table t as
      2  select least(rownum,3) as val, '*' as pad
      3    from dual
      4  connect by level <= 1000000;
    Table created
    SQL> create index idx on t(val);
    Index created
    SQL> select val, count(*)
      2    from t
      3   group by val;
           VAL   COUNT(*)
             1          1
             2          1
             3     999998So, we have table with very skewed data distribution.
    Let's gather statistics without histograms.
    SQL> exec dbms_stats.gather_table_stats( user, 'T', estimate_percent => 100, method_opt => 'for all columns size 1', cascade => true);
    PL/SQL procedure successfully completed
    SQL> select blocks, num_rows  from dba_tab_statistics
      2   where table_name = 'T';
        BLOCKS   NUM_ROWS
          3106    1000000
    SQL> select blevel, leaf_blocks, clustering_factor
      2    from dba_ind_statistics t
      3   where table_name = 'T'
      4     and index_name = 'IDX';
        BLEVEL LEAF_BLOCKS CLUSTERING_FACTOR
             2        4017              3107
    SQL> select column_name,
      2         num_distinct,
      3         density,
      4         num_nulls,
      5         low_value,
      6         high_value
      7    from dba_tab_col_statistics
      8   where table_name = 'T'
      9     and column_name = 'VAL';
    COLUMN_NAME  NUM_DISTINCT    DENSITY  NUM_NULLS      LOW_VALUE      HIGH_VALUE
    VAL                     3 0,33333333          0           C102            C104So, Oracle suggests that values between 1 and 3 (raw C102 and C104) are distributed uniform and the density of the distribution is 0.33.
    Let's try to explain plan
    SQL> explain plan for
      2  select --+ no_cpu_costing
      3         *
      4    from t
      5   where val = 1
      6  ;
    Explained
    SQL> @plan
    | Id  | Operation         | Name | Rows  | Cost  |
    |   0 | SELECT STATEMENT  |      |   333K|   300 |
    |*  1 |  TABLE ACCESS FULL| T    |   333K|   300 |
    Predicate Information (identified by operation id):
       1 - filter("VAL"=1)
    Note
       - cpu costing is off (consider enabling it)Below is an excerpt from trace 10053
    BASE STATISTICAL INFORMATION
    Table Stats::
      Table:  T  Alias:  T
        #Rows: 1000000  #Blks:  3106  AvgRowLen:  5.00
    Index Stats::
      Index: IDX  Col#: 1
        LVLS: 2  #LB: 4017  #DK: 3  LB/K: 1339.00  DB/K: 1035.00  CLUF: 3107.00
    SINGLE TABLE ACCESS PATH
      BEGIN Single Table Cardinality Estimation
      Column (#1): VAL(NUMBER)
        AvgLen: 3.00 NDV: 3 Nulls: 0 Density: 0.33333 Min: 1 Max: 3
      Table:  T  Alias: T
        Card: Original: 1000000  Rounded: 333333  Computed: 333333.33  Non Adjusted: 333333.33
      END   Single Table Cardinality Estimation
      Access Path: TableScan
        Cost:  300.00  Resp: 300.00  Degree: 0
          Cost_io: 300.00  Cost_cpu: 0
          Resp_io: 300.00  Resp_cpu: 0
      Access Path: index (AllEqRange)
        Index: IDX
        resc_io: 2377.00  resc_cpu: 0
        ix_sel: 0.33333  ix_sel_with_filters: 0.33333
        Cost: 2377.00  Resp: 2377.00  Degree: 1
      Best:: AccessPath: TableScan
             Cost: 300.00  Degree: 1  Resp: 300.00  Card: 333333.33  Bytes: 0Cost of FTS here is 300 and cost of Index Range Scan here is 2377.
    I have disabled cpu costing, so selectivity does not affect the cost of FTS.
    cost of Index Range Scan is calculated as
    blevel + (leaf_blocks * selectivity + clustering_factor * selecivity) = 2 + (4017*0.33333 + 3107*0.33333) = 2377.
    Oracle considers that it has to read 2 root/branch blocks of the index, 1339 leaf blocks of the index and 1036 blocks of the table.
    Pay attention that selectivity is the major component of the cost of the Index Range Scan.
    Let's try to gather histograms:
    SQL> exec dbms_stats.gather_table_stats( user, 'T', estimate_percent => 100, method_opt => 'for columns val size 3', cascade => true);
    PL/SQL procedure successfully completedIf you look at dba_tab_histograms you will see following
    SQL> select endpoint_value,
      2         endpoint_number
      3    from dba_tab_histograms
      4   where table_name = 'T'
      5     and column_name = 'VAL'
      6  ;
    ENDPOINT_VALUE ENDPOINT_NUMBER
                 1               1
                 2               2
                 3         1000000ENDPOINT_VALUE is the column value (in number for any type of data) and ENDPOINT_NUMBER is cumulative number of rows.
    Number of rows for any ENDPOINT_VALUE = ENDPOINT_NUMBER for this ENDPOINT_VALUE - ENDPOINT_NUMBER for the previous ENDPOINT_VALUE.
    explain plan and 10053 trace of the same query:
    | Id  | Operation                   | Name | Rows  | Cost  |
    |   0 | SELECT STATEMENT            |      |     1 |     4 |
    |   1 |  TABLE ACCESS BY INDEX ROWID| T    |     1 |     4 |
    |*  2 |   INDEX RANGE SCAN          | IDX  |     1 |     3 |
    Predicate Information (identified by operation id):
       2 - access("VAL"=1)
    Note
       - cpu costing is off (consider enabling it)
    BASE STATISTICAL INFORMATION
    Table Stats::
      Table:  T  Alias:  T
        #Rows: 1000000  #Blks:  3106  AvgRowLen:  5.00
    Index Stats::
      Index: IDX  Col#: 1
        LVLS: 2  #LB: 4017  #DK: 3  LB/K: 1339.00  DB/K: 1035.00  CLUF: 3107.00
    SINGLE TABLE ACCESS PATH
      BEGIN Single Table Cardinality Estimation
      Column (#1): VAL(NUMBER)
        AvgLen: 3.00 NDV: 3 Nulls: 0 Density: 5.0000e-07 Min: 1 Max: 3
        Histogram: Freq  #Bkts: 3  UncompBkts: 1000000  EndPtVals: 3
      Table:  T  Alias: T
        Card: Original: 1000000  Rounded: 1  Computed: 1.00  Non Adjusted: 1.00
      END   Single Table Cardinality Estimation
      Access Path: TableScan
        Cost:  300.00  Resp: 300.00  Degree: 0
          Cost_io: 300.00  Cost_cpu: 0
          Resp_io: 300.00  Resp_cpu: 0
      Access Path: index (AllEqRange)
        Index: IDX
        resc_io: 4.00  resc_cpu: 0
        ix_sel: 1.0000e-06  ix_sel_with_filters: 1.0000e-06
        Cost: 4.00  Resp: 4.00  Degree: 1
      Best:: AccessPath: IndexRange  Index: IDX
             Cost: 4.00  Degree: 1  Resp: 4.00  Card: 1.00  Bytes: 0Pay attention on selectivity, ix_sel: 1.0000e-06
    Cost of the FTS is still the same = 300,
    but cost of the Index Range Scan is 4 now: 2 root/branch blocks + 1 leaf block + 1 table block.
    Thus, conclusion: histograms allows to calculate selectivity more accurate. The aim is to have more efficient execution plans.
    Alexander Anokhin
    http://alexanderanokhin.wordpress.com/

  • Forms 6.0 - how to have forms utilization statistics?

    I would like to know how i can retrieve statistics of form utilization during one period without recompile any form. I wanna know for instance the 20 forms more used on a Oracle Forms 6.0 app. Is it possible?
    thanks
    Luis Reis

    Frank
    thanks for your answer but the problem is that one . I have 800 forms and we have no time to open all them and make this script ( I am a CIO and not a tech guy so I presume that you are saying that we must write code in the Form, no?)
    Luis Reis

  • DIAdem 9.1; Windows2000; right problem; "Configuration-Drive Write Protected" for USER

    Hallo community !
    similar to these threads
    http://forums.ni.com/ni/board/message?board.id=60&message.id=322&requireLogin=False
    http://forums.ni.com/ni/board/message?board.id=60&message.id=647&requireLogin=False
    i run into problems working with DIAdem 9.1 on Windows 2000 SP4!
    The Diadem installation went quite well.
    My problem:
    only "Administrator" and "Power User" have the privilige to "WRITE" to the DIAdem install directory.
    Trying to start Diadem as non priviliged USER I get something like:
    DIAdem Configuration-Drive "C:\PROGRAM FILES\ NATIONAL INSTRUMENT\DIADEM 9.10" WRITE PROTECTED!
    I installed & started DIAdem for one
    time with Administrator rights. I activated the Product and exported
    the KEY to a License Server. Everything works fine for Power Users an
    Administrators. Only "simple" USERs are not able to start Diadem due to
    missing WRITE permissions.  i can't grand this permissions to
    normal USERs :-(
    Playing around in Registry -> "RegWriteAccessAllow" ("0","1") (DWORD) (HKEY_LOCAL_MACHINE/Software/National Instruments/Diadem/9.10),
    didn't solved the problem.
    I switched back to "local License" Managment (local Key files) but the Probelm persists.
    thanks in advance
    P.S.
    Quote: "You have to be at least a "Power User" in Windows 2000 to run DIAdem in Windows 2000."
    http://forums.ni.com/ni/board/message?board.id=60&message.id=102&requireLogin=False
    is this right ?
    Message Edited by DirkT on 08-01-2005 03:58 PM
    Message Edited by DirkT on 08-01-2005 04:02 PM

    Hi DirkT,
    Running DIAdem with User rights has different aspects we need to look at.
    1) Registry Write Access
    Launching DIAdem the first time with administrative rights registers some OLE components in the registry. After this step no more Admin rights are necessary for the registry. However running DIAdem as user might result in error messages as DIAdem tries to re-register the components again. You can avoid this by setting the RegWriteAccessAllow Key to 0 (see online help under "Administrator Notes")
    2) File / Folder Write Access to the DIAdem Installation Folder
    DIAdem needs write access to certain files e.g. the error logfile DEM_ERRS.001 or DESKTOP.DDD when saving configurations. You may change the folder where these file are being stored and searched for to a writeable folder by launching DIAdem with the "/d<NewFolder\MyDesktop.ddd>" start parameter or by setting the environment variable "DemDDD" to that folder. Then move the following files from the DIAdem installation folder to your new writeable folder:
    *.ddd
    userpar.par
    dem_errs.*
    3) File / Folder Write Access to other Folders
    There are a number of other folders that DIAdem 9.x needs write access to and which cannot be changed. All of these folders are located by default under "%ProgramFiles%\National Instruments\Shared". Currently you need your Admin to give you write access to that folder and below
    For future versions of DIAdem the team is working to make running under user rights easier.
    Hope that helps,
    Marcus

  • Upload data from excel to Ztable with statistics

    Hi,
    I have a requirement to upload data from excel sheet to ztable .
    Here i need tp provide the user with the execution statistics like
    1.Number of records read from the Excel spread-sheet
    2. Number records processed successfully
    3. Number records with Error
    4. Name and location of Error Log-file (text-file format)
    5. Name and location of the file containing error records (Excel spread-sheet format)
    I would appreciate if any of you have code written for the same

    See the below example code to upload from xl file to sap
    REPORT ZLWMI151_UPLOAD no standard page heading
                           line-size 100 line-count 60.
    *tables : zbatch_cross_ref.
    data : begin of t_text occurs 0,
           werks(4) type c,
           cmatnr(15) type c,
           srlno(12) type n,
           matnr(7) type n,
           charg(10) type n,
           end of t_text.
    data: begin of t_zbatch occurs 0,
          werks like zbatch_cross_ref-werks,
          cmatnr like zbatch_cross_ref-cmatnr,
          srlno like zbatch_cross_ref-srlno,
          matnr like zbatch_cross_ref-matnr,
          charg like zbatch_cross_ref-charg,
          end of t_zbatch.
    data : g_repid like sy-repid,
           g_line like sy-index,
           g_line1 like sy-index,
           $v_start_col         type i value '1',
           $v_start_row         type i value '2',
           $v_end_col           type i value '256',
           $v_end_row           type i value '65536',
           gd_currentrow type i.
    data: itab like alsmex_tabline occurs 0 with header line.
    data : t_final like zbatch_cross_ref occurs 0 with header line.
    selection-screen : begin of block blk with frame title text.
    parameters : p_file like rlgrap-filename obligatory.
    selection-screen : end of block blk.
    initialization.
      g_repid = sy-repid.
    at selection-screen on value-request for p_file.
      CALL FUNCTION 'F4_FILENAME'
           EXPORTING
                PROGRAM_NAME = g_repid
           IMPORTING
                FILE_NAME    = p_file.
    start-of-selection.
    Uploading the data into Internal Table
      perform upload_data.
      perform modify_table.
    top-of-page.
      CALL FUNCTION 'Z_HEADER'
      EXPORTING
        FLEX_TEXT1       =
        FLEX_TEXT2       =
        FLEX_TEXT3       =
    *&      Form  upload_data
          text
    FORM upload_data.
      CALL FUNCTION 'ALSM_EXCEL_TO_INTERNAL_TABLE'
           EXPORTING
                FILENAME                = p_file
                I_BEGIN_COL             = $v_start_col
                I_BEGIN_ROW             = $v_start_row
                I_END_COL               = $v_end_col
                I_END_ROW               = $v_end_row
           TABLES
                INTERN                  = itab
           EXCEPTIONS
                INCONSISTENT_PARAMETERS = 1
                UPLOAD_OLE              = 2
                OTHERS                  = 3.
      IF SY-SUBRC <> 0.
        write:/10 'File '.
      ENDIF.
      if sy-subrc eq 0.
        read table itab index 1.
        gd_currentrow = itab-row.
        loop at itab.
          if itab-row ne gd_currentrow.
            append t_text.
            clear t_text.
            gd_currentrow = itab-row.
          endif.
          case itab-col.
            when '0001'.
              t_text-werks = itab-value.
            when '0002'.
              t_text-cmatnr = itab-value.
            when '0003'.
              t_text-srlno = itab-value.
            when '0004'.
              t_text-matnr = itab-value.
            when '0005'.
              t_text-charg = itab-value.
          endcase.
        endloop.
      endif.
      append t_text.
    ENDFORM.                    " upload_data
    *&      Form  modify_table
          Modify the table ZBATCH_CROSS_REF
    FORM modify_table.
      loop at t_text.
        t_final-werks = t_text-werks.
        t_final-cmatnr = t_text-cmatnr.
        t_final-srlno = t_text-srlno.
        t_final-matnr = t_text-matnr.
        t_final-charg = t_text-charg.
        t_final-erdat = sy-datum.
        t_final-erzet = sy-uzeit.
        t_final-ernam = sy-uname.
        t_final-rstat = 'U'.
        append t_final.
        clear t_final.
      endloop.
      delete t_final where werks = ''.
      describe table t_final lines g_line.
      sort t_final by werks cmatnr srlno.
    Deleting the Duplicate Records
      perform select_data.
      describe table t_final lines g_line1.
      modify zbatch_cross_ref from table t_final.
      if sy-subrc ne 0.
        write:/ 'Updation failed'.
      else.
        Skip 1.
        Write:/12 'Updation has been Completed Sucessfully'.
        skip 1.
        Write:/12 'Records in file ',42 g_line .
        write:/12 'Updated records in Table',42 g_line1.
      endif.
      delete from zbatch_cross_ref where werks = ''.
    ENDFORM.                    " modify_table
    *&      Form  select_data
          Deleting the duplicate records
    FORM select_data.
      select werks
             cmatnr
             srlno from zbatch_cross_ref
             into table t_zbatch for all entries in t_final
             where werks = t_final-werks
             and  cmatnr = t_final-cmatnr
             and srlno = t_final-srlno.
      sort t_zbatch by werks cmatnr srlno.
      loop at t_zbatch.
        read table t_final with key werks = t_zbatch-werks
                                    cmatnr = t_zbatch-cmatnr
                                    srlno = t_zbatch-srlno.
        if sy-subrc eq 0.
          delete table t_final .
        endif.
        clear: t_zbatch,
               t_final.
      endloop.
    ENDFORM.                    " select_data
    Reward Points if it is helpful
    Thanks
    Seshu

  • UCCE 8.5.3/8.5.4 call volume statistics not matching in interval tables

    hello,
    We have just migrated a call center to UCCE 8.5.3 that runs roggers and ICM call flow scripts that contain very basic flows. In each flow there is basically a one-to-one ratio of call type elements to select skill group elements.
    Generally you would expect the call volume (as our scripts are written) to have almost the same call volume in the call type interval table and the skill group interval table, with allowable differences because of RONAs, etc. But basically somewhat close.
    In general this does work but after a reboot of the roggers (A and B sides) (logger down first, then router and reversed to start), the skill group interval data lags greatly from the call type interval data, meaning maybe 10% of the call type interval volume. We have learned that completely shutting down both A and B side and then briinging it back up in order from scratch seems to alleviate the problem (until the next restart). We cannot just leave the servers alone though due to company security patching policies.
    Cisco TAC had recommended patching to UCCE 8.5.4 which we have recently done, but the probelm persists.
    I was wondering if this data discrepancy has ever been seen by anyone else or if possibly some system config issue might be self-deating us? Rebooting the roggers leaves the phone system itself working just fine for the call centers, but the recording of statistics to the skill group interval table is greatly impacted with almost no recording of volume in relation to the call type interval (and for comparison, the call type skill group interval table as well).
    We would generally not worry about it, but the workforce management vendor that takes its feed from UCCE only uses the skill group interval, which is basically reporting almost no volume in the mentioned scenarios.
    If anyone can provide any information it would be most appreciated/
    Thanks.
    Greg

    Thank you for the response. The time source check is a great idea. We ran into problems before when internal web service servers did not match the PGs (CTIOS service) and CTI provideed stats did not match.
    We will continue to work through TAC, but I was just wondering if anyone else had seen this (as 8.5.4 sid not fix it) and if maybe it could have been something self-defeating in our system configuration or scripting. We did not immediately know this was happening until our 3rd party workforce management vendor made us aware.
    Thanks,
    Greg

  • Test Cases required for BW Statistics to test in QA annd DEV.

    HI All,
    I am currently working on a support Project.  My client has completed installing of Bw statistics in DEV and transported it to QA way back in 2006.Currrently before moving the BI Statistics data to PRD we have to test it in DEV and QA.
    How to prepare sample test case for testing it in DEV and QA? Please Sugggest.

    Hi,
    this forum is for the SAP BusinessObjects BI Solution architecture. I would suggest you post your question to the BW forum.
    ingo

  • Y axis auto scale not working in DIAdem 11

    Hi
    A few years ago I created a user dialog box for selecting and displaying channels in DIAdem view.
    It worked fine for all DIAdem 10.X releases, but for DIAdem 11.X there is a problem with the y-axis scaling.
    The channels are displayd as [n  Systems (Linear)]
    In DIAdem 10.X the y-axis for each separate graph would be scaled automatically accoring to the min. and max. values of the displayed channel.
    In DIAdem 11.X the y-axis for each graph is scaled from 55.0397 to 133.7655.... for all channels, no matter what the values are for the displayed channel.
    When I click the "Zoom Off" button the scaling becomes correct.
    What do I need to put in to my script to make the y-axis scale correctly?
    Below is the code that is used to set the scaling (no scaling for Y axis is set):
        R1 = SelectedStartTime 'this is a variable defined earlier in the code
        R2 = SelectedEndTime 'this is a variable defined earlier in the code
        View.ActiveSheet.Cursor.X1 = R1
        View.ActiveSheet.Cursor.X2 = R2
        View.ActiveSheet.Areas(1).DisplayObj.XScaling = "RangeZoom"
    I've tried adding this code, to no effect:
        View.ActiveSheet.Areas(1).DisplayObj.YScaling = "n systems [phys.]"
        View.ActiveSheet.Areas(1).DisplayObj.YScalingMode = "automatic"
    Solved!
    Go to Solution.

    I tried with Macro recording and Ctrl+A and the result was
    Option Explicit 'Forces the explicit declaration of all the variables in a script.
    R1 = 30
    R2 = 52
    Call View.NewLayout()
    View.Sheets("Sheet 1").Areas("Area : 1").DisplayObjType = "CurveChart2D"
    Dim oMyDisplayObj
    Set oMyDisplayObj = View.Sheets("Sheet 1").Areas("Area : 1").DisplayObj
    oMyDisplayObj.Curves.Add "[1]/Time", "[1]/Speed"
    oMyDisplayObj.YScalingMode ="automatic"
    oMyDisplayObj.Abscissa ="X-Cursor"
    oMyDisplayObj.Ordinate ="Y-Cursor"
    View.ActiveSheet.Cursor.X1 = R1
    View.ActiveSheet.Cursor.X2 = R2
    View.Sheets("Sheet 1").Areas("Area : 1").DisplayObj.XScalingMode = "RangeZoom"
    WndShow "VIEW", ""
    which seems to work fine. So there is only one difference to your code which is
    .XScalingMode
    instead of XScaling. Does changing this work with your code?

  • Contract net value for Header Statistics is not correct

    There is issue with value contract. The net value for Header Statistics is not showing correctly for some contracts. Especially, when we delete the PO line items or reverse all entries (GR and IR) for PO line item.
    Contract has one line with account category ‘U’. The target value is 300,000.00 and total quantity released to date is 160,000. The net value for Header Statistics should be 140,000 but it is showing 600,000 which is over (double) the target value and user cannot release any further PO reference to this contract.
    Earlier I defined net price for line item 300,000 and I changed net price to zero and execute report RM06ENP0 but it doesn’t work.
    Please share your experience and thoughts.
    Thanks,
    Shah.

    Hi Jurgen,
    There are few Purchace orders with multiple line items and each line item for Purchase orders referencing the same line and same contract.
    There is only one Purchase order has two deleted lines against this contract.
    Theses deleted line's net price has changed to zero and there is no PO history.
    Contract released order value is correct as there is only one line, but net price is wrong. and user is getting error for target value is excedeed by $nnn when trying to create purchase order.
    Thanks,
    Shah.

  • Report using BI Statistics

    People, I activated the BI Statistics in SAP/BW.
    The Multicube 0BWTC_C10 is active.
    Will I can generate a report related to load data (time of loading, the load status, etc), using this multicube (0BWTC_C10)?

    Hi Akash,
    The below standard queries will provide you the required details,
    Info
      Provider
    Bex Query
      Technical Name
    Bex Query
      description
    0TCT_MC11
    0TCT_MC11_Q0140
    InfoCube Status
    0TCT_MC11
    0TCT_MC11_Q0141
    InfoCube
      Correctness
    0TCT_MC11
    0TCT_MC11_Q0240
    InfoCube Status: Analysis
    0TCT_MC22
    ZCS_CUBE_DATADETAIL_STAT
    ZCS_CUBE_DATADETAIL_STAT
    0TCT_MC21
    ZTCT_MC21_Q_FB_01
    Dashboard - process
      chain historical loading time
    0TCT_MC22
    ZCS_STAT_SPEND_LOADS
    Statistics Spend
      Overview Loads
    0TCT_MC22
    ZCS_STAT_SPEND_LOADS_PERF
    Statistics Spend
      Overview Loads Performance
    0TCT_MC22
    ZTCT_MC22_Q_FB_02
    Dashboard - DSO
      loading time top 10
    0TCT_MC22
    ZTCT_MC22_Q_FB_04
    Dashboard - IC
      loading time top 10
    0TCT_MC22
    ZTCT_MC22_Q_FB_06
    Dashboard - IO
      loading time top 10
    -Arun.M.D

  • Improving graphics performance in DIAdem 11.0 Report?

    Dear Forum,
    when working on a Report layout, the poor graphic performance of DIAdem is a real problem for me.
    For example, many diagrams with lots of data are on one page. One opens the proficiencies of one of the diagrams and changes e.g. the curves color or whatever and closes the dialog window. After this, the whole Report page is refreshed - wich takes up to one minute, depending on the amount of data displayed.
    The strange thing is, when I close the given dialog window, and while the report page is being refreshed sent DIAdem to the taskbar and open any other application window, the process of refreshing the screen takes only 1 second instead of 30 seconds.
    The question is: Is there a tweak to improve the speed how DIAdem refreshes the report page? Or is there any workaround other than switching applications during refreshing?
    Thanks a lot.
    Greetings from Germany
    Oliver

    Hello Oliver,
    The easiest way to speed up the graphics performance in Report with large data files (or 3D graphs with lots of nodes) is to turn off the "Data Layer" (German: "Datenebene") in the "View" menu (German: "Ansicht").
    That function basically turns off the display of all data in Report, but retains the axis systems, texts, etc., so you can still move them, place them and edit them without having to guess sizes. Once you're done with the report, simply turn the data layer back on and you're ready to create a PDF, print or save the result as an image file (JPG, BMB, etc.).
    I think the performance difference you described (1 vs. 30 seconds) is due to the fact that once DIAdem has drawn a page, it retains an image in memory and will simply draw that image when you bring DIAdem to the foreground (or maximize it). After any edits you make to the report (e.g. change the color of a graph), DIAdem redraws the complete page from scratch, which takes much longer then just re-displaying an image that has already been drawn previously. Someone from the DIAdem R&D team could probably do a better job explaining this, but I think this is the explanation for the time difference in drawing.
    Anyway, the "Data Layer" trick should hopefully solve your problem, let us know if that works for you.
    Greeting from Austin, TX.
    Otmar D. Foehner
    Business Development Manager
    DIAdem and Test Data Management
    National Instruments
    Austin, TX - USA
    "For an optimist the glass is half full, for a pessimist it's half empty, and for an engineer is twice bigger than necessary."

  • Index Statistics Update - Problem

    We had performance problem yesterday with FI report FAGLL03, it timed out in online execution and in background mode it took 5000+ sec to execute. Result was no more than 100 records.
    Later with some investigation problem drill down to index usage of table FAGLFLEXA. We then updated the index statistics of table from DB02. After that report worked fine with execution time of 10-15 sec for same set of input.
    However user, in morning , was complaining again about performance problem with same report FAGLL03. We did that update index statistics again and as it was the case yesterday it fixed the problem.
    Later today I checked SQL server the job SAP CCMS_xxx_xxx_Update_Tabstats, which I guess is updating index statistics daily at 0400 hours, is working fine. I can't see any error log there. Daily job to check database consistency is also not reporting anything.
    Anyidea what could be going wrong.
    Basis Consultants are looking into problem however I am putting this case here if anyone of you had same problem and fixed it.
    Thanks,
    Pawan.
    Edited by: Pawan Kesari on Dec 11, 2009 4:05 PM

    Hi,
    Appears the stats are dropped eveytime the job runs @04:00
    Have a look at the table DBSTATC in trx: DB21 to see if it's setup to dropped the stats..
    Mark

Maybe you are looking for

  • Ipod shuffle 2 nd gen itunes 9 ipod not recognised (windows 7 64 bit OS)

    Hi I see that others are having issue. My problem is that I have windows 7 ultimate installed (64 bit) along with itunes 9 (64 version) . When i plug in my ipod shuffle 2nd gen 1gb I get the following message " the software required for communicating

  • H.264 Vimeo Export Questions

    Hello, I exported a H.264 according to Vimeo's guidelines ( Compression Tutorials on Vimeo ) for Premiere Pro CC.  I'm a bit dissapointed in the quality of the export - contrast is down, midtones are brighter, a noticeable loss of detail and added no

  • Dv2941se hard disk no boot with error Err2Err3

    On my dv2941se laptop with AMD 64 x2 TK-57,4096MB total Memory Bios versionF.26 Running Vista 64 bit home premium. When I turn on to boot  I get the HP screen with function keys and then goes to black screen with cursor in upper left corner. When boo

  • Linksys e2000 router Open NAT troubleshooting multiple xbox 360's

    Hello, I've been searching for about 4 hours now how to fix my NAT problems. I thought it had it working on one xbox in my bedroom, as it said OPEN NAT. But, when I looked at my Xbox in the living room, it said MODERATE NAT. How can I get these both

  • CCS3 3D rendering not working correctly

    I create online virtual tours. These tours use HTML5 CCS3 to display the cube faced images as a spherical panorama. All my tours work correctly in FF10 on mac and PC, but testing in FF11 on PC has shown a problem. Here is an example link http://obero