Discoverer 4i - Time Series Data type support

Does Discoverer 4i support time-series data type i.e. the ability to store an entire string of
numbers representing for example daily or weekly data points?
Thanks & Regards,
Deepti

Hi O G-M,
Each model must contain one numeric or date column that is used as the case series, which defines the time slices that the model will use. The data type for the key time column can be either a datetime data type or a numeric data type. However, the column must
contain continuous values, and the values must be unique for each series. The case series for a time series model cannot be stored in two columns, such as a Year column and a Month column. For more information about it, please see:
http://msdn.microsoft.com/en-us/library/ms174923(v=sql.100).aspx
Thanks,
Eileen
Eileen Zhao
TechNet Community Support

Similar Messages

  • Read optimization time-series data

    I am using Berkeley DB JE to store fairly high frequency (10hz) time-series data collected from ~80 sensors. The idea is to import a large number of csv files with this data, and allow quick access to time ranges of data to plot with a web front end. I have created a "sample" entity to hold these sampled metrics, indexed by the time stamp. My entity looks like this.
    @Entity
    public class Sample {
         // Unix time; seconds since Unix epoch
         @PrimaryKey
         private double time;
         private Map<String, Double> metricMap = new LinkedHashMap<String, Double>();
    as you can see, there is quite a large amount of data for each entity (~70 - 80 doubles), and I'm not sure storing them in this way is best. This is my first question.
    I am accessing the db from a web front end. I am not too worried about insertion performance, as this doesn't happen that often, and generally all at one time in bulk. For smaller ranges (~1-2 hr worth of samples) the read performance is decent enough for web calls. For larger ranges, the read operations take quite a while. What would be the best approach for configuring this application?
    Also, I want to define granularity of samples. Basically, If the number of samples returned by a query is very large, I want to only return a fraction of the samples. Is there an easy way to count the number of entities that will be iterated over with a cursor without actually iterating over them?
    Here are my current configuration params.
    environmentConfig.setAllowCreateVoid(true);
              environmentConfig.setTransactionalVoid(true);
              environmentConfig.setTxnNoSyncVoid(true);
              environmentConfig.setCacheModeVoid(CacheMode.EVICT_LN);
              environmentConfig.setCacheSizeVoid(1000000000);
              databaseConfig.setAllowCreateVoid(true);
              databaseConfig.setTransactionalVoid(true);
              databaseConfig.setCacheModeVoid(CacheMode.EVICT_LN);

    Hi Ben, sorry for the slow response.
    as you can see, there is quite a large amount of data for each entity (~70 - 80 doubles), and I'm not sure storing them in this way is best. This is my first question.That doesn't sound like a large record, so I don't see a problem. If the map keys are repeated in each record, that's wasted space that you might want to store differently.
    For larger ranges, the read operations take quite a while. What would be the best approach for configuring this application?What isolation level do you require? Do you need the keys and the data? If the amount you're reading is a significant portion of the index, have you looked at using DiskOrderedCursor?
    Also, I want to define granularity of samples. Basically, If the number of samples returned by a query is very large, I want to only return a fraction of the samples. Is there an easy way to count the number of entities that will be iterated over with a cursor without actually iterating over them?Not currently. Using the DPL, reading with a key-only cursor is the best available option. If you want to drop down to the base API, you can use Cursor.skipNext and skipPrev, which are further optimized.
    environmentConfig.setAllowCreateVoid(true);Please use the method names without the Void suffix -- those are just for bean editors.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Time-Series Data storage

    What's the most effiecient way to store time-series data in oracle?.
    Thanks,
    Jay.

    937054 wrote:
    Hello,
    1. Usally time-series data goes in multiple millions, so timeseries databases like FAME,KDB,SYBASE-IQ are used. Does oracle11gr2 provide storage optmizations, compressions, columnar database like FAME,KDB,SYBASE-IQ?
    The only methods of optimization are partitioning of the data by some date or if data set is narrow (few columns)l enough, partitioned IOT.
    2. http://www.oracle.com/us/corporate/press/1515738
    Link is about R statistical language and data mining integration with Oracle database 11gR2. Does this come default during installation or with BigData - EXADATA? OR this is a seperate license?I am not sure about the licensing, you will need to ask your sales person, but it looks like it might be apart of ODM (oracle data mining - a licensed product)
    Take a read through this case study.
    http://www.oracle.com/technetwork/database/options/advanced-analytics/odm/odmtelcowhitepaper-326595.pdf?ssSourceSiteId=ocomen
    >
    Thanks

  • Power Spectrum Density conversion to Time Series Data

    Hi,
    This may seem an odd request but is there a way to convert power spectrum density data back to the time series data that generated it in the first place. I have lost the original time series data but still have the PSD and need the time series to do other analysis.
    Thanks,
    Rhys Williams

    Hate to be the bearer of bad news, but there are an infinite number of time series that will generate a given PSD.  You lose all phase information upon taking the PSD.  For this reason I almost always save time domain data, or at least complex FFT values.  

  • Relevance of Time series data in PPDS

    Hi, we have an SNP planning area with some times series KFs in it. if we delete the time series for the PA what impact it would have on PPDS..
    or what would be the relevance of times series data in PPDS?..
    kindly explain.

    the only relation of time series data to PP/DS I know of is  Time-dependent safety stock planning in PP/DS.
    Safety and Target Stock Level Planning in PP/DS
    In Customizing for PP/DS, define the following:
    ·         Make SNP key figures available
    You define which SNP key figures PP/DS should use as safety stock or daysu2019 supply. (Make SNP Key Figures Available)
    ·         Define planning area
    You specify an SNP planning area that contains the key figures for the time-dependent safety stock/daysu2019 supply levels. (Global Settings ® Maintain Global Parameters and Defaults)
    if you did not use the planning area and key figure there should be no influence.
    Frank

  • Time Series Data Mining

    All,
    I'd like to know how to model a time series table to submit it to Oracle Data Mining. I've already tried a reverse pivot, but is there something different?
    Regards,
    Paulo de Tarso Costa de Sousa

    If you are trying to do something like create a demand forecasting model like ARIMA, ODM does not have explicit support for this type of modeling. Data mining is usually more about creating a "prediction" rather than a forecast. You may want to look at Oracle's OLAP capabilities for this.
    If you are trying to include variables that contain the element of "time", such as "blood pressure before" and "blood pressure after", you can include these a variables (attributes) in the model. ODM has no real limit on the number of variables it can include in the model, so you don't have to worry about creating too many of them (usually).
    You may want to "clump" the data so as to create a set of variables at certain check points in time like the "before" and "after" approach above. Rather than entering for example the measurement off an instrument ever 10 seconds (which would ordinarily create new variables for each time period), you may want to only detect "events". That is, only record the amount of time between events--sort of Mean Time Between Failure (MTBF) type of modeling.
    Hope this helps with your thinking about how to approach your problem

  • Tg4msql - Blob Data Type Support(Oracle 8i - sql server 2000)

    hi,
    We are trying to replicate the data from oracle 8i to sql server 2000 through tg4msql.if we want to replicate the data thet contains the blob field , we are getting invalid expression in the procedure marked with error line.
    the procedure i am using is as follows
    CREATE OR REPLACE PROCEDURE INSERT_SQL
    AS
    BEGIN
    FOR VAL_CURSOR IN
    --Get the Data from the table with status 0
    SELECT rs_report_rowid FROM BIJU.rs_report where rs_report_status = 0
    LOOP
    for new_cursor in
    --get the details of the record with the row id
    select id ,name,DOB from biju.sample where rowid = val_cursor.rs_report_rowid
    LOOP
    ---insert the data from oracle to sql where the oracle data type contains blob(here photo field data type is blob).
    Error Line: insert into [email protected](id,name,photo) values(new_cursor.id,new_cursor.NAME,new_cursor.DOB);
    END LOOP;
    END LOOP;
    COMMIT;
    END;
    i want to know that tg4msql won't support the blob data type on the oracle end?we configured blob eqivalent in sql server as binary.Any help or suggestion regarding this are welcome..
    regards
    Srini

    Srini,
    Support for LOB was added in v9.2. It is not supported in 8i.

  • Data Types supported by Express Edition

    Dear All,
    Is BLOB compatable in Express edition.
    What are the data-types are supported by Express edition.
    Thanks & regards,
    Abk

    http://download.oracle.com/docs/cd/B25329_01/doc/appdev.102/b25108/xedev_design.htm#XEDEV02000
    HTH...

  • Data types supported in weblogic web service

    I have used byte[] (base64) array to pass binary data using weblogic in a web service, since byte[] is a java datatype that is supported, is there a definite guideline available that byte [][] ( multi-dimensional array} is not supported in weblogic 8.1? (I have tried and failed with byte [][])
    Thanks

    Hi gchirrav ,
    Multi-Dimentional array support not listed in supported Data Types in weblogic 8.1.
    Go through the following link, you will find supported data types.
    http://e-docs.bea.com/wls/docs81/webserv/implement.html#1054236
    ----Anilkumar kari

  • Object Data Type support and JDBC2 SQLData interface

    Another useful and probably easy to implement feature is to support
    persistent-aware classes based on JDBC SQLData interface._
    It will let us persist such a member of persistent class to an object
    type column._
    The interface is very simple and quite limited (no support for arrays
    etc) but useful.
    The only thing we would need is to maintain jdbc object maps for our
    database connections and tell Kodo metadata not to enhance this
    particular class member but treat is as a simple type_
    Alex

    In SPROXY you can open your proxy and then view the Activation Log under the GoTo menu.  The log will explain better what the problems might be.  In my case I needed to activate another data type enhancement first.
    Thanks,
    Matt

  • SAP BW Time characteristics data type issue in BO OLAP Universe

    Hi Ingo,
    We have time characteristics in our SAP BW Queries, for example OCALMONTH with data type NUMC.
    When we create OLAP Universe these objects are having Character as data type and not working properly when we use these objects as filters. The filter working on string instead of calender month.
    If i try to change the data type in Universe we are not able to execute Web Intelligence queries due to Driver Not Capable error.
    Appreciate you help.
    Regards,
    Ravi Kumar Garre

    Hi,
    Please find my linline answers:
    - are you entering the value manually or did you select the value from a list of values ?
    I am selecting the values from list.
    - based on which item is the list of values ?
    List of valules are coming from BI infoObject 0CALMONTH
    - what about creating a variable for this in the underlying BW query ?
    We have created a BI Variable, found that the Condition operator is Between and hence when execute the WebI query for this object there are two prompts one is for FROM value another is for TO value. If i give a value Dec, 2007 for FROM and leave TO as blank then i am getting data for all months instead of DEC 2007 onwards.
    - did you trace the Web Intleligence part to see the details ?
    I do not have authorization to login on server and find the trace. I will ask our BO Administrator for MDX log file.
    Thanks & Regards,
    Ravi Kumar Garre

  • How to do an average on time series data?

    I need to generate average hold times for various stock of companies as follows:
    The data looks like:
    stock        timestamp (sec)            quantity
    GOOG          12459.6                    -100        <-- SALE
    GOOG          12634.0                    +100        <-- PURCHASE
    GOOG          12636.2                    +200
    GOOG          12464.8                    -100
    GOOG          12568.3                    -300
    GOOG          12678.0                    +200
    The rules are
    1. begin and end day with balance 0
    2. can short sell, i.e. can sell shares even if balance is currently 0
    3. hold time is defined as number of seconds stock was held before it was sold
    4. first stock purchased are sold first
    I need to generate the average hold times seconds per share.  I'd prefer to do this using SQL and NOT a procedure.
    Any tips on how to go about calculating this?  I have looked at various analytic functions, but still not sure.
    Thank you.

    I'm afraid you might be after something like below:
    this is a simplified scenario where the quantity balance always reaches 0 before changing sign (not very probable in real life)
    Simple examples are reserved for the lecturer was a pretty common phrase in my university times
    I dont know how to deal with a general case yet
    select * from trade_0 order by position,time
    TIME
    POSITION
    DIRECTION
    QUANTITY
    8
    GOOG
    S
    100
    13
    GOOG
    B
    20
    16
    GOOG
    B
    30
    17
    GOOG
    B
    30
    19
    GOOG
    B
    20
    22
    GOOG
    B
    20
    25
    GOOG
    B
    30
    26
    GOOG
    B
    20
    30
    GOOG
    B
    30
    34
    GOOG
    B
    20
    38
    GOOG
    B
    30
    41
    GOOG
    S
    150
    7
    YHOO
    S
    10
    12
    YHOO
    S
    20
    15
    YHOO
    S
    30
    16
    YHOO
    S
    40
    18
    YHOO
    S
    60
    21
    YHOO
    S
    30
    24
    YHOO
    S
    10
    25
    YHOO
    B
    100
    29
    YHOO
    B
    300
    33
    YHOO
    S
    100
    37
    YHOO
    S
    80
    40
    YHOO
    S
    20
    your condition 4. first stock purchased are sold first requires a procedural solution so model clause must be used if you want to do it in SQL.
    Model Men, bear with me, please !
    select m.*,
           avg(abs(x_time - decode(kind,'B',time_b,time_s))) over (partition by position
                                                                       order by rn rows between unbounded preceding
                                                                                            and unbounded following
              ) average
      from (select *
              from (select nvl(b.position,s.position) position,
                           nvl(b.rn,s.rn) rn,
                           nvl(b.cnt,0) cnt_b,
                           nvl(s.cnt,0) cnt_s,
                           b.time time_b,
                           s.time time_s,
                           b.quantity qty_b,
                           s.quantity qty_s
                      from (select time,position,quantity,
                                   row_number() over (partition by position order by time) rn,
                                   count(*) over (partition by position) cnt
                              from trade_0
                             where direction = 'B'
                           ) b
                           full outer join
                           (select time,position,quantity,
                                   row_number() over (partition by position order by time) rn,
                                   count(*) over (partition by position) cnt
                              from trade_0
                             where direction = 'S'
                           ) s
                        on b.position = s.position
                       and b.rn = s.rn
             model
             partition by (position)
             dimension by (rn)
             measures (0 loc,
                       case when cnt_b >= cnt_s then 'B' else 'S' end kind,
                       time_b,
                       time_s,
                       qty_b,
                       qty_s,
                       0 qty_left,
                       0 x_time
             rules iterate (1000000) until (loc[iteration_number] is null)
              loc[0] = nvl2(loc[0],loc[0],1),
              qty_left[loc[0]] = case when iteration_number > 0
                                      then qty_left[loc[0]] + case when kind[iteration_number] = 'B'
                                                                   then qty_b[iteration_number]
                                                                   else qty_s[iteration_number]
                                                              end
                                      else 0
                                 end,
              x_time[iteration_number] = case when kind[iteration_number] = 'B'
                                              then time_s[loc[0]]
                                              else time_b[loc[0]]
                                         end,
              loc[0] = loc[0] + case when qty_left[loc[0]] = case when kind[iteration_number] = 'B'
                                                                  then qty_s[loc[0]]
                                                                  else qty_b[loc[0]]
                                                             end
                                     then 1
                                     else 0
                                end
           ) m
    where kind is not null
    order by position,rn
    POSITION
    RN
    LOC
    KIND
    TIME_B
    TIME_S
    QTY_B
    QTY_S
    QTY_LEFT
    X_TIME
    AVERAGE
    GOOG
    1
    0
    B
    13
    8
    20
    100
    100
    8
    10.4
    GOOG
    2
    0
    B
    16
    41
    30
    150
    150
    8
    10.4
    GOOG
    3
    0
    B
    17
    30
    8
    10.4
    GOOG
    4
    0
    B
    19
    20
    0
    8
    10.4
    GOOG
    5
    0
    B
    22
    20
    0
    41
    10.4
    GOOG
    6
    0
    B
    25
    30
    0
    41
    10.4
    GOOG
    7
    0

  • JAX-RPC 9.0.4   --- Data Types Supported

    Hi,
    does any body know is java.util.List is a supported Datatype on OC4j 9.0.4.2 for JAX-RPC.
    Cheers
    Gaurav

    Gaurav,
    Not sure which version of JAX-RPC you are intended to run on top of 9.0.4.2 ?
    In any way, if you want your service to work with the majority of the clients, you need to focus on the XML types; not the java types.
    To my knowledge, the List will be mapped to a proprietary type, at best, which make it harder to re-use.
    All the best,
    Eric

  • Select time from date type without to_char

    I have table T(myDate date, + about 15 columns). When I insert values into it, i also insert the time of a date.
    If I wanna see all the data in the table, I need to call to_char on column myDate. So need to write all the 15 column names in the select.
    I'd like to simply write a select * from T, but see the time for myDate as well. Is this possible?
    Thanks

    You can use alter session in a SQLPlus session
    See this example:
    SQL>
    SQL> select * from emp
      2  where rownum <=2;
         EMPNO ENAME      JOB              MGR HIREDATE         SAL       COMM     DEPTNO
          7369 SMITH      CLERK           7902 17-DEC-80        800                    20
          7499 ALLEN      SALESMAN        7698 20-FEB-81       1600        300         30
    SQL>
    SQL> alter session set nls_date_format='DD-MON-YYYY HH24:MI:SS';
    Session altered.
    SQL>
    SQL> select * from emp
      2  where rownum <=2;
         EMPNO ENAME      JOB              MGR HIREDATE                    SAL       COMM     DEPTNO
          7369 SMITH      CLERK           7902 17-DEC-1980 00:00:00        800                    20
          7499 ALLEN      SALESMAN        7698 20-FEB-1981 00:00:00       1600        300         30Regards.
    Al

  • Data type support  9i Release 2

    Does oracle 9i relase 2 support the float datatype for replication .

    As a matter of fact oracle 9i has two types of api for external tables one is oracle_loader and the other is oracle_internal which in 10g become oracle_datapump.
    What my question is how can I use the other type in case of external table as we are doing in 10g? Do I need to run any script to enable this feature or is there any privillege issue?
    I've given the read, write privillege to the user and the oracle user has read/write permission in os level on the location of the directory.

Maybe you are looking for