BW statistics for ods objects

Hi Everyone,
In the search criteria I found documents only for bw statistics for infocubes. but not for ods objects.
After executing the queries for the ods objects in the bex, the entries are not made in rsddstat table, then where else it is store.
Can anyone please let me know how can I find the query runtime of ods objects.
Thanks,
Prashant.

Hi,
Yes I marked for the ods objects also for olap and whm, and executed the query in bex analyzer but still I do not get any entries for these objects in RSDDSTAT table.
What could be the problem. Is there a different procedure to find the query runtime of ods objects, please let me know.
Thanks,
Prashant.

Similar Messages

  • BW statistics for ODS?

    Hi,
    Is there any Statistics report which gives list of queries on ODS and query performances?
    I found reports for infocubes.
    I need to look at reports on ODS which run slow.
    Thanks,
    Sai.

    Hi,
    Look at this
    Have you mark these ods objects in rsa1->tools->bw statistics for infoprovider ? checkbox for olap and whm ticked
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/5401ab90-0201-0010-b394-99ffdb15235b

  • Unable tp create multi provider for  ODS objects

    Hello Gurus,
    I want to create a multi provider or an Z ods object. When I do the same it does not show me the Z ODS objects. But it shows me the Standard SAP ODS for creating multi provider.
    thanks in advance
    rgds

    Hi,
    For the ODS that you are trying to use, check if the BEx reporting has been turned on in the ODS object settings. If it is not, then you cannot use it in a multiprovider.
    Hope this helps...

  • Exporting and Importing Statistics for Schema objects.

    Hello All,
    I am trying to gather stats for optimization using, gather_schema_stats for all objects under schema. The manual what i am reading says it covers both Tables and Indexes and we can also include the all partions too. For the safety reasons i was informed by a friend that we should take a back up of old stats into a table (user defined), so as in case of messing up with new stats for performance issue we can use it as an backup copy. So, i have created a user table and exported the stats from existing schema. My Question is, do i have to create one for the indexes too. Or its all in one table. I hope my question is clear.
    Generally manuals don't teach all this kind of stuff until unless we learn it by ourselves by trial and error Or getting solutions from great resolvers like you people.
    Hope to hear soon.
    Thanks in Advance.

    One table for table and indexes stats is enough, a little test for you :
    SCOTT@demo102> exec DBMS_STATS.CREATE_STAT_TABLE('SCOTT','MYTBL');
    PL/SQL procedure successfully completed.
    SCOTT@demo102> desc mytbl
    Name                                      Null?    Type
    STATID                                             VARCHAR2(30)
    TYPE                                               CHAR(1)
    VERSION                                            NUMBER
    FLAGS                                              NUMBER
    C1                                                 VARCHAR2(30)
    C2                                                 VARCHAR2(30)
    C3                                                 VARCHAR2(30)
    C4                                                 VARCHAR2(30)
    C5                                                 VARCHAR2(30)
    N1                                                 NUMBER
    N2                                                 NUMBER
    N3                                                 NUMBER
    N4                                                 NUMBER
    N5                                                 NUMBER
    N6                                                 NUMBER
    N7                                                 NUMBER
    N8                                                 NUMBER
    N9                                                 NUMBER
    N10                                                NUMBER
    N11                                                NUMBER
    N12                                                NUMBER
    D1                                                 DATE
    R1                                                 RAW(32)
    R2                                                 RAW(32)
    CH1                                                VARCHAR2(1000)
    SCOTT@demo102> create table mytable_obj as select * from all_objects;
    Table created.
    SCOTT@demo102> create index myindex on mytable_obj(object_id);
    Index created.
    SCOTT@demo102> exec dbms_stats.gather_table_stats(ownname=>'SCOTT',tabname=>'MYTABLE_OBJ',cascade=>true);
    PL/SQL procedure successfully completed.
    SCOTT@demo102> select last_analyzed from all_tables where table_name='MYTABLE_OBJ';
    07/06/06
    SCOTT@demo102> select last_analyzed from all_indexes where table_name='MYTABLE_OBJ';
    07/06/06
    SCOTT@demo102> exec dbms_stats.export_table_stats('SCOTT','MYTABLE_OBJ',stattab=>'MYTBL')
    PL/SQL procedure successfully completed.
    SCOTT@demo102> exec dbms_stats.delete_table_stats('SCOTT','MYTABLE_OBJ');
    PL/SQL procedure successfully completed.
    SCOTT@demo102> select last_analyzed from all_tables where table_name='MYTABLE_OBJ';
    -->No value
    SCOTT@demo102> select last_analyzed from all_indexes where table_name='MYTABLE_OBJ';
    -->No value
    SCOTT@demo102>  exec dbms_stats.import_table_stats('SCOTT','MYTABLE_OBJ',stattab=>'MYTBL');
    PL/SQL procedure successfully completed.
    SCOTT@demo102> select last_analyzed from all_tables where table_name='MYTABLE_OBJ';
    07/06/06
    SCOTT@demo102> select last_analyzed from all_indexes where table_name='MYTABLE_OBJ';
    07/06/06
    SCOTT@demo102> Nicolas.

  • Time Charastrices for ODS objects

    Hello ALL
    When v are modelling ods object,I want to add <b>0calmonth</b> time charastrices to this ods.
    But Iam confused where to add this either to Key fields or Data fields
    Pl help me out
    Iam updating data from ODS to Cube.
    Regards
    BAlaji

    Hi,
    In the update rule of 0calmonth, generally we use automatic time conversion or routine to find the month from the date object which represents the record created date. So you can take it  either in data fields or in key fields.
    But if ocalmonth is being filled from the field which represents the Modification date of record ,it is better to include 0calmonth in data fields.
    With rgds,
    Anil Kumar Sharma .P
    Message was edited by: Anil Kumar Sharma

  • Error while activating ODS Object

    Hi,
    I have loaded the data from R/3 to a ODS Object and the activation of ODS failed due to the following errors as shown below..
    -Value '20060 ' of characteristic 0CALQUARTER is     not  plausible
    -Error when assigning SID (details in long text)
    -Activation of data records from ODS object ZODS_02 terminated
    -Error getting SID for ODS object ZODS_02
    I could see that Calendar Quarter as 20060 in PSA. How should i get rid of this error. Any idea would be appreciated.
    Regards,
    Raj

    Raj,
       you are getting this calender qurter from SAP right... check in RSA3 fro the your Datasource. if you are getting same value like '20060'. you can do one thing... you can set 'No Update' in Transfer Rules. you can write a routine in Update Rules for populating the data for calender Week.
    you can call some standard function modules or you can write a small routine. If you want to write a routine check this thread...Re: Date to Quarter
    Possibilites:
    1. updtae rules
    2. if RSA3(PSA) shows the value of the Quarter as '20050', if it is custom extractor(Generic using FM) you can change this.
    3. If it is Standard Extractor or Generic Extractor (using either view/table or Infoset) you can enhance this. you can do the same thing in SAP Side.
    All the best.
    Regards,
    Nagesh Ganisetti.

  • Error while activation of ODS Object.

    Hi,
    I have loaded the data from R/3 to a ODS Object and the activation of ODS failed due to the following errors as shown below..
    -Value '20060 ' of characteristic 0CALQUARTER is not plausible
    -Error when assigning SID (details in long text)
    -Activation of data records from ODS object ZODS_02 terminated
    -Error getting SID for ODS object ZODS_02
    I could see that Calendar Quarter as 20060 in table /BIC/AZINV_O0240. There is a calendarday mapped to quarter and year in update rules of ODS object. How the calendar quarter is populated in update table in BW. How should i get rid of this error. Any idea would be appreciated.
    Regards,
    Raj

    don't post same issue twice. Bhanu working on your thread.
    check this ...
    Raj,
    you are getting this calender qurter from SAP right... check in RSA3 fro the your Datasource. if you are getting same value like '20060'. you can do one thing... you can set 'No Update' in Transfer Rules. you can write a routine in Update Rules for populating the data for calender Week.
    you can call some standard function modules or you can write a small routine. If you want to write a routine check this thread...Re: Date to Quarter
    Possibilites:
    1. updtae rules
    2. if RSA3(PSA) shows the value of the Quarter as '20050', if it is custom extractor(Generic using FM) you can change this.
    3. If it is Standard Extractor or Generic Extractor (using either view/table or Infoset) you can enhance this. you can do the same thing in SAP Side.
    All the best.
    Regards,
    Nagesh Ganisetti.

  • Generated ODS Objects

    Hello All,
    I have 2 questions:
    1) I have loaded all my data from R/3 to ODS and now want to load from ODS to Cube. SO I have generated an export datasource by right clicking on the ODS and selecting "Generate Export..." and it gave a successful message.
       - Then I went into Infosources tree and from the menu Settings -> Display Generated Objects -> Choose "Show Generated ODS Objects".
       - Then I searched for the infosources starting with 8 but couldn't find any.
       Am I doing something wrong here? Is it the correct way to do that? Can I also upload into Cube from ODS by RC on the ODS and selecting "Update ODS Data in Data Target"? What is the difference?
    2) I have loaded the Sales Orders and Item data from R/3 by selecting "Initialization" in the Infopackage. I want to do the deltas now. I checked in LBWQ and found some records for application 11 & 12. But when I check in RSA7, it is showing '0' in Total field. Why is this happening? Do I need to do anything to before it gets updated in RSA7?
    Rgds,
    Sharath.

    Hi
    There is no need for you to genereted ODS object...it will be readily available under datamart infosource for ODS objects...only for cubes we need to do that...
    Search only under datamart infosources
    If it is showing delta records in RSA7, try executing the infopackage with delta
    if it is not showing delta records here.....for delta data records are not there in R/3
    Hope this helps
    Assign points if useful
    Regards
    N Ganesh

  • How to create DB partitioning in active data tables for ods?

    hi all,
    Can anyone let me know how to create DB partitioning in active data tables for ods. if any docs pls share with me at my email id : [email protected]
    regds
    haritha

    Haritha,
    The following steps will briefly explain you to improve the performance in terms of DB partitioning as well as loading. Please find the same,
    transaction RSCUSTA2,
    oss note 120253 565725 670208
    and remove 'bex reporting' setting in ods if that ods not used for reporting.
    hope this helps.
    565725
    Symptom
    This note contains recommendations for improving the load performance of ODS objects in Business Information Warehouse Release 3.0B and 3.1 Content.
    Other terms
    Business Information Warehouse, ODS object, BW, RSCUSTA2, RSADMINA
    Solution
    To obtain a good load performance for ODS objects, we recommend that you note the following:
    1. Activating data in the ODS object
    In the Implementation Guide in the BW Customizing, you can implement different settings under Business Information Warehouse -> General BW settings -> Settings for the ODS object that will improve performance when you activate data in the ODS object.
    1. Creating SIDs
    The creation of SIDs is time-consuming and may be avoided in the following cases:
    a) You should not set the indicator for BEx Reporting if you are only using the ODS object as a data store.Otherwise, SIDs are created for all new characteristic values by setting this indicator.
    b) If you are using line items (for example, document number, time stamp and so on) as characteristics in the ODS object, you should mark these as 'Attribute only' in the characteristics maintenance.
    SIDs are created at the same time if parallel activation is activated (see above).They are then created using the same number of parallel processes as those set for the activation. However:if you specify a server group or a special server in the Customizing, these specifications only apply to activation and not the creation of SIDs.The creation of SIDs runs on the application server on which the batch job is also running.
    1. DB partitioning on the table for active data (technical name:
    The process of deleting data from the ODS object may be accelerated by partitioning on the database level.Select the characteristic after which you want deletion to occur as a partitioning criterion.For more details on partitioning database tables, see the database documentation (DBMS CD).Partitioning is supported with the following databases:Oracle, DB2/390, Informix.
    1. Indexing
    Selection criteria should be used for queries on ODS objects.The existing primary index is used if the key fields are specified.As a result, the characteristic that is accessed more frequently should be left justified.If the key fields are only partially specified in the selection criteria (recognizable in the SQL trace), the query runtime may be optimized by creating additional indexes.You can create these secondary indexes in the ODS object maintenance.
    1. Loading unique data records
    If you only load unique data records (that is, data records with a one-time key combination) into the ODS object, the load performance will improve if you set the 'Unique data record' indicator in the ODS object maintenance.
    Hope this helps..
    ****Assign Points****
    Thanks,
    Gattu

  • Unable to see ODS Object in BEx.

    Hi all,
    I am unable to see an ODS object in BEx, which is an active ODS Object with data in modelling.
    Can any one show me a way out, how to find this ODS object in BEx as well.
    Thanks in advance,
    Regards.

    Hi Robert Zovic
    Thanks for your answer,
    When I switced on th same..following is the eror message:
    The BEx Reporting was switched on for ODS object 0RPM_DS02 ---> long text
    Message no. RSDODSO157
    System response
    When activating ODS object 0RPM_DS02 SIDs have to be got for all characteristic values in the ODS tables, provided that these characteristic values are not already assigned any SIDs. This is required because the SIDs are used in BEx Reporting.
    The activation of ODS object 0RPM_DS02 can thus take a while longer.
    Procedure
    If the activation of ODS object 0RPM_DS02 terminates due to timeout you have to activate the ODS object with the program RSDG_ODSO_ACTIVATE in the background.
    I'm working on this, is there any other way out for the
    same.
    Thanks.
    Regards.

  • BEx Authorisation: ODS objects

    Hi,
    RS_COMP only has a line for specifying authorisation for InfoCubes. Is there not an equivalent for ODS objects. RS_ODSO does not provide the same nature of authorisation.
    Any idea?
    Regards, Lea

    Hi Lea,
    specify the ODS name: the matchcode is a little bit unfaithfull. For more infos see OSS Note 540720 - "FAQ: Information on S_RS_COMP and S_RS_COMP1". There's written:
    1. How can I create authorizations for other InfoProviders besides InfoCubes? The authorization objects only provide an INFOCUBE field.
    The authorization object was created when only InfoCubes existed in BW, which explains why this field was selected.Since you can no longer change authorization objects, this name remains even though, technically speaking, the field is used to check the InfoProvider.This means that you can specify any InfoProvider (such as InfoCube, InfoSet, MultiCube) with the INFOCUBE field.         
    ... and I would add "ODSs"!
    Hope it helps
    GFV

  • Managing statistics for object collections used as table types in SQL

    Hi All,
    Is there a way to manage statistics for collections used as table types in SQL.
    Below is my test case
    Oracle Version :
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    PL/SQL Release 11.2.0.3.0 - Production
    CORE    11.2.0.3.0      Production
    TNS for IBM/AIX RISC System/6000: Version 11.2.0.3.0 - Production
    NLSRTL Version 11.2.0.3.0 - Production
    SQL> Original Query :
    SELECT
         9999,
         tbl_typ.FILE_ID,
         tf.FILE_NM ,
         tf.MIME_TYPE ,
         dbms_lob.getlength(tfd.FILE_DATA)
    FROM
         TG_FILE tf,
         TG_FILE_DATA tfd,
              SELECT
              FROM
                   TABLE
                        SELECT
                             CAST(TABLE_ESC_ATTACH(OBJ_ESC_ATTACH( 9999, 99991, 'file1.png', NULL, NULL, NULL),
                             OBJ_ESC_ATTACH( 9999, 99992, 'file2.png', NULL, NULL, NULL)) AS TABLE_ESC_ATTACH)
                        FROM
                             dual
         )     tbl_typ
    WHERE
         tf.FILE_ID     = tfd.FILE_ID
    AND tf.FILE_ID  = tbl_typ.FILE_ID
    AND tfd.FILE_ID = tbl_typ.FILE_ID;
    Elapsed: 00:00:02.90
    Execution Plan
    Plan hash value: 3970072279
    | Id  | Operation                                | Name         | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                         |              |     1 |   194 |  4567   (2)| 00:00:55 |
    |*  1 |  HASH JOIN                               |              |     1 |   194 |  4567   (2)| 00:00:55 |
    |*  2 |   HASH JOIN                              |              |  8168 |   287K|   695   (3)| 00:00:09 |
    |   3 |    VIEW                                  |              |  8168 |   103K|    29   (0)| 00:00:01 |
    |   4 |     COLLECTION ITERATOR CONSTRUCTOR FETCH|              |  8168 | 16336 |    29   (0)| 00:00:01 |
    |   5 |      FAST DUAL                           |              |     1 |       |     2   (0)| 00:00:01 |
    |   6 |    TABLE ACCESS FULL                     | TG_FILE      |   565K|    12M|   659   (2)| 00:00:08 |
    |   7 |   TABLE ACCESS FULL                      | TG_FILE_DATA |   852K|   128M|  3863   (1)| 00:00:47 |
    Predicate Information (identified by operation id):
       1 - access("TF"."FILE_ID"="TFD"."FILE_ID" AND "TFD"."FILE_ID"="TBL_TYP"."FILE_ID")
       2 - access("TF"."FILE_ID"="TBL_TYP"."FILE_ID")
    Statistics
              7  recursive calls
              0  db block gets
          16783  consistent gets
          16779  physical reads
              0  redo size
            916  bytes sent via SQL*Net to client
            524  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              2  rows processed Indexes are present in both the tables ( TG_FILE, TG_FILE_DATA ) on column FILE_ID.
    select
         index_name,blevel,leaf_blocks,DISTINCT_KEYS,clustering_factor,num_rows,sample_size
    from
         all_indexes
    where table_name in ('TG_FILE','TG_FILE_DATA');
    INDEX_NAME                     BLEVEL LEAF_BLOCKS DISTINCT_KEYS CLUSTERING_FACTOR     NUM_ROWS SAMPLE_SIZE
    TG_FILE_PK                          2        2160        552842             21401       552842      285428
    TG_FILE_DATA_PK                     2        3544        852297             61437       852297      852297 Ideally the view should have used NESTED LOOP, to use the indexes since the no. of rows coming from object collection is only 2.
    But it is taking default as 8168, leading to HASH join between the tables..leading to FULL TABLE access.
    So my question is, is there any way by which I can change the statistics while using collections in SQL ?
    I can use hints to use indexes but planning to avoid it as of now. Currently the time shown in explain plan is not accurate
    Modified query with hints :
    SELECT    
        /*+ index(tf TG_FILE_PK ) index(tfd TG_FILE_DATA_PK) */
        9999,
        tbl_typ.FILE_ID,
        tf.FILE_NM ,
        tf.MIME_TYPE ,
        dbms_lob.getlength(tfd.FILE_DATA)
    FROM
        TG_FILE tf,
        TG_FILE_DATA tfd,
            SELECT
            FROM
                TABLE
                        SELECT
                             CAST(TABLE_ESC_ATTACH(OBJ_ESC_ATTACH( 9999, 99991, 'file1.png', NULL, NULL, NULL),
                             OBJ_ESC_ATTACH( 9999, 99992, 'file2.png', NULL, NULL, NULL)) AS TABLE_ESC_ATTACH)
                        FROM
                             dual
        tbl_typ
    WHERE
        tf.FILE_ID     = tfd.FILE_ID
    AND tf.FILE_ID  = tbl_typ.FILE_ID
    AND tfd.FILE_ID = tbl_typ.FILE_ID;
    Elapsed: 00:00:00.01
    Execution Plan
    Plan hash value: 1670128954
    | Id  | Operation                                 | Name            | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                          |                 |     1 |   194 | 29978   (1)| 00:06:00 |
    |   1 |  NESTED LOOPS                             |                 |       |       |            |          |
    |   2 |   NESTED LOOPS                            |                 |     1 |   194 | 29978   (1)| 00:06:00 |
    |   3 |    NESTED LOOPS                           |                 |  8168 |  1363K| 16379   (1)| 00:03:17 |
    |   4 |     VIEW                                  |                 |  8168 |   103K|    29   (0)| 00:00:01 |
    |   5 |      COLLECTION ITERATOR CONSTRUCTOR FETCH|                 |  8168 | 16336 |    29   (0)| 00:00:01 |
    |   6 |       FAST DUAL                           |                 |     1 |       |     2   (0)| 00:00:01 |
    |   7 |     TABLE ACCESS BY INDEX ROWID           | TG_FILE_DATA    |     1 |   158 |     2   (0)| 00:00:01 |
    |*  8 |      INDEX UNIQUE SCAN                    | TG_FILE_DATA_PK |     1 |       |     1   (0)| 00:00:01 |
    |*  9 |    INDEX UNIQUE SCAN                      | TG_FILE_PK      |     1 |       |     1   (0)| 00:00:01 |
    |  10 |   TABLE ACCESS BY INDEX ROWID             | TG_FILE         |     1 |    23 |     2   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       8 - access("TFD"."FILE_ID"="TBL_TYP"."FILE_ID")
       9 - access("TF"."FILE_ID"="TBL_TYP"."FILE_ID")
           filter("TF"."FILE_ID"="TFD"."FILE_ID")
    Statistics
              0  recursive calls
              0  db block gets
             16  consistent gets
              8  physical reads
              0  redo size
            916  bytes sent via SQL*Net to client
            524  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              2  rows processed
    Thanks,
    B

    Thanks Tubby,
    While searching I had found out that we can use CARDINALITY hint to set statistics for TABLE funtion.
    But I preferred not to say, as it is currently undocumented hint. I now think I should have mentioned it while posting for the first time
    http://www.oracle-developer.net/display.php?id=427
    If we go across the document, it has mentioned in total 3 hints to set statistics :
    1) CARDINALITY (Undocumented)
    2) OPT_ESTIMATE ( Undocumented )
    3) DYNAMIC_SAMPLING ( Documented )
    4) Extensible Optimiser
    Tried it out with different hints and it is working as expected.
    i.e. cardinality and opt_estimate are taking the default set value
    But using dynamic_sampling hint provides the most correct estimate of the rows ( which is 2 in this particular case )
    With CARDINALITY hint
    SELECT
        /*+ cardinality( e, 5) */*
    FROM
         TABLE
              SELECT
                   CAST(TABLE_ESC_ATTACH(OBJ_ESC_ATTACH( 9999, 99991, 'file1.png', NULL, NULL, NULL),
                   OBJ_ESC_ATTACH( 9999, 99992, 'file2.png', NULL, NULL, NULL)) AS TABLE_ESC_ATTACH)
              FROM
                   dual
         ) e ;
    Elapsed: 00:00:00.00
    Execution Plan
    Plan hash value: 1467416936
    | Id  | Operation                             | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                      |      |     5 |    10 |    29   (0)| 00:00:01 |
    |   1 |  COLLECTION ITERATOR CONSTRUCTOR FETCH|      |     5 |    10 |    29   (0)| 00:00:01 |
    |   2 |   FAST DUAL                           |      |     1 |       |     2   (0)| 00:00:01 |
    With OPT_ESTIMATE hint
    SELECT
         /*+ opt_estimate(table, e, scale_rows=0.0006) */*
    FROM
         TABLE
              SELECT
                   CAST(TABLE_ESC_ATTACH(OBJ_ESC_ATTACH( 9999, 99991, 'file1.png', NULL, NULL, NULL),
                   OBJ_ESC_ATTACH( 9999, 99992, 'file2.png', NULL, NULL, NULL)) AS TABLE_ESC_ATTACH)
              FROM
                   dual
         ) e ;
    Execution Plan
    Plan hash value: 4043204977
    | Id  | Operation                              | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                       |      |     5 |   485 |    29   (0)| 00:00:01 |
    |   1 |  VIEW                                  |      |     5 |   485 |    29   (0)| 00:00:01 |
    |   2 |   COLLECTION ITERATOR CONSTRUCTOR FETCH|      |     5 |    10 |    29   (0)| 00:00:01 |
    |   3 |    FAST DUAL                           |      |     1 |       |     2   (0)| 00:00:01 |
    With DYNAMIC_SAMPLING hint
    SELECT
        /*+ dynamic_sampling( e, 5) */*
    FROM
         TABLE
              SELECT
                   CAST(TABLE_ESC_ATTACH(OBJ_ESC_ATTACH( 9999, 99991, 'file1.png', NULL, NULL, NULL),
                   OBJ_ESC_ATTACH( 9999, 99992, 'file2.png', NULL, NULL, NULL)) AS TABLE_ESC_ATTACH)
              FROM
                   dual
         ) e ;
    Elapsed: 00:00:00.00
    Execution Plan
    Plan hash value: 1467416936
    | Id  | Operation                             | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                      |      |     2 |     4 |    11   (0)| 00:00:01 |
    |   1 |  COLLECTION ITERATOR CONSTRUCTOR FETCH|      |     2 |     4 |    11   (0)| 00:00:01 |
    |   2 |   FAST DUAL                           |      |     1 |       |     2   (0)| 00:00:01 |
    Note
       - dynamic sampling used for this statement (level=2)I will be testing the last option "Extensible Optimizer" and put my findings here .
    I hope oracle in future releases, improve the statistics gathering for collections which can be used in DML and not just use the default block size.
    By the way, are you aware why it uses the default block size ? Is it because it is the smallest granular unit which oracle provides ?
    Regards,
    B

  • Linking two ODS objects using a multiprovider

    Hello Gurus,
    I need to link two ODS objects and both are having 3 common characteristics and the relationship between the two objects are one to many.
    My questions would be....
    1)When the relationship is one to many can I set it up in the multiprovider to sum it up and show in the report?
    2)I have 5 other charaterstics that I need to report(Slice and dice) on, but these are not common ones to both ODS objects. Is it possible for the user to get the report based on these charaterstics also? Or do I need to make sure that these 5 also exist in both DOS objects?
    Thank you,
    DM

    Hi Sh,
    Thank you very much for your reply. But I have few concerns.
    1)One to many relationship will not have any effect on the data linking since the multiprovider clubs the key figures on common characteristics. I tried this on the system and it worked.
    2)The problem would be not having similar characteristics on both infoproviders. So if I want to report on the chars individually it will have a problem( Will come as non assigned values)
    But the documentation says the multiprovider supports Homogenous and Heterogeneous cases. Please refer the following link.
    http://help.sap.com/saphelp_nw04/helpdata/en/52/1ddc37a3f57a07e10000009b38f889/frameset.htm
    I am not sure what I can do now....whether to get all the characteristics in both infoproviders by enhancing the them. Or is there another way?
    And lastly the scenario I am looking at is billing and cost of goods. So I might find records in cost of goods which may not have corresponding records in billing.(There can be goods in stock which need to be billed)The problem would be the way multiprovider link records(union) so it will have all the records from both sides even if they do not correspond to each other.
    Let me know what you think.
    Cheers,
    DM

  • How to create ods object with overwrite option

    Hi all,
    I have two ods objects for Billing and Delivery and one cube for Sales.
    Billing and Delivery ODS objects are based on item level data source while sales cube is based on Schedule line level data source.
    Now i want to create a new ODS object which will get the data from these there object and update the ods object.Basically i want all the sales order attributes and billing quantity and delivered quantity from the ODS objects.
    how to update the new ods object in overwrite mode ??/
    Regards,
    Viren.

    HI,
    yes OK but what i mean is that (you might have a perfect reason for this) normally it's nota good idea to have an info cube as a data source (for Performance, look-up reasons etc). If your goal is to load Billing/Delivery/Sales data into one ODS, then make sure that you model this correct.
    If you want to use the ODS'es as data sources I would recommend you to use and ODS as data source also for the Sales data.
    Kind regards
    /martin

  • Error activating ODS object.

    Hi Experts,
    None of the ODS objects gets activated and following are the error messages that i get when i try to activate a ODS object :
    Object ZRSMODS could not be activated
    <b>
    "</b>
    Message no. R7I028
    Transfer structure prefix for source system T90CLNT090 is not defined
    Message no. RSAR325
    Diagnosis
    A prefix is defined for the source system for determining cross-system unique transfer structure names. This prefix is not defined between your source system T90CLNT090 and Warehouse T90CLNT090. There was probably an error when creating the source system.
    Procedure
    If you have not yet loaded data from this source system, delete the source system and then recreate it.
    If the problem occurs again, check the SAP Support Portal. Create a new problem message, if necessary.
    Error when creating the export DataSource and dependent Objects
    Message no. RSDODSO169
    <b>"</b>                                                                               
    Pls send a solution asap. Thank you for the same.

    Hi Aravind,
    Have a look on the following points and see if they are satisfied in your case:
    Check all characteristics involved as described below:
    1. Check the Compounding and Attributes tabs in the characteristics maintenance.
    2. All characteristics here must be active. This is displayed by a green icon.
    If this is not the case for a characteristic involved (a gray icon is displayed on one of the two tabs), activate the characteristic involved. This also activates the inactive dependent characteristics.
    3. After you have checked and repaired all the characteristics involved as described above, the loading/maintenance of master data should work again.
    Because of this you may sometimes get the same error while activating the ODS object.
    Hope it helps...
    Best Regards,
    DMK
    *Assign points if it helps...

Maybe you are looking for