BW Queries mismatch in table and ST03

Hi
In St03, 3 queries has run in one infocube with 145 navigation step yesterday. When i checked the table 'RSDDSTAT' , there are infformation about 445 queries that run yesetrday for the same infocube.
is it the right way to check the number of queries that has been executed in BW system for any infocubes?
Thanks
Amar

Did you check the statistics settings ? RSA1 > Tools > BI statistics settings > Query tab

Similar Messages

  • Ora-01219 database not open; only queries on fixed tables and views.

    Hi everyone
    There is another post stating this error, but since the source is different, I think is okay to make a different post as well.
    Our database was working fine until the other day, when suddenly power went off, UPS was damaged and didn't hold, and when we restarted the database this is what happens:
    Using Enterprise Manager Console, we can make a succesful connection in the same machine, but if we choose Schema, the following error appears:
    ora-01219 database not open; only queries on fixed tables and views.
    When accesing from a different machine:
    ORA-01033: ORACLE initialization or shutdown in progress
    Thanks in advance

    Here is the last entries of alert log
    Fatal NI connect error 12638, connecting to:
    (DESCRIPTION=(LOCAL=YES)(ADDRESS=(PROTOCOL=beq)))
    VERSION INFORMATION:
         TNS for 32-bit Windows: Version 11.1.0.7.0 - Production
         Oracle Bequeath NT Protocol Adapter for 32-bit Windows: Version 11.1.0.7.0 - Production
    Time: 20-JUL-2010 15:48:31
    Tracing not turned on.
    Tns error struct:
    ns main err code: 12638
    TNS-12638: Fallo de recuperación de credenciales
    ns secondary err code: 0
    nt main err code: 0
    nt secondary err code: 0
    nt OS err code: 0
    Tue Jul 20 15:50:50 2010
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 2
    Using LOG_ARCHIVE_DEST_10 parameter default value as USE_DB_RECOVERY_FILE_DEST
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =18
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    Starting up ORACLE RDBMS Version: 11.1.0.7.0.
    Using parameter settings in server-side spfile E:\ORACLE\CRIMINALISTICA\PRODUCT\11.1.0\DB_1\DATABASE\SPFILECRIM.ORA
    System parameters with non-default values:
    processes = 150
    memory_target = 816M
    control_files = "E:\ORACLE\ORADATA\CRIM\CONTROL01.CTL"
    control_files = "E:\ORACLE\ORADATA\CRIM\CONTROL02.CTL"
    control_files = "E:\ORACLE\ORADATA\CRIM\CONTROL03.CTL"
    db_block_size = 8192
    compatible = "11.1.0.0.0"
    db_recovery_file_dest = "E:\Oracle\flash_recovery_area"
    db_recovery_file_dest_size= 2G
    undo_tablespace = "UNDOTBS1"
    remote_login_passwordfile= "EXCLUSIVE"
    db_domain = ""
    dispatchers = "(PROTOCOL=TCP) (SERVICE=CRIMXDB)"
    audit_file_dest = "E:\ORACLE\ADMIN\CRIM\ADUMP"
    audit_trail = "DB"
    db_name = "CRIM"
    open_cursors = 300
    diagnostic_dest = "E:\ORACLE"
    Tue Jul 20 15:50:55 2010
    PMON started with pid=2, OS id=1688
    Tue Jul 20 15:50:55 2010
    VKTM started with pid=3, OS id=2236 at elevated priority
    Tue Jul 20 15:50:55 2010
    DIAG started with pid=4, OS id=2332
    Tue Jul 20 15:50:55 2010
    DBRM started with pid=5, OS id=2328
    Tue Jul 20 15:50:55 2010
    PSP0 started with pid=6, OS id=2432
    VKTM running at (20)ms precision
    Tue Jul 20 15:50:55 2010
    DIA0 started with pid=7, OS id=2160
    Tue Jul 20 15:50:55 2010
    MMAN started with pid=8, OS id=2156
    Tue Jul 20 15:50:55 2010
    DBW0 started with pid=9, OS id=2064
    Tue Jul 20 15:50:55 2010
    LGWR started with pid=10, OS id=3176
    Tue Jul 20 15:50:55 2010
    CKPT started with pid=11, OS id=3180
    Tue Jul 20 15:50:55 2010
    SMON started with pid=12, OS id=3184
    Tue Jul 20 15:50:55 2010
    RECO started with pid=13, OS id=3188
    Tue Jul 20 15:50:55 2010
    MMON started with pid=14, OS id=3192
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    starting up 1 shared server(s) ...
    ORACLE_BASE from environment = E:\Oracle
    Tue Jul 20 15:50:55 2010
    MMNL started with pid=15, OS id=3196
    Tue Jul 20 15:50:56 2010
    alter database mount exclusive
    Setting recovery target incarnation to 2
    Successful mount of redo thread 1, with mount id 1304879008
    Database mounted in Exclusive Mode
    Lost write protection disabled
    Completed: alter database mount exclusive
    alter database open
    Errors in file e:\oracle\diag\rdbms\crim\crim\trace\crim_dbw0_2064.trc:
    ORA-01157: cannot identify/lock data file 9 - see DBWR trace file
    ORA-01110: data file 9: 'E:\ORACLE\ORADATA\STOPCRIM5\STOPCRIM5TABLESPACE.ORA'
    ORA-27047: unable to read the header block of file
    OSD-04006: fallo de ReadFile(); no se ha podido leer del archivo
    O/S-Error: (OS 38) Se ha alcanzado el final del archivo.
    Errors in file e:\oracle\diag\rdbms\crim\crim\trace\crim_dbw0_2064.trc:
    ORA-01157: cannot identify/lock data file 16 - see DBWR trace file
    ORA-01110: data file 16: 'E:\ORACLE\ORADATA\PRUEBACARLOS\PRUEBACARLOSTABLESPACE.ORA'
    ORA-27047: unable to read the header block of file
    OSD-04006: fallo de ReadFile(); no se ha podido leer del archivo
    O/S-Error: (OS 38) Se ha alcanzado el final del archivo.
    Errors in file e:\oracle\diag\rdbms\crim\crim\trace\crim_dbw0_2064.trc:
    ORA-01157: cannot identify/lock data file 17 - see DBWR trace file
    ORA-01110: data file 17: 'E:\ORACLE\ORADATA\PRUEBACARLOS\PRUEBACARLOSINDEX.ORA'
    ORA-27047: unable to read the header block of file
    OSD-04006: fallo de ReadFile(); no se ha podido leer del archivo
    O/S-Error: (OS 38) Se ha alcanzado el final del archivo.
    ORA-1157 signalled during: alter database open...
    Tue Jul 20 16:04:39 2010
    db_recovery_file_dest_size of 2048 MB is 0.00% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.

  • Data mismatch between TABLE and Std. Extract str. (AFPO & 2LIS_04_P_MATNR)

    Hi,
    I am getting data mismatch in ECC standard table and ECC standard extract structure. Please help me.
    I am using 2LIS_04_P_P_MATNR data source (standard extract structure).
    <b>I have following data in ECC Table (AFPO - Order Item).</b>
    AUFNR (Order Number) --> 4000460
    PGMNG (Total planned quantity)  --> 0
    PSMNG (ORDER QTY) --> 2000
    <b>Extract Structure is showing following data.</b>
    AUFNR (Order Number) --> 4000460
    PGMNG (Total planned quantity)  --> 2000 (It should be zero)
    PSMNG (ORDER QTY) --> 2000
    I have checked with many other order numbers. <b>The extract structure in RSA3 is showing correct values only if PGMNG is non-zero values. If PGMNG is zero, then it is automatically assigning PSMNG value.</b>
    Is this common? I want to display correct values for PGMNG from AFPO Table. Why this mismatch between table data and extract data. I have done  all required steps refreshing data (deleting setup tables, filling set up tables etc.).
    pls suggest. Poinsts will be assigned to every useful answer.
    Thanks in advance.

    Hi Avneet,
    I have a similar problem where in AFPO shows value 1000 in PSMNG field and 1200 in RSA3 for extractor 2LIS_04_P_MATNR.
    I have found that the standard feature of the extractor extracts the data at the time of release of process order. If the process order is changed after the release then the extractor will not picup the chagne.
    Ex: AFPO value 1000 ---BI Value 1000
    Now released in ECC
    AFPO value 1000 ---BI Value 1000
    In ECC value changed to 1500 after the release. BI will not pickup the change.
    Extractor is pulling the data from MCAFPOV table. You can check the values, there it matches with RSA3 values.
    I was just going thru forums for the solution. After i found the solution. I am just updating it.
    Hope it helps you.
    Thaks
    Srikanth

  • How to handle Datatype mismatch in table and EO

    Hi
    I have one table which store employee details, where empId of employee is of data type varchar2. I have its equivalent EO and VO with empId as Number type. when i create a row with empId as 100 its seen as -530000000000 something of this form.
    How to resolve this issue without disturbing database.

    Hello Suresh,
    Change the attribute in the entity object to VARCHAR2.
    Regards.

  • INITRANS mismatch for Table and its indexes

    What will happen if the INITRANS for a table is set as 50 and its index is set as 2?
    The number of simultaneous inserts is 32.
    Thanks in advance
    Aji

    This is an example that Oracle performs leaf node splits instead of waiting for ITL:
    SQL> SQL> drop table t cascade constraints purge;
    Table dropped.
    SQL> create table t(id int, pad char(50));
    Table created.
    SQL> insert into t select rownum, 'x' from dual connect by level <= 150;
    150 rows created.
    SQL> create index t_indx on t(pad) pctfree 0;
    Index created.
    SQL> analyze index t_indx validate structure;
    Index analyzed.
    SQL> select lf_rows, lf_blks from index_stats;
       LF_ROWS    LF_BLKS
           150          2
    SQL> set serveroutput on
    SQL>
    SQL> select name, value from v$statname n, v$sesstat s
      2   where name like '%splits%'
      3     and sid = userenv('sid')
      4     and n.statistic# = s.statistic#;
    NAME                                                                  VALUE
    leaf node splits                                                          0
    leaf node 90-10 splits                                                    0
    branch node splits                                                        0
    queue splits                                                              0
    SQL>
    SQL> declare
      2      procedure do_update(p_id number)
      3      is
      4          pragma autonomous_transaction;
      5      begin
      6          update t set pad = 'y' where id = p_id;
      7          if p_id <= 35 then
      8              do_update(p_id + 1);
      9          end if;
    10          commit;
    11      end;
    12  begin
    13      do_update(1);
    14  end;
    15  /
    PL/SQL procedure successfully completed.
    SQL>
    SQL> select name, value from v$statname n, v$sesstat s
      2   where name like '%splits%'
      3     and sid = userenv('sid')
      4     and n.statistic# = s.statistic#;
    NAME                                                                  VALUE
    leaf node splits                                                          1
    leaf node 90-10 splits                                                    0
    branch node splits                                                        0
    queue splits                                                              0
    SQL>
    SQL> analyze index t_indx validate structure;
    Index analyzed.
    SQL> select lf_rows, lf_blks from index_stats;
       LF_ROWS    LF_BLKS
           186          3
    SQL> select statistic_name, value from v$segment_statistics where object_name = 'T' and owner=user;
    STATISTIC_NAME                                                        VALUE
    logical reads                                                           176
    buffer busy waits                                                         0
    gc buffer busy                                                            0
    db block changes                                                        688
    physical reads                                                            0
    physical writes                                                           0
    physical reads direct                                                     0
    physical writes direct                                                    0
    gc cr blocks received                                                     0
    gc current blocks received                                                0
    ITL waits                                                                 0
    row lock waits                                                            0
    space used                                                                0
    space allocated                                                     1048576
    segment scans                                                             0
    15 rows selected.But increasing number of concurrent transactions fails with a deadlock:
    SQL> declare
      2      procedure do_update(p_id number)
      3      is
      4          pragma autonomous_transaction;
      5      begin
      6          update t set pad = 'y' where id = p_id;
      7          if p_id <= 50 then
      8              do_update(p_id + 1);
      9          end if;
    10          commit;
    11      end;
    12  begin
    13      do_update(1);
    14  end;
    15  /
    declare
    ERROR at line 1:
    ORA-00060: deadlock detected while waiting for resource
    ORA-06512: at line 6
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8
    ORA-06512: at line 8V$SEGMENT_STATISTICS does not reflect ITL waits (but they present in generated trace file) and leaf block split still happens. I don't know how to explain this.

  • Combine 2 Queries (from SAME table) into a SINGLE query

    I have this two queries (from SAME table), and want to combine into one SINGLE query, how?
    How can we use CASE WHEN THEN for such situation? 
    Query1:
    SELECT t_inner.*,
    Floor(t_inner.ProductiveTime/ 3600) || 'hr:' || LPAD(Floor(Mod(t_inner.ProductiveTime,3600) / 60),2,0) || 'min:' AS Productive_Time,
    Floor(t_inner.OperatorDownTime/ 3600) || 'hr:' || LPAD(Floor(Mod(t_inner.OperatorDownTime,3600) / 60),2,0) || 'min:' AS OperatorDown_Time
    FROM
    (SELECT SYSTEMTYPE,
    sum(TIME_TEST + TIME_STEP) AS ProductiveTime,
    sum(TIME_IDLE) AS OperatorDownTime
    FROM PFODS.PPL_TESTSYSTEMS_UTILISATION
    WHERE (SYSTEMTYPE = '0005-072')
    AND (TS_START >= to_date('13/01/2014', 'DD/MM/YYYY'))
    AND TS_End <= to_date('17/01/2014', 'DD/MM/YYYY') + 1 + (1/1440) +(59/86400)
    AND MONO != '9999999999'
    GROUP BY SYSTEMTYPE ) t_inner
    Query 2:
    SELECT t_inner.*,
    Floor(t_inner.MachineDownTime/ 3600) || 'hr ' || LPAD(Floor(Mod(t_inner.MachineDownTime,3600) / 60),2,0) || 'min' AS MachineDown_Time
    FROM
    (SELECT SYSTEMTYPE,
    sum(TIME_IDLE) AS MachineDownTime
    FROM PFODS.PPL_TESTSYSTEMS_UTILISATION
    WHERE (SYSTEMTYPE = '0005-072')
    AND (TS_START >= to_date('13/01/2014', 'DD/MM/YYYY'))
    AND TS_End <= to_date('17/01/2014', 'DD/MM/YYYY') + 1 + (1/1440) +(59/86400)
    AND MONO = '9999999999'
    GROUP BY SYSTEMTYPE) t_inner
    see http://postimg.org/image/koq87iyyz/  and
    http://postimg.org/image/fv3zxa38n

    with the first query, 
    SELECT t_inner.*,
    Floor(t_inner.ProductiveTime/ 3600) || 'hr:' || LPAD(Floor(Mod(t_inner.ProductiveTime,3600) / 60),2,0) || 'min' AS Productive_Time
    FROM
    (SELECT SYSTEMTYPE,
    --sum(TIME_TEST) AS TIME_TEST,
    --sum(TIME_SYSTEM) AS TIME_SYSTEM,
    --sum(TIME_STEP) AS TIME_STEP,
    --sum(TIME_IDLE) AS TIME_IDLE,
    sum(TIME_TEST + TIME_STEP) AS ProductiveTime
    FROM PFODS.PPL_TESTSYSTEMS_UTILISATION
    WHERE (SYSTEMTYPE = '0005-072')
    AND (TS_START >= to_date('13/01/2014', 'DD/MM/YYYY'))
    AND TS_End <= to_date('17/01/2014', 'DD/MM/YYYY') + 1 + (1/1440) +(59/86400)
    AND MONO != '9999999999'
    GROUP BY SYSTEMTYPE) t_inner
    it gives output as from 
    http://postimg.org/image/koq87iyyz/
    with the second query,
    SELECT t_inner.*,
    Floor(t_inner.MachineDownTime/ 3600) || 'hr ' || LPAD(Floor(Mod(t_inner.MachineDownTime,3600) / 60),2,0) || 'min' AS MachineDown_Time
    FROM
    (SELECT SYSTEMTYPE,
    sum(TIME_IDLE) AS MachineDownTime
    FROM PFODS.PPL_TESTSYSTEMS_UTILISATION
    WHERE (SYSTEMTYPE = '0005-072')
    AND (TS_START >= to_date('13/01/2014', 'DD/MM/YYYY'))
    AND TS_End <= to_date('17/01/2014', 'DD/MM/YYYY') + 1 + (1/1440) +(59/86400)
    AND MONO = '9999999999'
    GROUP BY SYSTEMTYPE) t_inner
    it gives output as from 
    http://postimg.org/image/fv3zxa38n/
    I want to come those 2 queries into a single query, such that it gives both outputs as above. Let me know if you need any other information. thanks.

  • 3 queries 3 tables and in a NON-relational DBMS

    3 queries 3 tables and in a NON-relational DBMS
    i got this msg box error:
    Error prop_id is a invalid identifier.
    basically i have querying form a non rel_dbms.
    with this statement
    select a.case_nbr, a.Aplc_id
    from Tcase a
    , tloc b
    , tla_prop c
    where c.plan_area_nbr = '1'
    and b.loc_id = a.Prop_id
    and a.Aplc_id = b.aplc_id
    thanks for responding
    here is the first
    query.
    select prop_id, plan_area_nbr
    from tla_prop
    where plan_area_nbr = '1'
    then i got a bunch of prop_id's from tla_prop that has plan_area_nbr of 1
    the second query will have
    select aplc_id, loc_id
    from tloc
    where loc_id is equal/the same prop_id from the table tla_prop
    the third selection
    will contain aplc_id from table tcase = aplc_is from the table tloc
    select case_nbr, Aplc_id
    from Tcase
    where Aplc_id = Secondselection.prop_id
    the goal is really to get the case_nbr with all criterias from query1,2,3 and table 1,2,3.
    any ideas?
    hmm, thanks in advance

    Hi
    Try this:
    select a.case_nbr, a.Aplc_id
    from Tcase a , tloc b , tla_prop c
    where c.plan_area_nbr = '1'
    and b.loc_id = c.Prop_id
    and a.Aplc_id = b.aplc_id
    Ott Karesz
    http://www.trendo-kft.hu

  • OBIEE bypasses smaller aggregate table and queries largest aggregate table

    Hello,
    Currently we are experiencing something strange regarding queries that are generated.
    Scenario:
    We have 1 detail table and 3 aggregate tables in the RPD. For this scenario I will only refer to 2 of the Aggregates.
    Aggregate 1 (1 million rows):
    Contains data - Division, Sales Rep, Month, Sales
    Aggregate 2 (13 milliion rows):
    Contains data - Division, Product, Month, Sales
    Both tables are set at the appropriate dimension levels in the Business Model. Row counts have been updated in the physical layer in the RPD.
    When we create an answers query that contains Division, Month and Sales, one would think that OBIEE would query the smaller and faster of the two tables. However, obiee wants to query the table with 13 million records completely bypassing the smaller table. If we make the larger aggregate inactive, then OBIEE queries the smaller table. We can't figure out why OBIEE wants to immediately go to the larger table.
    Has anyone experienced something such as this? Any help would be greatly appreciated.
    Edited by: gwb on Aug 19, 2009 7:45 AM

    Have you try to change the sort order of the logical table source in your logical table ?
    !http://gerardnico.com/wiki/_media/temp/obiee_logical_table_sources_sort.jpg!
    Set the Aggregate 1 first.
    Cheers
    Nico

  • Spatial Tables and Remote Queries

    Hi Experts,
    Can anyone explain me when and where to use Spatial Tables and Remote Queries with some examples.
    Thanks in advance.

    Hi,
    there are some nice demos from Andrejus:
    http://andrejusb.blogspot.com/search/label/Spatial
    Hope it helps,
    Friedhold

  • Mismatch in last entry in OJDT table and Nextval in series numbering setup

    Dear Support Team,
    We are getting FIRST DEFINE THE NUMBERING SERIES IN THE ADMINISTRATION MODULE ERROR ON ADDING ANY DOCUMENT.
    We checked for authorisation , group and posting periods. Everything in it's place.
    There is a mismatch in the last transid of OJDT table and the Nextval generated in document numbering series setup.
    How do we correct this.
    Any help would be greatly appreciated.
    Thanking you
    Pradnya

    Hi,
    We solved it by locking the erroneous series and by creating a new one. But your suggestion was more right and practical.
    Thanks for the valuable suggestion. We will keep this in mind.
    Thanking you
    Pradnya

  • How to join two tables and get the supply delivery date next to order?

    So there are two tables. One has customer's order no, ordered date, order quantity, available quantity and code of article-
    The other table comes form supply side where we have supply order no, article number, ordered qty, and delivery date.
    We keep stock so this can not be MOT (made to order) system.
    What i need is correct date of arrival to appear next to cusotmers spoecirfic order. The older cusotmers order get's the parts first, second oldest order is next in line etc.
    here is any example
    customer's order
    ref order
    art. code
    ordered qty
    available qty
    order date
    1809202491
    700497
    60
    0
    3.7.2014
    1809200528
    700497
    13
    0
    20.6.2014
    1809198640
    700497
    7
    0
    9.6.2014
    supply order
    supply order
    art. code
    qty orderd
    date of arrival
    4501243378
    700497
    50
    4.8.2014
    4501263437
    700497
    20
    6.10.2014
    There is actually a 3rd "table" and that sort of connects the two and that is stock on hand per art. code.
    The main issue is that stock is assigned to purchase orders only when it actually arrives in the warehouse.
    A human can easilly connect the dates of when the stock will arrive and quantities with correct customer's order. In this case the firts order will get 50 pcs in August while 10 pcs will remain on backorders. The missing 10 pcs Will arrive in October. The second order will get 10 pcs in october and 3 will remain on backorders with no delivery date. While the third customer orders does not have a delivery date.
    So how to make the SAP do this calculations and display the arrival date next to date of customer's order?

    I checked the instructions as i do not have access to this part. It seem this is a query. We had issues with queries in the past as not all codes from orders would appear in them. They never found the reason why that is happening.
    However, I think the main issue is that the information here is not connected and is separately provided for supply and for sales. So i doubt it can be connected in this query.
    edit: as you can see the only connection is stock on hand.
    and total number of various items we have is close to 100.000 of various article codes.

  • How to get table and column names thats being used in SQL , that's generating all my reports on SSRS.

    Good day,
    I searched through the forum and cant find anything.
    I have around 300 published reports on SSRS and we are busy migrating to a new system.
    They have already setup their tables on the new system and I need to provide them with a list of table names and column names that are being used currently to generate the 300 reports on SSRS.
    We use various tables and databases to generate these reports, and will take me forever to go through each query to get this info.
    Is it at all possible to write a query in SQL 2008 that will give me all the table names and columns being used?
    Your assistance is greatly appreciated.
    I thank you.
    Andre.

    There's no straightforward method for that I guess. There are couple of things you can use to get these details
    1. query the ReportServer.dbo.Catalog table
    for getting details
    you may use script below for that
    http://gallery.technet.microsoft.com/scriptcenter/42440a6b-c5b1-4acc-9632-d608d1c40a5c
    2. Another method is to run the reports and run sql profiler trace on background to retrieve queries used.
    But in some of these cases the report might be using a procedure and you will get only procedure. Then its upto you to get the other details from procedure like tables used, columns etc
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Two queries on same table in sender JDBC adapter

    Hi all..
    My requirement is as follows....
    query a table and fetch rows where column Error="true"
    If there are no records in the table satisfying the above condition ,i want to send a mail stating  that there are not records with error flag set to "true"
    But the problem is i wont get a message to PI if there are no records.
    If i write a query to get the count of records where column Error="true",Obviously the count would be zero. Using this message from database i can send a mail.
    Now please let me know if i can write both the queries in query SQL statement? if so howshould be the source structure?
    Thanks
    Ram..

    Hello ram,
                     As far as i know it is not possible to use two different SQL statements in the sender adapter, though you can try with the join statement.
    For further information refer this discussion
    [Multiple Select Possible?|How to call multiple SQL using the JDBC adapter;
    Regards,
    Prasanna

  • Join multiple fact tables and dimensions and use all tables in report issue

    Hi,
    I have a report requirements and need to use multiple fact tables and unconformed dimensions as described below
    Fact table: F1,F2,F3
    Dimensions tables: D1.....D9
    F1:(joined to) D1,D2,D3,D4
    F2::(joined to)D1,D2,D5,D6
    F3::(joined to)D1,D2,D7,D8
    D7::(joined to)D9,D8 (dimension D7 joined to two other dimensions D9 and D8
    I'm trying to use columns from almost all the fact and dimension tables but getting "Unable to navigate requested expression. Please fix the metadata consistency warnings."
    Repository is consistent and no errors and warnings.
    How can I configure the repository to develop reports using all fact tables and dimensions?
    Appreciate for your help.
    Thanks
    Jay.
    Edited by: Jay on Feb 9, 2012 4:14 PM

    So you want me to convert snowflake schema to star. does it solve my problem? individual star queries are working find but when I query multiple stars together getting inconsistency errors. I removed content tables dim level totals for unconformed dimensions in logical fact LTS and set level for measures at total level for unconformed dimensions. it is still in progress and need to test.
    Thanks
    Jay.

  • Are there any (dis)advantages in building a universe on fully normalized tables instead of building dimensional model/tables and then universe on top of them?

    Hello,
    I’m hoping someone can help me with understanding advantages and disadvantages if we want to build a universe on top of a fully normalized tables, compared to using a dimensional model with star schemas.
    I’ve read some discussions here that say that it is possible to create a universe on top of normalized tables. Then, can we avoid building of dimensional tables (a data mart), and just use normalized tables? I would say that it is easier to use star schema dimensions and facts tables to build a universe, but our end users might ask “why do we have to go through building a dimensional data mart, if we can have same reports with hierarchies and drill-down functionality based on a universe built on top of our already existing normalized tables?”
    Can you point me to some established best practices regarding using normalized tables to build a universe? Any documents with some examples for this?
    Any expected difficulties during design/development phase of our universe, related to using normalized tables?
    Any expected performance degradation if we use normalized tables compared to using dimensional tables?
    Can I build universe more easily if I transform (modify) our normalized model (by using alias tables and views) to look like snowflake model?
    I’m using BOE XI 3.1, tables are in Oracle 11.2.
    Thank you

    Few Disadvantages that I usually face when building universe on Normalized tables,
    1. Performance - Read operations have to suffer because indexing strategies do not go well with table joins
    2. Derived Tables - Due to complex Queries/ Logic, most of the time, end-up creating derived tables, which doesn't use back-end table indexes and slows down the report runtime.
    3. Normalized table/ Transaction tables may not always have proper cardinality and often results in Cartesian products
    4. Normalized tables may not have tight referential integrity and may have to join more than one column and join on varchar, etc whereas, good Dimensional model datawarehouse will have proper keys/ integer joins and not always necessary to join on multiple fields
    5. Often deal with Fan and Chasm Traps
    6. Dealing with Database fields with nulls, blanks, date in numeric format, etc.,
    7. No Facts, Dimensions separated and most of the time they are in same table
    and More...
    If performance is not a matter and building Datawarehouse is a big deal, then I will start building Universe on normalized tables by having the database diagram as reference for Joins, contexts, etc
    Note: After dealing with universes based on normalized tables for few years (by creating views, complex sql queries data loading to tables and unv on these tables, derived tables), I ended up creating star-schema dimensional model (couple of months extra ETL work), users/ developers felt lot better when they have to create standard/ ad-hoc reports and they are super fast compared to previous universes.

Maybe you are looking for