Stats in Cube Operation Log

Hello,
Does anyone know what does the 'STAT_COMP_TIME' operation that is logged in the Cube_Operations_log table mean?
The 'TOTAL_TIME' stats in the Cube_Operations_log table for one of my queries is about 6 secs but 'STAT_COMP_TIME' is around 7 secs.
Thanks

Thanks
One of the reasons why I am asking these questions is that there seem to be quite a big overhead querying the dimension/cube views through SQL
For e.g
select dim_key from time_view where level_name = 'YEAR'; takes around ~ 4sec on my current hardware and database configuration
But same query in the OLAP Worksheet using DML e.g. limit TIME to (inlist('YEAR' TIME_LEVELREL)) and report TIME returns almost instantly.
Another e.g. querying a cube view takes roughly around 5 secs (using Cube_Table, and ensuring LOOP Optimized is on)
select sales from sales_cube_view where customer = 'ALL' and time = '2010' and location = 'ALL'
But same query in the OLAP Worksheet using DML
limit SALES_CUBE_MEASURE_DIM to 'SALES'
limit customer to 'ALL'
limit location to 'ALL'
limit time to '2010'
report sales_cube_stored
The data is returned almost instantly
I'm on version 11.2.0.2. It will be good to understand what is the additional overhead that's taking quite a significant portion of the time.
Thanks

Similar Messages

  • Rejcted records cube build log.

    I am on 11.1.0.7 db with 11.1.0.7B AWM and loading the dimensions and cube. The olapsys.xml_load_log table is not populated with the build log. I can query cube_build_log but xml_load_log give me better information(rejected/processed records). I know there is a bug for this issue on 11.1.
    Can I somehow know the rejected records information other than doing a minus between fact and dimensions? Any workaroud for this?
    Thanks,

    With 11gr2 we have the following logs which will give all the information including rejected dimensions without having to do a minus from fact table.
    Maintenance Logs
    The first time you load data into a cube or dimension using Analytic Workspace Manager, it creates several logs. These logs are stored in tables in the same schema as the analytic workspace:
    •     Cube Build Log: Contains information about what happened during a build. Use this log to determine whether the build produced the results you were expecting, and if not, why not. The log is continually updated whenever a cube or dimension is refreshed, whether by Analytic Workspace Manager, the database materialized view refresh subsystem, or a PL/SQL procedure. You can query the log at any time to evaluate the progress of the build and to estimate the time to completion. The default table name is CUBE_BUILD_LOG.
    •     Cube Dimension Compile Log: Contains errors that occur during the validation of the dimension hierarchies when OLAP is aggregating a cube. The default table name is CUBE_DIMENSION_COMPILE.
    •     Cube Operations Log: Contains messages and debugging information for all OLAP engine events. The default table name is CUBE_OPERATIONS_LOG.
    •     Cube Rejected Records Log: Identifies any records that were rejected because they did not meet the expected format. The default table name is CUBE_DIMENSION_COMPILE.
    You can also run the $ORACLE_HOME/olap/admin/utlolaplog.sql script to create the build log with some useful views.

  • Using cube operator in mapping

    Hi everyone,
    I am just building my first owb map with a cube (relational fact table) but I can't find any description anywhere on how to actually use/connect the cube operator in a mapping. I have managed to build a cube, linked to two dimensions, but when I place it on the canvas, I am not sure what it expects.
    Does anybody know of any good examples or a bit of documentation anywhere?
    Thanks!
    Ed
    PS In a SQL statement I would populate that fact table by simply joining the external table with the dimension, but the input signature of the cube is just different from what I expected.

    hi Patrick
    i have already done this but it is not working as well and response is same "you are using two different sources in expression".
    in fact i am using few tables from source schema and another table(time dimension table) from target schema. i think either this error is due to using time dimension's table or it could be due to using table from two different schema's
    what you say... i need your point of view
    regards,
    imran

  • Select statement on cube

    Hi
    I am writing a routine (select statement) on the characterstics which has to get the data from other cube.
    My scenairo:
    Extracting the data from the flat file.
    In the cube i have a characteristics for which I need to pull the data from other cube.
    Routine:
    Select single xxxx into result from /bic/fsd_c03
    where bill_no = source_fields_rule-bill_no..
    When i click on the check, I am getting the following error" Unknown column name "XXXX" until runtime, you cannot specify a
    field list.
    Can we write select statements on cubes? or is there any issue in my routine?
    Thanks
    Annie

    Hi Annie,
    I would suggest not because of the following reasons:
    1) The way data is stored in the Info cube is multi-dimensional, which means to say that all the characteristic are grouped into dimensions and they result in unique dimension ID's which connect to your fact table (/BI0/F*)
    This would mean, your characteristics and key figures would not be available in the same table for doing the select.
    2) If you happen to compress the data in the infocube, the data in the fact table would be  moved to the E fact table (/BI0/E) from the f fact table (/BI0/F).
    This would need change of code.
    I would rather suggest to go about this way.
    --> Check if the cube which you are trying to do a lookup is fed from a DSO.
    --> If so, use the Active table(/BI0/A*00) of the DSO for the lookup.
    Hope this helps!
    Edited by: Suchindra B R on Feb 10, 2010 7:45 AM

  • Cube operator and aggregation of measures on existing records

    Hi there,
    I'm using a cube operator with loading type LOAD in order to perform merge on fact table.
    We have following situation
    Record already exists on fact table for customer and product with qty 8.
    Incoming record has qty 2, so tried using cube operator hoping that because aggragtion specfied on qty measure was sum and set to solve cube, new qty would be 10.
    However, looked at sql generated and record simply gets updated with new qty rather than adding new to existing qty.
    I can achieve our aim by simply reading for any existing record and adding new record qty to existing record qty, but was hoping the cube operator would do thi s for me.
    Anybody achivede anything similar using sinmply cube operator.
    Many Thanks

    Do you mean that you want to load the data into the AW using the AVG function instead of SUM? If this is true, are you planning to use AVG as the aggregation operator in the AW as well? Will this give the answer you want? The code currently defaults to SUM for load even if you aggregate the cube using AVG since AVG of AVG is not usually what people want. If you want to do it anyway then there it is possible if you hand edit the XML to add an attribute named AggregationMethod to the CubeMap. E.g.
        <CubeMap
          Name="MAP1"
          Query="SALES_FACT"
          AggregationMethod="AVG">But the simpler way to do it is to define a SQL View that aggregates to the load level using AVG and then map the cube to the view.

  • Using dimension and cube operator

    hi
    I am facing problem in using this dimension and cube operator..
    Actually I want to create a sales data with dimensions as time,customer and product
    After creating a dimension in owb I get a table created attached with it same is the case with cube operator
    so where should I store my measures in the cube or the table created and where to store the data about dimensions in the table created.
    and finally how to use them in mapping to retrieve the data stored
    Thanks

    It sounds like you are not getting any matching keys for loading into the cube.
    Do you have a time dimension created by OWB in the cube? The key used by the cube operator for an OWB time dimension is a formatted number. The time dimension keys are stored as follows;
    Day Level - YYYYMMDD
    Month Level - YYYYMM
    Week Level - YYYYWW
    Quarter - YYYYQ
    Year - YYYY
    If you have a source that has a SQL date datatype for example and want to construct the key for a cube's time dimension at the day level something like the following expression can be used to construct the time reference from a SQL date...
    to_number(to_char( time_key, 'YYYYMMDD'))
    It may not be this but just a thought.
    Cheers
    David

  • /* + ordered */ hint used in relational cube operator

    We are using OWB 10.2.0.3 build a relational datamart. The mapping which loads the fact table has a joiner joining three tables, the result of the joiner goes into the cube operator. We realized that the code created by the cube operator contains the /* + ordered */ hint. This might make sense if you do not have any joiner outside the cube operator, but in our case the hint results into a sub-optimal explain plain.
    Has anyone any idea how to get rid of this hint? What is the reason for this hint?
    Thanks in advance, Maren

    Hi,
    if you open the configuration of the mapping, select cube operators, select your target cube, select table operators, selct your cube table -> what do you see in the property 'Loading hints'?
    Regards,
    Detlef

  • Missing boot events on the Diagnostic performance operational log (event-id 100)

    I noticed that event (100) is not always reported or not reported at all on several machines and I would like to understand how can I force reporting it.
    In addition I would appreciate if you can explain regarding the keys under:
    [HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Diagnostics\Performance\Boot]
    Things I've already tested on win7:
    1. Verified that the diagnostic performance > operational log is enabled
    2. Services: ‘Diagnostic Policy Service’ (auto) and ‘Diagnostic service host’ (manual) were running
    3. Verified that GPO is not disabled for:
    - Computer Configuration\Administrative Templates\System\Troubleshooting
    - Diagnostics\Windows Boot Performance Diagnostics Computer Configuration\Administrative Templates\System\Troubleshooting and Diagnostics\Diagnostics: Configuration scenario executing level
    Thanks,

    For keys under [HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows\CurrentVersion\Diagnostics\Performance\Boot]
    Please take a look of discussions in this link(reply posted by David J)
    Diagnostics-Performance log Event 100 - Critical, Error, or Warning - when and why?
    http://social.technet.microsoft.com/Forums/windows/en-US/48005f5d-5f66-439a-af51-3a2ebb894e31/diagnosticsperformance-log-event-100-critical-error-or-warning-when-and-why?forum=w7itproperf
    My understanding to this issue—“event (100) is not always reported or not reported at all on several machines”
    is that : this kind event log occurs when an time parameter during the boot process reaches a threshold in the registry key, then a warning, critical or error message will appear in the event viewer. If it doesn’t reach the threshold, then there will be
    no report.
    Regards
    Yolanda
    TechNet Community Support

  • Nokia C2-01 How to Active Operator logs?

    I'm Sudip from Silchar
    I want to active Operator logs in My Nokia C2-01
    How to active Operator Logs? Tell Me Quickly.
    Solved!
    Go to Solution.

    Look in display settings if its there and greyed out its not possible this way,or you can google OPERATOR LOGO and your phone plenty of help there.
    If  i have helped at all a click on the white star below would be nice thanks.
    Now using the Lumia 1520

  • Exporting Operations Log to file

    Hi,
    We have a requirement to retain the Management Operations logs i.e. who created what, and when. Can anyone point a noob in the right direction please?

    Have you seen any sample apps on the web?  I don't have too much time to code something, so wondering if there is anything like a client app to download logged events with a GUI similar to the web site, where I can plug in a subscription and get logs.
    Thanks, Jason
    Didn't get enough help here? Submit a case with the Microsoft Customer Support teams for deeper investigation - Azure service support: https://manage.windowsazure.com/?getsupport=true For on Premise software support go here instead: http://support.microsoft.com/select/default.aspx?target=assistance

  • SQl statement cause extreme Portal log growth

    Our Portal server log file is growing very fast (30 - 50 Mb per minute). When examining the log file we see that the log file does contain only the error messages:
    #1.5#00096BF521750058005A9CA900001E480003E80BD8A2C263#1099562506968#/System/Database/sql/jdbc/common#sap.com/irj#com.sap.sql.jdbc.common.StatementAnalyzerImpl#Guest#18####655e44612e4811d9b41300096bf52175#Thread[ThreadPool-Dispatcher,5,SAPEngine_Application_Thread[impl:3]_Group]##0#0#Error#1#com.sap.sql.jdbc.common.StatementAnalyzerImpl#Java#com.sap.sql.jdbc.common_1123#com.sap.sql.log.OpenSQLResourceBundle#The SQL statement "" contains the semantics error[s]: #2#SELECT MIN("ID"),"PRIORITY" FROM "KMC_TQ_QUEUE" WHERE "NAMESPACE_HASH" = ? AND "NAMESPACE" = ? AND "TASK_STATE" = 1 GROUP BY "PRIORITY" ORDER BY 2 DESC#type check error: the expression >>"NAMESPACE"<< (LONGVARCHAR) is not comparable and must not be used with "="
    This message occurs more than 500 times per log file. It has something to do with Knowledge management, but we cannot find anything that triggers this sql statement.
    Any ideas?
    Noel Hendrikx & Pascal Rijnart

    SAP tells us in reaction to our customer message:
    "The table's definition is correct. What has to be changed is the
    computed type for field NAMESPACE and the generated file for the table.
    As I wrote before, if you make any change in the field NAMESPACE the
    new computation is started (so change the description which has no
    effect on the table's version on the database). Make the change and
    save the table in the SapNetWeaver Developer Studio. You will then
    see that the field's JDBC-type changes to VARCHAR.
    Choose 'Create archive' for Dictionary Project the table belongs to
    and deploy the sda on the database. Both can be done in the studio."
    Can anyone tell us what to do now?
    Noel Hendrikx

  • Need help in using a case statement in expression operator

    Hi All,
    I am using OWB version 10.2.0.1.0.
    My requirement is to add a new column called call _zone_key in expression operator and map it to the target table. I need to use the below expression for populating call_zone_key values
    Expression:
    case when (INGRP1.CHARGETYPE in ('O','F') or  INGRP1.TARIFF_GROUP in ('SMSINT','MMSINT')or ( INGRP1.CALL_TYPE = '002' and   INGRP1.TARIFF_GROUP  = 'MTV'))
    then
    (select call_zone_reltn_key from call_zone_reltn where
    call_zone_cd=substr(case
      when substr( INGRP1.B_SUBNO,1,2)='00'
      then
      substr( INGRP1.B_SUBNO,3)
      else substr( INGRP1.B_SUBNO,1)
      end,1,length(call_zone_cd))and rownum=1)
    else -1
    end
    All the columns needed for using the above expression is available in INGRP1 but still I am unable to deploy the mapping using above expression. Call_zone_reltn table is also imported to the module. I am getting below error
    Error:
    Warning
    ORA-06550: line 4980, column 2:
    PLS-00103: Encountered the symbol "SELECT" when expecting one of the following:
       ( - + case mod new not null others <an identifier>
       <a double-quoted delimited-identifier> <a bind variable> avg
       count current exists max min prior sql stddev sum variance
       execute forall merge time timestamp interval date
       <a string literal with character set specification>
       <a number> <a single-quoted SQL string> pipe
       <an alternatively-quoted string literal with character set specification>
       <an alternativ
    NEW_MOD_MAP_CELL_USAGE_FCT_PRE
    Create
    Warning
    ORA-06550: line 4989, column 43:
    PLS-00103: Encountered the symbol ")" when expecting one of the following:
       * & - + ; / at for mod remainder rem <an exponent (**)> and
       or group having intersect minus order start union where
       connect || multiset
    If i replace the expression with numbers such as 1 or 2, I am able to deploy the mapping.
    Kindly help in fixing this issue.
    Thanks,
    Kabilan

    You can't use the SELECT statement inside the expression, you need to join both tables before the expression. Use a Join operator with this JOIN condition:
    CALCULATED_CALL_ZONE_CD = call_zone_reltn.call_zone_cd ( + )
    Where Calculated_call_zone_cd proceed from a previous expression:
    CALCULATED_CALL_ZONE_CD = substr(case when substr( INGRP1.B_SUBNO,1,2)='00' then substr( INGRP1.B_SUBNO,3) else substr( INGRP1.B_SUBNO,1) end,1,length(call_zone_cd))
    And after joining both tables, you can use another expression to get the rownum, then another filter operator to keep only the rownum = 1, and now you can use your expression without the SELECT, using the call_zone_cd column from the outgroup in the joiner operator (you need to include that column in the filter operator to use it).
    Regards
    ANA GH

  • Prepared Statement with IN Operator

    Hello all, do anyone know how to use IN operator in a prepared statement? For ex.,
    "select name from user where user_no in (?)"
    If I do preparedStatement.setString(1, "2, 3, 4") and execute it, it would give me an ORA-01722: invalid number exception. Any help or pointer will be appreciated. Thanks.
    yien

    You can't (easily) pass a comma-separated list to a SQL statement. If you want to pass N values straight to a PreparedStatement, you would need N bind variables, i.e.
    select name from user where user_no in (?, ?, ?)I would tend to either pass an array to a stored procedure that returned a REF CURSOR for the SELECT or create a pipelined table function that parsed the comma-separated list into a table structure.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Select Statement Blocking Delete Operation

    Hi Experts,
    In a Single table am trying to do delete operation but it has been blocked by simple select query.
    Database setting : Read Committed Snapshot already set to ON for that database.
    Any Idea why SELECT blocking DELETE.
    FYI : Before Select statement I could noticed the following setting "set transaction isolation level  read committed".
    Kindly Advise
    Regards,
    Nandhu

    FYI : Before Select statement I could noticed the following setting "set transaction isolation level  read committed".
    Hi NandhuJana,
    1. Did you read
    Uri Dimant's response?
    Please try to remove this statement and check again
    2. When we ask you for DDL+DML , the idea is to let us the option to (1) reproduce the issue in some cases, (2) get the best information which is the real database structure, queries, etc'. In this case for example I do not see the isolation level statement
    in your code and we need it to understand what is happening :-) The order and structure (nested transactions for example) of statements is what important here.
    3. you can check your isolet level using one of those options to make sure what is going in your script:
    SELECT TRANSACTION_ISOLATION_LEVEL = CASE transaction_isolation_level
    WHEN 0 THEN 'Unspecified'
    WHEN 1 THEN 'ReadUncommitted'
    WHEN 2 THEN 'ReadCommitted'
    WHEN 3 THEN 'Repeatable'
    WHEN 4 THEN 'Serializable'
    WHEN 5 THEN 'Snapshot'
    END
    FROM sys.dm_exec_sessions
    where session_id = @@SPID
    GO
    DBCC USEROPTIONS WITH NO_INFOMSGS
    GO
    I hope this is helpful :-)
      Ronen Ariely
     [Personal Site]    [Blog]    [Facebook]

  • Difference beween Case Statement and Pivot Operator in a OWB mapping

    Hi ,
    Kindly clarify what is the difference between using a PIVOT operator and the CASE statement in a Expression Operator in a particular mapping.
    rgds
    Arinjit

    Hi
    With PIVOT operator you can transform columns to rows. For example if oyu have 12 column (one for every months) than you can create 12 rows with one column of month data.
    If you want to do this without PIVOT you can do it with union (12 times) or with
    something like this.
    In this case you can't use CASE statement I think.
    If you want to do the reverse of this, create 12 column from 12 rows, you can use UNPIVOT operator. In this case you can use CASE statement and aggregating.
    Ott Karesz
    http://www.trendo-kft.hu

Maybe you are looking for

  • Data acquisition with control references

    I'm a new LabView user and I would like to acquire data in a subVI and display that data in the main VI. I think that I need to use control references and refnums, but I just haven't been able to figure it out. I would GREATLY appreciate any help I c

  • End User Login Message

    Hi Folks, Our production IDM setup has 3 IDM instances sharing the same repository. The IDM End user is used for setting passwords in IDM and LDAP. Recently, couple of users have reported this error while trying to log in into IDM end user login page

  • Query to find out which process consuming more cpu?

    Hi All, some one please tell me the query to find out which process is consuming more cpu resources? Thanks and Regards Srikanth

  • Upload songs from multiple computers?

    Has anyone figured out if there is a way to UPLOAD songs from multiple computers. Note, I don't want to sync. I want to be able to plug my iPhone in at my office and put mp3s which I buy at work onto the phone. This works on the iPod, but not the iPh

  • Configuration error in PID synthesis step of control design assistant

    Hello,         I am exploring an evaluation version of NI LabVIEW Control Design Assistant 2011. For  getting started with it ; I am using getting started tutorial provided in the labview signal express help. Upto the step of frequency domain analysi