Webanalysis : Query Log File

Hi,
Plz suggest me where can i the path for report's query log file. i.e I run a webanalysis report and now i want to see the query that it fire to the source.
I want to see the detail level log.
Is there any other setting that i need to do like entry in property file or something.
Plz suggest the way.
Thanks

Create two env variables
ADM_TRACE_LEVEL=0
REDIRECTOR_TRACE_LEVEL=0
Also you can add this line to WA.prop file
LogQueries=true
Hope it hepls
Regards
CK

Similar Messages

  • Location of query log files in OBIEE 11g (version 11.1.1.5)

    Hi,
    I wish to know the Location of query log files in OBIEE 11g (version 11.1.1.5)??

    Hi,
    Log Files in OBIEE 11g
    Login to the URL http://server.domain:7001/em and navigate to:
    Farm_bifoundation_domain-> Business Intelligence-> coreapplications-> Dagnostics-> Log Messages
    You will find the available files:
    Presentation Services Log
    Server Log
    Scheduler Log
    JavaHost Log
    Cluster Controller Log
    Action Services Log
    Security Services Log
    Administrator Services Log
    However, you can also review them directly on the hard disk.
    The log files for OBIEE components are under <OBIEE_HOME>/instances/instance1/diagnostics/logs.
    Specific log files and their location is defined in the following table:
    Log Location
    Installation log                     <OBIEE_HOME>/logs
    nqquery log <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIServerComponent/coreapplication_obis1
    nqserver log <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIServerComponent/coreapplication_obis1
    servername_NQSAdminTool log      <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIServerComponent/coreapplication_obis1
    servername_NQSUDMLExec log                          <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIServerComponent/coreapplication_obis1
    servername_obieerpdmigrateutil log (Migration log)           <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIServerComponent/coreapplication_obis1
    sawlog0 log (presentation)                          <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIPresentationServicesComponent/coreapplication_obips1
    jh log (Java Host)                               <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIJavaHostComponent\coreapplication_obijh
    webcatupgrade log (Web Catalog Upgrade)                <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIPresentationServicesComponent/coreapplication_obips1
    nqscheduler log (Agents)                          <OBIEE_HOME>/instances/instance1/diagnostics/logsOracleBISchedulerComponent/coreapplication_obisch1
    nqcluster log                                    <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIClusterControllerComponent\coreapplication_obiccs1
    ODBC log                                    <OBIEE_HOME>/instances/instance1/diagnostics/logs/OracleBIODBCComponent/coreapplication_obips1
    opmn log                                    <OBIEE_HOME>/instances/instance1/diagnostics/logs/OPMN/opmn
    debug log                                    <OBIEE_HOME>/instances/instance1/diagnostics/logs/OPMN/opmn
    logquery log                               <OBIEE_HOME>/instances/instance1/diagnostics/logs/OPMN/opmn
    service log                                    <OBIEE_HOME>/instances/instance1/diagnostics/logs/OPMN/opmn
    opmn out                              <OBIEE_HOME>/instances/instance1/diagnostics/logs/OPMN/opmn
    Upgrade Assistant log                         <OBIEE_HOME>Oracle_BI1/upgrade/logs
    Regards
    MuRam

  • Query log file missing

    I am trying create a query log file for an ASO cube. I have created the database.cfg file in the database directory and restarted the application. Then I run a few queries, and stop the application. Query log file (database.qlg) is not created. My database.cfg file looks like this:<BR>QUERYLOG [LOB]<BR>QUERYLOG LOGFILESIZE 2<BR>QUERYLOG TOTALLOGFILESIZE 1024<BR>QUERYLOG ON<BR><BR>Thanks

    I am trying create a query log file for an ASO cube. I have created the database.cfg file in the database directory and restarted the application. Then I run a few queries, and stop the application. Query log file (database.qlg) is not created. My database.cfg file looks like this:<BR>QUERYLOG [LOB]<BR>QUERYLOG LOGFILESIZE 2<BR>QUERYLOG TOTALLOGFILESIZE 1024<BR>QUERYLOG ON<BR><BR>Thanks

  • Query log file location?

    Is log file is created when query has completed execution, if yes plz tell its location.

    Put a set timing on before executing the query in SQL *Plus like:
    SQL> set timing on
    SQL> select 1 from dual;
    1
    1
    Elapsed: 00:00:00.00
    To make trace on just do the following:
    SQL> SET AUTOTRACE TRACEONLY
    SQL> SELECT 1 FROM DUAL;
    Elapsed: 00:00:00.01
    Execution Plan
    0 SELECT STATEMENT Optimizer=ALL_ROWS (Cost=2 Card=1)
    1 0 FAST DUAL (Cost=2 Card=1)
    Statistics
    1 recursive calls
    0 db block gets
    0 consistent gets
    0 physical reads
    0 redo size
    419 bytes sent via SQL*Net to client
    508 bytes received via SQL*Net from client
    2 SQL*Net roundtrips to/from client
    0 sorts (memory)
    0 sorts (disk)
    1 rows processed
    SQL>
    Regards
    Edited by: Virendra.k.Yadav on Aug 20, 2010 2:37 AM

  • SQL query log File..

    Hi All,
    I am using a standalone BIP environment... Is there way we can see the actual query log the BIP generates? which location on the server..or how to get it activated.
    I want to see what is the actual sql query that is generated along with variables in it.
    Fyi : I want to have the query log n not the server error log..
    -dev

    Hi Eric,
    I would like you to re-check on the content level settings here as they are the primary causes of this kind of behavior. You could notice that the same information might have written down in the logical plan of the query too.
    Also, as per your description
    "In the SOURCES for this logical table, I've set the logical level of the content for E2 appropriately (detail level, same as E1)"
    I would like to check on this point again, as if you had mapped E2 to E1 in the same logical source with an inner join, you would get to set the content level at E1 levels themselves but not E2 (Now, that E2 would become a part of the E1 hierarchy too). This might be the reason, the BI Server is choosing to elimiate(null) the values from E2 too (even you could see them in the sql client)
    Hope this helps.
    Thank you,
    Dhar

  • Query Logging Enabled but cannot find

    As per the DBAG I should be able to find the query log hereHYPERION_HOME/logs/essbase/app/appname/dbname/dbname00001.qlg
    When I use this path to I can only get to to app D:\EPM\EPMSystem11R1\logs\essbase\app there is nothing inside the app folder. Am I looking at the wrong folder?
    Thanks for the help.

    Hello,
    From the DBA guide
    To enable query logging, create a query configuration file ... and add to the file the configuration settings that control how query logging is performed.
    In the ARBORPATH/app/appname/dbname directory, create a query log configuration file. The configuration file must be named dbname.cfg, where dbname matches the name of the database. For example, the query log configuration file for Sample.Basic is basic.cfg. The output query log file is located, by default, at ARBORPATH/app/appname/dbname00001.qlg.
    You must stop and start the application, so it can read the config file.
    Contents of a query log config file:
    http://docs.oracle.com/cd/E40248_01/epm.1112/essbase_tech_ref/frameset.htm?launch.html
    Query Log Settings File SyntaxThe query log settings filename must be of the form dbname.cfg, where dbname represents the name of a database. The dbname.cfg file must be located in the ARBORPATH\App\appname\dbname directory of Essbase. The dbname.cfg file consists of the following syntax:
    QUERYLOG [dimension_name]
    QUERYLOG NONE GENERATION generation-range
    QUERYLOG NONE LEVEL level-range
    QUERYLOG GENERATION generation-range
    QUERYLOG LEVEL level-range
    QUERYLOG LOGHAMBRS ON | OFF
    QUERYLOG LOGPATH path-expression
    QUERYLOG LOGFORMAT CLUSTER | TUPLE
    QUERYLOG LOGFILESIZE n
    QUERYLOG TOTALLOGFILESIZE n
    QUERYLOG ON | OFF
    Regards,
    Philip Hulsebosch.
    www.trexco.nl
    p.s. To all users, close questions which were answered. Then we do not have to open them to see if we can help = saves us time to help others.

  • Query logging?

    Hi,
    A customer of mine picked up this in the essbase log
    Missing Database Config File [C:\Hyperion\AnalyticServices\APP\<app name>\OHead\OHead.cfg], Query logging disabled
    The essbase app is used for planning.
    I cannot find the file concerned. The other databases do not have a config file.
    Any ideas??
    Thanks,
    Nathan

    You will only have a query loggin cfg file if you set one up. It is related to the query logging capability.
    From the DBAG
    Implementing Query Logs
    Query logging provides a way for Essbase administrators to track query patterns of an Essbase database. The query log file tracks all queries performed against the database regardless of whether the query originated from Spreadsheet Add-in or Report Writer. Query logging can track members, generation or level numbers of members belonging to specific generations or levels, and Hybrid Analysis members. Query logging also offers the flexibility to exclude logging of certain dimensions and members belonging to generations or levels. Because the query log file output is an XML document, you can import the log file to any XML-enabled tool to view the log. For details about the query log file structure, refer to querylog.dtd in the ARBORPATH/bin directory.
    To enable query logging, create a query configuration file (distinct from the essbase.cfg file) and add to the file the configuration settings that control how query logging is performed.
    In the ARBORPATH\App\appname\dbname directory of Essbase, create a query log configuration file. The configuration file must be named dbname.cfg, where dbname matches the name of the database. For example, the query log configuration file for Sample Basic is basic.cfg. The output query log file is located at, by default, ARBORPATH\App\appname\dbname00001.qlg.
    For more information about query logging and how to set up the query log configuration file, see the Technical Reference.

  • Query Log and Exception log

    Hi All,
    I have running essbase application in my system. I execute some query now, what i need to do? to see query log file and exception log file.

    An exception log is only produced if the essbase server or application/database shuts down abnormally, more information :- http://download.oracle.com/docs/cd/E17236_01/epm.1112/esb_dbag/dlogs.html#dlogs1038688
    Query log I provided the details to set up in your previous post :- Log Files
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Log file Growth query

    Hi All,
    Is there any query to increase the logfile growth to unlimited in SQL 2000 . Am unable to do that through GUI mode and getting some transaction error. Please suggest me any query.
    Thanks & Regards,
    Venkat.

                     
    "As I said, you cannot set log files to "unrestricted".  You must set them to a number.  "
    Not true. From
    http://msdn.microsoft.com/en-us/library/bb522469.aspx:
    MAXSIZE { max_size| UNLIMITED }
    Specifies the maximum file size to which the file can grow.
    max_size                              
    Is the maximum file size. The KB, MB, GB, and TB suffixes can be used to specify kilobytes, megabytes, gigabytes, or terabytes. The default is MB. Specify a whole number and do not include a decimal. If
    max_size is not specified, the file size will increase until the disk is full.
    UNLIMITED                            
    Specifies that the file grows until the disk is full. In SQL Server, a log file specified with unlimited growth has a maximum size of 2 TB, and a data file has a maximum size of 16 TB. There is no maximum size when this option is specified for a FILESTREAM
    container. It continues to grow until the disk is full.
    In other words, you _can_ set a log file's MAXSIZE to UNLIMITED and you do _not_ have to specify a number, but SQL Server will _not_ grow a log file beyond 2TB (even when you try to allow it)

  • Query runs every minute from B1 client - read B1 log file

    Hi all,
    We found this query has been run every minute from B1 client. It slows down the system, as we have over 100,000 records in table OCLG - activities. How do we stop this query to be run from B1 client. Also, anyone know what is the number for Duration in B1 log file, does the number mean CPU time? We got Duration=3581 for this query. but it takes about 10-15 seconds to run the same query from SQL 2005 management studio. Any idea?
    Thanks,
    David
    ============================================================================================
    16/02/2010  12:10:42:295830    SQLMessage    Note           ExecDirectInt     C:\Program Files\SAP\SAP Business One\SAP Business One.exe            PID=1972             TID=1920             Duration=3581  Fetched=0
                                                                                    Query     
    SELECT T0.[ClgCode], T0.[Action], T0.[Details], T1.[Name], T0.[Recontact], T0.[BeginTime], T0.[AttendUser] FROM  [dbo].[OCLG] T0   LEFT OUTER  JOIN [dbo].[OCLO] T1  ON  T1.[Code] = T0.[Location]   WHERE T0.[Reminder] = (N'Y' )  AND  T0.[RemSented] = (N'N' )  AND  (T0.[RemDate] < (CONVERT(DATETIME, '20100216', 112) )  OR  (T0.[RemDate] = (CONVERT(DATETIME, '20100216', 112) )  AND  T0.[RemTime] <= (1210 ) ))

    Thanks Paulo,
    I found another two queries that also run every minute on each B1 workstation. What is the measurement for the duration number in the log file, like  Duration=1391?
    David
    ============================================================================================
    17/02/2010  11:32:46:620527    SQLMessage    Note           ExecDirectInt     C:\Program Files\SAP\SAP Business One\SAP Business One.exe     PID=1868     TID=3236     Duration=1391     Fetched=21
                                  Query      SELECT T0.[ClgCode], T0.[AttendUser], T0.[Closed], T0.[Recontact], T0.[endDate], T0.[Action], T0.[BeginTime], T0.[ENDTime], T0.[Duration], T0.[DurType], T0.[Details], T0.[Notes], T0.[personal] FROM [dbo].[OCLG] T0 WHERE (T0.[Recontact] >= CONVERT(DATETIME, '20100215', 112)  AND T0.[Recontact] <= CONVERT(DATETIME, '20100221', 112)  )  AND  T0.[endDate] = T0.[Recontact]  AND  T0.[inactive] = (N'N' )  AND  T0.[BeginTime] IS NOT NULL   AND  T0.[ENDTime] IS NOT NULL   AND  (T0.[Action] = (N'C' )  OR  T0.[Action] = (N'M' )  OR  T0.[Action] = (N'N' ) ) AND  (T0.[AttendUser] = (612 ) )  ORDER BY T0.[Recontact],T0.[BeginTime]
    17/02/2010  11:32:54:917455    SQLMessage    Note           ExecDirectInt     C:\Program Files\SAP\SAP Business One\SAP Business One.exe     PID=1868     TID=3236     Duration=487     Fetched=0
                                  Query      SELECT T0.[ClgCode], T0.[AttendUser], T0.[Closed], T0.[Recontact], T0.[endDate], T0.[Action], T0.[BeginTime], T0.[ENDTime], T0.[Duration], T0.[DurType], T0.[Details], T0.[Notes], T0.[personal] FROM [dbo].[OCLG] T0 WHERE (((T0.[Recontact] >= CONVERT(DATETIME, '20100215', 112)  AND T0.[Recontact] <= CONVERT(DATETIME, '20100221', 112)  )  OR  (T0.[endDate] >= CONVERT(DATETIME, '20100215', 112)  AND T0.[endDate] <= CONVERT(DATETIME, '20100221', 112)  ) ) OR  (T0.[Recontact] < (CONVERT(DATETIME, '20100215', 112) )  AND  T0.[endDate] > (CONVERT(DATETIME, '20100221', 112) ) )) AND  T0.[endDate] <> T0.[Recontact]  AND  T0.[inactive] = (N'N' )  AND  T0.[BeginTime] IS NOT NULL   AND  T0.[ENDTime] IS NOT NULL   AND  (T0.[Action] = (N'C' )  OR  T0.[Action] = (N'M' )  OR  T0.[Action] = (N'N' ) ) AND  (T0.[AttendUser] = (612 ) )  ORDER BY T0.[Recontact],T0.[BeginTime]

  • WebDAV Query generates a high number of transaction log files

    Hi all,
    I have a program that launch WebDAV queries to search for contacts on an Exchange 2007 server. The number of contacts returned for each user's mailbox is quite high (about 4500).
    I've noticed that each time the query is launched, about 15 transaction log files are generated on the Exchange server (each of them 1Mb). If I ask only for 2 properties on the contacts, this number is reduced to about 8.
    This is a problem since our program is supposed to launch often (about every 3/5min) as It will synchronize Exchange mailboxes with a SQL Server DB. The result is that the logs increase very quickly on the server side, even if there are not so many updates.
    Any idea why so many transaction logs are generated when doing a WebDAV search returning many items? I would understand that logs are created when an update is done on the server, but here it's only a search with many contacts items returned.
    Is there maybe a setting on the Exchange server to control what kind of logs to generate?
    Thank for your help,
    Alexandre

    Hi Alex,
    Actually circular logging/backup was not a solution, I was just explaining that there is an option like that on server but it is not recommended hence not useful in our case :)
    - I am not a developer but AFAIK, WebDAV search query shouldn't generate transaction log because it just searches the mailboxes and gives the result in HTTP format and doesn't produce any Exchange transaction.
    - I wouldn't open transaction logs since it is being used by Exchange which may generate errors and may corrupt Exchange database sometime too. However it is not readable, as you observed, other than Exchange Information Store service (store.exe).
    - You can post this query in development forum to get better idea on this, if any other programmer observed similar symptom while using WebDAV contact search query in Exchange 2007 or can validate your query.
    Microsoft TechNet > Forums Home > Exchange Server > Development
    Well, I just saw that you are using Exchange 2007, in that case why don't you use Exchange Web Service which is better and improved method to access/query mailboxes where as WebDAV is also de-emphasized in Exchange 2007 and might be disappeared in next version of Exchange. Checkout below article for further detail.
    Development: Overview
    http://technet.microsoft.com/en-us/library/aa997614.aspx
    Amit Tank | MVP - Exchange | MCITP:EMA MCSA:M | http://ExchangeShare.WordPress.com

  • Cancel the query which uses full transaction log file

    Hi,
    We have reindexing job to run every week sunday. During the last run, the transaction log got full and the subsequent transactions to the database got errorred out stating 'Transaction log is full'. I want to restrict the utilization of the log file, that
    is, when the reindexing job reaches the utilization of the log file to a certain threshold automatically the job should get cancelled. Is there any way to get it.

    Hi,
    We have reindexing job to run every week sunday. During the last run, the transaction log got full and the subsequent transactions to the database got errorred out stating 'Transaction log is full'. I want to restrict the utilization of the log file, that
    is, when the reindexing job reaches the utilization of the log file to a certain threshold automatically the job should get cancelled. Is there any way to get it.
    Hello,
    Instead of Putting limit on trn log it would be good to find out cause causing high utilization.Even if you find that your log is growing because of some transaction it would be a blunder to rollback its little easy to do it for Index rebuild but if you
    cancel for some delete operation you would end up in mess.Please don't create a program to delete or kill running operation.
    You can create custom job for alert for trn log file growth.That would be good.
    From 2008 onwards Index rebuild is fully logged so sometimes it causes trn log issue.To solve this either you run index rebuild for specific tables or for selective tables.
    Other option is widely accepted Ola Hallengren script for index rebuild.I suggest you try this
    http://ola.hallengren.com/
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Logical sql in log file.

    Can someone please tell me how to see the complete sql query in the log file. If I run the same query the sql is not being produced I looked in the server log file and also manage sessions log file. It just says all columns from 'Subject Area'. I want to see all the joins and filters as well. Even for repeated queries how can I see complete sql. I set my logging level to 2.

    http://lmgtfy.com/?q=obiee+disable+query+caching
    http://catb.org/esr/faqs/smart-questions.html#homework

  • DATE fields and LOG files  in context with external tables

    I am facing two problems when dealing with the external tables feature in Oracle 9i.
    I created an External Table with some fileds with the DATE data type . There were no issues during the creation part. But when i query the table, the DATE fields are not properly selected though the data is there in the files. Is there any ideas to deal with this ?
    My next question is regarding the log files. The contents in the log file seems to be growing when querying the external tables. Is there a way to control this behaviour?
    Suggestions / Advices on the above two issues are welcome.
    Thanks
    Lakshminarayanan

    Hi
    If you have date datatypes than:
    select
    greatest(TABCASER1.CASERRECIEVEDDATE, EVCASERS.FINALEVDATES, EVCASERS.PUBLICATIONDATE, EVCASERS.PUBLICATIONDATE, TABCASER.COMPAREACCEPDATE)
    from TABCASER, TABCASER1, EVCASERS
    where ...-- join and other conditions
    1. greatest is good enough
    2. to_date creates date dataype from string with the format of format string ('mm/dd/yyyy')
    3. decode(a, b, c, d) is a function: if a = b than return c else d. NULL means that there is no data in the cell of the table.
    6. to format the date for display use to_char function with format modell as in the to_date function.
    Ott Karesz
    http://www.trendo-kft.hu

  • Performance Issue: Wait event "log file sync" and "Execute to Parse %"

    In one of our test environments users are complaining about slow response.
    In statspack report folowing are the top-5 wait events
    Event Waits Time (cs) Wt Time
    log file parallel write 1,046 988 37.71
    log file sync 775 774 29.54
    db file scattered read 4,946 248 9.47
    db file parallel write 66 248 9.47
    control file parallel write 188 152 5.80
    And after runing the same application 4 times, we are geting Execute to Parse % = 0.10. Cursor sharing is forced and query rewrite is enabled
    When I view v$sql, following command is parsed frequently
    EXECUTIONS PARSE_CALLS
    SQL_TEXT
    93380 93380
    select SEQ_ORDO_PRC.nextval from DUAL
    Please suggest what should be the method to troubleshoot this and if I need to check some more information
    Regards,
    Sudhanshu Bhandari

    Well, of course, you probably can't eliminate this sort of thing entirely: a setup such as yours is inevitably a compromise. What you can do is make sure your log buffer is a good size (say 10MB or so); that your redo logs are large (at least 100MB each, and preferably large enough to hold one hour or so of redo produced at the busiest time for your database without filling up); and finally set ARCHIVE_LAG_TARGET to something like 1800 seconds or more to ensure a regular, routine, predictable log switch.
    It won't cure every ill, but that sort of setup often means the redo subsystem ceases to be a regular driver of foreground waits.

Maybe you are looking for