Contents of a red log file

Hi
Just want to know about the internal of redo log file. My objective is to find out some dump of redo log file to understand how the data is stored in it. what is stored in it. deos redo log file has a header. if yes what is stored in it. what is the structure of redo log file.
The objective of this thread is to learn the internal of redo log file using some sort of dump. I hope Guru will help me
thanks
Nick

internal of redo log file using some sort of dumpHey Nick,
Yes you can . Basically its a binary file. So you cannot learn about it directly.
Some of the contents of the redologs are
1) every DML operation and DDL operation is logged here. (redo means replay. Whatever you run in the database will be logged here, except SQL select statements
2) it contains the timestamp too of the statements and also the SCN numbers
3) It will also contain COMMIT and ROLLBACK commands.
You have to use a utility called LOGMINER that will "mine" the redologs and then you can view the contents in an ascii format using "v$logminer_contents" view.
All the best.

Similar Messages

  • RAID for control and red log files

    I'm considering one of these two configurations
    Option 1:
    2 disks kept as separate disks (not using RAID), storing control01.ctl, redo log group 1 on first disk, and storing control02.ctl, and redo log group 2 on second disk.
    Option 2:
    2 disks configured as a RAID 1+0 volume (mirroring and striping) holding control01.ctl and a redo log group.
    (Data files and third control file are elsewhere on the disk array.)
    With option 1, I have the option of dropping a control file and dropping a redo log group if a disk fails and then continuing processing. But with option 2, if the RAID array fails, I'm sunk.
    I'm thinking option 1 is better. What do others think? And why?
    Thanks in advance,
    Chris Curzon

    Your Raid option has the following advantage: since it mirrors things, a person can [accidently or maliciously] simultaneously remove all control files and redo logs with a single remove command, effectively crashing the database into near unrecoverability.
    Option 1 would have the disadvantage - for the person identified above - in that they would need two separate commands to ruin the system.
    But, with simple mirroring, between the first and second command, the system would have halted and hopefully alerted someone that a weird but recoverable thing is happening. And the DBA would be able to get going simply by copying the mirrored file to the other, removed, location.
    Message was edited by: Hans Forbrich
    To make it clear - the advantage is NOT on the DBA's side ...

  • Redo log file contents

    How can I check the contents of the redo log files?
    I recently install 10g. We haven't used the db yet, but the redo logs are very active.
    Is there a query to determine what activity is filling the redo log files?
    Thank you for any help.
    Take Care.
    S

    The Automatic Workload Repository takes snapshots of
    statistics, via scheduled job, at hourly intervals.
    This is probably responsible for some of the redo.
    You can see if this feature is turned on by checking
    the job queue (simply: select * from dba_jobs).
    The default settings on AWR can be managed using the
    DBMS_WORKLOAD_REPOSITORY built-in package.
    Hope this helps.
    Kailash.

  • Reading .log file & Sorting input

    Hello all
    Currently I'm working on a projekt were I have to read a "in.log" file, sort it and save it to another "out.log" file, the contents of the in.log file is:
    [204.0.44.73]: Dir: path
    [204.0.44.73]: Dir: path
    [204.0.44.74]: Dir: path
    [204.0.44.73]: Dir: path
    [204.0.44.74]: Dir: path
    and so on, now what I have to end up with is this in the out.log:
    #1
    [204.0.44.73]: Dir: path
    [204.0.44.73]: Dir: path
    [204.0.44.73]: Dir: path
    Count = 3
    #2
    [204.0.44.74]: Dir: path
    [204.0.44.74]: Dir: path
    Count = 2
    It's for the system administrator at school who doesn't want to do it himself but wants to pay an amount for me to do it (very small amount). I'll pay max duke dollars for a reply which can help me..
    Please help, thx in advance

    I must be really bored... Use it with "java Parser <logname>"
    import java.io.*;
    import java.util.*;
    public class Parser{
         public void splitLog( File input ){
              try{
                   FileInputStream fis = new FileInputStream( input );
                   BufferedReader br = new BufferedReader( new InputStreamReader( fis ) );
                   Hashtable hosts = new Hashtable();
                   String line;
                   while( ( line = br.readLine() ) != null ){
                        StringTokenizer st = new StringTokenizer( line, "]" );
                        if( st.hasMoreTokens() ){
                             String host = st.nextToken();
                             if( host.trim().startsWith("[") ){
                                  host = host.trim().substring(1);
                             if( !hosts.containsKey( host ) ){
                                  hosts.put( host, new Vector() );
                             Vector v = (Vector)hosts.get( host );
                             v.addElement( line );
                   Enumeration enum = hosts.keys();
                   while( enum.hasMoreElements() ){
                        String host = (String)enum.nextElement();
                        Vector v = (Vector)hosts.get(host);
                        FileOutputStream fos = new FileOutputStream( host + ".log" );
                        for( int i = 0; i < v.size(); i++ ){
                             line = (String)v.elementAt( i );
                             fos.write( (line + "\r\n").getBytes());
                        fos.close();
              catch( Exception e ){
                   e.printStackTrace();
         public static void main( String[] args ){
              File input = new File( args[0] );
              Parser parser = new Parser();
              parser.splitLog( input );
    }

  • How to fetch and display arbitrary log files from a managed host?

    We are creating a small management portal for a custom application and need a way to display, in a browser, the contents of an arbitrary log file located on an EM managed host. The effect we are looking for is similar to what you get when you show the contents of a database alert log using EM.
    I want to extend the EM agent to get the log and display the contents, but see two problems: (1) it seems like we would have to schedule the fetch of the log contents to happen on a regular basis and would not be able to do it on demand and (2) storing the results in the EM repository seems like it could consume a pile of storage, even if it only stores the logs for 24 hours.
    I am pretty sure we can bypass the repository somehow because I think the "show database alert log" EM process does so. Of course, I can't really be sure how this code is getting the alert log, but it seems reasonable to assume it is using the management agent.
    Is there some API for the EM Agent that I am missing?
    Any ideas would be appreciated.

    From within OEM it doesnt look possible. User defined metrics (UDM) can only return number or a string. The alertlog fetch is done via alertlogViewer.pl while being passed a couple of parameters.
    OEM does have a preliminary/rough api available at http://www.oracle.com/technology/products/oem/emx/index.html but I havent seen anyone make use of it yet.
    If you use just a web connection then you will probably run into security issues as a web server can usually only view/access stuff under their htmldoc dir.
    Seems like some non-OEM very custom code is what you seek..

  • Need help in understanding the result logged in gateway log file

    I have installed unixODBC.x86_64 (2.2.14) and mysql-connector-odbc-5.3.4-linux-el6-x85-64 on Oracle Linux Server 6.5 64 bit with Oracle 12c and MySQL Community Server 5..6.14.  Oracle database character set is AL32UTF8 and MySQL database character set is latin1. I have copied the content of a gateway log file to the bottom of this entry and boldfaced the lines that I hope someone can explain to me what they mean and what should I do.
    Gateway init file has the following configuration:
    HS_FDS_CONNECT_INFO = myodbc5
    HS_FDS_TRACE_LEVEL = DEBUG
    HS_FDS_SHAREABLE_NAME = /usr/lib64/libodbc.so
    HS_LANGUAGE=AMERICAN_AMERICA.WE8ISO8859P15
    set ODBCSYSINI=/tmp/shared
    ODBC.ini located in /tmp/shared has the following configuration:
    [myodbc5]
    Driver          = /tmp/shared/mysql-odbc/lib/libmyodbc5w.so
    DATABASE        = peter
    DESCRIPTION     = MySQL ODBC 5.3 Unicode Driver
    SERVER          = localhost
    UID             = peter
    PASSWORD        = peter
    SOCKET          = /var/lib/mysql/mysql.sock
    Listener.ora has the following configuration:
    LISTENER =
      (ADDRESS_LIST=
    (ADDRESS=(PROTOCOL=tcp)(HOST=localhost)(PORT=1521))
    (ADDRESS=(PROTOCOL=ipc)(KEY=PNPKEY)))
    SID_LIST_LISTENER =
       (SID_LIST =
         (SID_DESC =
           (GLOBAL_DBNAME = PETER)
           (ORACLE_HOME = /opt/oracle/12c)
           (SID_NAME = PETER)
         (SID_DESC =
           (SID_NAME = mysqlodbc)
           (ORACLE_HOME = /opt/oracle/12c)
           (PROGRAM = dg4odbc)
    Tnsnames.ora has the following configuration:
    mysqlodbc =
    (DESCRIPTION=
    (ADDRESS=
    (PROTOCOL=TCP) (HOST=localhost) (PORT=1521)
    (CONNECT_DATA=
    (SID=mysqlodbc))
    (HS=OK)
    Gateway log file has the following content:
    Oracle Corporation --- TUESDAY   FEB 17 2015 14:08:39.306
    Heterogeneous Agent Release
    12.1.0.1.0
    Oracle Corporation --- TUESDAY   FEB 17 2015 14:08:39.306
        Version 12.1.0.1.0
    Entered hgogprd
    HOSGIP for "HS_FDS_TRACE_LEVEL" returned "DEBUG"
    Entered hgosdip
    setting HS_OPEN_CURSORS to default of 50
    setting HS_FDS_RECOVERY_ACCOUNT to default of "RECOVER"
    setting HS_FDS_RECOVERY_PWD to default value
    setting HS_FDS_TRANSACTION_LOG to default of HS_TRANSACTION_LOG
    setting HS_IDLE_TIMEOUT to default of 0
    setting HS_FDS_TRANSACTION_ISOLATION to default of "READ_COMMITTED"
    setting HS_NLS_NCHAR to default of "AL32UTF8"
    setting HS_FDS_TIMESTAMP_MAPPING to default of "DATE"
    setting HS_FDS_DATE_MAPPING to default of "DATE"
    setting HS_RPC_FETCH_REBLOCKING to default of "ON"
    setting HS_FDS_FETCH_ROWS to default of "100"
    setting HS_FDS_RESULTSET_SUPPORT to default of "FALSE"
    setting HS_FDS_RSET_RETURN_ROWCOUNT to default of "FALSE"
    setting HS_FDS_PROC_IS_FUNC to default of "FALSE"
    setting HS_FDS_MAP_NCHAR to default of "TRUE"
    setting HS_NLS_DATE_FORMAT to default of "YYYY-MM-DD HH24:MI:SS"
    setting HS_FDS_REPORT_REAL_AS_DOUBLE to default of "FALSE"
    setting HS_LONG_PIECE_TRANSFER_SIZE to default of "65536"
    setting HS_SQL_HANDLE_STMT_REUSE to default of "FALSE"
    setting HS_FDS_QUERY_DRIVER to default of "TRUE"
    setting HS_FDS_SUPPORT_STATISTICS to default of "FALSE"
    setting HS_FDS_QUOTE_IDENTIFIER to default of "TRUE"
    setting HS_KEEP_REMOTE_COLUMN_SIZE to default of "OFF"
    setting HS_FDS_GRAPHIC_TO_MBCS to default of "FALSE"
    setting HS_FDS_MBCS_TO_GRAPHIC to default of "FALSE"
    Default value of 64 assumed for HS_FDS_SQLLEN_INTERPRETATION
    setting HS_CALL_NAME_ISP to "gtw$:SQLTables;gtw$:SQLColumns;gtw$:SQLPrimaryKeys;gtw$:SQLForeignKeys;gtw$:SQLProcedures;gtw$:SQLStatistics;gtw$:SQLGetInfo"
    setting HS_FDS_DELAYED_OPEN to default of "TRUE"
    setting HS_FDS_WORKAROUNDS to default of "0"
    Exiting hgosdip, rc=0
    ORACLE_SID is "mysqlodbc"
    Product-Info:
      Port Rls/Upd:1/0 PrdStat:0
      Agent:Oracle Database Gateway for ODBC
      Facility:hsa
      Class:ODBC, ClassVsn:12.1.0.1.0_0017, Instance:mysqlodbc
    Exiting hgogprd, rc=0
    Entered hgoinit
    HOCXU_COMP_CSET=1
    HOCXU_DRV_CSET=46
    HOCXU_DRV_NCHAR=873
    HOCXU_DB_CSET=873
    HS_LANGUAGE is AMERICAN_AMERICA.WE8ISO8859P15
    LANG=en_US.UTF-8
    HOCXU_SEM_VER=121000
    HOCXU_VC2_MAX=4000
    HOCXU_RAW_MAX=2000
    Entered hgolofn at 2015/02/17-14:08:39
    HOSGIP for "HS_FDS_SHAREABLE_NAME" returned "/usr/lib64/libodbc.so"
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac07fe0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac08110
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac089d0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac09d40
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac11c00
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac120a0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac148d0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac15dd0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac16610
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac18060
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac18070
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac197c0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1cb80
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1cf60
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1eb70
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1f800
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1fb80
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac21af0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac21f10
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac23b20
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac23960
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac0a360
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac0bc20
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac0f710
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac11470
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac12c00
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac15810
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac16f70
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac18400
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac19ec0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1a390
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1b5f0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1c320
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1d9c0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1dcc0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac1e7c0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac20380
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac208d0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac20eb0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac21520
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac22210
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac24fb0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac23610
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac26910
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Entered hgolofns at 2015/02/17-14:08:39
    symbol_peflctx=0xac275e0
    hoaerr:0
    Exiting hgolofns at 2015/02/17-14:08:39
    Exiting hgolofn, rc=0 at 2015/02/17-14:08:39
    HOSGIP for "HS_OPEN_CURSORS" returned "50"
    HOSGIP for "HS_FDS_FETCH_ROWS" returned "100"
    HOSGIP for "HS_LONG_PIECE_TRANSFER_SIZE" returned "65536"
    HOSGIP for "HS_NLS_NUMERIC_CHARACTERS" returned ".,"
    HOSGIP for "HS_KEEP_REMOTE_COLUMN_SIZE" returned "OFF"
    HOSGIP for "HS_FDS_DELAYED_OPEN" returned "TRUE"
    HOSGIP for "HS_FDS_WORKAROUNDS" returned "0"
    HOSGIP for "HS_FDS_MBCS_TO_GRAPHIC" returned "FALSE"
    HOSGIP for "HS_FDS_GRAPHIC_TO_MBCS" returned "FALSE"
    Invalid value of 64 given for HS_FDS_SQLLEN_INTERPRETATION
    treat_SQLLEN_as_compiled = 1
    Exiting hgoinit, rc=0 at 2015/02/17-14:08:39
    Entered hgolgon at 2015/02/17-14:08:39
    reco:0, name:peter, tflag:0
    Entered hgosuec at 2015/02/17-14:08:39
    Exiting hgosuec, rc=0 at 2015/02/17-14:08:39
    HOSGIP for "HS_FDS_RECOVERY_ACCOUNT" returned "RECOVER"
    HOSGIP for "HS_FDS_TRANSACTION_LOG" returned "HS_TRANSACTION_LOG"
    HOSGIP for "HS_FDS_TIMESTAMP_MAPPING" returned "DATE"
    HOSGIP for "HS_FDS_DATE_MAPPING" returned "DATE"
    HOSGIP for "HS_FDS_MAP_NCHAR" returned "TRUE"
    HOSGIP for "HS_FDS_RESULTSET_SUPPORT" returned "FALSE"
    HOSGIP for "HS_FDS_RSET_RETURN_ROWCOUNT" returned "FALSE"
    HOSGIP for "HS_FDS_PROC_IS_FUNC" returned "FALSE"
    HOSGIP for "HS_FDS_REPORT_REAL_AS_DOUBLE" returned "FALSE"
    using peter as default schema
    HOSGIP for "HS_SQL_HANDLE_STMT_REUSE" returned "FALSE"
    Entered hgocont at 2015/02/17-14:08:39
    HS_FDS_CONNECT_INFO = "myodbc5"
    RC=-1 from HOSGIP for "HS_FDS_CONNECT_STRING"
    Entered hgogenconstr at 2015/02/17-14:08:39
    dsn:myodbc5, name:peter
    optn:
    Entered hgocip at 2015/02/17-14:08:39
    dsn:myodbc5
    Exiting hgocip, rc=0 at 2015/02/17-14:08:39
    Exiting hgogenconstr, rc=0 at 2015/02/17-14:08:39
    Entered hgolosf at 2015/02/17-14:08:39
    Exiting hgolosf, rc=0 at 2015/02/17-14:08:39
    DriverName:libmyodbc5w.so, DriverVer:05.03.0004
    DBMS Name:MySQL, DBMS Version:5.6.14
    Exiting hgocont, rc=0 at 2015/02/17-14:08:39
    SQLGetInfo returns Y for SQL_CATALOG_NAME
    SQLGetInfo returns 192 for SQL_MAX_CATALOG_NAME_LEN
    Exiting hgolgon, rc=0 at 2015/02/17-14:08:39
    Entered hgoulcp at 2015/02/17-14:08:39
    Entered hgowlst at 2015/02/17-14:08:39
    Exiting hgowlst, rc=0 at 2015/02/17-14:08:39
    SQLGetInfo returns 0x0 for SQL_SCHEMA_USAGE
    TXN Capable:3, Isolation Option:0xf
    SQLGetInfo returns 0 for SQL_MAX_SCHEMA_NAME_LEN
    SQL_SU_DML_STATEMENTS bit is not set. Schemas are not supported by FDS.
    SQLGetInfo returns 192 for SQL_MAX_TABLE_NAME_LEN
    SQLGetInfo returns 192 for SQL_MAX_PROCEDURE_NAME_LEN
    HOSGIP returned value of "TRUE" for HS_FDS_QUOTE_IDENTIFIER
    SQLGetInfo returns ` (0x60) for SQL_IDENTIFIER_QUOTE_CHAR
    Entered hgopoer at 2015/02/17-14:08:39
    Exiting hgopoer, rc=0 at 2015/02/17-14:08:39 with error ptr FILE:hgopoer.c LINE:195 ID:GetDiagRec error
    hgoulcp, line 2138: calling SQLFetch got sqlstate 00000
    3 instance capabilities will be uploaded
      capno:5964, context:0x00000000, add-info:        0
      capno:5989, context:0x00000000, add-info:        0
      capno:5992, context:0x0001ffff, add-info:        1, translation:"`"
    Exiting hgoulcp, rc=0 at 2015/02/17-14:08:39 with error ptr FILE:hgoulcp.c LINE:2140 ID:Translation text for Unicode literal not supported. Leaving HOACOPTTSTN1 capability off.
    Entered hgouldt at 2015/02/17-14:08:39
    NO instance DD translations were uploaded
    Exiting hgouldt, rc=0 at 2015/02/17-14:08:39
    Entered hgobegn at 2015/02/17-14:08:39
    tflag:0 , initial:1
    hoi:0xd991a328, ttid (len 24) is ...
      00: 50455445 522E3864 63666537 31332E31  [PETER.8dcfe713.1]
      10: 2E33312E 32303235                    [.31.2025]
                     tbid (len 21) is ...
      00: 50455445 525B312E 33312E32 3032355D  [PETER[1.31.2025]]
      10: 5B312E34 5D                          [[1.4]]
    Exiting hgobegn, rc=0 at 2015/02/17-14:08:39
    Entered hgodtab at 2015/02/17-14:08:39
    count:1
      table: mysql_table1
    Allocate hoada[0] @ 0x10b6300
    Entered hgopcda at 2015/02/17-14:08:39
    Column:1(dest): dtype:1 (CHAR), prc/scl:5/0, nullbl:1, octet:5, sign:1, radix:0
    Exiting hgopcda, rc=0 at 2015/02/17-14:08:39
    The hoada for table mysql_table1 follows...
    hgodtab, line 1079: Printing hoada @ 0x10b6300
    MAX:1, ACTUAL:1, BRC:1, WHT=6 (TABLE_DESCRIBE)
    hoadaMOD bit-values found (0x200:TREAT_AS_CHAR)
    DTY      NULL-OK  LEN  MAXBUFLEN   PR/SC  CST IND MOD NAME
      1 CHAR Y          5          5   0/  0    0   0 200 dest
    Exiting hgodtab, rc=0 at 2015/02/17-14:08:39
    Entered hgodafr, cursor id 0 at 2015/02/17-14:08:39
    Free hoada @ 0x10b6300
    Exiting hgodafr, rc=0 at 2015/02/17-14:08:39
    Entered hgopars, cursor id 1 at 2015/02/17-14:08:39
    type:0
    SQL text from hgopars, id=1, len=39 ...
         00: 53454C45 43542041 312E6064 65737460  [SELECT A1.`dest`]
         10: 2046524F 4D20606D 7973716C 5F746162  [ FROM `mysql_tab]
         20: 6C653160 204131                      [le1` A1]
    Exiting hgopars, rc=0 at 2015/02/17-14:08:39
    Entered hgoopen, cursor id 1 at 2015/02/17-14:08:39
    hgoopen, line 87: NO hoada to print
    Deferred open until first fetch.
    Exiting hgoopen, rc=0 at 2015/02/17-14:08:39
    Entered hgodscr, cursor id 1 at 2015/02/17-14:08:39
    Allocate hoada @ 0x10b6300
    Entered hgodscr_process_sellist_description at 2015/02/17-14:08:39
    Entered hgopcda at 2015/02/17-14:08:39
    Column:1(dest): dtype:1 (CHAR), prc/scl:5/0, nullbl:1, octet:5, sign:1, radix:0
    Exiting hgopcda, rc=0 at 2015/02/17-14:08:39
    hgodscr, line 470: Printing hoada @ 0x10b6300
    MAX:1, ACTUAL:1, BRC:100, WHT=5 (SELECT_LIST)
    hoadaMOD bit-values found (0x200:TREAT_AS_CHAR)
    DTY      NULL-OK  LEN  MAXBUFLEN   PR/SC  CST IND MOD NAME
      1 CHAR Y          5          5   0/  0    0   0 200 dest
    Exiting hgodscr, rc=0 at 2015/02/17-14:08:39
    Entered hgoftch, cursor id 1 at 2015/02/17-14:08:39
    hgoftch, line 135: Printing hoada @ 0x10b6300
    MAX:1, ACTUAL:1, BRC:100, WHT=5 (SELECT_LIST)
    hoadaMOD bit-values found (0x200:TREAT_AS_CHAR)
    DTY      NULL-OK  LEN  MAXBUFLEN   PR/SC  CST IND MOD NAME
      1 CHAR Y          5          5   0/  0    0   0 200 dest
    Performing delayed open.
    SQLBindCol: column 1, cdatatype: 1, bflsz: 6
    SQLFetch: row: 1, column 1, bflsz: 6, bflar: 5
    SQLFetch: row: 1, column 1, bflsz: 6, bflar: 5, (bfl: 5, mbl: 5)
    1 rows fetched
    Exiting hgoftch, rc=0 at 2015/02/17-14:08:39
    Entered hgoftch, cursor id 1 at 2015/02/17-14:08:39
    hgoftch, line 135: Printing hoada @ 0x10b6300
    MAX:1, ACTUAL:1, BRC:1, WHT=5 (SELECT_LIST)
    hoadaMOD bit-values found (0x200:TREAT_AS_CHAR)
    DTY      NULL-OK  LEN  MAXBUFLEN   PR/SC  CST IND MOD NAME
      1 CHAR Y          5          5   0/  0    0   0 200 dest
    0 rows fetched
    Exiting hgoftch, rc=1403 at 2015/02/17-14:08:39
    Entered hgoclse, cursor id 1 at 2015/02/17-14:23:33
    Exiting hgoclse, rc=0 at 2015/02/17-14:23:33
    Entered hgodafr, cursor id 1 at 2015/02/17-14:23:33
    Free hoada @ 0x10b6300
    Exiting hgodafr, rc=0 at 2015/02/17-14:23:33
    Entered hgocomm at 2015/02/17-14:23:33
    keepinfo:0, tflag:1
       00: 50455445 522E3864 63666537 31332E31  [PETER.8dcfe713.1]
       10: 2E33312E 32303235                    [.31.2025]
                     tbid (len 21) is ...
       00: 50455445 525B312E 33312E32 3032355D  [PETER[1.31.2025]]
       10: 5B312E34 5D                          [[1.4]]
    cmt(0):
    Entered hgocpctx at 2015/02/17-14:23:33
    Exiting hgocpctx, rc=0 at 2015/02/17-14:23:33
    Exiting hgocomm, rc=0 at 2015/02/17-14:23:33
    Entered hgolgof at 2015/02/17-14:23:33
    tflag:1
    Exiting hgolgof, rc=0 at 2015/02/17-14:23:33
    Entered hgoexit at 2015/02/17-14:23:33
    Exiting hgoexit, rc=0
    Entered horcrces_CleanupExtprocSession at 2015/02/17-14:23:33
    Entered horcrpooe_PopOciEnv at 2015/02/17-14:23:33
    Entered horcrfoe_FreeOciEnv at 2015/02/17-14:23:33
    Exiting horcrfoe_FreeOciEnv at 2015/02/17-14:23:33
    Entered horcrfse_FreeStackElt at 2015/02/17-14:23:33
    Exiting horcrfse_FreeStackElt at 2015/02/17-14:23:33
    Exiting horcrpooe_PopOciEnv at 2015/02/17-14:23:33
    Exiting horcrces_CleanupExtprocSession at 2015/02/17-14:23:33

    Hi Matt,
    There is no error and data is returned.  It is just my curiosity to know about those lines in the log file and wonder whether they imply some underlying issues.
    Thanks,
    Peter

  • Any ideas creating app to read log file when updated?

    Afternoon all,
    I have been asked to write a java app which will read the contents of the server log file every time the log file is updated by the server. The app will be deployed onto WebSphere Application Server.
    Can anyone point me in the right direction as I have never written anything like this before and I don't know where to start. Any help will be much appreciated.
    Thanks in advance,
    A.

    alex@work wrote:
    I agree with most of what you've said but unfortunately I don't have a say in what the company wants. However, I am interested in the appender idea, perhaps they may go for that if I suggest it.I'd say it'll take you a day to read up about Log4J and how to write basic appenders, and another day to write your own appender for this problem. Compare that to the effort of writing something to poll a log file, re-read it constantly and update another file, operations which will get slower and slower as they go along. That's a fair amount more code than a single appender would be. There's how to sell it to your company.
    Can you give me a brief overview in how it works?Log4J uses objects called appenders, which take logging info - generated by your container - and do something with it. It ships with some appenders already in it, for writing to stdout, files, sockets and databases. You can also write your own appenders that do something more than these standard ones do. You write logging code in your application - in this case, your container already does this so you don't have to - and the configuration of Log4J decides what happens to these logging messages. That's what you're interested in. You could write an appender - a simple class - that takes raw logging messages in, and writes them out to file in whatever format you want
    Come to think of it, depending on how complex the required XML is, you may even be able to do this without writing any code at all. You can write formatting patterns in the Log4J config that existing file appenders will use to write your XML files
    A bit of an abstract explanation, I guess. Your best bet is to first ascertain that Log4J is indeed in use, and then read the documentation, which is surprisingly good for an Apache project :-)
    [http://logging.apache.org/log4j/1.2/index.html]

  • Reg:printing contents to log file

    Hello all iam creating a debug log file for my program.. debug will be based on the level set which will be taken as a input from the user ..when i try to print the contents to a file gen.out i could not print anything ..file is empty..can some one pls help me ..there is no error thrown by the program
    // my main propram
    public class ff {
      public static void main(String[] args) {
        PrintStream out = System.out;
        int dbgLevel = Integer.parseInt(args[1]);
        String path ="gen.out";
        debug.setLevel(dbgLevel);
        try {
            debug.setLogFile(path);
        catch(Exception e1) {
          e1.printStackTrace();
    debug.msg(0, "Printing contents to a file  ");
    if (Debug.allowed(0)) {
                    Debug.msg(0, "1.printing th econtents );
    //debug class
    public class debug {
      private static int ourLevel = 0;
      private static FileWriter c_writer;
      private static final String k_nl = System.getProperty("line.separator");
        public static void msg(int reqLevel, String debugMessage) {
        if (reqLevel <= ourLevel) {
          String s = "DBG(" + reqLevel + ") " + getTS() + " " + debugMessage + k_nl;
          if (c_writer == null) {
            System.out.println("writer is null "+c_writer.toString());
            System.err.println(s);
            System.err.flush();
          else {
              System.out.println("writer is not null "+c_writer.toString());
              try {
                         c_writer.write(s);
            catch (IOException e) {
                e.printStackTrace();
          } // if (c_writer...) else ...
        } // if (reqLevel...
      } // msg()
    public static void setLogFile(String path) throws IOException {
        c_writer = new FileWriter(path);
    public static void setLevel(int l) {
        ourLevel = l;
        public static boolean allowed(int reqLevel) {
        return reqLevel <= ourLevel;
    file is getting created but its empty ..wat am i doing wrong here .?

    thanks mkoryak got it i was flushing a wrong stream
    thanks for the help

  • What is the best way to merge a file content into log file

    What is the best way to merge a file content into log file.
    In worst case, I will read the file line by line as string, then use
    logger.info(lineString)to output to log file.
    However, is there better way to do this?
    The eventual log file will be something like:
    log message 1
    log message 2
    content from file line 1
    content from file line 2
    content from file line 3
    log message 3
    log message 4Thanks

    John618 wrote:
    Thank you and let me explain:
    1. What do you mean by better?
    I would like to see better performance. read line by line and log each line as string can be slow. Did you measure this and determine that it is actually a problem for your application? Or are you guessing?
    Regardless of what you do you are still going to need to read the file.
    >
    2.The only better way I can think of is not having to do it, but I assume you have a very good reason to want to do this.
    Yes, I have to do it beacuse the requirement is to have that file content be part of logging.
    Any idea?How is it supposed to be part of it? For example which of the following is better?
            File AAA - contents
                       First Line
                       Second Line XXX
            Log 1
                    2009-03-27 DEBUG: Random preceding line
                    2009-03-27 DEBUG: First Line
                    2009-03-27 DEBUG: Second Line XXX
                    2009-03-27 DEBUG: Random following line
            Log 2
                    2009-03-27 DEBUG: Random preceding line
                    2009-03-27 DEBUG: ----- File: AAA -------------
                    First Line
                    Second Line XXX
                    2009-03-27 DEBUG: Random following lineBoth of the above have some advantages and disadvantages.
    The first in a mult-threaded app can end up with intermittent log entries in between lines, so having log lines with thread ids becomes important.
    The first can be created by reading one line at a time and posting one at a time.
    The second can be created by reading the entire file as a single string and then posting using a single log statement.

  • Where are SmartView BI+ Content Independent Provider Log Files

    We are on Oracle/Hyperion v11.1.1.3 (BI/Foundation/Planning Services on Solaris Unix; Weblogic; Essbase on Windows).
    Our users have created some MS Powepoint presentations that contain pages that run a Financial Report (FR) which queries data from a planning application using a "Hyperion Reporting and Analysis" Independent provider connection (example: "http://server_name:19000/workspace/browse/listxml").
    I am unable to find the location of the log files that shows where logon/activity is tracked. We have looked and only find tracing information from Essbase log files when the query runs and authentication happens at the Essbase level but need to find where the other log information (including APS service activity) is logged to.
    Thanks.

    Hello Ramkumar,
    Go through the link https://help.sap.com/saphelp_nw73/helpdata/en/21/e77dc01e694f4f89ad6a833420c7d5/content.htm?frameset=/en/6a/f247bf0ce745…
    for more information..
    Regards
    Subbarao M

  • Meaning of Log File Contents

    MSWindowsXP/SP3, MSIE8.0,
    Flashplayer11.6 - update installed successfully today. 
    I want to have a plain english explanation for each of the entries in the log file created as part of the download / install.  Contents of the log appear below.
    I prefer to receive responses vai E-mail [removed] with explanation or with pointer to accessible source that contains complete information.  Thank you.
    Begin Log File
    ¿  TCSO  settings panel?ð
    autoUpdateInterval@
    crossdomainAllow
    safefullscreen
    windowlessDisable
    crossdomainAlways
    autoUpdateDisabled
    disallowP2PUplink defaultalways
      autoUpdateDefaultUpdated 
    secureCrossDomainCacheSize¿ð
    defaultmicrophone
    echosuppression
    allowThirdPartyLSOAccess
    trustedPaths
          domains     
    defaultcamera
    autoUpdateLastCheckBsÜ×w p gain@I
    defaultklimit authorizedFeaturesExpiry
    End Log File

    That is not a FlashInstall.log file; what is it?

  • How to check the contents of Control file, Log file & Parameter file

    Can anybody help me how i can check the contents of Control file, Log file & Parameter file.
    Arif

    OK ...
    Parameter file:
    It will normally be in the $ORACLE_HOME/dbs directopry. It could be aan init{sid}.ora or a spfile{sid}.ora ... do not edit an SPFILE as yu could corrupt it.
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14231/create.htm#sthref434
    If it's an init{sid}.ora, use any editor to see the actual content.
    If it's a spfile{sid}.ora, use one of V$PARAMETER, V$PARAMETER2, V$SYSTEM_PARAMETER, V$SYSTEM_PARAMETER2 as aappropriate (read the Reference manual to get an idea about when each is appropriate)
    Control File:
    You should not edit a control file. It consists of many areas, as described here http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14220/physical.htm#i10135 and each are is interrogated by V$views as given in the reference manual. The following V$views all deal with some aspect of the control file or provide info about the control files or the contents of control files:
    V$ARCHIVE_DEST
    V$ARCHIVED_LOG
    V$BACKUP_ARCHIVELOG_DETAILS
    V$BACKUP_CONTROLFILE_DETAILS
    V$BACKUP_CONTROLFILE_SUMMARY
    V$BACKUP_COPY_DETAILS
    V$BACKUP_COPY_SUMMARY
    V$BACKUP_CORRUPTION
    V$BACKUP_DATAFILE
    V$BACKUP_DATAFILE_DETAILS
    V$BACKUP_FILES
    V$BACKUP_PIECE
    V$BACKUP_REDOLOG
    V$BACKUP_SET
    V$BACKUP_SET_DETAILS
    V$BACKUP_SPFILE
    V$CONTROLFILE
    V$CONTROLFILE_RECORD_SECTION
    V$COPY_CORRUPTION
    V$DATABASE
    V$DATABASE_INCARNATION
    V$DATAFILE
    V$DATAFILE_COPY
    V$DATAFILE_HEADER
    V$DELETED_OBJECT
    V$LOCK
    V$LOG
    V$LOG_HISTORY
    V$LOGHIST
    V$OBSOLETE_BACKUP_FILES
    V$OFFLINE_RANGE
    V$PROXY_COPY_DETAILS
    V$PROXY_COPY_SUMMARY
    V$PROXY_DATAFILE
    V$RMAN_BACKUP_JOB_DETAILS
    V$RMAN_BACKUP_SUBJOB_DETAILS
    V$RMAN_CONFIGURATION
    V$SESSION
    V$TABLESPACE
    V$THREAD
    V$UNUSABLE_BACKUPFILE_DETAILS
    3) Log Files
    The contents of the log files are viewed using LogMiner - read Chapter 11 of the Oracle Utilities manual

  • Upgrade log file in MaintenanceLogs library breaks content deployment

    I have two SharePoint 2013 farms: authoring and publishing target.  Content deployment jobs are used to push changes from authoring.  This has been working just fine until I applied the latest updates.  The upgrade process create a log
    files in the _catalog/MaintenanceLogs library.  The Content Deployment process wants to deploy this list now.  Since it already exists on the target, the import fails.
    Has anyone run into this?  Is there any way to tell Content Deployment to not export this?  There's really no reason to publish this library.  Any way to delete it on the target so it just gets re-created?
    Ray Riopel

    Hi,
    Thank you for your question.
    We are currently looking into this issue and will give you an update as soon as possible.
    Thank you for your understanding and support.
    Thanks,
    Linda Li
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Linda Li
    TechNet Community Support

  • Delete content in alert.log file

    hii,
    Presently, i am working at oracle 10g database on windows 2003 server.
    alter.log file size is 432 MB around which take several time time to open.
    so, how can i delete content form alert.log file to reduce size
    Regards
    Vaibhav

    Hi,
    No need to delete contents of alertlog.
    just rename the alert log file or copy to another location and delete old one,then it will automatically creates new alert log file with format alert_<sid>.log
    note:
    no need of database bounce.

  • Accessing the script log file contents

    I would like to know the contents of the script log at a customer's site.   since we run DIADEM as a service,  it is not possible to  'see' the log.
    are the contents stored somewhere?
    is it possible to access (programatically) and write out contents?
    jim

    Hello Jim!
    The log is stored in a file. The filename with path is the concation of the variables DeskDrv and LogFilename. How verbose DIAdem is can be set via the LogFileExtent variable (see help) and in your script you can write to the log with the LogfileWrite command.
    Hope this helps
    Matthias
    Matthias Alleweldt
    Project Engineer / Projektingenieur
    Twigeater?  

Maybe you are looking for

  • Issue with reversal of MIRO for Planned Delivery Cost

    Hi Xperts We have found out an issue while reversing the MIRO document for Planned Del costs. When we have done the MIRO, the accounting entry got correctly posted with correct account keys.Conditions are not inventoried. However, when we had reverse

  • Where can I obtain a genuine Mac OS X 10.5.8 Installation Disc?

    Hi everyone, I bought a 2nd hand Apple MacBook about 18 months ago (via eBay).  It is my first Mac.  It did not come with its original OS X 10.5.8 Installation Disc/s and I now have one or more problems with my Hard Drive, so I need the Installation

  • Line items  AND GENERIC EXTRACTION

    what does line items exactly mean in DATASOURCES AND WHAT ARE DELTA TYPE EXTRACTIONS IN GENERIC TIME STAMPING CALENDAR DAY NUMERICAL POINTER COULD ANY PLEASE EXPLAIN THE DIFFERNCE  AND IN WHAT SCENARIOS WE USE IT ... LOOKING FOR YOUR REPLY

  • Error when editing region of the placeholder

    I defined a placeholder using the following documentation http://www.oracle.com/technology/products/content-management/ucm/sitestudio10gr4tutorials/AddingContent/addingcontent_viewlet_swf.html To summarize, following are the steps that I followed 1.

  • HTTPS with EV cert errors with embedded content

    <blockquote>Locking duplicate thread.<br> Please continue here: [[/questions/925376]]</blockquote> I have a site hosted under an EV cert. Once logged in, there is an embedded video that is hosted on our CDN. The Video has a valid cert, but it's not a