Idoc types that generate the log files

Hi All,
Can anybody provide me the list of IDOC types that generated application log files.
As I found that material master IDOCs generate log files. Is there IDOC types other
than material master that generate log files.
Regards,
Azra.

Hi Azra,
You can tell whether an IDOC does create the application log from the function module it uses.
Go to SE37, and type in IDOC_INPUT_<message_type> for inbound or IDOC_OUTPUT_<message_type> for outbound, and search for any function module which start with 'BAL_'. If you can find it, then it creates application log files.
Regards,
Lim...

Similar Messages

  • MDIS failed to generate the Log file!!!

    Hello All,
    Having a issue where MDIS is not generating the log file..
    The scenario is something like this-
    The files are getting archived and the records not flowing into MDM
    Basis team says-
    2014-06-30T14:11:33.339,47083231971072,24,"[MDS=sapdpm1 Repository=REAL_ESTATE ClientSystem=MDM_REAL_ESTATE Port=Building]: Nigerian Building updates part 2 - SLKDDY.txt is empty, the file will be skipped
    But the source file was having data it was not empty(bit strange!!)
    Also its not generating the LOG to analyze
    Regards,
    Girish

    Hi Shenoy,
    Let me explain the scenario--
    User uploads the file through Portal and through FTP records resides in MDM...the issues is when i tried to import through IM it worked and i tried manually push file through Filezilla FTP it worked.
    But when we upload file through portal, the file resides in Archive and generating the message-
    2014-06-30T14:11:33.339,47083231971072,24,"[MDS=sapdpm1 Repository=REAL_ESTATE ClientSystem=MDM_REAL_ESTATE Port=Building]: Nigerian Building updates part 2 - SLKDDY.txtis empty, the file will be skipped
    But the file is having data.
    Regards,
    Girish

  • Where can I file the log file generate by RMAN using EM 10g.

    Hi
    I trying to find the log file that is generated using RMAN invoked from EM.
    I can only see the file using Internet Explorer with the URL:
    em/console/database/rec/bkpMgmt?skey=257&type=oracle_database&target=isatprod.dla_dns.com&event=showJobDe
    But I need to find where the log files are located in the filesystem because in other server I will not have EM with OC4J.
    Thanks.
    Juan.

    When use OEM for 10g and choose the option / Maintance/Backup Reports
    I can see information of all my backups, that includes:
    Backup Name - Start Time - Time Taken - Status - Type- Output Devices - Input Size .....
    When I click on Status field I can see the log file of this Backup.
    (whe click in status one URL will be invoked ,something like below URL)
    http://10.5.0.86:1158/em/console/database/rec/bkpMgmt?skey=259&type=oracle_database&target=isatprod.dla_dns.com&event=showJobDet&objType=jobDtl
    So the log file exist in any place for every backup made , the problem that I can not find it.
    The log has approximately 500 lines, if you want I can send you by email the log.
    Currently I don't have a repository catalog, I use a control file as repository.
    I think that 500 lines of log is not include in any dynamic performance views.
    Thanks
    Juan

  • Type of error in the log file while using using call transaction mode u2018Eu2019

    Hi Gurus,
    Please Answer for this qusetion urgently
    what type of error exactly  you will be seeing in the log file while using call transaction mode u2018Eu2019?
    Thanks/
    Radha.

    Hi,
    Can you be clear.
    In call transaction , no error logs  are created, you have to handle the errors explicitly using the structure BDCMSGCOLL.
    Whenever you use E mode then if the transaction encounters any of the errors i.e. data type mismatching or invalid values etc, it will stop at that screen.
    You can handle the errors in call transaction in the following method.
    create a table using the structure BDCMSGCOLL.
    then
    loop at ......
          CALL TRANSACTION 'XK01' USING I_BDCDATA MODE 'N' UPDATE 'S' MESSAGES INTO I_MESGTAB.
    endloop.
      SORT I_MESGTAB BY MSGID MSGV1 ASCENDING.
      DELETE ADJACENT DUPLICATES FROM I_MESGTAB.
      LOOP AT I_MESGTAB.
        CALL FUNCTION 'FORMAT_MESSAGE'
          EXPORTING
            ID   = I_MESGTAB-MSGID
            LANG = I_MESGTAB-MSGSPRA
            NO   = I_MESGTAB-MSGNR
            V1   = I_MESGTAB-MSGV1
            V2   = I_MESGTAB-MSGV2
            V3   = I_MESGTAB-MSGV3
            V4   = I_MESGTAB-MSGV4
          IMPORTING
            MSG  = MESG1.
        IF I_MESGTAB-MSGTYP = 'S' .
          WA_SUCCMESG-MESG = MESG1.
          APPEND WA_SUCCMESG TO I_SUCCMESG.
    else     IF I_MESGTAB-MSGTYP = 'E' .
          WA_ERRMESG-MESG = MESG1.
          APPEND WA_ERRMESG TO I_ERRMESG.
        ENDIF.
      ENDLOOP.
    Hope this is clear.
    Thanks and Regards.

  • Edit the database that RH10 uses to generate the CPD file

    Is it possible to edit the database that RH10 uses to generate the CPD file? I've got one weird entry in the topics list in the Project Manager pod. I've deleted the CPD file and let RH10 rebuild it but the entry is still there.  My team leader would like me to get rid of the entry before we send the project to translation. Thanks.

    Dang, I saw that same problem occur with a project I was trying to fix but eventually I had to start over on a fresh copy of the broken project and did not create the problem again.
    If you are using an agency that translates and uses RoboHelp, the last thing I would have done is delete the snippet. They would have translated that once instead of everywhere it is used thus keeping costs down. That is an aside though.
    Do you have a backup of the project so that you can recreate the Snippet in the hope that the project is good with that in and you can then delete again to check you used the right option?
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • When using wusa.exe to install MSU update package and enabling logging using the /log switch, what format are the log files in?

    I have been installing a number of hotfixes for Windows 7 using MSU files and the wusa.exe utility included in Windows. I thought it would be a good idea to generate separate log files for each update as it was installed since wusa.exe now supports this
    option using /log:<file name>. However, the log files created do not seem to be regular text files or any other log file format that I immediately recognize. When opened in Notepad or Wordpad you can see that they contain a lot of additional binary data
    which can't be read by a regular text viewer.
    Does anyone know what format these log files are in? What tool should you use to read them?

    I have been installing a number of hotfixes for Windows 7 using MSU files and the wusa.exe utility included in Windows. I thought it would be a good idea to generate separate log files for each update as it was installed since wusa.exe now supports this
    option using /log:<file name>. However, the log files created do not seem to be regular text files or any other log file format that I immediately recognize. When opened in Notepad or Wordpad you can see that they contain a lot of additional binary data
    which can't be read by a regular text viewer.
    Does anyone know what format these log files are in? What tool should you use to read them?
    Only Microsoft can manage to design something as stupid as this. If you start wusa from the command line, it pops up the alternative command line switches. For log, it just says: " /log     - installer will enable logging". It doesn't say that you
    should specify the log file name, hence not HOW you specify it (" /log:<path\filename>.") It doesn't say what extension to use for the file, alas which file type it is (" /log:<path\filename.evtx>."). You open it in notepad, you cannot read it.
    You open it in sccm's cmtrace.exe/trace32.exe, you get nothing. IT IS NOT POSSIBLE TO IMPLEMENT THIS IN A WORSE WAY. How can something be so bad? How can Microsoft get such stupidity inhouse, when you need to go through six interviews or something to get in??
    I cannot believe it - unfortunely, this is seen again and again. 

  • Bursting Program Ends in Error with nothing in the Log file

    Hi All
    I have a RDF which calls a bursting program in After report trigger, the problem i'm facing is that the bursting program is completing successfully for a set of parameters and when again the program is run for the same set of parameters the bursting program is ending in error, i have checked the log file but nothing in the log file
    below is my log file
    +---------------------------------------------------------------------------+
    XML Publisher: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    XDOBURSTREP module: XML Publisher Report Bursting Program
    +---------------------------------------------------------------------------+
    Current system time is 30-SEP-2013 15:11:15
    +---------------------------------------------------------------------------+
    XML/BI Publisher Version : 5.6.3
    Request ID: 574533
    All Parameters: Dummy for Data Security=N:ReportRequestID=574532:DebugFlag=Y
    Report Req ID: 574532
    Debug Flag: Y
    Updating request description
    Updated description
    Retrieving XML request information
    Node Name:T1DEVEBSAPP1
    Preparing parameters
    null output =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574533.out
    inputfilename =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Data XML File:/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Set Bursting parameters..
    Temp. Directory:/tmp
    [093013_031118656][][STATEMENT] Oracle XML Parser version ::: Oracle XML Developers Kit 10.1.3.130 - Production
    [093013_031118657][][STATEMENT] setOAProperties called..
    Bursting propertes.....
    {user-variable:cp:territory=US, user-variable:cp:ReportRequestID=574532, user-variable:cp:language=en, user-variable:cp:responsibility=20678, user-variable.OA_MEDIA=http://t1devebsapp1.travelzoo.com:8001/OA_MEDIA, burstng-source=EBS, user-variable:cp:DebugFlag=Y, user-variable:cp:parent_request_id=574532, user-variable:cp:locale=en-US, user-variable:cp:user=SETUPUSER, user-variable:cp:application_short_name=XDO, user-variable:cp:request_id=574533, user-variable:cp:org_id=81, user-variable:cp:reportdescription=Travelzoo Invoice Print Selected Invoices-Child, user-variable:cp:Dummy for Data Security=N}
    Start bursting process..
    Bursting process complete..
    Generating Bursting Status Report..
    +---------------------------------------------------------------------------+
    Start of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    End of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    Executing request completion options...
    Output file size:
    602
    Finished executing request completion options.
    +---------------------------------------------------------------------------+
    Concurrent request completed
    Current system time is 30-SEP-2013 15:11:41
    +---------------------------------------------------------------------------+
    and the output says..
    "Error!! Could not deliver the output for Delivery channel:null "
    Wondering What might be the issue, any help on this is greatly appreciated.
    -Ragul

    Ram,
    Is this a custom or standard concurrent program?
    Was this working properly before? If yes, what changes have been done recently?
    Did you try to relink the concurrent program executable file and see if this helps? Also, you could enable trace/debug and submit the request again and see if more details are collected in the logs -- See (Note: 296559.1 - FAQ: Common Tracing Techniques within the Oracle Applications 11i/R12).
    Regards,
    Hussein

  • Bursting Program Errors Out with nothing in the Log file

    Hi All
    I have a RDF which calls a bursting program in After report trigger, the problem i'm facing is that the bursting program is completing successfully for a set of parameters and when again the program is run for the same set of parameters the bursting program is ending in error, i have checked the log file but nothing in the log file
    below is my log file
    +---------------------------------------------------------------------------+
    XML Publisher: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    XDOBURSTREP module: XML Publisher Report Bursting Program
    +---------------------------------------------------------------------------+
    Current system time is 30-SEP-2013 15:11:15
    +---------------------------------------------------------------------------+
    XML/BI Publisher Version : 5.6.3
    Request ID: 574533
    All Parameters: Dummy for Data Security=N:ReportRequestID=574532:DebugFlag=Y
    Report Req ID: 574532
    Debug Flag: Y
    Updating request description
    Updated description
    Retrieving XML request information
    Node Name:T1DEVEBSAPP1
    Preparing parameters
    null output =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574533.out
    inputfilename =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Data XML File:/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Set Bursting parameters..
    Temp. Directory:/tmp
    [093013_031118656][][STATEMENT] Oracle XML Parser version ::: Oracle XML Developers Kit 10.1.3.130 - Production
    [093013_031118657][][STATEMENT] setOAProperties called..
    Bursting propertes.....
    {user-variable:cp:territory=US, user-variable:cp:ReportRequestID=574532, user-variable:cp:language=en, user-variable:cp:responsibility=20678, user-variable.OA_MEDIA=http://t1devebsapp1.travelzoo.com:8001/OA_MEDIA, burstng-source=EBS, user-variable:cp:DebugFlag=Y, user-variable:cp:parent_request_id=574532, user-variable:cp:locale=en-US, user-variable:cp:user=SETUPUSER, user-variable:cp:application_short_name=XDO, user-variable:cp:request_id=574533, user-variable:cp:org_id=81, user-variable:cp:reportdescription=Travelzoo Invoice Print Selected Invoices-Child, user-variable:cp:Dummy for Data Security=N}
    Start bursting process..
    Bursting process complete..
    Generating Bursting Status Report..
    +---------------------------------------------------------------------------+
    Start of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    End of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    Executing request completion options...
    Output file size:
    602
    Finished executing request completion options.
    +---------------------------------------------------------------------------+
    Concurrent request completed
    Current system time is 30-SEP-2013 15:11:41
    +---------------------------------------------------------------------------+
    and the output says..
    "Error!! Could not deliver the output for Delivery channel:null "
    Wondering What might be the issue, any help on this is greatly appreciated.
    -Ragul

    Can you find any details about the error from the "View Detail" button (the same window where you check the log and output files)?
    I found the Workflow logs, I am not sure what I am looking for, but i am not seeing any errors reported.The event viewer is supposed to send an email, so do you see anything in the logs that could be related?
    Thanks,
    Hussein

  • I wonder what is included in the log files?

    My application generates 1G log files per day. when I use logmnr to
    analyze the log files, I find that 2000 rows which is performed by my
    application user is in the v$logmnr_contents view. And 55000 rows
    which is performed by the Oracle sys user. In these 5000 rows I can
    see SQL statements( Redo, Undo), and 50000 rows I can't see SQL
    statements which is identified by "unsupport statemet" or null.
    I wonder if my application generates 2000 rows, why oracle generates
    25 times than it. Can I tuning the log file that log file can be
    minimized.
    With best regard!

    Your problem may be related to the '''BLNGBAR''' shown at the end of your User Agent String. This answer from '''SafeBrowser''' may help -
    http://support.mozilla.org/en-US/questions/900875

  • Character Garbled in the log file

    HP-UX ia64
    Netweaver 7.0
    CPS M28.17-35130
    DB2 database
    HP-UX locale: en_US.utf8
    DB2 database encoding: utf8
    1: I have Chinese characters directly in the Redwoodscript(Job Definition), which would be printed into jcsErr/jcsOut.
       then in the log file, the Chinese characters are garbled.
    2: there is a job parameter as input string, and it will be printed into jcsErr/jscOut in the RedwoodScript.
       when submitting the job definition, set the parameter with some Chinese Characters,
       then in the log file, the Chinese characters are shown properly.
    I get down the java file in the /usr/sap/CPT/J00/j2ee/cluster/server0/java/compile0, and open it by Eclipse(utf-8),
    and see that the Chinese Characters in the java source file are shown properly.
    this is tested in our Product env.
    Once in our dev/test env, all is ok.
    and now I don't know what the difference causes issue 1.
    I have no idear how to trace this problem.
    hope you can shed some light to me.

    System_info output like below:
    System branch:M28.17
    System Id: SAP:CPT_00
    Build Id: M28.17-35130
    Server Root Directory  : null
    Job File Directory     : /usr/sap/CPT/J00/j2ee/cluster/server0/scheduler
    Java Work Directory    : /usr/sap/CPT/J00/j2ee/cluster/server0/java
    System Log Directory   : /usr/sap/CPT/J00/j2ee/cluster/server0/log
    Database type: (Open SQL) DatabaseType:DB2_ULW
    Database Driver Version: IBM DB2 JDBC Universal Driver Architecture 3.52.95
    Database Name: DB2/HPUX-IA64 (SQL09052)
    Database Connection URL: jdbc:db2://cpstprd:5912/CPT
    SAP_J2EE_Engine_Version=7.11.3301.221185.20081011173236
    SKIP_CLASS_WRAPPERS=false
    file.encoding=utf8
    file.encoding.pkg=sun.io
    file.separator=/
    http.nonProxyHosts=10.|127.|192.168.*
    j2ee.dbhost=cpstprd
    j2ee.engine.apps.folder=file:/usr/sap/CPT/J00/j2ee/cluster/apps/
    java.awt.graphicsenv=sun.awt.X11GraphicsEnvironment
    java.awt.headless=true
    java.awt.printerjob=sun.print.PSPrinterJob
    java.class.version=49.0
    java.endorsed.dirs=/usr/sap/CPT/J00/exe/sapjvm_5/jre/lib/endorsed
    java.ext.dirs=/usr/sap/CPT/J00/exe/sapjvm_5/jre/lib/ext
    java.home=/usr/sap/CPT/J00/exe/sapjvm_5/jre
    java.io.tmpdir=./temp
    java.protocol.handler.pkgs=iaik.protocol
    java.runtime.name=Java(TM) 2 Runtime Environment, Standard Edition
    java.runtime.version=5.1.024
    java.security.auth.login.config=./temp/security/fallback.config
    java.security.egd=file:/dev/urandom
    java.security.policy=./../bin/kernel/java.policy
    java.specification.name=Java Platform API Specification
    java.specification.vendor=Sun Microsystems Inc.
    java.specification.version=1.5
    java.vendor=SAP AG
    java.vendor.url=http://www.sap.com/
    java.vendor.url.bug=http://service.sap.com/support
    java.version=1.5.0_14
    java.vm.info=Sep  2 2008 10:46:19 - 51_REL - optU - hpux ia64 - bas2:106386 (mixed mode)
    java.vm.name=SAP Java 64-Bit Server VM
    java.vm.specification.name=Java Virtual Machine Specification
    java.vm.specification.vendor=Sun Microsystems Inc.
    java.vm.specification.version=1.0
    java.vm.vendor=SAP AG
    java.vm.version=5.1.024
    javax.rmi.CORBA.PortableRemoteObjectClass=com.sap.engine.system.PortableRemoteObjectProxy
    javax.security.jacc.policy.provider=com.sap.engine.system.SystemPolicy
    jco.jarm=1
    jmx.invoke.getters=true
    jstartup.debuggable=yes
    jstartup.mode=JSTART
    jstartup.ownHardwareId=S2011879030
    jstartup.ownProcessId=6447
    jstartup.whoami=server
    line.separator=

  • What are the log files created while running a mapping?

    Hi Sasi , I have a doubt who actually generates the " Log Events ". Is it the Log Manager or Application services ?

    Hi Anitha, The Integration service  will be generate two logs when the mapping runs 1) Session log  -- Has the details of the task ,session errors and load statistics..2) Workflow log -- Has the details of the workflow processing, and workflow errors..  The workflow log will be generated when the workflow started and the session log will be generated once the session intiated.For more detail Please refer the infa help docs... Normally the services will generate log  ex: IS and RS will log their activity...  The below process will happen the when the workflow inititated..[Copied fropm infa help docs] 1.#The Integration Service writes binary log files on the node. It sends  information about the sessions and workflows to the Log Manager.2.#The Log Manager stores information about workflow and session logs in the  domain configuration database. The domain configuration database stores  information such as the path to the log file location, the node that contains  the log, and the Integration Service that created the log.3.#When you view a session or workflow in the Log Events window, the Log Manager  retrieves the information from the domain configuration database to determine  the location of the session or workflow logs.4.#The Log Manager dispatches a Log Agent to retrieve the log events on each  node to display in the Log Events window.    ThanksSasiramesh

  • Generating rman log file

    hi
    How can I generate Rman log file? Oracle xe on windows
    My rman backup script is like:
    run{
    backup device type disk tag '%TAG' database;
    SQL 'ALTER SYSTEM ARCHIVE LOG CURRENT';
    backup device type disk tag '%TAG' archivelog all not backed up delete all input;
    delete noprompt obsolete device type disk;
    }

    In addition to what Paul mentioned, you can also consider the LOG option from the command line.

  • Confused about the log files

    I have written an application that has a Primary and Secondary database. The application creates tens-of-thousands of records in the Primary database, with a 1-to-1 relationship in the Secondary. On subsequent runs it will either update existing Primary records (which should not update the secondary as that element does not change) or it will create new records.
    The application actually works correctly, with the right data, the right updates and the right logical processing. The problem is the log files.
    The input data I am testing with is originally 2Mb as a CSV file and with a fresh database it creates almost 20Mb of data. This is about right for the way it splits the information up and indexes it. If I run the application again with exactly the same data, it should just update all the entries and create nothing new. My understanding is that the updated records will be written to the end of the logs, and the old ones in the earlier logs would be redundant and the cleaner thread would clean them up. I am explicitly cleaning as per the examples. The issue is that each run, the data just doubles in size! Logically it is fine, physically it is taking a ridiculous amount of space. RUnning DbSpace shows that the logs are mostly full (over 90%) where I would expect most to be empty, or sparsely occupied as the new updates are written to new files. cleanLog() does nothing. I am at a total loss!
    Generally the processing I am doing on the primary is looking up the key, if it is there updating the entry, if not creating one. I have been using a cursor to do this, and using the putCurrent() method for existing updates, and put() for new records. I have even tried using Database.delete() and the full put() in place of putCurrent() - but no difference (except it is slower).
    Please help - it is driving me nuts!

    Let me provide a little more context for the questions I was asking. If this doesn't lead us further into understanding your log situation, perhaps we should take this offline. When log cleaning doesn't occur, the basic questions are:
    a. is the application doing anything that prohibits log cleaning? (in your case, no)
    b. has the utilization level fallen to the point where log cleaning should occur? (not on the second run, but it should on following runs)
    c. does the log utilization level match what the application expects? (no, it doesn't match what you expect).
    1) Ran DbDump with and withour -r. I am expecting the
    data to stay consistent. So, after the first run it
    creates the data, and leaves 20mb in place, 3 log
    files near 100% used. After the second run it should
    update the records (which it does from the
    applications point of view) but I now have 40mb
    across 5 log files all near 100% usage.I think that it's accurate to say that both of us are not surprised that the second run (which updates data but does not change the number of records) creates a second 20MB of log, for a total of 40MB. What we do expect though, is that the utilization reported by DbSpace should fall to closer to 50%. Note that since JE's default minimum utilization level is 50%, we don't expect any automatic log cleaning even after the second run.
    Here's the sort of behavior we'd expect from JE if all the basics are taken care of (there are enough log files, there are no open txns, the application stays up long enough for the daemon to run, or the application does batch cleanLog calls itself, etc).
    run 1 - creates 20MB of log file, near 100% utilization, no log cleaning
    run 2 - updates every record, creates another 20MB of log file, utilization falls, maybe to around 60%. No log cleaning yet, because the utilization is still above the 50% threshold.
    run 3 - updates every record, creates another 20MB of log file, utilization falls below 50%, log cleaning starts running, either in the background by the daemon thread, or because the app calls Environment.cleanLog(), without any need to set je.cleaner.forceCleanFiles.
    So the question here is (c) from above -- you're saying that your DbSpace utilization level doesn't match what you believe your application is doing. There are three possible answers -- your application has a bug :-), or with secondaries and whatnot, JE is representing your data in a fashion you didn't expect, or JE's disk space utilization calculation is inaccurate.
    I suggested using DbDump -r as a first sanity check of what data your application holds. It will dump all the valid records in the environment (though not in key order, no -r is slower, but dumps in key order). Keys and data should up on different lines, so the number of lines in the dump files should be twice the number of records in the environment. You've done this already in your application, but this is an independent way of checking. It also makes it easier to see what portion of data is in primary versus secondary databases, because the data is dumped into per-database files. You could also load the data into a new, blank environment to look at it.
    I think asked you about the size of your records because a customer recently reported a JE disk utilization bug, which we are currently working on. It turns out that if your data records are very different in size (in this case, 4 orders of magnitude) and consistently only the larger or the smaller records are made obsolete, the utilization number gets out of whack. It doesn't really sound like your situation, because you're updating all your records, and they don't sound like they're that different in size. But nevertheless, here's a way of looking at what JE thinks your record sizes are. Run this command:
    java -jar je.jar DbPrintLog -h <envhome> -S
    and you'll see some output that talks about different types of log entries, and their sizes. Look at the lines that say LN and LN_TX at the top. These are data records. Do they match the sizes you expect? These lines do include JE's per-record headers. How large that is depends on whether your data is transactional or not. Non-transactional data records have a header of about 35 bytes, whereas transactional data records have 60 bytes added to them. If your data is small, that can be quite a large percentage. This is quite a lot more than for BDB (Core), partly because BDB (Core) doesn't have record level locking, and partly because we store a number of internal fields as 64 bit rather than 16 or 32 bit values.
    The line that's labelled "key/data" shows what portion JE thinks is the application's data. Note that DbPrintLog, unlike DbSpace, doesn't account for obsoleteness, so while you'll see a more detailed picture of what the records look like in the log, you may see more records than you expect.
    A last step we can take is to send you a development version of DbSpace that has a new feature to recalculate the utilization level. It runs more slowly than the vanilla DbSpace, but is a way of double checking the utilization level.
    In my first response, I suggested trying je.cleaner.forceCleanFiles just to make it clear that the cleaner will run, and to see if the problem is really around the question of what the utilization level should be. Setting that property lets the cleaner bypass the utilization trigger. If using it really reduced the size of your logs, it reinforces that your idea of what your application is doing is correct, and casts suspicion on the utilization calculation.
    So in summary, let's try these steps
    - use DbDump and DbPrintLog to double check the amount and size of your application data
    - make a table of runs, that shows the log size in bytes, number of log files, and the utilization level reported by DbSpace
    - run a je.cleaner.forceCleanFiles cleanLog loop on one of the logs that seems to have a high utilization level, and see how much it reduces to, and what the resulting utilization level is
    If it all points to JE, we'll probably take it offline, and ask for your test case.
    Regards,
    Linda

  • OWB does not store errors in the log files?

    I've specified the log file on the mapping. When I execute the mapping, I see that it does create the log file i specified. However, there are no error logs in the file itself. Is there a step I'm missing. What do I need to do in order for OWB to generate the error codes into the log file?

    The description and instructions for all the different log files and audit records are in this document
    http://otn.oracle.com/products/warehouse/pdf/Cases/case10.pdf
    Nikolai Rochnik

  • History report error: | An Exceptional Error occurred. Application exiting. Check the log file for error 5022

    Hi all
    I've got a error msg when try to generate a report using Cisco history report tool:
    Error | An Exceptional Error occurred. Application exiting.  Check the log file for error 5022
    It only happens when choose report template: ICD_Contact_Service_Queue_Activity_by_CSQ_en_us.
    User tried samething on other PC, it working fine.
    only on user' own PC and only choose this report, error appears.
    user runing windows 7 and do not have crystal report installed
    tried reinstalled the software, doesn't work.
    also tried this: (https://cisco-support.hosted.jivesoftware.com/thread/2041254) - doesn't work
    then tried https://supportforums.cisco.com/docs/DOC-6209  - doesn't work
    attached the log file.
    thanks.

    wenqianyu wrote:From the log file:Looks like you get a Login Window.Error message showed up after username/password be enteredThere is an error in the log: Error happened in comparing UCCX version and HRC versionYou may need to do a clean uninstall, download the Historical report from the server, and install it again on the PC.Does this only happen to one PC or to every PC with this application?Wenqian 
    I have completely uninstalled the HRC, and download from server install again -- still doesn't work with exactly same error.
    this matter only happens on this PC, when user try same thing on other PC, it works.
    so i think it not relate to server or account.

Maybe you are looking for