I wonder what is included in the log files?

My application generates 1G log files per day. when I use logmnr to
analyze the log files, I find that 2000 rows which is performed by my
application user is in the v$logmnr_contents view. And 55000 rows
which is performed by the Oracle sys user. In these 5000 rows I can
see SQL statements( Redo, Undo), and 50000 rows I can't see SQL
statements which is identified by "unsupport statemet" or null.
I wonder if my application generates 2000 rows, why oracle generates
25 times than it. Can I tuning the log file that log file can be
minimized.
With best regard!

Your problem may be related to the '''BLNGBAR''' shown at the end of your User Agent String. This answer from '''SafeBrowser''' may help -
http://support.mozilla.org/en-US/questions/900875

Similar Messages

  • (Cisco Historical Reporting / HRC ) All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054

    Hi All,
    I am getting an error message "All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054"  when trying to log into HRC (This user has the reporting capabilities) . I checked the log files this is what i found out 
    The log file stated that there were ongoing connections of HRC with the CCX  (I am sure there isn't any active login to HRC)
    || When you tried to login the following error was being displayed because the maximum number of connections were reached for the server .  We can see that a total number of 5 connections have been configured . ||
    1: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Current number of connections (5) from historical Clients/Scheduler to 'CRA_DATABASE' database exceeded the maximum number of possible connections (5).Check with your administrator about changing this limit on server (wfengine.properties), however this might impact server performance.
    || Below we can see all 5 connections being used up . ||
    2: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:[DB Connections From Clients (count=5)]|[(#1) 'username'='uccxhrc','hostname'='3SK5FS1.ucsfmedicalcenter.org']|[(#2) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#3) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#4) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#5) 'username'='uccxhrc','hostname'='47BMMM1.ucsfmedicalcenter.org']
    || Once the maximum number of connection was reached it threw an error . ||
    3: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Number of max connection to 'CRA_DATABASE' database was reached! Connection could not be established.
    4: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Database connection to 'CRA_DATABASE' failed due to (All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054.)
    Current exact UCCX Version 9.0.2.11001-24
    Current CUCM Version 8.6.2.23900-10
    Business impact  Not Critical
    Exact error message  All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054
    What is the OS version of the PC you are running  and is it physical machine or virtual machine that is running the HRC client ..
    OS Version Windows 7 Home Premium  64 bit and it’s a physical machine.
    . The Max DB Connections for Report Client Sessions is set to 5 for each servers (There are two servers). The no of HR Sessions is set to 10.
    I wanted to know if there is a way to find the HRC sessions active now and terminate the one or more or all of that sessions from the server end ? 

    We have had this "PRX5" problem with Exchange 2013 since the RTM version.  We recently applied CU3, and it did not correct the problem.  We have seen this problem on every Exchange 2013 we manage.  They are all installations where all roles
    are installed on the same Windows server, and in our case, they are all Windows virtual machines using Windows 2012 Hyper-V.
    We have tried all the "this fixed it for me" solutions regarding DNS, network cards, host file entries and so forth.  None of those "solutions" made any difference whatsoever.  The occurrence of the temporary error PRX5 seems totally random. 
    About 2 out of 20 incoming mail test by Microsoft Connectivity Analyzer fail with this PRX5 error.
    Most people don't ever notice the issue because remote mail servers retry the connection later.  However, telephone voice mail systems that forward voice message files to email, or other such applications such as your scanner, often don't retry and
    simply fail.  Our phone system actually disables all further attempts to send voice mail to a particular user if the PRX5 error is returned when the email is sent by the phone system.
    Is Microsoft totally oblivious to this problem?
    PRX5 is a serious issue that needs an Exchange team resolution, or at least an acknowledgement that the problem actually does exist and has negative consequences for proper mail flow.
    JSB

  • When using wusa.exe to install MSU update package and enabling logging using the /log switch, what format are the log files in?

    I have been installing a number of hotfixes for Windows 7 using MSU files and the wusa.exe utility included in Windows. I thought it would be a good idea to generate separate log files for each update as it was installed since wusa.exe now supports this
    option using /log:<file name>. However, the log files created do not seem to be regular text files or any other log file format that I immediately recognize. When opened in Notepad or Wordpad you can see that they contain a lot of additional binary data
    which can't be read by a regular text viewer.
    Does anyone know what format these log files are in? What tool should you use to read them?

    I have been installing a number of hotfixes for Windows 7 using MSU files and the wusa.exe utility included in Windows. I thought it would be a good idea to generate separate log files for each update as it was installed since wusa.exe now supports this
    option using /log:<file name>. However, the log files created do not seem to be regular text files or any other log file format that I immediately recognize. When opened in Notepad or Wordpad you can see that they contain a lot of additional binary data
    which can't be read by a regular text viewer.
    Does anyone know what format these log files are in? What tool should you use to read them?
    Only Microsoft can manage to design something as stupid as this. If you start wusa from the command line, it pops up the alternative command line switches. For log, it just says: " /log     - installer will enable logging". It doesn't say that you
    should specify the log file name, hence not HOW you specify it (" /log:<path\filename>.") It doesn't say what extension to use for the file, alas which file type it is (" /log:<path\filename.evtx>."). You open it in notepad, you cannot read it.
    You open it in sccm's cmtrace.exe/trace32.exe, you get nothing. IT IS NOT POSSIBLE TO IMPLEMENT THIS IN A WORSE WAY. How can something be so bad? How can Microsoft get such stupidity inhouse, when you need to go through six interviews or something to get in??
    I cannot believe it - unfortunely, this is seen again and again. 

  • When I look at the storage on my Mac under about this Mac, what is included in the "other" category?

    When I look at the storage on my Mac under about this Mac, what is included in the "other" category?

    Hi, just about everything else...
    OmniDiskSweeper is likely the easiest/best, and is now free...
    http://www.omnigroup.com/applications/omnidisksweeper/download/
    And if needed, you can use Terminal to run it in admin mode to see files/folders your user can't see...
    http://www.macobserver.com/tmo/article/how_to_recover_missing_hard_drive_space/

  • HT1386 what is included in the "other" category

    I am trying to find out what is included in the yellow "other" category on my iPad. I have more space taken up by that than anything else.

    This is copied from one of the countless discussions about the "other" issue on iDevices.
    That large of "other" is corrupt data. Try restoring your iPad from backup first, followed by syncing your content back to the device. If restoring from backup does not fix things, you will have to restore as a new device to get rid of it. There is no way to directly delete it, other than restoring your iPad.
    Other includes app data and files, Safari Bookmarks and history, Notes, Calendar entries, text messages and email messages - basically the non media items on the iPad. If it is taking up that much room - more than anything else on your iPad - you should follow the suggestion I posted above.

  • What is included with the ipad 32bit?

    What is included with the Ipad2 wi-fi 3G?

    What do you mean by, what is included?  Do you mean what is included in the box?  If you do, you get an iPad,  a USB cable, a 10W charger, an Apple Logo Sticker and a very small booklet that gives you the most basic set up instructions.

  • What's Included in the Server Package

    I have been using SBS08 for the past few years and we are now ready to upgrade our system to Server 2012. We have more users than would be allowed by the Essentials package.
    When we purchased SBS08, it came with Exchange and SQL.
    My question is what is included in the server package? Does it come with Exchange and SQL or do these applications have to be purchased separately?
    Thanks!

    Server comes with Server.  You can see what is there by looking at the TechNet library.
    http://technet.microsoft.com/en-us/library/bb625087.aspx
    Applications like Exchange, SQL, SharePoint, etc. are all extras.
    Personally (I do not work for Microsoft) when making a move like you are making, I would seriously consider looking into Office 365 or some other hosted Exchange service rather than paying to set one up internally.  My logic on this is that it is a
    commodity.  I have seen very, very, very few Exchange installations that were different enough that the management of it was unique to the company and provided a strategic value to the company.  If your Exchange implementation is not of that
    flavor, why bother with the internal expense of setting it up and managing it?  Initial reaction is often that one can do it for less than a subscription service, but if you look at all the costs of running it and the cost of the time it takes away from
    working on applications that do provide strategic value to your company, you might find the cost of a subscription service to be less costly.
    . : | : . : | : . tim

  • How can i know what is included in the start up time of firefox ?

    hi. where can i find a list of what is included in the start up of my browser ?
    If for example, it takes almost 2 minutes for my browser to start..
    i want to reduce the time where it's Possible...
    please direct me to a place that gives that info..
    '''thanks''''''bold text'''

    Your problem may be related to the '''BLNGBAR''' shown at the end of your User Agent String. This answer from '''SafeBrowser''' may help -
    http://support.mozilla.org/en-US/questions/900875

  • What is included in the 1 year free warranty for an iphone4?

    What is included in the 1 year free warranty for an iphone4?

    Like most any warranty it covers manufacturer defects.
    It does not cover anything else.
    Apple's Limited Warranty for iPhone excludes coverage for damage resulting from accident, disassembly, unauthorized service and unauthorized modifications. Please review the warranty for further details.
    An iPhone that is ineligible for warranty service may be eligible for Out-of-Warranty (OOW) Service. For example, an iPhone that has failed due to damage or liquid damage is ineligible for warranty service but eligible for OOW service. However, certain damage is ineligible for OOW service, including catastrophic damage, such as the device separating into multiple pieces, and inoperability caused by unauthorized modifications. Please see Apple's Repair Terms And Conditions for further details. Refer to this article to check if the Liquid Contact indicator on your product has been triggered.
    If your iPhone is not eligible for warranty service and you are a Verizon customer, you may wish to contact Verizon for other options.
    http://support.apple.com/kb/index?page=servicefaq&geo=United_States&product=ipho ne

  • My http frequently down ,what is wrong from the log ?

    my message server is iplanet message server 5.2 for solaris9 ,and the mail express usually down ,and the time span maybe 10 day or 1 day ,from the http log i can not see the useful log .and when the express can not access the pop and smtp service is ok .and the use is not very large. when i see the http log ,i have some doubt ?
    1. what is the unauthenticated?
    2. why i can not see the open session only the close session ?
    3. why has the 127.0.0.1 access ,and i have not access the express via 127.0.0.1 ?
    4. how to config the http log then it will prompt more detail log ?
    thanks for you !
    [25/Oct/2006:14:34:20 +0800] v880 httpd[8324]: General Information: Log created (1161758060)
    [25/Oct/2006:14:34:20 +0800] v880 httpd[8324]: Account Notice: close [40.15.22.27] [unauthenticated] 2006/10/25 14:32:59 0:01:21 735 80679 0
    [25/Oct/2006:14:34:26 +0800] v880 httpd[8324]: Account Notice: close [222.189.61.178] [unauthenticated] 2006/10/25 14:33:04 0:01:22 2980 23573 0
    [25/Oct/2006:14:34:36 +0800] v880 httpd[8324]: Account Notice: close [127.0.0.1] [unauthenticated] 2006/10/25 14:34:36 0:00:00 0 0 0
    [25/Oct/2006:14:35:43 +0800] v880 httpd[8324]: Account Notice: close 40.15.44.34 chuyj 2006/10/25 14:05:06 0:30:37 0 0 1
    [25/Oct/2006:14:44:09 +0800] v880 httpd[8324]: General Notice: hanjj[40.40.40.35] created message <[email protected]>
    [25/Oct/2006:14:44:36 +0800] v880 httpd[8324]: Account Notice: close [127.0.0.1] [unauthenticated] 2006/10/25 14:44:36 0:00:00 0 0 0
    [25/Oct/2006:14:44:43 +0800] v880 httpd[8324]: Account Notice: close 40.15.43.12 zhoun 2006/10/25 14:13:25 0:31:18 0 0 1

    It's also a waste of my time. Thank you, Shane.
    Your questions about connections from 127.0.0.1 are valid questions. The server tests itself, every 10 minutes. That's what you see in the logs.
    The log only shows disconnect, at this loglevel. That's why you only see that.
    Both are normal.
    you really need to apply 5.2patch2. It's free, and you get it here:
    http://www.sun.com/download/products.xml?id=411bb50d

  • What are the log files created while running a mapping?

    Hi Sasi , I have a doubt who actually generates the " Log Events ". Is it the Log Manager or Application services ?

    Hi Anitha, The Integration service  will be generate two logs when the mapping runs 1) Session log  -- Has the details of the task ,session errors and load statistics..2) Workflow log -- Has the details of the workflow processing, and workflow errors..  The workflow log will be generated when the workflow started and the session log will be generated once the session intiated.For more detail Please refer the infa help docs... Normally the services will generate log  ex: IS and RS will log their activity...  The below process will happen the when the workflow inititated..[Copied fropm infa help docs] 1.#The Integration Service writes binary log files on the node. It sends  information about the sessions and workflows to the Log Manager.2.#The Log Manager stores information about workflow and session logs in the  domain configuration database. The domain configuration database stores  information such as the path to the log file location, the node that contains  the log, and the Integration Service that created the log.3.#When you view a session or workflow in the Log Events window, the Log Manager  retrieves the information from the domain configuration database to determine  the location of the session or workflow logs.4.#The Log Manager dispatches a Log Agent to retrieve the log events on each  node to display in the Log Events window.    ThanksSasiramesh

  • Confused about the log files

    I have written an application that has a Primary and Secondary database. The application creates tens-of-thousands of records in the Primary database, with a 1-to-1 relationship in the Secondary. On subsequent runs it will either update existing Primary records (which should not update the secondary as that element does not change) or it will create new records.
    The application actually works correctly, with the right data, the right updates and the right logical processing. The problem is the log files.
    The input data I am testing with is originally 2Mb as a CSV file and with a fresh database it creates almost 20Mb of data. This is about right for the way it splits the information up and indexes it. If I run the application again with exactly the same data, it should just update all the entries and create nothing new. My understanding is that the updated records will be written to the end of the logs, and the old ones in the earlier logs would be redundant and the cleaner thread would clean them up. I am explicitly cleaning as per the examples. The issue is that each run, the data just doubles in size! Logically it is fine, physically it is taking a ridiculous amount of space. RUnning DbSpace shows that the logs are mostly full (over 90%) where I would expect most to be empty, or sparsely occupied as the new updates are written to new files. cleanLog() does nothing. I am at a total loss!
    Generally the processing I am doing on the primary is looking up the key, if it is there updating the entry, if not creating one. I have been using a cursor to do this, and using the putCurrent() method for existing updates, and put() for new records. I have even tried using Database.delete() and the full put() in place of putCurrent() - but no difference (except it is slower).
    Please help - it is driving me nuts!

    Let me provide a little more context for the questions I was asking. If this doesn't lead us further into understanding your log situation, perhaps we should take this offline. When log cleaning doesn't occur, the basic questions are:
    a. is the application doing anything that prohibits log cleaning? (in your case, no)
    b. has the utilization level fallen to the point where log cleaning should occur? (not on the second run, but it should on following runs)
    c. does the log utilization level match what the application expects? (no, it doesn't match what you expect).
    1) Ran DbDump with and withour -r. I am expecting the
    data to stay consistent. So, after the first run it
    creates the data, and leaves 20mb in place, 3 log
    files near 100% used. After the second run it should
    update the records (which it does from the
    applications point of view) but I now have 40mb
    across 5 log files all near 100% usage.I think that it's accurate to say that both of us are not surprised that the second run (which updates data but does not change the number of records) creates a second 20MB of log, for a total of 40MB. What we do expect though, is that the utilization reported by DbSpace should fall to closer to 50%. Note that since JE's default minimum utilization level is 50%, we don't expect any automatic log cleaning even after the second run.
    Here's the sort of behavior we'd expect from JE if all the basics are taken care of (there are enough log files, there are no open txns, the application stays up long enough for the daemon to run, or the application does batch cleanLog calls itself, etc).
    run 1 - creates 20MB of log file, near 100% utilization, no log cleaning
    run 2 - updates every record, creates another 20MB of log file, utilization falls, maybe to around 60%. No log cleaning yet, because the utilization is still above the 50% threshold.
    run 3 - updates every record, creates another 20MB of log file, utilization falls below 50%, log cleaning starts running, either in the background by the daemon thread, or because the app calls Environment.cleanLog(), without any need to set je.cleaner.forceCleanFiles.
    So the question here is (c) from above -- you're saying that your DbSpace utilization level doesn't match what you believe your application is doing. There are three possible answers -- your application has a bug :-), or with secondaries and whatnot, JE is representing your data in a fashion you didn't expect, or JE's disk space utilization calculation is inaccurate.
    I suggested using DbDump -r as a first sanity check of what data your application holds. It will dump all the valid records in the environment (though not in key order, no -r is slower, but dumps in key order). Keys and data should up on different lines, so the number of lines in the dump files should be twice the number of records in the environment. You've done this already in your application, but this is an independent way of checking. It also makes it easier to see what portion of data is in primary versus secondary databases, because the data is dumped into per-database files. You could also load the data into a new, blank environment to look at it.
    I think asked you about the size of your records because a customer recently reported a JE disk utilization bug, which we are currently working on. It turns out that if your data records are very different in size (in this case, 4 orders of magnitude) and consistently only the larger or the smaller records are made obsolete, the utilization number gets out of whack. It doesn't really sound like your situation, because you're updating all your records, and they don't sound like they're that different in size. But nevertheless, here's a way of looking at what JE thinks your record sizes are. Run this command:
    java -jar je.jar DbPrintLog -h <envhome> -S
    and you'll see some output that talks about different types of log entries, and their sizes. Look at the lines that say LN and LN_TX at the top. These are data records. Do they match the sizes you expect? These lines do include JE's per-record headers. How large that is depends on whether your data is transactional or not. Non-transactional data records have a header of about 35 bytes, whereas transactional data records have 60 bytes added to them. If your data is small, that can be quite a large percentage. This is quite a lot more than for BDB (Core), partly because BDB (Core) doesn't have record level locking, and partly because we store a number of internal fields as 64 bit rather than 16 or 32 bit values.
    The line that's labelled "key/data" shows what portion JE thinks is the application's data. Note that DbPrintLog, unlike DbSpace, doesn't account for obsoleteness, so while you'll see a more detailed picture of what the records look like in the log, you may see more records than you expect.
    A last step we can take is to send you a development version of DbSpace that has a new feature to recalculate the utilization level. It runs more slowly than the vanilla DbSpace, but is a way of double checking the utilization level.
    In my first response, I suggested trying je.cleaner.forceCleanFiles just to make it clear that the cleaner will run, and to see if the problem is really around the question of what the utilization level should be. Setting that property lets the cleaner bypass the utilization trigger. If using it really reduced the size of your logs, it reinforces that your idea of what your application is doing is correct, and casts suspicion on the utilization calculation.
    So in summary, let's try these steps
    - use DbDump and DbPrintLog to double check the amount and size of your application data
    - make a table of runs, that shows the log size in bytes, number of log files, and the utilization level reported by DbSpace
    - run a je.cleaner.forceCleanFiles cleanLog loop on one of the logs that seems to have a high utilization level, and see how much it reduces to, and what the resulting utilization level is
    If it all points to JE, we'll probably take it offline, and ask for your test case.
    Regards,
    Linda

  • Bursting Program Ends in Error with nothing in the Log file

    Hi All
    I have a RDF which calls a bursting program in After report trigger, the problem i'm facing is that the bursting program is completing successfully for a set of parameters and when again the program is run for the same set of parameters the bursting program is ending in error, i have checked the log file but nothing in the log file
    below is my log file
    +---------------------------------------------------------------------------+
    XML Publisher: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    XDOBURSTREP module: XML Publisher Report Bursting Program
    +---------------------------------------------------------------------------+
    Current system time is 30-SEP-2013 15:11:15
    +---------------------------------------------------------------------------+
    XML/BI Publisher Version : 5.6.3
    Request ID: 574533
    All Parameters: Dummy for Data Security=N:ReportRequestID=574532:DebugFlag=Y
    Report Req ID: 574532
    Debug Flag: Y
    Updating request description
    Updated description
    Retrieving XML request information
    Node Name:T1DEVEBSAPP1
    Preparing parameters
    null output =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574533.out
    inputfilename =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Data XML File:/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Set Bursting parameters..
    Temp. Directory:/tmp
    [093013_031118656][][STATEMENT] Oracle XML Parser version ::: Oracle XML Developers Kit 10.1.3.130 - Production
    [093013_031118657][][STATEMENT] setOAProperties called..
    Bursting propertes.....
    {user-variable:cp:territory=US, user-variable:cp:ReportRequestID=574532, user-variable:cp:language=en, user-variable:cp:responsibility=20678, user-variable.OA_MEDIA=http://t1devebsapp1.travelzoo.com:8001/OA_MEDIA, burstng-source=EBS, user-variable:cp:DebugFlag=Y, user-variable:cp:parent_request_id=574532, user-variable:cp:locale=en-US, user-variable:cp:user=SETUPUSER, user-variable:cp:application_short_name=XDO, user-variable:cp:request_id=574533, user-variable:cp:org_id=81, user-variable:cp:reportdescription=Travelzoo Invoice Print Selected Invoices-Child, user-variable:cp:Dummy for Data Security=N}
    Start bursting process..
    Bursting process complete..
    Generating Bursting Status Report..
    +---------------------------------------------------------------------------+
    Start of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    End of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    Executing request completion options...
    Output file size:
    602
    Finished executing request completion options.
    +---------------------------------------------------------------------------+
    Concurrent request completed
    Current system time is 30-SEP-2013 15:11:41
    +---------------------------------------------------------------------------+
    and the output says..
    "Error!! Could not deliver the output for Delivery channel:null "
    Wondering What might be the issue, any help on this is greatly appreciated.
    -Ragul

    Ram,
    Is this a custom or standard concurrent program?
    Was this working properly before? If yes, what changes have been done recently?
    Did you try to relink the concurrent program executable file and see if this helps? Also, you could enable trace/debug and submit the request again and see if more details are collected in the logs -- See (Note: 296559.1 - FAQ: Common Tracing Techniques within the Oracle Applications 11i/R12).
    Regards,
    Hussein

  • Bursting Program Errors Out with nothing in the Log file

    Hi All
    I have a RDF which calls a bursting program in After report trigger, the problem i'm facing is that the bursting program is completing successfully for a set of parameters and when again the program is run for the same set of parameters the bursting program is ending in error, i have checked the log file but nothing in the log file
    below is my log file
    +---------------------------------------------------------------------------+
    XML Publisher: Version : 12.0.0
    Copyright (c) 1979, 1999, Oracle Corporation. All rights reserved.
    XDOBURSTREP module: XML Publisher Report Bursting Program
    +---------------------------------------------------------------------------+
    Current system time is 30-SEP-2013 15:11:15
    +---------------------------------------------------------------------------+
    XML/BI Publisher Version : 5.6.3
    Request ID: 574533
    All Parameters: Dummy for Data Security=N:ReportRequestID=574532:DebugFlag=Y
    Report Req ID: 574532
    Debug Flag: Y
    Updating request description
    Updated description
    Retrieving XML request information
    Node Name:T1DEVEBSAPP1
    Preparing parameters
    null output =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574533.out
    inputfilename =/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Data XML File:/u01/oracle/CRP2/inst/apps/CRP2_t1devebsapp1/logs/appl/conc/out/o574532.out
    Set Bursting parameters..
    Temp. Directory:/tmp
    [093013_031118656][][STATEMENT] Oracle XML Parser version ::: Oracle XML Developers Kit 10.1.3.130 - Production
    [093013_031118657][][STATEMENT] setOAProperties called..
    Bursting propertes.....
    {user-variable:cp:territory=US, user-variable:cp:ReportRequestID=574532, user-variable:cp:language=en, user-variable:cp:responsibility=20678, user-variable.OA_MEDIA=http://t1devebsapp1.travelzoo.com:8001/OA_MEDIA, burstng-source=EBS, user-variable:cp:DebugFlag=Y, user-variable:cp:parent_request_id=574532, user-variable:cp:locale=en-US, user-variable:cp:user=SETUPUSER, user-variable:cp:application_short_name=XDO, user-variable:cp:request_id=574533, user-variable:cp:org_id=81, user-variable:cp:reportdescription=Travelzoo Invoice Print Selected Invoices-Child, user-variable:cp:Dummy for Data Security=N}
    Start bursting process..
    Bursting process complete..
    Generating Bursting Status Report..
    +---------------------------------------------------------------------------+
    Start of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    End of log messages from FND_FILE
    +---------------------------------------------------------------------------+
    +---------------------------------------------------------------------------+
    Executing request completion options...
    Output file size:
    602
    Finished executing request completion options.
    +---------------------------------------------------------------------------+
    Concurrent request completed
    Current system time is 30-SEP-2013 15:11:41
    +---------------------------------------------------------------------------+
    and the output says..
    "Error!! Could not deliver the output for Delivery channel:null "
    Wondering What might be the issue, any help on this is greatly appreciated.
    -Ragul

    Can you find any details about the error from the "View Detail" button (the same window where you check the log and output files)?
    I found the Workflow logs, I am not sure what I am looking for, but i am not seeing any errors reported.The event viewer is supposed to send an email, so do you see anything in the logs that could be related?
    Thanks,
    Hussein

  • How to add a date suffix to the log file name

    In Windows, I want to run certain commands and save the output to a logfile every day. How to add a suffix to the log file name so I can distinguish which log file for which day?
    e.g. cmd >> logfile.date

    AZ wrote:
    In Windows, I want to run certain commands and save the output to a logfile every day. How to add a suffix to the log file name so I can distinguish which log file for which day?
    e.g. cmd >> logfile.datemy best friend name is "google", refer to this [url | http://stackoverflow.com/questions/203090/how-to-get-current-datetime-on-windows-command-line-in-a-suitable-format-for-usi]
    This is what i did
    1) created a dummy file in c drive
    2) copy pasted below lines, you can play around more with the format
    set _my_datetime=%date%_%time%
    set _my_datetime=%_my_datetime: =_%
    set _my_datetime=%_my_datetime::=%
    set _my_datetime=%_my_datetime:/=_%
    set _my_datetime=%_my_datetime:.=_%3) Rename the file from dos
    ren some.txt dummy_file_%_my_datetime%.txt4) Here goes the output
    C:\dir
    dummy_file_Mon_09_20_2010_161347_21.txt
    Most of the code i copied from above url, you can tweak a little bit based on ur requirement and format.
    Regards
    Learner                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

Maybe you are looking for


HashFlare