J2ee1.4 and log file problem

Hello all,
I am running J2ee1.4 on Win 2000 Pro. I have 2 problems:
1) The server.log file becomes too big even I set its size. Seem like it does not rotate when the limit size is reached. (I set it for 2MB)
2) Under Logging tab, I enabled the option "Log messages to standard error", but when I restart the server It cannot start the server1 instance.
Can anyone please help me on this. I would like to see messages on standard output screen and I do not want the log file take all my hardrive space. Thank you so much!

Thank you for responding. When I enable the option "Log messages to standard error" and restart the j2ee1.4 server I got the following message:
Starting Domain domain1, please wait.
Log redirected to C:\Sun\AppServer\domains\domain1\logs\server.log
Subprocess timed out after 600000mS
CLI156 Could not start the domain domain1
Press any key to countinue...
When I look at processes under Window Task Manager, java.exe and javaw.exe is running. At the end, I cannot start it in "verbose" mode. Is there any other thing I need to do beside enable the "Log messages to standard error" ?
I am appreciated any help.

Similar Messages

  • Process Flow ignores name and location for Control- and Log-Files

    Hi!
    Our OWB Version is 10.1.0.3.0 - DB Version 9.2.0.7.0 - OWF Version 2.6.2
    Clients and server are running on Windows. Database contains target schemas as well as OWB Design and Runtime, plus OWF repositories. The source files to load reside on the same server as the database.
    I have for example a SQL*Loader Mapping MAP_TEXT which loads one flat file "text.dat" into a table stg_text.
    The mapping MAP_TEXT is well configured and runs perfect. i.e. Control file "text.ctl" is generated to location LOC_CTL, flat file "text.dat" read from another location LOC_DATA, bad file "text.bad" is written to LOC_BAD and the log file "text.log" is placed into LOC_LOG. All locations are registered in runtime repository.
    When I integrate this mapping into a WorkFlow Process PF_TEXT, then only LOC_DATA and LOC_BAD are used. After deploying PF_TEXT, I execute it and found out, that Control and Log file are placed into the directory <OWB_HOME>\owb\temp and got generic names <Mapping Name>.ctl and <Mapping Name>.log (in this case MAP_TEXT.ctl and MAP_TEXT.log).
    How can I influence OWB to execute the Process Flow using the locations configured for the mapping placed inside?
    Has anyone any helpfull idea?
    Thx,
    Johann.

    I didn't expect to be the only one to encounter this misbehaviour of OWB.
    Meanwhile I found out what the problem is and had to recognize that it is like it is!
    There is no solution for it till Paris Release.
    Bug Nr. 3099551 at Oracle MetaLink adresses this issue.
    Regards,
    Johann Lodina.

  • Steps to move Data and Log file for clustered SQL Server

    Hi guys 
    we have Active'passive SQL 2008R2 cluster environment.
    looking for steps to move Data and log files from user Database  and System Database for  SQL Server Clustered Instance. 
    Currently Data and log  files resides on same drive for user and system Databases..
    Thanks
    Please Mark As Answer if it is helpful. \\Aim To Inspire Rather to Teach A.Shah

    Try the below link
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/468de435-3432-45c2-a50b-23519cd2686e/moving-the-system-databases-in-a-sql-cluster?forum=sqldisasterrecovery
    -Prashanth

  • Shell Script to grep Job File name and Log File name from crontab -l

    Hello,
    I am new to shell scripting. I need to write a shell script where i can grep the name of file ie. .sh file and log file from crontab -l.
    #51 18 * * * /home/oracle/refresh/refresh_ug634.sh > /home/oracle/refresh/refresh_ug634.sh.log 2>&1
    #40 17 * * * /home/oracle/refresh/refresh_ux634.sh > /home/oracle/refresh/refresh_ux634.sh.log 2>&1
    In crontab -l, there are many jobs, i need to grep job name like 'refresh_ug634.sh' and corresponding log name like 'refresh_ug634.sh.log '.
    I am thinking of making a universal script that can grep job name, log name and hostname for one server.
    Then, suppose i modify the refresh_ug634.sh script and call that universal script and echo those values when the script gets executed.
    Please can anyone help.
    All i need to do is have footer in all the scripts running in crontab in one server.
    job file name
    log file name
    hostname
    Please suggest if any better solution. Thanks.

    957704 wrote:
    I need help how to grep that information from crontab -l
    Please can you provide some insight how to grep that shell script name from list of crontab -l jobs
    crontab -l > cron.log -- exporting the contents to a file
    cat cron.log|grep something -- need some commands to grep that infoYou are missing the point. This forum is for discussion of SQL and PL/SQL questions. What does your question have to do with SQL or PL/SQL.
    It's like you just walked into a hardware store and asked where they keep the fresh produce.
    I will point out one thing about your question. You are assuming every entry in the crontab has exactly the same format. consider this crontab:
    #=========================================================================
    # NOTE:  If this is on a clustered environment, all changes to this crontab
    #         must be replicated on all other nodes of the cluster!
    # minute        (0 thru 59)
    # hour          (0 thru 23)
    # day-of-month  (1 thru 31)
    # month         (1 thru 12)
    # weekday       (0 thru 6, sunday thru saturday)
    # command
    #=========================================================================
    00 01 1-2 * 1,3,5,7 /u01/scripts/myscript01  5 orcl  dev
    00 04 * * * /u01/scripts/myscript02 hr 365 >/u01/logs/myscript2.lis
    00 6 * * * /u01/scripts/myscript03  >/u01/logs/myscript3.lisThe variations are endless.
    When you get to an appropriate forum (this on is not it) it will be helpful to explain your business requirement, not just your proposed technical solution.

  • System Center 2012 R2 install: SQL server Data file and log file

    This might be a dumb question, but I can't find the answer anywhere.  
    I'm installing a new instance of  System Center 2012 R2 on a new server, I'm stuck on the SQL Server data file section.  Everytime I put in a path, it says that tne path does not exist.  I'm I supposed to be creating some sort of SQL Server
    data file and log file before this installation, I didn't get this prompt when installing System Center 2012 SP1 or hen I upgraded from System Center 2012 SP1 to System Center 2012 R2
    My SQL is on a different server
    Thank you in advanced

    Have you reviewed the setup.log?
    On a side note, why would you put the database file on the same drive as the OS? That defeats the whole purpose of having a remote SQL Server. Why use a remote SQL Server in the first place.
    Jason | http://blog.configmgrftw.com

  • Sql server data file and log file

    hello experts
    what is the best way to save data file and log file in a two node cluster environment. i have an active\passive cluster with windows server 2008r2 enterprise and sql server 2008r2. i am new to the environment and i noticed that all system and user databases
    including their data and log files are stored in one drive, just curious, what is the best practise in such kinds of scenario, thank you as always for your help.

    Make sure  you have valid/tested  backup strategy for both system and user databases.
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • How to design SQL server data file and log file growth

    how to design SQL DB data file and log file growth- SQL server 2012
    if my data file is having 10 GB sizze and log file is having 5 GB size
    what should be the size in MB (not in %) of autogrowth. based on what we have to determine the ideal size of file auto growth.

    It's very difficult to give a definitive answer on this. Best principal is to size your database correctly in advance so that you never have to autogrow, of course in reality that isn't always practical.
    The setting you use is really dictated by the expected growth in your files. Given that the size is relatively small why not set it to 1gb on the datafile(s) and 512mb on the log file? The important thing is to monitor it on an on-going basis to see if that's
    the appropriate amount.
    One thing you should do is enable instant file initialization by granting the service account Perform Volume Maintenance tasks in group policy. This will allow the data files to grow quickly when required, details here:
    https://technet.microsoft.com/en-us/library/ms175935%28v=sql.105%29.aspx?f=255&MSPPError=-2147217396
    Also, it possible to query the default trace to find autogrowth events, if you wanted you could write an alert/sql job based on this 
    SELECT
    [DatabaseName],
    [FileName],
    [SPID],
    [Duration],
    [StartTime],
    [EndTime],
    CASE [EventClass]
    WHEN 92 THEN 'Data'
    WHEN 93 THEN 'Log' END
    FROM sys.fn_trace_gettable('c:\path\to\trace.trc', DEFAULT)
    WHERE
    EventClass IN (92,93)
    hope that helps

  • Do we need to format data and log files with 64k cluster size for sql server 2012?

    Do we need to format data and log files with 64k cluster size for sql server 2012?
    Does this best practice still applies to sql server 2012 & 2014?

    Yes.  The extent size of SQL Server data files, and the max log block size have not changed with the new versions, so the guidance should remain the same.
    Microsoft SQL Server Storage Engine PM

  • Change the Data and Log file locations in livecache

    Hi
    We have installed livecache in unix systems in the /sapdb mount directory where the installer have created sapdata and sapdblog directories. But the unix team has already created two mount direcotries as follows:
    /sapdb/LC1/lvcdata and /sapdb/LC1/lvclog mount points.
    While installing livecache we had selected this locations for creating the DATA and LOG volumes. Now they are asking to move the DATA and LOG volumes created in sapdata and saplog directories to these mount points. How to move the data and log file and make the database consistent. Is there any procedure to move the files to the mount point directories and change the pointers of livecahce to these locations.
    regards
    bala

    Hi Lars
    Thanks for the link. I will try it and let u know.
    But this is livecache (even it uses MaxDB ) database which was created by
    sapinst and morover is there any thing to be adjusted in SCM and as well as
    any modification ot be done in db level.
    regards
    bala

  • 2005 database and log file locations

    Is there a SQL query to list where exactly the database and log files reside for all databases on an instance (sql server 2005)?

    You can query the Information form DMV
    sys.master_files (Transact-SQL)
    SELECT *
    FROM [sys].[master_files]
    Olaf Helper
    [ Blog] [ Xing] [ MVP]

  • Problem with File Handler and log files

    I'm developing a polymorphic structure for file handling specially for the log file for a server based App... Its working fine except for one thing. The log file that goes into the File Handler comes as a parameter to the class, the problem is that when it writes de file, though it DOES know where it should go, it doesn't do it and it writes the message into some other Log file belonging to another process...
    Does someone know how to avoid this or fix it?? any ideas or tips would be great!!

    Immediately below the Tabs saying "Files" and Assets" is a small box 
    with arrow on the right to show the drop down list.        In the box 
    on the right there's an icon of two networked computers.  Then it 
    says, "ftp://Hill farm Web Site"  which is the name of my website.     
    If I click on the arrows to pull up the drop-down box,  I get four 
    options divided by a line.   Above the line the options are Computer, 
    HD and ftp://Hill farm Web Site.  Below the line it says manage sites.
    Below this is list of files that make up my website in a directory 
    structure.   The header for the first column reads, "Local Files",  
    which appears to be untrue, because the top line in the directory 
    structure below reads,  "ftp://Hill farm Web Site".
    Does this help?
    regards
    David

  • DATE fields and LOG files  in context with external tables

    I am facing two problems when dealing with the external tables feature in Oracle 9i.
    I created an External Table with some fileds with the DATE data type . There were no issues during the creation part. But when i query the table, the DATE fields are not properly selected though the data is there in the files. Is there any ideas to deal with this ?
    My next question is regarding the log files. The contents in the log file seems to be growing when querying the external tables. Is there a way to control this behaviour?
    Suggestions / Advices on the above two issues are welcome.
    Thanks
    Lakshminarayanan

    Hi
    If you have date datatypes than:
    select
    greatest(TABCASER1.CASERRECIEVEDDATE, EVCASERS.FINALEVDATES, EVCASERS.PUBLICATIONDATE, EVCASERS.PUBLICATIONDATE, TABCASER.COMPAREACCEPDATE)
    from TABCASER, TABCASER1, EVCASERS
    where ...-- join and other conditions
    1. greatest is good enough
    2. to_date creates date dataype from string with the format of format string ('mm/dd/yyyy')
    3. decode(a, b, c, d) is a function: if a = b than return c else d. NULL means that there is no data in the cell of the table.
    6. to format the date for display use to_char function with format modell as in the to_date function.
    Ott Karesz
    http://www.trendo-kft.hu

  • Installing script and .log file

    Hello eveyone, I've the following scriopt that installs 2 packages et copies 2 files into a specific location.
    echo installation Firebird
    sudo installer -pkg /Autonome/FirebirdSS-1.5.5-4926-powerpc.pkg -target /
    echo "Copie des fichiers necessaires pour l'application (Conf et UDF)"
    cp /Autonome/FreeAdhocUDF /Library/Frameworks/Firebird.framework/Versions/A/Resources/English.lproj/var
    cp -f /Autonome/firebird.conf /Library/Frameworks/Firebird.framework/Versions/A/Resources/English.lproj/var
    echo installation Autonome
    sudo installer -pkg /Autonome/ClientTheriaqueAutonome.pkg -target /Applications/ClientTheriaqueAutonome
    echo suppression des fichiers temporaires
    rm -rf /Autonome
    rm /Applications/ClientTheriaqueAutonome/LiveUpdate/Test.xml
    It works fine for most users, but forme some others who have Mac OS 10.5.4 there's a problem when installing the ClientTheriaqueAutonome package (I'm not sure but after the investigation,I've made I found that it is not installed in its location).
    So I'm asking your help to tell me if :
    1. There's some specification in the /Applications in the 10.5.4 Mac OS version
    2. How I must to change my script to have a log file ( to know if a command exection was succes or fail)
    3. There's an other best alternative to do the installation as I've made in the script.
    Many thanks in advance for your help and collbaoration.
    Best reguads.

    there's a problem when installing the ClientTheriaqueAutonome package
    That's probably because you're saying:
    sudo installer -pkg /Autonome/ClientTheriaqueAutonome.pkg -target /Applications/ClientTheriaqueAutonome
    According to man installer, the -target parameter is the device you want to install on, not the directory. Therefore you should specify -target / and the installer package should contain all the necessary data to install the file in /Application/ClientTheriaqueAutonome/
    2. How I must to change my script to have a log file ( to know if a command exection was succes or fail).
    How are you calling the script?
    If you're calling it in the command line you can just append '> /path/to/your.log', e.g.:
    ./install > /path/to/your.log
    If not, you can append '>> /path/to/your.log' to each command:
    echo 'Installation Firebird' >> /path/to/your.log
    sudo installer -pkg blah blah blah >> /path/to/your.log
    ... etc.
    As for an AppleScript-based solution there are several ways, but what's not clear is how the files are getting setup in the first place. Your script defines the source files as being in /Autonome - how are they getting there?

  • Adobe document service issue attached he trace and log file

    hi friends
    adobe pdf is not working.
    i am  sending the trace file and server log  file
    1.5 #00145ED2ABF2006C00000087000EF0AC00045A8A76851A40#1225450647263#com.adobe.ads.operation.support
    .RenderErrorLog#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#487
    #SAP J2EE Engine JTA Transaction : [18363cffffffa200813]#n47a#DEMO1#772fd940a73611dda59700145ed2abf2#
    SAPEngine_Application_Thread[impl:3]_6##0#0#Error#1#/System/Server#Plain###Encountered a problem
    embedding Render collateral:
    General exception while merging XMLForm Module Render log data into log.pdf#
    #1.5 #00145ED2ABF2006C00000089000EF0AC00045A8A76852292#1225450647265#com.adobe.ads.request.Request#com.
    adobe/AdobeDocumentServices#com.adobe.ads.request.Request#ADSUSER#487#SAP J2EE Engine JTA Transaction :
    [18363cffffffa200813]#n47a#DEMO1#772fd940a73611dda59700145ed2abf2#SAPEngine_Application_Thread[impl:3]_6##0#0#
    Error#1#/System/Server#Plain###Client: 48316 AppName: com.crane.episode.EpisodeApp
    Unhandled exception detected during a "Render" operation.
    Request start time: Fri Oct 31 15:57:27 PKT 2008
    java.lang.NullPointerException
    Exception Stack Trace:
    java.lang.NullPointerException
         at com.adobe.ads.operation.Render.addErrorPdfToStreams(Unknown Source)
         at com.adobe.ads.operation.Render.handleRenderException(Unknown Source)
         at com.adobe.ads.operation.Render.execute(Unknown Source)
         at com.adobe.ads.operation.ADSOperation.doWork(Unknown Source)
         at com.adobe.ads.request.Request.processOperations(Unknown Source)
         at com.adobe.ads.request.Request.process(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.processRequest(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.rpData(Unknown Source)
         at com.adobe.AdobeDocumentServicesLocalLocalObjectImpl0_0.rpData(AdobeDocumentServicesLocalLocalObjectImpl0_0.java:120)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:85)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:58)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:60)
         at java.lang.reflect.Method.invoke(Method.java:391)
         at com.sap.engine.services.webservices.runtime.EJBImplementationContainer.invokeMethod(EJBImplementationContainer.java:126)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:157)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:79)
         at com.sap.engine.services.webservices.runtime.servlet.ServletDispatcherImpl.doPost(ServletDispatcherImpl.java:92)
         at SoapServlet.doPost(SoapServlet.java:51)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:401)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:387)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:365)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:944)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:266)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:95)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:175)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(AccessController.java:215)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)
    server0.log file
    #1.5 #00145ED2ABF20079000007260005206200045AEB49C2D359#1225866508161#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Warning#1#com.adobe.ads.operation.support.RenderErrorLog#Plain###Cannot generate Log.pdf#
    #1.5 #00145ED2ABF20079000007270005206200045AEB49C2D414#1225866508162#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Error#1#com.adobe.ads.operation.support.RenderErrorLog#Plain###Encountered a problem embedding Render collateral:
    General exception while merging XMLForm Module Render log data into log.pdf#
    #1.5 #00145ED2ABF20079000007290005206200045AEB49C2D982#1225866508163#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.request.Request#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Error#1#com.adobe.ads.request.Request#Plain###Client: eb582 AppName: samplesearchlab.chdp.com.SampleSearchLabApp
    Unhandled exception detected during a "Render" operation.
    Request start time: Wed Nov 05 11:28:27 PKT 2008
    java.lang.NullPointerException
    Exception Stack Trace:
    java.lang.NullPointerException
         at com.adobe.ads.operation.Render.addErrorPdfToStreams(Unknown Source)
         at com.adobe.ads.operation.Render.handleRenderException(Unknown Source)
         at com.adobe.ads.operation.Render.execute(Unknown Source)
         at com.adobe.ads.operation.ADSOperation.doWork(Unknown Source)
         at com.adobe.ads.request.Request.processOperations(Unknown Source)
         at com.adobe.ads.request.Request.process(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.processRequest(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.rpData(Unknown Source)
         at com.adobe.AdobeDocumentServicesLocalLocalObjectImpl0_0.rpData(AdobeDocumentServicesLocalLocalObjectImpl0_0.java:120)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:85)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:58)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:60)
         at java.lang.reflect.Method.invoke(Method.java:391)
         at com.sap.engine.services.webservices.runtime.EJBImplementationContainer.invokeMethod(EJBImplementationContainer.java:126)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:157)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:79)
         at com.sap.engine.services.webservices.runtime.servlet.ServletDispatcherImpl.doPost(ServletDispatcherImpl.java:92)
         at SoapServlet.doPost(SoapServlet.java:51)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:401)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:387)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:365)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:944)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:266)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:95)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:175)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(AccessController.java:215)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)
    regards
    raja

    Hello,
    Were you able to resolve this?  We are experiencing the same problem and have had no luck finding a solution.
    Thanks,
    Anthony

  • OSB logging - Process writes logs in both osb_server1.out and log file ..?

    Hello,
    I have few OSB proxy service and we have configured few log operations for logging but while testing, I noticed logs are getting written in both osb_server1.out and osb_server1.log file. I dont want to write logs in osb_server1.out file.
    I am running my weblogic server in development mode.
    Could someone please advice me whats wrong I am doing here.

    Hello,
    I have few OSB proxy service and we have configured few log operations for logging but while testing, I noticed logs are getting written in both osb_server1.out and osb_server1.log file. I dont want to write logs in osb_server1.out file.
    I am running my weblogic server in development mode.
    Could someone please advice me whats wrong I am doing here.

Maybe you are looking for