Write log file on error

With alert configuration we can send notify user when there is an error.
Is there a way to write a log file when there is an error (assume no BPM used).

There are a couple of things you can do to create a log file:
1. You can create a custom alert and trigger it by generating an exception in your message mapping. This can be done using a user defined function. In the same user defined function you can call a communication channel to write the error log file.
2. You can create 2 interfaces the first one is your main interface, in that interface donot error out the interface in the message mapping. Populate a certain error value in the target and use that value in another message map to create another map with the error
IntfA -
> UDF (Error --> X) ---> Target (<field Z>X</field>). --> IntfB
IntfB -
> Check value for X in field Z to populate the eror records to the target structure --> intFC.
Same condition u need to put in the rcvr determination so that it doesnot give an error.

Similar Messages

  • History report error: | An Exceptional Error occurred. Application exiting. Check the log file for error 5022

    Hi all
    I've got a error msg when try to generate a report using Cisco history report tool:
    Error | An Exceptional Error occurred. Application exiting.  Check the log file for error 5022
    It only happens when choose report template: ICD_Contact_Service_Queue_Activity_by_CSQ_en_us.
    User tried samething on other PC, it working fine.
    only on user' own PC and only choose this report, error appears.
    user runing windows 7 and do not have crystal report installed
    tried reinstalled the software, doesn't work.
    also tried this: (https://cisco-support.hosted.jivesoftware.com/thread/2041254) - doesn't work
    then tried https://supportforums.cisco.com/docs/DOC-6209  - doesn't work
    attached the log file.
    thanks.

    wenqianyu wrote:From the log file:Looks like you get a Login Window.Error message showed up after username/password be enteredThere is an error in the log: Error happened in comparing UCCX version and HRC versionYou may need to do a clean uninstall, download the Historical report from the server, and install it again on the PC.Does this only happen to one PC or to every PC with this application?Wenqian 
    I have completely uninstalled the HRC, and download from server install again -- still doesn't work with exactly same error.
    this matter only happens on this PC, when user try same thing on other PC, it works.
    so i think it not relate to server or account.

  • (Cisco Historical Reporting / HRC ) All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054

    Hi All,
    I am getting an error message "All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054"  when trying to log into HRC (This user has the reporting capabilities) . I checked the log files this is what i found out 
    The log file stated that there were ongoing connections of HRC with the CCX  (I am sure there isn't any active login to HRC)
    || When you tried to login the following error was being displayed because the maximum number of connections were reached for the server .  We can see that a total number of 5 connections have been configured . ||
    1: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Current number of connections (5) from historical Clients/Scheduler to 'CRA_DATABASE' database exceeded the maximum number of possible connections (5).Check with your administrator about changing this limit on server (wfengine.properties), however this might impact server performance.
    || Below we can see all 5 connections being used up . ||
    2: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:[DB Connections From Clients (count=5)]|[(#1) 'username'='uccxhrc','hostname'='3SK5FS1.ucsfmedicalcenter.org']|[(#2) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#3) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#4) 'username'='uccxhrc','hostname'='PFS-HHXDGX1.ucsfmedicalcenter.org']|[(#5) 'username'='uccxhrc','hostname'='47BMMM1.ucsfmedicalcenter.org']
    || Once the maximum number of connection was reached it threw an error . ||
    3: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Number of max connection to 'CRA_DATABASE' database was reached! Connection could not be established.
    4: 6/20/2014 9:13:49 AM %CHC-LOG_SUBFAC-3-UNK:Database connection to 'CRA_DATABASE' failed due to (All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054.)
    Current exact UCCX Version 9.0.2.11001-24
    Current CUCM Version 8.6.2.23900-10
    Business impact  Not Critical
    Exact error message  All available connections to database server are in use by other client machines. Please try again later and check the log file for error 5054
    What is the OS version of the PC you are running  and is it physical machine or virtual machine that is running the HRC client ..
    OS Version Windows 7 Home Premium  64 bit and it’s a physical machine.
    . The Max DB Connections for Report Client Sessions is set to 5 for each servers (There are two servers). The no of HR Sessions is set to 10.
    I wanted to know if there is a way to find the HRC sessions active now and terminate the one or more or all of that sessions from the server end ? 

    We have had this "PRX5" problem with Exchange 2013 since the RTM version.  We recently applied CU3, and it did not correct the problem.  We have seen this problem on every Exchange 2013 we manage.  They are all installations where all roles
    are installed on the same Windows server, and in our case, they are all Windows virtual machines using Windows 2012 Hyper-V.
    We have tried all the "this fixed it for me" solutions regarding DNS, network cards, host file entries and so forth.  None of those "solutions" made any difference whatsoever.  The occurrence of the temporary error PRX5 seems totally random. 
    About 2 out of 20 incoming mail test by Microsoft Connectivity Analyzer fail with this PRX5 error.
    Most people don't ever notice the issue because remote mail servers retry the connection later.  However, telephone voice mail systems that forward voice message files to email, or other such applications such as your scanner, often don't retry and
    simply fail.  Our phone system actually disables all further attempts to send voice mail to a particular user if the PRX5 error is returned when the email is sent by the phone system.
    Is Microsoft totally oblivious to this problem?
    PRX5 is a serious issue that needs an Exchange team resolution, or at least an acknowledgement that the problem actually does exist and has negative consequences for proper mail flow.
    JSB

  • How to manually write log file when tranform xslt by using Transformer?

    I want to ask experts that are there any way to write the information(such as template name or number of template used) into log file while performing transformation with javax.xml.transform.Transformer.
    Below is my sample code
    import javax.xml.transform.Result;
    import javax.xml.transform.Source;
    import javax.xml.transform.Transformer;
    // declare and assign value of transform, source, and result
    transformer.transform(source, result);
    Thanks for advance

    I think it will be from FDM, if I remember correctly FDM will generate a text file in the background and then load data into essbase using the text file.
    The codes at each line are standard essbase generated codes which relate to an operation e.g. 1013162 = Received Command [Calculate] from user [%s] using [%s]
    If you have a search on the web you will be able to find a full list of codes from numerous locations.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Wait Events "log file parallel write" / "log file sync" during CREATE INDEX

    Hello guys,
    at my current project i am performing some performance tests for oracle data guard. The question is "How does a LGWR SYNC transfer influences the system performance?"
    To get some performance values, that i can compare i just built up a normal oracle database in the first step.
    Now i am performing different tests like creating "large" indexes, massive parallel inserts/commits, etc. to get the bench mark.
    My database is an oracle 10.2.0.4 with multiplexed redo log files on AIX.
    I am creating an index on a "normal" table .. i execute "dbms_workload_repository.create_snapshot()" before and after the CREATE INDEX to get an equivalent timeframe for the AWR report.
    After the index is built up (round about 9 GB) i perform an awrrpt.sql to get the AWR report.
    And now take a look at these values from the AWR
                                                                       Avg
                                                 %Time  Total Wait    wait     Waits
    Event                                 Waits  -outs    Time (s)    (ms)      /txn
    log file parallel write              10,019     .0         132      13      33.5
    log file sync                           293     .7           4      15       1.0
    ......How can this be possible?
    Regarding to the documentation
    -> log file sync: http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/waitevents003.htm#sthref3120
    Wait Time: The wait time includes the writing of the log buffer and the post.-> log file parallel write: http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/waitevents003.htm#sthref3104
    Wait Time: Time it takes for the I/Os to complete. Even though redo records are written in parallel, the parallel write is not complete until the last I/O is on disk.This was also my understanding .. the "log file sync" wait time should be higher than the "log file parallel write" wait time, because of it includes the I/O and the response time to the user session.
    I could accept it, if the values are close to each other (maybe round about 1 second in total) .. but the different between 132 seconds and 4 seconds is too noticeable.
    Is the behavior of the log file sync/write different when performing a DDL like CREATE INDEX (maybe async .. like you can influence it with the initialization parameter COMMIT_WRITE??)?
    Do you have any idea how these values come about?
    Any thoughts/ideas are welcome.
    Thanks and Regards

    Surachart Opun (HunterX) wrote:
    Thank you for Nice Idea.
    In this case, How can we reduce "log file parallel write" and "log file sync" waited time?
    CREATE INDEX with NOLOGGINGA NOLOGGING can help, can't it?Yes - if you create index nologging then you wouldn't be generating that 10GB of redo log, so the waits would disappear.
    Two points on nologging, though:
    <ul>
    it's "only" an index, so you could always rebuild it in the event of media corruption, but if you had lots of indexes created nologging this might cause an unreasonable delay before the system was usable again - so you should decide on a fallback option, such as taking a new backup of the tablespace as soon as all the nologging operatons had completed.
    If the database, or that tablespace, is in +"force logging"+ mode, the nologging will not work.
    </ul>
    Don't get too alarmed by the waits, though. My guess is that the +"log file sync"+ waits are mostly from other sessions, and since there aren't many of them the other sessions are probably not seeing a performance issue. The +"log file parallel write"+ waits are caused by your create index, but they are happeninng to lgwr in the background which is running concurrently with your session - so your session is not (directly) affected by them, so may not be seeing a performance issue.
    The other sessions are seeing relatively high sync times because their log file syncs have to wait for one of the large writes that you have triggered to complete, and then the logwriter includes their (little) writes with your next (large) write.
    There may be a performance impact, though, from the pure volume of I/O. Apart from the I/O to write the index you have LGWR writting (N copies) of the redo for the index and ARCH is reading and writing the completed log files caused by the index build. So the 9GB of index could easily be responsible for vastly more I/O than the initial 9GB.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    To post code, statspack/AWR report, execution plans or trace files, start and end the section with the tag {noformat}{noformat} (lowercase, curly brackets, no spaces) so that the text appears in fixed format.
    "Science is more than a body of knowledge; it is a way of thinking"
    Carl Sagan                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • SQL Server not Starting Up - Could not open error log file system error = 5(Access is denied.)

    Hi guys,
     I am working on issue with SQL Server 2008 R2 .. we rebooted the server and after that all SQL services is not coming up. below is the error from eventlog...
    SQL Server running under NetworkService account and has full control on this files..
    from Config Manager ....
    The request failed or the service did not respond in a timely fashion....
    From Event log..
    initerrlog: Could not open error log file 'C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Log\ERRORLOG'. Operating system error = 5(Access is denied.).
    Any idea or workaround..?
    Thanks..
    Please Mark As Answer if it is helpful. \\Aim To Inspire Rather to Teach

    Finally I found solution and was able to bring back online the SQL Services...
    I had a very similar issue with the following error (event log):
    Description:
    The SQL Server (MSSQLSERVER) service terminated with service-specific error 17058 (0x42A2).
    For more information, see Help and Support Center at http://go.microsoft.com/fwlink/events.asp.
    Event Type:      Error
    Event Source:      MSSQLSERVER
    Event Category:      (2)
    Event ID:      17058
    Date:            19/05/2012
    Time:            10:20:14
    User:            N/A
    Computer:      NSQLXX
    Description:
    initerrlog: Could not open error log file 'C:\Program Files\Microsoft SQL Server\MSSQL.1\MSSQL\LOG\ERRORLOG'.
    Operating system error = 5(error not found).
    In this case the ERRORLOG file had been marked as read only. This was following a scheduled automated boot.  -This
    was culprit
    MSSQL wouldn't not start until this file had been changed.
    http://www.experts-exchange.com/Database/MS-SQL-Server/Q_23951723.html
    Thanks everyone for your kind replies and time...
    Please Mark As Answer if it is helpful. \\Aim To Inspire Rather to Teach

  • How can I write log file???

    Hi every body.
    I have a project built by ADF BC. I want that when user performs INSERT, UPDATE, DELETE,... into DB from client tier in some form, before performing those queries I will get the user information (who login the system) and write out a log file about that user from that information in each entity influenced by the user.
    How can I do this???

    I understand what you mean from your link. But the question I meant is not that. What I meant is that, I can get the user information (user_name,password,...) who performs the queries to DB and write those information to log file.
    Thanks.

  • Setup log file Critical errors

    if i start the setup it says Critical Errors Found In Setup Please See The Setup Log.File For Details

    You need to provide the logs and tell us more about your system. Please refer to this page for info on where to find the logs (in the CS5 section).#
    Mylenium

  • Trans.log file showing  'ERROR = CONNECT failed with sql error 12541'

    Dear All
    We have upgraded our datatabse from Oracle 9.2.0.8 to Oracel 10.2.0.4 on our ECC 5.0 IS-AFS SAP system. When we check version from SQL, it's showing 10.2.0.4. That means database is successfully upgraded. Now, when we login through our <SID>adm user and checking R3trans -d, it is showing
    'This is R3trans version 6.13 (release 640 - 07.01.08 - 14:25:00).
    unicode enabled version
    2EETW169 no connect possible: "DBMS = ORACLE                           --- dbs_ora_tnsname = 'V
    AP'"
    R3trans finished (0012).'
    In the trans.log file it is showing,
    4 ETW000 R3trans version 6.13 (release 640 - 07.01.08 - 14:25:00).
    4 ETW000 unicode enabled version
    4 ETW000 ===============================================
    4 ETW000
    4 ETW000 date&time   : 22.06.2009 - 11:56:40
    4 ETW000 control file: <no ctrlfile>
    4 ETW000 R3trans was called as follows: R3trans -d
    4 ETW000  trace at level 2 opened for a given file pointer
    4 ETW000  [dev trc     ,00000]  Mon Jun 22 11:56:40 2009                             110  0.000
    110
    4 ETW000  [dev trc     ,00000]  db_con_init called                                    27  0.000
    137
    4 ETW000  [dev trc     ,00000]  create_con (con_name=R/3)                             62  0.000
    199
    4 ETW000  [dev trc     ,00000]  Loading DB library '/usr/sap/VAP/SYS/exe/run/dboraslib.o' ...
    4 ETW000                                                                              77  0.000
    276
    4 ETW000  [dev trc     ,00000]  load shared library (/usr/sap/VAP/SYS/exe/run/dboraslib.o), hdl
    0
    4 ETW000                                                                            8142  0.008
    418
    4 ETW000  [dev trc     ,00000]  Library '/usr/sap/VAP/SYS/exe/run/dboraslib.o' loaded
    4 ETW000                                                                              43  0.008
    461
    4 ETW000  [dev trc     ,00000]  function DbSlExpFuns loaded from library /usr/sap/VAP/SYS/exe/r
    un/dboraslib.o
    4 ETW000                                                                              61  0.008
    522
    4 ETW000  [dev trc     ,00000]  Version of '/usr/sap/VAP/SYS/exe/run/dboraslib.o' is "640.00",
    patchlevel (0.276)
    4 ETW000                                                                             421  0.008
    943
    4 ETW000  [dev trc     ,00000]  function dsql_db_init loaded from library /usr/sap/VAP/SYS/exe/
    run/dboraslib.o
    4 ETW000                                                                              46  0.008
    989
    4 ETW000  [dev trc     ,00000]  function dbdd_exp_funs loaded from library /usr/sap/VAP/SYS/exe
    /run/dboraslib.o
    4 ETW000                                                                              75  0.009
    064
    4 ETW000  [dev trc     ,00000]  New connection 0 created                              48  0.009
    112
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = -000000001 state = DISCONNECTED, perm =
    YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              53  0.009
    165
    4 ETW000  [dev trc     ,00000]  db_con_connect (con_name=R/3)                         55  0.009
    220
    4 ETW000  [dev trc     ,00000]  find_con_by_name found the following connection for reuse:
    4 ETW000                                                                              53  0.009
    273
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = 000000000 state = DISCONNECTED, perm =
    YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              49  0.009
    322
    4 ETW000  [dev trc     ,00000]  Got ORACLE_HOME=/oracle/VAP/102_64 from environment
    4 ETW000                                                                             521  0.009
    843
    4 ETW000  [dev trc     ,00000]  -->oci_initialize (con_hdl=0)                         84  0.009
    927
    4 ETW000  [dev trc     ,00000]  got NLS_LANG='AMERICAN_AMERICA.UTF8' from environment
    4 ETW000                                                                              65  0.009
    992
    4 ETW000  [dev trc     ,00000]  Client NLS settings: AMERICAN_AMERICA.UTF8          2033  0.012
    025
    4 ETW000  [dev trc     ,00000]  Logon as OPS$-user to get SAPVAP's password           49  0.012
    074
    4 ETW000  [dev trc     ,00000]  Connecting as /@VAP on connection 0 (nls_hdl 0) ... (dbsl 640 1
    80309)
    4 ETW000                                                                              57  0.012
    131
    4 ETW000  [dev trc     ,00000]  Nls CharacterSet                 NationalCharSet              C
          EnvHp      ErrHp ErrHpBatch
    4 ETW000                                                                              49  0.012
    180
    4 ETW000  [dev trc     ,00000]    0 UTF8                                                      1
    0x1113a4400 0x1113b6b00 0x1113b63b8
    4 ETW000                                                                              58  0.012
    238
    4 ETW000  [dev trc     ,00000]  Allocating service context handle for con_hdl=0       48  0.012
    286
    4 ETW000  [dev trc     ,00000]  Allocating server context handle                      42  0.012
    328
    4 ETW000  [dev trc     ,00000]  Attaching to DB Server VAP (con_hdl=0,svchp=0x1113b97d8,srvhp=0
    x1113b99f8)
    4 ETW000                                                                              95  0.012
    423
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCIServerAttach' failed: rc = 12541
    4 ETW000                                                                            4556  0.016
    979
    4 ETW000  [dbsloci.    ,00000]  *** ERROR => CONNECT failed with sql error 12541.
    4 ETW000                                                                              59  0.017
    038
    4 ETW000  [dev trc     ,00000]     set_ocica() -> OCI or SQL return code 12541        39  0.017
    077
    4 ETW000  [dev trc     ,00000]  Try to connect with default password                 184  0.017
    261
    4 ETW000  [dev trc     ,00000]  Connecting as SAPVAP/<pwd>@VAP on connection 0 (nls_hdl 0) ...
    (dbsl 640 180309)
    4 ETW000                                                                              46  0.017
    307
    4 ETW000  [dev trc     ,00000]  Nls CharacterSet                 NationalCharSet              C
          EnvHp      ErrHp ErrHpBatch
    4 ETW000                                                                              48  0.017
    355
    4 ETW000  [dev trc     ,00000]    0 UTF8                                                      1
    0x1113a4400 0x1113b6b00 0x1113b63b8
    4 ETW000                                                                              48  0.017
    403
    4 ETW000  [dev trc     ,00000]  server_detach(con_hdl=0,stale=0,svrhp=0x1113b99f8)
    4 ETW000                                                                              54  0.017
    457
    4 ETW000  [dev trc     ,00000]  Detaching from DB Server (con_hdl=0,svchp=0x1113b97d8,srvhp=0x1
    113b99f8)
    4 ETW000                                                                              54  0.017
    511
    4 ETW000  [dev trc     ,00000]  Deallocating server context handle 0x1113b99f8        42  0.017
    553
    4 ETW000  [dev trc     ,00000]  Allocating server context handle                      33  0.017
    586
    4 ETW000  [dev trc     ,00000]  Attaching to DB Server VAP (con_hdl=0,svchp=0x1113b97d8,srvhp=0
    x1113b99f8)
    4 ETW000                                                                              54  0.017
    640
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCIServerAttach' failed: rc = 12541
    4 ETW000                                                                            2547  0.020
    187
    4 ETW000  [dbsloci.    ,00000]  *** ERROR => CONNECT failed with sql error 12541.
    4 ETW000                                                                              45  0.020
    232
    4 ETW000  [dev trc     ,00000]     set_ocica() -> OCI or SQL return code 12541        28  0.020
    260
    4 ETW000  [dblink      ,00428]  ***LOG BY2=>sql error 12541  performing CON [dblink#3 @ 428]
    4 ETW000                                                                             220  0.020
    480
    4 ETW000  [dblink      ,00428]  ***LOG BY0=>ORA-12541: TNS:no listener [dblink#3 @ 428]
    4 ETW000                                                                              42  0.020
    522
    2EETW169 no connect possible: "DBMS = ORACLE                           --- dbs_ora_tnsname = 'V
    AP'"
    Checked the listener service. It is already started and status is OK. Please help.
    Regards
    Shashank Shekhar

    Hi
    Yes, I am trying it with orasid user also. There in trans.log file it is showing:
    '4 ETW000 R3trans version 6.13 (release 640 - 07.01.08 - 14:25:00).
    4 ETW000 unicode enabled version
    4 ETW000 ===============================================
    4 ETW000
    4 ETW000 date&time   : 22.06.2009 - 15:54:42
    4 ETW000 control file: <no ctrlfile>
    4 ETW000 R3trans was called as follows: R3trans -d
    4 ETW000  trace at level 2 opened for a given file pointer
    4 ETW000  [dev trc     ,00000]  Mon Jun 22 15:54:42 2009                             109  0.000
    109
    4 ETW000  [dev trc     ,00000]  db_con_init called                                    27  0.000
    136
    4 ETW000  [dev trc     ,00000]  create_con (con_name=R/3)                             61  0.000
    197
    4 ETW000  [dev trc     ,00000]  Loading DB library '/usr/sap/VAP/SYS/exe/run/dboraslib.o' ...
    4 ETW000                                                                              79  0.000
    276
    4 ETW000  [dev trc     ,00000]  load shared library (/usr/sap/VAP/SYS/exe/run/dboraslib.o), hdl
    0
    4 ETW000                                                                            8174  0.008
    450
    4 ETW000  [dev trc     ,00000]  Library '/usr/sap/VAP/SYS/exe/run/dboraslib.o' loaded
    4 ETW000                                                                              43  0.008
    493
    4 ETW000  [dev trc     ,00000]  function DbSlExpFuns loaded from library /usr/sap/VAP/SYS/exe/r
    un/dboraslib.o
    4 ETW000                                                                              61  0.008
    554
    4 ETW000  [dev trc     ,00000]  Version of '/usr/sap/VAP/SYS/exe/run/dboraslib.o' is "640.00",
    patchlevel (0.276)
    4 ETW000                                                                             430  0.008
    984
    4 ETW000  [dev trc     ,00000]  function dsql_db_init loaded from library /usr/sap/VAP/SYS/exe/
    run/dboraslib.o
    4 ETW000                                                                              46  0.009
    030
    4 ETW000  [dev trc     ,00000]  function dbdd_exp_funs loaded from library /usr/sap/VAP/SYS/exe
    /run/dboraslib.o
    4 ETW000                                                                              76  0.009
    106
    4 ETW000  [dev trc     ,00000]  New connection 0 created                              48  0.009
    154
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = -000000001 state = DISCONNECTED, perm =
    YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              53  0.009
    207
    4 ETW000  [dev trc     ,00000]  db_con_connect (con_name=R/3)                         57  0.009
    264
    4 ETW000  [dev trc     ,00000]  find_con_by_name found the following connection for reuse:
    4 ETW000                                                                              54  0.009
    318
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = 000000000 state = DISCONNECTED, perm =
    YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              49  0.009
    367
    4 ETW000  [dev trc     ,00000]  Got ORACLE_HOME=/oracle/VAP/102_64 from environment
    4 ETW000                                                                             526  0.009
    893
    4 ETW000  [dev trc     ,00000]  -->oci_initialize (con_hdl=0)                         84  0.009
    977
    4 ETW000  [dev trc     ,00000]  got NLS_LANG='AMERICAN_AMERICA.UTF8' from environment
    4 ETW000                                                                              64  0.010
    041
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCIEnvCreate(mode=16384)' failed: rc = -
    1
    4 ETW000                                                                            1464  0.011
    505
    4 ETW000  [dev trc     ,00000]     set_ocica() -> OCI or SQL return code -1           40  0.011
    545
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCIErrorGet' failed: rc = -2
    4 ETW000                                                                              58  0.011
    603
    4 ETW000  [dblink      ,00428]  ***LOG BY2=>sql error -1     performing CON [dblink#3 @ 428]
    4 ETW000                                                                              74  0.011
    677
    4 ETW000  [dblink      ,00428]  ***LOG BY0=>Cannot get Oracle error text. [dblink#3 @ 428]
    4 ETW000                                                                              55  0.011
    732
    2EETW169 no connect possible: "DBMS = ORACLE                           --- dbs_ora_tnsname = 'V
    AP'"
    Someone suggested 'The solution is to download the new oracle client dll into the kernel'. Can anyone tell from where i can get the client dll for the kernel.
    Regards
    Shashank Shekhar

  • Error Details Missing in log files for errors on U.I.

    I am working with Weblogic Portal Server 9.2 on windows XP system with oracle 10g database. I have kept separate schema for appGroupspace data-source, and all other tables in main.
    All my collaboration portlets are working but all groupspace portlets are showing error "An error occured while processing your request. Please try again or see your System Administrator".
    All that I can fing in error logs in message "Cant find the primary instance for the portlet [{0}]. The portlet will be served in file mode".
    Is there some way I can find the stack trace of these errors?
    Any clues can be helpful as I'm stuck..

    Hi, I followed all the steps correctly and also the owner of the directory is correct.
    But i have not installed the patches which is listed at the bottom of the link page.
    I had done this on RHEl and i had no problem but now i am getting problem on solaris.

  • Node.js loss of permission to write/create log files

    We have been operating Node.js as a worker role cloud service. To track server activity, we write log files (via log4js) to C:\logs
    Originally the logging was configured with size-based roll-over. e.g. new file every 20MB. I noticed on some servers the sequencing was uneven
    socket.log <-- current active file
    socket.log.1
    socket.log.3
    socket.log.5
    socket.log.7
    it should be
    socket.log.1
    socket.log.2
    socket.log.3
    socket.log.4
    Whenever there is uneven sequence, i realise the beginning of each file revealed the Node process was restarted. From Windows Azure event log, it further indicated worker role hosting mechanism found node.exe to have terminated abruptly.
    With no other information to clue what is exactly happening, I thought there was some fault with log4js roll over implementation (updating to latest versions did not help). Subsequently switched to date-based roll-over mode; saw that roll-over happened every
    midnight and was happy with it.
    However some weeks later I realise the roll-over was (not always, but pretty predictably) only happening every alternate midnight.
    socket.log-2014-06-05
    socket.log-2014-06-07
    socket.log-2014-06-09
    And each file again revealed that midnight the roll-over did not happen, node.exe was crashing again. Additional logging on uncaughtException and exit happens showed nothing; which seems to suggest node.exe was killed by external influence (e.g. process
    kill) but it was unfathomable anything in the OS would want to kill node.exe.
    Additionally, having two instances in the cloud service, we observe the crashing of both node.exe within minutes of each other. Always. However if we had two server instances brought up on different days, then the "schedule" for crashing would
    be offset by the difference of the instance launch dates.
    Unable to trap more details what's going on, we tried a different logging library - winston. winston has the additional feature of logging uncaughtExceptions so it was not necessary to manually log that. Since winston does not have date-based roll-over it
    went back to size-based roll-over; which obviously meant no more midnight crash. 
    Eventually, I spotted some random midday crash today. It did not coincide with size-based rollover event, but winston was able to log an interesting uncaughtException.
    "date": "Wed Jun 18 2014 06:26:12 GMT+0000 (Coordinated Universal Time)",
    "process": {
    "pid": 476,
    "uid": null,
    "gid": null,
    "cwd": "E:
    approot",
    "execPath": "E:\\approot
    node.exe",
    "version": "v0.8.26",
    "argv": ["E:\\approot\\node.exe", "E:\\approot\\server.js"],
    "memoryUsage":
    { "rss": 80433152, "heapTotal": 37682920, "heapUsed": 31468888 }
    "os":
    { "loadavg": [0, 0, 0], "uptime": 163780.9854492 }
    "trace": [],
    "stack": ["Error: EPERM, open 'c:\\logs\\socket1.log'"],
    "level": "error",
    "message": "uncaughtException: EPERM, open 'c:\\logs\\socket1.log'",
    "timestamp": "2014-06-18T06:26:12.572Z"
    Interesting question: the Node process _was_ writing to socket1.log all along; why would there be a sudden EPERM error?
    On restart it could resume writing to the same log file. Or in previous cases it would seem like the lack of permission to create a new log file. 
    Any clues on what could possibly cause this? On a "scheduled" basis per server? Given that it happens so frequently and in sync with sister instances in the cloud service, something is happening in the back scenes which I cannot put a finger to.
    thanks
    The melody of logic will always play out the truth. ~ Narumi Ayumu, Spiral

    Hi,
    It is  strange. From your description, how many instances of your worker role? Do you store the log file on your VM local disk? To avoid this question, the best choice is you could store your log file into azure storage blob . If you do this, all log
    file will be stored on blob storage. About how to use azure blob storage, please see this docs:
    http://azure.microsoft.com/en-us/documentation/articles/storage-introduction/
    Please try it.
    If I misunderstood, please let me know.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Error:41453 in log file even when the account data is imported successfully

    Hi,
    As a beginner, I tried importing ‘Accounts’ record type with 2 records, by taking below 4 fields:
    Account Name, Account Type, City, Owner
    The email received for the account import request completion has below details:
    Total Records: 2
    Successfully Imported: 0
    Partially Imported: 0
    Duplicate Records Ignored: 0
    Failed: 2
    Against both the records, error : 41453 is mentioned in the log file, with error description as - An unexpected error occurred during the import of the following row: 'Account Name: <name of account>'
    However when I query for the records under Account tab, the records are all created properly (with all the values correctly present). I am not sure why am I getting this error if the records are successfully imported? Any help on this would be appreciated.
    In import wizard below selection was made-
    Select the method Import should use to uniquely identify matching records. You can use either the External Unique ID or On Demand Row ID or the following On Demand predefined fields: Account Name, Location
    * External Unique ID
    If the unique record identifier of the record being imported matches a record already in On Demand:
    * Don't Import Duplicate Records
    If the unique record identifier of the record being imported does not match a record already in On Demand:
    * Create New Record
    The import process may encounter field values that do not match the picklist values for fields in Oracle CRM On Demand. Please decide how to resolve these conflicts:
    * Don't Import Field Value
    This option allows you to decide if import should create a new record for missing associations (related records) in your data file.
    For the current import, this option will apply to the following associated record fields: Parent Account
    * Create Associated Record
    Thanks,
    Gunjan

    *'Create Associated Record’ & ‘Don't Create Associated Record’*
    If you are inserting Contacts. Account for that Contact is an Associated Record.
    And if you gave 'Create Associated Record’ it will create Accounts for that Contact if it didn't find a matching Account and insert the Contact.
    If you gave ‘Don't Create Associated Record’ it will not create Account and the Contact.
    Its better to give ‘Don't Create Associated Record’ since the Account you gave in contact is case sensitive.
    So if u give 'Create Associated Record’ then while inserting contact if the account is not exactly matching even though it exist in some other case(lower/upper)
    it wont map to that account which you meant and will create a new Account and map it to that contact.
    Hope you got some idea.
    Regards,
    Lemu

  • Check Event Alert failed with error - No errors in the log file.

    Hi All,
    I am developing a simple event based alert on PO_HEADERS table. I want to send alerts when a PO is created.
    I did all the steps according to the metalink note How To Send An Email In A Simple Periodic Or Event Alert? [ID 1162153.1]
    When i create the PO, the alert is triggering, and Check Event Alert concurrent program is running. But the program completes with error.
    Checking the output file (empty) log file (no errors)
    What can i do here to find out what is the problem? There is nothing in the Alert Manager - History form also. I have kept 7 days as days to keep.
    Thanks!
    M

    Can you find any details about the error from the "View Detail" button (the same window where you check the log and output files)?
    I found the Workflow logs, I am not sure what I am looking for, but i am not seeing any errors reported.The event viewer is supposed to send an email, so do you see anything in the logs that could be related?
    Thanks,
    Hussein

  • SMTP Error in concurrent program log file

    Hi All,
    Please help me on this asap.
    I need to send an excel file as attachment in a mail through a concurrent request submission. But after run the program I am getting the following error in log file.
    'Error in Send Mail: ORA-29279: SMTP permanent error: 501 5.5.4 Invalid Address' for any email id.
    Please let me know how can I resolve this issue. Appreciate your help.
    Thanks,
    JP

    Thanks for the reply.
    My OS is windows, EBS is 11.5.10.2. It is a custom concurrent program. We have SMTP server and this the first time I am working on mail sending. I am using the same code existed in my production env which sends mail for different purpose. I did not identified where the issue casues.
    IF fexists THEN
    fnd_file.put_line (FND_FILE.LOG,'File Exist');
    if file_length > 1 then
    fnd_file.put_line (FND_FILE.LOG,'File Size: '||file_length);
    fnd_file.put_line (FND_FILE.LOG,'Block Size: '||block_size);
    l_mail_error:= SendMailJPkg.SENDMAIL(
    SMTPServerName => 'mocrsmtp',
    Sender => '[email protected]',
    Recipient => [email protected],
    CcRecipient => '',
    BccRecipient => '',
    Subject => 'testmail',
    Body => 'Dear Super User, '
    ||chr(13)
    ||chr(13)
    ||'Testing mail sending'
    ||chr(13)
    ||chr(13)
    ||'Thank you for your cooperation.',
    ErrorMessage => l_mail_error,
    Attachments => SendMailJPkg.ATTACHMENTS_LIST(v_file_path, NULL));
    ELSE
    fnd_file.put_line (FND_FILE.LOG,'File Size Else: '||file_length);
    fnd_file.put_line (FND_FILE.LOG,'Block Size Else: '||block_size);
    ots_send_mail_pkg.send
    ( p_sender_email => '[email protected]',
    p_from => '[email protected]',
    p_to => ots_send_mail_pkg.array('[email protected]'),
    -- p_cc => ots_send_mail_pkg.array('[email protected]'),
    p_bcc => ots_send_mail_pkg.array(''),
    p_subject => v_subject_error,
    p_body => v_body_error||v_file_name||'PDF' );
    END IF;
    ELSE
    fnd_file.put_line (FND_FILE.LOG,'File Does Not Exist 1');
    ots_send_mail_pkg.send
    ( p_sender_email => '[email protected]',
    p_from => '[email protected]',
    p_to => ots_send_mail_pkg.array('[email protected]'),
    -- p_cc => ots_send_mail_pkg.array('[email protected]'),
    p_bcc => ots_send_mail_pkg.array(''),
    p_subject => v_subject_error_c,
    p_body => v_body_error_c||v_file_name||'PDF' );
    END IF;
    ELSIF v_file_type = 'EXCEL' then
    v_file_path := 'e:\oracle\uatcomn\admin\out\UAT_a7erpdb\'||v_file_name||'EXCEL';
    utl_file.fgetattr('e:\oracle\uatcomn\admin\out\UAT_a7erpdb',
    v_file_name||'EXCEL',
    fexists,
    file_length,
    block_size);
    I changed the original mail id. Please look into this code and correct me...thank you so much
    Regards
    JP

  • How to create different log files for different levels

    Hi,
    Can some one please help me with my problem I have here?
    I want to send log data to two diffrent files depending on the logging level such as DEBUG and WARN using the same logger instance.
    How can you configure this in log4j.properties.
    Please post sample code for log4j.properties to achieve this.
    Thanks in advance.
    Anurag SIngh

    Hi,
    I have tried your code. the issue is other log file for error is blank. its not writting log to that file.
    following is the code of my log4j.properties
    Please read the code first
    # Set root logger level to DEBUG and its only appender to A1.
    log4j.rootCategory=DEBUG,A1
    #Appender and its layout for A1
    log4j.appender.A1=org.apache.log4j.FileAppender
    log4j.appender.A1.layout=org.apache.log4j.PatternLayout
    log4j.appender.A1.layout.ConversionPattern=%-4r [%t] %-5p %c %x - %log4j.appender.A1.Threshold=DEBUG
    log4j.appender.A1.File=./LogonApplication_Debug.log
    log4j.appender.A1.Append=true
    #Appender and its layout for A2
    log4j.appender.A2=org.apache.log4j.FileAppender
    log4j.appender.A2.layout=org.apache.log4j.PatternLayout
    log4j.appender.A2.layout.ConversionPattern=%-4r [%t] %-5p %c %x - %
    log4j.appender.A2.Threshold=ERROR
    log4j.appender.A2.File=./LogonApplication_Error.log
    log4j.appender.A2.Append=true
    In my logger class i have written something like this:
    public static Logger myLogger = Logger.getLogger (LoggerTest.class.getName());
         public static void main(String args[]){
              myLogger.debug("I have logged a debug message");
              myLogger.error("I have logged an error message");
    Issue:
    Now the problem is that it is creating two files specified in the configuration but it is logging messages both the messages debug and error in only ./LogonApplication_Debug.log file.
    Though it is creating the ./LogonApplication_Error.log file, but it is only blank.
    Can you please trace whats missing/wrong.
    Thanks a lot.
    Shanu

Maybe you are looking for

  • Error while running a OAF page with 12.1.1 (jdeveloper patch 8431482)

    Hi, I am working on Oracle apps R12.1.1 i m using jdeveloper with patch 8431482 which is mentioned in metalink Doc Id 416708.1 i have tried to run the page from jdeveloper. it is giving me the following error. i have checked the DBC file and it is co

  • Audio begins lagging behind video in iTunes;  other video works fine

    Something that appears to have started after iTunes 10.x as I don't normally play too many videos on the PC from iTunes is regarding video playback in iTunes. The longer a video plays the longer the lag between what's showing up on the screen and the

  • Oracle API for sale rep update

    Hi , I want to update sales rep information by using the API -HZ_CUST_ACCOUNT_SITE_V2PUb.update_cust_site_use. I have four fields which i want to update through the form..please help me in writing the syntax to achieve this.I want to write this synta

  • How to cleanup the mess?

    OK, PC users since the days of the original real IBM PC...over the last six months or so our household has drunk the cool-aid, as they say: two iPhones, two iPod touch and two nicely-loaded Mac Pro's. We are not abandoning PC's, just adding on Macs t

  • Preset Not Behaving Properly

    I have a preset which tweaks clarity, vibrance, sharpening and lens correction.  I recently applied it to a new photo and it didn't seem to do much.  I went in to edit it ( right clicked on preset and selected UPDATE WITH CURRENT SETTINGS).  Only whi