OPM과 OM ISSUE 해결을 위해 LOG FILE을 생성 및 검색하는 방법

제품 : MFG_OM
작성날짜 : 2005-11-09
OPM과 OM ISSUE 해결을 위해 LOG FILE을 생성 및 검색하는 방법
===========================================================
PURPOSE
Customer에게 Oralce Processing Manufacturing(OPM)/Order Management(OM)
issue 해결을 위해 Log file을 생성하고 검색하는 방법을 알고 이를 고객에게
전달할 수 있다.
Explanation
OPM 고객이 Order process 중 fail이 발생 했을 경우,fails 된 Order process
에 따라 우리는 관련된 log file을 받기를 원한다.
예를 들면;
a) Order pad에 문제가 있다면 사용자는 Tools > Debug을 'On'으로 setting
함으로써 log file을 생성할 수 있다.
이는 Order pad 자체에서 행할수 있고 좀 더 자세한 사항은 아래의 step2 에
설명되어져 있다.
b) 만약 pick release가 fails 된다면 OPM support는 'pick release
generation' concurrent manager log file을 요구하게 된다.
좀 더 자세한 사항은 아래 step4 에 설명되어져 있다.
c) 만약 Transact process move order form(manual allocation screen)의
allocation 작업이나 pick confirm이 fails 이라면 support는 생성되는
OPMLOG***를 받기를 원할 것이다.
좀 더 자세한 사항은 아래 step5 에 설명되어져 있다.
d) 만약 Order Management/Inventory Interface가 fail 이라면 support는
'Interface Trip Stop' concurrent 실행중에 생성된 log file을 받기를 원할
것이다. 자세한 사항은 step5 에 설명되어져 있다.
OM/OPM issues를 위한 Log files의 생성 및 검색 방법
1. utl_file_dir parameter 값을 얻기 위해 아래의 sql을 실행한다.
select value
from v$parameter
where name like 'utl_file_dir';
2. OM: Debug Directory profile option 값으로 step1의 결과값이 setting
되어져 있는지 확인한다.
3. Order pad에서 log file 생성:
1) OM: Debug level profile을 '5'로 setting
2) Order pad에서 Tools > Debug를 ON, lof file name을 적어 놓는다.
3) Issue를 재현하고 step1 결과로 나온 경로에서 log file을 검색한다.
4. Pick release concurrent log file을 생성하기 위해서는 profile option
OM: Debug level을 '5' setting 후 pick release를 concurrent manager
mode로 수행한다.
Application의 view requests 화면에서 log file을 검색할 수 있다.
5. Transact move order form에서 'manual allocation' 진행시의 log file을
생성하기 위해서 아래의 steps을 따른다.
1) Step2에서 지정한 경로에서 'OPMLOG****' file을 remove 한다.
2) Manual allocation을 수행한다.
3) 'OPMLOG***' log file이 생성될 것이며, '****'는 session id 값이다.
Session id 값은 application 화면의 Help > Tools > Examine의
$Profiles/db_session_id 위에서 찾을 수 있다.
4) Pick confirm에 관련된 log는 2)번 step을 제외하고, pick confirm을
수행하여 생성할 수 있다.
6. Interface Trip Stop(ITS)가 실행되는 중의 log file을 생성하기 위해서는
OM: Debug level profile option 값을 Interface Trip Stop을 실행하기
전에 '5'로 setting 한다.
Interface Trip Stop을 실행하기전에 'Defer' checkbox가 uncheck 되어
있는지 확인해야 하며, 이 log file은 Application의 view requests 화면
에서 검색할 수 있다.
Reference Documents
Note 290069.1

Similar Messages

  • Performance issues - Log file parallel write

    Hi there,
    Since a few months I have big performance issues with my Oracle 11.2.0.1.0.
    If I look in the Enterprise manager (in blocking sessions) I see al lot of "log file paralles writes" and a lot of "log file sync" .
    We have configured an active data guard environment and are using ASM.
    We are not stressing out the database with heavy queries or commits or something, but sometimes during the day this happens on non specific times...
    We've investigated everything (performance to SAN / heavy queries / oracle problems etc etc) and we really don't know what to do anymore so i thought.. let's try a post on the Forum.....
    Perhaps someone had similar things?
    Thanks,
    BR
    Mark

    mwevromans wrote:
    See blow a tail of alertlog.
    Tue Apr 24 15:12:17 2012
    Thread 1 cannot allocate new log, sequence 194085
    Checkpoint not complete
    Current log# 1 seq# 194084 mem# 0: +DATA/kewillprd/onlinelog/group_1.262.712516155
    Current log# 1 seq# 194084 mem# 1: +FRA/kewillprd/onlinelog/group_1.438.756466165
    LGWR: Standby redo logfile selected to archive thread 1 sequence 194085
    LGWR: Standby redo logfile selected for thread 1 sequence 194085 for destination LOG_ARCHIVE_DEST_2
    Thread 1 advanced to log sequence 194085 (LGWR switch)
    Current log# 2 seq# 194085 mem# 0: +DATA/kewillprd/onlinelog/group_2.264.712516155
    Current log# 2 seq# 194085 mem# 1: +FRA/kewillprd/onlinelog/group_2.418.756466215
    Tue Apr 24 15:12:21 2012
    Archived Log entry 388061 added for thread 1 sequence 194084 ID 0x90d7aa62 dest 1:
    Tue Apr 24 15:14:09 2012
    Thread 1 cannot allocate new log, sequence 194086
    Checkpoint not complete
    Current log# 2 seq# 194085 mem# 0: +DATA/kewillprd/onlinelog/group_2.264.712516155
    Current log# 2 seq# 194085 mem# 1: +FRA/kewillprd/onlinelog/group_2.418.756466215
    LGWR: Standby redo logfile selected to archive thread 1 sequence 194086
    LGWR: Standby redo logfile selected for thread 1 sequence 194086 for destination LOG_ARCHIVE_DEST_2
    Thread 1 advanced to log sequence 194086 (LGWR switch)
    Current log# 3 seq# 194086 mem# 0: +DATA/kewillprd/onlinelog/group_3.266.712516155
    Current log# 3 seq# 194086 mem# 1: +FRA/kewillprd/onlinelog/group_3.435.756466241
    Tue Apr 24 15:14:14 2012
    Archived Log entry 388063 added for thread 1 sequence 194085 ID 0x90d7aa62 dest 1:
    Tue Apr 24 15:16:46 2012
    Thread 1 cannot allocate new log, sequence 194087
    Checkpoint not complete
    Current log# 3 seq# 194086 mem# 0: +DATA/kewillprd/onlinelog/group_3.266.712516155
    Current log# 3 seq# 194086 mem# 1: +FRA/kewillprd/onlinelog/group_3.435.756466241
    Thread 1 cannot allocate new log, sequence 194087
    Private strand flush not complete
    Current log# 3 seq# 194086 mem# 0: +DATA/kewillprd/onlinelog/group_3.266.712516155
    Current log# 3 seq# 194086 mem# 1: +FRA/kewillprd/onlinelog/group_3.435.756466241
    LGWR: Standby redo logfile selected to archive thread 1 sequence 194087
    LGWR: Standby redo logfile selected for thread 1 sequence 194087 for destination LOG_ARCHIVE_DEST_2
    Thread 1 advanced to log sequence 194087 (LGWR switch)
    Current log# 1 seq# 194087 mem# 0: +DATA/kewillprd/onlinelog/group_1.262.712516155
    Current log# 1 seq# 194087 mem# 1: +FRA/kewillprd/onlinelog/group_1.438.756466165
    Tue Apr 24 15:16:54 2012
    Archived Log entry 388065 added for thread 1 sequence 194086 ID 0x90d7aa62 dest 1:
    Tue Apr 24 15:18:59 2012
    Thread 1 cannot allocate new log, sequence 194088
    Checkpoint not complete
    Current log# 1 seq# 194087 mem# 0: +DATA/kewillprd/onlinelog/group_1.262.712516155
    Current log# 1 seq# 194087 mem# 1: +FRA/kewillprd/onlinelog/group_1.438.756466165
    Thread 1 cannot allocate new log, sequence 194088
    Private strand flush not complete
    Current log# 1 seq# 194087 mem# 0: +DATA/kewillprd/onlinelog/group_1.262.712516155
    Current log# 1 seq# 194087 mem# 1: +FRA/kewillprd/onlinelog/group_1.438.756466165
    LGWR: Standby redo logfile selected to archive thread 1 sequence 194088
    LGWR: Standby redo logfile selected for thread 1 sequence 194088 for destination LOG_ARCHIVE_DEST_2
    Thread 1 advanced to log sequence 194088 (LGWR switch)
    Current log# 2 seq# 194088 mem# 0: +DATA/kewillprd/onlinelog/group_2.264.712516155
    Current log# 2 seq# 194088 mem# 1: +FRA/kewillprd/onlinelog/group_2.418.756466215
    Tue Apr 24 15:19:06 2012
    Archived Log entry 388067 added for thread 1 sequence 194087 ID 0x90d7aa62 dest 1:
    Tue Apr 24 15:22:00 2012
    Thread 1 cannot allocate new log, sequence 194089
    Checkpoint not complete
    Current log# 2 seq# 194088 mem# 0: +DATA/kewillprd/onlinelog/group_2.264.712516155
    Current log# 2 seq# 194088 mem# 1: +FRA/kewillprd/onlinelog/group_2.418.756466215
    Thread 1 cannot allocate new log, sequence 194089
    Private strand flush not complete
    Current log# 2 seq# 194088 mem# 0: +DATA/kewillprd/onlinelog/group_2.264.712516155
    Current log# 2 seq# 194088 mem# 1: +FRA/kewillprd/onlinelog/group_2.418.756466215
    LGWR: Standby redo logfile selected to archive thread 1 sequence 194089
    LGWR: Standby redo logfile selected for thread 1 sequence 194089 for destination LOG_ARCHIVE_DEST_2
    Thread 1 advanced to log sequence 194089 (LGWR switch)
    Current log# 3 seq# 194089 mem# 0: +DATA/kewillprd/onlinelog/group_3.266.712516155
    Current log# 3 seq# 194089 mem# 1: +FRA/kewillprd/onlinelog/group_3.435.756466241
    Tue Apr 24 15:19:06 2012
    Archived Log entry 388069 added for thread 1 sequence 194088 ID 0x90d7aa62 dest 1:Hi
    1st switch time ==> Tue Apr 24 15:18:59 2012
    2nd switch time ==> Tue Apr 24 15:19:06 2012
    3rd switch time ==> Tue Apr 24 15:19:06 2012
    Redo log file switch has good impact on the performance of the database. Frequent log switches may lead to the slowness of the database . Oracle documents suggests to resize the redolog files so that log switches happen more like every 15-30 min (roughly depending on the architecture and recovery requirements).
    AS i check the alertlog file and find that the log are switchinh very fequent which is one of the reason that you are getting checkpoint  not complete message . i have face this issue many times and i generally increase the size of the logfile and set the archive_lag_time parameter as i have suggested above . If you further want to go root cause and more details then above guys will help you more because i don't have much experience in database tunning . If you looking for aworkarounf then you must go through it .
    Good Luck
    --neeraj                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Apex/OHS issue:log files in opmn consume disk space

    Hi.
    I have installed OHS from Companion CD (10g Rel1) and the new Apex 3.0.
    There seems to be a problem because log files in opmn/logs are eating up my disk space. Apparently there's something missing or not configured yet.
    The exact error message is:
    07/04/12 10:34:13 [4] Falta el factor de formato de la conexión local 0,127.0.0.1,6100
    The contents of etc/hosts is:
    [oracle@caman bin]$ cat /etc/hosts
    # Do not remove the following line, or various programs
    # that require network functionality will fail.
    127.0.0.1 localhost.localdomain localhost
    172.17.1.8 caman.rioturbio.com.ve caman
    172.17.1.1 tepuy.rioturbio.com.ve tepuy
    OS info:
    [oracle@caman bin]$ cat /etc/redhat-release
    Red Hat Enterprise Linux AS release 4 (Nahant Update 2)
    Please help ...! Thanks in advance ....!

    Hi,
    I believe this is due to a conflict with something else using that port number (6100). One solution would be to edit the opmn config file and change the port from 6100 to something else (such as 6101).

  • Log File Issue In SQL server 2005 standard Edition

    We have database of size 375GB .The data file has 80 GB free space within .When trying to rebuild the index we had 450 GB free space on the disk where Log file is residing.The rebuild index activity failed due to space issue.added more space and got the
    job done successfully
    The Log file has grow up to 611GB to complete the rebuild index.
    version :SQL server 2005 Standard Edition .Is ther a way to estimate the space required for rebuild index in this version.
    I am aware we notmaly allocate 1.5 times of data file.But in this case It was totaly wrong.
    Any suggestion with examples would be appreciated.
    Raghu

    OK, there's a few things here.
    Can you outline for everybody the recovery model you are using, the frequency with which you take full, differential and transaction log backups.
    Are you selectively rebuilding your indexes or are you rebuilding everything?
    How often are you doing this? Do you need to?
    There are some great resources on automated index maintenance, check out
    this post by Kendra Little.
    Depending on your recovery point objectives I would expect a production database to be in the full recovery mode and as part of this you need to be taking regular log backups otherwise your log file will just continue to grow. By taking a log backup it will
    clear out information from inactive VLF's and therefore allow SQL Server to write back to those VLF's rather than having to grow the log file. This is a simplified version of events, there are caveats.
    A VLF will be marked as active if it still has an open transaction in it or there is a HA option that still requires that data to be available as that data has not been copied to another node yet.
    Most customers that I see take transaction log backups every 15 - 30 minutes, but this really does depend upon how much data your company can afford to lose. That's another discussion for another day.
    Make sure that you take a transaction log backup prior to your job that does your index rebuilds (hopefully a smart job not a sledge hammer job).
    As mentioned previously swapping to bulk logged can help to reduce the size of the amount of information logged during index rebuilds. If you do this make sure to swap back into the full recovery model straight after and perform a full backup. There are
    problems with the ability to do point in time restores whilst in the bulk logged recovery model, so you need to reduce the amount of time you use it.
    Really you also need to look at how your indexes are created does the design of them lead to them being fragmented on a regular basis? Are they being used? Are there better indexes out there that can help performance?
    Hopefully that should put you on the right track.
    If you find this helpful, please mark the post as helpful,
    If you think this solves the problem, please propose or mark it an an answer.
    Please provide details on your SQL Server environment such as version and edition, also DDL statements for tables when posting T-SQL issues
    Richard Douglas
    My Blog: Http://SQL.RichardDouglas.co.uk
    Twitter: @SQLRich

  • Standby creating archives log files issue!

    Hello Everyone,
    Working on oracle 10g R2/Windows, I have created a dataguard with one standby database, but there is a strange issue that happen, and I'll need someone to shed the light to me.
    By default archived log created from the primary database should be the  sent to the stanndby database, but I found that the standby database has plus one archived log file.
    From the primary database:
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\local_destination1_orcl
    Oldest online log sequence 1021
    Next log sequence to archive 1023
    Current log sequence 1023
    contents of C:\local_destination1_orcl
    1_1022_623851185.ARC
    from the standby database:
    SQL> archive log list
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\local_destination1_orcl
    Oldest online log sequence 1022
    Next log sequence to archive 0
    Current log sequence 1023
    contents of C:\local_destination1_orcl
    1_1022_623851185.ARC
    1_1023_623851185.ARC ---> this is the extra archive file created in the standby database, should someone let me know how to avoid this?
    Thanks for your help

    SELECT FROM v$version;*
    BANNER
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bi
    PL/SQL Release 10.2.0.1.0 - Production
    CORE 10.2.0.1.0 Production
    TNS for 64-bit Windows: Version 10.2.0.1.0 - Production
    NLSRTL Version 10.2.0.1.0 - Production
    The standby database is a physical standby database (not logical standby)
    Thanks against for your contribution, but I'm still not know why standby create and arhive files too?

  • WLS 7.0.4 - JMS Connection Factory - Server Affinity - issues in log file

    <b>WLS 7.0.4 - JMS Connection Factory - Server Affinity - issues in log file</b>
              We are using WLS 7.0.4 - One of JMS connection factory setting in admin console we selected "Server Affinity" options.
              We see this messages appear in Weblogic log file,
              ####<Apr 24, 2006 1:56:53 AM EDT> <Error> <Cluster> <liberatenode4.dc2.adelphia.com> <node4_svr> <ExecuteThrea
              d: '4' for queue: '__weblogic_admin_rmi_queue'> <kernel identity> <> <000123> <Conflict start: You tried to bi
              nd an object under the name sbetrmi2 in the JNDI tree. The object you have bound from liberatenode2.dc2.adelp
              hia.com is non clusterable and you have tried to bind more than once from two or more servers. Such objects ca
              n only deployed from one server.>
              and then,
              ####<Apr 24, 2006 1:58:12 AM EDT> <Error> <Cluster> <liberatenode5.dc2.adelphia.com> <node5_svr> <ExecuteThrea
              d: '7' for queue: '__weblogic_admin_rmi_queue'> <kernel identity> <> <000125> <Conflict Resolved: sbetrmi2 for
              the object from liberatenode5.dc2.adelphia.com under the bind name sbetrmi2 in the JNDI tree.>
              Should we use 'load balancing option' instead of 'server affinity' ?
              Any thuoghts?
              Thanks in adv.
              Vijay

    Test Reply
              <Vijay Kumar> wrote in message news:[email protected]..
              > <b>WLS 7.0.4 - JMS Connection Factory - Server Affinity - issues in log
              > file</b>
              >
              > We are using WLS 7.0.4 - One of JMS connection factory setting in admin
              > console we selected "Server Affinity" options.
              >
              > We see this messages appear in Weblogic log file,
              > ####<Apr 24, 2006 1:56:53 AM EDT> <Error> <Cluster>
              > <liberatenode4.dc2.adelphia.com> <node4_svr> <ExecuteThrea
              > d: '4' for queue: '__weblogic_admin_rmi_queue'> <kernel identity> <>
              > <000123> <Conflict start: You tried to bi
              > nd an object under the name sbetrmi2 in the JNDI tree. The object you have
              > bound from liberatenode2.dc2.adelp
              > hia.com is non clusterable and you have tried to bind more than once from
              > two or more servers. Such objects ca
              > n only deployed from one server.>
              >
              > and then,
              > ####<Apr 24, 2006 1:58:12 AM EDT> <Error> <Cluster>
              > <liberatenode5.dc2.adelphia.com> <node5_svr> <ExecuteThrea
              > d: '7' for queue: '__weblogic_admin_rmi_queue'> <kernel identity> <>
              > <000125> <Conflict Resolved: sbetrmi2 for
              > the object from liberatenode5.dc2.adelphia.com under the bind name
              > sbetrmi2 in the JNDI tree.>
              >
              >
              > Should we use 'load balancing option' instead of 'server affinity' ?
              >
              > Any thuoghts?
              >
              > Thanks in adv.
              > Vijay

  • Transactional log file issue

    Dear All,
    There have been issues in the past where the transactional log file has grown too big that it made the drive to limit its size. I would like to know the answers to the following
    please:
    1. To resolve the space issue, is the correct way to first take a backup of the transactional log then shrink the transactional log file?
    2. What would be the recommended auto growth size, for example if I have a DB which is 1060 GB?
    3. At the moment, the transactional log backup is done every 1 hour, but I'm not sure if it should be taken more regularly?
    4. How often should the update stat job should run please?
    Thank you in advance!

    Hi
    My answers might be very similar to geeks already answer, but hope it will add something more
    1. To resolve the space issue, is the correct way to first take a backup of the transactional log then shrink the transactional log file?
     --> If database recovery model is full \ bulk then tlog backup is helpful, and it doesnt help try to increase frequency of log backup and you can refer :
    Factors That Can Delay Log Truncation
    2. What would be the recommended auto growth size, for example if I have a DB which is 1060 GB?
    Auto grow for very large db is very crucial if its too high can cause active vlf and too less can cause fragmentation. In your case your priority is to control space utilizatiuon.
    i suggest you to keep minimum autogrowth and it must be in size not in percentage.
    /*******Auto grow formula for log file**********/
    Auto grow less than 64MB = 4 VLFs 
    Autogrow  of 64MB and less than 1GB = 8 VLFs 
    Autogrow of 1GB and larger = 16 VLFs
    3. At the moment, the transactional log backup is done every 1 hour, but I'm not sure if it should be taken more regularly?
    ---> If below query returns log_backup for respective database then yes you can to increase log backup frequency. But if it returns some other factor , please check above
    mention link
    "select name as [database] ,log_reuse_wait , log_reuse_wait_desc from sys.databases"
    4. How often should the update stat job should run please?
    this totaly depend on ammount of dml operation you are performing. you can select auto update stats and weekly you can do update stats with full scan.
    Thanks Saurabh Sinha
    http://saurabhsinhainblogs.blogspot.in/
    Please click the Mark as answer button and vote as helpful
    if this reply solves your problem

  • Issues o delete log files

    Hi All,
    I am using oracle AS 10g wireless for messaging services. Now one log file is taking too much size and i want to delete it. Can anyone let me know...can I delete it
    Log file name -----Oracle Home\j2ee\OC4J_Wireless\log\OC4J_Wireless_default_island_1
    I think there should not be any issue to delete it but i want to make sure plz let me ASAP.
    Thanks in advance
    Saqib

    Ok, as nobody wanted to risk answering this question a close the message.

  • File system task - Move file issue, skip the current log file thats open by another process - IIS logs

    Hi,
    I have created a package to move IIS log files from their source directory to another directory called processing. However, when the package tries to move the current log file it errors because the file is in use by another process (by IIS as its still writing
    to it). I would like the package to just ignore the "in-use" file and process the rest and show the package completing successfully. The file in use will be rolled by IIS at midnight and the file will be picked up by the next package run.
    However, when the file is rolled and becomes free a new log file is created which will be marked as "in-use"...... which should be ignored at the next run.
    Is there some way that I can tell the file system task to ignore in use files or add an event handler to do the same sort of thing?
    Any assistance is appreciated

    Hi Arthur,
    Thank you for your reply.
    I have resolved the issue with the following example:
    http://www.timmitchell.net/post/2010/08/23/ssis-conditional-file-processing-in-a-foreach-loop/
    What I realised was that I needed to just ignore the current log file which gets created daily, so using the above example (and changing the last written date to created date) I am able to ignore/skip the log thats in use.
    Adam

  • Mpd issue not able to open log file

    I've been trying to setup mpd on my netbook arch install and have got config and everything setup but I keep getting this error,
    log: problem opening log file "/var/log/mpd/mpd.log" (config line 37) for writing
    My config line 37 is
    log_file "/var/log/mpd/mpd.log"
    the mpd user owns /var/log/mpd so I don't know what the issue is.
    Let me know if you need anything else from my system.

    Okay I'm at a loss here for whatever reason music does not show up. I tried starting mpd as root and I get this output
    [maxmarze@titan ~]$ sudo mpd
    listen: bind to '0.0.0.0:6600' failed: Address already in use (continuing anyway, because binding to '[::]:6600' succeeded)
    music directory is not a directory: "/home/maxmarze/music"
    output: No "audio_output" defined in config file
    output: Attempt to detect audio output device
    output: Attempting to detect a alsa audio device
    output: Successfully detected a alsa audio device
    My mpd config file is as such
    music_directory "/home/maxmarze/music/" # Your music dir.
    playlist_directory "/var/lib/mpd/playlists"
    db_file "/var/lib/mpd/mpd.db"
    log_file "/var/log/mpd/mpd.log"
    pid_file "/var/run/mpd/mpd.pid"
    state_file "/var/lib/mpd/mpdstate"
    user "mpd"
    # Binding to address and port causing problems in mpd-0.14.2 best to leave
    # commented.
    # bind_to_address "127.0.0.1"
    # port "6600"
    and my ncmpcpp config is
    mpd_host = "127.0.0.1"
    mpd_port = "6600"
    mpd_music_dir = "/home/maxmarze/music/"
    mpd_connection_timeout = "5"
    mpd_crossfade_time = "5"
    The music directory is owned by me and mpd as I assumed it would be.

  • Performance Issue: Wait event "log file sync" and "Execute to Parse %"

    In one of our test environments users are complaining about slow response.
    In statspack report folowing are the top-5 wait events
    Event Waits Time (cs) Wt Time
    log file parallel write 1,046 988 37.71
    log file sync 775 774 29.54
    db file scattered read 4,946 248 9.47
    db file parallel write 66 248 9.47
    control file parallel write 188 152 5.80
    And after runing the same application 4 times, we are geting Execute to Parse % = 0.10. Cursor sharing is forced and query rewrite is enabled
    When I view v$sql, following command is parsed frequently
    EXECUTIONS PARSE_CALLS
    SQL_TEXT
    93380 93380
    select SEQ_ORDO_PRC.nextval from DUAL
    Please suggest what should be the method to troubleshoot this and if I need to check some more information
    Regards,
    Sudhanshu Bhandari

    Well, of course, you probably can't eliminate this sort of thing entirely: a setup such as yours is inevitably a compromise. What you can do is make sure your log buffer is a good size (say 10MB or so); that your redo logs are large (at least 100MB each, and preferably large enough to hold one hour or so of redo produced at the busiest time for your database without filling up); and finally set ARCHIVE_LAG_TARGET to something like 1800 seconds or more to ensure a regular, routine, predictable log switch.
    It won't cure every ill, but that sort of setup often means the redo subsystem ceases to be a regular driver of foreground waits.

  • Database Engine won't start - master DB log file issue

    I had a server patched and reboot yesterday and now the SQL services won't start because it's saying it can't find the master log file. I checked the path and the log file is present. Any suggestions.
    2014-04-21 09:48:55.22 Server      Microsoft SQL Server 2008 R2 (SP2) - 10.50.4000.0 (X64)
     Jun 28 2012 08:36:30
     Copyright (c) Microsoft Corporation
     Standard Edition (64-bit) on Windows NT 6.1 <X64> (Build 7601: Service Pack 1)
    2014-04-21 09:48:55.22 Server      (c) Microsoft Corporation.
    2014-04-21 09:48:55.22 Server      All rights reserved.
    2014-04-21 09:48:55.22 Server      Server process ID is 4008.
    2014-04-21 09:48:55.22 Server      System Manufacturer: 'HP', System Model: 'ProLiant DL380 G5'.
    2014-04-21 09:48:55.22 Server      Authentication mode is MIXED.
    2014-04-21 09:48:55.22 Server      Logging SQL Server messages in file 'C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Log\ERRORLOG'.
    2014-04-21 09:48:55.22 Server      This instance of SQL Server last reported using a process ID of 4080 at 4/21/2014 9:47:41 AM (local) 4/21/2014 2:47:41 PM (UTC). This is an informational message only; no user action is required.
    2014-04-21 09:48:55.22 Server      Registry startup parameters:
      -d C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATA\master.mdf
      -e C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Log\ERRORLOG
      -l C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATA\mastlog.ldf -g512
    2014-04-21 09:48:55.22 Server      SQL Server is starting at normal priority base (=7). This is an informational message only. No user action is required.
    2014-04-21 09:48:55.22 Server      Detected 4 CPUs. This is an informational message; no user action is required.
    2014-04-21 09:48:55.33 Server      Using dynamic lock allocation.  Initial allocation of 2500 Lock blocks and 5000 Lock Owner blocks per node.  This is an informational message only.  No user action is required.
    2014-04-21 09:48:55.35 Server      Node configuration: node 0: CPU mask: 0x000000000000000f:0 Active CPU mask: 0x000000000000000f:0. This message provides a description of the NUMA configuration for this computer. This is an informational
    message only. No user action is required.
    2014-04-21 09:48:55.38 spid7s      Starting up database 'master'.
    2014-04-21 09:48:55.39 spid7s      Error: 17204, Severity: 16, State: 1.
    2014-04-21 09:48:55.39 spid7s      FCB::Open failed: Could not open file C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATA\mastlog.ldf -g512 for file number 2.  OS error: 2(The system cannot find the file
    specified.).
    2014-04-21 09:48:55.39 spid7s      Error: 5120, Severity: 16, State: 101.
    2014-04-21 09:48:55.39 spid7s      Unable to open the physical file "C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\DATA\mastlog.ldf -g512". Operating system error 2: "2(The system cannot find the
    file specified.)".

    That did it.  Not sure why that switch was there.
    It might have been intended, but it was added incorrectly. -g512 is a legal switch. Then again, since you have 64-bit SQL Server, it is not likely that you will needed.
    (The -g switch increases the area known as memtoleave, which is the virtual address space (VAS) not used for the buffer cache. On 32-bit SQL Server you may need to increase this area from the default of 256 MB. On 64-bit SQL Server the full physical memory
    is available for VAS; on 32-bit it's only 2GB.)
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Disk space issue with log files

    I noticed that I was running out of disk space, so I downloaded OmniDiskSweeper and ran it. Turns out that the main culprit is log files. They are located in /private/var/log/asl. However, OmniDiskSweeper won't let me delete them (it doesn't give an error message or anything--it just won't delete the files). What are these .asl files and how can I delete them? I cannot see them when I open up the Finder, so I assume they are hidden. Any help would be most appreciated. Thanks!

    plarkin wrote:
    I noticed that I was running out of disk space, so I downloaded OmniDiskSweeper and ran it. Turns out that the main culprit is log files. They are located in /private/var/log/asl. However, OmniDiskSweeper won't let me delete them (it doesn't give an error message or anything--it just won't delete the files). What are these .asl files and how can I delete them? I cannot see them when I open up the Finder, so I assume they are hidden.
    Those log files are created by the "Apple System Log" process. The first thing I'd do is to use the Console utility to examine what's in them to see if there's anything obvious that should be fixed. In Console, open the "/var/log" item on the left, open the "asl" item, then select the first item in the list. You'll see a lot of gibberish, but you'll also see some text.
    To navigate to those files in the Finder, use the menu item Go -> Go to Folder and type "/var/log/asl" (without the quotes) in the field, and click "Go".

  • Sql server 2005 log file shrinking issue.

    hi,
    one of my production server database  log file size is increasing continuously,we can take log backup but no use,
    we can change the database recovery model from  full to simple we shrink the log file, log file shrink successfully but space is not released,and we are getting the alerts continuously your physical disk is 95% like that.
    here what we can do how reduced the log file size and how to control the log file size.
    please give me the answer  ASAP.
    Thanks,

    Hello,
    In SQL Server, log truncation occurs automatically as follows:
    • Under the simple recovery model, after a checkpoint.
    • Under the full recovery model or bulk-logged recovery model, after a log backup, if a checkpoint has occurred since the previous backup.
    As Dan posted earlier, when we run the above code, we will get an integer value. The log_reuse_wait_desc value tells us why the transaction log space is not reused. You can refer to the article below to check the root cause. 
    http://msdn.microsoft.com/en-us/library/ms345414.aspx
    Regards,
    Alisa Tang
    Alisa Tang
    TechNet Community Support

  • Adobe document service issue attached he trace and log file

    hi friends
    adobe pdf is not working.
    i am  sending the trace file and server log  file
    1.5 #00145ED2ABF2006C00000087000EF0AC00045A8A76851A40#1225450647263#com.adobe.ads.operation.support
    .RenderErrorLog#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#487
    #SAP J2EE Engine JTA Transaction : [18363cffffffa200813]#n47a#DEMO1#772fd940a73611dda59700145ed2abf2#
    SAPEngine_Application_Thread[impl:3]_6##0#0#Error#1#/System/Server#Plain###Encountered a problem
    embedding Render collateral:
    General exception while merging XMLForm Module Render log data into log.pdf#
    #1.5 #00145ED2ABF2006C00000089000EF0AC00045A8A76852292#1225450647265#com.adobe.ads.request.Request#com.
    adobe/AdobeDocumentServices#com.adobe.ads.request.Request#ADSUSER#487#SAP J2EE Engine JTA Transaction :
    [18363cffffffa200813]#n47a#DEMO1#772fd940a73611dda59700145ed2abf2#SAPEngine_Application_Thread[impl:3]_6##0#0#
    Error#1#/System/Server#Plain###Client: 48316 AppName: com.crane.episode.EpisodeApp
    Unhandled exception detected during a "Render" operation.
    Request start time: Fri Oct 31 15:57:27 PKT 2008
    java.lang.NullPointerException
    Exception Stack Trace:
    java.lang.NullPointerException
         at com.adobe.ads.operation.Render.addErrorPdfToStreams(Unknown Source)
         at com.adobe.ads.operation.Render.handleRenderException(Unknown Source)
         at com.adobe.ads.operation.Render.execute(Unknown Source)
         at com.adobe.ads.operation.ADSOperation.doWork(Unknown Source)
         at com.adobe.ads.request.Request.processOperations(Unknown Source)
         at com.adobe.ads.request.Request.process(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.processRequest(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.rpData(Unknown Source)
         at com.adobe.AdobeDocumentServicesLocalLocalObjectImpl0_0.rpData(AdobeDocumentServicesLocalLocalObjectImpl0_0.java:120)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:85)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:58)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:60)
         at java.lang.reflect.Method.invoke(Method.java:391)
         at com.sap.engine.services.webservices.runtime.EJBImplementationContainer.invokeMethod(EJBImplementationContainer.java:126)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:157)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:79)
         at com.sap.engine.services.webservices.runtime.servlet.ServletDispatcherImpl.doPost(ServletDispatcherImpl.java:92)
         at SoapServlet.doPost(SoapServlet.java:51)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:401)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:387)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:365)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:944)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:266)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:95)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:175)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(AccessController.java:215)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)
    server0.log file
    #1.5 #00145ED2ABF20079000007260005206200045AEB49C2D359#1225866508161#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Warning#1#com.adobe.ads.operation.support.RenderErrorLog#Plain###Cannot generate Log.pdf#
    #1.5 #00145ED2ABF20079000007270005206200045AEB49C2D414#1225866508162#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.operation.support.RenderErrorLog#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Error#1#com.adobe.ads.operation.support.RenderErrorLog#Plain###Encountered a problem embedding Render collateral:
    General exception while merging XMLForm Module Render log data into log.pdf#
    #1.5 #00145ED2ABF20079000007290005206200045AEB49C2D982#1225866508163#/System/Server#com.adobe/AdobeDocumentServices#com.adobe.ads.request.Request#ADSUSER#1317#SAP J2EE Engine JTA Transaction : [18363cffffffa200b4e]#n47a#DEMO#eb569ff0ab0211ddce8400145ed2abf2#SAPEngine_Application_Thread[impl:3]_17##0#0#Error#1#com.adobe.ads.request.Request#Plain###Client: eb582 AppName: samplesearchlab.chdp.com.SampleSearchLabApp
    Unhandled exception detected during a "Render" operation.
    Request start time: Wed Nov 05 11:28:27 PKT 2008
    java.lang.NullPointerException
    Exception Stack Trace:
    java.lang.NullPointerException
         at com.adobe.ads.operation.Render.addErrorPdfToStreams(Unknown Source)
         at com.adobe.ads.operation.Render.handleRenderException(Unknown Source)
         at com.adobe.ads.operation.Render.execute(Unknown Source)
         at com.adobe.ads.operation.ADSOperation.doWork(Unknown Source)
         at com.adobe.ads.request.Request.processOperations(Unknown Source)
         at com.adobe.ads.request.Request.process(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.processRequest(Unknown Source)
         at com.adobe.AdobeDocumentServicesEJB.rpData(Unknown Source)
         at com.adobe.AdobeDocumentServicesLocalLocalObjectImpl0_0.rpData(AdobeDocumentServicesLocalLocalObjectImpl0_0.java:120)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:85)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:58)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:60)
         at java.lang.reflect.Method.invoke(Method.java:391)
         at com.sap.engine.services.webservices.runtime.EJBImplementationContainer.invokeMethod(EJBImplementationContainer.java:126)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:157)
         at com.sap.engine.services.webservices.runtime.RuntimeProcessor.process(RuntimeProcessor.java:79)
         at com.sap.engine.services.webservices.runtime.servlet.ServletDispatcherImpl.doPost(ServletDispatcherImpl.java:92)
         at SoapServlet.doPost(SoapServlet.java:51)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:401)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:387)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:365)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:944)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:266)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:95)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:175)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(AccessController.java:215)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:100)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:170)
    regards
    raja

    Hello,
    Were you able to resolve this?  We are experiencing the same problem and have had no luck finding a solution.
    Thanks,
    Anthony

Maybe you are looking for