Managing Home Group log file sizes

Log Name:     
Microsoft-Windows-Kernel-EventTracing/Admin
Source:        Microsoft-Windows-Kernel-EventTracing
Date:          1.10.13 18:39:13
Event ID:      4
Level:         Warning
Description:
The maximum file size for session "P2PLog" has been reached. As a result, events might be lost (not logged) to file "C:\WINDOWS\Logs\Homegroup\p2p.etl". The maximum files size is currently set to 5242880 bytes.
Name="FileName">C:\WINDOWS\Logs\Homegroup\p2p.etl</Data>
    <Data Name="ErrorCode">3221225864</Data>
    <Data Name="LoggingMode">293601285</Data>
    <Data Name="MaxFileSize">5242880</Data>
How can you change to overwrite?
How do you read one of these etl files?
TIA, Gerry

What about making changes referring to this one?  Why doesn't it refer to the Homegroup though?
C:\>wevtutil gp Microsoft-Windows-Kernel-EventTracing
BTW the help link fails.  What a surprise!
http://go.microsoft.com/fwlink/events.asp?CoName=Microsoft%20Corporation&ProdName=Microsoft%c2%ae%20Windows%c2%ae%20Operating%20System&ProdVer=6.3.9600.16384&FileName=Microsoft-Windows-System-Events.dll&FileVer=6.3.9600.16384
Where the heck are they thinking that would go?
Ah.  I might have something.  I stripped off all the detail from &ProdVer= on and got this
http://www.microsoft.com/products/ee/transform.aspx?ProdName=Windows%20Operating%20System&ProdVer=&EvtID=&EvtSrc=&FileVer=&FileName=&EvtType=&LCID=
So, if that's where the Go link was supposed to be taking us, let's try filling it in with those details that were stripped off to get there.  What are the chances I'm going to get this right?...
FWIW here is my stab at it.  ID: and Source:  are still not filled in.
http://www.microsoft.com/products/ee/transform.aspx?ProdName=Windows%20Operating%20System&ProdVer=6.3.9600.16384&EvtID=&EvtSrc=&FileVer=6.3.9600.16384&FileName=Microsoft-Windows-System-Events.dll&EvtType=&LCID=
But it looks like that link probably wouldn't be taking us any place useful anyway even if it was working correctly.
Looks like a red herring.
Robert Aldwinckle

Similar Messages

  • Log file size in Sun Access Manager

    Does anyone have an idea about the how the Sun Access Manager's log file size will increase in size with respective to the actions performed?
    Can someone give a data regarding this? If someone has a better scenario and the supportive data w.r.t log file size it will be helpful.
    Thanks,

    I would like to take the log files backup daily (for future reference).
    I need to know the following:
    1) What log files need to be backup? Do I need to take all am*.* files (It will be around 3.5 GB in size)?
    2) Ideally, I believe that only few MB of data goes into these am*.* files but, I cannot store every day 3.5 + GB of logs
    3) I observed that these files (am*.* have every days activity added to them), So I would like to enable the Log rotation
    Please let me know how can I proceed.
    Thanks

  • Enterprise manager log file sizes

    Hi,
    I was wondering if there is a way of managing the size of the emdb.nohup file in Enterprise Manager. Looking at the documentation, it looks as though you can control the emoms trace and log file sizes, but I can't find anything about the nohup log.
    Ideally I would like to be able to purge the log file.
    Thanks very much!

    Hi again,
    I found the emdb.nohup file in my log directory at the location you noted. It is apparently created when I stopped and restarted my dbconsole (emctl start dbconsole) and is updated each time I make a connection to that database (for which you are connecting), and for every time the page refreshes.
    I think the name of the file using a suffix of nohup is probably intentional on Oracle's part to indicate that this is a log file that is 'active', but in reality, it is not a true nohup file as in the sense of using the unix nohup command. (at least that is what I'm thinking)
    Sorry I'm not knowledgeable enough on this to be sure of my theory, but that is what I basically theorize.
    According to man pages on nohup, it states nohup is "a utility immune to hangups".
    "nohup - run a command immune to hangups, with output to a non-tty"
    To answer your question, there is no problem purging or pruning this file.
    I just cleared it out by sending a date to the file which effectively clears it out except for a new entry in the file with the current date/timestamp.
    e.g., $ date > emdb.nohup
    Then, I reconnected to my OEM console for this database and it updated the file with new entries for the new connection. No problem....
    Wed Aug 6 09:46:53 EDT 2008
    08/08/06 09:47:07 ## oracle.sysman.db.adm.inst.SitemapController: event="doLoad"
    08/08/06 09:47:07 ## 1. newPage = /database/instance/sitemap/sitemap
    08/08/06 09:47:07 ## 2. newPage = /database/instance/sitemap/sitemap
    Ji Li

  • Problem about space management of archived log files

    Dear friends,
    I have a problem about space management of archived log files.
    my database is Oracle 10g release 1 running in archivelog mode. I use OEM(web based) to config all the backup and recovery settings.
    I config "Flash Recovery Area" to do backup and recovery automatically. my daily backup schedule is every night at 2:00am. and my backup setting is "disk settings"--"compressed backup set". the following is the RMAN script:
    Daily Script:
    run {
    allocate channel oem_disk_backup device type disk;
    recover copy of database with tag 'ORA$OEM_LEVEL_0';
    backup incremental level 1 cumulative copies=1 for recover of copy with tag 'ORA$OEM_LEVEL_0' database;
    the retention policy is the second choice, that is "Retain backups that are necessary for a recovery to any time within the specified number of days (point-in-time recovery)". the recovery window is 1 day.
    I assign enough space for flash recovery area. my database size is about 2G. I assign 20G as flash recovery area.
    now here is the problem, through oracle online manual, it said oracle can manage the flash recovery area automatically, that is, when the space is full, it can delete the obsolete archived log files. but in fact, it never works! whenever the space is full, the database will hang up! besides, the status of archived log files is very strange, for example, it can change "obsolete" stauts from "yes" to "no", and then from "no" to "yes". I really have no idea about this! even though I know oracle usually keep archived files for some longer days than retention policy, but I really don't know why the obsolete status can change automatically. although I can write a schedule job to delete obsolete archived files every day, but I just want to know the reason. my goal is to backup all the files on disk and let oracle automatically manage them.
    also, there is another problem about archive mode. I have two oracle 10g databases(release one), the size of db1 is more than 20G, the size of db2 is about 2G. both of them have the same backup and recovery policy, except I assign more flash recovery area for db1. both of them are on archive mode. both of nearly nobody access except for the schedule backup job and sometime I will admin through oem. the strange thing is that the number of archived log files of smaller database, db2, are much bigger than ones of bigger database. also the same situation for the size of the flashback logs for point-in-time recovery. (I enable flashback logging for fast database point-in-time recovery, the flashback retention time is 24 hours.) I found the memory utility of smaller database is higher than bigger database. nearly all the time the smaller database's memory utility keeps more than 99%. while the bigger one's memory utility keeps about 97%. (I enable "Automatic Shared Memory Management" on both databases.) but both database's cup and queue are very low. I'm nearly sure no one hack the databases. so I really have no idea why the same backup and recovery policy will result so different result, especially the smaller one produces more redo logs than bigger one. does there anyone happen to know the reason or how should I do to check the reason?
    by the way, I found web based OEM can't reflect the correct database status when the database shutdown abnormally. for example, if the database hang up because of out of flash recovery area, after I assign more flash recovery area space and then restart the database, the OEM usually can't reflect the correct database status. I must restart OEM manually to correctly reflect the current database status. does there anyone know in what situation I should restart OEM to reflect the correct database status?
    sorry for the long message, I just want to describe in details to easy diagnosis.
    any hint will be greatly appreciated!
    Sammy

    thank you very much, in fact, my site's oracle never works about managing archive files automatically although I have tried all my best. at last, I made a job running daily to check the archive files and delete them.
    thanks again.

  • Roblem about space management of archived log files

    Dear friends,
    I have a problem about space management of archived log files.
    my database is Oracle 10g release 1 running in archivelog mode. I use OEM(web based) to config all the backup and recovery settings.
    I config "Flash Recovery Area" to do backup and recovery automatically. my daily backup schedule is every night at 2:00am. and my backup setting is "disk settings"--"compressed backup set". the following is the RMAN script:
    Daily Script:
    run {
    allocate channel oem_disk_backup device type disk;
    recover copy of database with tag 'ORA$OEM_LEVEL_0';
    backup incremental level 1 cumulative copies=1 for recover of copy with tag 'ORA$OEM_LEVEL_0' database;
    the retention policy is the second choice, that is "Retain backups that are necessary for a recovery to any time within the specified number of days (point-in-time recovery)". the recovery window is 1 day.
    I assign enough space for flash recovery area. my database size is about 2G. I assign 20G as flash recovery area.
    now here is the problem, through oracle online manual, it said oracle can manage the flash recovery area automatically, that is, when the space is full, it can delete the obsolete archived log files. but in fact, it never works! whenever the space is full, the database will hang up! besides, the status of archived log files is very strange, for example, it can change "obsolete" stauts from "yes" to "no", and then from "no" to "yes". I really have no idea about this! even though I know oracle usually keep archived files for some longer days than retention policy, but I really don't know why the obsolete status can change automatically. although I can write a schedule job to delete obsolete archived files every day, but I just want to know the reason. my goal is to backup all the files on disk and let oracle automatically manage them.
    also, there is another problem about archive mode. I have two oracle 10g databases(release one), the size of db1 is more than 20G, the size of db2 is about 2G. both of them have the same backup and recovery policy, except I assign more flash recovery area for db1. both of them are on archive mode. both of nearly nobody access except for the schedule backup job and sometime I will admin through oem. the strange thing is that the number of archived log files of smaller database, db2, are much bigger than ones of bigger database. also the same situation for the size of the flashback logs for point-in-time recovery. (I enable flashback logging for fast database point-in-time recovery, the flashback retention time is 24 hours.) I found the memory utility of smaller database is higher than bigger database. nearly all the time the smaller database's memory utility keeps more than 99%. while the bigger one's memory utility keeps about 97%. (I enable "Automatic Shared Memory Management" on both databases.) but both database's cup and queue are very low. I'm nearly sure no one hack the databases. so I really have no idea why the same backup and recovery policy will result so different result, especially the smaller one produces more redo logs than bigger one. does there anyone happen to know the reason or how should I do to check the reason?
    by the way, I found web based OEM can't reflect the correct database status when the database shutdown abnormally. for example, if the database hang up because of out of flash recovery area, after I assign more flash recovery area space and then restart the database, the OEM usually can't reflect the correct database status. I must restart OEM manually to correctly reflect the current database status. does there anyone know in what situation I should restart OEM to reflect the correct database status?
    sorry for the long message, I just want to describe in details to easy diagnosis.
    any hint will be greatly appreciated!
    Sammy

    thank you very much, in fact, my site's oracle never works about managing archive files automatically although I have tried all my best. at last, I made a job running daily to check the archive files and delete them.
    thanks again.

  • SQL log file size is extending rapidly

    Hello All,
    We are using ECC 6.0, our database is SQL 2005 & operating system is Windows NT 4x AMD64 L.
    Our database log file size is increasing rapidly, now its size is more than the all 4 data files (near about 300gb).
    Last week I tried to shrink log file but it didn't worked.
    Now less space is remained on disk, please help me.
    Now the system is started giving dump at the time of log in, & the dump is like "START_CALL_SICK ".
    I am attaching dump error text file.
    Please help why is this happening.
    Thanks in advance
    Mahendra

    Hi,
    I have backed up log file & shrink the file but it didn't worked for me
    What is the result? It shrinks the log and release all the space (for all committed transactions).
    How can i add another log file?
    Can i delete old log file after adding new log file.
    You can add another log file by following below steps. but in your case, this is not the right solution because you have good amount of log file configuration for your database (now its size is more than the all 4 data files (near about 300gb)).
    Open SQL server management studio > Expand database > Right click on database > Select Files > Click on Add > Give the input parameters (Logical file name, path, initial size etc.) click on OK
    If system is not allowing you to shrink the log file, it means you have active transactions in system which are continuously using your log file.
    Regards,
    Nick Loy

  • Archive log file size is varying in RAC 10g database.

    ---- Environment oracle 10g rac 9 node cluster database, with 3 log groups for each node with 500 mb size for each redo log file.
    Question is why would be the archive log file size is varying, i know when ever there is log file switch the redo log will be archived, So as our redo log file size is of 500 MB
    isn't the archive log file size should be the same as 500 MB?
    Instead we are seeing the archive log file is varying from 20 MB to 500 MB this means the redo log file is not using the entire 500 MB space? What would be causing this to happen? how can we resolve this?
    Some init parameter values.(just for information)
    fast_start_mttr_target ----- 400
    log_checkpoint_timeout ----- 0
    log_checkpoint_interval ----- 0
    fast_start_io_target ----- 0

    There was a similar discussion a few days back,
    log file switch before it filled up
    The guy later claimed it's because their log_buffer size. It's remain a mystery to me still.

  • Optimal online redo log file size

    hello to all,
    I have installed just now 11gr2 patching 2.
    I found
    The Redo Logfile Size Advisor can be used to determine the least optimal online redo log file size based on the current FAST_START_MTTR_TARGET setting and MTTR statistics.
    Which means that Redo Logfile Size Advisor is enabled only if FAST_START_MTTR_TARGET is set.
    for now this installation is not a production instance (only for my testing) I am still on 10g
    my question is:
    is't important to set FAST_START_MTTR_TARGET, in 11g or that all old infos?, because from my query I see OPTIMAL_LOGFILE_SIZE isn't set.
    SELECT TARGET_MTTR,ESTIMATED_MTTR,WRITES_MTTR,WRITES_LOGFILE_SIZE, OPTIMAL_LOGFILE_SIZE FROM V$INSTANCE_RECOVERY;TARGET_MTTR     ESTIMATED_MTTR     WRITES_MTTR     WRITES_LOGFILE_SIZE     OPTIMAL_LOGFILE_SIZE
    0     20     0     0
    and
    SELECT a.group#, b.member, a.status, a.bytes
    FROM v$log a, v$logfile b
    WHERE a.group#=b.group#
    GROUP#     MEMBER     STATUS     BYTES
    6     /u03/oradata/TEST/redo06.log     CURRENT     1048576000
    5     /u03/oradata/TEST/redo05.log     INACTIVE     1048576000
    4     /u03/oradata/TEST/redo04.log     INACTIVE     1048576000
    3     /u03/oradata/TEST/redo03.log     INACTIVE     1048576000
    2     /u03/oradata/TEST/redo02.log     INACTIVE     1048576000
    1     /u03/oradata/TEST/redo01.log     INACTIVE     1048576000
    thanks for any doc or info you will pointing me

    for now this installation is not a production instance (only for my testing) I am still on 10g
    my question is:
    is't important to set FAST_START_MTTR_TARGET, in 11g or that all old infos?,There is no difference for FAST_START_MTTR_TARGET in 10g and 11g.
    For 10g
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14237/initparams068.htm#i1127412
    For 11g
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28320/initparams079.htm
    Both documentation have exact text, means it will work exactly in 11g as 10g.
    But do'nt see this like (because its not a official documentation link... as it is different from official doc's text though)
    http://www.stanford.edu/dept/itss/docs/oracle/10g/server.101/b10755/initparams069.htm
    Regards
    Girish Sharma

  • How to set up PopProxy* log file size ?

    Dear All,
    Does anybody know how to set up MMP PopProxy* log file size and rollovertime ?
    ./imsimta version
    Sun Java(tm) System Messaging Server 7.0-3.01 64bit (built Dec 9 2008)
    libimta.so 7.0-3.01 64bit (built 09:24:13, Dec 9 2008)
    Steve

    SteveHibox wrote:
    Does anybody know how to set up MMP PopProxy* log file size and rollovertime ?Details on these settings are available here:
    http://wikis.sun.com/display/CommSuite6U1/Communications+Suite+6+Update+1+What%27s+New#CommunicationsSuite6Update1What%27sNew-MMPLogging
    Regards,
    Shane.

  • Get Total DB size , Total DB free space , Total Data & Log File Sizes and Total Data & Log File free Sizes from a list of server

    how to get SQL server Total DB size , Total DB free space , Total Data  & Log File Sizes and Total Data  & Log File free Sizes from a list of server 

    Hi Shivanq,
    To get a list of databases, their sizes and the space available in each on the local SQL instance.
    dir SQLSERVER:\SQL\localhost\default\databases | Select Name, Size, SpaceAvailable | ft -auto
    This article is also helpful for you to get DB and Log File size information:
    Checking Database Space With PowerShell
    I hope this helps.

  • SQL LOG FILE SIZE INCREASING

    Hi DBA's
    SQL Log file size occupies more disk space on the server, the overall database size is 8GB
    How to decrease the SQL LDF file size on the server, please explain the suitable steps to perform
    Thanks
    DBA

    use master
    go
    dump transaction <YourDBName>
    with no_log
    go
    use <YourDBName>
    go
    DBCC SHRINKFILE (<YourDBNameLogFileName>,
    100) -- where 100 is the size you may want to shrink it to in MB, change it to your needs
    go
    -- then you can call to check that all went fine
    dbcc checkdb(<YourDBName>)
    Andy ,
    what point in asking user to use No_log and you did not even motioned about what this eveil command will do. Actually its
    seriously not required reason being initial size of log file set to 8 G.
    Plus what is point in running checkdb ?
    I don't agree to any part you pointed
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Log file size

    We have a DNS Server running on solaris 9, it's generating huge logs hence /var/adm/messages file size is vey big. Is there any way to create seperate log file for everyday or can I restrict the log file size for a single file.
    Thank you

    Hmmm,
    For what type environment is this DNS server used for? How many domains/delegated domains are configured on the host?
    I think by default BIND allows 1000 recursive lookup connections. (That is already plenty and if you have that amount of legitimate traffic you will have to add more DNS servers and configure the nodes accordingly)
    Is the server listed as a Name Server for your domain and used externally for name resolution for your domain host entries, maybe the SOA?
    nslookup (enter)
    set type=ns (enter)
    you_domain_mane (i.e. your_domain.com) (enter)Or
    dig �q NS your_domain.com
    If the affected server returns in the list it is NEVER EVER a good idea to allow recursive lookups.
    My guess is that you are subject to denial of service, unless you host a fairly large environment with 1000s of hosts.
    Change the recursive-cient connection back (you system cannot handle 5000 recursive lookups and your system utilization shows this.)
    Then configure
    �category queries { your_query_file; };� In your namd.conf
    restart BIND
    Use �rndc� to change the trace level to 1
    Let it run for 2 -5 min and stop BIND entirely
    Then run something like:
    �cat your_query_file | cut -d'/' -f2 | sort | uniq �c | more� (depends on the log file format, better yet use nwak)
    take a quick look to see if there is one IP that is hammering your system.

  • Archive Log file size

    I am using Oracle database 9.2.0.1.0, My OS is Linux AS4 Update version.
    My database is in archive log mode, the archive file size generated on disk is 100 MB. I want to monitor the reason that why the size of redo generated is too big.
    Kindly suggest.
    Regards

    Archived log file size will always be the same size as redo log or less than the redo log size (but never bigger than redo log size)
    ARCHIVE_LAG_TARGET is the reason (apart from manual archiving ALTER SYSTEM ARCHIVE LOG CURRENT/ALL) why you see archived logs with lesser size than redo log
    why  archive log file size constanly change?

  • Log file size in Sun Directory Server

    Does anyone have an idea about the how the Sun Directory Server's log file size will increase in size with respective to the actions performed?
    Can someone give a data regarding this? If someone has a better scenario and the supportive data w.r.t log file size it will be helpful.
    Thanks,

    AFAIK No its based on time "At a certain time, or after a specified interval, the server rotates your access logs. "
    More info in Archiving Log Files in [http://docs.sun.com/app/docs/doc/820-7985/gczxv?l=en&a=vie]
    It should be easy to write such a script to be run as a daemon in logs directory. Here is the pseudo code :
    while [1]
    do
    get size of the access/error log file
    If size of file > max_size
    <ws-install-dir>/https-<instance>/bin/rotate
    sleep for sometime
    done

  • Managing and configuring log files for Oracle 9ias

    Hi all,
    I'm wondering where I can find documentation on managing and configuring log files like:
    ORACLE_HOME/admin/ sid/*dump/* ORACLE_HOME/assistants/opca/install.log ORACLE_HOME/webcache/logs/*
    ORACLE_HOME/dcm/logs/*
    ORACLE_HOME/ldap/log/*
    ORACLE_HOME/opmn/logs/*
    ORACLE_HOME/sysman/log/*
    ORACLE_HOME/j2ee/ OC4J_instance/log/*/* ORACLE_HOME/config/schemaload.log ORACLE_HOME/config/useinfratool.log
    because I didn't find anything in document like:
    Oracle9 i Application Server
    Administrator�s Guide
    Release 2 (9.0.2)
    May 2002
    Part No. A92171-02
    So, if anyone has any idea...
    Thanks in advance

    Does anyone know how or if it is possible to send the stdout and/or stderr to the log4j type of logging? I already capture the stdout and stderr to the flat file. But I would like to timestamp every line to compare and diagnose problems with each application that encounters problems. Each web app is using log4j to their own application log4j log file. When they encounter errors or resource draining problems, I would like to see the container logs and see what was occuring inside the container at around that same time.
    Any ideas?

Maybe you are looking for

  • Internal Order for AUC Asset not getting created Automatically

    As per SAP Standard , when I create an  AUC Internal Order (KO01), and save , AUC asset is gets created automatically ted via AS01 and stores the AUC Asset in Internal Order settlement rule.  However when tried with the client Im not able to create,

  • Inventory Aging Data, Duplication in SAP B1 8.81

    Hello Everyone, I have been working on Inventory Aging Report creation and has been facing with a problem. The data seems to get duplicated when i execute the report. I have attached the screen shot below for your reference. I have taken the followin

  • Design Reports in Word or Excel

    [url http://www.windwardreports.com/]Windward Reports, the leader in enterprise report generation, has released the world’s first enterprise reporting software that enables use of Excel as a design tool. The Windward Reports Server Engine is perfect

  • Can I import hyperlinks in Power Point 2007 to Captivate?

      I have a presentation in Power Point 2007 I imported into Captivate 6.  Within the Power Point presentaiton, the user cannot navigate through the presentation without clicking one of two objects on the slide.  Each of these objects is hyperlinked t

  • Using xsl:import in Weblogic

    All, I have a requirement to import an XSL template into another XSL template and use one of the methods defined in the imported file into this importing file. The problem is that it works perfectly fine if I do this in XMLSPY .. but if I do this thr