Log file audit script to search and collect

Hi guys,
I'm trying to figure out the best way to complete this log file audit, so I would like to scripted, but can't seem to get a grasp on it and how best to do it. I need to search for the log files (all log files OS and App logs) on a few dozen systems on a
few different drives per system. I'm looking to collect the log location, name, size, last log file event in that log and then export info that to a CSV file and email it to myself monthly to report on.  

Please read the following:
Posting guidelines
Handy tips for posting to this forum
How to ask questions in a technical forum
Rubber duck problem solving
How to write a bad forum post
Help Vampires: A Spotter's Guide
-- Bill Stewart [Bill_Stewart]

Similar Messages

  • Script to search and relink linked image files that were moved?

    Good afternoon
    My indesign product catalog has links with a lot of different folders around in my hard disk, not a single folder. And to keep my catalog up to date I must package it.
    When I move my linked files in my computer because I change the folder structure or do some cleaning, my indd loses track of its linked files. Currently I need to relink one by one.
    Therefore I would like to find a script which would scan the hard disk with the all broken links in mind to detect their new location and relink. Yes I am a lazy person.
    Thanks for any clue,

    Thanks so much! I will try soonest.
    How does your script react in the unlikely case of duplicates?
    Date: Thu, 18 Jun 2009 12:43:47 -0600
    From: [email protected]
    To: [email protected]
    Subject: Script to search and relink linked image files that were moved?
    At least your honest about your laziness.... Here's your reward:
    Just a hint, don't select root of your volume, it will loop through every folder in existence, but key to this script is, if you know the general location, select into that folder hierarchy and the script will test for a relative path, otherwise, come back in a few days, and it will be finished.
    ~mike
    var processed = 0
    var skipped = 0
    var updated = 0
    if (app.documents.length > 0){
        if(app.activeDocument.links.length > 0){
            var mydoc = app.activeDocument;
            var mylinks = mydoc.links;
            var myRoot = Folder.selectDialog("Choose the volume or server where assets are located", undefined, false);
            main();
    else
    { alert("No Links present")
    else{alert("No Documents Open")
    function main(){
        if(myRoot != null){
            for(var i = 0; i < mylinks.length ; i++){
                if(mylinks.item(i).status == LinkStatus.linkMissing){
                        var linkdata= mylinks.item(i).filePath;
                        var my_result = linkRepair(linkdata);
                            if (my_result == false){
                                var filetype = "." + mylinks.item(i).linkType;
                                var mysearch =   search(linkdata, filetype);
                                    if(mysearch != undefined){
    var myswitch = confirm(mylinks.item(i).name + " has been found in a different location, Relink and Update?", "Relinker")
                                            if(myswitch){
                                                mylinks.item(i).relink(File(mysearch));
                                                mylinks.item(i).update();
                                else{
                                    alert("" + mylinks.item(i).name " was not found\nFolders processed: "processed + "\nFiles skipped: "+ skipped)
                            else{ alert( my_result + " has been found");
                                mylinks.item(i).relink(File(my_result));
    function linkRepair(linkdata){
    var mypath = linkdata.split(":")
            my_status = analyzePath(mypath)
            if(my_status == true){alert("UPDATED");
                    updated++       
        return my_status
    function analyzePath(mypath){
    var num = mypath.length ;
    for(var i = 0; i < num-1; i++){
        mypath.shift()
        var newpath = pathRebuild(mypath)
        newpath = (myRoot + newpath)
            if(File(newpath).exists){
                return newpath
        return false
    function pathRebuild(pathArray){
    var solidPath = ""
        for(var i = 0; i < pathArray.length ; i++){
            solidPath +=  "/" + pathArray+ ;
        return solidPath;
    function search(linkdata, filetype){
        var mypath = linkdata.split(":")
        var mylink = mypath.pop();
        var OK = confirm("Relative Path does not exist, search folder hiearchy for file?", true, "File Scanner")
        if(OK){
    var myscanresult = getfiles(mylink, myRoot)
    return myscanresult
    function getfiles(mylink, myBase){
    myBase = Folder(myBase);
    var files = myBase.getFiles("*")
    for(var i = 0; i < files.length; i++){
    try{
    var foldertest = files+.getFiles();
    processed++
    var myfile =     File(files+ + "/" + mylink)
        if(myfile.exists == true){
            return myfile
            break;
        else{
        var myscan = getfiles(mylink, Folder(files+))    
            if(myscan != undefined){
            return myscan
            break;
    catch(myerror){
        skipped++
    >

  • 45 min long session of log file sync waits between 5000 and 20000 ms

    45 min long log file sync waits between 5000 and 20000 ms
    Encountering a rather unusual performance issue. Once every 4 hours I am seeing a 45 minute long log file sync wait event being reported using Spotlight on Oracle. For the first 30 minutes the event wait is for approx 5000 ms, followed by an increase to around 20000 ms for the next 15 min before rapidly dropping off and normal operation continues for the next 3 hours and 15 minutes before the cycle repeats itself. The issue appears to maintain it's schedule independently of restarting the database. Statspack reports do not show an increase in commits or executions or any new sql running during the time the issue is occuring. We have two production environments both running identicle applications with similar usage and we do not see the issue on the other system. I am leaning towards this being a hardware issue, but the 4 hour interval regardless of load on the database has me baffled. If it were a disk or controller cache issue one would expect to see the interval change with database load.
    I cycle my redo logs and archive them just fine with log file switches every 15-20 minutes. Even during this unusally long and high session of log file sync waits I can see that the redo log files are still switching and are being archived.
    The redo logs are on a RAID 10, we have 4 redo logs at 1 GB each.
    I've run statspack reports on hourly intervals around this event:
    Top 5 Wait Events
    ~~~~~~~~~~~~~~~~~ Wait % Total
    Event Waits Time (cs) Wt Time
    log file sync 756,729 2,538,034 88.47
    db file sequential read 208,851 153,276 5.34
    log file parallel write 636,648 129,981 4.53
    enqueue 810 21,423 .75
    log file sequential read 65,540 14,480 .50
    And here is a sample while not encountering the issue:
    Top 5 Wait Events
    ~~~~~~~~~~~~~~~~~ Wait % Total
    Event Waits Time (cs) Wt Time
    log file sync 953,037 195,513 53.43
    log file parallel write 875,783 83,119 22.72
    db file sequential read 221,815 63,944 17.48
    log file sequential read 98,310 18,848 5.15
    db file scattered read 67,584 2,427 .66
    Yes I know I am already tight on I/O for my redo even during normal operations yet, my redo and archiving works just fine for 3 hours and 15 minutes (11 to 15 log file switches). These normal switches result in a log file sync wait of about 5000 ms for about 45 seconds while the 1GB redo log is being written and then archived.
    I welcome any and all feedback.
    Message was edited by:
    acyoung1
    Message was edited by:
    acyoung1

    Lee,
    log_buffer = 1048576 we use a standard of 1 MB for our buffer cache, we've not altered the setting. It is my understanding that Oracle typically recommends that you not exceed 1MB for the log_buffer, stating that a larger buffer normally does not increase performance.
    I would agree that tuning the log_buffer parameter may be a place to consider; however, this issue last for ~45 minutes once every 4 hours regardless of database load. So for 3 hours and 15 minutes during both peak usage and low usage the buffer cache, redo log and archival processes run just fine.
    A bit more information from statspack reports:
    Here is a sample while the issue is occuring.
    Snap Id Snap Time Sessions
    Begin Snap: 661 24-Mar-06 12:45:08 87
    End Snap: 671 24-Mar-06 13:41:29 87
    Elapsed: 56.35 (mins)
    Cache Sizes
    ~~~~~~~~~~~
    db_block_buffers: 196608 log_buffer: 1048576
    db_block_size: 8192 shared_pool_size: 67108864
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 615,141.44 2,780.83
    Logical reads: 13,241.59 59.86
    Block changes: 2,255.51 10.20
    Physical reads: 144.56 0.65
    Physical writes: 61.56 0.28
    User calls: 1,318.50 5.96
    Parses: 210.25 0.95
    Hard parses: 8.31 0.04
    Sorts: 16.97 0.08
    Logons: 0.14 0.00
    Executes: 574.32 2.60
    Transactions: 221.21
    % Blocks changed per Read: 17.03 Recursive Call %: 26.09
    Rollback per transaction %: 0.03 Rows per Sort: 46.87
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.99 Redo NoWait %: 100.00
    Buffer Hit %: 98.91 In-memory Sort %: 100.00
    Library Hit %: 98.89 Soft Parse %: 96.05
    Execute to Parse %: 63.39 Latch Hit %: 99.87
    Parse CPU to Parse Elapsd %: 90.05 % Non-Parse CPU: 85.05
    Shared Pool Statistics Begin End
    Memory Usage %: 89.96 92.20
    % SQL with executions>1: 76.39 67.76
    % Memory for SQL w/exec>1: 72.53 63.71
    Top 5 Wait Events
    ~~~~~~~~~~~~~~~~~ Wait % Total
    Event Waits Time (cs) Wt Time
    log file sync 756,729 2,538,034 88.47
    db file sequential read 208,851 153,276 5.34
    log file parallel write 636,648 129,981 4.53
    enqueue 810 21,423 .75
    log file sequential read 65,540 14,480 .50
    And this is a sample during "normal" operation.
    Snap Id Snap Time Sessions
    Begin Snap: 671 24-Mar-06 13:41:29 88
    End Snap: 681 24-Mar-06 14:42:57 88
    Elapsed: 61.47 (mins)
    Cache Sizes
    ~~~~~~~~~~~
    db_block_buffers: 196608 log_buffer: 1048576
    db_block_size: 8192 shared_pool_size: 67108864
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 716,776.44 2,787.81
    Logical reads: 13,154.06 51.16
    Block changes: 2,627.16 10.22
    Physical reads: 129.47 0.50
    Physical writes: 67.97 0.26
    User calls: 1,493.74 5.81
    Parses: 243.45 0.95
    Hard parses: 9.23 0.04
    Sorts: 18.27 0.07
    Logons: 0.16 0.00
    Executes: 664.05 2.58
    Transactions: 257.11
    % Blocks changed per Read: 19.97 Recursive Call %: 25.87
    Rollback per transaction %: 0.02 Rows per Sort: 46.85
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.99 Redo NoWait %: 100.00
    Buffer Hit %: 99.02 In-memory Sort %: 100.00
    Library Hit %: 98.95 Soft Parse %: 96.21
    Execute to Parse %: 63.34 Latch Hit %: 99.90
    Parse CPU to Parse Elapsd %: 96.60 % Non-Parse CPU: 84.06
    Shared Pool Statistics Begin End
    Memory Usage %: 92.20 88.73
    % SQL with executions>1: 67.76 75.40
    % Memory for SQL w/exec>1: 63.71 68.28
    Top 5 Wait Events
    ~~~~~~~~~~~~~~~~~ Wait % Total
    Event Waits Time (cs) Wt Time
    log file sync 953,037 195,513 53.43
    log file parallel write 875,783 83,119 22.72
    db file sequential read 221,815 63,944 17.48
    log file sequential read 98,310 18,848 5.15
    db file scattered read 67,584 2,427 .66

  • Log file is growing too big and too quick on ECC6 EHP4 system

    Hello there,
    i have ECC6EHP6 system based on NW 7.01 and MSSQL as back-end.
    My log file is growing very fast. i seen a number of threads but does not got some perfect idea what to do. Please see DEV_RFC0 and DEV_W0 log file excerpt below. Please help
    DEV_RFC0
    Trace file opened at 20100618 065308 Eastern Daylight Time, SAP-REL 701,0,14 RFC-VER U 3 1016574 MT-SL
    ======> CPIC-CALL: 'ThSAPOCMINIT' : cmRc=2 thRc=679
    Transaction program not registered                                      
    ABAP Programm: SAPLSADC (Transaction: )
    User: DDIC (Client: 000)
    Destination: SAPDB_DBM_DAEMON (handle: 2, , )
    SERVER> RFC Server Session (handle: 1, 59356166, {ADC77ADF-C575-F1CC-B797-0019B9E204CC})
    SERVER> Caller host:
    SERVER> Caller transaction code:  (Caller Program: SAPLSADC)
    SERVER> Called function module: DBM_CONNECT_PUR
    Error RFCIO_ERROR_SYSERROR in abrfcpic.c : 1501
    CPIC-CALL: 'ThSAPOCMINIT' : cmRc=2 thRc=679
    Transaction program not registered                                      
    DEST =SAPDB_DBM_DAEMON
    HOST =%%RFCSERVER%%
    PROG =dbmrfc@sapdb
    Trace file opened at 20100618 065634 Eastern Daylight Time, SAP-REL 701,0,14 RFC-VER U 3 1016574 MT-SL
    ======> CPIC-CALL: 'ThSAPOCMINIT' : cmRc=2 thRc=679
    Transaction program not registered                                      
    ABAP Programm: SAPLSADC (Transaction: )
    User: DDIC (Client: 000)
    Destination: SAPDB_DBM_DAEMON (handle: 2, , )
    SERVER> RFC Server Session (handle: 1, 59587535, {28C87ADF-4577-F16D-B797-0019B9E204CC})
    SERVER> Caller host:
    SERVER> Caller transaction code:  (Caller Program: SAPLSADC)
    SERVER> Called function module: DBM_CONNECT_PUR
    Error RFCIO_ERROR_SYSERROR in abrfcpic.c : 1501
    CPIC-CALL: 'ThSAPOCMINIT' : cmRc=2 thRc=679
    Transaction program not registered                                      
    DEST =SAPDB_DBM_DAEMON
    HOST =%%RFCSERVER%%
    PROG =dbmrfc@sapdb
    DEV_W0

    X Fri Jun 18 18:05:15 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    N Fri Jun 18 18:05:25 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    X Fri Jun 18 18:05:39 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    X Fri Jun 18 18:05:46 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    X Fri Jun 18 18:05:52 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    N Sun Jun 20 10:55:32 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    X Sun Jun 20 11:00:02 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    N Sun Jun 20 11:00:32 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    C Sun Jun 20 11:03:34 2010
    C  Thread ID:1956
    C  dbmssslib.dll patch info
    C    patchlevel   0
    C    patchno      13
    C    patchcomment Errors when running with par_stmt_prepare set to zero (1253696)
    C  Local connection used on SAPFIVE to named instance: np:SAPFIVE\ECT

    C Sun Jun 20 11:03:49 2010
    C  OpenOledbConnection: line 23839. hr: 0x8000ffff Login timeout expired
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Login timeout expired
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), An error has occurred while establishing a connection to the server. When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server does not allow remote connections.
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), SQL Network Interfaces: Server doesn't support requested protocol [xFFFFFFFF].
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Invalid connection string attribute
    C  Procname: [OpenOledbConnection - no proc]

    C Sun Jun 20 11:04:04 2010
    C  OpenOledbConnection: line 23839. hr: 0x8000ffff Login timeout expired
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Login timeout expired
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), An error has occurred while establishing a connection to the server. When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server does not allow remote connections.
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), SQL Network Interfaces: Server doesn't support requested protocol [xFFFFFFFF].
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Invalid connection string attribute
    C  Procname: [OpenOledbConnection - no proc]

    C Sun Jun 20 11:04:19 2010
    C  OpenOledbConnection: line 23839. hr: 0x8000ffff Login timeout expired
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Login timeout expired
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), An error has occurred while establishing a connection to the server. When connecting to SQL Server 2005, this failure may be caused by the fact that under the default settings SQL Server does not allow remote connections.
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err -1, sev 0), SQL Network Interfaces: Server doesn't support requested protocol [xFFFFFFFF].
    C  Procname: [OpenOledbConnection - no proc]
    C  sloledb.cpp [OpenOledbConnection,line 23839]: Error/Message: (err 0, sev 0), Invalid connection string attribute
    C  Procname: [OpenOledbConnection - no proc]
    C  failed to establish conn to np:SAPFIVE\ECT.
    C  Retrying without protocol specifier: SAPFIVE\ECT
    C  Connected to db server : [SAPFIVE\ECT] server_used : [SAPFIVE\ECT], dbname: ECT, dbuser: ect
    C  pn_id:SAPFIVE_ECT_ECTECT_ECT
    B  Connection 4 opened (DBSL handle 2)
    B  Wp  Hdl ConName          ConId     ConState     TX  PRM RCT TIM MAX OPT Date     Time   DBHost         
    B  000 000 R/3              000000000 ACTIVE       YES YES NO  000 255 255 20100618 061458 SAPFIVE\ECT    
    B  000 001 +DBO+0050      000002156 INACTIVE     NO  NO  NO  004 255 255 20100618 101017 SAPFIVE\ECT    
    B  000 002 +DBO+0050      000001981 DISCONNECTED NO  NO  NO  000 255 255 20100620 070023 SAPFIVE\ECT    
    B  000 003 +DBO+0050      000001982 DISCONNECTED NO  NO  NO  000 255 255 20100620 070023 SAPFIVE\ECT    
    B  000 004 R/3*INACT_PACK   000002157 ACTIVE       NO  NO  NO  004 255 255 20100620 110334 SAPFIVE\ECT  
    N Sun Jun 20 17:35:35 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    M Sun Jun 20 17:47:29 2010
    M  ThAlarmHandler (1)
    M  ThAlarmHandler (1)
    M  ThAlarmHandler: set CONTROL_TIMEOUT/DP_CONTROL_JAVA_EXIT and break sql

    C Sun Jun 20 17:47:33 2010
    C  SQLBREAK: DBSL_CMD_SQLBREAK: CbOnCancel was not set. rc: 15
    M  program canceled
    M    reason   = max run time exceeded
    M    user     = SAPSYS     
    M    client   = 000
    M    terminal =                    
    M    report   = SAPMSSY2                               
    M  ThAlarmHandler: return from signal handler

    A Sun Jun 20 17:48:33 2010
    A  TH VERBOSE LEVEL FULL
    A  ** RABAX: level LEV_RX_PXA_RELEASE_MTX entered.
    A  ** RABAX: level LEV_RX_PXA_RELEASE_MTX completed.
    A  ** RABAX: level LEV_RX_COVERAGE_ANALYSER entered.
    A  ** RABAX: level LEV_RX_COVERAGE_ANALYSER completed.
    A  ** RABAX: level LEV_RX_ROLLBACK entered.
    A  ** RABAX: level LEV_RX_ROLLBACK completed.
    A  ** RABAX: level LEV_RX_DB_ALIVE entered.
    A  ** RABAX: level LEV_RX_DB_ALIVE completed.
    A  ** RABAX: level LEV_RX_HOOKS entered.
    A  ** RABAX: level LEV_RX_HOOKS completed.
    A  ** RABAX: level LEV_RX_STANDARD entered.
    A  ** RABAX: level LEV_RX_STANDARD completed.
    A  ** RABAX: level LEV_RX_STOR_VALUES entered.
    A  ** RABAX: level LEV_RX_STOR_VALUES completed.
    A  ** RABAX: level LEV_RX_C_STACK entered.

    A Sun Jun 20 17:48:42 2010
    A  ** RABAX: level LEV_RX_C_STACK completed.
    A  ** RABAX: level LEV_RX_MEMO_CHECK entered.
    A  ** RABAX: level LEV_RX_MEMO_CHECK completed.
    A  ** RABAX: level LEV_RX_AFTER_MEMO_CHECK entered.
    A  ** RABAX: level LEV_RX_AFTER_MEMO_CHECK completed.
    A  ** RABAX: level LEV_RX_INTERFACES entered.
    A  ** RABAX: level LEV_RX_INTERFACES completed.
    A  ** RABAX: level LEV_RX_GET_MESS entered.
    A  ** RABAX: level LEV_RX_GET_MESS completed.
    A  ** RABAX: level LEV_RX_INIT_SNAP entered.
    A  ** RABAX: level LEV_RX_INIT_SNAP completed.
    A  ** RABAX: level LEV_RX_WRITE_SYSLOG entered.
    A  ** RABAX: level LEV_RX_WRITE_SYSLOG completed.
    A  ** RABAX: level LEV_RX_WRITE_SNAP_BEG entered.
    A  ** RABAX: level LEV_RX_WRITE_SNAP_BEG completed.
    A  ** RABAX: level LEV_RX_WRITE_SNAP entered.

    A Sun Jun 20 17:48:48 2010
    A  ** RABAX: level LEV_SN_END completed.
    A  ** RABAX: level LEV_RX_WRITE_SNAP_END entered.
    A  ** RABAX: level LEV_RX_WRITE_SNAP_END completed.
    A  ** RABAX: level LEV_RX_SET_ALERT entered.
    A  ** RABAX: level LEV_RX_SET_ALERT completed.
    A  ** RABAX: level LEV_RX_COMMIT entered.
    A  ** RABAX: level LEV_RX_COMMIT completed.
    A  ** RABAX: level LEV_RX_SNAP_SYSLOG entered.
    A  ** RABAX: level LEV_RX_SNAP_SYSLOG completed.
    A  ** RABAX: level LEV_RX_RESET_PROGS entered.
    A  ** RABAX: level LEV_RX_RESET_PROGS completed.
    A  ** RABAX: level LEV_RX_STDERR entered.
    A  Sun Jun 20 17:48:48 2010

    A  ABAP Program SAPMSSY2                                .
    A  Source RSBTCTRC                                 Line 131.
    A  Error Code TIME_OUT.
    A  Module  $Id: //bas/701_REL/src/krn/runt/abinit.c#1 $ SAP.
    A  Function ab_chstat Line 1941.
    A  ** RABAX: level LEV_RX_STDERR completed.
    A  ** RABAX: level LEV_RX_RFC_ERROR entered.
    A  ** RABAX: level LEV_RX_RFC_ERROR completed.
    A  ** RABAX: level LEV_RX_RFC_CLOSE entered.
    A  ** RABAX: level LEV_RX_RFC_CLOSE completed.
    A  ** RABAX: level LEV_RX_IMC_ERROR entered.
    A  ** RABAX: level LEV_RX_IMC_ERROR completed.
    A  ** RABAX: level LEV_RX_DATASET_CLOSE entered.
    A  ** RABAX: level LEV_RX_DATASET_CLOSE completed.
    A  ** RABAX: level LEV_RX_RESET_SHMLOCKS entered.
    A  ** RABAX: level LEV_RX_RESET_SHMLOCKS completed.
    A  ** RABAX: level LEV_RX_ERROR_SAVE entered.
    A  ** RABAX: level LEV_RX_ERROR_SAVE completed.
    A  ** RABAX: level LEV_RX_ERROR_TPDA entered.
    A  ** RABAX: level LEV_RX_ERROR_TPDA completed.
    A  ** RABAX: level LEV_RX_PXA_RELEASE_RUDI entered.
    A  ** RABAX: level LEV_RX_PXA_RELEASE_RUDI completed.
    A  ** RABAX: level LEV_RX_LIVE_CACHE_CLEANUP entered.
    A  ** RABAX: level LEV_RX_LIVE_CACHE_CLEANUP completed.
    A  ** RABAX: level LEV_RX_END entered.
    A  ** RABAX: level LEV_RX_END completed.
    A  ** RABAX: end no http/smtp
    A  ** RABAX: end RX_BTCHLOG|RX_VBLOG
    A  Time limit exceeded..


    X Sun Jun 20 17:49:03 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]
    N Sun Jun 20 18:35:40 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    N Sun Jun 20 18:50:35 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    N Sun Jun 20 18:55:31 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    N Sun Jun 20 19:00:31 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.

    X Sun Jun 20 19:01:59 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    X Sun Jun 20 19:02:05 2010
    X  *** ERROR => EmActiveData: Invalid Context Handle -1 [emxx.c       2214]

    M Sun Jun 20 19:04:02 2010
    M  ***LOG R49=> ThReceive, CPIC-Error (020223) [thxxhead.c   7488]
    M  ***LOG R5A=> ThReceive, CPIC-Error (76495728) [thxxhead.c   7493]
    M  ***LOG R64=> ThReceive, CPIC-Error ( CMSEND(SAP)) [thxxhead.c   7498]

    N Sun Jun 20 19:05:31 2010
    N  RSEC: The entry with identifier /RFC/DIICLNT800
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    N  RSEC: The entry with identifier /RFC/T90CLNT090
    N  was encrypted by a system
    N  with different SID and cannot be decrypted here.
    Thanks
    Mani

    i have few BG jobs running at backend, but i dont think that could be a problem. I have these jobs running
    ESH_IX_PROCESS_CP_20100621051308
    EU_PUT
    EU_REORG
    RPTMC_CREATE_CHANGEPOINT_AUTH
    i cancelled  the last one and first job is for trex indexing for changed object i SAP hR master data. middle two i dont know the purpose.
    do you think these could be problem?
    Mani

  • I can't able to open the *.log files from the firefox browser and i need to open the file inside the frame

    If i open the log file from firefox browser it is not open and throws an error as
    "The address wasn't understood
    Firefox doesn't know how to open this address, because one of the following protocols (e) isn't associated with any program or is not allowed in this context."

    What type of log file?
    What is the file name?
    What is the file path or URI?

  • Script to Search and Print

    Hi.
    I would like to automate the search of a pdf document and the subsequent printing (or extraction) of all the pages on which the searched text appears.  I think a script would be appropriate but I have no idea where to begin,  Can someone help me with this?  I deal routinely with large, mostly text pdf documents where I need to print many non-sequential pages.
    Thanks

    this piece of script will help you get there:
    [code]
    set search_for to "part" ---search for this name
    set search_here to POSIX path of (path to desktop folder) as string ---- too much to handle
    set search_here to "/Users/timmason/Desktop/spotlight scripts/csacsacs/alta__001 copy 2.pdf" ------file to search -not working
    set search_here to "/Users/timmason/Desktop/spotlight scripts/csacsacs/" ------folder to search
    display dialog searcher(search_for, search_here) ---returns location of doc
    on searcher(search_for, search_here)
    set the_script to {"mdfind -onlyin '", search_here, "' '", search_for, "'"} as string --converted for vars
    set the_text to (do shell script the_script)
    set num_records to count of paragraphs in the_text
    set x to 0
    set all_recs to ""
    repeat num_records times
    set x to x + 1
    set all_recs to all_recs & return & return & paragraph x of the_text
    end repeat
    return all_recs
    end searcher

  • Apple script to search and replace in notes

    I'd much appreciate such script or any suggestions as how to achieve that. I have hundreds of notes and decided to change my tagging system.
    Thanks a loT!

    Well, color me confused.  But if you're just concerned about Mail notes, here's a script that uses the example you gave above.
    if application "Mail" is running then
      -- mail should not be running when you use this script
              display alert "Whoops!" message "Please quit Mail.app before running this script" as critical
              return
    end if
    tell application "Finder"
      -- locate the correct mail folder for the given account
              try
                        set accountFolder to folder "[email protected]" of folder "V2" of folder "Mail" of folder "Library" of home
              on error
      -- folders changed from snow leopard to lion, check the folder hierarchy visually
                        display alert "Whoops!" message "something's wrong with the folder hierarchy." as critical
                        open folder "Mail" of folder "Library" of home
                        return
              end try
              tell accountFolder
                        with timeout of 600 seconds
      -- gather notes file references. set a long timeout because this can take a while if there are a large number of notes
                                  set theNotes to (every file of (entire contents of folder "Notes.mbox") whose name extension is "emlx") as alias list
                        end timeout
              end tell
    end tell
    -- set up keywords to be swapped out here.
    set swapPairs to {{"#essay", "kbEssay"}, {"#another key", "its replacement"}}
    repeat with thisNote in theNotes
      -- open and read the emlx file into a variable
              set fp to open for access thisNote with write permission
              set noteHTMLText to read fp
      -- loop through the swap pairs making replacements
              repeat with thisPair in swapPairs
                        set noteParts to tid(noteHTMLText, first item of thisPair)
                        set noteHTMLText to tid(noteParts, second item of thisPair)
              end repeat
      -- zero out the file and replace with the revised text
              set eof fp to 0
              write noteHTMLText to fp
              close access fp
    end repeat
    on tid(input, delim)
      -- generic text items handler
              set {oldTID, my text item delimiters} to {my text item delimiters, delim}
              if class of input is list then
                        set output to input as text
              else
                        set output to text items of input
              end if
              set my text item delimiters to oldTID
              return output
    end tid
    I've tested this on my laptop and it works, but I STRONGLY suggest that you back up your mail folder before trying it yourself, and (if possible) that you set up a few notes in a little used or dummy account and test it before you run it on your actual notes. 
    I also make no guarantees about what will happen if these notes are synched to other devices.  It probably won't explode your computer, but...

  • Script to search all files in specified folder for multiple string text values listed in a source file and output each match to one single results txt file

    I have been searching high and low for this one.  I have a vbscript that can successfully perform the function if one file is listed.  It does a Wscript.echo on the results and if I run this via command using cscript, I can output to a text file
    that way.  However, I cannot seem to get it to work properly if I want it to search ALL the files in the folder.  At one point, I was able to have it create the output file and appear as if it worked, but it never showed any results when the script
    was executed and folder was scanned.  So I am going back to the drawing board and starting from the beginning.
    I also have a txt file that contains the list of string text entries I would like it to search for.  Just for testing, I placed 4 lines of sample text and one single matching text in various target files and nothing comes back.  The current script
    I use for each file has been executed with a few hundred string text lines I want it to search against to well over one thousand.  It might take awhile, but it works every time. The purpose is to let this run against various log files in a folder and
    let it search.  There is no deleting, moving, changing of either the target folder/files to run against, nor of the file that contains the strings to search for.  It is a search (read) only function, going thru the entire contents of the folder and
    when done, performs the loop function and onto the next file to repeat the process until all files are searched.  When completed, instead of running a cscript to execute the script and outputting the results to text, I am trying to create that as part
    of the overall script.  Saving yet another step for me to do.
    My current script is set to append to the same results file and will echo [name of file I am searching]:  No errors found.  Otherwise, the
    output shows the filename and the string text that matched.  Because the results append to it, I can only run the script against each file separately or create individual output names.  I would rather not do that if I could include it all in one.
     This would also free me from babysitting it and running each file script separately upon the other's completion.  I can continue with my job and come back later and view the completed report all in one.  So
    if I could perform this on an entire folder, then I would want the entries to include the filename, the line number that the match occurred on in that file and the string text that was matched (each occurrence).  I don't want the entire line to be listed
    where the error was, just the match itself.
    Example:  (In the event this doesn't display correctly below, each match, it's corresponding filename and line number all go together on the same line.  It somehow posted the example jumbled when I listed it) 
    File1.txt Line 54 
    Job terminated unexpectedly
     File1.txt Line 58 Process not completed
    File1.txt
    Line 101 User input not provided
    File1.txt
    Line 105  Process not completed
    File2.txt
    No errors found
    File3.txt
    Line 35 No tape media found
    File3.txt
    Line 156 Bad surface media
    File3.txt Line 188
    Process terminated
    Those are just random fake examples for this post.
    This allows me to perform analysis on a set of files for various projects I am doing.  Later on, when the entire search is completed, I can go back to the results file and look and see what files had items I wish to follow up on.  Therefore, the
    line number that each match was found on will allow me to see the big picture of what was going on when the entry was logged.
    I actually import the results file into a spreadsheet, where further information is stored regarding each individual text string I am using.  Very useful.
    If you know how I can successfully achieve this in one script, please share.  I have seen plenty of posts out there where people have requested all different aspects of it, but I have yet to see it all put together in one and work successfully.
    Thanks for helping.

    I'm sorry.  I was so consumed in locating the issue that I completely overlooked posting what exactly I was needing  help with.   I did have one created, but I came across one that seemed more organized than what I originally created.  Later
    on I would learn that I had an error in log location on my original script and therefore thought it wasn't working properly.  Now that I am thinking that I am pretty close to achieving what I want with this one, I am just going to stick with it.
    However, I could still use help on it.  I am not sure what I did not set correctly or perhaps overlooking as a typing error that my very last line of this throws an "Expected Statement" error.  If I end with End, then it still gives same
    results.
    So to give credit where I located this:
    http://vbscriptwmi.uw.hu/ch12lev1sec7.html
    I then adjusted it for what I was doing.
    What this does does is it searches thru log files in a directory you specify when prompted.  It looks for words that are contained in another file; objFile2, and outputs the results of all matching words in each of those log files to another file:  errors.log
    Once all files are scanned to the end, the objects are closed and then a message is echoed letting you know (whether there errors found or not), so you know the script has been completed.
    What I had hoped to achieve was an output to the errors.log (when matches were found) the file name, the line number that match was located on in that file and what was the actual string text (not the whole line) that matched.  That way, I can go directly
    to each instance for particular events if further analysis is needed later on.
    So I could use help on what statement should I be closing this with.  What event, events or error did I overlook that I keep getting prompted for that.  Any help would be appreciated.
    Option Explicit
    'Prompt user for the log file they want to search
    Dim varLogPath
    varLogPath = InputBox("Enter the complete path of the logs folder.")
    'Create filesystem object
    Dim oFSO
    Set oFSO = WScript.CreateObject("Scripting.FileSystemObject")
    'Creates the output file that will contain errors found during search
    Dim oTSOut
    Set oTSOut = oFSO.CreateTextFile("c:\Scripts\errors.log")
    'Loop through each file in the folder
    Dim oFile, varFoundNone
    VarFoundNone = True
    For Each oFile In oFSO.GetFolder(varLogPath).Files
        'Verifies files scanned are log files
        If LCase(Right(oFile.Name,3)) = "log" Then
            'Open the log file
            Dim oTS
            oTS = oFSO.OpenTextFile(oFile.Path)
            'Sets the file log that contains error list to look for
            Dim oFile2
            Set oFile2 = oFSO.OpenTextFile("c:\Scripts\livescan\lserrors.txt", ForReading)
            'Begin reading each line of the textstream
            Dim varLine
            Do Until oTS.AtEndOfStream
                varLine = oTS.ReadLine
                Set objRegEx = CreateObject("VBScript.RegExp")
                objRegEx.Global = True  
                Dim colMatches, strName, strText
                Do Until oErrors.AtEndOfStream
                    strName = oFile2.ReadLine
                    objRegEx.Pattern = ".{0,}" & strName & ".{0,}\n"
                    Set colMatches = objRegEx.Execute(varLine)  
                    If colMatches.Count > 0 Then
                        For Each strMatch in colMatches 
                            strText = strText & strMatch.Value
                            WScript.Echo "Errors found."
                            oTSOut.WriteLine oFile.Name, varLine.Line, varLine
                            VarFoundNone = False
                        Next
                    End If
                Loop
                oTS.Close
                oFile2.Close
                oTSOut.Close
                Exit Do
                If VarFoundNone = True Then
                    WScript.Echo "No errors found."
                Else
                    WScript.Echo "Errors found.  Check logfile for more info."
                End If
        End if

  • Script for generate randomize administrator password which make log file to recorded new administrator password with associated computer name on it

    Hi
    I Need VDS script in order to change domain client local administrator password in my domain ,and put this script in startup script via group policy, but for security purpose  I want to randomize local administrator password and log new password
    set for each computer on a text file, I want to over write the old password of eachcomputer in log file with new one in order to have the update log file ,my support team some times need administrator password for troubleshooting.
    I need a script for generate randomize administrator password  which make log file to recorded new administrator password with associated computer name on it and each time new administrator password set it over write the old record on
    the log file and update the content of log file automatically.
    Regards

    Hi
    I need a script for generate randomize administrator password  which record new password on a  log file with associated computer name  and each time new administrator password set for a computer it  over write the old record
    on the log file and update the content of log file automatically.
    Regards

  • Dataguard lost both Primary redo log and standby redo log files

    Hi,
    I am new to data guard, i came acorss a scenario where we loose both primary redo log file and standby redo log files.
    Can someone please help me understand how to recover from this situation.
    Thanks!

    >loose both primary redo log file and standby redo log files
    We have to be very clear.
    There are (set A) online redo log files  and (set B) standby redo log files at (location 1) Primary and (location 2) Standby.
    The standby redo log files, depending on the configuration, aren't strictly mandatory.  The standby can be applying redo without online redo log files present as well, depending on how it was setup.
    So, the question is  : Did you lose online redo log files at the primary ?  Didn't the primary shutdown itself then ? If so, you have to do an incomplete recovery at the primary OR switch over to the standby (which may or may not have received the last transaction, depending on how it was configured and operating)   OR restore from the standby (again, with possible loss of transactions) to the primary.
    Hemant K Chitale

  • VMX20 - Limit virtual machine log file size and number

    Hi all,
    Here is parameter file
    log.rotateSize=1000000
    log.keepOld=10
    that I have apply on virtual machine VMX. After I run the compliance checker again and it still showing fail.
    I also have try restart the virtual machine machine few time but the log file keep increase.
    For my understanding if we apply this parameter setting once the virtual machine log file over 1,000kb or 10 log file the new log file will replace the old one, correct me if Ii wrong.
    Here is my question
    1. Did the parameter file have restrict on any version of VMware?
    2. If the log file less than 1,000kb and it already have 10 log file, so the new log file will still replace the old log file?
    Thk.
    Best regards.
    Wong Pak Lian

    Hi a_nut_in,
    The problem have been solved in VMware vsphere version 5 with compliance checker for vsphere 5.0.
    But under VMware vsphere version 4 I have found out the parameter is ok and workable but the results is still showing fail, will this is software bug?
    Thk.
    Best regards.
    Wong Pak Lian

  • Log file size is huge and cannot shrink.

    I have a live database that is published with merge replication and a couple of push subscriptions. I just noticed that the log file has grown to 500GB. The database is in full recovery. We do weekly full backups, daily log backups. I cannot shrink the log
    file back down to normal proportions. It should only be about 5GB. The file properties show an initial size equal to the current size, I cannot change that number and I don't know why it is so big now? How do I go about shrinking the log file. the normal DBCC
    shrink and the SSMS GUI to shrink are not doing anything and say there is 0MB space free!!

    As per your first posing log_reuse_wait_desc was LOG_BACKUP
    and in 2nd REPLICATION . i am consfused
    if
    log_reuse_wait_desc  column
    shows LOG_BACKUP then table log backup of your db, if it is REPLICATION 
    and you are sure that your replications are in sync then reset the status of replicated transactions
    You can reset this by first turning the Reader agent off ( turn the whole SQL Server Agent off), and then run that query on the database for which you want to fix the replication issue:
    EXEC sp_repldone @xactid = NULL, @xact_segno = NULL, @numtrans = 0, @time= 0, @reset = 1
    vt
    Please mark answered if I've answered your question and vote for it as helpful to help other user's find a solution quicker

  • Shrink Log file in log shipping and change the database state from Standby to No recovery mode

    Hello all,
    I have configured sql server 2008 R2 log shipping for some databases and I have two issues:
    can I shrink the log file for these databases: If I change the primary database from full to simple and shrink the log file then change it back to full recovery mode the log shipping will fail, I've seen some answers talked about using "No
    Truncate" option, but as I know this option will not affect the log file and it will shrink the data file only.
          I also can't create maintenance to reconfigure the log shipping every time I want to shrink the log file because the database size is huge and it will take time to restore in the DR site, so the reconfiguration
    is not an option :( 
    how can I change the secondary database state from Standby to No recovery mode? I tried to change it from the wizard and wait until the next restore for the transaction log backup, but the job failed and the error was: "the step failed". I need
    to do this to change the mdf and ldf file location for the secondary databases.
    can any one help?
    Thanks in advance,
    Faris ALMasri
    Database Administrator

    1. can I shrink the log file for these databases: If I change the primary database from full to simple and shrink the log file then change it back to full recovery mode the log shipping will fail, I've seen some answers talked about using "No Truncate"
    option, but as I know this option will not affect the log file and it will shrink the data file only.
          I also can't create maintenance to reconfigure the log shipping every time I want to shrink the log file because the database size is huge
    and it will take time to restore in the DR site, so the reconfiguration is not an option :( 
    2. how can I change the secondary database state from Standby to No recovery mode? I tried to change it from the wizard and wait until the next restore for the transaction log backup, but the job failed and the error was: "the step failed". I need to do
    this to change the mdf and ldf file location for the secondary databases.
    can any one help?
    Thanks in advance,
    Faris ALMasri
    Database Administrator
    1. If you change recovery model of database in logshipping to simple and back to full Logshipping will break and logs wont be resored on Secondary server as log chain will be broken.You can shrink log file of primary database but why would you need that
    what is schedule of log backup. Frequent log backup is already taking care of log files why to shrink it and create performance load on system when log file will ultimately grow and since because instant file initilaization is not for Log files it takes time
    to grow and thus slows performace.
    You said you want to shrink as Database size is huge is it huge or does it have lots of free space. dont worry about data file free space it will eventually be utilized by SQL server when more data comes
    2. You are following wrong method changing state to no recovery would not even allow you to run select queries which you can run in Standby mode. Please refer below link to move Secondary data and log files
    http://www.mssqltips.com/sqlservertip/2836/steps-to-move-sql-server-log-shipping-secondary-database-files/
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it.
    My TechNet Wiki Articles

  • Ttcwerrors.log and ttcwmesg.log file locations

    Is there an option we can set in the ttcrsagent.options file to change the location of the ttcwerrors.log and ttcwmesg.log files?
    We can change the location of the daemon user and error log files by setting the -supportlog and -userlog options in ttendaemon.options, so we were hoping we could do the same in ttcrsagent.options.
    Thanks,
    -Jules

    Unfortunately, no. The ttcw log file location is not configurable
    Simon

  • Has 903 oc4j fixed / added rotating log files; -out and -err ?

    To be 24/7 an appserver needs to at least support managing the System.out traffic
    that apps, frameworks (BC4J) spew.
    It was suggested that 903 OC4J will support rotating log files?
    Is this true and any help or doc pointers would be great.
    BTW, I'm moving to 903 enterprise with OEM.
    Thanks, curt

    To be 24/7 an appserver needs to at least support managing the System.out traffic
    that apps, frameworks (BC4J) spew.
    It was suggested that 903 OC4J will support rotating log files?
    Is this true and any help or doc pointers would be great.
    BTW, I'm moving to 903 enterprise with OEM.
    Thanks, curt Curt - there is no way in 903 to rotate log files for a specific OC4J instance.
    There is a feature request in place for this functionality but I am not sure which version it will be delivered with.
    -steve-

Maybe you are looking for