Logs are creating frequently

Our production server logs are creating frequently and system is down simultaneously.
Any suggestions

Hi,
you could also consider using infinite logging, (as of DB2 v 8). This means is possible to support environments with large jobs that require more log space than you would normally allocate to the primary logs. For this, the database configuration parameter LOGSECOND must be set to -1
Please refer to SAP note  1493587
regards, Javier

Similar Messages

  • Logs are lost frequently in Remote Agent Server

    Hi All,
    ACS 4.2 and remote agent was working properly two months before. But in past two months we are facing weird issue in RA server.For Somedays we are missing logs from both ACS and RA server. Once we notice this we use to restart the services in ACS to give workaround. But due to this we loose our daily logs intermittently and facing risk in without having logs.
    This is not like communication between ACS and RA is not at all happening. It happens properly for a week or month, but again it is going bad without any config change.
    CSAgent.ini file is properly configured.
    Full version is 4.2.1.15 and patch is 10 in acs and ra.
    ACS and Remote Agent Major and Patch version are same.
    Any help would be highly appreciated.
    Thanks.

    Do you see the same intermittent logging issue with all services like csauth, rds, tcs etc.
    Could you plz. provide the cslogagent file from the remote agent server when you see this issue happening.
    C:\Program Files\Cisco\CiscoSecure ACS Agent\CSLogAgent\Logs
    Steps to gather msinfo32
    - Click the Start button.
    - Click the Run option.
    - Type "msinfo32" into the box and press Enter.
    - Click "System Information."
    - Choose "Save as System Information File" from the "Action" menu.
    - Open an appropriate directory, or something easy like the desktop.
    - Type "MSInfo32" into the Filename box and press Enter.
    - Wait several minutes for the program to collect and save information
    Did you check whether cisco remote agent server is excluded for being scanned by AV?
    Jatin Katyal
    - Do rate helpful posts -

  • HTTP logs not created when using software load balancer

    This is my setup:
    1. I have a server running a software load balancer - call it A
    2. There are two servers clustered under A - call it X and Y
    3. I am able to access the application deployed on X and Y through A.
    Fail-over also works. Shutting down X when the application is up and running routes the requests to Y and vice-versa. All is well until this point.
    Problem is:
    Let us take server X.
    $ORACLE_HOME/Apache/Apache/logs does not have an access log in it. (It
    however has an error log.)
    Access logs are not created when accessing the servers in the cluster (X and Y) through the load balancer URL.
    However, when accessing the individual servers in the cluster (with their URLs), access logs are created.
    Specifics on the setup:
    1. A, X, Y - all three are OAS 10.1.3.2
    2. All three are deployed on individual server machines
    3. All three servers have one OC4J instance each running in them
    4. The software load balancer used is from WebLogic 8.1sp5 - basically to create a *.war file that has the details of X and Y (servers in the cluster)
    5. This *.war file is deployed on A
    6. Server A, the one that has the software load balancer, DOES NOT have AS control. One of the other servers in the cluster does.
    Has anybody encountered this problem before? Or is there anything wrong
    with the setup? What is the resolution for the same?
    Any help is greatly appreciated.
    Thank you.

    Thank you, Jacco. You are right in saying that the software load balancer does not route the request to Apache and hence the logs are not created.
    While the logs are not critical to the component we are working on, was just wondering why they were not being created.
    As regards your comment on FastCGI, there is a FastCGI directory created under $ORACLE_HOME/Apache/Apache/logs - but there is nothing in it. Neither do the error logs have any information pertaining to FastCGI.
    Anyway, I am not going to worry about this issue anymore - may be sometime in the future will try to use a different software load balancer and see if it makes any difference.

  • Why multiple  log files are created while using transaction in berkeley db

    we are using berkeleydb java edition db base api, we have already read/write CDRFile of 9 lack rows with transaction and
    without transaction implementing secondary database concept the issues we are getting are as follows:-
    with transaction----------size of database environment 1.63gb which is due to no. of log files created each of 10 mb.
    without transaction-------size of database environment 588mb and here only one log file is created which is of 10mb. so we want to know how REASON CONCRETE CONCLUSION ..
    how log files are created and what is meant of using transaction and not using transaction in db environment and what are this db files db.001,db.002,_db.003,_db.004,__db.005 and log files like log.0000000001.....plz reply soon

    we are using berkeleydb java edition db base api, If you are seeing __db.NNN files in your environment root directory, these are environment's shared region files. And since you see these you are using Berkeley DB Core (with the Java/JNI Base API), not Berkeley DB Java Edition.
    with transaction ...
    without transaction ...First of all, do you need transactions or not? Review the documentation section called "Why transactions?" in the Berkeley DB Programmer's Reference Guide.
    without transaction-------size of database environment 588mb and here only one log file is created which is of 10mb.There should be no logs created when transactions are not used. That single log file has likely remained there from the previous transactional run.
    how log files are created and what is meant of using transaction and not using transaction in db environment and what are this db files db.001,db.002,_db.003,_db.004,__db.005 and log files like log.0000000001Have you reviewed the basic documentations references for Berkeley DB Core?
    - Berkeley DB Programmer's Reference Guide
    in particular sections: The Berkeley DB products, Shared memory regions, Chapter 11. Berkeley DB Transactional Data Store Applications, Chapter 17. The Logging Subsystem.
    - Getting Started with Berkeley DB (Java API Guide) and Getting Started with Berkeley DB Transaction Processing (Java API Guide).
    If so, you would have had the answers to these questions; the __db.NNN files are the environment shared region files needed by the environment's subsystems (transaction, locking, logging, memory pool buffer, mutexes), and the log.MMMMMMMMMM are the log files needed for recoverability and created when running with transactions.
    --Andrei                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • SHADOW_IMPORT_UPG1 is very very slow, no log files are created

    Hi all
    We are now doing our production upgrade, during the SHADOW_IMPORT_UPG1 phase system is very slow
    no log files are created in the /usr/sap/put/log directory.
    only three files are growing in /usr/sap/tmp directory
    orar3p> ls -lrt
    total 219176
    -rw-rw-rw-   1 r3padm     sapsys        2693 Aug 15 18:42 UCMIG_DE.ECO
    -rw-rw-rw-   1 r3padm     sapsys        2374 Aug 15 18:42 R3trans.out
    -rw-rw-rw-   1 r3padm     sapsys        2685 Aug 15 18:46 ADDON_TR.ECO
    -rw-rw-rw-   1 r3padm     sapsys         726 Aug 15 20:04 crshdusr.log
    -rw-rw-rw-   1 r3padm     sapsys        3915 Aug 15 21:53 EU_IMTSK.ECO
    -rw-rw-r--   1 r3padm     sapsys         257 Aug 15 22:09 SAPKKLFRN18.R3P
    -rw-rw-r--   1 r3padm     sapsys         257 Aug 15 22:09 SAPKKLPTN18.R3P
    -rw-rw-r--   1 r3padm     sapsys         257 Aug 15 22:09 SAPKKLESN18.R3P
    -rw-rw-r--   1 r3padm     sapsys     36433272 Aug 15 23:44 SAPKLESN18.R3P
    -rw-rw-r--   1 r3padm     sapsys     36807577 Aug 15 23:44 SAPKLFRN18.R3P
    -rw-rw-r--   1 r3padm     sapsys     35372350 Aug 15 23:44 SAPKLPTN18.R3P
    orar3p> date
    Fri Aug 15 23:44:54 PDT 2008
    Can anyone help what to do
    Thanks
    Senthil

    Hello,
    did you discover what the cause was for this phase running so slow? And  how long did it take to complete in the end?
    We are currently running an upgrade of our Development system and have struck the same issue.
    I killed the upgrade after the phase had been running for 4 hours and restarted it, but it looks like it is still going to run for a long time.
    Regards....John

  • Logs are not getting applied in the standby database

    Hello,
    I have created a physical standby database and in it the logs are not getting applied..
    following is an extract of the standby alert log
    Wed Sep 05 07:53:59 2012
    Media Recovery Log /u01/oracle/oradata/ABC/archives/1_37638_765704228.arc
    Error opening /u01/oracle/oradata/ABC/archives/1_37638_765704228.arc
    Attempting refetch
    Media Recovery Waiting for thread 1 sequence 37638
    Fetching gap sequence in thread 1, gap sequence 37638-37643
    Wed Sep 05 07:53:59 2012
    RFS[46]: Assigned to RFS process 3081
    RFS[46]: Allowing overwrite of partial archivelog for thread 1 sequence 37638
    RFS[46]: Opened log for thread 1 sequence *37638* dbid 1723205832 branch 765704228
    Wed Sep 05 07:55:34 2012
    RFS[42]: Possible network disconnect with primary database
    However, the archived files are getting copied to the standby server.
    I tried registering and recovering the logs but it also failed..
    Follows some of the information,
    Primary
    Oralce 11R2 EE
    SQL> select max(sequence#) from v$log where archived='YES';
    MAX(SEQUENCE#)
    37668
    SQL> select DEST_NAME, RECOVERY_MODE,DESTINATION,ARCHIVED_SEQ#,APPLIED_SEQ#, SYNCHRONIZATION_STATUS,SYNCHRONIZED,GAP_STATUS from v$archive_dest_status where DEST_NAME = 'LOG_ARCHIVE_DEST_3';
    DEST_NAME RECOVERY_MODE DESTINATION ARCHIVED_SEQ# APPLIED_SEQ# SYNCHRONIZATION_STATUS SYNCHRONIZED GAP_STATUS
    LOG_ARCHIVE_DEST_3 MANAGED REAL TIME APPLY ABC 37356 0 CHECK CONFIGURATION NO RESOLVABLE GAP
    Standby
    Oralce 11R2 EE
    SQL> select max(sequence#) from v$archived_log where applied='YES';
    MAX(SEQUENCE#)
    37637
    SQL> select * from v$archive_gap;
    no rows selected
    SQL> select open_mode, database_role from v$database;
    OPEN_MODE DATABASE_ROLE
    READ ONLY WITH APPLY PHYSICAL STANDBY
    Please help me to troubleshoot this and get the standby in sync..
    Thanks a lot..

    the results are as follows..
    SQL> select process, status, group#, thread#, sequence# from v$managed_standby order by process, group#, thread#, sequence#;
    PROCESS STATUS GROUP# THREAD# SEQUENCE#
    ARCH CLOSING 1 1 37644
    ARCH CLOSING 1 1 37659
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    ARCH CONNECTED N/A 0 0
    MRP0 WAIT_FOR_GAP N/A 1 37638
    RFS IDLE N/A 0 0
    RFS IDLE N/A 0 0
    RFS IDLE N/A 0 0
    RFS RECEIVING N/A 1 37638
    RFS RECEIVING N/A 1 37639
    RFS RECEIVING N/A 1 37640
    RFS RECEIVING N/A 1 37641
    RFS RECEIVING N/A 1 37642
    RFS RECEIVING N/A 1 37655
    RFS RECEIVING N/A 1 37673
    RFS RECEIVING N/A 1 37675
    42 rows selected.
    SQL>
    SQL> select name,value, time_computed from v$dataguard_stats;
    NAME VALUE TIME_COMPUTED
    transport lag +00 02:44:33 09/05/2012 09:25:58
    apply lag +00 03:14:30 09/05/2012 09:25:58
    apply finish time +00 00:01:09.974 09/05/2012 09:25:58
    estimated startup time 12 09/05/2012 09:25:58
    SQL> select timestamp , facility, dest_id, message_num, error_code, message from v$dataguard_status order by timestamp
    TIMESTAMP FACILITY DEST_ID MESSAGE_NUM ERROR_CODE MESSAGE
    05-SEP-12 Remote File Server 0 60 0 RFS[13]: Assigned to RFS process 2792
    05-SEP-12 Remote File Server 0 59 0 RFS[12]: Assigned to RFS process 2790
    05-SEP-12 Log Apply Services 0 61 16037 MRP0: Background Media Recovery cancelled with status 16037
    05-SEP-12 Log Apply Services 0 62 0 MRP0: Background Media Recovery process shutdown
    05-SEP-12 Log Apply Services 0 63 0 Managed Standby Recovery Canceled
    05-SEP-12 Log Apply Services 0 64 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 65 0 Attempt to start background Managed Standby Recovery process
    05-SEP-12 Log Apply Services 0 66 0 MRP0: Background Managed Standby Recovery process started
    05-SEP-12 Log Apply Services 0 67 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 68 0 Media Recovery Waiting for thread 1 sequence 37638 (in transit)
    05-SEP-12 Network Services 0 69 0 RFS[5]: Possible network disconnect with primary database
    05-SEP-12 Network Services 0 70 0 RFS[6]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 71 0 RFS[14]: Assigned to RFS process 2829
    05-SEP-12 Remote File Server 0 72 0 RFS[15]: Assigned to RFS process 2831
    05-SEP-12 Network Services 0 73 0 RFS[9]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 74 0 RFS[16]: Assigned to RFS process 2833
    05-SEP-12 Network Services 0 75 0 RFS[1]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 76 0 RFS[17]: Assigned to RFS process 2837
    05-SEP-12 Network Services 0 77 0 RFS[3]: Possible network disconnect with primary database
    05-SEP-12 Network Services 0 78 0 RFS[2]: Possible network disconnect with primary database
    05-SEP-12 Network Services 0 79 0 RFS[7]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 80 0 RFS[18]: Assigned to RFS process 2848
    05-SEP-12 Network Services 0 81 0 RFS[16]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 82 0 RFS[19]: Assigned to RFS process 2886
    05-SEP-12 Network Services 0 83 0 RFS[19]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 84 0 RFS[20]: Assigned to RFS process 2894
    05-SEP-12 Log Apply Services 0 85 16037 MRP0: Background Media Recovery cancelled with status 16037
    05-SEP-12 Log Apply Services 0 86 0 MRP0: Background Media Recovery process shutdown
    05-SEP-12 Log Apply Services 0 87 0 Managed Standby Recovery Canceled
    05-SEP-12 Remote File Server 0 89 0 RFS[22]: Assigned to RFS process 2900
    05-SEP-12 Remote File Server 0 88 0 RFS[21]: Assigned to RFS process 2898
    05-SEP-12 Remote File Server 0 90 0 RFS[23]: Assigned to RFS process 2902
    05-SEP-12 Remote File Server 0 91 0 Primary database is in MAXIMUM PERFORMANCE mode
    05-SEP-12 Remote File Server 0 92 0 RFS[24]: Assigned to RFS process 2904
    05-SEP-12 Remote File Server 0 93 0 RFS[25]: Assigned to RFS process 2906
    05-SEP-12 Log Apply Services 0 94 0 Attempt to start background Managed Standby Recovery process
    05-SEP-12 Log Apply Services 0 95 0 MRP0: Background Managed Standby Recovery process started
    05-SEP-12 Log Apply Services 0 96 0 Managed Standby Recovery starting Real Time Apply
    05-SEP-12 Log Apply Services 0 97 0 Media Recovery Waiting for thread 1 sequence 37638 (in transit)
    05-SEP-12 Log Transport Services 0 98 0 ARCa: Beginning to archive thread 1 sequence 37644 (7911979302-7912040568)
    05-SEP-12 Log Transport Services 0 99 0 ARCa: Completed archiving thread 1 sequence 37644 (0-0)
    05-SEP-12 Network Services 0 100 0 RFS[8]: Possible network disconnect with primary database
    05-SEP-12 Log Apply Services 0 101 16037 MRP0: Background Media Recovery cancelled with status 16037
    05-SEP-12 Log Apply Services 0 102 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 103 0 MRP0: Background Media Recovery process shutdown
    05-SEP-12 Log Apply Services 0 104 0 Managed Standby Recovery Canceled
    05-SEP-12 Network Services 0 105 0 RFS[20]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 106 0 RFS[26]: Assigned to RFS process 2930
    05-SEP-12 Network Services 0 107 0 RFS[24]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 108 0 RFS[27]: Assigned to RFS process 2938
    05-SEP-12 Network Services 0 109 0 RFS[14]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 110 0 RFS[28]: Assigned to RFS process 2942
    05-SEP-12 Network Services 0 111 0 RFS[15]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 112 0 RFS[29]: Assigned to RFS process 2986
    05-SEP-12 Network Services 0 113 0 RFS[17]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 114 0 RFS[30]: Assigned to RFS process 2988
    05-SEP-12 Log Apply Services 0 115 0 Attempt to start background Managed Standby Recovery process
    05-SEP-12 Log Apply Services 0 116 0 MRP0: Background Managed Standby Recovery process started
    05-SEP-12 Log Apply Services 0 117 0 Managed Standby Recovery starting Real Time Apply
    05-SEP-12 Log Apply Services 0 118 0 Media Recovery Waiting for thread 1 sequence 37638 (in transit)
    05-SEP-12 Network Services 0 119 0 RFS[18]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 120 0 RFS[31]: Assigned to RFS process 3003
    05-SEP-12 Network Services 0 121 0 RFS[26]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 122 0 RFS[32]: Assigned to RFS process 3005
    05-SEP-12 Network Services 0 123 0 RFS[27]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 124 0 RFS[33]: Assigned to RFS process 3009
    05-SEP-12 Remote File Server 0 125 0 RFS[34]: Assigned to RFS process 3012
    05-SEP-12 Log Apply Services 0 127 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 126 16037 MRP0: Background Media Recovery cancelled with status 16037
    05-SEP-12 Log Apply Services 0 128 0 MRP0: Background Media Recovery process shutdown
    05-SEP-12 Log Apply Services 0 129 0 Managed Standby Recovery Canceled
    05-SEP-12 Network Services 0 130 0 RFS[32]: Possible network disconnect with primary database
    05-SEP-12 Log Apply Services 0 131 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 132 0 Attempt to start background Managed Standby Recovery process
    05-SEP-12 Log Apply Services 0 133 0 MRP0: Background Managed Standby Recovery process started
    05-SEP-12 Log Apply Services 0 134 0 Managed Standby Recovery not using Real Time Apply
    05-SEP-12 Log Apply Services 0 135 0 Media Recovery Waiting for thread 1 sequence 37638 (in transit)
    05-SEP-12 Network Services 0 136 0 RFS[33]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 137 0 RFS[35]: Assigned to RFS process 3033
    05-SEP-12 Log Apply Services 0 138 16037 MRP0: Background Media Recovery cancelled with status 16037
    05-SEP-12 Log Apply Services 0 139 0 MRP0: Background Media Recovery process shutdown
    05-SEP-12 Log Apply Services 0 140 0 Managed Standby Recovery Canceled
    05-SEP-12 Remote File Server 0 141 0 RFS[36]: Assigned to RFS process 3047
    05-SEP-12 Log Apply Services 0 142 0 Attempt to start background Managed Standby Recovery process
    05-SEP-12 Log Apply Services 0 143 0 MRP0: Background Managed Standby Recovery process started
    05-SEP-12 Log Apply Services 0 144 0 Managed Standby Recovery starting Real Time Apply
    05-SEP-12 Log Apply Services 0 145 0 Media Recovery Waiting for thread 1 sequence 37638 (in transit)
    05-SEP-12 Network Services 0 146 0 RFS[35]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 147 0 RFS[37]: Assigned to RFS process 3061
    05-SEP-12 Network Services 0 148 0 RFS[36]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 149 0 RFS[38]: Assigned to RFS process 3063
    05-SEP-12 Remote File Server 0 150 0 RFS[39]: Assigned to RFS process 3065
    05-SEP-12 Network Services 0 151 0 RFS[25]: Possible network disconnect with primary database
    05-SEP-12 Network Services 0 152 0 RFS[21]: Possible network disconnect with primary database
    05-SEP-12 Remote File Server 0 153 0 Archivelog record exists, but no file is found
    05-SEP-12 Remote File Server 0 154 0 RFS[40]: Assigned to RFS process 3067
    05-SEP-12 Network Services 0 155 0 RFS[37]: Possible network disconnect with primary database

  • Where to see activities, which are created for the campaign with target group

    Hi
    I have created a Communication Method in SPRO -> CRM -> Marketing -> Marketing Planning & Campaign Management -> Campaign Executtion - Define Communication Method : Activity has been configured
    Activities are not appearing after creation of campaign with target group.
    Where to see activities, which are created for the campaign with target group?
    Regards
    Hamid

    Hello,
    alternatively to check  table CRMD_ORDERADM_H directly you can use the related transactions search for the campaign:
    This should bring all activities generated by the campaign. If this is empty first check the job log of the campaign execution job again:
    This gives information if any target group members have business transactions created. If the log does not contain the required information you can debug from SM37. The business transactions (activities, leads, sales orders) are generated from function module CRM_MKTTGGRP_PROC_CREATE.
    best regards,
    Johannes

  • Restored standby database from primary; now no logs are shipped

    Hi
    We recently had a major network/SAN issue and had to restore our standby database from a backup of the primary. To do this, we restored the database to the standby, created a standby controlfile on the primary, copied this across to the control file locations and started in standby recover and applied the logs manually/registered to get it back up to speed.
    However, no new logs are being shipped across from the primary.
    Have we missed a step somewhere?
    One thing we've noticed is that there is no RFS process running on the standby:
    SQL> SELECT PROCESS, CLIENT_PROCESS, SEQUENCE#, STATUS FROM V$MANAGED_STANDBY;
    PROCESS CLIENT_P SEQUENCE# STATUS
    ARCH ARCH 0 CONNECTED
    ARCH ARCH 0 CONNECTED
    MRP0 N/A 100057 WAIT_FOR_LOG
    How do we start this? Or will it only show if the arc1 process on the primary is sending files?
    The arc1 process is showing at OS level on the primary but I'm wondering if its faulty somehow?
    There are NO errors in the alert logs in the primary or the standby. There's not even the normal FAL gap sequence type error - in the standby it's just saying 'waiting for log' and a number from ages ago. It's like the primary isn't even talking to the standby. The listener is up and running ok though...
    What else can we check/do?
    If we manually copy across files and do an 'alter database register' then they are applied to the standby without issue; there's just no automatic log shipping going on...
    Thanks
    Ross

    Hi all
    Many thanks for all the responses.
    The database is 10.2.0.2.0, on AIX 6.
    I believe the password files are ok; we've had issues previously and this is always flagged in the alert log on the primary - not the case here.
    Not set to DEFER on primary; log_archive_dest_2 is set to service="STBY_PHP" optional delay=720 reopen=30 and log_archive_dest_state_2 is set to ENABLE.
    I ran those troubleshooting scripts, info from standby:
    SQL> @troubleshoot
    NAME DISPLAY_VALUE
    db_file_name_convert
    db_name PHP
    db_unique_name PHP
    dg_broker_config_file1 /oracle/PHP/102_64/dbs/dr1PHP.dat
    dg_broker_config_file2 /oracle/PHP/102_64/dbs/dr2PHP.dat
    dg_broker_start FALSE
    fal_client STBY_PHP
    fal_server PHP
    local_listener
    log_archive_config
    log_archive_dest_2 service=STBY_PHP optional delay=30 reopen=30
    log_archive_dest_state_2 DEFER
    log_archive_max_processes 2
    log_file_name_convert
    remote_login_passwordfile EXCLUSIVE
    standby_archive_dest /oracle/PHP/oraarch/PHParch
    standby_file_management AUTO
    NAME DB_UNIQUE_NAME PROTECTION_MODE DATABASE_R OPEN_MODE
    PHP PHP MAXIMUM PERFORM PHYSICAL S MOUNTED
    ANCE TANDBY
    THREAD# MAX(SEQUENCE#)
    1 100149
    PROCESS STATUS THREAD# SEQUENCE#
    ARCH CONNECTED 0 0
    ARCH CONNECTED 0 0
    MRP0 WAIT_FOR_LOG 1 100150
    NAME VALUE UNIT TIME_COMPUTED
    apply finish time day(2) to second(1) interval
    apply lag day(2) to second(0) interval
    estimated startup time 8 second
    standby has been open N
    transport lag day(2) to second(0) interval
    NAME Size MB Used MB
    0 0
    On the primary, the script has froze!! How long should it take? Got as far as this:
    SQL> @troubleshoot
    NAME DISPLAY_VALUE
    db_file_name_convert
    db_name PHP
    db_unique_name PHP
    dg_broker_config_file1 /oracle/PHP/102_64/dbs/dr1PHP.dat
    dg_broker_config_file2 /oracle/PHP/102_64/dbs/dr2PHP.dat
    dg_broker_start FALSE
    fal_client STBY_R1P
    fal_server R1P
    local_listener
    log_archive_config
    log_archive_dest_2 service="STBY_PHP" optional delay=720 reopen=30
    log_archive_dest_state_2 ENABLE
    log_archive_max_processes 2
    log_file_name_convert
    remote_login_passwordfile EXCLUSIVE
    standby_archive_dest /oracle/PHP/oraarch/PHParch
    standby_file_management AUTO
    NAME DB_UNIQUE_NAME PROTECTION_MODE DATABASE_R OPEN_MODE SWITCHOVER_STATUS
    PHP PHP MAXIMUM PERFORMANCE PRIMARY READ WRITE SESSIONS ACTIVE
    THREAD# MAX(SEQUENCE#)
    1 100206
    NOW - before you say it - :) - yes, I'm aware that fal_client as STBY_R1P and fal_server as R1P are incorrect - should be PHP - but it looks like it's always been this way! Well, as least for the last 4 years where it's worked fine, as I found an old SP file and it still has R1P set in there...?!?
    Any ideas?
    Ross

  • Change document log not created for customer in Batch input/BDC

    Hi Guys,
    Change document log is created when you manually change the customer using XD02 but when run the Batch/BDC the values are Changed but the change document is not created.
    Please someone can give me some hints?
    Luca.

    Any update guys????
    Regards,
    guru singh

  • BPC Detail error/reject log not created in the PrivatePub/user/temp folder

    Hi,
    We have a custom import package that imports data from a comma seperated txt file in to BPC. If all the data members in the data file are valid dimension members the package is successful and the package log file is created in the PrivatePub/user/Temp folder. But if there are any rejects the reject log is not getting created under the temp folder and moreover the package status is "Error". The package log is created but the reject log is not. The Webfolders has the full share permissions.
    The same package runs fine in our Test environment and creates the package log and reject log.
    We are on BPC 5.1 SP3.
    Is there a configuration that is missing.
    Thanks in advance.
    Swaroop

    I'm pretty sure that when you do not explicitly indicate a transformation file, a default file (import.xls I think?) is used.
    Also, there are a number of temp files & variables that must be created in the MODIFYSCRIPT variable, and certain conversion tasks that are part of the standard import process. You can see the proper syntax in the sample import pacakges.
    If you've created a new package of your own, you may be missing some key pieces. If you don't understand what all the pieces are in the sample packages (and why you don't need them), you should start by debugging with one of them, and then see what you need to add back into your package.
    I'd normally recommend to start with the sample import package, and only modify it if necessary, and as little as possible.
    Last thought - file permission... does the user running the package (not the end user logged into BPC, but the BPC data manager user, which I believe is the "USER" (not admin or sysadmin) set up during the server install) have appropriate access rights to that network share where the data file is located? Try changing it to a local path, and copy the test.txt to that location.

  • Archive logs are missing in hot backup

    Hi All,
    We are using the following commands to take hot backup of our database. Hot backup is fired by "backup" user on Linux system.
    =======================
    rman target / nocatalog <<EOF
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '$backup_dir/$date/%F';
    run {
    allocate channel oem_backup_disk1 type disk format '$backup_dir/$date/%U';
    #--Switch archive logs for all threads
    sql 'alter system archive log current';
    backup as COMPRESSED BACKUPSET database include current controlfile;
    #--Switch archive logs for all threads
    sql 'alter system archive log current';
    #--Backup Archive logs and delete what we've backedup
    backup as COMPRESSED BACKUPSET archivelog all not backed up delete all input;
    release channel oem_backup_disk1;
    allocate channel for maintenance type disk;
    delete noprompt obsolete device type disk;
    release channel;
    exit
    EOF
    =======================
    Due to which after command (used 2 times) "sql 'alter system archive log current';" I see the following lines in alert log 2 times. Because of this all the online logs are not getting archived (Missing 2 logs per day), the backup taken is unusable when restoring. I am worried about this. I there any to avoid this situation.
    =======================
    Errors in file /u01/oracle/admin/rac/udump/rac1_ora_3546.trc:
    ORA-19504: failed to create file "+DATA/rac/1_32309_632680691.dbf"
    ORA-17502: ksfdcre:4 Failed to create file +DATA/rac/1_32309_632680691.dbf
    ORA-15055: unable to connect to ASM instance
    ORA-01031: insufficient privileges
    =======================
    Regards,
    Kunal.

    All thanks you for help, pleas find additional information. I goth the following error as log sequence was missing. Everyday during hotbackup, there are 2 missing archive logs, which makes our backup inconsistent and useless.
    archive log filename=/mnt/xtra-backup/ora_archivelogs/1_32531_632680691.dbf thread=1 sequence=32531
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28768_632680691.dbf thread=2 sequence=28768
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28769_632680691.dbf thread=2 sequence=28769
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28770_632680691.dbf thread=2 sequence=28770
    archive log filename=/mnt/xtra-backup/ora_archivelogs/1_32532_632680691.dbf thread=1 sequence=32532
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28771_632680691.dbf thread=2 sequence=28771
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28772_632680691.dbf thread=2 sequence=28772
    archive log filename=/mnt/xtra-backup/ora_archivelogs/2_28772_632680691.dbf thread=2 sequence=28773
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of recover command at 12/13/2012 04:22:56
    RMAN-11003: failure during parse/execution of SQL statement: alter database recover logfile '/mnt/xtra-backup/ora_archivelogs/2_28772_632680691.dbf'
    ORA-00310: archived log contains sequence 28772; sequence 28773 required
    ORA-00334: archived log: '/mnt/xtra-backup/ora_archivelogs/2_28772_632680691.dbf'
    Let me try the susggestions provided above.

  • How to identify what are all the events are created in background jobs?

    Hi all,
    how to identify what are all the events are created for  background jobs. And what events gets triggered for a particular job.
    thanxs
    haritha

    Hi Haritha,
    JOB is a program which starts to a determined point of time and executes some standard programs in the system. JOBs can be planed to a determined point of time on the regular basis (every night, for example) or to some discret time moments. So, the JOB can be planed and then will be started automatically without the manual start.
    Realtime programs are understood in the most cases as actual program execution which is started by somebody to the actual moment of time.
    Typically per JOBs some special processes will be started that should be executed automatically and regularly: for example, IDOC application, some correction reports, statistic updates etc.
    Standard jobs are those background jobs that should be run regularly in a production SAP System These jobs are usually jobs that clean up parts of the system, such as by deleting old spool requests.
    Use
    As of Release 4.6C, the Job Definition transaction ( sm36 ) provides a list of important standard jobs, which you can schedule, monitor, and edit.
    Standard jobs are those background jobs that should be run regularly in a production SAP System. These jobs are usually jobs that clean up parts of the system, such as by deleting old spool requests.
    for more information you can go thru the following thread:
    http://help.sap.com/saphelp_nw70/helpdata/en/24/b884388b81ea55e10000009b38f842/frameset.htm
    About Events:
    Events have meaning only in the background processing system. You can use events only to start background jobs.
    Triggering an event notifies the background processing system that a named condition has been reached. The background processing system reacts by starting any jobs that were waiting for the event.
    Types of Events:
    There are two types of events:
    1.)System events are defined by SAP. These events are triggered automatically when such system changes as the activation of a new operation mode take place.
    2.)User events are events that you define yourself. You must trigger these events yourself from ABAP or from external programs. You could, for example, signal the arrival of external data to be read into the SAP system by using an external program to trigger a background processing event.The event scheduler processes an event if the event is defined in the system.
    For example, if a system (System 1) receives an event from another system (System 2), the event scheduler of System 1 processes the event only if it is defined in System 1. That event does not need to be defined in System 2 (the sending system).
    You define an event by assigning a name (EVENTID) to it. When defining an event, you do not define the event arguments.
    for more information you can go thru the following thread:
    http://help.sap.com/saphelp_nw04s/helpdata/en/fa/096e2a543b11d1898e0000e8322d00/frameset.htm
    When you schedule the process chain or infopackages the jobs associated with it run in the background mode. In case you want to create a job for a specific activity you can do so in SM36. You would be creating jobs that would get executed in any one of the options:
    1. Immediate
    2. Date & Time
    3. After event.
    4. After job.
    5. At Operation mode.
    In case you want to view the job logs go to sm37.
    Also Pls check DB02 for database performance and ST03 for workload .
    Analyse u will have an idea ,
    *pls assign points,if info is useful**
    Regards
    CSM reddy
    null

  • Soft errors/warnings are creating idocs in 51 status, and creating orders

    In our SAP 4.7 environment, inbound orders that encounter credit checks, materials in the wrong status, blocked customers, and several other issues that should just be warnings or messages rather than resulting in status 51 idocs, are hitting 51 error status, and they are creating orders in SAP as well (51 status idocs are still creating orders, and nuisance messages are creating 51 status idocs, and still creating orders). I assume someone made some programming changes years ago to cause this to happen. This is really polluting workflow and BD87, and customer service is duplicating orders as they process 51 status idocs through workflow, unaware that orders already created immediately regardless of the 51 status.
    Can anyone point me in the direction of where I can specify what warnings or errors constitute 51 status, and where I can stop 51 status idocs from creating orders until the error is repaired?
    Would it be in program RBDAPP01, or the inbound function module EDI_DATA_INCOMING?
    Any help would be appreciated. Thanks!

    Hi Kevin,
    so then it is time to check it out in WE19 debug mode. After each PERFORM in the EDIDC Loop in the FM IDOC_INPUT_ORDERS, check the control and status data, this will lead you to the point where the unexpected and/or unwanted behavior appears.
    In the LOOP AT IDOC_CONTRL, in our system line 183 you find the comment fill IDOC Status. Check if field OK is not cleared here. If OK is set although you should have errors, then it's already too late.
    But please double-check EXIT_SAPL_VEDA_004 if it changes the tables parameter didoc_status = idoc_status.
    What I did not mention yet as I think you are talking only about IDoc not about online processing, applies to both:
    Customizing SPRO, SAP Reference IMG, Sales and Distribution -> Sales -> Sales Documents -> Define variable messages:
    Here you may have entries for application area V4, Sales messages (variable messages and EDI messages). Check if you find suspicious messages set to 'W' here.
    If this is the case and table logging is switched on as it should be, you can use menu uitilities -> change logs to find out who changed what when.
    Good luck!
    Regards
    Clemens

  • Users are created but Roles are not Provisioned in the Target System

    Hi,
    It would be great if somebody would provided solution to my problem. The problem is when I try to create the Users in Identity Managment UI then the Users are created in the Target systems but the Roles are not provisioned to the Users.
    In the provisioning job SetABAPRole&ProfileForUser,
    It is says In the Error putNextEntry failed storing
    Exception from Modify operation:com.sap.idm.ic.ToPassException: User does not exist
    MSKEY 58437
    Please note the When we create the User, the user is created however the Roles is not provisioned to the user.
    Regards,
    Hakim

    Hello Nits,
    since this thread is from 2010 and the OP was logged on last in 2012 (as you can see in the profile), I don't think you'll get an answer here.
    Please create a new thread to explain your problem (with version and SP numbers, logs etc). You can add a link to this thread to show, that the problem is similar.
    Regards,
    Steffi.

  • To retieve tasks when user tasks are created using Jdeveloper BPELDesigner

    I am trying to get the list of tasks which are created using Jdeveloper BPel designer.
    I have used the following code :
    Locator loc = new Locator("default","bpel");
              IWorklistService taskService = (IWorklistService)loc.lookupService(IWorklistService.SERVICE_NAME);
    ITask[] tasks = taskService.listTasksByAssignee("rekha");
    if(tasks!=null){
    for (int j = 0; j < tasks.length; j++) {
    ITask thisTask = tasks[j];
    out.println("Task Title " + thisTask.getTitle() +"Assignee:"+thisTask.getAssignee()+"$$$$"+"\n");
    but its not returning any tests.
    But the same code works for the tasks which are created through sample bpel processes which are provided by oracle which uses TaskManager service.
    But the Jdeveloper bpel designer user task creates service which is TaskActionHandler..
    please help me how can I get the List of tasks for a particular user for this problem
    Please help me ............

    There is sample code that is supposed to do this here: http://www.codeguru.com/cpp/sample_chapter/article.php/c10789__12/
    I have tried using the login-style task retrieval with the oracle.tip.pc packages - this works logging in as, for example, fkafka.
    Working Code:
    IWorklistContext ctx = wlSvc.authenticateUser("fkafka", "welcome");
    Map filterMap = new HashMap();
    filterMap.put(IWorklistService.FILTER_TYPE_TASK_FILTER,
    IWorklistService.TASK_FILTER_MY);
    filterMap.put(IWorklistService.FILTER_TYPE_STATUS_FILTER, IWorklistService.STATUS_FILTER_ASSIGNED);
    String keywords = "";
    java.util.List tasks = wlSvc.getWorklistTasks(ctx, keywords, filterMap, IWorklistService.SORT_FIELD_TASK_TITLE, IWorklistService.SORT_ORDER_ASCENDING);
    Using the sample code and retrieving tasks with
    ITask[] tasks = worklist.listTasksByAssignee("fkafka");
    returns nothing, however.
    (&lt;h1>&lt;%= tasks.length %> tasks waiting for approval&lt;/h1>
    Prints: ' 0 tasks...' )
    I notice that the JDeveloper generated JSPs use the oracle.tip.pc packages - seeing as I'm deploying to the same BPEL server I guess I don't need to worry about the other methods of access (i.e. remote) so hopefully I can manage my workflow app with the oracle.tip.pc packages only.

Maybe you are looking for

  • How to get multiple values from the db using select

    hi all, i am using oracle 9i database i created a procedure create or replace procedure sample_proce as to_i varchar2(50) begin select to_id into to_i from sample where flag in('y'); dbms_output.put_line('hai'|| to_i); end; when i executed this proce

  • Events on iCal not showing up on my iPhone 4

    I downloaded iCloud and synced my iPhone calendar with my calendars on my mac. Before I did not have mobile me, so they were two seperate calendars that I could sync and update when I synced my phone to my computer. Now that I've merged the two toget

  • Lightroom Mobile Crop Results in Black Screens

    I have been successfully using LR Mobile to quickly crop about 1,600 Canon RAW images. Part way through this effort, the cropped images start showing up as black screens. I can three finger tap and get the correctly cropped image to show up, but if I

  • My adjustment brush suddenly stopped working. LR 5.

    My flow and density are set to 100 and I've tried using an overlay just to be sure. Nothing is working. Any ideas?

  • Unable to access to Business rules from planning

    HI all, We are using Hyperion Planning 9.3.1, Essbase 9.3.1 and AAS 9.3.1. Unfortunately, we are not able to access to business rules from planning. All rights are set properly, the business rules are visible through AAS 9.3.1 and are accessible from