Problems with Sybase Database for Collection Manager in SCE2020

We have problems with Sybase Database for Collection Manager in a SCE2020. The status is:
[[email protected] log]# ~scmscm/setup/alive.sh STATUS OK [[email protected] monitor]# ./monitor.sh -a -d Test: 01db_up.sh. Status: FAIL. Message: DB is not running Test: 02cm_up.sh. Status: PASS. Message: CM is running Test: 03free_db.sh. Status: PASS. Message: 99% free space in data db Test: 04free_log.sh. Status: PASS. Message: 99% free space in log db Test: 05cm_persistent_buffers.sh. Status: FAIL. Message: The following directory/ies have more than 500 files in them - JDBCAdapter TAAdapter [[email protected] monitor]# ~scmscm/scripts/dbtables.sh /home/scmscm/scripts/common.sh: line 43: /root/cm/bin/cm: is a directory Executing query ... /home/scmscm/scripts/dbtables.sh: line 83: /root/cm/bin/cm: is a directory [[email protected] monitor]# df -k Filesystem           1K-blocks      Used Available Use% Mounted on /dev/mapper/VolGroup00-LogVol00                      149559596 138280700   3681636  98% / /dev/sda1               101086     20685     75182  22% /boot none                   1036624         0   1036624   0% /dev/shm
We restored the Sybase database, but we have problems to access the database from Collection Manager and we can't obtain reports from SCE 2020.
Here is the info:
Problem Details: The SCA Reporter cannot generate Reports, the CM diagnostics show the following:
[[email protected] ~]# ~scmscm/setup/alive.sh STATUS OK
[[email protected] ~]# ~scmscm/scripts/dbfree.sh
Name                    % Data Free % Log Free
Database                        55      99
[[email protected]m ~]# ~scmscm/scripts/dbtables.sh
/home/scmscm/scripts/common.sh: line 43: /root/cm/bin/cm: No such file or directory Executing query ...
/home/scmscm/scripts/dbtables.sh: line 83: /root/cm/bin/cm: No such file or directory
[[email protected] ~]# ~scmscm/setup/monitor/monitor.sh -d -a
Test: 01db_up.sh. Status: FAIL. Message: DB is not running
Test: 02cm_up.sh. Status: PASS. Message: CM is running
Test: 03free_db.sh. Status: PASS. Message: 55% free space in data db
Test: 04free_log.sh. Status: PASS. Message: 99% free space in log db
Test: 05cm_persistent_buffers.sh. Status: FAIL. Message: The following directory/ies have more than 500 files in them - JDBCAdapter
Message was edited by: EMILIO MENCIA

Tomo:
we have problems with our reports in Collection Manager again. We reboot the CM, but the problems continue.
This is the log of the CM. What can be the problem? Thanks tomo
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum number of User Accounts during current sample period: 3.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum number of User Accounts since startup: 3.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum Configured Number of User Connections during current sample period: 200.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum Configured Number of User Connections since startup: 200.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum Number of User Connections during current sample period: 15.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum Number of User Connections since startup: 18.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum number of user seat licenses used during current sample period: 1.
Nov  3 11:49:18 localhost sybase_init: 00:00000:00012:2011/11/03 11:49:18.00 server  Maximum number of user seat licenses used since startup: 3.
Nov  9 19:21:49 localhost sybase_init: 00:00000:00069:2011/11/09 19:21:49.78 kernel  Cannot send, host process disconnected: btllt0012  suid: 3
Nov  9 19:21:49 localhost sybase_init: 00:00000:00069:2011/11/09 19:21:49.86 kernel  Cannot send, host process disconnected: btllt0012  suid: 3
Nov  9 19:21:49 localhost sybase_init: 00:00000:00069:2011/11/09 19:21:49.92 server  Error: 1608, Severity: 18, State: 4
Nov  9 19:21:49 localhost sybase_init: 00:00000:00069:2011/11/09 19:21:49.94 server  A client process exited abnormally, or a network error was encountered. Unless other errors occurred, continue processing normally.
Nov  9 19:21:49 localhost sybase_init: 00:00000:00069:2011/11/09 19:21:49.94 kernel  extended error information: hostname: btllt0012 login: pqb_admin
Nov  9 19:41:10 localhost sybase_init: 00:00000:00086:2011/11/09 19:41:10.01 kernel  Cannot send, host process disconnected: btllt0012  suid: 3
Nov  9 19:41:10 localhost sybase_init: 00:00000:00086:2011/11/09 19:41:10.01 kernel  Cannot send, host process disconnected: btllt0012  suid: 3
Nov  9 19:41:10 localhost sybase_init: 00:00000:00086:2011/11/09 19:41:10.01 server  Error: 1608, Severity: 18, State: 4
Nov  9 19:41:10 localhost sybase_init: 00:00000:00086:2011/11/09 19:41:10.01 server  A client process exited abnormally, or a network error was encountered. Unless other errors occurred, continue processing normally.
Nov  9 19:41:10 localhost sybase_init: 00:00000:00086:2011/11/09 19:41:10.01 kernel  extended error information: hostname: btllt0012 login: pqb_admin
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 server  Error: 632, Severity: 20, State: 2
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 server  Attempt to move memory with an incorrect length
of -794444483. Maximum allowed length is 16384.
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  ************************************
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  SQL causing error : =
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  ************************************
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  curdb = 4 tempdb = 2 pstat = 0x10000
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  lasterror = 632 preverror = 0 transtate = 3
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  curcmd = 0 program =                          
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.16 kernel  extended error information: hostname: btl-sce-cm login: pqb_admin
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08d40ca3 pcstkwalk+0x31b(0x9a78fdbc, 0x9a78fb6c, 0x0000270f, 0x00000002, 0x9a78fb6c)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08d40832 ucstkgentrace+0x13a(0x68a4006f, 0x00000002, 0x0000270f, (nil), (nil))
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08ce346d ucbacktrace+0x5d((nil), 0x00000001, (nil), 0x00000003, 0x20202020)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x082f19b8 terminate_process+0xa5c((nil), 0xffffffff, 0x9a7904c4, 0x08349a96, 0x00000278)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08349acb close_network+0xf(0x00000002, 0x9ced1ea0, 0x9a790544, 0x0834902e, 0x00000006)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08349aad hdl_default+0x45(0x00000006, 0x00000020, 0x00000014, 0x00000002, 0x9a7904fc)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x0834902e ex_raise+0x18a(0x00000006, 0x00000020, 0x00000014, 0x00000002, 0xd0a5bd3d)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08357c9b memmove_error+0x27(0xd0a5bd3d, 0x00004000, 0xffffffff, 0x0896b8a4, 0x9e3d7cfc)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x0833e54e recvhost+0xbe(0x9a7909ec, 0xd0a5bd3d, 0x00000018, 0x9e44d5d8, 0x9e44d52c)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08339160 recvchars+0x74(0x9ced2a1c, 0xd0a5bd3d, 0x9a7909ec, 0x000000ff, (nil))
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x0831e237 tdsrecv_declare+0x207(0x00000010, 0x9cecb914, 0x9a7911b4, 0x0832dd5f, (nil))
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  [Handler pc: 0x0x0863eca4 ut_handle installed by the following function:-]
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x0832e3a9 conn_hdlr+0xe49(0x00000030, 0x9a7911c8, 0x895eed31, (nil), (nil))
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  pc: 0x08d33984 kpexit((nil), (nil), (nil), 0x9a020900, 0x00000070)
Nov 23 08:35:23 localhost sybase_init: 00:00000:00011:2011/11/23 08:35:23.25 kernel  end of stack trace, spid 11, kpid 1755578479, suid 3
AFTER RESTART
ov 28 09:19:06 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:06.50 kernel  SySAM: Checked out license for 1 ASE_CORE (2010.04040/permanent/148F 853E 92A9 E302).
Nov 28 09:19:06 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:06.50 kernel  This product is licensed to: CISCO SYSTEMS, and OEM license from Sybase, Inc.
Nov 28 09:19:06 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:06.50 kernel  Checked out license ASE_CORENov 28 09:19:06 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:06.50 kernel  Adaptive Server Enterprise (Small Business Edition)
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.02 kernel  Using config area from primary master device.
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.17 kernel  Locking shared memory into physical memory.
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.22 kernel  Internal run-time model set for Linux  - Native
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Using 1024 file descriptors.
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Adaptive Server Enterprise/15.0.2/EBF 14331/P/Linux Intel/Linux 2.4.21-47.ELsmp i686/ase1502/2486/32-bit/FBO/Thu May 24 08:15:50 2007
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Confidential property of Sybase, Inc.Nov 28 09:19:07 localhost messagebus: messagebus startup succeeded
Nov 28 09:19:07 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Copyright 1987, 2007
Nov 28 09:19:08 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Sybase, Inc.  All rights reserved.
Nov 28 09:19:08 localhost rhnsd: Red Hat Network Services Daemon running with check_in interval set to 240 minutes.
Nov 28 09:19:08 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Unpublished rights reserved under U.S. copyright laws.
Nov 28 09:19:08 localhost rhnsd: Red Hat Network Services Daemon running with check_in interval set to 240 minutes.
Nov 28 09:19:08 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel 
Nov 28 09:19:08 localhost rhnsd[27742]: Red Hat Network Services Daemon starting up.
Nov 28 09:19:08 localhost rhnsd: rhnsd startup succeeded
Nov 28 09:19:08 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  This software contains confidential and trade secret information of Sybase,
Nov 28 09:19:09 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Inc.   Use,  duplication or disclosure of the software and documentation by
Nov 28 09:19:09 localhost cups-config-daemon: cups-config-daemon startup succeeded
Nov 28 09:19:09 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  the  U.S.  Government  is  subject  to  restrictions set forth in a license
Nov 28 09:19:09 localhost haldaemon: haldaemon startup succeeded
Nov 28 09:19:09 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  agreement  between  the  Government  and  Sybase,  Inc.  or  other  written
Nov 28 09:19:09 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  agreement  specifying  the  Government's rights to use the software and any
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  applicable FAR provisions, for example, FAR 52.227-19.
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Sybase, Inc. One Sybase Drive, Dublin, CA 94568, USA
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Using /opt/sybase as the 'SYBASE' environment variable, found during startup.
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.26 kernel  Using OCS-15_0 as the 'SYBASE_OCS' environment variable, found during startup.
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.29 kernel  ASE booted on Linux release 2.6.9-78.0.13.ELsmp version #1 SMP Wed Jan 7 17:52:47 EST 2009.
Nov 28 09:19:10 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.29 kernel  Using '/opt/sybase/ASE-15_0/pqbsyb1.cfg' for configuration information.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.29 kernel  Logging ASE messages in file '/opt/sybase/ASE-15_0/install/pqbsyb1.log'.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.29 kernel  Platform TCP network is forced to IPv4-only.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.32 kernel  ASE booted with TCP_NODELAY enabled.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.35 kernel  SSL Plus v5.0.4 security modules loaded successfully.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.35 kernel  Network and device connection limit is 1009.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.65 server  Number of blocks left for proc headers: 12760.
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:07.65 server  Proc header memory allocated 2552 pages for each per engine cache
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.16 server  Size of the 16K memory pool: 307200 Kb
Nov 28 09:19:11 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.16 server  Memory allocated for the default data cache cachelet 1: 307200 Kb
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.16 kernel  Enabling Linux Native Kernel asynchronous disk I/O strategy.
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.16 kernel  Initializing virtual device 0, '/opt/sybase/data/master.dat' with dsync 'on'.
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.16 kernel  Virtual device 0 started using asynchronous i/o.
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.17 server  Loaded default Unilib conversion handle.
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.56 kernel  Worker Thread Manager is not enabled for use in ASE.
Nov 28 09:19:12 localhost fstab-sync[28568]: removed all generated mount points
Nov 28 09:19:12 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.58 kernel  Either the config parameter 'use security services' is set to 0, or ASE does not support use of external security mechanisms on this platform. The Security Control Layer will not be initialized. No external security mechanisms will be supported.
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.58 kernel  Unix interval timer enabled for sysclk interrupts.
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.85 kernel  Begin processing to generate RSA keypair.
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.94 kernel  Completed processing to generate RSA keypair.
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.94 kernel  Encryption provider initialization succeeded on engine 0.
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.94 kernel  engine 0, os pid 27618  online
Nov 28 09:19:13 localhost sybase_init: 00:00000:00000:2011/11/28 09:19:08.94 server  No active traceflags
Nov 28 09:19:13 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.08 kernel  libomni1 - Component Integration Services: usin
g 'Sybase Client-Library/15.0/P-EBF14165 ESD #7/DRV.15.0.3/Linux Intel/Linux 2.4.21-47.0.1.ELsmp i686/BUILD1500-093/OPT/Wed Dec 13 21:46:44 2006'
Nov 28 09:19:13 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.10 server  Opening Master Database ...
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.66 server  Loading ASE's default sort order and character set
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.83 server  Recovering database 'master'.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.92 server  Started estimating recovery log boundaries for database 'master'.
Nov 28 09:19:14 localhost kernel: mtrr: type mismatch for d8000000,2000000 old: uncachable new: write-combining
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Database 'master', checkpoint=(1831, 20), first=(1831, 20), last=(1831, 32).
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Completed estimating recovery log boundaries for database 'master'.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Started ANALYSIS pass for database 'master'.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Completed ANALYSIS pass for database 'master'.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Log contains all committed transactions until 2011/11/27 13:28:54.20 for database master.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:09.96 server  Started REDO pass for database 'master'. The total number of log records to process is 13.
Nov 28 09:19:14 localhost fstab-sync[28716]: added mount point /media/cdrecorder for /dev/scd0
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.12 server  Redo pass of recovery has processed 4 committed and 0 aborted transactions.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.12 server  Completed REDO pass for database 'master'.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.12 server  Recovery of database 'master' will undo incomplete nested top actions.
Nov 28 09:19:14 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.12 server  Started recovery checkpoint for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.38 server  Completed recovery checkpoint for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:10.56 server  Started filling free space info for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:11.24 server  Completed filling free space info for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:11.26 server  Started cleaning up the default data cache for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:11.26 server  Completed cleaning up the default data cache for database 'master'.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:11.42 server  Checking external objects.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.40 server  Database 'master' is now online.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.40 server  The transaction log in the database 'master' will use I/O size of 16 Kb.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.67 server  Warning: ASE_HA has no valid license and therefore is not initialized.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.73 server  server name is 'pqbsyb1'
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.90 server  Activating disk 'sysprocsdev' of size 126976 KB.
Nov 28 09:19:15 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.95 kernel  Initializing virtual device 1, '/opt/sybase/data/sysprocs.dat' with dsync 'on'.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.95 kernel  Virtual device 1 started using asynchronous i/o.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.95 server  Activating disk 'systemdbdev' of size 49152 KB.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.99 kernel  Initializing virtual device 2, '/opt/sybase/data/sybsysdb.dat' with dsync 'on'.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.99 kernel  Virtual device 2 started using asynchronous i/o.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:12.99 server  Activating disk 'apricot_data1' of size 35082660 KB.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.02 kernel  Initializing virtual device 3, '/opt/sybase_data/apticotdata.dat' with dsync 'off'.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.02 kernel  Virtual device 3 started using asynchronous (with DIRECTIO) i/o.
Nov 28 09:19:16 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.02 server  Activating disk 'apricot_log1' of size 2980002 KB.
Nov 28 09:19:17 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.04 kernel  Initializing virtual device 4, '/opt/sybase_data/apricotlog.dat' with dsync 'off'.
Nov 28 09:19:17 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.04 kernel  Virtual device 4 started using asynchronous (with DIRECTIO) i/o.
Nov 28 09:19:17 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.04 server  Activating disk 'tempdb_dev' of size 1048576 KB.
Nov 28 09:19:17 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.09 kernel  Initializing virtual device 5, '/opt/sybase_data/tempdb.dat' with dsync 'off'.
Nov 28 09:19:17 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.09 kernel  Virtual device 5 started using asynchronous i/o.
Nov 28 09:19:18 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.30 server  Recovering database 'sybsystemdb'.
Nov 28 09:19:18 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.32 server  Started estimating recovery log boundaries for database 'sybsystemdb'.
Nov 28 09:19:18 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Database 'sybsystemdb', checkpoint=(843, 106), first=(843, 106), last=(843, 106).
Nov 28 09:19:18 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Completed estimating recovery log boundaries for database 'sybsystemdb'.
Nov 28 09:19:19 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Started ANALYSIS pass for database 'sybsystemdb'.
Nov 28 09:19:19 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Completed ANALYSIS pass for database 'sybsystemdb'.
Nov 28 09:19:19 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Log contains all committed transactions until 2011/10/25 09:55:36.72 for database sybsystemdb.
Nov 28 09:19:20 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.40 server  Started REDO pass for database 'sybsystemdb'. The total number of log records to process is 1.
Nov 28 09:19:20 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.41 server  Completed REDO pass for database 'sybsystemdb'.
Nov 28 09:19:20 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.41 server  Recovery of database 'sybsystemdb' will undo incomplete nested top actions.
Nov 28 09:19:20 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.41 server  Started recovery checkpoint for database 'sybsystemdb'.
Nov 28 09:19:21 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.46 server  Completed recovery checkpoint for database 'sybsystemdb'.
Nov 28 09:19:21 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.58 server  Started filling free space info for database 'sybsystemdb'.
Nov 28 09:19:21 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.69 server  Completed filling free space info for database 'sybsystemdb'.
Nov 28 09:19:21 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.70 server  Started cleaning up the default data cache for database 'sybsystemdb'.
Nov 28 09:19:21 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.70 server  Completed cleaning up the default data cache for database 'sybsystemdb'.
Nov 28 09:19:22 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.70 server  Boot Count: 13
Nov 28 09:19:22 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:13.81 server  Checking external objects.
Nov 28 09:19:22 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:14.18 server  The transaction log in the database 'sybsystemdb' will use I/O size of 16 Kb.
Nov 28 09:19:22 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.59 server  Completed recovery checkpoint for database 'model'.
Nov 28 09:19:22 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.70 server  Started filling free space info for database 'model'.
Nov 28 09:19:23 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.72 server  Completed filling free space info for database 'model'.
Nov 28 09:19:23 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.74 server  Started cleaning up the default data cache for database 'model'.
Nov 28 09:19:23 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.74 server  Completed cleaning up the default data cache for database 'model'.
Nov 28 09:19:23 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:15.77 server  Checking external objects.
Nov 28 09:19:23 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:16.07 server  The transaction log in the database 'model' will use I/O size of 16 Kb.
Nov 28 09:19:24 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:16.08 server  Database 'model' is now online.
Nov 28 09:19:24 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:16.08 server  The logical pagesize of the server is 16 Kb.
Nov 28 09:19:24 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:16.08 server  0 dump conditions detected at boot time
Nov 28 09:19:24 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:16.08 server  Clearing temp db
Nov 28 09:19:24 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:18.37 server  Processed 27 allocation unit(s) out of 262 units (allocation page 6656). 10% completed.
Nov 28 09:19:25 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:19.51 server  Processed 53 allocation unit(s) out of 262 units (allocation page 13312). 20% completed.
Nov 28 09:19:25 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:20.81 server  Processed 79 allocation unit(s) out of 262 units (allocation page 19968). 30% completed.
Nov 28 09:19:25 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:22.19 server  Processed 105 allocation unit(s) out of 262 units (allocation page 26624). 40% completed.
Nov 28 09:19:25 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:23.52 server  Processed 131 allocation unit(s) out of 262 units (allocation page 33280). 50% completed.
Nov 28 09:19:25 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:24.96 server  Processed 158 allocation unit(s) out of 262 units (allocation page 40192). 60% completed.
Nov 28 09:19:27 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:27.08 server  Processed 184 allocation unit(s) out of 262 units (allocation page 46848). 70% completed.
Nov 28 09:19:29 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:29.94 server  Processed 210 allocation unit(s) out of 262 units (allocation page 53504). 80% completed.
Nov 28 09:19:32 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:32.01 server  Processed 236 allocation unit(s) out of 262 units (allocation page 60160). 90% completed.
Nov 28 09:19:32 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:32.57 server  Processed 262 allocation unit(s) out of 262 units (allocation page 66816). 100% completed.
Nov 28 09:19:32 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:32.84 server  The transaction log in the database 'tempdb' will use I/O size of 16 Kb.
Nov 28 09:19:32 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:32.86 server  Database 'tempdb' is now online.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.06 server  Recovering database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.08 server  Started estimating recovery log boundaries for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Database 'sybsystemprocs', checkpoint=(6333, 93), first=(6333, 93), last=(6333, 93).
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Completed estimating recovery log boundaries for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Started ANALYSIS pass for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Completed ANALYSIS pass for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Log contains all committed transactions until 2011/10/25 09:55:36.72 for database sybsystemprocs.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Started REDO pass for database 'sybsystemprocs'. The total number of log records to process is 1.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Completed REDO pass for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Recovery of database 'sybsystemprocs' will undo incomplete nested top actions.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.14 server  Started recovery checkpoint for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.18 server  Completed recovery checkpoint for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.20 server  Started filling free space info for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.39 server  Completed filling free space info for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.40 server  Started cleaning up the default data cache for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.40 server  Completed cleaning up the default data cache for database 'sybsystemprocs'.
Nov 28 09:19:33 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:33.41 server  Checking external objects.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.04 server  The transaction log in the database 'sybsystemprocs' will use I/O size of 16 Kb.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.09 server  Database 'sybsystemprocs' is now online.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00008:2011/11/28 09:19:34.23 kernel  network name localhost.localdomain, interface IPv4, address 10.1.1.33, type tcp, port 4100, filter NONE
Nov 28 09:19:34 localhost sybase_init: 00:00000:00008:2011/11/28 09:19:34.23 kernel  network name localhost.localdomain, interface IPv4, address 127.0.0.1, type tcp, port 4100, filter NONE
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.74 server  Recovery has tuned the size of '128K' pool in 'default data cache' to benefit recovery performance. The original configuration will be restored at the end of recovery.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.74 server  Recovery has tuned the size of '16K' pool in 'default data cache' to benefit recovery performance. The original configuration will be restored at the end of recovery.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.74 server  Recovery has tuned the '128K' pool in 'default data cache' by changing its 'local async prefetch limit' from 10 to 80. The original configuration will be restored at the end of recovery.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.74 server  Recovery has tuned the '16K' pool in 'default data cache' by changing its 'local async prefetch limit' from 10 to 80. The original configuration will be restored at the end of recovery.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.74 server  The server will recover databases serially.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.81 server  Recovering database 'apricot'.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.82 server  Started estimating recovery log boundaries for database 'apricot'.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Database 'apricot', checkpoint=(2358271, 12), first=(2358271, 11), last=(2358275, 172).
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Completed estimating recovery log boundaries for database 'apricot'.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Started ANALYSIS pass for database 'apricot'.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Completed ANALYSIS pass for database 'apricot'.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Log contains all committed transactions until 2011/11/28 09:16:20.68 for database apricot.
Nov 28 09:19:34 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:34.89 server  Started REDO pass for database 'apricot'. The total number of log records to process is 846.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.10 server  Redo pass of recovery has processed 2 committed and 138 aborted transactions.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.10 server  Completed REDO pass for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.11 server  Recovery of database 'apricot' will undo incomplete nested top actions.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.11 server  Started recovery checkpoint for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.29 server  Completed recovery checkpoint for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.35 server  Started filling free space info for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.47 server  Completed filling free space info for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.51 server  Started cleaning up the default data cache for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.51 server  Completed cleaning up the default data cache for database 'apricot'.
Nov 28 09:19:35 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:35.55 server  Checking external objects.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.30 server  The transaction log in the database 'apricot' will use I/O size of 16 Kb.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.31 server  Database 'apricot' is now online.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.34 server  Recovery has restored the value of 'local async prefetch limit' for '128K' pool in 'default data cache' from '80' to 'DEFAULT'.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.34 server  Recovery has restored the value of 'local async prefetch limit' for '16K' pool in 'default data cache' from '80' to 'DEFAULT'.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.40 server  Recovery has restored the original size for '128K' pool and '16K' pool in 'default data cache'.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.44 server  Recovery complete.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server  ASE's default unicode sort order is 'binary'.
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server  ASE's default sort order is:
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server     'bin_iso_1' (ID = 50)
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server  on top of default character set:
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server     'iso_1' (ID = 1).
Nov 28 09:19:36 localhost sybase_init: 00:00000:00001:2011/11/28 09:19:36.46 server  Master device size: 240 megabytes, or 122880 virtual pages. (A virtual page is 2048 bytes.)
Nov 28 09:34:40 localhost sybase_init: 00:00000:00109:2011/11/28 09:34:40.95 kernel  Cannot send, host process disconnected: btl-sce-cm  suid: 3
Nov 28 09:34:40 localhost sybase_init: 00:00000:00109:2011/11/28 09:34:40.95 server  Error: 1608, Severity: 18, State: 3
Nov 28 09:34:40 localhost sybase_init: 00:00000:00109:2011/11/28 09:34:40.95 server  A client process exited abnormally, or a network error was encountered. Unless other errors occurred, continue processing normally.
Nov 28 09:34:40 localhost sybase_init: 00:00000:00109:2011/11/28 09:34:40.95 kernel  extended error information: hostname: btl-sce-cm login: pqb_admin
Nov 28 09:38:27 localhost sshd(pam_unix)[29201]: session opened for user root by (uid=0)

Similar Messages

  • Problems with Stack 27 for Solution Manager

    We are trying to install the stack 27 ,on Solution Manager based on Nw 701, but despite we downloaded all the packages listed with the MOPZ  the SPAM transaction continue to arise consistency problems with the queue.
    The package affected seems to be the SAPKITL437, but it's not clear why.
    I'm not able to find any Oss notes for that.
    Any advise ?

    Hi,
    Did you check SAP Note No. 1274287 at section for your ST 400 version?
    Attention: The Prerequisites for Managing sytem from
    SAP Note No. 1274287 must be fulfilled at any time.
    Regards,
    Aidan

  • SSRS 2012, SQL Server 2012. Problem with changing database

    Hi all!
    I have a problem with changing database for SSRS.
    I have a SSRS 2012 and SQL Server 2012. These programs is on WIndows Server 2012.
    I need to link SSRS to the new database instance. For this I run SSRS Configeration Manager, open tab Database, click Change Database. Then I choose option "Choose an existing report server database", click Next, write the server name, test
    connection (it's successed) and push Next again. Then I see this error message:
    Error
    The feature: "Using other editions of SQL Server for report data sources and/or the report server database" is not supported in this edition of Reporting Services.
    OK   
    I'm confused, cause I have compatible versions of programs and THIS error.
    Please, help me.
    Thanks.

    Hi Kirill,
    From the error message, we can know that the issue may be caused by incorrect editions of SQL Server for report data sources are used as the Report Server Database.
    When creating a report server database, please be aware that not all editions of SQL Server can be used to host the database. So we should make sure that we have used the correct database as the
    Report Server Database. For more details, please refer to the “Report Server Database Server Edition Requirements” section in the following document:
    http://technet.microsoft.com/en-us/library/cc645993.aspx
    Hope this helps.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Primavera Expedition 8.5 standalone with Sybase database on windows 32 bits. We need to upgrade existing expedition to Oracle contract manager 13.1.

    Primavera Expedition 8.5 standalone with Sybase database on windows 32 bits. We need to upgrade existing expedition to Oracle contract manager 13.1.
         We are unable to upgrade Sybase database version from 8.5 to 10.0 before migration to oracle database. Please share the path and details of convert.exe we are unable to find in the oracle contract manager 13.1 binaries files.
    How To: Upgrade Expedition 8.5 Sybase databases to 10.x in preparation for upgrade to Contract Management 13 (Doc ID 914273.1)
    Regards,
    Amit

    The convert.exe to which the article refers is in the v13 setup files, and not in 13.1.

  • Problem with Configuring Tomcat for running jsp web applications..Plz HELP

    I am using Tomcat 5.5 and Jdk 1.5.0_12 and Oracle 10g. I am using jdbc-odbc bridge connection
    to connect to the database. I have placed my project folder called
    tdm under the webapps folder in Tomcat. This 'tdm' folder consists of
    a collection of html pages,jsp pages and images of my project. Also I created a
    WEB-INF folderand in that I have lib folder which contains catalina-root.jar
    , classes12.jar and nls_charset.jar files. And also in the WEB-INF folder I have the web.xml
    file which looks like this
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <!--
    Copyright 2004 The Apache Software Foundation
    Licensed under the Apache License, Version 2.0 (the "License");
    you may not use this file except in compliance with the License.
    You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
    Unless required by applicable law or agreed to in writing, software
    distributed under the License is distributed on an "AS IS" BASIS,
    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    See the License for the specific language governing permissions and
    limitations under the License.
    -->
    <web-app>
    <resource-ref>
    <description>Oracle Datasource example</description>
    <res-ref-name>jdbc/gdn</res-ref-name>
    <res-type>javax.sql.DataSource</res-type>
    <res-auth>Container</res-auth>
    </resource-ref>
    </web-app>
    My Server.xml file in Tomcat\conf folder is as follows
    <!-- Example Server Configuration File -->
    <!-- Note that component elements are nested corresponding to their
    parent-child relationships with each other -->
    <!-- A "Server" is a singleton element that represents the entire JVM,
    which may contain one or more "Service" instances. The Server
    listens for a shutdown command on the indicated port.
    Note: A "Server" is not itself a "Container", so you may not
    define subcomponents such as "Valves" or "Loggers" at this level.
    -->
    <Server port="8005" shutdown="SHUTDOWN">
    <!-- Comment these entries out to disable JMX MBeans support used for the
    administration web application -->
    <Listener className="org.apache.catalina.mbeans.ServerLifecycleListener" />
    <Listener className="org.apache.catalina.mbeans.GlobalResourcesLifecycleListener" />
    <!-- Global JNDI resources -->
    <GlobalNamingResources>
    <!-- Test entry for demonstration purposes -->
    <Environment name="simpleValue" type="java.lang.Integer" value="30"/>
    <!-- Editable user database that can also be used by
    UserDatabaseRealm to authenticate users -->
    <Resource name="UserDatabase" auth="Container"
    type="org.apache.catalina.UserDatabase"
    description="User database that can be updated and saved"
    factory="org.apache.catalina.users.MemoryUserDatabaseFactory"
    pathname="conf/tomcat-users.xml" />
    <Resource name="jdbc/gdn" auth="Container"
    type="javax.sql.DataSource" driverClassName="sun.jdbc.odbc.JdbcOdbcDriver"
    url="jdbc:odbc:gdn"
    username="system" password="tiger" maxActive="20" maxIdle="10"
    maxWait="-1"/>
    </GlobalNamingResources>
    <!-- A "Service" is a collection of one or more "Connectors" that share
    a single "Container" (and therefore the web applications visible
    within that Container). Normally, that Container is an "Engine",
    but this is not required.
    Note: A "Service" is not itself a "Container", so you may not
    define subcomponents such as "Valves" or "Loggers" at this level.
    -->
    <!-- Define the Tomcat Stand-Alone Service -->
    <Service name="Catalina">
    <!-- A "Connector" represents an endpoint by which requests are received
    and responses are returned. Each Connector passes requests on to the
    associated "Container" (normally an Engine) for processing.
    By default, a non-SSL HTTP/1.1 Connector is established on port 8080.
    You can also enable an SSL HTTP/1.1 Connector on port 8443 by
    following the instructions below and uncommenting the second Connector
    entry. SSL support requires the following steps (see the SSL Config
    HOWTO in the Tomcat 5 documentation bundle for more detailed
    instructions):
    * If your JDK version 1.3 or prior, download and install JSSE 1.0.2 or
    later, and put the JAR files into "$JAVA_HOME/jre/lib/ext".
    * Execute:
    %JAVA_HOME%\bin\keytool -genkey -alias tomcat -keyalg RSA (Windows)
    $JAVA_HOME/bin/keytool -genkey -alias tomcat -keyalg RSA (Unix)
    with a password value of "changeit" for both the certificate and
    the keystore itself.
    By default, DNS lookups are enabled when a web application calls
    request.getRemoteHost(). This can have an adverse impact on
    performance, so you can disable it by setting the
    "enableLookups" attribute to "false". When DNS lookups are disabled,
    request.getRemoteHost() will return the String version of the
    IP address of the remote client.
    -->
    <!-- Define a non-SSL HTTP/1.1 Connector on port 8080 -->
    <Connector
    port="5050" maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" redirectPort="8443" acceptCount="100"
    connectionTimeout="20000" disableUploadTimeout="true" />
    <!-- Note : To disable connection timeouts, set connectionTimeout value
    to 0 -->
         <!-- Note : To use gzip compression you could set the following properties :
                   compression="on"
                   compressionMinSize="2048"
                   noCompressionUserAgents="gozilla, traviata"
                   compressableMimeType="text/html,text/xml"
         -->
    <!-- Define a SSL HTTP/1.1 Connector on port 8443 -->
    <!--
    <Connector port="8443"
    maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" disableUploadTimeout="true"
    acceptCount="100" scheme="https" secure="true"
    clientAuth="false" sslProtocol="TLS" />
    -->
    <!-- Define an AJP 1.3 Connector on port 8009 -->
    <Connector port="8009"
    enableLookups="false" redirectPort="8443" protocol="AJP/1.3" />
    <!-- Define a Proxied HTTP/1.1 Connector on port 8082 -->
    <!-- See proxy documentation for more information about using this. -->
    <!--
    <Connector port="8082"
    maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" acceptCount="100" connectionTimeout="20000"
    proxyPort="80" disableUploadTimeout="true" />
    -->
    <!-- An Engine represents the entry point (within Catalina) that processes
    every request. The Engine implementation for Tomcat stand alone
    analyzes the HTTP headers included with the request, and passes them
    on to the appropriate Host (virtual host). -->
    <!-- You should set jvmRoute to support load-balancing via AJP ie :
    <Engine name="Standalone" defaultHost="localhost" jvmRoute="jvm1">
    -->
    <!-- Define the top level container in our container hierarchy -->
    <Engine name="Catalina" defaultHost="localhost">
    <!-- The request dumper valve dumps useful debugging information about
    the request headers and cookies that were received, and the response
    headers and cookies that were sent, for all requests received by
    this instance of Tomcat. If you care only about requests to a
    particular virtual host, or a particular application, nest this
    element inside the corresponding <Host> or <Context> entry instead.
    For a similar mechanism that is portable to all Servlet 2.4
    containers, check out the "RequestDumperFilter" Filter in the
    example application (the source for this filter may be found in
    "$CATALINA_HOME/webapps/examples/WEB-INF/classes/filters").
    Request dumping is disabled by default. Uncomment the following
    element to enable it. -->
    <!--
    <Valve className="org.apache.catalina.valves.RequestDumperValve"/>
    -->
    <!-- Because this Realm is here, an instance will be shared globally -->
    <!-- This Realm uses the UserDatabase configured in the global JNDI
    resources under the key "UserDatabase". Any edits
    that are performed against this UserDatabase are immediately
    available for use by the Realm. -->
    <Realm className="org.apache.catalina.realm.UserDatabaseRealm"
    resourceName="UserDatabase"/>
    <!-- Comment out the old realm but leave here for now in case we
    need to go back quickly -->
    <!--
    <Realm className="org.apache.catalina.realm.MemoryRealm" />
    -->
    <!-- Replace the above Realm with one of the following to get a Realm
    stored in a database and accessed via JDBC -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="org.gjt.mm.mysql.Driver"
    connectionURL="jdbc:mysql://localhost/authority"
    connectionName="test" connectionPassword="test"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="oracle.jdbc.driver.OracleDriver"
    connectionURL="jdbc:oracle:thin:@ntserver:1521:ORCL"
    connectionName="scott" connectionPassword="tiger"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="sun.jdbc.odbc.JdbcOdbcDriver"
    connectionURL="jdbc:odbc:CATALINA"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!-- Define the default virtual host
    Note: XML Schema validation will not work with Xerces 2.2.
    -->
    <Host name="localhost" appBase="webapps"
    unpackWARs="true" autoDeploy="true"
    xmlValidation="false" xmlNamespaceAware="false">
    <!-- Defines a cluster for this node,
    By defining this element, means that every manager will be changed.
    So when running a cluster, only make sure that you have webapps in there
    that need to be clustered and remove the other ones.
    A cluster has the following parameters:
    className = the fully qualified name of the cluster class
    name = a descriptive name for your cluster, can be anything
    mcastAddr = the multicast address, has to be the same for all the nodes
    mcastPort = the multicast port, has to be the same for all the nodes
    mcastBindAddr = bind the multicast socket to a specific address
    mcastTTL = the multicast TTL if you want to limit your broadcast
    mcastSoTimeout = the multicast readtimeout
    mcastFrequency = the number of milliseconds in between sending a "I'm alive" heartbeat
    mcastDropTime = the number a milliseconds before a node is considered "dead" if no heartbeat is received
    tcpThreadCount = the number of threads to handle incoming replication requests, optimal would be the same amount of threads as nodes
    tcpListenAddress = the listen address (bind address) for TCP cluster request on this host,
    in case of multiple ethernet cards.
    auto means that address becomes
    InetAddress.getLocalHost().getHostAddress()
    tcpListenPort = the tcp listen port
    tcpSelectorTimeout = the timeout (ms) for the Selector.select() method in case the OS
    has a wakup bug in java.nio. Set to 0 for no timeout
    printToScreen = true means that managers will also print to std.out
    expireSessionsOnShutdown = true means that
    useDirtyFlag = true means that we only replicate a session after setAttribute,removeAttribute has been called.
    false means to replicate the session after each request.
    false means that replication would work for the following piece of code: (only for SimpleTcpReplicationManager)
    <%
    HashMap map = (HashMap)session.getAttribute("map");
    map.put("key","value");
    %>
    replicationMode = can be either 'pooled', 'synchronous' or 'asynchronous'.
    * Pooled means that the replication happens using several sockets in a synchronous way. Ie, the data gets replicated, then the request return. This is the same as the 'synchronous' setting except it uses a pool of sockets, hence it is multithreaded. This is the fastest and safest configuration. To use this, also increase the nr of tcp threads that you have dealing with replication.
    * Synchronous means that the thread that executes the request, is also the
    thread the replicates the data to the other nodes, and will not return until all
    nodes have received the information.
    * Asynchronous means that there is a specific 'sender' thread for each cluster node,
    so the request thread will queue the replication request into a "smart" queue,
    and then return to the client.
    The "smart" queue is a queue where when a session is added to the queue, and the same session
    already exists in the queue from a previous request, that session will be replaced
    in the queue instead of replicating two requests. This almost never happens, unless there is a
    large network delay.
    -->
    <!--
    When configuring for clustering, you also add in a valve to catch all the requests
    coming in, at the end of the request, the session may or may not be replicated.
    A session is replicated if and only if all the conditions are met:
    1. useDirtyFlag is true or setAttribute or removeAttribute has been called AND
    2. a session exists (has been created)
    3. the request is not trapped by the "filter" attribute
    The filter attribute is to filter out requests that could not modify the session,
    hence we don't replicate the session after the end of this request.
    The filter is negative, ie, anything you put in the filter, you mean to filter out,
    ie, no replication will be done on requests that match one of the filters.
    The filter attribute is delimited by ;, so you can't escape out ; even if you wanted to.
    filter=".*\.gif;.*\.js;" means that we will not replicate the session after requests with the URI
    ending with .gif and .js are intercepted.
    The deployer element can be used to deploy apps cluster wide.
    Currently the deployment only deploys/undeploys to working members in the cluster
    so no WARs are copied upons startup of a broken node.
    The deployer watches a directory (watchDir) for WAR files when watchEnabled="true"
    When a new war file is added the war gets deployed to the local instance,
    and then deployed to the other instances in the cluster.
    When a war file is deleted from the watchDir the war is undeployed locally
    and cluster wide
    -->
    <!--
    <Cluster className="org.apache.catalina.cluster.tcp.SimpleTcpCluster"
    managerClassName="org.apache.catalina.cluster.session.DeltaManager"
    expireSessionsOnShutdown="false"
    useDirtyFlag="true"
    notifyListenersOnReplication="true">
    <Membership
    className="org.apache.catalina.cluster.mcast.McastService"
    mcastAddr="228.0.0.4"
    mcastPort="45564"
    mcastFrequency="500"
    mcastDropTime="3000"/>
    <Receiver
    className="org.apache.catalina.cluster.tcp.ReplicationListener"
    tcpListenAddress="auto"
    tcpListenPort="4001"
    tcpSelectorTimeout="100"
    tcpThreadCount="6"/>
    <Sender
    className="org.apache.catalina.cluster.tcp.ReplicationTransmitter"
    replicationMode="pooled"
    ackTimeout="15000"/>
    <Valve className="org.apache.catalina.cluster.tcp.ReplicationValve"
    filter=".*\.gif;.*\.js;.*\.jpg;.*\.htm;.*\.html;.*\.txt;"/>
    <Deployer className="org.apache.catalina.cluster.deploy.FarmWarDeployer"
    tempDir="/tmp/war-temp/"
    deployDir="/tmp/war-deploy/"
    watchDir="/tmp/war-listen/"
    watchEnabled="false"/>
    </Cluster>
    -->
    <!-- Normally, users must authenticate themselves to each web app
    individually. Uncomment the following entry if you would like
    a user to be authenticated the first time they encounter a
    resource protected by a security constraint, and then have that
    user identity maintained across all web applications contained
    in this virtual host. -->
    <!--
    <Valve className="org.apache.catalina.authenticator.SingleSignOn" />
    -->
    <!-- Access log processes all requests for this virtual host. By
    default, log files are created in the "logs" directory relative to
    $CATALINA_HOME. If you wish, you can specify a different
    directory with the "directory" attribute. Specify either a relative
    (to $CATALINA_HOME) or absolute path to the desired directory.
    -->
    <!--
    <Valve className="org.apache.catalina.valves.AccessLogValve"
    directory="logs" prefix="localhost_access_log." suffix=".txt"
    pattern="common" resolveHosts="false"/>
    -->
    <!-- Access log processes all requests for this virtual host. By
    default, log files are created in the "logs" directory relative to
    $CATALINA_HOME. If you wish, you can specify a different
    directory with the "directory" attribute. Specify either a relative
    (to $CATALINA_HOME) or absolute path to the desired directory.
    This access log implementation is optimized for maximum performance,
    but is hardcoded to support only the "common" and "combined" patterns.
    -->
    <!--
    <Valve className="org.apache.catalina.valves.FastCommonAccessLogValve"
    directory="logs" prefix="localhost_access_log." suffix=".txt"
    pattern="common" resolveHosts="false"/>
    -->
    <Context path="/tdm" docBase="tdm" debug="0" reloadable="true" />
    </Host>
    </Engine>
    </Service>
    </Server>
    I have set the context path to /tdm in the server.xml file. Should this be placed in context.xml?
    My first page in the project is called Homepage.html. To start my project I give http://localhost:5050/tdm/homepage.html
    in a browser. Here I accept a username and password from the user and then do the validation in
    a valid.jsp file, where I connect to the database and check and use jsp:forward to go to next pages
    accordingly. However when I enter the username and password and click Go in the homepage, nothing is
    displayed on the next page. The URL in the browser says valid.jsp but a blank screen appears.
    WHY DOES IT HAPPEN SO? DOES IT MEAN THAT TOMCAT IS NOT RECOGNIZING JAVA IN MY SYSTEM OR IS IT A PROBLEM
    WITH THE DATABASE CONNECTION OR SOMETHING ELSE? I FEEL THAT TOMCAT IS NOT EXECUTING JSP COMMANDS?
    IS IT POSSIBLE?WHY WILL THIS HAPPEN?
    I set the JAVA_HOME and CATALINA_HOME environment to the jdk and tomcat folders resp.
    Is there any other thing that I need to set in classpath? Should I have my project as a
    WAR file in the webapps of TOMCAT or just a folder i.e. directory structure will fine?

    I am using Tomcat 5.5 and Jdk 1.5.0_12 and Oracle 10g. I am using jdbc-odbc bridge connection
    to connect to the database. I have placed my project folder called
    tdm under the webapps folder in Tomcat. This 'tdm' folder consists of
    a collection of html pages,jsp pages and images of my project. Also I created a
    WEB-INF folderand in that I have lib folder which contains catalina-root.jar
    , classes12.jar and nls_charset.jar files. And also in the WEB-INF folder I have the web.xml
    file which looks like this
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <!--
    Copyright 2004 The Apache Software Foundation
    Licensed under the Apache License, Version 2.0 (the "License");
    you may not use this file except in compliance with the License.
    You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
    Unless required by applicable law or agreed to in writing, software
    distributed under the License is distributed on an "AS IS" BASIS,
    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    See the License for the specific language governing permissions and
    limitations under the License.
    -->
    <web-app>
    <resource-ref>
    <description>Oracle Datasource example</description>
    <res-ref-name>jdbc/gdn</res-ref-name>
    <res-type>javax.sql.DataSource</res-type>
    <res-auth>Container</res-auth>
    </resource-ref>
    </web-app>
    My Server.xml file in Tomcat\conf folder is as follows
    <!-- Example Server Configuration File -->
    <!-- Note that component elements are nested corresponding to their
    parent-child relationships with each other -->
    <!-- A "Server" is a singleton element that represents the entire JVM,
    which may contain one or more "Service" instances. The Server
    listens for a shutdown command on the indicated port.
    Note: A "Server" is not itself a "Container", so you may not
    define subcomponents such as "Valves" or "Loggers" at this level.
    -->
    <Server port="8005" shutdown="SHUTDOWN">
    <!-- Comment these entries out to disable JMX MBeans support used for the
    administration web application -->
    <Listener className="org.apache.catalina.mbeans.ServerLifecycleListener" />
    <Listener className="org.apache.catalina.mbeans.GlobalResourcesLifecycleListener" />
    <!-- Global JNDI resources -->
    <GlobalNamingResources>
    <!-- Test entry for demonstration purposes -->
    <Environment name="simpleValue" type="java.lang.Integer" value="30"/>
    <!-- Editable user database that can also be used by
    UserDatabaseRealm to authenticate users -->
    <Resource name="UserDatabase" auth="Container"
    type="org.apache.catalina.UserDatabase"
    description="User database that can be updated and saved"
    factory="org.apache.catalina.users.MemoryUserDatabaseFactory"
    pathname="conf/tomcat-users.xml" />
    <Resource name="jdbc/gdn" auth="Container"
    type="javax.sql.DataSource" driverClassName="sun.jdbc.odbc.JdbcOdbcDriver"
    url="jdbc:odbc:gdn"
    username="system" password="tiger" maxActive="20" maxIdle="10"
    maxWait="-1"/>
    </GlobalNamingResources>
    <!-- A "Service" is a collection of one or more "Connectors" that share
    a single "Container" (and therefore the web applications visible
    within that Container). Normally, that Container is an "Engine",
    but this is not required.
    Note: A "Service" is not itself a "Container", so you may not
    define subcomponents such as "Valves" or "Loggers" at this level.
    -->
    <!-- Define the Tomcat Stand-Alone Service -->
    <Service name="Catalina">
    <!-- A "Connector" represents an endpoint by which requests are received
    and responses are returned. Each Connector passes requests on to the
    associated "Container" (normally an Engine) for processing.
    By default, a non-SSL HTTP/1.1 Connector is established on port 8080.
    You can also enable an SSL HTTP/1.1 Connector on port 8443 by
    following the instructions below and uncommenting the second Connector
    entry. SSL support requires the following steps (see the SSL Config
    HOWTO in the Tomcat 5 documentation bundle for more detailed
    instructions):
    * If your JDK version 1.3 or prior, download and install JSSE 1.0.2 or
    later, and put the JAR files into "$JAVA_HOME/jre/lib/ext".
    * Execute:
    %JAVA_HOME%\bin\keytool -genkey -alias tomcat -keyalg RSA (Windows)
    $JAVA_HOME/bin/keytool -genkey -alias tomcat -keyalg RSA (Unix)
    with a password value of "changeit" for both the certificate and
    the keystore itself.
    By default, DNS lookups are enabled when a web application calls
    request.getRemoteHost(). This can have an adverse impact on
    performance, so you can disable it by setting the
    "enableLookups" attribute to "false". When DNS lookups are disabled,
    request.getRemoteHost() will return the String version of the
    IP address of the remote client.
    -->
    <!-- Define a non-SSL HTTP/1.1 Connector on port 8080 -->
    <Connector
    port="5050" maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" redirectPort="8443" acceptCount="100"
    connectionTimeout="20000" disableUploadTimeout="true" />
    <!-- Note : To disable connection timeouts, set connectionTimeout value
    to 0 -->
         <!-- Note : To use gzip compression you could set the following properties :
                   compression="on"
                   compressionMinSize="2048"
                   noCompressionUserAgents="gozilla, traviata"
                   compressableMimeType="text/html,text/xml"
         -->
    <!-- Define a SSL HTTP/1.1 Connector on port 8443 -->
    <!--
    <Connector port="8443"
    maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" disableUploadTimeout="true"
    acceptCount="100" scheme="https" secure="true"
    clientAuth="false" sslProtocol="TLS" />
    -->
    <!-- Define an AJP 1.3 Connector on port 8009 -->
    <Connector port="8009"
    enableLookups="false" redirectPort="8443" protocol="AJP/1.3" />
    <!-- Define a Proxied HTTP/1.1 Connector on port 8082 -->
    <!-- See proxy documentation for more information about using this. -->
    <!--
    <Connector port="8082"
    maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
    enableLookups="false" acceptCount="100" connectionTimeout="20000"
    proxyPort="80" disableUploadTimeout="true" />
    -->
    <!-- An Engine represents the entry point (within Catalina) that processes
    every request. The Engine implementation for Tomcat stand alone
    analyzes the HTTP headers included with the request, and passes them
    on to the appropriate Host (virtual host). -->
    <!-- You should set jvmRoute to support load-balancing via AJP ie :
    <Engine name="Standalone" defaultHost="localhost" jvmRoute="jvm1">
    -->
    <!-- Define the top level container in our container hierarchy -->
    <Engine name="Catalina" defaultHost="localhost">
    <!-- The request dumper valve dumps useful debugging information about
    the request headers and cookies that were received, and the response
    headers and cookies that were sent, for all requests received by
    this instance of Tomcat. If you care only about requests to a
    particular virtual host, or a particular application, nest this
    element inside the corresponding <Host> or <Context> entry instead.
    For a similar mechanism that is portable to all Servlet 2.4
    containers, check out the "RequestDumperFilter" Filter in the
    example application (the source for this filter may be found in
    "$CATALINA_HOME/webapps/examples/WEB-INF/classes/filters").
    Request dumping is disabled by default. Uncomment the following
    element to enable it. -->
    <!--
    <Valve className="org.apache.catalina.valves.RequestDumperValve"/>
    -->
    <!-- Because this Realm is here, an instance will be shared globally -->
    <!-- This Realm uses the UserDatabase configured in the global JNDI
    resources under the key "UserDatabase". Any edits
    that are performed against this UserDatabase are immediately
    available for use by the Realm. -->
    <Realm className="org.apache.catalina.realm.UserDatabaseRealm"
    resourceName="UserDatabase"/>
    <!-- Comment out the old realm but leave here for now in case we
    need to go back quickly -->
    <!--
    <Realm className="org.apache.catalina.realm.MemoryRealm" />
    -->
    <!-- Replace the above Realm with one of the following to get a Realm
    stored in a database and accessed via JDBC -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="org.gjt.mm.mysql.Driver"
    connectionURL="jdbc:mysql://localhost/authority"
    connectionName="test" connectionPassword="test"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="oracle.jdbc.driver.OracleDriver"
    connectionURL="jdbc:oracle:thin:@ntserver:1521:ORCL"
    connectionName="scott" connectionPassword="tiger"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!--
    <Realm className="org.apache.catalina.realm.JDBCRealm"
    driverName="sun.jdbc.odbc.JdbcOdbcDriver"
    connectionURL="jdbc:odbc:CATALINA"
    userTable="users" userNameCol="user_name" userCredCol="user_pass"
    userRoleTable="user_roles" roleNameCol="role_name" />
    -->
    <!-- Define the default virtual host
    Note: XML Schema validation will not work with Xerces 2.2.
    -->
    <Host name="localhost" appBase="webapps"
    unpackWARs="true" autoDeploy="true"
    xmlValidation="false" xmlNamespaceAware="false">
    <!-- Defines a cluster for this node,
    By defining this element, means that every manager will be changed.
    So when running a cluster, only make sure that you have webapps in there
    that need to be clustered and remove the other ones.
    A cluster has the following parameters:
    className = the fully qualified name of the cluster class
    name = a descriptive name for your cluster, can be anything
    mcastAddr = the multicast address, has to be the same for all the nodes
    mcastPort = the multicast port, has to be the same for all the nodes
    mcastBindAddr = bind the multicast socket to a specific address
    mcastTTL = the multicast TTL if you want to limit your broadcast
    mcastSoTimeout = the multicast readtimeout
    mcastFrequency = the number of milliseconds in between sending a "I'm alive" heartbeat
    mcastDropTime = the number a milliseconds before a node is considered "dead" if no heartbeat is received
    tcpThreadCount = the number of threads to handle incoming replication requests, optimal would be the same amount of threads as nodes
    tcpListenAddress = the listen address (bind address) for TCP cluster request on this host,
    in case of multiple ethernet cards.
    auto means that address becomes
    InetAddress.getLocalHost().getHostAddress()
    tcpListenPort = the tcp listen port
    tcpSelectorTimeout = the timeout (ms) for the Selector.select() method in case the OS
    has a wakup bug in java.nio. Set to 0 for no timeout
    printToScreen = true means that managers will also print to std.out
    expireSessionsOnShutdown = true means that
    useDirtyFlag = true means that we only replicate a session after setAttribute,removeAttribute has been called.
    false means to replicate the session after each request.
    false means that replication would work for the following piece of code: (only for SimpleTcpReplicationManager)
    <%
    HashMap map = (HashMap)session.getAttribute("map");
    map.put("key","value");
    %>
    replicationMode = can be either 'pooled', 'synchronous' or 'asynchronous'.
    * Pooled means that the replication happens using several sockets in a synchronous way. Ie, the data gets replicated, then the request return. This is the same as the 'synchronous' setting except it uses a pool of sockets, hence it is multithreaded. This is the fastest and safest configuration. To use this, also increase the nr of tcp threads that you have dealing with replication.
    * Synchronous means that the thread that executes the request, is also the
    thread the replicates the data to the other nodes, and will not return until all
    nodes have received the information.
    * Asynchronous means that there is a specific 'sender' thread for each cluster node,
    so the request thread will queue the replication request into a "smart" queue,
    and then return to the client.
    The "smart" queue is a queue where when a session is added to the queue, and the same session
    already exists in the queue from a previous request, that session will be replaced
    in the queue instead of replicating two requests. This almost never happens, unless there is a
    large network delay.
    -->
    <!--
    When configuring for clustering, you also add in a valve to catch all the requests
    coming in, at the end of the request, the session may or may not be replicated.
    A session is replicated if and only if all the conditions are met:
    1. useDirtyFlag is true or setAttribute or removeAttribute has been called AND
    2. a session exists (has been created)
    3. the request is not trapped by the "filter" attribute
    The filter attribute is to filter out requests that could not modify the session,
    hence we don't replicate the session after the end of this request.
    The filter is negative, ie, anything you put in the filter, you mean to filter out,
    ie, no replication will be done on requests that match one of the filters.
    The filter attribute is delimited by ;, so you can't escape out ; even if you wanted to.
    filter=".*\.gif;.*\.js;" means that we will not replicate the session after requests with the URI
    ending with .gif and .js are intercepted.
    The deployer element can be used to deploy apps cluster wide.
    Currently the deployment only deploys/undeploys to working members in the cluster
    so no WARs are copied upons startup of a broken node.
    The deployer watches a directory (watchDir) for WAR files when watchEnabled="true"
    When a new war file is added the war gets deployed to the local instance,
    and then deployed to the other instances in the cluster.
    When a war file is deleted from the watchDir the war is undeployed locally
    and cluster wide
    -->
    <!--
    <Cluster className="org.apache.catalina.cluster.tcp.SimpleTcpCluster"
    managerClassName="org.apache.catalina.cluster.session.DeltaManager"
    expireSessionsOnShutdown="false"
    useDirtyFlag="true"
    notifyListenersOnReplication="true">
    <Membership
    className="org.apache.catalina.cluster.mcast.McastService"
    mcastAddr="228.0.0.4"
    mcastPort="45564"
    mcastFrequency="500"
    mcastDropTime="3000"/>
    <Receiver
    className="org.apache.catalina.cluster.tcp.ReplicationListener"
    tcpListenAddress="auto"
    tcpListenPort="4001"
    tcpSelectorTimeout="100"
    tcpThreadCount="6"/>
    <Sender
    className="org.apache.catalina.cluster.tcp.ReplicationTransmitter"
    replicationMode="pooled"
    ackTimeout="15000"/>
    <Valve className="org.apache.catalina.cluster.tcp.ReplicationValve"
    filter=".*\.gif;.*\.js;.*\.jpg;.*\.htm;.*\.html;.*\.txt;"/>
    <Deployer className="org.apache.catalina.cluster.deploy.FarmWarDeployer"
    tempDir="/tmp/war-temp/"
    deployDir="/tmp/war-deploy/"
    watchDir="/tmp/war-listen/"
    watchEnabled="false"/>
    </Cluster>
    -->
    <!-- Normally, users must authenticate themselves to each web app
    individually. Uncomment the following entry if you would like
    a user to be authenticated the first time they encounter a
    resource protected by a security constraint, and then have that
    user identity maintained across all web applications contained
    in this virtual host. -->
    <!--
    <Valve className="org.apache.catalina.authenticator.SingleSignOn" />
    -->
    <!-- Access log processes all requests for this virtual host. By
    default, log files are created in the "logs" directory relative to
    $CATALINA_HOME. If you wish, you can specify a different
    directory with the "directory" attribute. Specify either a relative
    (to $CATALINA_HOME) or absolute path to the desired directory.
    -->
    <!--
    <Valve className="org.apache.catalina.valves.AccessLogValve"
    directory="logs" prefix="localhost_access_log." suffix=".txt"
    pattern="common" resolveHosts="false"/>
    -->
    <!-- Access log processes all requests for this virtual host. By
    default, log files are created in the "logs" directory relative to
    $CATALINA_HOME. If you wish, you can specify a different
    directory with the "directory" attribute. Specify either a relative
    (to $CATALINA_HOME) or absolute path to the desired directory.
    This access log implementation is optimized for maximum performance,
    but is hardcoded to support only the "common" and "combined" patterns.
    -->
    <!--
    <Valve className="org.apache.catalina.valves.FastCommonAccessLogValve"
    directory="logs" prefix="localhost_access_log." suffix=".txt"
    pattern="common" resolveHosts="false"/>
    -->
    <Context path="/tdm" docBase="tdm" debug="0" reloadable="true" />
    </Host>
    </Engine>
    </Service>
    </Server>
    I have set the context path to /tdm in the server.xml file. Should this be placed in context.xml?
    My first page in the project is called Homepage.html. To start my project I give http://localhost:5050/tdm/homepage.html
    in a browser. Here I accept a username and password from the user and then do the validation in
    a valid.jsp file, where I connect to the database and check and use jsp:forward to go to next pages
    accordingly. However when I enter the username and password and click Go in the homepage, nothing is
    displayed on the next page. The URL in the browser says valid.jsp but a blank screen appears.
    WHY DOES IT HAPPEN SO? DOES IT MEAN THAT TOMCAT IS NOT RECOGNIZING JAVA IN MY SYSTEM OR IS IT A PROBLEM
    WITH THE DATABASE CONNECTION OR SOMETHING ELSE? I FEEL THAT TOMCAT IS NOT EXECUTING JSP COMMANDS?
    IS IT POSSIBLE?WHY WILL THIS HAPPEN?
    I set the JAVA_HOME and CATALINA_HOME environment to the jdk and tomcat folders resp.
    Is there any other thing that I need to set in classpath? Should I have my project as a
    WAR file in the webapps of TOMCAT or just a folder i.e. directory structure will fine?

  • I have a problem with the indicator for power on the side of my HP DV6-1341el.

    I have a problem with the indicator for power on the side of my HP DV6-1341el. Insert the adapter's power jack on the notebook in front of the LED battery charging works, windows detects that the battery is charged, but if I use the diagnostic programs, they tell me they found that the computer is running on battery power and not AC mode.
    Where is the problem?
    I have also made the maneuver for the recognition by the operating system of the battery and does not seem to be there the problem.
    If I use the laptop without battery in AC mode happens that led to the side of the notebook blinks twice, three times and then goes off when I put the power jack.

    Go into the Device Manager. Expand the battery device, select the ACPI-Compliant Battery Control Method device, right-click it and select uninstall. After you have done that, right-click any device in the Device Manager and select scan for hardware changes. Your operating system will reinstall the  ACPI-Compliant Battery Control Method device. 
    Post your results whether they are negative or positive to keep us posted.
    ****Please click on Accept As Solution if a suggestion solves your problem. It helps others facing the same problem to find a solution easily****
    2015 Microsoft MVP - Windows Experience Consumer

  • Problem with Logical Database DDF

    Hey All,
    I have a problem with Logical Database DDF.
    In my program I am typing this
    GET KNA1 based on selection screen. It is bringing back everything I want except BRAN1 whereby the field is ''. However in SE16 for KNA1 and for the same KUNNR, the field is populated with 'T180'
    I am not understanding this?! I thought LDB where menat to make things easy!!

    I think it's not used by DDF.
    so you must check it with:
    check kna1-bran1 in s_bran1.
    A.

  • Problem with ORACLE database

    Hi all,
    I am facing some problem with ORACLE database while configuring one JDBC scenario in the receiver side. I have configured one JDBC scenario like FILE TO JDBC. As JDBC is at receiver side i have provided the input file with the following format ,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>employee</table>
              <access>
                    <flag>NO</flag> 
                    <ID>567</ID>
             <NAME>134</NAME>
                 </access>
         </dbTableName>
      </StatementName1>
    </root>
    But in the ORACLE database all my columns are in the uppercase.
    So when i have tried to send this input file to process through JDBC receiver side then i have faced the below error :
    FATAL ERROR: Column 'flag' does not exist in table 'employee'
    Then I have tried to check with sender side processing of JDBC adapter. By using the SQL statements as
    SELECT * FROM EMPLOYEE WHERE flag='NO'
    UPDATE EMPLOYEE SET Flag='YES'  WHERE flag='NO'
    Here it worked fine.
    Now my question is,
    If the same adapter is working when we are writing a query directly(sender side), then we must also allow case independence in the auto generated SQL part(receiver side).
    The JDBC drivers for both SQL server and Oracle, supports the same java code, for accessing the DB, irrespective of the case of the column names.
    Please clarify this doubt as soon as possible.
    Thanks,
    Soorya.

    Hi Dharmaveer,
    When I go with the following input file,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>EMPLOYEE</table>
              <access>
                    <FLAG>NO</FLAG> 
                    <ID>666</ID>
                    <NAME>SSS</NAME>
                 </access>
             </dbTableName>
      </StatementName1>
    </root>
    I will get the following query with succesfull processing.
    INSERT INTO  EMPLOYEE (FLAG, ID, NAME) VALUES (NO, 666, SSS)
    when i go with this input file,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>EMPLOYEE</table>
              <access>
                                    <flag>NO</flag> 
                                   <ID>777</ID>
                        <NAME>TTTT</NAME>
                        </access>
                   </dbTableName>
      </StatementName1>
    </root>
    Then i am facing this problem,
    Message processing failed. Cause: com.sap.aii.af.ra.ms.api.RecoverableException: Error processing request in sax parser: Error when executing statement for table/stored proc. 'EMPLOYEE' (structure 'StatementName1'): java.sql.SQLException: FATAL ERROR: Column 'flag' does not exist in table 'EMPLOYEE'
    Here I cant get any SQL statement as it is not successfull.
    I have gone through your link but it also says to modify the column in the input file which is not possible for me. Can you please help me in this regard?
    Thanks,
    Soorya

  • TS5249 Problems with the Korg Legacy Collection in the new version of Garageband in Mavericks.

    Hello!
    I'm having problems with the Korg Legacy Collection in the new version of Garageband in Mavericks. When I try to load the AU, I get this error message:
    "Failed to load Audio Unit, please contact the manufacturer for an updated version or further assistance. (-10110)"
    They seem to work fine in Ableton Live 8 Lite, also running in Mavericks. So, I think it is something with Garageband. In addition, Korg updated to 64bit early last year.
    Anyone know how to fix the problem? Or is Korg or Apple's problem?
    Thanks.

    They seem to work fine in Ableton Live 8 Lite, also running in Mavericks. So, I think it is something with Garageband. In addition, Korg updated to 64bit early last year.
    Have you checked, if you have the newest 64-bit version installed? There may still be an older version sitting around - check both plug-in folders, the one in your Home folder and the one in the system library:
    ~/Library/Audio/Plug-Ins/
    /Library/Audio/Plug-Ins/
    Some programs may still be able to use 64-bit plug-ins, but GarageBand will not use them:  http://support.apple.com/kb/TS5249

  • I have been a loyal customer for years and have a hot spot because internet options are very limited in our rural area.  I have stayed with 10G since the beginning.  Most months I don't use the 10G.  Verizon has no problem with me paying for Data I don't

    I have been a loyal customer for years and have a hot spot because internet options are very limited in our rural area.  I have stayed with 10G since the beginning.  Most months I don't use the 10G.  Verizon has no problem with me paying for Data I don't use.  This month my daughter comes home from College and she accidentally uses 24G.....14 over get blind sided with $140.00 in overages.  I called Verizon today......stopped in to the store today......My neighbor told me they waved charges for her one time.  Lots of charges.....But I'm am getting no help.  Not even an offer of a payment plan to help me out.  I went back just 6 months and I have over paid for 19G.....Seems like they would like to help out their loyal customers!!!!     Does anybody have any suggestions on how to deal with them?   We are not wealthy....or I would just pay this and walk away......

    There is a big misconception in what customers believe a cell carrier is obligated to do.
    You pay a set price to use up to that amount of xx data. It makes no difference if you use it to the paid limit or way under. Its like peace of mind when you don't have to worry about a data counter.
    Your daughter used the data, your plan is quite clear of what overage charges are. Why should or would Verizon wireless just forgive the charges because you are a customer? Your daughter used the data, get the money from her. That is the responsible thing to do.
    There is no "I have been a loyal customer so please remove the $120, or $250, or $2,000.00 since I did not mean to use it"
    Your electric company, or gas company or any other company does not remove valid charges. Why should Verizon wireless?
    Just pay the invoice and don't think you are being mistreated because Verizon is a business and not a charity.
    Good Luck

  • Problem with Revision level for a Material

    Hi All,
    I have a problem with Revision Levels for a Material.
    I want to know the Revision Level for a particular Material for BOM Creation.
    Can anyone please let me know from which Table or Funtion Module can I get the Rev Level, once I pass MATNR (Material) and AENNR (Change Number).
    <REMOVED BY MODERATOR>
    Thanks and Regards,
    Harish Kalla
    Edited by: Alvaro Tejada Galindo on Apr 14, 2008 4:30 PM

    HI,
    You can use table AEOI.
    Hope this helps.
    <REMOVED BY MODERATOR>
    Regards,
    Sipra
    Edited by: Alvaro Tejada Galindo on Apr 14, 2008 4:30 PM

  • Haviing problem with the facebbok for 9800

    i hava a problem with the facebook for blackberry 9800, i download it many times but i can not see the i con, i used many ways to find out it, but failed to find it,

    Hello nomanmahmoo, 
    Welcome to the forums. 
    Have a look at this article which may help you out with your issue. 
    Facebook icon is missing on the BlackBerry smartphone after upgrading to Facebook
    -SR
    Come follow your BlackBerry Technical Team on twitter! @BlackBerryHelp
    Be sure to click Kudos! for those who have helped you.Click Solution? for posts that have solved your issue(s)!

  • Litlle problem with nokia maps for 5800 xm

    hi evryone,
    i have a problem with nokia maps for my 5800 xpressmusi which is i cant find my acces point when itry to GO ONLINE
    please help thank you

    Hi,
    When you switch on your device for the first time, the access points may be configured automatically based on the service provider information in your SIM card. Otherwise contact your operator or service provider to receive the access point settings if you cannot find it..

  • App-V: A Configuration Template for Deploying to Stateless RDS Clients on Citrix Published Desktops with Citrix UPM for Profile Management

    Please Vote if you find this to be helpful!
    App-V:  A Configuration Template for Deploying to Stateless RDS Clients on Citrix Published Desktops with Citrix UPM for Profile Management
    Just posted this to the wiki:
    http://social.technet.microsoft.com/wiki/contents/articles/25318.app-v-a-configuration-template-for-deploying-to-stateless-rds-clients-on-citrix-published-desktops-with-citrix-upm-for-profile-management.aspx

    I would not recommend this and keep the package cache and the client on the same non-persistent drive and enable the Shared Content Store. If you separate the cache and the App-V client they could get out of sync and strange behaviour can occur. 
    You can use a temporary local profile with Citrix UPM or UE-V and specify what to roam/save.
    You can use the Shared Content Store so packages will stream over the network. When the user logs on there is a publishing phase where shortcuts etc are created for the user, this will take some time.
    Are you using the App-V full infrastructure?
    Are you using a boot disk, partition or PXE in combination with PVS?

  • Problem with restoring database from backupset

    Hello,
    I'm newie in working with RMAN and I have problem with restoring database from backup set in my testcase.
    I've restored controlfile, but I couldn't restore database - it fails with:
    RMAN-00571: =============================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS
    RMAN-00571: =============================================
    RMAN-03002: failure of restore command at 08/31/2006 12:06:47
    ORA-01180: can not create datafile 1
    ORA-01110: data file 1: 'C:\ORACLE\ORADATA\LOCA10G2\SYSTEM01.DBF'
    List of backupsets from restored controlfile
    (I restored controlfile by command: restore controlfile from 'C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP'; -- I have disabled controlfile autobackup, therefore I couldn't restore from autobackup);
    List of Backup Sets
    ===================
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    36 Full 6.98M DISK 00:00:02 29-AUG-06
    BP Key: 36 Status: AVAILABLE Compressed: NO Tag: 01_CTL
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCNNF_01_CTL_2H82NDQT_.BKP
    Control File Included: Ckp SCN: 578469 Ckp time: 29-AUG-06
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    37 Full 322.96M DISK 00:00:27 29-AUG-06
    BP Key: 37 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113622
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NNNDF_TAG20060829T113622_2H82NQ58_.BKP
    List of Datafiles in backup set 37
    File LV Type Ckp SCN Ckp Time Name
    1 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\SYSTEM01.DBF
    2 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\UNDOTBS01.DBF
    3 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\SYSAUX01.DBF
    4 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\USERS01.DBF
    BS Key Size Device Type Elapsed Time Completion Time
    38 650.50K DISK 00:00:00 29-AUG-06
    BP Key: 38 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113804
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_ANNNN_TAG20060829T113804_2H82QYOV_.BKP
    List of Archived Logs in backup set 38
    Thrd Seq Low SCN Low Time Next SCN Next Time
    1 32 577277 29-AUG-06 578529 29-AUG-06
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    39 Full 7.02M DISK 00:00:00 29-AUG-06
    BP Key: 39 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113622
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP
    Control File Included: Ckp SCN: 578493 Ckp time: 29-AUG-06
    SPFILE Included: Modification time: 28-AUG-06
    I can successfully crosscheck backup sets (by command crosscheck backup), but I couldn't restore database.
    Path C:\ORACLE\ORADATA\LOCA10G2\ exists and I have right privileges.
    Output of command restore database validate:
    RMAN> restore database validate;
         Starting restore at 31-AUG-06
         allocated channel: ORA_DISK_1
         channel ORA_DISK_1: sid=102 devtype=DISK
         data file 1 will be created automatically during restore operation
         data file 2 will be created automatically during restore operation
         data file 3 will be created automatically during restore operation
         data file 4 will be created automatically during restore operation
         restore not done; all files readonly, offline, or already restored
         Finished restore at 31-AUG-06
    What's wrong?
    Thanks
         Paul

    Hi,
    I think, that everything seems to be OK.
    Commands which I use to validation of existence backup:
    RMAN> crosscheck backup;
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=102 devtype=DISK
    crosschecked backup piece: found to be 'AVAILABLE'
    backup piece handle=C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCNNF_01_CTL_2H82NDQT_.BKP recid=36 stamp=599744172
    crosschecked backup piece: found to be 'AVAILABLE'
    backup piece handle=C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NNNDF_TAG20060829T113622_2H82NQ58_.BKP recid=37 stamp=599744183
    crosschecked backup piece: found to be 'AVAILABLE'
    backup piece handle=C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_ANNNN_TAG20060829T113804_2H82QYOV_.BKP recid=38 stamp=599918805
    crosschecked backup piece: found to be 'AVAILABLE'
    backup piece handle=C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP recid=39 stamp=599918805
    Crosschecked 4 objects
    RMAN> list backup summary;
    List of Backups
    ===============
    Key TY LV S Device Type Completion Time #Pieces #Copies Compressed Tag
    36 B F A DISK 29-AUG-06 1 1 NO 01_CTL
    37 B F A DISK 29-AUG-06 1 1 NO TAG20060829T113622
    38 B A A DISK 29-AUG-06 1 1 NO TAG20060829T113804
    39 B F A DISK 29-AUG-06 1 1 NO TAG20060829T113622
    RMAN> list backup;
    List of Backup Sets
    ===================
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    36 Full 6.98M DISK 00:00:02 29-AUG-06
    BP Key: 36 Status: AVAILABLE Compressed: NO Tag: 01_CTL
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCNNF_01_CTL_2H82NDQT_.BKP
    Control File Included: Ckp SCN: 578469 Ckp time: 29-AUG-06
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    37 Full 322.96M DISK 00:00:27 29-AUG-06
    BP Key: 37 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113622
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NNNDF_TAG20060829T113622_2H82NQ58_.BKP
    List of Datafiles in backup set 37
    File LV Type Ckp SCN Ckp Time Name
    1 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\SYSTEM01.DBF
    2 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\UNDOTBS01.DBF
    3 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\SYSAUX01.DBF
    4 Full 578481 29-AUG-06 C:\ORACLE\ORADATA\LOCA10G2\USERS01.DBF
    BS Key Size Device Type Elapsed Time Completion Time
    38 650.50K DISK 00:00:00 29-AUG-06
    BP Key: 38 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113804
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_ANNNN_TAG20060829T113804_2H82QYOV_.BKP
    List of Archived Logs in backup set 38
    Thrd Seq Low SCN Low Time Next SCN Next Time
    1 32 577277 29-AUG-06 578529 29-AUG-06
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    39 Full 7.02M DISK 00:00:00 29-AUG-06
    BP Key: 39 Status: AVAILABLE Compressed: NO Tag: TAG20060829T113622
    Piece Name: C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP
    Control File Included: Ckp SCN: 578493 Ckp time: 29-AUG-06
    SPFILE Included: Modification time: 28-AUG-06
    RMAN> restore database validate;
    Starting restore at 04-SEP-06
    using channel ORA_DISK_1
    data file 1 will be created automatically during restore operation
    data file 2 will be created automatically during restore operation
    data file 3 will be created automatically during restore operation
    data file 4 will be created automatically during restore operation
    restore not done; all files readonly, offline, or already restored
    Finished restore at 04-SEP-06
    Is something wrong? After crosscheck backup set with system datafile is AVAILABLE.
    If i tried to test existence of backup pieces on disk and permissisons - everything is OK too:
    C:\>dir C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\
    Volume in drive C has no label.
    Volume Serial Number is E003-9FC6
    Directory of C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29
    29.08.2006 11:38 <DIR> .
    29.08.2006 11:38 <DIR> ..
    29.08.2006 11:38 666 624 O1_MF_ANNNN_TAG20060829T113804_2H82QYOV_.BKP
    29.08.2006 11:36 7 340 032 O1_MF_NCNNF_01_CTL_2H82NDQT_.BKP
    29.08.2006 11:37 7 372 800 O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP
    29.08.2006 11:36 338 657 280 O1_MF_NNNDF_TAG20060829T113622_2H82NQ58_.BKP
    4 File(s) 354 036 736 bytes
    2 Dir(s) 56 865 202 176 bytes free
    C:\>copy C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\ C:\ORACLE\ORADATA\LOCA10G2\
    C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_ANNNN_TAG20060829T113804_2H82QYOV_.BKP
    C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCNNF_01_CTL_2H82NDQT_.BKP
    C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NCSNF_TAG20060829T113622_2H82OW77_.BKP
    C:\ORACLE\FLASH_RECOVERY_AREA\LOCA10G2\BACKUPSET\2006_08_29\O1_MF_NNNDF_TAG20060829T113622_2H82NQ58_.BKP
    4 file(s) copied.
    Thanks
    Pavel

Maybe you are looking for

  • Tricky Help

    I have simple sql query in my report, when i execute report result will vertically, i need to show horizontally in out put. To better understand i will eg below. select abc,ddd from dual; Report Output: abc xxx yyy zzz I need to show report out put l

  • 5508 WLC redundancy issue.

    I am trying to set up fail-over for Cisco 5508 Wireless LAN Controller which is using 7.4 version software. I setup the redundancy port and peer redundancy. While checking,one becomes active and other becomes standby. Its showing everything perfectly

  • Alot of unknown harddrive activity

    hi for the last 2 week i had very much harddrive activity when im using my computer. it sometimes takes more than 3 minutes to get up a explorer window. i tried unified driver 2.03 and 2.45 no difference. im using the same program as i have done for

  • Why can't I save my work as a document?

    The first thing I did when I downloaded Pages was create a default template with Times New Roman. Now when I try to save my work, it seems to automatically want to save it as a template. There is no option to save as a document and I can't choose to

  • Ibook Clamshell Airport issues...

    hello everyone! I have an iBook Clamshell, Graphite color, OS 10.4.3, 320MB RAM, 466MHz Processor. I also have a macbook with OS 10.6.4 2GB RAM, 2.4 GHz Processor. I am trying to connect to the internet on my clamshell via it's Airport. The Macbook w


HashFlare