Configuring Panther to run an Agfa Arcus II scanner
Greetings!
For reasons I won't bore you with, I have to configure an Agfa Arcus II flatbed scanner to run under 10.3. Agfa stopped actively supporting its scanners several years ago, but they still maintain support info on their web site and do provide information on how to get the scanner up and running.
However, doing so requires two original system files from a system running 10.2.8. I have no access to such a system (or to the software to set up the recommended partition) and was wondering if any of you might.
For the benefit of others who may be facing this or a similar problem, I've included part of Agfa's instructions below. (To read the entire procedure, go to www.agfa.com and follow the links on the support page.)
Thanks in advance.
How to make FotoLook 3.6 work on Mac OS 10.3
In order to make FotoLook 3.6 work on OS 10.3, you need to replace some sytem files by older system files from OS 10.2.8.
This action is only recommended if you really want to use your scanner on 10.3. A better and safer solution is to keep a partition with 10.2.x and use your scanner on that partition.
First make a backup of your system folder before proceeding. Modifications on system/library can destabilize your system.
If you are familiar with Terminal, make yourself root user and copy the following two files from an old 10.2.8 into the system folder of 10.3
IOSCSIArchitectureModelFamily.kext
IOSCSIParallelFamily.kext
Location: System/Library/Extensions
-------------------------------------------------
Hi KET,
Kext swapping for compatability with scsi is not a real problem as long as you understand that it may change the way that those kext files interact with other devices. If you are only using the Arcus with scsi I don't think that you will find any problems.
As for getting the files into the system it would have to be done while booted onto another partition or hard drive. You could also boot target mode on another mac and do the swap. I have tried swapping kext files on an active/booted system and it caused problems. I'm sure that there is a way to do it in Terminal but I don't know what it is.
As for the files themselves, I have them on an old computer in the garage and I thinkd it still boots. I would have to network it to my current mac to get the files. Alternatively, you might be able to use Pacifist to get the files off of a combo update from Apple.
Similar Messages
-
[SOLVED] configure: error: cannot run C compiled programs
I'm trying to build lib32-libxkbcommon 0.5.0-1 from AUR with makepkg. I already tried installing pacman (setting the default makepkg.conf) and multilib-devel with no luck.
makepkg messages:
==> Making package: lib32-libxkbcommon 0.5.0-1 (Mon May 11 00:17:05 EEST 2015)
==> Checking runtime dependencies...
==> Checking buildtime dependencies...
==> Retrieving sources...
-> Found libxkbcommon-0.5.0.tar.xz
==> Validating source files with sha256sums...
libxkbcommon-0.5.0.tar.xz ... Passed
==> Extracting sources...
-> Extracting libxkbcommon-0.5.0.tar.xz with bsdtar
bsdtar: Failed to set default locale
==> Starting prepare()...
==> Removing existing $pkgdir/ directory...
==> Starting build()...
checking for a BSD-compatible install... /usr/bin/install -c
checking whether build environment is sane... yes
checking for a thread-safe mkdir -p... /usr/bin/mkdir -p
checking for gawk... gawk
checking whether make sets $(MAKE)... yes
checking whether make supports nested variables... yes
checking whether to enable maintainer-specific portions of Makefiles... yes
checking for style of include used by make... GNU
checking for gcc... gcc -m32
checking whether the C compiler works... yes
checking for C compiler default output file name... a.out
checking for suffix of executables...
checking whether we are cross compiling... configure: error: in `/mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0':
configure: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details
==> ERROR: A failure occurred in build().
Aborting...
makepkg.conf:
# /etc/makepkg.conf
# SOURCE ACQUISITION
#-- The download utilities that makepkg should use to acquire sources
# Format: 'protocol::agent'
DLAGENTS=('ftp::/usr/bin/curl -fC - --ftp-pasv --retry 3 --retry-delay 3 -o %o %u'
'http::/usr/bin/curl -fLC - --retry 3 --retry-delay 3 -o %o %u'
'https::/usr/bin/curl -fLC - --retry 3 --retry-delay 3 -o %o %u'
'rsync::/usr/bin/rsync --no-motd -z %u %o'
'scp::/usr/bin/scp -C %u %o')
# Other common tools:
# /usr/bin/snarf
# /usr/bin/lftpget -c
# /usr/bin/wget
#-- The package required by makepkg to download VCS sources
# Format: 'protocol::package'
VCSCLIENTS=('bzr::bzr'
'git::git'
'hg::mercurial'
'svn::subversion')
# ARCHITECTURE, COMPILE FLAGS
CARCH="x86_64"
CHOST="x86_64-unknown-linux-gnu"
#-- Compiler and Linker Flags
# -march (or -mcpu) builds exclusively for an architecture
# -mtune optimizes for an architecture, but builds for whole processor family
CPPFLAGS="-D_FORTIFY_SOURCE=2"
CFLAGS="-march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4"
CXXFLAGS="${CFLAGS}"
LDFLAGS="-Wl,-O1,--sort-common,--as-needed,-z,relro"
#-- Make Flags: change this for DistCC/SMP systems
MAKEFLAGS="-j5"
#-- Debugging flags
DEBUG_CFLAGS="-g -fvar-tracking-assignments"
DEBUG_CXXFLAGS="-g -fvar-tracking-assignments"
# BUILD ENVIRONMENT
# Defaults: BUILDENV=(!distcc color !ccache check !sign)
# A negated environment option will do the opposite of the comments below.
#-- distcc: Use the Distributed C/C++/ObjC compiler
#-- color: Colorize output messages
#-- ccache: Use ccache to cache compilation
#-- check: Run the check() function if present in the PKGBUILD
#-- sign: Generate PGP signature file
BUILDENV=(!distcc color !ccache check !sign)
#-- If using DistCC, your MAKEFLAGS will also need modification. In addition,
#-- specify a space-delimited list of hosts running in the DistCC cluster.
#DISTCC_HOSTS=""
#-- Specify a directory for package building.
#BUILDDIR=/tmp/makepkg
# GLOBAL PACKAGE OPTIONS
# These are default values for the options=() settings
# Default: OPTIONS=(strip docs !libtool !staticlibs emptydirs zipman purge !upx !debug)
# A negated option will do the opposite of the comments below.
#-- strip: Strip symbols from binaries/libraries
#-- docs: Save doc directories specified by DOC_DIRS
#-- libtool: Leave libtool (.la) files in packages
#-- staticlibs: Leave static library (.a) files in packages
#-- emptydirs: Leave empty directories in packages
#-- zipman: Compress manual (man and info) pages in MAN_DIRS with gzip
#-- purge: Remove files specified by PURGE_TARGETS
#-- upx: Compress binary executable files using UPX
#-- debug: Add debugging flags as specified in DEBUG_* variables
OPTIONS=(strip docs !libtool !staticlibs emptydirs zipman purge !upx !debug)
#-- File integrity checks to use. Valid: md5, sha1, sha256, sha384, sha512
INTEGRITY_CHECK=(md5)
#-- Options to be used when stripping binaries. See `man strip' for details.
STRIP_BINARIES="--strip-all"
#-- Options to be used when stripping shared libraries. See `man strip' for details.
STRIP_SHARED="--strip-unneeded"
#-- Options to be used when stripping static libraries. See `man strip' for details.
STRIP_STATIC="--strip-debug"
#-- Manual (man and info) directories to compress (if zipman is specified)
MAN_DIRS=({usr{,/local}{,/share},opt/*}/{man,info})
#-- Doc directories to remove (if !docs is specified)
DOC_DIRS=(usr/{,local/}{,share/}{doc,gtk-doc} opt/*/{doc,gtk-doc})
#-- Files to be removed from all packages (if purge is specified)
PURGE_TARGETS=(usr/{,share}/info/dir .packlist *.pod)
# PACKAGE OUTPUT
# Default: put built package and cached source in build directory
#-- Destination: specify a fixed directory where all packages will be placed
#PKGDEST=/home/packages
#-- Source cache: specify a fixed directory where source files will be cached
#SRCDEST=/home/sources
#-- Source packages: specify a fixed directory where all src packages will be placed
#SRCPKGDEST=/home/srcpackages
#-- Log files: specify a fixed directory where all log files will be placed
#LOGDEST=/home/makepkglogs
#-- Packager: name/email of the person or organization building packages
#PACKAGER="John Doe <[email protected]>"
#-- Specify a key to use for package signing
#GPGKEY=""
# COMPRESSION DEFAULTS
COMPRESSGZ=(gzip -c -f -n)
COMPRESSBZ2=(bzip2 -c -f)
COMPRESSXZ=(xz -c -z -)
COMPRESSLRZ=(lrzip -q)
COMPRESSLZO=(lzop -q)
COMPRESSZ=(compress -c -f)
# EXTENSION DEFAULTS
# WARNING: Do NOT modify these variables unless you know what you are
# doing.
PKGEXT='.pkg.tar.xz'
SRCEXT='.src.tar.gz'
# vim: set ft=sh ts=2 sw=2 et:
config.log:
This file contains any messages produced by compilers while
running configure, to aid debugging if configure makes a mistake.
It was created by libxkbcommon configure 0.5.0, which was
generated by GNU Autoconf 2.69. Invocation command line was
$ ./configure --prefix=/usr --libdir=/usr/lib32 --disable-docs --disable-static
## Platform. ##
hostname = Arch
uname -m = x86_64
uname -r = 4.0.1-1-ARCH
uname -s = Linux
uname -v = #1 SMP PREEMPT Wed Apr 29 12:00:26 CEST 2015
/usr/bin/uname -p = unknown
/bin/uname -X = unknown
/bin/arch = unknown
/usr/bin/arch -k = unknown
/usr/convex/getsysinfo = unknown
/usr/bin/hostinfo = unknown
/bin/machine = unknown
/usr/bin/oslevel = unknown
/bin/universe = unknown
PATH: /usr/local/sbin
PATH: /usr/local/bin
PATH: /usr/bin
PATH: /usr/lib/jvm/default/bin
PATH: /usr/bin/site_perl
PATH: /usr/bin/vendor_perl
PATH: /usr/bin/core_perl
## Core tests. ##
configure:2424: checking for a BSD-compatible install
configure:2492: result: /usr/bin/install -c
configure:2503: checking whether build environment is sane
configure:2558: result: yes
configure:2709: checking for a thread-safe mkdir -p
configure:2748: result: /usr/bin/mkdir -p
configure:2755: checking for gawk
configure:2771: found /usr/bin/gawk
configure:2782: result: gawk
configure:2793: checking whether make sets $(MAKE)
configure:2815: result: yes
configure:2844: checking whether make supports nested variables
configure:2861: result: yes
configure:2987: checking whether to enable maintainer-specific portions of Makefiles
configure:2996: result: yes
configure:3023: checking for style of include used by make
configure:3051: result: GNU
configure:3122: checking for gcc
configure:3149: result: gcc -m32
configure:3378: checking for C compiler version
configure:3387: gcc -m32 --version >&5
gcc (GCC) 5.1.0
Copyright (C) 2015 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
configure:3398: $? = 0
configure:3387: gcc -m32 -v >&5
Using built-in specs.
COLLECT_GCC=gcc
COLLECT_LTO_WRAPPER=/usr/lib/gcc/x86_64-unknown-linux-gnu/5.1.0/lto-wrapper
Target: x86_64-unknown-linux-gnu
Configured with: /build/gcc-multilib/src/gcc-5-20150505/configure --prefix=/usr --libdir=/usr/lib --libexecdir=/usr/lib --mandir=/usr/share/man --infodir=/usr/share/info --with-bugurl=https://bugs.archlinux.org/ --enable-languages=c,c++,ada,fortran,go,lto,objc,obj-c++ --enable-shared --enable-threads=posix --enable-libmpx --with-system-zlib --with-isl --enable-__cxa_atexit --disable-libunwind-exceptions --enable-clocale=gnu --disable-libstdcxx-pch --disable-libssp --enable-gnu-unique-object --enable-linker-build-id --enable-lto --enable-plugin --enable-install-libiberty --with-linker-hash-style=gnu --enable-gnu-indirect-function --enable-multilib --disable-werror --enable-checking=release --with-default-libstdcxx-abi=c++98
Thread model: posix
gcc version 5.1.0 (GCC)
configure:3398: $? = 0
configure:3387: gcc -m32 -V >&5
gcc: error: unrecognized command line option '-V'
gcc: fatal error: no input files
compilation terminated.
configure:3398: $? = 1
configure:3387: gcc -m32 -qversion >&5
gcc: error: unrecognized command line option '-qversion'
gcc: fatal error: no input files
compilation terminated.
configure:3398: $? = 1
configure:3418: checking whether the C compiler works
configure:3440: gcc -m32 -march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4 -D_FORTIFY_SOURCE=2 -Wl,-O1,--sort-common,--as-needed,-z,relro conftest.c >&5
configure:3444: $? = 0
configure:3492: result: yes
configure:3495: checking for C compiler default output file name
configure:3497: result: a.out
configure:3503: checking for suffix of executables
configure:3510: gcc -m32 -o conftest -march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4 -D_FORTIFY_SOURCE=2 -Wl,-O1,--sort-common,--as-needed,-z,relro conftest.c >&5
configure:3514: $? = 0
configure:3536: result:
configure:3558: checking whether we are cross compiling
configure:3566: gcc -m32 -o conftest -march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4 -D_FORTIFY_SOURCE=2 -Wl,-O1,--sort-common,--as-needed,-z,relro conftest.c >&5
In file included from /usr/include/stdio.h:27:0,
from conftest.c:11:
/usr/include/features.h:365:25: fatal error: sys/cdefs.h: No such file or directory
compilation terminated.
configure:3570: $? = 1
configure:3577: ./conftest
./configure: line 3579: ./conftest: No such file or directory
configure:3581: $? = 127
configure:3588: error: in `/mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0':
configure:3590: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details
## Cache variables. ##
ac_cv_env_CC_set=set
ac_cv_env_CC_value='gcc -m32'
ac_cv_env_CFLAGS_set=set
ac_cv_env_CFLAGS_value='-march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4'
ac_cv_env_CPPFLAGS_set=set
ac_cv_env_CPPFLAGS_value=-D_FORTIFY_SOURCE=2
ac_cv_env_CPP_set=
ac_cv_env_CPP_value=
ac_cv_env_DOT_set=
ac_cv_env_DOT_value=
ac_cv_env_DOXYGEN_set=
ac_cv_env_DOXYGEN_value=
ac_cv_env_LDFLAGS_set=set
ac_cv_env_LDFLAGS_value=-Wl,-O1,--sort-common,--as-needed,-z,relro
ac_cv_env_LIBS_set=
ac_cv_env_LIBS_value=
ac_cv_env_PKG_CONFIG_LIBDIR_set=
ac_cv_env_PKG_CONFIG_LIBDIR_value=
ac_cv_env_PKG_CONFIG_PATH_set=set
ac_cv_env_PKG_CONFIG_PATH_value=/usr/lib32/pkgconfig
ac_cv_env_PKG_CONFIG_set=
ac_cv_env_PKG_CONFIG_value=
ac_cv_env_XCB_XKB_CFLAGS_set=
ac_cv_env_XCB_XKB_CFLAGS_value=
ac_cv_env_XCB_XKB_LIBS_set=
ac_cv_env_XCB_XKB_LIBS_value=
ac_cv_env_XORG_MALLOC_DEBUG_ENV_set=
ac_cv_env_XORG_MALLOC_DEBUG_ENV_value=
ac_cv_env_YACC_set=
ac_cv_env_YACC_value=
ac_cv_env_YFLAGS_set=
ac_cv_env_YFLAGS_value=
ac_cv_env_build_alias_set=
ac_cv_env_build_alias_value=
ac_cv_env_host_alias_set=
ac_cv_env_host_alias_value=
ac_cv_env_target_alias_set=
ac_cv_env_target_alias_value=
ac_cv_path_install='/usr/bin/install -c'
ac_cv_path_mkdir=/usr/bin/mkdir
ac_cv_prog_AWK=gawk
ac_cv_prog_ac_ct_CC='gcc -m32'
ac_cv_prog_make_make_set=yes
am_cv_make_support_nested_variables=yes
## Output variables. ##
ACLOCAL='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/missing aclocal-1.14'
ADMIN_MAN_DIR=''
ADMIN_MAN_SUFFIX=''
AMDEPBACKSLASH='\'
AMDEP_FALSE='#'
AMDEP_TRUE=''
AMTAR='$${TAR-tar}'
AM_BACKSLASH='\'
AM_DEFAULT_V='$(AM_DEFAULT_VERBOSITY)'
AM_DEFAULT_VERBOSITY='1'
AM_V='$(V)'
APP_MAN_DIR=''
APP_MAN_SUFFIX=''
AR=''
AUTOCONF='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/missing autoconf'
AUTOHEADER='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/missing autoheader'
AUTOMAKE='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/missing automake-1.14'
AWK='gawk'
BASE_CFLAGS=''
BUILD_LINUX_TESTS_FALSE=''
BUILD_LINUX_TESTS_TRUE=''
CC='gcc -m32'
CCDEPMODE=''
CFLAGS='-march=core2 -m64 -mfpmath=sse -O2 -fomit-frame-pointer -pipe -fstack-protector --param=ssp-buffer-size=4'
CHANGELOG_CMD=''
CPP=''
CPPFLAGS='-D_FORTIFY_SOURCE=2'
CWARNFLAGS=''
CYGPATH_W='echo'
DEFS=''
DEPDIR='.deps'
DLLTOOL=''
DOT=''
DOXYGEN=''
DRIVER_MAN_DIR=''
DRIVER_MAN_SUFFIX=''
DSYMUTIL=''
DUMPBIN=''
ECHO_C=''
ECHO_N='-n'
ECHO_T=''
EGREP=''
ENABLE_DOCS_FALSE=''
ENABLE_DOCS_TRUE=''
ENABLE_X11_FALSE=''
ENABLE_X11_TRUE=''
EXEEXT=''
FGREP=''
FILE_MAN_DIR=''
FILE_MAN_SUFFIX=''
GREP=''
HAVE_DOT=''
HAVE_DOT_FALSE=''
HAVE_DOT_TRUE=''
HAVE_DOXYGEN_FALSE=''
HAVE_DOXYGEN_TRUE=''
HAVE_NO_UNDEFINED_FALSE=''
HAVE_NO_UNDEFINED_TRUE=''
INSTALL_CMD=''
INSTALL_DATA='${INSTALL} -m 644'
INSTALL_PROGRAM='${INSTALL}'
INSTALL_SCRIPT='${INSTALL}'
INSTALL_STRIP_PROGRAM='$(install_sh) -c -s'
LD=''
LDFLAGS='-Wl,-O1,--sort-common,--as-needed,-z,relro'
LIBOBJS=''
LIBS=''
LIBTOOL=''
LIB_MAN_DIR=''
LIB_MAN_SUFFIX=''
LIPO=''
LN_S=''
LTLIBOBJS=''
MAINT=''
MAINTAINER_MODE_FALSE='#'
MAINTAINER_MODE_TRUE=''
MAKEINFO='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/missing makeinfo'
MANIFEST_TOOL=''
MAN_SUBSTS=''
MISC_MAN_DIR=''
MISC_MAN_SUFFIX=''
MKDIR_P='/usr/bin/mkdir -p'
NM=''
NMEDIT=''
OBJDUMP=''
OBJEXT=''
OTOOL64=''
OTOOL=''
PACKAGE='libxkbcommon'
PACKAGE_BUGREPORT='https://bugs.freedesktop.org/enter_bug.cgi?product=libxkbcommon'
PACKAGE_NAME='libxkbcommon'
PACKAGE_STRING='libxkbcommon 0.5.0'
PACKAGE_TARNAME='libxkbcommon'
PACKAGE_URL='http://xkbcommon.org'
PACKAGE_VERSION='0.5.0'
PATH_SEPARATOR=':'
PKG_CONFIG=''
PKG_CONFIG_LIBDIR=''
PKG_CONFIG_PATH='/usr/lib32/pkgconfig'
RANLIB=''
RT_LIBS=''
SED=''
SET_MAKE=''
SHELL='/bin/sh'
STRICT_CFLAGS=''
STRIP=''
VERSION='0.5.0'
XCB_XKB_CFLAGS=''
XCB_XKB_LIBS=''
XKBCONFIGROOT=''
XLOCALEDIR=''
XORG_MALLOC_DEBUG_ENV=''
XORG_MAN_PAGE=''
YACC=''
YACC_INST=''
YFLAGS=''
ac_ct_AR=''
ac_ct_CC='gcc -m32'
ac_ct_DUMPBIN=''
am__EXEEXT_FALSE=''
am__EXEEXT_TRUE=''
am__fastdepCC_FALSE=''
am__fastdepCC_TRUE=''
am__include='include'
am__isrc=''
am__leading_dot='.'
am__nodep='_no'
am__quote=''
am__tar='$${TAR-tar} chof - "$$tardir"'
am__untar='$${TAR-tar} xf -'
bindir='${exec_prefix}/bin'
build=''
build_alias=''
build_cpu=''
build_os=''
build_vendor=''
datadir='${datarootdir}'
datarootdir='${prefix}/share'
docdir='${datarootdir}/doc/${PACKAGE_TARNAME}'
dvidir='${docdir}'
exec_prefix='NONE'
host=''
host_alias=''
host_cpu=''
host_os=''
host_vendor=''
htmldir='${docdir}'
includedir='${prefix}/include'
infodir='${datarootdir}/info'
install_sh='${SHELL} /mnt/tmp/yaourt-tmp-tsester/aur-lib32-libxkbcommon/src/libxkbcommon-0.5.0/build-aux/install-sh'
libdir='/usr/lib32'
libexecdir='${exec_prefix}/libexec'
localedir='${datarootdir}/locale'
localstatedir='${prefix}/var'
mandir='${datarootdir}/man'
mkdir_p='$(MKDIR_P)'
oldincludedir='/usr/include'
pdfdir='${docdir}'
prefix='/usr'
program_transform_name='s,x,x,'
psdir='${docdir}'
sbindir='${exec_prefix}/sbin'
sharedstatedir='${prefix}/com'
sysconfdir='${prefix}/etc'
target_alias=''
## confdefs.h. ##
/* confdefs.h */
#define PACKAGE_NAME "libxkbcommon"
#define PACKAGE_TARNAME "libxkbcommon"
#define PACKAGE_VERSION "0.5.0"
#define PACKAGE_STRING "libxkbcommon 0.5.0"
#define PACKAGE_BUGREPORT "[url]https://bugs.freedesktop.org/enter_bug.cgi?product=libxkbcommon[/url]"
#define PACKAGE_URL "[url]http://xkbcommon.org[/url]"
#define PACKAGE "libxkbcommon"
#define VERSION "0.5.0"
configure: exit 1
other info:
core/pacman 4.2.1-1
multilib/gcc-multilib 4.9.2-4 (multilib-devel) [installed]
The GNU Compiler Collection - C and C++ frontends for multilib
multilib/lib32-fakeroot 1.20.2-1 (multilib-devel) [installed]
Tool for simulating superuser privileges (32-bit)
multilib/lib32-libltdl 2.4.5-1 (multilib-devel) [installed]
A generic library support script (32-bit)
Last edited by tsester (2015-05-10 22:10:28)tsester wrote:P.S.: I recently transfered the linux system between failing disks
In that case you should probably check that no other packages are missing files with
pacman -Qkk 2>&1 | grep "No such file or directory"
Any packages that report that they're missing files, you should reinstall. -
Oracle 9i database configuration assisstant not running
Hi All,
My database is Oracle 9i.The configuration assisstant not running in this .can any body help?
ReenaHi Reena, this got lost in the wrong forum. try the database forum
Barry -
Configure: error: cannot run C compiled programs.
Hi,
I am new to Solaris. I have installed Solaris 10 and compiled packages without any error.
I have tried to compile Wireshark 1.7.1 and it was exist indicating '/usr/include/sys/feature_tests.h no such file or directory'.
So I have download feature_tests.h file and uploaded it to the relevant location and run ./configure again.
Now I am not able to compile any package and it indicate error
'configure: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details'
I have attached portion of the config.log
configure:3739: checking whether the C compiler works
configure:3761: gcc conftest.c >&5
configure:3765: $? = 0
configure:3813: result: yes
configure:3816: checking for C compiler default output file name
configure:3818: result: a.out
configure:3824: checking for suffix of executables
configure:3831: gcc -o conftest conftest.c >&5
configure:3835: $? = 0
configure:3857: result:
configure:3879: checking whether we are cross compiling
configure:3887: gcc -o conftest conftest.c >&5
In file included from conftest.c:11:
/usr/include/stdio.h:21:31: sys/feature_tests.h: No such file or directory
In file included from /usr/include/stdio.h:66,
from conftest.c:11:
/usr/include/iso/stdio_iso.h:90: error: syntax error before "fpos_t"
/usr/include/iso/stdio_iso.h:208: error: syntax error before "fpos_t"
/usr/include/iso/stdio_iso.h:210: error: syntax error before '*' token
In file included from /usr/include/stdio.h:135,
from conftest.c:11:
/usr/include/iso/stdio_c99.h:54: error: conflicting types for '_RESTRICT_KYWD'
/usr/include/iso/stdio_c99.h:54: error: previous definition of '_RESTRICT_KYWD' was here
/usr/include/iso/stdio_c99.h:56: error: redefinition of parameter '_RESTRICT_KYWD'
/usr/include/iso/stdio_c99.h:56: error: previous definition of '_RESTRICT_KYWD' was here
/usr/include/iso/stdio_c99.h:68: error: conflicting types for '_RESTRICT_KYWD'
/usr/include/iso/stdio_c99.h:68: error: previous definition of '_RESTRICT_KYWD' was here
/usr/include/iso/stdio_c99.h:70: error: conflicting types for '_RESTRICT_KYWD'
/usr/include/iso/stdio_c99.h:70: error: previous definition of '_RESTRICT_KYWD' was here
configure:3891: $? = 1
configure:3898: ./conftest
./configure: line 3900: ./conftest: No such file or directory
configure:3902: $? = 127
configure:3909: error: in `/wireshark-1.6.7':
configure:3911: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details934802 wrote:
Hi,
I am new to Solaris. I have installed Solaris 10 and compiled packages without any error.
I have tried to compile Wireshark 1.7.1 and it was exist indicating '/usr/include/sys/feature_tests.h no such file or directory'.
So I have download feature_tests.h file and uploaded it to the relevant location and run ./configure again.
Now I am not able to compile any package and it indicate error Where did you get this file? -
Windows setup could not configure Windows to run on this computer's hardware.
Just over a month ago, I purchased 6 identical Lenovo B5400 laptop computers for a small charity. Each is preconfigured with Windows 7 Professional, with a Win 7 recovery partition, and is supplied with a set of Windows 8 Pro recovery media (the charity
requires Windows 7 installed). I unpacked three laptops, configuring the first one as required by the charity, and using the second and third to experiment with the sysprep / imagex process before applying it to the configured system for migration to the other
laptops. After securing the second and third laptops by taking image backups, I applied sufficient configuration to the second to test the process, then sysprep'd it, and imagex'd it to a network drive, and to a USB stick. So far so good. After using imagex
to apply the custom .wim file to the third laptop, I found it booted up through the "updating registry settings", "starting services", and "installing devices" phases, but when it got to "applying system settings", it
displayed a messagebox titled "Install Windows", with the message "Windows Setup could not configure Windows to run on this computer's hardware". In the last few days, I have explored all the other incidences of problems with similar
symptoms found by a google search, including those on this forum, and tried numerous experiments without getting any further (in most cases not even getting as far)". The sysprep'd image on the second laptop boots ok, going through the same phases up
to "applying system settings", then reboots normally, requests username and PC name, etc, the only defect being that it loses the Realtek display driver (which only takes 2 minutes to restore). I have tried imagex /apply to put the custom .wim file
back on the same laptop it was taken off (the second laptop), and it works exactly the same as it did for the same system immediately after sysprep'ing. The only difference between laptops 2 and 3 is a serial number (QB07523967, and QB07524012). Is there any
reason that this should not work? If it will not, I'm faced with about 100 hours to configure the 6 laptops individually to the charity's requirement.Hi,
This issue can be caused by hard disk settings in bios or DVD drive or bad SATA cable. Please try those steps below:
Upgrade the BIOS to the latest version; go into the BIOS and activate "Load Fail-Safe Defaults".
If you connect the SATA drives, please set the SATA controller to compatible mode in BIOS.
For more information about this issue, please refer to the article:
http://support.microsoft.com/kb/2466753
Andy Altmann
TechNet Community Support -
Respected Sir,
My name is Md. Naushad, contact No- 9436507010.
Sir, hamara ek shopping mall hain, jismein ek server and 5 nos. cash counter clint pc hein, server ju computer hein ismain hamne WINDOWS SMALL BUSINESS SERVER 2008 original os use kar raha ho, yein server calte hue humnein pehele bhi 3-4 bar format korke
dubara ready kiya tha, job koi problem huta tha tu, kuch din pehele se server mein firse problem ho rahi thi (server open nehi hu rahi thi, tu hamne data backup leke pehele jis tarha se server install kiya tha usi tarha se humne install karne ki kusis ki tha,
tu jab humne windows small business server 2008 software ka original disc-1 se installation start karti hu tu kuch der install huke ek error message aaa rahi hein, ju ki "WINDOWS SETUP COULD NOT CONFIGURE WINDOWS TO RUN ON THIS COMPUTER'S HARDWARE."
yein message aa rahi hein, jiske baad server installation bech mein hi band hi jati hein, jiske karon hum server ready kar nahi pa rahi ho, so plz help me, ta ki mein server ku firse install kar sakti ho.
we wait for ur valuable response, plz help us sir.
Thank u.Hi,
Sorry for my delay.
Based on the error message, please refer to following similar thread and KB, then check if can help you.
SBS
2008 SP2 " Windows Setup could not configure Windows on this computer’s hardware"
“Windows Setup could not configure Windows on this computer’s
hardware” installation error on a Windows 7-based or a Windows Server 2008 R2-based computer
However,
please understand that these forums will provide support in English. Since we do not familiar with this language which you use to describe this issue, it is very difficult
for us to help you with your question. If you could ask your question in English, I will be very happy to help.
If the issue is urgent to your business, it is recommended that you contact Microsoft Product Support Services
via telephone so that a dedicated Support Professional can assist you recover the server in a more efficient manner. Please be advised that contacting phone support will be a charged call.
To obtain the phone numbers for specific technology request please take a look at the web site listed below.
http://support.microsoft.com/contactus/?ws=support&ln=en-us
If you are outside the US please see
http://support.microsoft.com/ for
regional support phone numbers.
Thanks for your understanding.
Hope this helps.
Best regards,
Justin Gu -
"Windows Setup could not configure Windows to run on this computer's hardware" error using MDT
Good day. I am using Microsoft Deployment Toolkit to deploy images to a group of ClearCube R3080D bladed PCs. Below are the server details:
VMware virtual server with Server 2008 R2 Enterprise Service Pack 1
Microsoft Deployment Workbench Version 6.1.2373.0
Management Console 3.0 Version 6.1(Build 7601: Service Pack 1)
Windows Deployment Services Version 6.1.7600.16385
The image I am deploying to the blades is that of Windows 7 Enterprise x64 Service Pack 1 with Office 2010 and a few other third-party applications.
I have pushed this image to literally hundreds of blades successfully. I had a problem with one blade and found that the HDD was faulty. I RMA’d that HDD and installed the replacement. I pushed the same image to the blade and it works fine to the very end
then I receive the error ”Windows Setup could not configure Windows to run on this computer’s hardware.”
I have checked the BIOS configurations on this blade and they are identical to those of several known good blades. In fact, I swapped HDDs between a known good blade and the blade in question. The problem seemed to follow the HDD. Therefore, I assumed that
ClearCube sent a faulty HDD. I had the identical problems with a second HDD from ClearCube so I’m sure that the HDD is fine.
I have read through http://support.microsoft.com/kb/2466753
as well as many other questions posed by other folks and I have yet to find a good solution.
Again, this image works on all other blades that are of the same model and the HDDs are also of the same model. I'm quite confused!
TonyTony?
What are you doing posting the same question 5 times, within 3 hours???
http://social.technet.microsoft.com/Forums/en-US/67f72721-da1b-42bc-b373-9064fe1a8245/windows-setup-could-not-configure-windows-to-run-on-this-computers-hardware-error-using-mdt
http://social.technet.microsoft.com/Forums/en-US/9e7ca626-5b67-4a3f-ade0-4bac767e060f/windows-setup-could-not-configure-windows-to-run-on-this-computers-hardware-error-using-mdt
http://social.technet.microsoft.com/Forums/en-US/8c508acc-ad40-4608-911c-fbfbc5e293fe/windows-setup-could-not-configure-windows-to-run-on-this-computers-hardware-error-using-mdthttp://social.technet.microsoft.com/Forums/en-US/a8a98fce-0ddf-4a59-ba8a-9acab7de0c5e/windows-setup-could-not-configure-windows-to-run-on-this-computers-hardware-error-using-mdt
If this post is helpful please click "Mark for answer", thanks! Kind regards -
IPhone configuration utility wont run anymore
hiya all!
my iPhone configuration utility wont run anymore. in the log i found this:
[2015-02-03 16:59:56.4529 iPCU] Error: Could not get InstallDir key from 'HKEY_LOCAL_MACHINE\SOFTWARE\Apple Inc.\Apple Mobile Device Support'
[2015-02-03 16:59:57.8758 iPCU] Error: Exception from OnStartupInitial: Das angegebene Modul wurde nicht gefunden. (Ausnahme von HRESULT: 0x8007007E)
[2015-02-03 17:01:01.8904 iPCU] Error: Could not get InstallDir key from 'HKEY_LOCAL_MACHINE\SOFTWARE\Apple Inc.\Apple Mobile Device Support'
[2015-02-03 17:01:03.1672 iPCU] Error: Exception from OnStartupInitial: Das angegebene Modul wurde nicht gefunden. (Ausnahme von HRESULT: 0x8007007E)
i disinstalled and reinstalled every software of apple and deleted all folders in the user folder.
anyone can help me?I get the same error(s) in my log:
[2015-02-10 12:43:46.1048 iPCU] Error: Could not get InstallDir key from 'HKEY_LOCAL_MACHINE\SOFTWARE\Apple Inc.\Apple Mobile Device Support'
[2015-02-10 12:43:49.9609 iPCU] Error: Exception from OnStartupInitial: The specified module could not be found. (Exception from HRESULT: 0x8007007E)
While attempting to start IPCU, I see the very common dialog error "iPhone Configuration Utility failed to locate 'Apple Mobile Device Support," but trying some of the .dll copying/replacing techniques seen in other threads doesn't fix it.
I believe there is something about the Log statements that will lead to the root cause, as it seems completely different than the other threads with similar errors. I'll report back if I figure it out. Good luck. -
Installing Agfa Duoscan T1200 scanner with latest OS X (10.4.8)
Does anyone know if it is possible to install my Agfa Duoscan T1200 scanner in my G4 without causing much trouble?
My G4 has a Ultra SCSI card installed. When I check the contents of my PCI Cards it says that I have ADPT,2930CU installed. I am assuming that is the Adaptec 2930 SCSI card.
When I went to download the latest driver for the Duoscan at
http://www.agfa.com/digicamscannerdrivers/index.html
there is a section that says: "Not every scanner supported!" and it says that Mac OS X 10.3 does not support the Duoscan T1200. On that site I was able to download the driver, but I don't want to install it if it will cause problems.
If I don't use FotoLook, but instead I use TWAIN from Photoshop 7.0.1, will it work?
Was anyone successful at all with this, or should I buy another scanner? I hate to do that, as this scanner works very well.Yes, you may have to get SANE and install it. That worked fine for me.
Vuescan is another way to go and there is a third program than will work.
Google for scanning software for OS X. -
Bar code using Form6i run in c\s using scanner
Dear all
Want to ask if possible to used bar code in form6i run in c\s using scanner as bar code reader..
Do i need to install something to work this bar code?
Any step by step documentation on how to do a bar code?
thanksA scanner is a hardware device that basically works like a keyboard. When you use a scanner it reads and interprets the bar code and sends the character to the Operating System (OS). Forms does not read characters directly from keyboard or scanner. It reads the characters from the OS. Therefore, Forms doesn't know where the characters came from, and it does not care.
Think of it this way...
When you change out your simple keyboard and replace it with a fancy keyboard, does Forms know or care. No.
Or here is a even simplier way to think about it...
Think of a scanner as being a box that sends characters to the computer. When you open the box and look inside you see a trained monkey who looks at the bar code and types the correct character on the keyboard. (But this trained monkey does not need to be fed.)
The question you should be asking is, not if Forms will work with the scanner, but whether the scanner will work with my OS. Only the scanner manufacturer can answer that. If you want to know how to install scanner or how to operate it, then ask the manufacturer of the scanner. -
Problem with Configuring Tomcat for running jsp web applications..Plz HELP
I am using Tomcat 5.5 and Jdk 1.5.0_12 and Oracle 10g. I am using jdbc-odbc bridge connection
to connect to the database. I have placed my project folder called
tdm under the webapps folder in Tomcat. This 'tdm' folder consists of
a collection of html pages,jsp pages and images of my project. Also I created a
WEB-INF folderand in that I have lib folder which contains catalina-root.jar
, classes12.jar and nls_charset.jar files. And also in the WEB-INF folder I have the web.xml
file which looks like this
<?xml version="1.0" encoding="ISO-8859-1"?>
<!--
Copyright 2004 The Apache Software Foundation
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-->
<web-app>
<resource-ref>
<description>Oracle Datasource example</description>
<res-ref-name>jdbc/gdn</res-ref-name>
<res-type>javax.sql.DataSource</res-type>
<res-auth>Container</res-auth>
</resource-ref>
</web-app>
My Server.xml file in Tomcat\conf folder is as follows
<!-- Example Server Configuration File -->
<!-- Note that component elements are nested corresponding to their
parent-child relationships with each other -->
<!-- A "Server" is a singleton element that represents the entire JVM,
which may contain one or more "Service" instances. The Server
listens for a shutdown command on the indicated port.
Note: A "Server" is not itself a "Container", so you may not
define subcomponents such as "Valves" or "Loggers" at this level.
-->
<Server port="8005" shutdown="SHUTDOWN">
<!-- Comment these entries out to disable JMX MBeans support used for the
administration web application -->
<Listener className="org.apache.catalina.mbeans.ServerLifecycleListener" />
<Listener className="org.apache.catalina.mbeans.GlobalResourcesLifecycleListener" />
<!-- Global JNDI resources -->
<GlobalNamingResources>
<!-- Test entry for demonstration purposes -->
<Environment name="simpleValue" type="java.lang.Integer" value="30"/>
<!-- Editable user database that can also be used by
UserDatabaseRealm to authenticate users -->
<Resource name="UserDatabase" auth="Container"
type="org.apache.catalina.UserDatabase"
description="User database that can be updated and saved"
factory="org.apache.catalina.users.MemoryUserDatabaseFactory"
pathname="conf/tomcat-users.xml" />
<Resource name="jdbc/gdn" auth="Container"
type="javax.sql.DataSource" driverClassName="sun.jdbc.odbc.JdbcOdbcDriver"
url="jdbc:odbc:gdn"
username="system" password="tiger" maxActive="20" maxIdle="10"
maxWait="-1"/>
</GlobalNamingResources>
<!-- A "Service" is a collection of one or more "Connectors" that share
a single "Container" (and therefore the web applications visible
within that Container). Normally, that Container is an "Engine",
but this is not required.
Note: A "Service" is not itself a "Container", so you may not
define subcomponents such as "Valves" or "Loggers" at this level.
-->
<!-- Define the Tomcat Stand-Alone Service -->
<Service name="Catalina">
<!-- A "Connector" represents an endpoint by which requests are received
and responses are returned. Each Connector passes requests on to the
associated "Container" (normally an Engine) for processing.
By default, a non-SSL HTTP/1.1 Connector is established on port 8080.
You can also enable an SSL HTTP/1.1 Connector on port 8443 by
following the instructions below and uncommenting the second Connector
entry. SSL support requires the following steps (see the SSL Config
HOWTO in the Tomcat 5 documentation bundle for more detailed
instructions):
* If your JDK version 1.3 or prior, download and install JSSE 1.0.2 or
later, and put the JAR files into "$JAVA_HOME/jre/lib/ext".
* Execute:
%JAVA_HOME%\bin\keytool -genkey -alias tomcat -keyalg RSA (Windows)
$JAVA_HOME/bin/keytool -genkey -alias tomcat -keyalg RSA (Unix)
with a password value of "changeit" for both the certificate and
the keystore itself.
By default, DNS lookups are enabled when a web application calls
request.getRemoteHost(). This can have an adverse impact on
performance, so you can disable it by setting the
"enableLookups" attribute to "false". When DNS lookups are disabled,
request.getRemoteHost() will return the String version of the
IP address of the remote client.
-->
<!-- Define a non-SSL HTTP/1.1 Connector on port 8080 -->
<Connector
port="5050" maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" redirectPort="8443" acceptCount="100"
connectionTimeout="20000" disableUploadTimeout="true" />
<!-- Note : To disable connection timeouts, set connectionTimeout value
to 0 -->
<!-- Note : To use gzip compression you could set the following properties :
compression="on"
compressionMinSize="2048"
noCompressionUserAgents="gozilla, traviata"
compressableMimeType="text/html,text/xml"
-->
<!-- Define a SSL HTTP/1.1 Connector on port 8443 -->
<!--
<Connector port="8443"
maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" disableUploadTimeout="true"
acceptCount="100" scheme="https" secure="true"
clientAuth="false" sslProtocol="TLS" />
-->
<!-- Define an AJP 1.3 Connector on port 8009 -->
<Connector port="8009"
enableLookups="false" redirectPort="8443" protocol="AJP/1.3" />
<!-- Define a Proxied HTTP/1.1 Connector on port 8082 -->
<!-- See proxy documentation for more information about using this. -->
<!--
<Connector port="8082"
maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" acceptCount="100" connectionTimeout="20000"
proxyPort="80" disableUploadTimeout="true" />
-->
<!-- An Engine represents the entry point (within Catalina) that processes
every request. The Engine implementation for Tomcat stand alone
analyzes the HTTP headers included with the request, and passes them
on to the appropriate Host (virtual host). -->
<!-- You should set jvmRoute to support load-balancing via AJP ie :
<Engine name="Standalone" defaultHost="localhost" jvmRoute="jvm1">
-->
<!-- Define the top level container in our container hierarchy -->
<Engine name="Catalina" defaultHost="localhost">
<!-- The request dumper valve dumps useful debugging information about
the request headers and cookies that were received, and the response
headers and cookies that were sent, for all requests received by
this instance of Tomcat. If you care only about requests to a
particular virtual host, or a particular application, nest this
element inside the corresponding <Host> or <Context> entry instead.
For a similar mechanism that is portable to all Servlet 2.4
containers, check out the "RequestDumperFilter" Filter in the
example application (the source for this filter may be found in
"$CATALINA_HOME/webapps/examples/WEB-INF/classes/filters").
Request dumping is disabled by default. Uncomment the following
element to enable it. -->
<!--
<Valve className="org.apache.catalina.valves.RequestDumperValve"/>
-->
<!-- Because this Realm is here, an instance will be shared globally -->
<!-- This Realm uses the UserDatabase configured in the global JNDI
resources under the key "UserDatabase". Any edits
that are performed against this UserDatabase are immediately
available for use by the Realm. -->
<Realm className="org.apache.catalina.realm.UserDatabaseRealm"
resourceName="UserDatabase"/>
<!-- Comment out the old realm but leave here for now in case we
need to go back quickly -->
<!--
<Realm className="org.apache.catalina.realm.MemoryRealm" />
-->
<!-- Replace the above Realm with one of the following to get a Realm
stored in a database and accessed via JDBC -->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="org.gjt.mm.mysql.Driver"
connectionURL="jdbc:mysql://localhost/authority"
connectionName="test" connectionPassword="test"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="oracle.jdbc.driver.OracleDriver"
connectionURL="jdbc:oracle:thin:@ntserver:1521:ORCL"
connectionName="scott" connectionPassword="tiger"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="sun.jdbc.odbc.JdbcOdbcDriver"
connectionURL="jdbc:odbc:CATALINA"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!-- Define the default virtual host
Note: XML Schema validation will not work with Xerces 2.2.
-->
<Host name="localhost" appBase="webapps"
unpackWARs="true" autoDeploy="true"
xmlValidation="false" xmlNamespaceAware="false">
<!-- Defines a cluster for this node,
By defining this element, means that every manager will be changed.
So when running a cluster, only make sure that you have webapps in there
that need to be clustered and remove the other ones.
A cluster has the following parameters:
className = the fully qualified name of the cluster class
name = a descriptive name for your cluster, can be anything
mcastAddr = the multicast address, has to be the same for all the nodes
mcastPort = the multicast port, has to be the same for all the nodes
mcastBindAddr = bind the multicast socket to a specific address
mcastTTL = the multicast TTL if you want to limit your broadcast
mcastSoTimeout = the multicast readtimeout
mcastFrequency = the number of milliseconds in between sending a "I'm alive" heartbeat
mcastDropTime = the number a milliseconds before a node is considered "dead" if no heartbeat is received
tcpThreadCount = the number of threads to handle incoming replication requests, optimal would be the same amount of threads as nodes
tcpListenAddress = the listen address (bind address) for TCP cluster request on this host,
in case of multiple ethernet cards.
auto means that address becomes
InetAddress.getLocalHost().getHostAddress()
tcpListenPort = the tcp listen port
tcpSelectorTimeout = the timeout (ms) for the Selector.select() method in case the OS
has a wakup bug in java.nio. Set to 0 for no timeout
printToScreen = true means that managers will also print to std.out
expireSessionsOnShutdown = true means that
useDirtyFlag = true means that we only replicate a session after setAttribute,removeAttribute has been called.
false means to replicate the session after each request.
false means that replication would work for the following piece of code: (only for SimpleTcpReplicationManager)
<%
HashMap map = (HashMap)session.getAttribute("map");
map.put("key","value");
%>
replicationMode = can be either 'pooled', 'synchronous' or 'asynchronous'.
* Pooled means that the replication happens using several sockets in a synchronous way. Ie, the data gets replicated, then the request return. This is the same as the 'synchronous' setting except it uses a pool of sockets, hence it is multithreaded. This is the fastest and safest configuration. To use this, also increase the nr of tcp threads that you have dealing with replication.
* Synchronous means that the thread that executes the request, is also the
thread the replicates the data to the other nodes, and will not return until all
nodes have received the information.
* Asynchronous means that there is a specific 'sender' thread for each cluster node,
so the request thread will queue the replication request into a "smart" queue,
and then return to the client.
The "smart" queue is a queue where when a session is added to the queue, and the same session
already exists in the queue from a previous request, that session will be replaced
in the queue instead of replicating two requests. This almost never happens, unless there is a
large network delay.
-->
<!--
When configuring for clustering, you also add in a valve to catch all the requests
coming in, at the end of the request, the session may or may not be replicated.
A session is replicated if and only if all the conditions are met:
1. useDirtyFlag is true or setAttribute or removeAttribute has been called AND
2. a session exists (has been created)
3. the request is not trapped by the "filter" attribute
The filter attribute is to filter out requests that could not modify the session,
hence we don't replicate the session after the end of this request.
The filter is negative, ie, anything you put in the filter, you mean to filter out,
ie, no replication will be done on requests that match one of the filters.
The filter attribute is delimited by ;, so you can't escape out ; even if you wanted to.
filter=".*\.gif;.*\.js;" means that we will not replicate the session after requests with the URI
ending with .gif and .js are intercepted.
The deployer element can be used to deploy apps cluster wide.
Currently the deployment only deploys/undeploys to working members in the cluster
so no WARs are copied upons startup of a broken node.
The deployer watches a directory (watchDir) for WAR files when watchEnabled="true"
When a new war file is added the war gets deployed to the local instance,
and then deployed to the other instances in the cluster.
When a war file is deleted from the watchDir the war is undeployed locally
and cluster wide
-->
<!--
<Cluster className="org.apache.catalina.cluster.tcp.SimpleTcpCluster"
managerClassName="org.apache.catalina.cluster.session.DeltaManager"
expireSessionsOnShutdown="false"
useDirtyFlag="true"
notifyListenersOnReplication="true">
<Membership
className="org.apache.catalina.cluster.mcast.McastService"
mcastAddr="228.0.0.4"
mcastPort="45564"
mcastFrequency="500"
mcastDropTime="3000"/>
<Receiver
className="org.apache.catalina.cluster.tcp.ReplicationListener"
tcpListenAddress="auto"
tcpListenPort="4001"
tcpSelectorTimeout="100"
tcpThreadCount="6"/>
<Sender
className="org.apache.catalina.cluster.tcp.ReplicationTransmitter"
replicationMode="pooled"
ackTimeout="15000"/>
<Valve className="org.apache.catalina.cluster.tcp.ReplicationValve"
filter=".*\.gif;.*\.js;.*\.jpg;.*\.htm;.*\.html;.*\.txt;"/>
<Deployer className="org.apache.catalina.cluster.deploy.FarmWarDeployer"
tempDir="/tmp/war-temp/"
deployDir="/tmp/war-deploy/"
watchDir="/tmp/war-listen/"
watchEnabled="false"/>
</Cluster>
-->
<!-- Normally, users must authenticate themselves to each web app
individually. Uncomment the following entry if you would like
a user to be authenticated the first time they encounter a
resource protected by a security constraint, and then have that
user identity maintained across all web applications contained
in this virtual host. -->
<!--
<Valve className="org.apache.catalina.authenticator.SingleSignOn" />
-->
<!-- Access log processes all requests for this virtual host. By
default, log files are created in the "logs" directory relative to
$CATALINA_HOME. If you wish, you can specify a different
directory with the "directory" attribute. Specify either a relative
(to $CATALINA_HOME) or absolute path to the desired directory.
-->
<!--
<Valve className="org.apache.catalina.valves.AccessLogValve"
directory="logs" prefix="localhost_access_log." suffix=".txt"
pattern="common" resolveHosts="false"/>
-->
<!-- Access log processes all requests for this virtual host. By
default, log files are created in the "logs" directory relative to
$CATALINA_HOME. If you wish, you can specify a different
directory with the "directory" attribute. Specify either a relative
(to $CATALINA_HOME) or absolute path to the desired directory.
This access log implementation is optimized for maximum performance,
but is hardcoded to support only the "common" and "combined" patterns.
-->
<!--
<Valve className="org.apache.catalina.valves.FastCommonAccessLogValve"
directory="logs" prefix="localhost_access_log." suffix=".txt"
pattern="common" resolveHosts="false"/>
-->
<Context path="/tdm" docBase="tdm" debug="0" reloadable="true" />
</Host>
</Engine>
</Service>
</Server>
I have set the context path to /tdm in the server.xml file. Should this be placed in context.xml?
My first page in the project is called Homepage.html. To start my project I give http://localhost:5050/tdm/homepage.html
in a browser. Here I accept a username and password from the user and then do the validation in
a valid.jsp file, where I connect to the database and check and use jsp:forward to go to next pages
accordingly. However when I enter the username and password and click Go in the homepage, nothing is
displayed on the next page. The URL in the browser says valid.jsp but a blank screen appears.
WHY DOES IT HAPPEN SO? DOES IT MEAN THAT TOMCAT IS NOT RECOGNIZING JAVA IN MY SYSTEM OR IS IT A PROBLEM
WITH THE DATABASE CONNECTION OR SOMETHING ELSE? I FEEL THAT TOMCAT IS NOT EXECUTING JSP COMMANDS?
IS IT POSSIBLE?WHY WILL THIS HAPPEN?
I set the JAVA_HOME and CATALINA_HOME environment to the jdk and tomcat folders resp.
Is there any other thing that I need to set in classpath? Should I have my project as a
WAR file in the webapps of TOMCAT or just a folder i.e. directory structure will fine?I am using Tomcat 5.5 and Jdk 1.5.0_12 and Oracle 10g. I am using jdbc-odbc bridge connection
to connect to the database. I have placed my project folder called
tdm under the webapps folder in Tomcat. This 'tdm' folder consists of
a collection of html pages,jsp pages and images of my project. Also I created a
WEB-INF folderand in that I have lib folder which contains catalina-root.jar
, classes12.jar and nls_charset.jar files. And also in the WEB-INF folder I have the web.xml
file which looks like this
<?xml version="1.0" encoding="ISO-8859-1"?>
<!--
Copyright 2004 The Apache Software Foundation
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-->
<web-app>
<resource-ref>
<description>Oracle Datasource example</description>
<res-ref-name>jdbc/gdn</res-ref-name>
<res-type>javax.sql.DataSource</res-type>
<res-auth>Container</res-auth>
</resource-ref>
</web-app>
My Server.xml file in Tomcat\conf folder is as follows
<!-- Example Server Configuration File -->
<!-- Note that component elements are nested corresponding to their
parent-child relationships with each other -->
<!-- A "Server" is a singleton element that represents the entire JVM,
which may contain one or more "Service" instances. The Server
listens for a shutdown command on the indicated port.
Note: A "Server" is not itself a "Container", so you may not
define subcomponents such as "Valves" or "Loggers" at this level.
-->
<Server port="8005" shutdown="SHUTDOWN">
<!-- Comment these entries out to disable JMX MBeans support used for the
administration web application -->
<Listener className="org.apache.catalina.mbeans.ServerLifecycleListener" />
<Listener className="org.apache.catalina.mbeans.GlobalResourcesLifecycleListener" />
<!-- Global JNDI resources -->
<GlobalNamingResources>
<!-- Test entry for demonstration purposes -->
<Environment name="simpleValue" type="java.lang.Integer" value="30"/>
<!-- Editable user database that can also be used by
UserDatabaseRealm to authenticate users -->
<Resource name="UserDatabase" auth="Container"
type="org.apache.catalina.UserDatabase"
description="User database that can be updated and saved"
factory="org.apache.catalina.users.MemoryUserDatabaseFactory"
pathname="conf/tomcat-users.xml" />
<Resource name="jdbc/gdn" auth="Container"
type="javax.sql.DataSource" driverClassName="sun.jdbc.odbc.JdbcOdbcDriver"
url="jdbc:odbc:gdn"
username="system" password="tiger" maxActive="20" maxIdle="10"
maxWait="-1"/>
</GlobalNamingResources>
<!-- A "Service" is a collection of one or more "Connectors" that share
a single "Container" (and therefore the web applications visible
within that Container). Normally, that Container is an "Engine",
but this is not required.
Note: A "Service" is not itself a "Container", so you may not
define subcomponents such as "Valves" or "Loggers" at this level.
-->
<!-- Define the Tomcat Stand-Alone Service -->
<Service name="Catalina">
<!-- A "Connector" represents an endpoint by which requests are received
and responses are returned. Each Connector passes requests on to the
associated "Container" (normally an Engine) for processing.
By default, a non-SSL HTTP/1.1 Connector is established on port 8080.
You can also enable an SSL HTTP/1.1 Connector on port 8443 by
following the instructions below and uncommenting the second Connector
entry. SSL support requires the following steps (see the SSL Config
HOWTO in the Tomcat 5 documentation bundle for more detailed
instructions):
* If your JDK version 1.3 or prior, download and install JSSE 1.0.2 or
later, and put the JAR files into "$JAVA_HOME/jre/lib/ext".
* Execute:
%JAVA_HOME%\bin\keytool -genkey -alias tomcat -keyalg RSA (Windows)
$JAVA_HOME/bin/keytool -genkey -alias tomcat -keyalg RSA (Unix)
with a password value of "changeit" for both the certificate and
the keystore itself.
By default, DNS lookups are enabled when a web application calls
request.getRemoteHost(). This can have an adverse impact on
performance, so you can disable it by setting the
"enableLookups" attribute to "false". When DNS lookups are disabled,
request.getRemoteHost() will return the String version of the
IP address of the remote client.
-->
<!-- Define a non-SSL HTTP/1.1 Connector on port 8080 -->
<Connector
port="5050" maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" redirectPort="8443" acceptCount="100"
connectionTimeout="20000" disableUploadTimeout="true" />
<!-- Note : To disable connection timeouts, set connectionTimeout value
to 0 -->
<!-- Note : To use gzip compression you could set the following properties :
compression="on"
compressionMinSize="2048"
noCompressionUserAgents="gozilla, traviata"
compressableMimeType="text/html,text/xml"
-->
<!-- Define a SSL HTTP/1.1 Connector on port 8443 -->
<!--
<Connector port="8443"
maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" disableUploadTimeout="true"
acceptCount="100" scheme="https" secure="true"
clientAuth="false" sslProtocol="TLS" />
-->
<!-- Define an AJP 1.3 Connector on port 8009 -->
<Connector port="8009"
enableLookups="false" redirectPort="8443" protocol="AJP/1.3" />
<!-- Define a Proxied HTTP/1.1 Connector on port 8082 -->
<!-- See proxy documentation for more information about using this. -->
<!--
<Connector port="8082"
maxThreads="150" minSpareThreads="25" maxSpareThreads="75"
enableLookups="false" acceptCount="100" connectionTimeout="20000"
proxyPort="80" disableUploadTimeout="true" />
-->
<!-- An Engine represents the entry point (within Catalina) that processes
every request. The Engine implementation for Tomcat stand alone
analyzes the HTTP headers included with the request, and passes them
on to the appropriate Host (virtual host). -->
<!-- You should set jvmRoute to support load-balancing via AJP ie :
<Engine name="Standalone" defaultHost="localhost" jvmRoute="jvm1">
-->
<!-- Define the top level container in our container hierarchy -->
<Engine name="Catalina" defaultHost="localhost">
<!-- The request dumper valve dumps useful debugging information about
the request headers and cookies that were received, and the response
headers and cookies that were sent, for all requests received by
this instance of Tomcat. If you care only about requests to a
particular virtual host, or a particular application, nest this
element inside the corresponding <Host> or <Context> entry instead.
For a similar mechanism that is portable to all Servlet 2.4
containers, check out the "RequestDumperFilter" Filter in the
example application (the source for this filter may be found in
"$CATALINA_HOME/webapps/examples/WEB-INF/classes/filters").
Request dumping is disabled by default. Uncomment the following
element to enable it. -->
<!--
<Valve className="org.apache.catalina.valves.RequestDumperValve"/>
-->
<!-- Because this Realm is here, an instance will be shared globally -->
<!-- This Realm uses the UserDatabase configured in the global JNDI
resources under the key "UserDatabase". Any edits
that are performed against this UserDatabase are immediately
available for use by the Realm. -->
<Realm className="org.apache.catalina.realm.UserDatabaseRealm"
resourceName="UserDatabase"/>
<!-- Comment out the old realm but leave here for now in case we
need to go back quickly -->
<!--
<Realm className="org.apache.catalina.realm.MemoryRealm" />
-->
<!-- Replace the above Realm with one of the following to get a Realm
stored in a database and accessed via JDBC -->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="org.gjt.mm.mysql.Driver"
connectionURL="jdbc:mysql://localhost/authority"
connectionName="test" connectionPassword="test"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="oracle.jdbc.driver.OracleDriver"
connectionURL="jdbc:oracle:thin:@ntserver:1521:ORCL"
connectionName="scott" connectionPassword="tiger"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!--
<Realm className="org.apache.catalina.realm.JDBCRealm"
driverName="sun.jdbc.odbc.JdbcOdbcDriver"
connectionURL="jdbc:odbc:CATALINA"
userTable="users" userNameCol="user_name" userCredCol="user_pass"
userRoleTable="user_roles" roleNameCol="role_name" />
-->
<!-- Define the default virtual host
Note: XML Schema validation will not work with Xerces 2.2.
-->
<Host name="localhost" appBase="webapps"
unpackWARs="true" autoDeploy="true"
xmlValidation="false" xmlNamespaceAware="false">
<!-- Defines a cluster for this node,
By defining this element, means that every manager will be changed.
So when running a cluster, only make sure that you have webapps in there
that need to be clustered and remove the other ones.
A cluster has the following parameters:
className = the fully qualified name of the cluster class
name = a descriptive name for your cluster, can be anything
mcastAddr = the multicast address, has to be the same for all the nodes
mcastPort = the multicast port, has to be the same for all the nodes
mcastBindAddr = bind the multicast socket to a specific address
mcastTTL = the multicast TTL if you want to limit your broadcast
mcastSoTimeout = the multicast readtimeout
mcastFrequency = the number of milliseconds in between sending a "I'm alive" heartbeat
mcastDropTime = the number a milliseconds before a node is considered "dead" if no heartbeat is received
tcpThreadCount = the number of threads to handle incoming replication requests, optimal would be the same amount of threads as nodes
tcpListenAddress = the listen address (bind address) for TCP cluster request on this host,
in case of multiple ethernet cards.
auto means that address becomes
InetAddress.getLocalHost().getHostAddress()
tcpListenPort = the tcp listen port
tcpSelectorTimeout = the timeout (ms) for the Selector.select() method in case the OS
has a wakup bug in java.nio. Set to 0 for no timeout
printToScreen = true means that managers will also print to std.out
expireSessionsOnShutdown = true means that
useDirtyFlag = true means that we only replicate a session after setAttribute,removeAttribute has been called.
false means to replicate the session after each request.
false means that replication would work for the following piece of code: (only for SimpleTcpReplicationManager)
<%
HashMap map = (HashMap)session.getAttribute("map");
map.put("key","value");
%>
replicationMode = can be either 'pooled', 'synchronous' or 'asynchronous'.
* Pooled means that the replication happens using several sockets in a synchronous way. Ie, the data gets replicated, then the request return. This is the same as the 'synchronous' setting except it uses a pool of sockets, hence it is multithreaded. This is the fastest and safest configuration. To use this, also increase the nr of tcp threads that you have dealing with replication.
* Synchronous means that the thread that executes the request, is also the
thread the replicates the data to the other nodes, and will not return until all
nodes have received the information.
* Asynchronous means that there is a specific 'sender' thread for each cluster node,
so the request thread will queue the replication request into a "smart" queue,
and then return to the client.
The "smart" queue is a queue where when a session is added to the queue, and the same session
already exists in the queue from a previous request, that session will be replaced
in the queue instead of replicating two requests. This almost never happens, unless there is a
large network delay.
-->
<!--
When configuring for clustering, you also add in a valve to catch all the requests
coming in, at the end of the request, the session may or may not be replicated.
A session is replicated if and only if all the conditions are met:
1. useDirtyFlag is true or setAttribute or removeAttribute has been called AND
2. a session exists (has been created)
3. the request is not trapped by the "filter" attribute
The filter attribute is to filter out requests that could not modify the session,
hence we don't replicate the session after the end of this request.
The filter is negative, ie, anything you put in the filter, you mean to filter out,
ie, no replication will be done on requests that match one of the filters.
The filter attribute is delimited by ;, so you can't escape out ; even if you wanted to.
filter=".*\.gif;.*\.js;" means that we will not replicate the session after requests with the URI
ending with .gif and .js are intercepted.
The deployer element can be used to deploy apps cluster wide.
Currently the deployment only deploys/undeploys to working members in the cluster
so no WARs are copied upons startup of a broken node.
The deployer watches a directory (watchDir) for WAR files when watchEnabled="true"
When a new war file is added the war gets deployed to the local instance,
and then deployed to the other instances in the cluster.
When a war file is deleted from the watchDir the war is undeployed locally
and cluster wide
-->
<!--
<Cluster className="org.apache.catalina.cluster.tcp.SimpleTcpCluster"
managerClassName="org.apache.catalina.cluster.session.DeltaManager"
expireSessionsOnShutdown="false"
useDirtyFlag="true"
notifyListenersOnReplication="true">
<Membership
className="org.apache.catalina.cluster.mcast.McastService"
mcastAddr="228.0.0.4"
mcastPort="45564"
mcastFrequency="500"
mcastDropTime="3000"/>
<Receiver
className="org.apache.catalina.cluster.tcp.ReplicationListener"
tcpListenAddress="auto"
tcpListenPort="4001"
tcpSelectorTimeout="100"
tcpThreadCount="6"/>
<Sender
className="org.apache.catalina.cluster.tcp.ReplicationTransmitter"
replicationMode="pooled"
ackTimeout="15000"/>
<Valve className="org.apache.catalina.cluster.tcp.ReplicationValve"
filter=".*\.gif;.*\.js;.*\.jpg;.*\.htm;.*\.html;.*\.txt;"/>
<Deployer className="org.apache.catalina.cluster.deploy.FarmWarDeployer"
tempDir="/tmp/war-temp/"
deployDir="/tmp/war-deploy/"
watchDir="/tmp/war-listen/"
watchEnabled="false"/>
</Cluster>
-->
<!-- Normally, users must authenticate themselves to each web app
individually. Uncomment the following entry if you would like
a user to be authenticated the first time they encounter a
resource protected by a security constraint, and then have that
user identity maintained across all web applications contained
in this virtual host. -->
<!--
<Valve className="org.apache.catalina.authenticator.SingleSignOn" />
-->
<!-- Access log processes all requests for this virtual host. By
default, log files are created in the "logs" directory relative to
$CATALINA_HOME. If you wish, you can specify a different
directory with the "directory" attribute. Specify either a relative
(to $CATALINA_HOME) or absolute path to the desired directory.
-->
<!--
<Valve className="org.apache.catalina.valves.AccessLogValve"
directory="logs" prefix="localhost_access_log." suffix=".txt"
pattern="common" resolveHosts="false"/>
-->
<!-- Access log processes all requests for this virtual host. By
default, log files are created in the "logs" directory relative to
$CATALINA_HOME. If you wish, you can specify a different
directory with the "directory" attribute. Specify either a relative
(to $CATALINA_HOME) or absolute path to the desired directory.
This access log implementation is optimized for maximum performance,
but is hardcoded to support only the "common" and "combined" patterns.
-->
<!--
<Valve className="org.apache.catalina.valves.FastCommonAccessLogValve"
directory="logs" prefix="localhost_access_log." suffix=".txt"
pattern="common" resolveHosts="false"/>
-->
<Context path="/tdm" docBase="tdm" debug="0" reloadable="true" />
</Host>
</Engine>
</Service>
</Server>
I have set the context path to /tdm in the server.xml file. Should this be placed in context.xml?
My first page in the project is called Homepage.html. To start my project I give http://localhost:5050/tdm/homepage.html
in a browser. Here I accept a username and password from the user and then do the validation in
a valid.jsp file, where I connect to the database and check and use jsp:forward to go to next pages
accordingly. However when I enter the username and password and click Go in the homepage, nothing is
displayed on the next page. The URL in the browser says valid.jsp but a blank screen appears.
WHY DOES IT HAPPEN SO? DOES IT MEAN THAT TOMCAT IS NOT RECOGNIZING JAVA IN MY SYSTEM OR IS IT A PROBLEM
WITH THE DATABASE CONNECTION OR SOMETHING ELSE? I FEEL THAT TOMCAT IS NOT EXECUTING JSP COMMANDS?
IS IT POSSIBLE?WHY WILL THIS HAPPEN?
I set the JAVA_HOME and CATALINA_HOME environment to the jdk and tomcat folders resp.
Is there any other thing that I need to set in classpath? Should I have my project as a
WAR file in the webapps of TOMCAT or just a folder i.e. directory structure will fine? -
Hi, I am administering a cutover from Office Pro 2010 to Office Pro 2013 via our Office 365 E3 licenses. Before I start migrating mailboxes from our Exchange 2003 server to Exchange Online, I want to have our staff install Office Pro 2013 including Outlook
2013 so they can have time to review it before we switch to the Exchange Online server in a couple of days. I know Outlook 2013 is not compatible with our Exchange 2003 server but I was hoping to be able to setup Outlook 2013 to connect to my Office 365 email
address while keeping my msi installed version of Outlook 2010 for my Exchange 2003 email address.
I am trying to test the capability of Outlook 2013 running side-by-side with my Outlook 2010 program. I have the "Click-to-Run" version installed on my Windows 7 64-bit machine however when I try to open Outlook 2013, I get an error message: "Another
version of Outlook is currently running. Close it and try again."
In reading in TechNet and other sources, it clearly states that both programs can run simultaneously which is what I need to do
but obviously something is not configured correctly.
Can you please help me get this configuration running or let me know if it is in fact not possible?
Thanks!
FrankYes, it should be able to install side by side.
The versions of Office that you install must be the same edition. For example, both installations of Office are the 32-bit edition.
When you get the error
"Another version of Outlook is currently running. Close it and try again.", go to task manager and end all existing
outlook process.
If the issue persists, do a repair for your Office suites in Control panel>Program and features>Office>Change>Repair
Thanks,
Tyor Wang
Forum Support
Come back and mark the replies as answers if they help and unmark them
if they provide no help.
If you have any feedback on our support, please click
here -
Trying to configure Portal to run off the network
Hi all,
I am trying to configure Oracle Portal 3.0.9.8.2 to run off the network. I have taken all of the recommended steps in terms of configuring hostname, using a static IP, running and checking the ssoodatan script, updating the hosts file, and configuring a loopback adapter, but can still not make it work. The Portal works and functions fine when connected to the network, but as soon as disconnected from the network Portal returns
Error: Timeout occurred whilst retrieving page meta data 0 ms.
or
Sometimes it will display the banner, but the Portlets and the Providers will instantly timeout. I have cranked up the timeout levels of JServ and the Providers but to no avail. I have put the recommended timeout, stall, requesttime, buffer, minStall, minTimeout in the zone.properties file, but no luck.
I have noticed within the jserv.log the error message below. The link that jumps out at me is UnknownHostException: www-proxy.us.oracle.com. I did have my proxy server set at one time, but I erased it from the global settings page, and I ran the procedure I found somewhere on this discussion board to make sure it was set back to null. When I run the diag.cmd program it shows my proxy server being empty. Where could this be picking this up from, I also have proxy servers turned off in my browsers. Any suggestions or help??
13/02/2002 16:04:08:752 PST] page/JNI: Exception when trying to connect in 1.
[13/02/2002 16:04:08:752 PST] page/Timeout occurred, label=147 url=http://jstrada-lap.us.oracle.com/pls/portal30/!PORTAL30.wwpro_app_provider.execute_portlet time=10ms
[13/02/2002 16:04:08:762 PST] page/ContentFetcher Unexpected Exception, name=content-fetcher0
java.net.UnknownHostException: www-proxy.us.oracle.com
at java.net.InetAddress.getAllByName0(InetAddress.java:575)
at java.net.InetAddress.getAllByName0(InetAddress.java:544)
at java.net.InetAddress.getAllByName(InetAddress.java:537)
at HTTPClient.HTTPConnection$EstablishConnection.run(HTTPConnection.java, Compiled Code)I think you must add an entry to your /etc/hosts file, as follows:
127.0.0.1 noa noa
and create a vhost entry on /etc/httpd/conf/extra/httpd-vhosts
<VirtualHost *:80>
ServerName noa
DocumentRoot "/srv/http/"
<Directory "/src/http/">
Options Indexes FollowSymLinks
AllowOverride None
Require all granted
</Directory>
</VirtualHost>
Although I'm not sure if the vhost is actually necesary, but it won't harm.
pd: if you use vhosts, remember to uncomment "Include conf/extra/httpd-vhosts.conf" con /etc/httpd/conf/httpd.conf and restart apache
# systemctl restart httpd -
Cannot configure webutil to run client_host command
I am using WLS 10.3.5 with FMw 11.1.2 (Forms/Reports 11 Release 2), Forms Builder, Windows 7, Chrome browser
1) every thing was working fine, i was creating forms and running them Ok.
2) then, I want to use client_host command but i am unable to (as i could not configure the webutil package).
runs fine: host('cmd /c E:\Oracle\file.pdf');
does not compile: client_host('cmd /c E:\Oracle\file.pdf');
3) I have searched the forums for this and the threads point to the webutil_106.zip.
4) I have tried to configure webutil on my own by using the readme with this package: [http://www.oracle.com/technology/software/products/forms/files/webutil/webutil_106.zip], but i have messed up. Before trying to configure webutil, my forms 11g was running fine, but now when i run any form it gives me the following error window:
Application Error
classNotFoundException
oracle.forms.engine.Main
This is my first post in oracle forums, if i have missed any detail kindly guide me how to get it to you.
Thanks
ZuhayerZuhayer wrote:
This is my first post in oracle forums, if i have missed any detail kindly guide me how to get it to you.
As salamualikum, Zuhayer
Welcome to the Oracle Forums. Please take a few minutes to review the following:
<ul>
<li>Oracle Forums FAQ
<li>Before posting on this forum please read
<li>10 Commandments for the OTN Forums Member
<li>How to ask questions the smart way
</ul>
Following these simple guidelines will ensure you have a positive experience in any forum; not just this one!
I am using WLS 10.3.5 with FMw 11.1.2 (Forms/Reports 11 Release 2), Forms Builder, Windows 7, Chrome browser
1) every thing was working fine, i was creating forms and running them Ok.
2) then, I want to use client_host command but i am unable to (as i could not configure the webutil package).
runs fine: host('cmd /c E:\Oracle\file.pdf');
does not compile: client_host('cmd /c E:\Oracle\file.pdf');
3) I have searched the forums for this and the threads point to the webutil_106.zip.
4) I have tried to configure webutil on my own by using the readme with this package: [http://www.oracle.com/technology/software/products/forms/files/webutil/webutil_106.zip], but i have messed up. Before trying to configure webutil, my forms 11g was running fine, but now when i run any form it gives me the following error window:
Application Error
classNotFoundException
oracle.forms.engine.MainCheck this.Configure Webutil 11g at Windows
Hope this helps
Hamid -
After configuring SSL, cannot run Report Manager but ReportServer works fine
I'm using SSRS 2014 as part of SQL Server 2014 Express. I've configured both the Web Service URL and the Report Manager URL to use a Certificate and only allow port 443 traffic. I did this by removing the port 80 identity and adding an SSL Identity.
All is working well for https://myserver.com/ReportServer. But when I try and access report manager, I get the following error when running remote:
The underlying connection was closed: Could not
establish trust relationship for the SSL/TLS secure channel.
When running directly on the server, the server, it just rejects my passwords and asks for it again.
If I remove the SSL Indentity and add back in port 80, it works fine over https.
Any ideas of what might be causing this problem?
Thanks,
Glenn
Please mark this post as Answered if it solved your problem, otherwise you can click Vote if it helped you out. -Glenn ThimmesHi Glenn,
Per my understanding that you have added an SSL Identity and you got some errors "The underlying connection was closed: Could not establish trust relationship for the SSL/TLS secure channel." when you logging remotely to the report manager,
but which work fine when login report server, and it also asked for credential when you login report manager locally, right?
The issue can be caused by many factors. Please check below points one by one to help you clarify:
In rsreportserver.config file(Path like this:C:\Program Files\Microsoft SQL Server\MSRS10.MSSQLSERVER\Reporting Services\ReportServer), ensure that SecureConnectionLevel is set to the value 2 or 3. If disable the SSL value is set to 0. Corrected
line in your rsreportserver.confiog file should look like:
<Add Key="SecureConnectionLevel" Value="3"/>
In rsreportserver.config, add the correct value to the <URLRoot> element. the ReportServerUrl value must be the name of the server that is registered for that certificate. with a hostname that is valid for your certificate. For example, if the
certificate is registered to https://adventure-works.com.onlinesales, the report server URL would be
https://adventure-works.com.onlinesales/reportserver.
<UrlRoot>
https://adventure-works.com.onlinesales/reportserver.
</UrlRoot>
Ensure that your certificate exists in Trusted Root Certification Authorities in certmgr for the local machine.
http://www.c-sharpcorner.com/uploadfile/anavijai/could-not-establish-trust-relationship-for-the-ssltls-secure-channel/
This is very important. In rsreportserver.config file, add the same value to the <ReportServerUrl> element the same as that you added in step 2.
<ReportServerUrl>
https://adventure-works.com.onlinesales/reportserver.
</ReportServerUrl>
Then restart your report server (stop & then start in Report Server Configuration Manager).
Article related to this issue also contain the troubleshooting for your reference:
http://msdn.microsoft.com/en-us/library/ms345223(v=sql.100).aspx
Similar threads for your reference:
Could not establish trust relationship for the SSL/TLS secure
channel
SSRS 2008- Underlying connection
was closed – could not establish trust relationship for SSL/TLS
If your problem still exists, please try to provide us snapshopt about the setting for the SSL Identity and also the Report Manager URL,etc.
Regards
Vicky Liu
Maybe you are looking for
-
I purchased PSCS3 Extended because it looked like it would really be great. I'm an architect and I was really looking forward to the measurement tool... I have to measure things off of job site pics all the time. I have to say the measurement tool in
-
windows 7: when I enlarge text or font size to my preference it goes back to original when I change a page, I want to set it and forget it thanks.
-
I keep getting error message -8084
I keep getting this error message -8084 when I copy from an external NAS drive. Does anyone know what this means and how to prevent it from happening?
-
To add items in purchase order
Hi abapers, I created a purchase order.In which i passed only single item.....Now if i want to add more items what should i do.....Which function module i have to use for it... Can anybody send me example code for bapi ....
-
Sales Order Line item characteristics (classes)
I am not 100% sure what I am talking about but I need to get characteristics for materials on sales orders (line items). I am told that these characteristics are stored in classes which get assigned to materials via the material master. I am also t