Create secondary node on BOBJ cluster environment

Hi All,
I have done migration from old BOBJ server to new BOBJ server. The primary nodes has successfully completed and manage to view reports. I currently having issue to proceed installing second node on BOBJ cluster environment. i stuck at phase repository database as no database entry was found.
Please help. I also attach my steps to install the second node.
Thanks.
Regards
Aiman

Hi,
im guessing you do an Expand installation on the second node and want to enter the credentials of your existing CMS DB, the one you configured during the installation of Node 1. Correct?
What is your CMS DB of a Kind? MS SQL? Oracle?
It looks like that the ODBC DSN entries are missing on Node 2.
Regards
-Seb.

Similar Messages

  • Can we create secondary index for a cluster table

    hi
    can we create secondary index for a cluster table

    Jyothsna,
    There seems to be some kind of misunderstanding here. You <i>cannot</i> create a secondary index on a cluster table. A cluster table does not exist as a separate physical table in the database; it is part of a "physical cluster". In the case of BSEG for instance, the physical cluster is RFBLG. The only fields of the cluster table that also exist as fields of the physical cluster are the leading fields of the primary key. Taking again BSEG as the example, the primary key includes the fields MANDT, BUKRS, BELNR, GJAHR, BUZEI. If you look at the structure of the RFBLG table, you will see that it has primary key fields MANDT, BUKRS, BELNR, GJAHR, PAGENO. The first four fields are those that all cluster tables inside BSEG have in common. The fifth field, PAGENO, is a "technical" field giving the sequence number of the current record in the series of cluster records sharing the same primary key.
    All the "functional" fields of the cluster table (for BSEG this is field BUZEI and everything beyond that) exist only inside a raw binary object. The database does not know about these fields, it only sees the raw object (the field VARDATA of the physical cluster). Since the field does not exist in the database, it is impossible to create a secondary index on it. If you try to create a secondary index on a cluster table in transaction SE11, you will therefore rightly get the error "Index maintenance only possible for transparent tables".
    Theoretically you could get around this by converting the cluster table to a transparent table. You can do this in the SAP dictionary. However, in practice this is almost never a good solution. The table becomes much larger (clusters are compressed) and you lose the advantage that related records are stored close to each other (the main reason for having cluster tables in the first place). Apart from the performance and disk space hit, converting a big cluster table like BSEG to transparent would take extremely long.
    In cases where "indexing" of fields of a cluster table is worthwhile, SAP has constructed "indexing tables" around the cluster. For example, around BSEG there are transparent tables like BSIS, BSAS, etc. Other clusters normally do not have this, but that simply means there is no reason for having it. I have worked with the SAP dictionary for over 12 years and I have never met a single case where it was necessary to convert a cluster to transparent.
    If you try to select on specific values of a non-transparent field in a cluster without also specifying selections for the primary key, then the database will have to do a serial read of the whole physical cluster (and the ABAP DB interface will have to decompress every single record to extract the fields). The performance of that is monstrous -- maybe that was the reason of your question. However, the solution then is (in the case of BSEG) to query via one of the index tables (where you are free to create secondary indexes since those tables are transparent).
    Hope this clarifies things,
    Mark

  • Creating Secondary Node for Oracle Apps

    I've got an interesting issue.
    Background -- We're running Oracle Apps 11.5.9 in an n-tier environment.
    Servers -- Running HP-UX 11i (for both DB tier, Application tier).
    Goad -- Implement MWA server. (mobile supply chain).
    We started an implementation before and it had problems. The MWA devices (wireless barcode scanners) kept on freezing and we could never find the processes to kill. The only way to fix was to restart the application tier (uncool). More testing went and we found lots of patches to apply and did so. Things seem to be working fine, but I'd like to create a secondary node JUST for the MWA devices.
    What it boils down to is creating a secondary node for Oracle Apps that ONLY the mobile devices will use. All other users will use the primary node. Once I've found that the MWA devices are working fine, then I'll load balance, but not for now.
    I'm fairly new to Oracle Apps (1 year exp -- boss just left ACK!) and I'm not certian on how to create the secondary node and make it unavailable to all other users.
    Any help is GREATLY APPRECIATED!
    Thanks in Advance!

    Hi Jeremy,
    Did you manage to solve your MWA devices (wireless barcode scanners) freezing by adding an extra node? there is a client that is experiencing the same problem, I have trive to increase the number of ports from 10200, adding 10202 & 10204 when I start the Mobile application and increasing memory from mx1024 - ms512 to mx2048 - ms1024 & all is fruitless, we rebbot our MWA everyday at night as the time goes on during the day approx 13h00 all hell break loose, slow response, freezing of barcode scanners, everything stops.
    Regards
    Sihle

  • Rapid clone 11.5.10  from single node to Veritas cluster environment

    Is anybody there who had done the same and working? Please shed some light.
    version 11.5.10
    Source
    DB tier : DB, CM and Admin ( OS : Solaris 10 ), single node
    App tier : FrmSrv, iAS ( OS : Solaris 10), single node
    Target
    DB tier : DB, CM, and Admin ( OS : Solaris 10), 2-nodes Veritas Cluster
    (active/passive)
    App tier : FrmSrv, iAS (OS : Solaris 10), 2-nodes load balancer
    I've many time single node to single node rapid clone , no problem.
    But with this single node to multi-nodes, no luck.
    your input is much appreciated

    Hi
    You can use the Advanced cloning (section 4) of the cloining document# 230672.1
    to clone signle node to multi node systems.
    for webserver load balancing follow the metalink note#:217368.1
    Regards
    Srinath

  • Publishing does not clear cache on secondary node on the cluster

    Hello,
    I am publishing assets from Management to Delivery. Delivery consists of 2 nodes, one of which is the publish destination for the management. So, after a publish, I could see the updated content on Delivery 1 (cs/Satellite) but Delivery 2 (cs/Satellite) still shows stale content.
    I checked the following properties on ft.ini and it's holding correcting values
    cc.cacheNoSync, ft.sync, ft.usedisksync
    Any help/pointers is very much appreciated.
    Thanks!

    Seems like your cache flushing is not working correctly - easiest way to see if your system is configured correctly with regard to the satellite servers - go into support tools / cache / list pages - pick Satellite - - it will log (remotely) onto both satellites and list the inventory - it validates the entries in your system satellite table - that table needs to list all of your delivery satellite servers and credentials -

  • Create directory on a two-node Oracle server cluster

    Hi,
    How can I create directory on a two-node Oracle server cluster (2 identical servers running the same Oracle database instances with shared disks)? Both of them run Windows 2008.
    I know this works for Oracle running in a single server. How about in failover cluster environment?
    CREATE OR REPLACE DIRECTORY g_vid_lib AS 'X:\video\library\g_rated';
    Thanks.
    Andy

    Using 11.2.0.? Using ASM? use ACFS - it is a SHARED file system.
    http://docs.oracle.com/cd/E16338_01/server.112/e10500/asmfilesystem.htm
    create [big empty disk group]
    create ACFS volume {container file that lives in that diskgroup]
    create ACFS file system
    mount the file system.
    Now, all nodes in the cluster can access that shared device at the OS level just as if it were any other device.

  • HOWTO: Create 2-node Solaris Cluster 4.1/Solaris 11.1(x64) using VirtualBox

    I did this on VirtualBox 4.1 on Windows 7 and VirtualBox 4.2 on Linux.X64. Basic pre-requisites are : 40GB disk space, 8GB RAM, 64-bit guest capable VirtualBox.
    Please read all the descriptive messages/prompts shown by 'scinstall' and 'clsetup' before answering.
    0) Download from OTN
    - Solaris 11.1 Live Media for x86(~966 MB)
    - Complete Solaris 11.1 IPS Repository Image (total 7GB)
    - Oracle Solaris Cluster 4.1 IPS Repository image (~73MB)
    1) Run VirtualBox Console, create VM1 : 3GB RAM, 30GB HDD
    2) The new VM1 has 1 NIC, add 2 more NICs (total 3). Setting the NIC to any type should be okay, 'VirtualBox Host Only Adapter' worked fine for me.
    3) Start VM1, point the "Select start-up disk" to the Solaris 11.1 Live Media ISO.
    4) Select "Oracle Solaris 11.1" in the GRUB menu. Select Keyboard layout and Language.
    VM1 will boot and the Solaris 11.1 Live Desktop screen will appear.
    5) Click <Install Oracle Solaris> from the desktop, supply necessary inputs.
    Default Disk Discovery (iSCSI not needed) and Disk Selection are fine.
    Disable the "Support Registration" connection info
    6) The alternate user created during the install has root privileges (sudo). Set appropriate VM1 name
    7) When the VM has to be rebooted after the installation is complete, make sure the Solaris 11.1 Live ISO is ejected or else the VM will again boot from the Live CD.
    8) Repeat steps 1-6, create VM2 and install Solaris.
    9) FTP(secure) the Solaris 11.1 Repository IPS and Solaris Cluster 4.1 IPS onto both the VMs e.g under /home/user1/
    10) We need to setup both the packages: Solaris 11.1 Repository and Solaris Cluster 4.1
    11) All commands now to be run as root
    12) By default the 'solaris' repository is of type online (pkg.oracle.com), that needs to be updated to the local ISO we downloaded :-
    +$ sudo sh+
    +# lofiadm -a /home/user1/sol-11_1-repo-full.iso+
    +//output : /dev/lofi/N+
    +# mount -F hsfs /dev/lofi/N /mnt+
    +# pkg set-publisher -G '*' -M '*' -g /mnt/repo solaris+
    13) Setup the ha-cluster package :-
    +# lofiadm -a /home/user1/osc-4_1-ga-repo-full.iso+
    +//output : /dev/lofi/N+
    +# mkdir /mnt2+
    +# mount -f hsfs /dev/lofi/N /mnt2+
    +# pkg set-publisher -g file:///mnt2/repo ha-cluster+
    14) Verify both packages are fine :-
    +# pkg publisher+
    PUBLISHER                   TYPE     STATUS P LOCATION
    solaris                     origin   online F file:///mnt/repo/
    ha-cluster                  origin   online F file:///mnt2/repo/
    15) Install the complete SC4.1 package by installing 'ha-cluster-full'
    +# pkg install ha-cluster-full+
    14) Repeat steps 12-15 on VM2.
    15) Now both VMs have the OS and SC4.1 installed.
    16) By default the 3 NICs are in the "Automatic" profile and have DHCP configured. We need to activate the Fixed profile and put the 3 NICs into it. Only 1 interface, the public interface, needs to be
    configured. The other 2 are for the cluster interconnect and will be automatically configured by scinstall. Execute the following commands :-
    +# netadm enable -p ncp defaultfixed+
    +//verify+
    +# netadm list -p ncp defaultfixed+
    +#Configure the public-interface+
    +#Verify none of the interfaces are listed, add all the 3+
    +# ipadm show-if+
    +# run dladm show-phys or dladm show-link to check interface names : must be net0/net1/net2+
    +# ipadm create-ip net0+
    +# ipadm create-ip net1+
    +# ipadm create-ip net2+
    +# ipadm show-if+
    +//select proper IP and configure the public interface. I have used 192.168.56.171 & 172+
    +# ipadm create-addr -T static -a 192.168.56.171/24 net0/publicip+
    +#IP plumbed, restart+
    +# ipadm down-addr -t net0/publicip+
    +# ipadm up-addr -t net0/publicip+
    +//Verify publicip is fine by pinging the host+
    +# ping 192.168.56.1+
    +//Verify, net0 should be up, net1/net2 should be down+
    +# ipadm+
    17) Repeat step 16 on VM2
    18) Verify both VMs can ping each other using the public IP. Add entries to each other's /etc/hosts
    Now we are ready to run scinstall and create/configure the 2-node cluster
    19)
    +# cd /usr/cluster/bin+
    +# ./scinstall+
    select 1) Create a new cluster ...
    select 1) Create a new cluster
    select 2) Custom in "Typical or Custom Mode"
    Enter cluster name : mycluster1 (e.g)
    Add the 2 nodes : solvm1 & solvm2 and press <ctrl-d>
    Accept default "No" for <Do you need to use DES authentication>"
    Accept default "Yes" for <Should this cluster use at least two private networks>
    Enter "No" for <Does this two-node cluster use switches>
    Select "1)net1" for "Select the first cluster transport adapter"
    If there is warning of unexpected traffic on "net"1, ignore it
    Enter "net1" when it asks corresponding adapter on "solvm2"
    Select "2)net2" for "Select the second cluster transport adapter"
    Enter "net2" when it asks corresponding adapter on "solvm2"
    Select "Yes" for "Is it okay to accept the default network address"
    Select "Yes" for "Is it okay to accept the default network netmask"Now the IP addresses 172.16.0.0 will be plumbed in the 2 private interfaces
    Select "yes" for "Do you want to turn off global fencing"
    (These are SATA serial disks, so no fencing)
    Enter "Yes" for "Do you want to disable automatic quorum device selection"
    (we will add quorum disks later)
    Enter "Yes" for "Proceed with cluster creation"
    Select "No" for "Interrupt cluster creation for cluster check errors"
    The second node will be configured and 2nd node rebooted
    The first node will be configured and rebootedAfter both nodes have rebooted, verify the cluster has been created and both nodes joined.
    On both nodes :-
    +# cd /usr/cluster/bin+
    +# ./clnode status+
    +//should show both nodes Online.+
    At this point there are no quorum disks, so 1 of the node's will be designated quorum vote. That node VM has to be up for the other node to come up and cluster to be formed.
    To check the current quorum status, run :-
    +# ./clquorum show+
    +//one of the nodes will have 1 vote and other 0(zero).+
    20)
    Now the cluster is in 'Installation Mode' and we need to add a quorum disk.
    Shutdown both the nodes as we will be adding shared disks to both of them
    21)
    Create 2 VirtualBox HDDs (VDI Files) on the host, 1 for quorum and 1 for shared filesystem. I have used a size of 1 GB for each :-
    *$ vboxmanage createhd --filename /scratch/myimages/sc41cluster/sdisk1.vdi --size 1024 --format VDI --variant Fixed*
    *0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100%*
    *Disk image created. UUID: 899147b9-d21f-4495-ad55-f9cf1ae46cc3*
    *$ vboxmanage createhd --filename /scratch/myimages/sc41cluster/sdisk2.vdi --size 1024 --format VDI --variant Fixed*
    *0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100%*
    *Disk image created. UUID: 899147b9-d22f-4495-ad55-f9cf15346caf*
    22)
    Attach these disks to both the VMs as shared type
    *$ vboxmanage storageattach solvm1 --storagectl "SATA" --port 1 --device 0 --type hdd --medium /scratch/myimages/sc41cluster/sdisk1.vdi --mtype shareable*
    *$ vboxmanage storageattach solvm1 --storagectl "SATA" --port 2 --device 0 --type hdd --medium /scratch/myimages/sc41cluster/sdisk2.vdi --mtype shareable*
    *$ vboxmanage storageattach solvm2 --storagectl "SATA" --port 1 --device 0 --type hdd --medium /scratch/myimages/sc41cluster/sdisk1.vdi --mtype shareable*
    *$ vboxmanage storageattach solvm2 --storagectl "SATA" --port 2 --device 0 --type hdd --medium /scratch/myimages/sc41cluster/sdisk2.vdi --mtype shareable*
    The disks are attached to SATA ports 1 & 2 of each VM. On my VirtualBox on Linux, the controller type is "SATA", whereas on Windows it is "SATA Controller".
    The "--mtype shareable' parameter is important
    23)
    Mark both disks as shared :-
    *$ vboxmanage modifyhd /scratch/myimages/sc41cluster/sdisk1.vdi --type shareable*
    *$ vboxmanage modifyhd /scratch/myimages/sc41cluster/sdisk2.vdi --type shareable*
    24) Start both VMs. We need to format the 2 shared disks
    25) From VM1, run format. In my case, the 2 new shared disks show up as 'c7t1d0' and 'c7t2d0'.
    +# format+
    select disk 1 (c7t1d0)
    [disk formated]
    FORMAT MENU
    fdisk
    Type 'y' to accept default partition
    partition
    0
    <enter>
    <enter>
    1
    995mb
    print
    label
    <yes>
    quit
    quit26) Repeat step 25) for the 2nd disk (c7t2d0)
    27) Make sure the shared disks can be used for quorum :-
    On VM1
    +# ./cldevice refresh+
    +# ./cldevice show+
    On VM2
    +# ./cldevice refresh+
    +# ./cldevice show+
    The shared disks should have the same DID (d2,d3,d4 etc). Note down the DID that you are going to use for quorum (e.g d2)
    By default, global fencing is enabled for these disks. We need to turn it off for all disks as these are SATA disks :-
    +# cldevice set -p default_fencing=nofencing-noscrub d1+
    +# cldevice set -p default_fencing=nofencing-noscrub d2+
    +# cldevice set -p default_fencing=nofencing-noscrub d3+
    +# cldevice set -p default_fencing=nofencing-noscrub d4+
    28) It is better to do one more reboot of both VMs, otherwise I got a error when adding the quorum disk
    29) Run clsetup to add quorum disk and to complete cluster configuration :-
    +# ./clsetup+
    === Initial Cluster Setup ===
    Enter 'Yes' for "Do you want to continue"
    Enter 'Yes' for "Do you want add any quorum devices"
    Select '1) Directly Attached Shared Disk' for the type of device
    Enter 'Yes' for "Is it okay to continue"
    Enter 'd2' (or 'd3') for 'Which global device do you want to use'
    Enter 'Yes' for "Is it okay to proceed with the update"
    The command 'clquorum add d2' is run
    Enter 'No' for "Do you want to add another quorum device"
    Enter 'Yes' for "Is it okay to reset "installmode"?"Cluster initialization is complete.!!!
    30) Run 'clquorum status' to confirm both nodes and the quorum disk have 1 vote each
    31) Run other cluster commands to explore!
    I will cover Data services and shared file system in another post. Basically the other shared disk
    can be used to create a UFS filesystem and mount it on all nodes.

    The Solaris Cluster 4.1 Installation and Concepts Guide are available at :-
    http://docs.oracle.com/cd/E29086_01/index.html
    Thanks.

  • Cloned two vm cluster nodes from development cluster to act as template to create production cluster

    Morning,
    There was so much done to setup the development cluster, I thought it would be easy to have the two nodes in the cluster cloned.  To my surprise, the development cluster was up and happily running on the new vm servers.  Stopping resources verifies
    it is stopping and starting resources on the original cluster.  I am not sure how to safely have the two new servers not manage the development cluster and create a new production cluster on them.  
    I am hesitant to destroy the cluster as I suspect it will destroy the real development cluster.  How do I do this?  How do I delete the windows cluster software and re-install without affecting the development cluster?
    Note that I tried to create a new cluster in the failover cluster manager and specify the new vm cluster servers, but it says they are part of a cluster already.  I do not see them listed as nodes.  I am not sure how to see what cluster it thinks
    the new servers are part of or how to not make them part of the development cluster.  That might be the path to my solution.

    This actually has worked out okay.  I found these steps and did them on both of the nodes that were claiming to be in a cluster already:
    powershell
    Import-Module FailoverClusters;
    clear-clusternode

  • Can we create secondary indexes in pooled and cluster table?

    hello all
    can we create secondary indexes in pooled and cluster table?

    Hi,
    Yes, you can.
    Refer the below links
    [http://www.sap-img.com/abap/the-different-types-of-sap-tables.htm|http://www.sap-img.com/abap/the-different-types-of-sap-tables.htm]
    [http://help.sap.com/saphelp_40b/helpdata/en/cf/21f083446011d189700000e8322d00/content.htm|http://help.sap.com/saphelp_40b/helpdata/en/cf/21f083446011d189700000e8322d00/content.htm]
    Regards,
    SB

  • Concurrent nodes reading from JMS topic (cluster environment)

    Hi.
    Need some help on this:
    Concurrent nodes reading from JMS topic (cluster environment)
    Thanks
    Denis

    After some thinking, I noted the following:
    1 - It's correct that only one node subscribes to a topic at a time. Otherwise, the same message would be processed by the nodes many times.
    2 - In order to solve the load balancing problem, I think that the Topic should be changed by a Queue. This way, each BPEL process from the node would poll for a message, and as soon as the message arrives, only one BPEL node gets the message and take if off the Queue.
    The legacy JMS provider I mentioned in the post above is actually the Retek Integration Bus (RIB). I'm integrating Retek apps with E-Business Suite.
    I'll try to configure the RIB to provide both a Topic (for the existing application consumers) and a Queue (an exclusive channel for BPEL)
    Do you guys have already tried to do an integration like this??
    Thanks
    Denis

  • Creating 4 Node SQL Cluster

    Hi
    I need to create a 4 Node - SQL 2012 Cluster (all in active Mode). 
    Node-1 will be having 1 Instance, Node-2 will be having 2 Instance, Node-3 will be running with 1 instance, and Node-4 will be running with 6 SQL Instance.
    Let me know whether can I install all the 10 Instance on Node-1 and then can i go ahead and add the  node-2 by selecting 10 Instances which is already installed and then node-3 and node-4.
    or Let me know if any other recommended way of Setting up the Cluster.
    Raj

    Hi,
    In the Server 2012 Failover cluster the node maximum is 64, the SQL server also have the maximum node limit, such as SQL 2008 clustering depends on which version of the software
    you purchase, along with which version of the operating system you intend to use. Therefore for the rigid plan I suggest you ask the issue in SQL forum.
    More information:
    SQL Server forum:
    http://social.technet.microsoft.com/Forums/sqlserver/en-us/home?category=sqlserver
    Before Installing Failover Clustering
    http://msdn.microsoft.com/en-us/library/ms189910.aspx
    Thanks for your understanding and support.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • OPMN fails in cluster environment after the machie reboots.

    Dear All:
    Have you ever meet the following problem or have any suggestion for it? Thanks a lot.
    There are machine A and B have been setup successfully in a cluster environment, everything is fine.
    However, after the machine B reboots, the Oracle BI Server and Presentation Server in machine B could not start.
    The error message for Oracle BI Server is:
    [2011-06-02T10:40:34.000+00:00] [OracleBIServerComponent] [NOTIFICATION:1] [] [] [ecid: ] [tid: 1060] Server start up failed: [nQSError: 43079] Oracle BI Server could not start because locating or importing newly published repositories failed.
    The error message for presentation server is:
    [2011-06-02T18:40:24.000+08:00] [OBIPS] [ERROR:1] [] [saw.sawserver.initializesawserver] [ecid: ] [tid: ] Another instance of Oracle Business Intelligence is trying to upgrade/update/create the catalog located at \\?\UNC\10.91.61.158\BIEEShare\CATALOG\SampleAppLite\root. Retry after it finishes.[
    The detail information is:
    1)     Machine A and B both run Windows 2008 server 64bit.
    2)     Machine A runs Admin Server+BI_Server1(managed server). Machine B is configured with “Scale out an existing installation” option and runs BI_Server2(managed server).
    3)     BIEE 11.1.1.3
    4)     In EM, the shared location of shared repository has been set to something like \\10.91.61.158\shared\RPD. The catalog path has been set to \\10.91.61.158\BIEEShare\CATALOG\SampleAppLite. Both path could be accessed by machine A and B.
    5)     Before machine B reboots, everything is fine. All the servers could be started.
    6)     After machine B reboots, the managed server BI_server2 could be started successfully. However, “opmnctl startall” would incur the above error message.
    Any suggestion is welcome. Thanks a lot!

    Hello,
    Thanks for answering.
    I am using OBIEE 11.1.1.5.0 enterprise software-only install scaleout existing Bi system option on HostMachine2. All services and coreapplication systems are running except coreapplication_obis1, coreapplication_obis2, coreapplication_obips1 and coreapplication_obips2 of bi_server2 on HostMachine2.
    Here is my ClusterConfig.xml, as you can see all changes in this file is done by em and primary and secondary controller settings are already applied. The same file also exist in the same location on HostMachine2.
    <?xml version="1.0" encoding="UTF-8" standalone="no"?>
    <Cluster xmlns="oracle.bi.cluster.services/config/v1.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="oracle.bi.cluster.services/config/v1.1 ClusterConfig.xsd">
    <ClusterProperties>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ClusterEnabled>true</ClusterEnabled>
    <ServerPollSeconds>5</ServerPollSeconds>
    <ControllerPollSeconds>5</ControllerPollSeconds>
    </ClusterProperties>
    <NodeList>
    <Node>
    <NodeType>PrimaryController</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obiccs1</NodeId>
    <!--HostNameOrIP can be a hostname, IP or virtual hostname-->
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9706</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9700</MonitorPort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ListenAddress>HostMachine1.localdomain.com</ListenAddress>
    </Node>
    <Node>
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MasterServer>true</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obis1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9703</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9701</MonitorPort>
    </Node>
    <Node>
    <NodeType>Scheduler</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obisch1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9705</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9708</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance1:coreapplication_obis2</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9702</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9709</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obis1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9761</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9762</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obis2</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9763</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9764</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>SecondaryController</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obiccs1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9765</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9766</MonitorPort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ListenAddress>HostMachine2</ListenAddress>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Scheduler</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obisch1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9770</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9771</MonitorPort>
    </Node>
    </NodeList>
    <SSLProperties>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><SSL>false</SSL>
    <SSLCertificateFile/>
    <SSLPrivateKeyFile/>
    <SSLCACertificateFile/>
    <SSLVerifyPeer>false</SSLVerifyPeer>
    </SSLProperties>
    </Cluster>

  • Run something once in a cluster environment

    I'm looking at using JBoss Cache to cache a large amount of data.  If you're unfamiliar with JC, it's a cachine system that will synchronize multiple remote caches in a clustered environment.  However, I don't want each cache to load itself at cluster startup, and then sync with all the other caches - there's a lot of data, and I don't want all the caches to become really chatty during startup for performance reasons.  I would rather have one server instance act as the master cache, and then have it sync all the other empty caches...
    I therefore need a way to determine what node in the cluster I'm running on.  How can I programatically do this?  I've looked through the API's, but it's like trying to find a needle in a haystack...

    Hi Ken,
    Ok, at least according t its documentation  the jboss cache will deal in its own with the failure detection, replication, new servers joining, which answers my curiosity in the upper memo
    I was speaking of servers later than NW 2004, but here is a suggestion how to deal with NW 2004:
    In order to achieve your goal - "prefetch everything only on one node", you could play with the locking feature. Here is a possible algorithm that could be put in code executed at application start (servlet init method, etc.):
    1) Check if the cache is empty, if it is not, you are secondary and someone has already updated you - exit
    2) Try to take a cluster unique logical lock in the locking service (here is the <a href="http://help.sap.com/saphelp_erp2005/helpdata/en/9a/4cdcc80fa4c747a2ccb5859f467412/frameset.htm">locking documentation</a>
    )If you cannot take the lock- you are the secondary server and someone is refreshing the cache at the moment, exit
    3) Update the cache
    4) Release the lock
    If there is no API to check if the cache is empty, you could potentially use some variable that is part of the cache and will be replicated to all servers. Before releasing the lock, you put some value "CacheUpdated" there.
    Another option would be to use some hardcoded logic and workarounds - do something only for specific server id as suggested in upper memo, do something if there is a specific property (System.getProperty()...) that you would put manually in only one server JVM startup options.
    Perhaps the simplest solution is to turn off all the cache prefetching - i.e. all nodes will start with empty cache and will load data on demand. You won't screw the performance of the "master" server as well.
    HTH
    Peter

  • ADS Configuration issue in HA cluster environment

    Hi All,
    We have HA cluster environment in which we have installed primary node, secondary node and additional application server. We have configured ADS in HA cluster but when we execute "fp_test_00" program in SE38 it throws "ADS: com.adobe.ProcessingException: com.adobe.Processin(200101)" exception. As well as RFC destination "FP_ICF_DATA_<SID>" in java system throws "Error during ping operation: Error while silently connecting: org.w3c.www.protocol.http.HttpException: Connection refused: connect"  while we ping destination.
    I have checked URL,Port and Logon data all information provided for RFC is fine. Please help me to resolve this issue.
    Thanks

    Hi Reagan
    As you suggested I have checked the URL(http://hostname:8000) has been used for connection, logon data is also fine. Below I have attached dev_icm trace file and there I can see that system throwing error while "reading data from message server"
    dev_icm
    Thr 14116] Mon May 19 01:11:29 2014
    [Thr 14116] {0000a598} Traffic Control: Nettimeout (30) exceeded by peer: 10.15.212.11:23327 [icxxthr.c 4303]
    [Thr 16228] Mon May 19 03:51:29 2014
    [Thr 16228] *** ERROR => MsINiRead: NiBufReceive failed (NIECONN_BROKEN) [msxxi.c      2829]
    [Thr 16228]
    [Thr 16228] *****************************************************************************
    [Thr 16228] *
    [Thr 16228] *  ERROR       Reading data from the message server (hostname / 3901)
    [Thr 16228] *              failed.
    [Thr 16228] *              Please check the trace file of the message server.
    [Thr 16228] *
    [Thr 16228] *  TIME        Mon May 19 03:51:29 2014
    [Thr 16228] *  RELEASE     721
    [Thr 16228] *  COMPONENT   MS (message handling interface, multithreaded)
    [Thr 16228] *  VERSION     4
    [Thr 16228] *  RC          -36
    [Thr 16228] *  MODULE      msxxi.c
    [Thr 16228] *  LINE        2848
    [Thr 16228] *  COUNTER     173
    [Thr 16228] *
    [Thr 16228] *****************************************************************************
    [Thr 16228]
    [Thr 16228] *** ERROR => MsIReadFromHdl: NiRead (rc=NIECONN_BROKEN) [msxxi.c      1867]
    [Thr 16228] JNCMIMsConnBrokenCallback: Message server connection broken
    [Thr 16228] IcmJncmConnBrokenCallback called
    [Thr 16228] JNCMReconnectAsync: Start Message Server reconnect, command MS_J2EE_RECO(2), allow 3600 sec for reconnect
    [Thr 16228] JNCMReconnectAsync: Allow 30 sec for consistent cluster reconnect
    [Thr 16228] Mon May 19 03:51:30 2014
    [Thr 16228]
    [Thr 16228] *****************************************************************************
    [Thr 16228] *
    [Thr 16228] *  ERROR       The connection to the specified message server
    [Thr 16228] *              (hostname / 3901) failed. Please check the trace file
    [Thr 16228] *              of the message server. If a network filter has
    [Thr 16228] *              been activated (see message server parameters
    [Thr 16228] *              ms/acl_file_...) please check the configuration
    [Thr 16228] *
    [Thr 16228] *  TIME        Mon May 19 03:51:30 2014
    [Thr 16228] *  RELEASE     721
    [Thr 16228] *  COMPONENT   MS (message handling interface, multithreaded)
    [Thr 16228] *  VERSION     4
    [Thr 16228] *  RC          -35
    [Thr 16228] *  MODULE      msxxi.c
    [Thr 16228] *  LINE        2783
    [Thr 16228] *  COUNTER     176
    [Thr 16228] *
    [Thr 16228] *****************************************************************************
    [Thr 16228]
    [Thr 16228] *** ERROR => MsIAttachEx: MsINiWrite (rc=-35) [msxxi.c      884]
    [Thr 16228] JNCMIMsConnBrokenCallback: Message server connection broken during reconnect
    [Thr 16228] JNCMReconnectAsync: cannot reconnect to message server [cmd=MS_J2EE_RECO(2),loop=1],  rc=-35(-35)
    dev_server0
    [Thr 10444] Mon May 19 03:51:30 2014
    M  [Thr 10444] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 6832] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 26/sock 4824
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 19208] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 42/sock 8104
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 4680] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 50/sock 368
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 10444] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 34/sock 7128
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 6832] Mon May 19 03:51:32 2014
    M  [Thr 6832] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 19208] Mon May 19 03:51:32 2014
    M  [Thr 19208] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 10444] Mon May 19 03:51:32 2014
    M  [Thr 10444] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 4680] Mon May 19 03:51:32 2014
    M  [Thr 4680] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 6832] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 35/sock 368
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 19208] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 27/sock 7592
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 10444] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 43/sock 8104
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 4680] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 51/sock 4824
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 6832] Mon May 19 03:51:35 2014
    M  [Thr 6832] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 10444] Mon May 19 03:51:35 2014
    M  [Thr 10444] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 10444] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 44/sock 7780
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 4680] Mon May 19 03:51:35 2014
    M  [Thr 4680] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 6832] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 36/sock 368
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 19208] Mon May 19 03:51:35 2014
    M  [Thr 19208] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 4680] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 52/sock 8104
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 19208] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 28/sock 7592
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 19208] Mon May 19 03:51:40 2014
    M  [Thr 19208] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 6832] Mon May 19 03:51:40 2014
    M  [Thr 6832] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 19208] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 53/sock 7128
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 6832] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 45/sock 7952
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 4680] Mon May 19 03:51:40 2014
    M  [Thr 4680] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 10444] Mon May 19 03:51:40 2014
    M  [Thr 10444] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 4680] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 29/sock 4824
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 10444] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 37/sock 8508
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]

    M [Thr 19208] Mon May 19 03:51:49 2014
    M  [Thr 19208] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 10444] Mon May 19 03:51:49 2014
    M  [Thr 10444] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 4680] Mon May 19 03:51:49 2014
    M  [Thr 4680] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]

    M [Thr 6832] Mon May 19 03:51:49 2014
    M  [Thr 6832] ***LOG Q0I=> NiPConnect2: IP-address:3301: connect (10061: WSAECONNREFUSED: Connection refused) [nixxi.cpp 3285]
    M  [Thr 19208] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 30/sock 4824
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 10444] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 54/sock 7592
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 4680] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 46/sock 7128
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    M  [Thr 6832] *** ERROR => NiPConnect2: SiPeekPendConn failed for hdl 38/sock 7952
        (SI_ECONN_REFUSE/10061; I4; ST; IP-address:3301) [nixxi.cpp    3285]
    Please see these logs and let me how I can resolve this issue.

  • TimesTen database in Sun Cluster environment

    Hi,
    Currently we have our application together with the TimesTen database installed at the customer on two different nodes (running on Sun Solaris 10). The second node acts as a backup to provide failover functionality, although right now only manual failover is supported.
    We are now looking into a hot-standby / high availability solution using Sun Cluster software. As understood from the documentation, applications can be 'plugged-in' to the Sun Cluster using Agents to monitor the application. Sun Cluster Agents should be already available for certain applications such as:
    # MySQL
    # Oracle 9i, 10g (HA and RAC)
    # Oracle 9iAS Application Server
    # PostgreSQL
    (See http://www.sun.com/software/solaris/cluster/faq.jsp#q_19)
    Our question is whether Sun Cluster Agents are already (freely) available for TimesTen? If so, where to find them. If not, should we write a specific Agent separately for TimesTen or handle database problems from the application.
    Does someone have any experience using TimesTen in a Sun Cluster environment?
    Thanks in advance!

    Yes, we use 2-way replication, but we don't use cache connect. The replication is created like this on both servers:
    create replication MYDB.REPSCHEME
    element SERVER01_DS datastore
    master MYDB on "SERVER01_REP"
    transmit nondurable
    subscriber MYDB on "SERVER02_REP"
    element SERVER02_DS datastore
    master MYDB on "SERVER02_REP"
    transmit nondurable
    subscriber MYDB on "SERVER01_REP"
    store MYDB on "SERVER01_REP"
    port 16004
    failthreshold 500
    store MYDB on "SERVER02_REP"
    port 16004
    failthreshold 500
    The application runs on SERVER01 and is standby on SERVER02. If an invalid state is detected in the application, the application on SERVER01 is stopped and the application on SERVER02 is started.
    In addition to this, we want to fail over if the database on the SERVER01 is in invalid state. What should we have monitored by the Clustering Agent to detect an invalid state in TT?

Maybe you are looking for

  • How to create a BI report overview page like ESS overview page.

    Hi Experts, I am implimenting FPN environment with BI portal.I have lot of BI reports in my system.now i would like to create a overview page to combine all BI reports in to one page Like our ESS overview page.can any body pleae suggest  me? Thanks ,

  • Custom tab order Acrobat Pro 8.0

    I just switched from Acrobat Pro 7.0 to 8.0 and I'm already lost. I have forms that have a hundred form fields on each page, and I need to set them to tab in a specific order. When I used Pro 7.0 it was a simple matter to skip any form field which di

  • How do you know the audio level of a keyframes

    I'm varying the audio levels of clips on the timeline by creating keyframes using the pen tool. I create four points: 1) -inf = beginning of fade-in 2) -5dB = end of fade-in 3) -5db = beginning of fade-out 4) -inf = end of fade-out But when I'm going

  • Scaling on WAD valuce axis

    Dear experts, can you please guide me how to apply scaling on value axis in WAD. Requirement is we need to have 0, 20000, 40000, 60000 .....  on Y-axis (value axis) in WAD. Thanking you in advance. Regards, Ravi Kumar VR Garre

  • WebLogic deployment on Unix or NT

    Hi, Does anybody have an idea of what is present percentage of Weblogic deployments on Unix and on NT. Curious to know what is most preferred platform for weblogic deployment. I appreciate your early response. thanks, Suresh.