Rapid clone 11.5.10  from single node to Veritas cluster environment

Is anybody there who had done the same and working? Please shed some light.
version 11.5.10
Source
DB tier : DB, CM and Admin ( OS : Solaris 10 ), single node
App tier : FrmSrv, iAS ( OS : Solaris 10), single node
Target
DB tier : DB, CM, and Admin ( OS : Solaris 10), 2-nodes Veritas Cluster
(active/passive)
App tier : FrmSrv, iAS (OS : Solaris 10), 2-nodes load balancer
I've many time single node to single node rapid clone , no problem.
But with this single node to multi-nodes, no luck.
your input is much appreciated

Hi
You can use the Advanced cloning (section 4) of the cloining document# 230672.1
to clone signle node to multi node systems.
for webserver load balancing follow the metalink note#:217368.1
Regards
Srinath

Similar Messages

  • How to migrate DB from single node 10gR2 to RAC 11gR2 on diff platform?

    How to migrate DB from single node 10gR2 to RAC 11gR2 on different platform with possible minimum downtime? We have a situation of upgrade/migrate oracle 10gR2 single instance DB to 2-node RAC 11gR. The source OS is Solaris 10 on SPARC and the target OS is Linux (the target servers could be changed to Solaris 11 x86 if needed). What is the best solution on that?
    Thanks,

    Technically, can do the following for upgrading and migration?
    1. Create 11gR2 oracle home on the same server and upgrade the database from 10gR2 to 11gR2 by running conversion (2 hour down time?)
    2. Set up Heterogeneous Primary and Physical Standbys by RMAN. The standby is the RAC with ASM. No need down time. (from Solaris Space to Linux - this may be a problem)
    3. At the cutoff time, activate the standby DB from the RAC ASM.
    If feasible, do we have some detail guild line for each step?

  • Migration from Single Node to RAC Node

    Hi,
    We are planning to migrate the Database from Single-Node to RAC Node. Are there any checklist list to be considered regarding Performance, System and Database ( or any other topic), before migrating to RAC.
    Thanks

    If it is already the same version, you will need to focus any "single-threadedness" of the application like sequences that are mandatory sequential, "semaphore-style" locking etc...
    My experience is that going RAC and in particular to ASM can give a performance boost. If it is the same version (11gR2 for example) then any statistics-related performance issues will still be there.

  • Migrating from single-node to multi-node (2 Node)

    Hi,
    We have decided to migrate our existing 11.5.9 application from single-node to 2 Node(multi-node). Please let me know what services should reside on each node (Recommended by Oracle).
    Forms, Reports, Apache on one node and CM, Admin, DB on another ?
    (or)
    DB on one node and all other on another node ?
    And please let me know the steps I should follow to achieve the task as well.
    Thank & Regards,
    Abdullah

    Hi;
    I suggest please review our previous posts and check notes:
    shared appltop/distributed appltop
    Re: two nodes
    In R12 shared appltop/distributed appltop
    shared appltop -Which node
    Re: shared appltop
    Shared appltop in R12
    Shared appltop in R12
    shared APPL TOP or shared application
    shared APPL TOP or shared application
    Regard
    Helios

  • Can we change the UCCX server from single node to the first node without affecting it's configurations

    Hi all,
    Firstly I was using the single CUCM7.0 server with the single UCCX 7.0 server. Now I am going to add the second CUCM and UCCX in the cluster. I was first configured the UCCX server as a single node. Now I have to configure this as a first node. How I could configure this from single node to the first node without affecting it's configurations..??????
    Regards
    Ali Raza

    Hi Aaron,
    How could we do the change on Imhost file. How we add the both servers in that file?
    Below is the Imhost file txt.
    # Copyright (c) 1993-1999 Microsoft Corp.
    # This is a sample LMHOSTS file used by the Microsoft TCP/IP for Windows.
    # This file contains the mappings of IP addresses to computernames
    # (NetBIOS) names.  Each entry should be kept on an individual line.
    # The IP address should be placed in the first column followed by the
    # corresponding computername. The address and the computername
    # should be separated by at least one space or tab. The "#" character
    # is generally used to denote the start of a comment (see the exceptions
    # below).
    # This file is compatible with Microsoft LAN Manager 2.x TCP/IP lmhosts
    # files and offers the following extensions:
    #      #PRE
    #      #DOM:
    #      #INCLUDE
    #      #BEGIN_ALTERNATE
    #      #END_ALTERNATE
    #      \0xnn (non-printing character support)
    # Following any entry in the file with the characters "#PRE" will cause
    # the entry to be preloaded into the name cache. By default, entries are
    # not preloaded, but are parsed only after dynamic name resolution fails.
    # Following an entry with the "#DOM:" tag will associate the
    # entry with the domain specified by . This affects how the
    # browser and logon services behave in TCP/IP environments. To preload
    # the host name associated with #DOM entry, it is necessary to also add a
    # #PRE to the line. The is always preloaded although it will not
    # be shown when the name cache is viewed.
    # Specifying "#INCLUDE " will force the RFC NetBIOS (NBT)
    # software to seek the specified and parse it as if it were
    # local. is generally a UNC-based name, allowing a
    # centralized lmhosts file to be maintained on a server.
    # It is ALWAYS necessary to provide a mapping for the IP address of the
    # server prior to the #INCLUDE. This mapping must use the #PRE directive.
    # In addtion the share "public" in the example below must be in the
    # LanManServer list of "NullSessionShares" in order for client machines to
    # be able to read the lmhosts file successfully. This key is under
    # \machine\system\currentcontrolset\services\lanmanserver\parameters\nullsessionshares
    # in the registry. Simply add "public" to the list found there.
    # The #BEGIN_ and #END_ALTERNATE keywords allow multiple #INCLUDE
    # statements to be grouped together. Any single successful include
    # will cause the group to succeed.
    # Finally, non-printing characters can be embedded in mappings by
    # first surrounding the NetBIOS name in quotations, then using the
    # \0xnn notation to specify a hex value for a non-printing character.
    # The following example illustrates all of these extensions:
    # 102.54.94.97     rhino         #PRE #DOM:networking  #net group's DC
    # 102.54.94.102    "appname  \0x14"                    #special app server
    # 102.54.94.123    popular            #PRE             #source server
    # 102.54.94.117    localsrv           #PRE             #needed for the include
    # #BEGIN_ALTERNATE
    # #INCLUDE \\localsrv\public\lmhosts
    # #INCLUDE \\rhino\public\lmhosts
    # #END_ALTERNATE
    # In the above example, the "appname" server contains a special
    # character in its name, the "popular" and "localsrv" server names are
    # preloaded, and the "rhino" server name is specified so it can be used
    # to later #INCLUDE a centrally maintained lmhosts file if the "localsrv"
    # system is unavailable.
    # Note that the whole file is parsed including comments on each lookup,
    # so keeping the number of comments to a minimum will improve performance.
    # Therefore it is not advisable to simply add lmhosts file entries onto the
    # end of this file.
    Regards
    Ali Raza

  • Create secondary node on BOBJ cluster environment

    Hi All,
    I have done migration from old BOBJ server to new BOBJ server. The primary nodes has successfully completed and manage to view reports. I currently having issue to proceed installing second node on BOBJ cluster environment. i stuck at phase repository database as no database entry was found.
    Please help. I also attach my steps to install the second node.
    Thanks.
    Regards
    Aiman

    Hi,
    im guessing you do an Expand installation on the second node and want to enter the credentials of your existing CMS DB, the one you configured during the installation of Node 1. Correct?
    What is your CMS DB of a Kind? MS SQL? Oracle?
    It looks like that the ODBC DSN entries are missing on Node 2.
    Regards
    -Seb.

  • How to change from single Node to RAC for AIA

    Recently I have installed AIA 11.1.1.5, but during installation I selected single node. But now I would like to configure it to use RAC. How to do it? Reinstalling is not an option.

    Hi,
    Hope this helps
    http://proskudin.blogspot.com/2007/06/two-level-tabs.html
    Br,Jari

  • Clustered enviroments -Starting/Stopping Managed servers, NodeManager & Admin server from single node

    Hi,
    We have 5 node Weblogic cluster. At the time of some mass activities like patching we have to stop all our managed, Admin Servers and node manager and it takes a lot of time to do this by logging in to all 5 managed servers and stopping all components. Though we have node manager configured so we can stop managed servers from console itself, but still have to login to each server to stop NodeManager and Oracle Http Servers.
    Things become worse when wee have to shutdown almost 15-20 instances.
    Is there a way (using wlst or shell script) that we can do all these tasks from by logging in to just one server?
    Thanks,
    Suraj

    The simple code below would shutdown all the running Servers. Some of the APIs for ex MBeanHome are deprecated..still work though..if your concerned about that..I am working on the same using JMX API..if you want I could share that too. And also as described in the below link..you can Cluster stop and start using WLST commands.  Managing the Server Life Cycle - 12c Release 1 (12.1.1)
    <This code is from Oracle Weblogic documentation>
    import java.util.Set;
    import java.util.Iterator;
    import java.rmi.RemoteException;
    import javax.naming.Context;
    import javax.management.ObjectName;
    import java.io.*;
    import java.net.*;
    import weblogic.jndi.Environment;
    import weblogic.management.MBeanHome;
    import weblogic.management.WebLogicMBean;
    import weblogic.management.configuration.ServerMBean;
    import weblogic.management.runtime.ServerRuntimeMBean;
    import weblogic.management.runtime.ServerStates;
    import weblogic.management.WebLogicObjectName;
    public class ServerStopper {
      public static void stop() throws Exception {
        MBeanHome home = null;
        //url of the Admin server
        String url = "t3://localhost:7001";
        String username = "weblogic";
        String password = "welcome1";
        ServerRuntimeMBean serverRuntime = null;
        Set mbeanSet = null;
        Iterator mbeanIterator = null;
        try {
          // Set ContextClassloader to prevent assertions
          URL[] urls = { new File("/").toURL() };
          Thread.currentThread().setContextClassLoader(new
             URLClassLoader(urls));
          Environment env = new Environment();
          env.setProviderUrl(url);
          env.setSecurityPrincipal(username);
          env.setSecurityCredentials(password);
          Context ctx = env.getInitialContext();
          home = (MBeanHome)
               ctx.lookup("weblogic.management.adminhome");
          mbeanSet = home.getMBeansByType("ServerRuntime");
          mbeanIterator = mbeanSet.iterator();
          while(mbeanIterator.hasNext()) {
            serverRuntime = (ServerRuntimeMBean)mbeanIterator.next();
            if(serverRuntime.getState().equals(ServerStates.RUNNING)){
                serverRuntime.shutdown();
        } catch (Exception e) {
          e.printStackTrace();

  • Long running servlet in a single node of a cluster

    Good evening,
    I am developing a J2EE application that consists of a number of web services that perform background processing of relatively long-running jobs.  Status of the jobs actually need to be reported back to the SAP ABAP system, which we do via JCo.  All communication back to ABAP occurs in a single long-running, JMS driven servlet.  This callback servlet uses both a timestamp (last time status was sent back) or a  count of status events to determine when to send data to the ABAP system.
    All of this works great in a single Java instance J2EE configuration.  However, when we introduce clustering into the configuration, we get one instance of the callback servlet per J2EE server process.  Is there an easy way to configure the servlet so that it only runs on a single J2EE server process in one instance?
    TIA,
    - Bill

    Hi Bill,
    Launching manually threads even from a web container is also not recommended as good J2EE practice.
    If each applications launches its own set of threads that may easily crash the server. Another drawback of such long running servlet is that you are blocking application thread from the server. One last point - it seems that you are storing the data in the servlet at some intermediate variables without any persistance. What will happen if the server crashes ? Can you afford to loose the data ?
    I am not implying that you launch threads from the MDB. I am just saying that you could define that single MDB and send messages directly to it. You do the job there inside the onMessage, no new thread. You will be guaranteed that no other call will be executed. You are guaranteed nothing will be lost in case server crashes, you will not use resources even if there is no traffic.
    Btw, please feel free to give me a phone call or write an email to further discuss the issue. You can take the details from your CSN I am currently processing
    Best Regards
    Peter

  • File Adapter not writing file when processed from one node in a clustered environment

    In our environment, we have 4 server nodes running. We are using a file adapter to write file to a location in the server. the adapter is writing files when its processing data in all nodes except one.
    Below is the JCA File
    <connection-factory location="eis/FileAdapter" adapterRef=""/>
    <endpoint-interaction portType="Write_ptt" operation="Write">
    <interaction-spec className="oracle.tip.adapter.file.outbound.FileInteractionSpec">
    <property name="PhysicalDirectory" value="/data/prd"/>
    <property name="FileNamingConvention" value="Sample.dat"/>
    <property name="Append" value="true"/>
    </interaction-spec>
    </endpoint-interaction>
    the File Adapter health is running OK. What would be the potential cause

    In our environment, we have 4 server nodes running. We are using a file adapter to write file to a location in the server. the adapter is writing files when its processing data in all nodes except one.
    Below is the JCA File
    <connection-factory location="eis/FileAdapter" adapterRef=""/>
    <endpoint-interaction portType="Write_ptt" operation="Write">
    <interaction-spec className="oracle.tip.adapter.file.outbound.FileInteractionSpec">
    <property name="PhysicalDirectory" value="/data/prd"/>
    <property name="FileNamingConvention" value="Sample.dat"/>
    <property name="Append" value="true"/>
    </interaction-spec>
    </endpoint-interaction>
    the File Adapter health is running OK. What would be the potential cause

  • Configuring Scheduler only on single node in a clustered environment

    Friends,
    I have OIM 11gr2 environment in clustered mode with 6 nodes in it.
    As per my client requirement, I need to configure scheduler only on one node and No on all other nodes. So, all schedule jobs OOTB and custom jobs should run only on one node.
    What is the process for this?
    Thanks,
    MM

    Check this: Managing the Scheduler - 11g Release 2 (11.1.2.1.0)
    -Bikash

  • Migrate Single Node 10g STreams to Multi Node RAC

    We are currently running 10g STreams. One capture process enqueing LCR into one queue. Propagation job dequeues messages and propagates them to a Destination queue. Then our Apply process at Destination database dequeues and apply changes.
    We have developed a single shell script STREAMSsetup.sh that does all the magic on setting up streams (capture, apply, propagator and the table queues.)
    Now, we are in the process of designing our new Horizontal architecture with Oracle RAC. We envision it is going to be a Data Guard Migration from Single node to the RAC environment.
    My question is, can we just let Data Guard takes care of that, expecting all Streams pieces will be in place within the RAC environment? OR do you recomment we remove all streams configuration and re-run our magic script STREAMSsetup.sh to set everything up again once migrated to the RAC environment. ???
    I now Streams follows an Instance Owner rule when it comes to where Capture or Apply will be started or stoped.
    I would appreciate if someone have had any experience with this type of migration provide any feedback.
    Thanks,
    Arqui

    Yes, we are opening an SR. Everything is pointing into the direction of recreating Streams configuration from RAC/ASM environment once migrated, using the instance with the lowest Instance number in the Cluster. That instance will be registered as the Primary instance owner of the Buffer queue. If primary instance fails queue ownership is transferred automatically to the next instance in the cluster and capture is started automatically in the other node.
    Thanks,
    ARqui

  • How to switch production single node EBS environment to multi node?

    We're considering moving our production EBS environment from single node to multi node to separate out the middle processing from the database. We've completed the steps via clone (adcfgclone.pl) in out test environment and are now testing, however we're leery of running the same process in production. I've researched this process greatly, but most links point to either multi to single or cloning. Please assist.

    Use Rapid Clone -- Rapid Clone Documentation Resources For Release 11i and 12 [ID 799735.1]
    All the steps are covered under "Cloning a Single-Node System to a Multi-Node System" section.
    Thanks,
    Hussein

  • Convert single node Multi Node

    We have one requirement, our single node is running on sun server ( 11.5.10.2 with 9.2.0.7)
    we want to convert the single node to 2 node, like this
    Node 1 = DB + CM + ADMIN
    Node 2 = FORMS, WEB
    I want to follow the following procedure
    (a) Copy the Total Application Tier File system from single node to Node 2
    (b) Run the config clone on Node 1 and specify to run the services CM and ADMIN only on this node
    (c) Run the config clone on Node 2 and specify to run Forms, web services only on this node 2
    is the above procedure correct?

    Node 1 = DB + CM + ADMIN
    Node 2 = FORMS, WEB
    I want to follow the following procedure
    (a) Copy the Total Application Tier File system from single node to Node 2
    (b) Run the config clone on Node 1 and specify to run the services CM and ADMIN only on this node
    (c) Run the config clone on Node 2 and specify to run Forms, web services only on this node 2It is correct.

  • BizTalk Mapper - Looping multiple nodes to map to a single node in a single row (flat file)

    Hi everybody,
    I'm still new in developing BizTalk app and require some help in this one problem. Appreciate your time and input to help me on this.
    Basically I have an XML document as input and a flat file as output. Example for input is as per below. The "Contact" node's maxOccurs here is set to unbounded and could be multiple. (phone, fax, website, telex ...)
    <root>
    <CustomerName>Company A</CustomerName>
    <Contact>
    <Type>Phone</Type>
    <Locator>03566789</Locator>
    <Type>Phone</Type>
    <Locator>03566790</Locator>
    <Type>Fax</Type>
    <Locator>03566795</Locator>
    <Type>Telex</Type>
    <Locator>03566798</Locator>
    <Type>Website</Type>
    <Locator>www.companyA.com</Locator>
    </Contact>
    </root>
    The expected output in XML would look like below. The final outcome would be a csv file. Strictly Phone 1, phone 2, fax and telex, the rest would be ignored.
    <root>
    <CustomerName>Company A</CustomerName>
    <Phone1>03566789</Phone1>
    <Phone2>03566790</Phone2>
    <Fax>03566795</Fax>
    <Telex>03566798</Telex>
    </root>
    Example of expected output result (csv file): CompanyName;Phone1;Phone2;Fax;Telex;
    In our case here: Company A;03566789;03566790;03566795;03566798;
    Another example could be: Company B;036778911;;036778912;; if only 1 phone number and 1 fax number provided.
    I've used Table Looping and Table Extractor and nearly got the desired result except that it is represented in multiple rows instead of one: Example:
    Company A;03566789;;;
    Company A;;03566790;;;
    Company A;;;03566795;;
    Company A;;;;03566798;
    Any idea how to do the mapping? I'm kind of stuck here and it sounds like an easy problem but i could not find any example to the solution that I need here. Table looping and table extractor is ok to map from single node flat file to multiple nodes but not
    the reverse like in this example.
    rgds,
    sportivo

    Hi,
    Please refer to below links where similar issue has been answered.
    http://social.msdn.microsoft.com/Forums/en-US/biztalkgeneral/thread/ecdff241-6795-4a95-bad7-48fca4410dfb
    http://www.epinaki.com/2011/05/other-options-to-using-biztalk-table-looping-functoid-par-i/
    I hope this helps you.
    Thanks With Regards,
    Shailesh Kawade
    MCTS BizTalk Server
    Please Mark This As Answer If This Helps You.
    http://shaileshbiztalk.blogspot.com/

Maybe you are looking for