Configuring Scheduler only on single node in a clustered environment

Friends,
I have OIM 11gr2 environment in clustered mode with 6 nodes in it.
As per my client requirement, I need to configure scheduler only on one node and No on all other nodes. So, all schedule jobs OOTB and custom jobs should run only on one node.
What is the process for this?
Thanks,
MM

Check this: Managing the Scheduler - 11g Release 2 (11.1.2.1.0)
-Bikash

Similar Messages

  • Get a list of users logged on each server node in a clustered environment

    Hi,
    does anybody know if there is way to get a list of users which are logged on each server node in a clustered environment. Or maybe is there api for that so I could write an application which does that?
    Regards
    Ladislav

    Hi,
    about the code I was looking at - easily you can find out that these iviews are components of <b>com.sap.portal.admin.monitor.par</b>. The thread Thread Overview is implemented by: <i>com.sapportals.portal.admin.psm.PortalServerMonitor</i>. In <i>getThreadsData</i> method there is a parameter <b>portalMonitor</b> so it led me to this code in <i>doContent</i>:
    /* 143*/          IPortalMonitor portalMonitor = (IPortalMonitor)PortalRuntime.getRuntimeResources().getService("com.sap.portal.runtime.application.monitor.PortalMonitor");
    /* 144*/          IPSMData psmData = portalMonitor.getIPSMDataInstance();
    So I have found <i>com.sap.portal.runtime.application.monitor.par</i> and there is a service implemented by <i>com.sapportals.portal.prt.service.monitor.PortalMonitor</i> where is <i>com.sapportals.portal.prt.service.monitor.MonitorCommunication</i>. And this is a service which calls all the nodes in the cluster around asking for info about the threads, requests, ... Here is a method <i>getMonitoringData</i> which works with object <i>com.sapportals.portal.prt.service.monitor.MonitorData</i>. Here in method <i>getThreadOverview</i>:
    /* 396*/          ApplThreadOverview oATOs[] = OverviewMonitor.getApplThreads();
    So in this <i>OverviewMonitor</i> the method <i>getApplThreads</i>:
                 public static ApplThreadOverview[] getApplThreads()
    /*  50*/          return ApplThreadMonitor.getApplThreads(false);
    from here:
    /* 130*/          ApplThreadMonitor hlpAppl = TaskMonitor.getApplThreadList();
    and this <i>ApplThreadMonitor</i> has:
                 private String userName;
                 private String reqName;
                 private String taskName;
                 private String compName;
                 private String action;
    So if you can get this (when the most of the classes mentioned is not in API but CORE part) on each node and ask for it from one place you got the user list at cluster nodes...
    In fact there can be a shortcut to get this info, I have found this ;o)
    Romano
    PS: and thanks for the stars!

  • Rapid clone 11.5.10  from single node to Veritas cluster environment

    Is anybody there who had done the same and working? Please shed some light.
    version 11.5.10
    Source
    DB tier : DB, CM and Admin ( OS : Solaris 10 ), single node
    App tier : FrmSrv, iAS ( OS : Solaris 10), single node
    Target
    DB tier : DB, CM, and Admin ( OS : Solaris 10), 2-nodes Veritas Cluster
    (active/passive)
    App tier : FrmSrv, iAS (OS : Solaris 10), 2-nodes load balancer
    I've many time single node to single node rapid clone , no problem.
    But with this single node to multi-nodes, no luck.
    your input is much appreciated

    Hi
    You can use the Advanced cloning (section 4) of the cloining document# 230672.1
    to clone signle node to multi node systems.
    for webserver load balancing follow the metalink note#:217368.1
    Regards
    Srinath

  • Configuring Mobile Access with single public ip in test environment

    Hi to all, I'm trying to configure a test lab environment for Lync 2013. Actually I installed DC+FE in 2 separate servers and domain and off-domain pc (with exported CA chain ad certificate) works perfectly. My goal is now to configure access for mobiles
    (iphone\ipad). In my country (Italy) Lync is not really diffused like in USA. Now my question. I've a lone public ip for testing, is it possible to make it works? What have i to install now? Edge? Reverse Proxy? Both?? I wanna have less resource possibile
    implicated in this lab environment also if this means "not supported".
    Thank you for the support.
    A.

    Edge server is needed when you need to support audio/video modality and push notification.
    You can deploy Lync mobility service without Edge Server in a lab environment.
    Lisa Zheng
    TechNet Community Support

  • File Adapter not writing file when processed from one node in a clustered environment

    In our environment, we have 4 server nodes running. We are using a file adapter to write file to a location in the server. the adapter is writing files when its processing data in all nodes except one.
    Below is the JCA File
    <connection-factory location="eis/FileAdapter" adapterRef=""/>
    <endpoint-interaction portType="Write_ptt" operation="Write">
    <interaction-spec className="oracle.tip.adapter.file.outbound.FileInteractionSpec">
    <property name="PhysicalDirectory" value="/data/prd"/>
    <property name="FileNamingConvention" value="Sample.dat"/>
    <property name="Append" value="true"/>
    </interaction-spec>
    </endpoint-interaction>
    the File Adapter health is running OK. What would be the potential cause

    In our environment, we have 4 server nodes running. We are using a file adapter to write file to a location in the server. the adapter is writing files when its processing data in all nodes except one.
    Below is the JCA File
    <connection-factory location="eis/FileAdapter" adapterRef=""/>
    <endpoint-interaction portType="Write_ptt" operation="Write">
    <interaction-spec className="oracle.tip.adapter.file.outbound.FileInteractionSpec">
    <property name="PhysicalDirectory" value="/data/prd"/>
    <property name="FileNamingConvention" value="Sample.dat"/>
    <property name="Append" value="true"/>
    </interaction-spec>
    </endpoint-interaction>
    the File Adapter health is running OK. What would be the potential cause

  • Can ASM be used in single node 11g?

    I tried to create DB in 11g using ASM, but I was asked to create oracle CSS services (which is for RAC). of course it failed. My platform is oracle linux 5. As I remember 10g can use ASM (in Solaris), any one has idea about 11g? how to use ASM (not RAC)? (I already used ASM in RAC 11g)
    Thank you in advance.

    Although this is for Oracle 10gR2, here's a web page that might shed some light on your situation: Note that is says CSS is created automatically even if you don't use it.
    http://download-west.oracle.com/docs/html/B10130_02/deinstall.htm#BCFBEEAJ
    6.2 Removing Oracle Cluster Synchronization Services (CSS)
    The first time you install Oracle Database, Oracle Universal Installer configures and starts a single-node version of the Oracle Cluster Synchronization Services (CSS) service. The CSS service enables synchronization between an Automatic Storage Management (ASM) instance and the database instances that rely on it for database file storage. It is configured and started even if you do not choose Automatic Storage Management as a storage mechanism for database files.
    If you do not choose Automatic Storage Management as a storage option, you can delete OracleCSService. To delete this service without deleting the Oracle home, perform the following:

  • Migrate Single Node 10g STreams to Multi Node RAC

    We are currently running 10g STreams. One capture process enqueing LCR into one queue. Propagation job dequeues messages and propagates them to a Destination queue. Then our Apply process at Destination database dequeues and apply changes.
    We have developed a single shell script STREAMSsetup.sh that does all the magic on setting up streams (capture, apply, propagator and the table queues.)
    Now, we are in the process of designing our new Horizontal architecture with Oracle RAC. We envision it is going to be a Data Guard Migration from Single node to the RAC environment.
    My question is, can we just let Data Guard takes care of that, expecting all Streams pieces will be in place within the RAC environment? OR do you recomment we remove all streams configuration and re-run our magic script STREAMSsetup.sh to set everything up again once migrated to the RAC environment. ???
    I now Streams follows an Instance Owner rule when it comes to where Capture or Apply will be started or stoped.
    I would appreciate if someone have had any experience with this type of migration provide any feedback.
    Thanks,
    Arqui

    Yes, we are opening an SR. Everything is pointing into the direction of recreating Streams configuration from RAC/ASM environment once migrated, using the instance with the lowest Instance number in the Cluster. That instance will be registered as the Primary instance owner of the Buffer queue. If primary instance fails queue ownership is transferred automatically to the next instance in the cluster and capture is started automatically in the other node.
    Thanks,
    ARqui

  • Invocation service only processing on a single node (Coherence 3.4).

    I'm trying to run a task on each node using the invocation service. The task should return a list of report files from each node.
    The service is setup as follows:<b>
        <invocation-scheme>
            <scheme-name>agents</scheme-name>
            <service-name>agents</service-name>
            <serializer>com.tangosol.io.DefaultSerializer</serializer>
        </invocation-scheme></b>
    And accessed as follows, using the deprecated getInvocationService method:<b>
        InvocationService service = CacheFactory.getInvocationService("agents");
        Map<Member, List<String>> reportQuery = service.query(new ReportFilesnameAgent(), null);</b>
    When I access it in this way, the ReportFilesnameAgent only runs on a single node, despite the following listing both cluster nodes:<b>
        Set memberSet = service.getCluster().getMemberSet();</b>
    I even tried to force it to run on all nodes by passing the member set to the query method, but this made no difference:<b>
        Map<Member, List<String>> reportQuery = service.query(new ReportFilesnameAgent(), memberSet);</b>
    If I access the service using the recommended (non-deprecated) method:<b>
        InvocationService service = (InvocationService) CacheFactory.getService("agents");</b>
    Coherences throws the following:<b>
        Caused by: java.lang.IllegalArgumentException: Missing scheme for service: "agents"</b>
    Any ideas?
    Many thanks in advance,
    Cormac.
    Edited by: user1744133 on Dec 21, 2009 4:15 AM

    Hi,
    I'm trying this invocation-schema with Version 3.5.2/463a and a pof serializer and it fails with the below stack trace suggesting that my nodes run of different configurations.
    I can see them loading the same cache-config and the same config though.
    Any ideas?
    many thanks,
    Christoph.
    The new invocation conig:
         <invocation-scheme>
              <scheme-name>agents</scheme-name>
              <service-name>agents</service-name>
              <serializer>
                   <class-name>
                        com.tangosol.io.pof.SafeConfigurablePofContext
              </class-name>
                   <init-params>
                        <init-param>
                             <param-type>string</param-type>
                             <param-value>
                                  orion-pof-config.xml
              </param-value>
                        </init-param>
                   </init-params>
              </serializer>          
              <autostart>true</autostart>
         </invocation-scheme>
    the stackt trace:
    2010-02-15 15:37:37.286/5.469 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:agents, member=2): Service agents joined the cluster with senior service member 1
    2010-02-15 15:37:37.333/5.516 Oracle Coherence GE 3.5.2/463 <Error> (thread=Invocation:agents, member=2): The service "agents" is configured to use serializer com.tangosol.io.DefaultSerializer {loader=sun.misc.Launcher$AppClassLoader@17590db}, which appears to be different from the serializer used by Member(Id=1, Timestamp=2010-02-15 15:37:20.268, Address=165.2.93.118:18001, MachineId=64118, Location=site:EMEA.AD.JPMORGANCHASE.COM,machine:LLDNCFI5SZW83J,process:2360, Role=TmpMain).
    java.io.StreamCorruptedException: invalid type: 78
         at com.tangosol.util.ExternalizableHelper.readObjectInternal(ExternalizableHelper.java:2231)
         at com.tangosol.util.ExternalizableHelper.readObject(ExternalizableHelper.java:2219)
         at com.tangosol.io.DefaultSerializer.deserialize(DefaultSerializer.java:60)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.Service.readObject(Service.CDB:4)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid$ServiceConfigMap.readObject(Grid.CDB:1)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid$MemberConfigResponse.read(Grid.CDB:13)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.onNotify(Grid.CDB:123)
         at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
         at java.lang.Thread.run(Thread.java:619)
    Stopping the agents service.
    2010-02-15 15:37:37.333/5.516 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:agents, member=2): Service agents left the cluster
    2010-02-15 15:37:37.333/5.516 Oracle Coherence GE 3.5.2/463 <Error> (thread=main, member=2): Error while starting service "agents": java.lang.RuntimeException: Failed to start Service "agents" (ServiceState=SERVICE_STOPPED)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.Service.start(Service.CDB:38)
         at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Grid.start(Grid.CDB:38)
         at com.tangosol.coherence.component.util.SafeService.startService(SafeService.CDB:28)
         at com.tangosol.coherence.component.util.SafeService.ensureRunningService(SafeService.CDB:27)
         at com.tangosol.coherence.component.util.SafeService.start(SafeService.CDB:14)
         at com.tangosol.net.CacheFactory.getInvocationService(CacheFactory.java:943)
         at com.tmp.InvocationClient.main(InvocationClient.java:37)

  • Can we change the UCCX server from single node to the first node without affecting it's configurations

    Hi all,
    Firstly I was using the single CUCM7.0 server with the single UCCX 7.0 server. Now I am going to add the second CUCM and UCCX in the cluster. I was first configured the UCCX server as a single node. Now I have to configure this as a first node. How I could configure this from single node to the first node without affecting it's configurations..??????
    Regards
    Ali Raza

    Hi Aaron,
    How could we do the change on Imhost file. How we add the both servers in that file?
    Below is the Imhost file txt.
    # Copyright (c) 1993-1999 Microsoft Corp.
    # This is a sample LMHOSTS file used by the Microsoft TCP/IP for Windows.
    # This file contains the mappings of IP addresses to computernames
    # (NetBIOS) names.  Each entry should be kept on an individual line.
    # The IP address should be placed in the first column followed by the
    # corresponding computername. The address and the computername
    # should be separated by at least one space or tab. The "#" character
    # is generally used to denote the start of a comment (see the exceptions
    # below).
    # This file is compatible with Microsoft LAN Manager 2.x TCP/IP lmhosts
    # files and offers the following extensions:
    #      #PRE
    #      #DOM:
    #      #INCLUDE
    #      #BEGIN_ALTERNATE
    #      #END_ALTERNATE
    #      \0xnn (non-printing character support)
    # Following any entry in the file with the characters "#PRE" will cause
    # the entry to be preloaded into the name cache. By default, entries are
    # not preloaded, but are parsed only after dynamic name resolution fails.
    # Following an entry with the "#DOM:" tag will associate the
    # entry with the domain specified by . This affects how the
    # browser and logon services behave in TCP/IP environments. To preload
    # the host name associated with #DOM entry, it is necessary to also add a
    # #PRE to the line. The is always preloaded although it will not
    # be shown when the name cache is viewed.
    # Specifying "#INCLUDE " will force the RFC NetBIOS (NBT)
    # software to seek the specified and parse it as if it were
    # local. is generally a UNC-based name, allowing a
    # centralized lmhosts file to be maintained on a server.
    # It is ALWAYS necessary to provide a mapping for the IP address of the
    # server prior to the #INCLUDE. This mapping must use the #PRE directive.
    # In addtion the share "public" in the example below must be in the
    # LanManServer list of "NullSessionShares" in order for client machines to
    # be able to read the lmhosts file successfully. This key is under
    # \machine\system\currentcontrolset\services\lanmanserver\parameters\nullsessionshares
    # in the registry. Simply add "public" to the list found there.
    # The #BEGIN_ and #END_ALTERNATE keywords allow multiple #INCLUDE
    # statements to be grouped together. Any single successful include
    # will cause the group to succeed.
    # Finally, non-printing characters can be embedded in mappings by
    # first surrounding the NetBIOS name in quotations, then using the
    # \0xnn notation to specify a hex value for a non-printing character.
    # The following example illustrates all of these extensions:
    # 102.54.94.97     rhino         #PRE #DOM:networking  #net group's DC
    # 102.54.94.102    "appname  \0x14"                    #special app server
    # 102.54.94.123    popular            #PRE             #source server
    # 102.54.94.117    localsrv           #PRE             #needed for the include
    # #BEGIN_ALTERNATE
    # #INCLUDE \\localsrv\public\lmhosts
    # #INCLUDE \\rhino\public\lmhosts
    # #END_ALTERNATE
    # In the above example, the "appname" server contains a special
    # character in its name, the "popular" and "localsrv" server names are
    # preloaded, and the "rhino" server name is specified so it can be used
    # to later #INCLUDE a centrally maintained lmhosts file if the "localsrv"
    # system is unavailable.
    # Note that the whole file is parsed including comments on each lookup,
    # so keeping the number of comments to a minimum will improve performance.
    # Therefore it is not advisable to simply add lmhosts file entries onto the
    # end of this file.
    Regards
    Ali Raza

  • R12 Single node to Multi-node Configuration

    Hi All!
    We have a single node setup (Applications and Database are on the same server). Now, our management decided to setup a Multi-node configuration and i'm the one who will do this.
    The plan is to use Oracle ASM and RAC for the database and for the Application is to have a multiple application servers. For the database, we will be using a minimum of 3 servers and another 3 servers for the application. Is this possible? How can this be done. Can someone point or give some link or infos on how to do this. We're going to use the existing data on the single node server.
    Thanks and Regards,
    cmadiam82

    Hi,
    We have a single node setup (Applications and Database are on the same server). Now, our management decided to setup a Multi-node configuration and i'm the one who will do this.Contratulations. This is a good oppurtunity.
    The plan is to use Oracle ASM and RAC for the database and for the Application is to have a multiple application servers. For the database, we will be using a minimum of 3 servers and another 3 servers for the application. Is this possible? How can this be done. Can someone point or give some link or infos on how to do this. We're going to use the existing data on the single node server.For a start you can read
    http://gjilevski.wordpress.com/2009/12/13/rac-enabling-ebs-12-1-1/
    http://www.appsdba.info/index.php?module=pagemaster&PAGE_user_op=view_page&PAGE_id=52&MMN_position=44:44
    http://www.appsdba.info/index.php?module=pagemaster&PAGE_user_op=view_page&PAGE_id=61&MMN_position=57:57
    Thanks

  • How to Parse only single node

    Hi:
    I am DOM parser and i want to parse only single node. That is,
    String xmlRecords = "<name>Anup</name>";
    Here i want to parse name node.
    So can any body help me in that?
    Edited by: user11688398 on Jul 22, 2011 1:12 AM

    Ram wrote:
    Peter Lawrey wrote:
    Using String manipulation is likely to be faster than using a DOM (assuming you had valid XML) however if performance were an issue you would use JDBC to access the database.What will happen, if OP has large number of nodes in the string? Will String manipulation be the faster one? String manipulation will be faster only if the string has one or two nodes. Otherwise it will be complicated.Direct String manipulation will always be faster. This is because you can make assumptions a full DOM will not make (as it needs to be able to read any type of XML document)
    For this reason its best avoided unless you really need the speed as any subtle change in the XML could break a simplified parser.
    Peter Lawrey wrote:
    Make sure you have proper XML (not just a fragment) and use a document parser as EJP suggests.EJP didn't suggest anything, rather he did changed the conversation in wrong direction.I have been reading too many posts which look the same. Sabre not EJP made a suggestion. ;)

  • Error - convert single node-RAC-ConvertTomydb.xml -

    my single node init.ora file:
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # Cache and I/O
    db_block_size=8192
    db_file_multiblock_read_count=16
    # Cursors and Library Cache
    open_cursors=300
    # Database Identification
    db_domain=""
    db_name=mydb
    # Diagnostics and Statistics
    background_dump_dest=/u01/app/oracle/admin/mydb/bdump
    core_dump_dest=/u01/app/oracle/admin/mydb/cdump
    user_dump_dest=/u01/app/oracle/admin/mydb/udump
    # File Configuration
    control_files=("/u01/app/oracle/oradata/mydb/control01.ctl", "/u01/app/oracle/oradata/mydb/control02.ctl", "/u01/app/oracle/oradata/mydb/control03.ctl")
    # Job Queues
    job_queue_processes=10
    # Miscellaneous
    compatible=10.2.0.1.0
    # Processes and Sessions
    processes=150
    # SGA Memory
    sga_target=1083179008
    # Security and Auditing
    audit_file_dest=/u01/app/oracle/admin/mydb/adump
    remote_login_passwordfile=EXCLUSIVE
    # Shared Server
    dispatchers="(PROTOCOL=TCP) (SERVICE=mydbXDB)"
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=360710144
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_tablespace=UNDOTBS1
    my ConvertTomydb.xml ------------which is copy of ConvertToRAC.xml file
    <?xml version="1.0" encoding="UTF-8" ?>
    - <n:RConfig xmlns:n="http://www.oracle.com/rconfig" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.oracle.com/rconfig">
    - <n:ConvertToRAC>
    - <!-- Verify does a precheck to ensure all pre-requisites are met, before the conversion is attempted. Allowable values are: YES|NO|ONLY
    -->
    - <n:Convert verify="ONLY">
    - <!-- Specify current OracleHome of non-rac database for SourceDBHome
    -->
    <n:SourceDBHome>/u01/app/oracle/product/10.2.0/db_1</n:SourceDBHome>
    - <!-- Specify OracleHome where the rac database should be configured. It can be same as SourceDBHome
    -->
    <n:TargetDBHome>/u01/app/oracle/product/10.2.0/db_1</n:TargetDBHome>
    - <!-- Specify SID of non-rac database and credential. User with sysdba role is required to perform conversion
    -->
    - <n:SourceDBInfo SID="mydb">
    - <n:Credentials>
    <n:User>sys</n:User>
    <n:Password>oracle</n:Password>
    <n:Role>sysdba</n:Role>
    </n:Credentials>
    </n:SourceDBInfo>
    - <!-- ASMInfo element is required only if the current non-rac database uses ASM Storage
    -->
    - <n:ASMInfo SID="+ASM1">
    - <n:Credentials>
    <n:User>sys</n:User>
    <n:Password>oracle</n:Password>
    <n:Role>sysdba</n:Role>
    </n:Credentials>
    </n:ASMInfo>
    - <!-- Specify the list of nodes that should have rac instances running. LocalNode should be the first node in this nodelist.
    -->
    - <n:NodeList>
    <n:Node name="linux1" />
    <n:Node name="linux2" />
    </n:NodeList>
    - <!-- Specify prefix for rac instances. It can be same as the instance name for non-rac database or different. The instance number will be attached to this prefix.
    -->
    <n:InstancePrefix>mydb</n:InstancePrefix>
    - <!-- Specify port for the listener to be configured for rac database.If port="", alistener existing on localhost will be used for rac database.The listener will be extended to all nodes in the nodelist
    -->
    <n:Listener port="1521" />
    - <!-- Specify the type of storage to be used by rac database. Allowable values are CFS|ASM. The non-rac database should have same storage type.
    -->
    - <n:SharedStorage type="ASM">
    - <!-- Specify Database Area Location to be configured for rac database.If this field is left empty, current storage will be used for rac database. For CFS, this field will have directory path.
    -->
    <n:TargetDatabaseArea>+ORCL_DATA1</n:TargetDatabaseArea>
    - <!-- Specify Flash Recovery Area to be configured for rac database. If this field is left empty, current recovery area of non-rac database will be configured for rac database. If current database is not using recovery Area, the resulting rac database will not have a recovery area.
    -->
    <n:TargetFlashRecoveryArea>+FLASH_RECOVERY_AREA</n:TargetFlashRecoveryArea>
    </n:SharedStorage>
    </n:Convert>
    </n:ConvertToRAC>
    </n:RConfig>
    Ran the xml file
    $ rconfig ConvertTomydb.xml
    Got this below error.
    [oracle@linux1 bin]$ sh rconfig ConvertTomydb.xml
    <?xml version="1.0" ?>
    <RConfig>
    <ConvertToRAC>
    <Convert>
    <Response>
    <Result code="1" >
    Operation Failed
    </Result>
    <ErrorDetails>
    Clusterware is not configured
    </ErrorDetails>
    </Response>
    </Convert>
    </ConvertToRAC></RConfig>
    [oracle@linux1 bin]$
    Log file from /u01/app/oracle/product/10.2.0/db_1/cfgtoollogs/rconfig/rconfig.log
    [main] [0:14:4:4] [RuntimeExec.runCommand:175] Returning from RunTimeExec.runCommand
    oracle.ops.mgmt.cluster.RemoteShellException: PRKC-1044 : Failed to check remote command execution setup for node linux2 using shells /usr/bin/ssh and /usr/bin/rsh
    linux2.com: Connection refused
         at oracle.ops.mgmt.nativesystem.UnixSystem.checkRemoteExecutionSetup(UnixSystem.java:1880)
         at oracle.ops.mgmt.nativesystem.UnixSystem.getRemoteShellCmd(UnixSystem.java:1634)
         at oracle.ops.mgmt.nativesystem.UnixSystem.createCommand(UnixSystem.java:614)
         at oracle.ops.mgmt.nativesystem.UnixSystem.removeFile(UnixSystem.java:622)
         at oracle.ops.mgmt.nativesystem.UnixSystem.isSharedPath(UnixSystem.java:1352)
         at oracle.ops.mgmt.cluster.Cluster.isSharedPath(Cluster.java:916)
         at oracle.ops.mgmt.cluster.Cluster.isSharedPath(Cluster.java:859)
         at oracle.sysman.assistants.util.ClusterUtils.areSharedPaths(ClusterUtils.java:570)
         at oracle.sysman.assistants.util.ClusterUtils.isShared(ClusterUtils.java:501)
         at oracle.sysman.assistants.util.ClusterUtils.isShared(ClusterUtils.java:457)
         at oracle.sysman.assistants.util.attributes.CommonOPSAttributes.updateShared(CommonOPSAttributes.java:724)
         at oracle.sysman.assistants.util.attributes.CommonOPSAttributes.setNodeNames(CommonOPSAttributes.java:207)
         at oracle.sysman.assistants.rconfig.engine.Context.<init>(Context.java:54)
         at oracle.sysman.assistants.rconfig.engine.ASMInstance.createUtilASMInstanceRAC(ASMInstance.java:109)
         at oracle.sysman.assistants.rconfig.engine.Step.execute(Step.java:245)
         at oracle.sysman.assistants.rconfig.engine.Request.execute(Request.java:73)
         at oracle.sysman.assistants.rconfig.engine.RConfigEngine.execute(RConfigEngine.java:65)
         at oracle.sysman.assistants.rconfig.RConfig.<init>(RConfig.java:85)
         at oracle.sysman.assistants.rconfig.RConfig.<init>(RConfig.java:51)
         at oracle.sysman.assistants.rconfig.RConfig.main(RConfig.java:130)
    [main] [0:14:4:16] [UnixSystem.isSharedPath:1356] UnixSystem.isShared: creating file /u01/app/oracle/product/10.2.0/db_1/CFSFileName126249561289258204.tmp
    [main] [0:14:4:17] [UnixSystem.checkRemoteExecutionSetup:1817] checkRemoteExecutionSetup:: Checking user equivalence using Secured Shell '/usr/bin/ssh'
    [main] [0:14:4:17] [UnixSystem.checkRemoteExecutionSetup:1819] checkRemoteExecutionSetup:: Running Unix command: /usr/bin/ssh -o FallBackToRsh=no -o PasswordAuthentication=no -o StrictHostKeyChecking=yes -o NumberOfPasswordPrompts=0 linux2 /bin/true
    oracle.ops.mgmt.cluster.SharedDeviceException: PRKC-1044 : Failed to check remote command execution setup for node linux2 using shells /usr/bin/ssh and /usr/bin/rsh
    linux2.com: Connection refused
         at oracle.ops.mgmt.nativesystem.UnixSystem.testCFSFile(UnixSystem.java:1444)
         at oracle.ops.mgmt.nativesystem.UnixSystem.isSharedPath(UnixSystem.java:1402)
         at oracle.ops.mgmt.cluster.Cluster.isSharedPath(Cluster.java:916)
         at oracle.ops.mgmt.cluster.Cluster.isSharedPath(Cluster.java:859)
         at oracle.sysman.assistants.util.ClusterUtils.areSharedPaths(ClusterUtils.java:570)
         at oracle.sysman.assistants.util.ClusterUtils.isShared(ClusterUtils.java:501)
         at oracle.sysman.assistants.util.ClusterUtils.isShared(ClusterUtils.java:457)
         at oracle.sysman.assistants.util.attributes.CommonOPSAttributes.updateShared(CommonOPSAttributes.java:724)
         at oracle.sysman.assistants.util.attributes.CommonOPSAttributes.setNodeNames(CommonOPSAttributes.java:207)
         at oracle.sysman.assistants.rconfig.engine.Context.<init>(Context.java:54)
         at oracle.sysman.assistants.rconfig.engine.ASMInstance.createUtilASMInstanceRAC(ASMInstance.java:109)
         at oracle.sysman.assistants.rconfig.engine.Step.execute(Step.java:245)
         at oracle.sysman.assistants.rconfig.engine.Request.execute(Request.java:73)
         at oracle.sysman.assistants.rconfig.engine.RConfigEngine.execute(RConfigEngine.java:65)
         at oracle.sysman.assistants.rconfig.RConfig.<init>(RConfig.java:85)
         at oracle.sysman.assistants.rconfig.RConfig.<init>(RConfig.java:51)
         at oracle.sysman.assistants.rconfig.RConfig.main(RConfig.java:130)
    [main] [0:14:35:152] [Version.isPre10i:189] isPre10i.java: Returning FALSE
    [main] [0:14:35:152] [UnixSystem.getCSSConfigType:1985] configFile=/etc/oracle/ocr.loc
    [main] [0:14:35:157] [Utils.getPropertyValue:221] keyName=ocrconfig_loc props.val=/u02/oradata/orcl/OCRFile propValue=/u02/oradata/orcl/OCRFile
    [main] [0:14:35:157] [Utils.getPropertyValue:221] keyName=ocrmirrorconfig_loc props.val=/u02/oradata/orcl/OCRFile_mirror propValue=/u02/oradata/orcl/OCRFile_mirror
    [main] [0:14:35:157] [Utils.getPropertyValue:292] propName=local_only propValue=FALSE
    [main] [0:14:35:157] [UnixSystem.getCSSConfigType:2029] configType=false
    [main] [0:14:35:158] [Version.isPre10i:189] isPre10i.java: Returning FALSE
    [main] [0:14:35:168] [OCRTree.init:201] calling OCRTree.init
    [main] [0:14:35:169] [Version.isPre10i:189] isPre10i.java: Returning FALSE
    [main] [0:14:35:177] [OCRTree.<init>:157] calling OCR.init at level 7
    [main] [0:14:35:177] [HASContext.getInstance:190] Module init : 24
    [main] [0:14:35:177] [HASContext.getInstance:214] Local Module init : 0
    [main] [0:14:35:177] [HASContext.getInstance:249] HAS Context Allocated: 4 to oracle.ops.mgmt.has.ClusterLock@f47bf5
    [main] [0:14:35:177] [ClusterLock.<init>:60] ClusterLock Instance created.
    [main] [0:14:35:178] [OCR.getKeyValue:411] OCR.getKeyValue(SYSTEM.local_only)
    [main] [0:14:35:178] [nativesystem.OCRNative.Native] getKeyValue: procr_open_key retval = 0
    [main] [0:14:35:179] [nativesystem.OCRNative.Native] getKeyValue: procr_get_value retval = 0, size = 6
    [main] [0:14:35:179] [nativesystem.OCRNative.Native] getKeyValue: value is [false] dtype = 3
    [main] [0:14:35:179] [OCRTreeHA.getLocalOnlyKeyValue:1697] OCRTreeHA localOnly string = false
    [main] [0:14:35:180] [HASContext.getInstance:190] Module init : 6
    [main] [0:14:35:180] [HASContext.getInstance:214] Local Module init : 0
    [main] [0:14:35:180] [HASContext.getInstance:249] HAS Context Allocated: 5 to oracle.ops.mgmt.has.Util@f6438d
    [main] [0:14:35:180] [Util.<init>:86] Util Instance created.
    [main] [0:14:35:180] [has.UtilNative.Native] prsr_trace: Native: hasHAPrivilege
    [main] [0:14:35:184] [HASContext.getInstance:190] Module init : 56
    [main] [0:14:35:184] [HASContext.getInstance:214] Local Module init : 32
    [main] [0:14:35:184] [has.HASContextNative.Native] prsr_trace: Native: allocHASContext
    [main] [0:14:35:184] [has.HASContextNative.Native]
    allocHASContext: Came in
    [main] [0:14:35:184] [has.HASContextNative.Native] prsr_trace: Native: prsr_initCLSR
    [main] [0:14:35:185] [has.HASContextNative.Native]
    allocHASContext: CLSR context [1]
    [main] [0:14:35:185] [has.HASContextNative.Native]
    allocHASContext: retval [1]
    [main] [0:14:35:185] [HASContext.getInstance:249] HAS Context Allocated: 6 to oracle.ops.mgmt.has.ClusterAlias@18825b3
    [main] [0:14:35:185] [ClusterAlias.<init>:85] ClusterAlias Instance created.
    [main] [0:14:35:185] [has.UtilNative.Native] prsr_trace: Native: getCRSHome
    [main] [0:14:35:186] [has.UtilNative.Native] prsr_trace: Native: getCRSHome crs_home=/u01/app/oracle/product/10.2.0/crs(**)
    [main] [0:14:35:280] [ASMTree.getASMInstanceOracleHome:1328] DATABASE.ASM.linux1.+asm1 does exist
    [main] [0:14:35:280] [ASMTree.getASMInstanceOracleHome:1329] Acquiring shared CSS lock SRVM.ASM.DATABASE.ASM.linux1.+asm1
    [main] [0:14:35:280] [has.ClusterLockNative.Native] prsr_trace: Native: acquireShared
    [main] [0:14:35:281] [OCR.getKeyValue:411] OCR.getKeyValue(DATABASE.ASM.linux1.+asm1.ORACLE_HOME)
    [main] [0:14:35:281] [nativesystem.OCRNative.Native] getKeyValue: procr_open_key retval = 0
    [main] [0:14:35:282] [nativesystem.OCRNative.Native] getKeyValue: procr_get_value retval = 0, size = 36
    [main] [0:14:35:282] [nativesystem.OCRNative.Native] getKeyValue: value is [u01/app/oracle/product/10.2.0/db_1] dtype = 3
    [main] [0:14:35:282] [ASMTree.getASMInstanceOracleHome:1346] getASMInstanceOracleHome:ohome=/u01/app/oracle/product/10.2.0/db_1
    [main] [0:14:35:282] [ASMTree.getASMInstanceOracleHome:1367] Releasing shared CSS lock SRVM.ASM.DATABASE.ASM.linux1.+asm1
    [main] [0:14:35:282] [has.ClusterLockNative.Native] prsr_trace: Native: unlock
    [main] [0:14:35:802] [nativesystem.OCRNative.Native] keyExists: procr_close_key retval = 0
    [main] [0:14:35:802] [ASMTree.getNodes:1236] DATABASE.ASM does exist
    [main] [0:14:35:802] [ASMTree.getNodes:1237] Acquiring shared CSS lock SRVM.ASM.DATABASE.ASM
    [main] [0:14:35:802] [has.ClusterLockNative.Native] prsr_trace: Native: acquireShared
    [main] [0:14:35:803] [OCR.listSubKeys:615] OCR.listSubKeys(DATABASE.ASM)
    [main] [0:14:35:803] [nativesystem.OCRNative.Native] listSubKeys: key_name=[DATABASE.ASM]
    [main] [0:14:35:809] [GetASMNodeListOperation.run:78] Got nodes=[Ljava.lang.String;@11a75a2
    [main] [0:14:35:809] [GetASMNodeListOperation.run:91] result status 0
    [main] [0:14:35:809] [LocalCommand.execute:56] LocalCommand.execute: Returned from run method
    [main] [0:14:35:810] [ASMInstanceRAC.loadDiskGroups:2260] diskgroup: instName=+ASM2, diskGroupName=FLASH_RECOVERY_AREA, size=95378, freeSize=88454, type=EXTERN, state=MOUNTED
    [main] [0:14:35:811] [ASMInstanceRAC.loadDiskGroups:2260] diskgroup: instName=+ASM1, diskGroupName=FLASH_RECOVERY_AREA, size=95378, freeSize=88454, type=EXTERN, state=MOUNTED
    [main] [0:14:35:811] [ASMInstanceRAC.loadDiskGroups:2260] diskgroup: instName=+ASM2, diskGroupName=ORCL_DATA1, size=95384, freeSize=39480, type=NORMAL, state=MOUNTED
    [main] [0:14:35:811] [ASMInstanceRAC.loadDiskGroups:2260] diskgroup: instName=+ASM1, diskGroupName=ORCL_DATA1, size=95384, freeSize=39480, type=NORMAL, state=MOUNTED
    [main] [0:14:35:858] [ASMInstance.setBestDiskGroup:1422] sql to be executed:=select name from v$asm_diskgroup where free_mb= (select max(free_mb) from v$asm_diskgroup)
    [main] [0:14:35:864] [ASMInstance.setBestDiskGroup:1426] Setting best diskgroup....
    [main] [0:14:35:888] [SQLEngine.doSQLSubstitution:2165] The substituted sql statement:=select t1.name from v$asm_template t1, v$asm_diskgroup t2 where t1.group_number=t2.group_number and t2.name='FLASH_RECOVERY_AREA'
    [main] [0:14:35:888] [ASMInstance.setTemplates:1345] sql to be executed:=select t1.name from v$asm_template t1, v$asm_diskgroup t2 where t1.group_number=t2.group_number and t2.name='FLASH_RECOVERY_AREA'
    [main] [0:14:35:892] [ASMInstance.setTemplates:1349] Getting templates for diskgroup: oracle.sysman.assistants.util.asm.DiskGroup@170888e
    [main] [0:14:35:892] [ASMInstance.setTemplates:1357] template: PARAMETERFILE
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: DUMPSET
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: DATAGUARDCONFIG
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: FLASHBACK
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: CHANGETRACKING
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: XTRANSPORT
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: AUTOBACKUP
    [main] [0:14:35:893] [ASMInstance.setTemplates:1357] template: BACKUPSET
    [main] [0:14:35:894] [ASMInstance.setTemplates:1357] template: TEMPFILE
    [main] [0:14:35:894] [ASMInstance.setTemplates:1357] template: DATAFILE
    [main] [0:14:35:894] [ASMInstance.setTemplates:1357] template: ONLINELOG
    [main] [0:14:35:894] [ASMInstance.setTemplates:1357] template: ARCHIVELOG
    [main] [0:14:35:894] [ASMInstance.setTemplates:1357] template: CONTROLFILE
    [main] [0:14:35:894] [ASMInstance.createUtilASMInstanceRAC:113] Diskgroups loaded
    [main] [0:14:35:894] [LocalNodeCheck.checkLocalNode:107] Performing LocalNodeCheck
    [main] [0:14:35:894] [OracleHome.getNodeNames:270] inside getNodeNames
    [main] [0:14:36:116] [OracleHome.isClusterInstalled:252] bClusterInstalled=false
    [main] [0:14:36:120] [Step.execute:251] STEP Result=Clusterware is not configured
    [main] [0:14:36:121] [Step.execute:280] Returning result:Operation Failed
    [main] [0:14:36:121] [RConfigEngine.execute:67] bAsyncJob=false
    [main] [0:14:36:124] [RConfigEngine.execute:76] Result=<?xml version="1.0" ?>
    <RConfig>
    <ConvertToRAC>
    <Convert>
    <Response>
    <Result code="1" >
    Operation Failed
    </Result>
    <ErrorDetails>
    Clusterware is not configured
    </ErrorDetails>
    </Response>
    </Convert>
    </ConvertToRAC></RConfig>
    Log file from /u01/app/oracle/product/10.2.0/db_1/cfgtoollogs/rconfig/mydb/sqllog
    MYDB     mydb                    2622254467
    10.2.0.1.0     ACTIVE
    cluster_database                                        FALSE
    undo_management                                         AUTO
    db_domain
    dispatchers                                             (PROTOCOL=TCP) (SERVICE=mydbXDB)
    background_dump_dest                                        /u01/app/oracle/admin/mydb/bdump
    user_dump_dest                                             /u01/app/oracle/admin/mydb/udump
    core_dump_dest                                             /u01/app/oracle/admin/mydb/cdump
    audit_file_dest                                         /u01/app/oracle/admin/mydb/adump
    MYDB     mydb                    2622254467
    10.2.0.1.0     ACTIVE
    cluster_database                                        FALSE
    undo_management                                         AUTO
    db_domain
    dispatchers                                             (PROTOCOL=TCP) (SERVICE=mydbXDB)
    background_dump_dest                                        /u01/app/oracle/admin/mydb/bdump
    user_dump_dest                                             /u01/app/oracle/admin/mydb/udump
    core_dump_dest                                             /u01/app/oracle/admin/mydb/cdump
    audit_file_dest                                         /u01/app/oracle/admin/mydb/adump
    MYDB     mydb                    2622254467
    10.2.0.1.0     ACTIVE
    cluster_database                                        TRUE
    undo_management                                         AUTO
    db_domain
    dispatchers                                             (PROTOCOL=TCP) (SERVICE=mydbXDB)
    background_dump_dest                                        /u01/app/oracle/admin/mydb/bdump
    user_dump_dest                                             /u01/app/oracle/admin/mydb/udump
    core_dump_dest                                             /u01/app/oracle/admin/mydb/cdump
    audit_file_dest                                         /u01/app/oracle/admin/mydb/adump
    MYDB     mydb                    2622254467
    10.2.0.1.0     ACTIVE
    cluster_database                                        TRUE
    undo_management                                         AUTO
    db_domain
    dispatchers                                             (PROTOCOL=TCP) (SERVICE=mydbXDB)
    background_dump_dest                                        /u01/app/oracle/admin/mydb/bdump
    user_dump_dest                                             /u01/app/oracle/admin/mydb/udump
    core_dump_dest                                             /u01/app/oracle/admin/mydb/cdump
    audit_file_dest                                         /u01/app/oracle/admin/mydb/adump
    MYDB     mydb                    2622254467
    10.2.0.1.0     ACTIVE
    cluster_database                                        TRUE
    undo_management                                         AUTO
    db_domain
    dispatchers                                             (PROTOCOL=TCP) (SERVICE=mydbXDB)
    background_dump_dest                                        /u01/app/oracle/admin/mydb/bdump
    user_dump_dest                                             /u01/app/oracle/admin/mydb/udump
    core_dump_dest                                             /u01/app/oracle/admin/mydb/cdump
    audit_file_dest                                         /u01/app/oracle/admin/mydb/adump
    Please help me where I am making mistake.
    Thanks

    1) I have created single node standard database called mydb in /u01/app/oracle/product/10.2.0/db_1 home (hostname linux1)
    2) installed crs and asm on linux1 and linux2 and shared storag on ASM(which external HD running ieee1294 cards and ports) . no database is created on linux1 or linux2
    3) I want to convert mydb to RAC DATABASE called mydb1 instance on linux1 and mydb2 on linux2 machine respectively.
    4) copied and modifed xml as you see above called ConvertTomydb.xml to $ORACLE_HOME/bin directory
    5) when I run
    $rconfig ConvertTomydb.xml from $ORACLE_HOME/bin directory , i get the following error
    <ConvertToRAC>
    <Convert>
    <Response>
    <Result code="1" >
    Operation Failed
    </Result>
    <ErrorDetails>
    Clusterware is not configured
    </ErrorDetails>
    </Response>
    </Convert>
    </ConvertToRAC>
    $
    Please see my crs_stat -t command output
    Name Type Target State Host
    ora....SM1.asm application ONLINE ONLINE linux1
    ora....X1.lsnr application ONLINE ONLINE linux1
    ora.linux1.gsd application ONLINE ONLINE linux1
    ora.linux1.ons application ONLINE ONLINE linux1
    ora.linux1.vip application ONLINE ONLINE linux1
    ora....SM2.asm application ONLINE ONLINE linux2
    ora....X2.lsnr application ONLINE ONLINE linux2
    ora.linux2.gsd application ONLINE ONLINE linux2
    ora.linux2.ons application ONLINE ONLINE linux2
    ora.linux2.vip application ONLINE ONLINE linux2
    ora.orcl.db application ONLINE ONLINE linux1
    ora....l1.inst application ONLINE ONLINE linux1
    ora....l2.inst application ONLINE ONLINE linux2
    ora....test.cs application ONLINE ONLINE linux1
    ora....cl1.srv application ONLINE ONLINE linux1
    ora....cl2.srv application ONLINE UNKNOWN linux2
    please see the output from olsnodes command
    [oracle@linux1 bin]$ olsnodes
    linux1
    linux2
    [oracle@linux1 bin]$
    What is your cache fusion interconnect strategy?
    I don't about this, please let me know where can i find the answers. what kind of command do i have to use get the answer
    damorgan , Please let me know, if I gave answers to your questions. if not please let me know, i can give as much as possible. i really appreciate for your help
    Thanks

  • 9ir2 Single Node RAC + RH80

    Hi all.
    I'm triying to install a single node RAC based in the single node rac for linux document in the metalink.
    The document was made based in a oracle 9iR1 database an RH71, I have followed the steps and i have a working kernel made by my own, the raw devices, and softdog hacked for being testing only. But the service does not go up.
    I don't know if someone have made a single node rac instalation and can help me to choose the right pieces.
    I don't have a RHAS2.1 but I do have a UnitedLinux 1.0 Server Patched.
    I want to try the RAC tech but here (ecuador-sudamerica) is dificult to find firewire cards and disks so I´m working around a single node instalation.
    Any help will be apreciate.
    Fernando

    Hi
    Database version and GI versions are 11.2.0.2.2. And these are not multinode RAC configuration at any given time only single instance will be there for any given database. Some thing like ACTIVE and PASSIVE in hardware clusters such are VCS (Veritas Clusters).
    I agree with you failover sinario in multi node (2 or more) RAC environments. In single node clustering only one instance will be there, like services in multinode, here whole instance will be re created on available nodes.
    Hi
    Database version and GI versions are 11.2.0.2.2. And these are not multi node RAC configuration at any given time only single instance will be there for any given database. Some thing like ACTIVE and PASSIVE in hardware clusters such are VCS (VERITAS Clusters).
    I agree with you failover scenario in multi node (2 or more) RAC environments. In single node clustering only one instance will be there, like services in multi node, here whole instance will be re created on available nodes.
    Thanks,
    Thanks,

  • Configurations, instances, nodes, virtual hosts, clusters ... oh my!

    Perhaps I am being dense, but I am having some trouble wrapping my head around the different "objects" of a SJS web server 7 installation. It is not clear to me how they relate/map to each other and to real processes on the server (logical architecture, mapped to physical architecture).
    I had the same difficulty with WebSphere App Server 6 terminology until, with one fell swipe of the Sword of Fundamental Truthiness, a very brilliant instructor cleared it up for me when he said (using WAS6 parlance) "a 'profile' is the on-disk configuration, files etc, of a node. 1 profile = 1 node. The node is the run-time object describe by the profile." Bingo. Everything (nodes, servers, cells, profiles) made sense after that.
    I've been through the docs, but it is still not clear to me. Any lucid descriptions would be appreciated. There's configurations, nodes, instances, virtual hosts, listeners, clusters.
    Is a cluster a collection of nodes? Or instances? Virtual server vs. instance? A "configuration" is the on-disk configuration/files etc describing a ???" Maybe there is a nice diagram out there someplace that maps all these bits together?
    Thanks,
    Mark

    Thanks all. Like I said, I've read the docs. They are not terribly clear, at least not to me, but maybe I am just being dense. Some detailed diagrams would be useful.
    [In the text below, I use leading CAPS to differentiate between generic use of a term and Web Server's use of a term ... instance vs. Instance for example.]
    To say "an Instance is a Configuration deployed to a Node" doesn't make any sense to me, since an "Instance" is a configuration item within the Configuration. Maybe it is technically correct
    I think I have the following correct:
    * A Configuration defines/describes an HTTP service that you wish to provide. It ties together all the elements required to define that service: ports to listen on, SSL or no SSL, document roots, access control and so on.
    * You deploy Configurations to Nodes. You define which Nodes to deploy the Configuration to using the Instances tab within the Configuration. The Instance is the run-time representation within a Node of the Configuration (as the Node is the run-time representation within a Cell of the on-disk Profile in WebSphere). You may only have one Instance of a Configuration per available Node. If you have multiple Configurations, you can deploy one Instance of each Configuration to a single Node. An important point here is that Instances really have nothing to do with Virtual Hosts or HTTP Listeners. They are related to Configurations. Configurations have a one-to-(potentially)-many relationship with Instances. Instances within a single Configuration have a one-to-one relationship with Nodes.
    * A Listener listens for incoming HTTP requests on a single port. You may have multiple Listeners per Configuration. A Listener forwards incoming HTTP requests to the "appropriate" Virtual Host.
    * You can map a single Listener (therefore a single port) onto multiple Virtual Hosts, (Why would you do this?) and of course, multiple Listeners onto a single Virtual Host. Listeners and Virtual Hosts have a many-to-many relationship.
    * A virtual host is really the definition of the "real HTTP service", by which I mean it defines the doc root, access controls etc.
    I think I have it straight. I apologize if this is amateur. I am hoping to replace all my Apache servers with SJSWS and need to have all my ducks lined up to make the argument for doing so.
    Thanks again,
    Mark

  • Install 10gr2 ASM on single node, need clusterware installe first?

    Install 10gr2 ASM on single node, Do I need install clusterware first?
    When I install clusterware on single node, it shows " the specified nodes are not clusterable"
    Thanks.

    ASM and clusterware are two separate things. Clusterware is required only if you are setting up a RAC environment, is this your case? If you are just planning to use ASM then proceed with the RDBMS (database disk) install and use the DBCA to perform ASM configuration. It is assumed you have the raw devices ready to setup the ASM instance. Once the ASM instance is ready you can create the database to use the ASM disks.
    ~ Madrid
    http://hrivera99.blogspot.com/

Maybe you are looking for

  • Adobe Reader bypasses "secure print" settings when printing from IE add-on

    We have the IE add-on for Adobe Reader enabled (Internet Explorer 9), and we've found that for some people this is causing PDFs to get printed without the "secure print" option, even though the print dialog shows it. Our printer settings all have the

  • Asset Management in Air App

    Hello All, I'm building an AIR application (2d platformer adventure game) which will be packaged on a CD for installation. I've embed the majority of the game assets into a static file to keep it simple and organized. This way I can reuse assets and

  • (approver) for PO approval workflow ?

    Hi everybody, Quick question: The WF guy told me that the PO approval workflow (1 step) works by default with the immediate manager (the manager is the approver). Im confused with this.. Is this true ?   If yes, What is then the purpose of Assign Age

  • Install error - database configuration assistant

    HI. During install a 10g oracle database, the universal installer prompt me that one of required component , DATABASE CONFIGURATION ASSISTANT, is not properly installed. The first warning has been submitted with popup that reported " ORA - 22288 , fi

  • R&R 4.5 cannot make bootable USB

    Hi, i have upgraded to R&R 4.5, but i have the problem. I have switched to UEFI, so my system SSD drive is GPT now. I have reformated my USB drive for backups to GPT as well. Now i started R&R to make first backup and make my USB disc bootable. When