Run something once in a cluster environment

I'm looking at using JBoss Cache to cache a large amount of data.  If you're unfamiliar with JC, it's a cachine system that will synchronize multiple remote caches in a clustered environment.  However, I don't want each cache to load itself at cluster startup, and then sync with all the other caches - there's a lot of data, and I don't want all the caches to become really chatty during startup for performance reasons.  I would rather have one server instance act as the master cache, and then have it sync all the other empty caches...
I therefore need a way to determine what node in the cluster I'm running on.  How can I programatically do this?  I've looked through the API's, but it's like trying to find a needle in a haystack...

Hi Ken,
Ok, at least according t its documentation  the jboss cache will deal in its own with the failure detection, replication, new servers joining, which answers my curiosity in the upper memo
I was speaking of servers later than NW 2004, but here is a suggestion how to deal with NW 2004:
In order to achieve your goal - "prefetch everything only on one node", you could play with the locking feature. Here is a possible algorithm that could be put in code executed at application start (servlet init method, etc.):
1) Check if the cache is empty, if it is not, you are secondary and someone has already updated you - exit
2) Try to take a cluster unique logical lock in the locking service (here is the <a href="http://help.sap.com/saphelp_erp2005/helpdata/en/9a/4cdcc80fa4c747a2ccb5859f467412/frameset.htm">locking documentation</a>
)If you cannot take the lock- you are the secondary server and someone is refreshing the cache at the moment, exit
3) Update the cache
4) Release the lock
If there is no API to check if the cache is empty, you could potentially use some variable that is part of the cache and will be replicated to all servers. Before releasing the lock, you put some value "CacheUpdated" there.
Another option would be to use some hardcoded logic and workarounds - do something only for specific server id as suggested in upper memo, do something if there is a specific property (System.getProperty()...) that you would put manually in only one server JVM startup options.
Perhaps the simplest solution is to turn off all the cache prefetching - i.e. all nodes will start with empty cache and will load data on demand. You won't screw the performance of the "master" server as well.
HTH
Peter

Similar Messages

  • OPMN fails in cluster environment after the machie reboots.

    Dear All:
    Have you ever meet the following problem or have any suggestion for it? Thanks a lot.
    There are machine A and B have been setup successfully in a cluster environment, everything is fine.
    However, after the machine B reboots, the Oracle BI Server and Presentation Server in machine B could not start.
    The error message for Oracle BI Server is:
    [2011-06-02T10:40:34.000+00:00] [OracleBIServerComponent] [NOTIFICATION:1] [] [] [ecid: ] [tid: 1060] Server start up failed: [nQSError: 43079] Oracle BI Server could not start because locating or importing newly published repositories failed.
    The error message for presentation server is:
    [2011-06-02T18:40:24.000+08:00] [OBIPS] [ERROR:1] [] [saw.sawserver.initializesawserver] [ecid: ] [tid: ] Another instance of Oracle Business Intelligence is trying to upgrade/update/create the catalog located at \\?\UNC\10.91.61.158\BIEEShare\CATALOG\SampleAppLite\root. Retry after it finishes.[
    The detail information is:
    1)     Machine A and B both run Windows 2008 server 64bit.
    2)     Machine A runs Admin Server+BI_Server1(managed server). Machine B is configured with “Scale out an existing installation” option and runs BI_Server2(managed server).
    3)     BIEE 11.1.1.3
    4)     In EM, the shared location of shared repository has been set to something like \\10.91.61.158\shared\RPD. The catalog path has been set to \\10.91.61.158\BIEEShare\CATALOG\SampleAppLite. Both path could be accessed by machine A and B.
    5)     Before machine B reboots, everything is fine. All the servers could be started.
    6)     After machine B reboots, the managed server BI_server2 could be started successfully. However, “opmnctl startall” would incur the above error message.
    Any suggestion is welcome. Thanks a lot!

    Hello,
    Thanks for answering.
    I am using OBIEE 11.1.1.5.0 enterprise software-only install scaleout existing Bi system option on HostMachine2. All services and coreapplication systems are running except coreapplication_obis1, coreapplication_obis2, coreapplication_obips1 and coreapplication_obips2 of bi_server2 on HostMachine2.
    Here is my ClusterConfig.xml, as you can see all changes in this file is done by em and primary and secondary controller settings are already applied. The same file also exist in the same location on HostMachine2.
    <?xml version="1.0" encoding="UTF-8" standalone="no"?>
    <Cluster xmlns="oracle.bi.cluster.services/config/v1.1" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="oracle.bi.cluster.services/config/v1.1 ClusterConfig.xsd">
    <ClusterProperties>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ClusterEnabled>true</ClusterEnabled>
    <ServerPollSeconds>5</ServerPollSeconds>
    <ControllerPollSeconds>5</ControllerPollSeconds>
    </ClusterProperties>
    <NodeList>
    <Node>
    <NodeType>PrimaryController</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obiccs1</NodeId>
    <!--HostNameOrIP can be a hostname, IP or virtual hostname-->
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9706</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9700</MonitorPort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ListenAddress>HostMachine1.localdomain.com</ListenAddress>
    </Node>
    <Node>
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MasterServer>true</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obis1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9703</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9701</MonitorPort>
    </Node>
    <Node>
    <NodeType>Scheduler</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><NodeId>instance1:coreapplication_obisch1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><ServicePort>9705</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><MonitorPort>9708</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance1:coreapplication_obis2</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine1.localdomain.com</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9702</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9709</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obis1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9761</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9762</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Server</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MasterServer>false</MasterServer>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obis2</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9763</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9764</MonitorPort>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>SecondaryController</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obiccs1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9765</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9766</MonitorPort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ListenAddress>HostMachine2</ListenAddress>
    </Node>
    <Node>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeType>Scheduler</NodeType>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <NodeId>instance2:coreapplication_obisch1</NodeId>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <HostNameOrIP>HostMachine2</HostNameOrIP>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <ServicePort>9770</ServicePort>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager-->
    <MonitorPort>9771</MonitorPort>
    </Node>
    </NodeList>
    <SSLProperties>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><SSL>false</SSL>
    <SSLCertificateFile/>
    <SSLPrivateKeyFile/>
    <SSLCACertificateFile/>
    <SSLVerifyPeer>false</SSLVerifyPeer>
    </SSLProperties>
    </Cluster>

  • Microsoft cluster environment and DR

    Hi all,
    I wonder if anyone can help me with this question. We are running SAP in a HA environment on two physical servers and have virtual servers as additional dialog servers.
    My undertsanding of the new 7.0 cluster architecture is the the message and enqueue server has been split out of the CI. So therefor a CI is built up of a enqueue server, a message server and a dialog server.
    So in our site we have two clustered servers vic1 and vic2 and 1 virtual server vicvm1 ECC6.0 CI install was done on vic1and the message server and enque server was set up on the cluster nodes running in separate resource groups. vicvm1 was deployed as a secondary application server. This is a pure abap environment
    Now to my question.
    We are building a dr environment where there is a db server vicdr1 and the application server vicvm1 is being replicated into the DR site.
    Is it possible to do a ASCS install on vicdr1 and bring up the replicated server vicvm1 for the system to work? The reason i ask this is there seems to be no clear information on this, and if the instance on vic1 crashed vicvm1 still works and users see no difference. One user has pointed out that there always needs to be a CI running but once you split the message and enqueue server out the CI becomes meaningless and simply another application server.

    Hi David
    1st - you asked "if the instance on vic1 crashed vicvm1 still works and users see no difference". So in your system if vic1 and vic2 are clustered and vic 1 crashes, doesn't the Ci automatically fail over to the vic 2 cluster node? That would be why the users can keep running transactions on the app server vicvm1.
    2nd - "Is it possible to do a ASCS install on vicdr1 and bring up the replicated server vicvm1 for the system to work?"  I'm not exactly sure what you are asking here, but I think you will need to re-install the application on the dr systems vicvm1, pointing it to the vicdr1 system. Not just replicate or copy the vicvm1 onto another server and have it work.
    Ken

  • SBO Mailer in a cluster environment

    Hi,
    I am currently facing an issue where when I install SBO Mailer on a cluster environment, the installation removes applications and license manager for some reason gets uninstalled.. When I want to uninstall the SBO Mailer, I do not see it under the Control Panel and neither do I see it in running in services. But if i look at the Service Manager the SBO Mailer Service is running.
    The strange part is that after this installation, there is an add-on that stopped working.
    Has anyone come across such a scenario?

    Hi,
    To uninstall you have to first run the Server tools and there you can uncheck the sbo mailer. There is not way you can uninstall SBO mailer service and you cannot find it under control panel its a service...not an applicaiton.
    Run the server tools and then choose repair.
    Regards,
    Rakesh N

  • How do I submit a job to run only once?

    I have a stored procedure that submits a job. If this job fails, I don't want it to continue trying to run. Does anyone know how to ensure that it will only run once, no matter what?
    I did try something similar to the following:
    procedure proc_submitted_as_job(p_job_no number) is
    begin
    dbms_job.broken(p_job_no, true);
    end;
    However, I found out that if you change the Broken state while a job is running, then it gets reset when the job completes.
    Basically, I'm using the dbms_job utility to run a long-running procedure in the background. If there's another way to obtain this same result, I'd love to hear it.

    I am still haveing problems so I will try and attach my program and see if you can help a little more that. It will probably be easier for you to look at this way. When you look at the diagram sequence 0 is the one the problem is in. In that sequence you will see a case structure and that is where the finite pulse train vi is. When case 1 is selected the finite pulse train will run again and again until I leave case 1. I need to have the program run the finite pulse train once and then wait and do nothing until the case is changed. The finite pulse train will be in all cases inside the case structure I just haven't done that yet, I am waiting until it works. The pulse train is used to move a stepper motor that is why running only once is im
    portant because the stepper motor is used to position something.
    Attachments:
    AMC_Eagle_Both_Inputs_Working_2.vi ‏266 KB

  • How do I get a case to run only once

    I am using an e-series DAQ board to bring a pwm signal, then I use labview to use that signal to pick one of three cases. depending on the case picked a finite pulse stream is sent out. Currently the finite pulse vi keeps repeating because the the program stays in the case. I need a way to get labview to run the finite pulse output only once, and not run again until a new case is picked. I don't have the case in a while loop.
    Thanks For The Help

    I am still haveing problems so I will try and attach my program and see if you can help a little more that. It will probably be easier for you to look at this way. When you look at the diagram sequence 0 is the one the problem is in. In that sequence you will see a case structure and that is where the finite pulse train vi is. When case 1 is selected the finite pulse train will run again and again until I leave case 1. I need to have the program run the finite pulse train once and then wait and do nothing until the case is changed. The finite pulse train will be in all cases inside the case structure I just haven't done that yet, I am waiting until it works. The pulse train is used to move a stepper motor that is why running only once is im
    portant because the stepper motor is used to position something.
    Attachments:
    AMC_Eagle_Both_Inputs_Working_2.vi ‏266 KB

  • PI 7.1 in a cluster environment (multiple ip-adresses): P4 port

    We want to  install PI 7.1 on unix in a cluster environment.Therefore we  installed also DEV+QA with virtual hostnames like the prodsystem, which will be later installed.
    At all sapinst installation screens we have used only the virtual hostname <virtual-hostname-server interface>.We have also set the SAPINST_USE_HOSTNAME=<virtual-hostname-server interface>. Although the P4-port seems to have used the physical hostname: in step 57 of sapinst we got problems and in dev_icm were:
    [Thr 05] *** ERROR => client with this banner already exists:
    1:<physical-hostname>:35644 {000306f5} [p4_plg_mt.c 2495]
    After we have set
    icm/server_port_1 = PROT=P4,PORT=5$$04, HOST=<virtual-hostname-server interface>
    icm/server_port_6 = PROT=P4,PORT=5$$04, HOST=<virtual-hostname-user interface>
    icm/server_port_7 = PROT=P4,PORT=5$$04, HOST=<physical hostname>
    icm/server_port_8 = PROT=P4,PORT=5$$04, HOST=127.0.0.1
    the sapinst was successfull.
    Now we're not sure how to set these P4-parameters in our  future productive cluster environment.
    Our productive system PX1 will live in a HA environment, so we don't want to use the physical hostnames in any profile.
    Our environment will look like:
    HOST-A (<physical-hostname-A>):
    <virtual-hostname-server interface>
    <virtual-hostname-user interface>
    HOST-B (<physical-hostname-B>):
    Normally our prodsystem will live on Host-A (physical-hostname-A). All parameters should
    only take the virtual hostname <virtual-hostname-server-interface>. During switchover the
    virtual hostnames (server and user interface) will be taken over to HOST-B, while the physical
    hostnames of HOST-A and HOST-B will stay like there are.
    How do the parameters have to be set here ?
    Have also the physical hostnames of both cluster nodes set in the
    instance profile, e.g:
    icm/server_port_1 = PROT=P4,PORT=5$$04, HOST=<virtual-hostname-server interface>
    icm/server_port_6 = PROT=P4,PORT=5$$04, HOST=<virtual-hostname-user interface>
    icm/server_port_7 = PROT=P4,PORT=5$$04, HOST=<physical-hostname-A>
    icm/server_port_8 = PROT=P4,PORT=5$$04, HOST=<physical-hostname-B>
    icm/server_port_9 = PROT=P4,PORT=5$$04, HOST=<localhost>
    Any recommendations ? In note 1158626 is some infomation regarding P4 ports with multiple network interfaces, but it's not 100% clear for us.
    Best regards,
    Uta

    Hi Uta!
    Obviously we are the only human beings in the SAP community having this problem. Nevertheless let's give it another try with a - hopefully - more simple problem description (and maybe it will be helpful to copy and paste this description into the open SAP CSN also).
    So here comes the scenario:
    We have one physical host:
    Physical hostname: physhost
    Physical IP address: 1.1.1.1
    On this physical host there is running one OS: SUN Solaris 10/SPARC
    On top of this we have two virtual hosts where we install 2 completely independent PI 7.1 instances with separate virtual hostnames and separate virtual IP addresses and separate DB2 9.1 databases. That is this is not an MCOD installation.
    Virtual Host no. 1 is PI 7.1 Development System:
    Virtual hostname: virthostdev
    Virtual IP address: 2.2.2.2
    Java Port numbers: 512xx
    Virtual Host no. 2 is PI 7.1 QA System:
    Virtual hostname: virthostqa
    Virtual IP address: 3.3.3.3
    Java Port numbers: 522xx
    With this constellation we face serious problems with the P4 port. Currently for example the JSPM for virthostdev does not start, because JSPM cannot connect to the P4 port.
    In SAP note 1158626 we have learned that by default always the physical hostname/IP address is used to address the P4 port and that we have to configure instance profile parameter icm/server_port_xx to avoid this.
    So how do we have to configure the instance profile parameter icm/server_port_xx for both systems to resolve these P4 port conflicts?
    Additionally: Is it important to use distinct server port slot numbers xx in both systems?
    Additionally: Is it possible to configure this parameter with hostnames instead of using IP addresses?
    So far we have tried several combinations, but with each combination at least one or even both systems have problems with that f.... P4 port.
    Please help! Thanx a lot in advance!
    Regards,
    Volker

  • SAP R/3 4.7 SR1 Migration to MSCS cluster environment

    Hi all,
    We are planning to move our existing standalone server R/3 production server to cluster environment.
    I have some queries regarding this migration
    Our R/3 system is running on SAP R/3 4.7x1.10 SR1 on Windows 2003 and Oracle 9.2
    1. Central Instance has to be installed in local disk or shared disk
    2. Is it correct to R3load procedures to export data from old server and import during Database installation
    3. Presently our server is running with kernel 640, but for the new installation kernel will be 620. So when can i upgrade kernel immediately after central instance installation or after CI and DB Instance installation.
    Give me some inputs
    Thanks & Regards
    KT

    Hi, the best place for your questions would have been the forum SAP on Windows.
    However, here the answers.
    1) CI must be on a shared disk
    2) Right, if you don't change the database platform and its release you can also use backup/restore
    3) Right. You can also perform the installation with the latest NW04 installation master

  • How to setup the cluster environment for BPM using weblogic

    want to setup the cluster environment for BPM using weblogic....
    i have installed the oracle weblogic server 10gr3 and oracle BPM enterprise for weblogic 10gR3
    i have used the Admin tools from the "oracle BPM enterprise for weblogic" to setup the configuration and create the weblogic domain servers.
    i can launch the process administrator and import the project exp file to domain server.
    but what should i do to setup cluster environment using weblogic?
    what i want to do is :
    setup one admin machine..
    setup two product machine..
    enable the cluster so the admin machine can monitor the status of the product machine..
    thanks a lot ...

    The install guide at http://download-llnw.oracle.com/docs/cd/E13154_01/bpm/docs65/config_guide/index.html gives a reasonable amount of info on how to do this.
    Personally I have not used the OBPM option to configure WebLogic instead I've used the information in the above install guide to create the weblogic domain in advance of configuring OBPM.
    Once you've setup WebLogic configure OBPM using the values I mention in the following thread: How to set the JMX Engine parameter in Process Administation?
    Let me know any specific config questions and I'll do my best to answer them for you.
    Thanks,
    Mike

  • Installation of CRM 2007 in Windows with oracle and cluster environment

    Dear Experts,
    We are about to start the installation of CRM 2007 (ABAP+JAVA) with
    Oracle 10g on Windows x64 in cluster environment. In the SAPINST dialog
    box under High availability option, I could see installation options like ASCS
    Instance, SCS Instance, First MSCS node, Database Instance, Additional
    MSCS Node, Enqueue Replication Server, Central Instance and Dialog
    Instance.
    I have gone through the installation guide. I have below queries
    regarding the same. Can you please clarify the same
    1) Our requirement is we want to have an ACTIVE-ACTIVE cluster setup
    with sap service running in Node A and Database running in Node B. Can
    we have this setup in ABAP+JAVA setup
    2) Also,in the SAPINST dialog box as said above except last two
    (Central and Dialoge instance) as per standard installation guide are
    to be installed in shared drives as per the requirement. But, central
    and dialog are said to be installed locally on any one node in MSCS or
    in separate server. As per my understanding Dialog instance will be
    used for load balancing which we do not require now. Hence I feel this
    is optional in our case. Where as Central Instance I am not able to
    understand why this option is for? Is it mandatory to be installed or optional. If
    so, when I installed it in one of the MSCS node the incase of failover
    how does it effect the system. As per my understanding ASCS and SCS
    comprise central instance.
    Please clarify and thanks in advance.
    Regards,
    Sharath Babu M
    +91-9003072891

    I am following as per the standard installation guide.
    Regards
    Sharath

  • Can Java SE run in a Window PE (Preinstallation Environment)

    Hi there,
    Can Java SE run in a Window PE (Preinstallation Environment).
    Thank you for all your help....

    One simple reason is that ME has a lot less core classes than SE. So there is a big possibility that you use something that will not be available on the ME.

  • How to set property for Cluster Environment for JMS Adapter

    Hi All,
    I am moving from DEV to Prod environment which is cluster.
    Can you Please explain me what property I need to Set for Cluster Environment for JMS Adapter, so that I could avoid race condition for Dequeue/enqueue.
    I am using soa suite 10.1..3.4
    Thanks in Advance.
    Edited by: vikky123 on Jul 12, 2010 7:03 AM

    put something like this
    <activationAgents>
          <activationAgent className="oracle.tip.adapter.fw.agent.jca.JCAActivationAgent" partnerLink="PARNERLINKNAME">
            <property name="clusterGroupId">SOMEUNIQUEVALUE</property>
            <property name="portType">PARTNERLINK_PORTTYPE</property>
          </activationAgent>
        </activationAgents>

  • Scheduling,Email Setup in cluster environment and multiple job servers

    Hello All,
        I have to schedule and email instances of crystal reports to users in PDF and excel format. A Clustered environment of Business Objects 3.1 SP3 is been setup on 2 servers. There is 2 instances of crystal reports Job server been created.
    My question is If there are multiple instances of job servers then should we have to configure mail settings on both instances of job server or not?
    And secondly, in cluster environment the email settings and scheduling of the reports should be setup on both CMS servers( CMS server 1 and CMS server 2) or should CMS server 1 should be enough?
    Thanks in advance.

    You have a couple of options here.
    1.  You can either set up email on all of the job servers on both CMS' in the cluster.  This is actually the easiest to manage.
    2.  You can set up email on some of the job servers on one or the other CMS.  However, at this point you would also have to set up a server group that contains the job servers that are set up for email, plus the RAS server - you might also have to include the Crystal Cache server and the Crystal Processing server.  Then ALL of the reports that are sent via email would have to be scheduled run ONLY on the server group that contains the email-enabled job server(s).
    -Dell

  • SQL PATCH IN CLUSTER ENVIRONMENT

    Dear All,
    One of my client is having SAP ECC6.0 , SQL 2005 , windows 2003 64 bit.
    They are having SQL 2005 version 9.00.1399 and according to SAP it is no more supported so they want to upgrade it to 9.00.4035.
    In DEV and QAS (non cluster environment) they are already having 9.00.4035 .
    - Can you guys please guide me about how to apply SQL patch in Cluster environment.
    - Shall I apply patch 9.00.4035 directly from 9.00.1399
    - Any recommendations before and after applying patched
    Kindly guide.
    Regards,
    Rmoihnitii

    Hello my friend
    Unfortunately this is not a zero down time solution since Database Services and Analysis Services checkboxes will not be available while patching SP3:
    1. Launch SP3 package on the active node;
    2. Select authentication mode for the instance you are going to update;
    3. Enter credentials of the login you want to run service pack on behalf of;
    4. Click all the way next to start installing;
    Then you need to update the inactive node it is necessary to execute all the steps except steps 2 and 3. There can be some problems with updating database service in the cluster. Restart of each node of the cluster and restart of the cluster service usually resolves all the problems. After restart try to update database service one more time. 
    Regards,
    Effan

  • How to run the Start_Script outside the cluster

    Hi
    We have a 3 node cluster and we have 4 RG created. All is working fine.
    In the situation , where the appliaction stops for some reason , I have created the probe script to only page the team and not do anything else. The probe script is always returning 0.
    Also , I have suppresed the PMF monitoring as described by http://blogs.sun.com/TF/entry/disabling_pmf_action_script_with
    This is the start of my start script
    =====================================
    #!/bin/ksh
    while getopts 'R:G:' opt
    do
    case "${opt}" in
    R) RESOURCE=${OPTARG};;
    G) RESOURCEGROUP=${OPTARG};;
    esac
    done
    sleep 60 &
    /usr/cluster/bin/pmfadm -s ${RESOURCEGROUP},${RESOURCE},0.svc
    [...rest of your code to start the application...]
    ===============================================
    The reason , we did this way is because
    1>When the application daemon goes down , only the master process goes down , all the child process are still up and running . The child process do not need the main daemon as long as they are UP. They only need main daemon if they need to restart.
    2>If we go by complete restart of the Resource , it will kill all the child process before starting them. This can be cause of concern for us since restarting of the child process required vendor interaction attimes as most of them are TCPIP connections with various applications(We are interface engine shop)
    3>However, We can just start the main daemon only and all the previous child process can connect back to it.
    Question is how? I tried starting the Start script outside the cluster by passing the parameters of Resource and RG , but it threw this error
    "pmfadm: ",,0.svc" No such <nametag> registered" , and I think this is because Iam running it outside the cluster as they do not have access to pmadm process. However, it starts the daemon , but I do not know if that will start PMF monitoring
    4>So , how can I start the process outside the cluster
    5>Also ,if I start the process outside the cluster , will it start the PMF monitoring
    thanks
    Dajiba
    Edited by: Cluster_newbie on Jul 29, 2008 11:06 AM
    Edited by: Cluster_newbie on Jul 29, 2008 11:08 AM

    thanks for the answer , Thats good then. Ok , now why I need to run it outside the SC
    1>Ours is a Egate envirnoment and the process that gets started by my START_SCRIPT is a boot strap kind of a daemon process. This process in turn launches multiple TCPIP , ODBC , FTP process and they run independently once they are launched.
    2>The daemon process is what we are doing a probe on. This daemon process sometimes can go down and more or less it is normal , however the child process still continues to run. In non cluster envirnoment , we just start the daemon again and all works fine
    3>However, if I take a restart option from Cluster when the daemon dies , It will first STOP the data service(stop all the child process) , then start it . My requirement is that I should NOT stop the Dataservice because stoping the DS will result in killing of child processes too My requirement is just to launch the START_SCRIPT so that only daemon is started. If the child process are still running , the daemon will not launch any duplicate child processes
    I can make Stop script more intellegent , but it is very difficult to know in what situation the Stop was called from. If I know it was from RESTART option , I can then ignore killing the child process. But I guess it is very difficult to know
    4>So , I thought , if the Daemon goes down , I will start the daemon outside the SC and my probe script will still do its job . I have anyway disabled the PMF monitoring , so I do not care about the PMF
    does it make sense?
    Do you know any other way of starting the start_script withing SC , without doing shutdown and start
    thanks
    Dajiba

Maybe you are looking for

  • I have a macbook air and since the last update, I cannot find camera. Ideas?

    I have a mac book air and since the last update, I cannot find my camera.

  • LSMW- line items thru bdc recording

    I m creating a lsmw thru bdc recodring for me21. the prob is ...how to upload the line items ? if i do in general it will only upload one line item for order....help  me in solving this.

  • Unknown error when trying to sync from itunes to ipod

    Does anyone know what the problem is? Everytime I try to sync music onto my 4GB ipod nano, a message pops up saying: Attempting to copy to the disk [ipod name] failed. An unknown error occured (-54). I can't seem to figure out what is wrong. I restor

  • Streams questions

    Oracle 11.1.7.0: Couple of quesitons I have: 1. When we remove streams configuration, oracle leaves behind DBA_APPLY_TABLE_COLUMNS contents. Why doesn't it remove it? 2. If we build index on source DB does the same index get build on destination DB a

  • No financial data on item

    Dear community, we are using PPM PFM items linked to PPM PRO with accounting integration to PS (Single-Obj. Controlling on WBS Elements / Multilevel Controlling - Role manual). Financial data like Budget, Planned costs and Actuals are correctly trans