CLUSTER  HELP RQD

i am confused with cluster... after creating cluster.. how to query..? physically it stores the values of joining records. but how to view it..? pls help
create cluster emp_dept(deptno number(10));
CREATE TABLE dept1
    deptno NUMBER(10)   ,
    name   VARCHAR2(20),
    loc    VARCHAR2(20)
  cluster emp_dept
    deptno
create index IDX on cluster emp_dept;
drop table dept1
CREATE TABLE emp1
    empno  NUMBER(5)   ,
    name   VARCHAR2(20),
    sal    NUMBER(10,2),
    deptno NUMBER(10)
  cluster emp_dept
    deptno
  )SELECT * FROM EMP_DEPT ( Table or view does not exists error)
also help me what is cluster id here..
select  * from tab  where tname like 'EMP%'  OR  TNAME LIKE 'DEPT%'S

You dont query the cluster you just query the related tables joining them by the cluster key.
Cluster is a method of storage where data from different table are stored in the same block based on the cluster key.
you should try a query like the below
select e.empno, e.name, d.deptno, d.name
  from emp e
  join dept d
    on e.deptno = d.deptnoAnd if you check the explain plan then oracle would have accessed the cluster.

Similar Messages

  • Flex session replication among cluster Help needed!

    Hi,
    I'am using Flex 3, LiceCycle Data Service 2.5.1, JBoss 4.2.1 in a web application. Our JBoss servers are in a cluster. We want to replicate Flex session among cluster, but it doesn't work. Did someone face issue with Flex session replication among application servers?
    Thank you for your help
    Gerard

    Gerard,<br /><br />You will need to cluster the LCDS 2.5.1 instances the same way you clustered <br />JBoss.<br /><br />Check out the "Configuring Data Services" chapter in the Admin guide:<br />http://livedocs.adobe.com/livecycle/es/sdkHelp/programmer/lcds/services_config_4.html<b r /><br />--<br />Tom Jordahl<br />Adobe<br /><br /><br /><[email protected]> wrote<br />> Hi,<br />><br />> I'am using Flex 3, LiceCycle Data Service 2.5.1, JBoss 4.2.1 in a web <br />> application. Our JBoss servers are in a cluster. We want to replicate Flex <br />> session among cluster, but it doesn't work. Did someone face issue with <br />> Flex session replication among application servers?<br />> Thank you for your help<br />><br />> Gerard

  • Cisco MXE 3500 Cluster Help

    Hello
    i need help for my Cisco MXE cluster . I have already installed RM license and RN licesnses and all are installed successfully. I need to know what are the next steps?. I have 1 X RM and 4 X RM , how can i add my 4 X RN to my cluster to can see RM successfully?.
    Thank you

    Hello Mohamed,
    The main step needed after licenses are properly applied is to add each RN as a host in the RM GUI Administration.
    You can actually follow the full MXE configuration tasks as documented at:
    http://www.cisco.com/en/US/docs/video/mxe/3500/sw/3_x/3_3/admin/guide/tasks.html
    This document is also valid for standalone deployments, but you  will notice that any extra step or different step needed in clustered  deployment will be added in a note starting like:
    "Note In a clustered deployment..."
    I hope this helps,
    Marco

  • Cluster help

    I downloaded Sun Cluster 3.2 for x86 and sparc. My intention is to learn this technology. What programs are available for free download to test and increase knowledge of clustering? I am looking for applications that I can use to test and learn upon. May I safely use a mix of x86 and sparc? I guess the application would have to be processor independent.
    If processor specific applications are more easily available for free/trial downloads or may be better to learn from, then let me know which those applications are, because I have more than one sparc based systems.
    Message was edited by:
    kartikvashishta

    Hi Kartikvashishta.
    You can use the following link for docs. http://docs.sun.com/app/docs/prod/sun.cluster32
    You can read the SC official blog in http://blogs.sun.com/sc.
    You cannot install cluster with a mix of x86 and SPARC. Instead try using single node cluster if you don't have 2 machines of the same architecture.
    You can use apache and nfs for a start and need not be downloaded. Kindly refer to the data services docs under sun cluster for more details.
    Best Regards,
    Madhan Kumar

  • How cluster help perofrmance?

    hello experts,
    below is code for cluster creation:
    CREATE CLUSTER personnel
    (department NUMBER(4))
    SIZE 512
    STORAGE (initial 100K next 50K);
    CREATE INDEX idx_personnel ON CLUSTER personnel;
    CREATE TABLE dept_10
    CLUSTER personnel (department_id)
    AS SELECT * FROM employees WHERE department_id = 10;
    CREATE TABLE dept_20
    CLUSTER personnel (department_id)
    AS SELECT * FROM employees WHERE department_id = 20;
    1. personnel cluster contain data of department 10 and 20. if we want to add new records in department 10 and 20
    then in which table we insert the record whether in table employees or in dept_10 and dept_20;
    2. if i want to fetch records with select sql statement then which table should we query for faster performance.
    thanks
    yash

    >
    1. personnel cluster contain data of department 10 and 20. if we want to add new records in department 10 and 20
    then in which table we insert the record whether in table employees or in dept_10 and dept_20;
    2. if i want to fetch records with select sql statement then which table should we query for faster performance.
    >
    As Tubby indicated these questions show that you don't understand what clusters or when and how they are used. Your design is a 'misuse' of clusters.
    Answer to the first question is 'a table is a table'. If you want to add a new record for department 10 you can put the record in either table and Oracle won't care at all. But because you misused the cluster technology and hard-coded values in the table names you will face maintenance and data problems for the life of the system.
    There is nothing in your architecture that will prevent a user from putting a department 10 record into table dept_20.
    Answer to the second question is 'a table is a table'. If you want data from the dept_10 table then you should query it. If you want data from dept_20 then query it. If you need data for all departments then you need to query them both.
    Read the documentation that Tubby referred you to and follow the examples provided in it and you should see why the example you provided is totally wrong.

  • ORACLE IMPORT HELP RQD

    Hai
    I want to take a export dump using subset of data. consider a emp table, only group id less than 5 i have take.
    exp user_name@schema_name \
                tables=emp query=\"where emp.groupid  \<5\"    file=emp.dmp\
                         log=emp.log \
                rows=y consistent=y feedback=10000 the other table emp1 also has the same groupid with the name groupid1.
    My question is
    can i import 2 tables(emp and emp1) at a time and put into dmp file like this
    exp user_name@schema_name \
                tables=\(emp,emp1\) query=\"where emp.groupid  \<5\"    query=\"where emp1.groupid1  \<5\"  file=emp.dmp\
                         log=emp.log \
                rows=y consistent=y feedback=10000 But this onne throwing error LRM-00112: multiple values not allowed for parameter 'query'
    Kindly give your sugestion
    S

    Hai
    i have done by using datapump
    expdp sen/hema@xe tables=DEPT,DEPT1 directory=expdp_dir dumpfile=EMP_DEPT.dmp logfile=expdpEMP_DEPT.log query=dept:\"where deptno=10\" dept1:\"where deptno1=20\"
    this is working.
    thanks.
    kindly confirm me whether it is possible to work in oracle 9i..? oracle 9i will support datapump..?
    Waiting
    S

  • Help! Build Standby DB from RAC to RAC

    I have had RAC 2 nodes. i want to migrant2 because some problem.
    so i think create standby DB , i want RAC 2 nodes(standby)
    help me, How i do?
    and if i created standby from init file and backup rman complete, and switch standby db to
    primary , i don't know what i find any problems?
    Because when i switch to primary ,
    How i register to Cluster
    help me,please

    http://www.oracle.com/technology/deploy/availability/pdf/MAA_WP_10g_RACPrimaryRACStandby.pdf

  • Remove KF value '*' from report

    Hi,
    In report * is coming in KF. I need to restrict those records which has KF value as '*' .
    Anybody suggest how to remove this.
    Points assured....
    Thanks,
    Debasish

    Hi Debasish,
    You may be interested to see the following threads ..
    Help Rqd Urgent: Remove star * from query
    * as result in query
    How to remove Star (*) Symbol
    Reporting data in alternate unit of measure
    Regards.

  • How to remove an SC authorization ?

    Hi,
    I want that ordinary user doesn't perform "list,show,status" operations.
    I know that there is "solaris.cluster.read" authorization, but it is included in "Basic Solaris User" profile. How can I remove it for particular user ?
    Thanks

    HI jjj,
    This is basically a Solaris RBAC Authorization management question. You are right that
    by default all cluster "read" authorizations are granted to all Solaris users. Taking away
    of individual authorizations (as you want to do), is not very straight-forward because
    RBAC is basically designed with the philosophy of "Start with the minimum and ADD
    what additional authorizations you need".
    That makes it a bit round about to "take away" Cluster read authorizations from
    basic Solaris users. Here is a quick and dirty procedure i use, you might wanna
    think about this a little more carefully (particularly if you are using NIS or other network
    based naming services).
    1) Create a new profile in /etc/security/prof_attr which grants all regular Solaris authorizations
    except the Sun Cluster related ones. eg
    NoSC:::Execute any command as the user or role except Sun CLuster:help=RtNoSC.html
    To define the set of authorizations this profile has, just copy the "Basic Solaris User" profile
    a little later in that file and remove all authorizations related to SC in there. Name it "Solaris Non Cluster User". Make sure the profiles keyword is NoSC. Do read man prof_attr(4).
    2) In /etc/security/policy.conf, change the line
    PROFS_GRANTED=Basic Solaris User
    To
    PROFS_GRANTED=Solaris Non Cluster User
    You are DONE!! Now all non-root users on your system cannot run any cluster commands, not
    even the commands to view things.
    3) For specific users, if you wanna give back cluster viewing priviledges, do, as root
    # usermod -A solaris.cluster.whatever.XXX mygooduser
    etc. etc.
    Hope you get the picture. Please don't do this on a production system without first carefully
    researching this. Also, treat that above just as a starting point, come up with your own
    procedure after having gone thru appropriate Solaris man pages (start with auths(1)).
    PS: Having said all that "philosphy" bit about RBAC and what not, don't i wish "useradd -A"
    supported a "!auth" syntax (the way you can do with priviledges to REMOVE A SPECIFIC PRIVILEDGE FROM A SPECIFIC USER)?? Absolutely!! :-) :-). If you figure that out, please post.
    Well... post your experience either way.
    HTH,
    -ashu

  • Need help to Configure Cisco ACE 4710 Cluster Deployment

    Dear Experts,
    I'm newbie for Cisco ACE 4710, and still I'm in learning stage. Meanwhile I got chance at my work place to deploy a Cisco ACE 4710 cluster which should load balance the traffic between  two Application Servers based on HTTP and HTTPS traffic. So I was looking for good deployment guide in Cisco SBA knowledge base then finall found this guide.
    http://www.cisco.com/en/US/docs/solutions/SBA/February2013/Cisco_SBA_DC_AdvancedServer-LoadBalancingDeploymentGuide-Feb2013.pdf
    This guide totally fine with my required deployment model. I have same deployment environment as this guide contains with ACE cluster that connects to two Cisco 3750X (Stack) switches. But I have some confusion places in this guide
    This guide follow the "One-armed mode" as a deployment method. But when I go through it further I have noticed that they have configured server VLAN as a 10.4.49.0/24 (all servers reside in it) and Client side VIP also in same VLAN which is 10.4.49.100/24 (even NAT pool also).
    My confusion is, as I have learned about Cisco ACE 4710 one-armed mode deployment method, it should has two VLAN segments, one for Client side which client request come and hit the VIP and then second one for Server side. which means besically two VLANs. So please be kind enough to go through above document then tell me where is wrong, what shoud I need to do for the best. Please this is an urgent, so need your help quickly.
    Thanks....!
    -Amal-

    Dear Kanwal,
    I need quick help for you. Following are the Application LB requirements which I received from my clinet side.
    Following detail required for configuring Oracle EBS Apps tier on HA:
    LBR IP and Name required to configure EBS APPS Tier (i.e, ap1ebs & ap2ebs nodes)
    Suggested IP and Name for LBR:
    IP : 172.25.45.x [should be on same 172.25.45 subnet of ap1ebs & ap2ebs nodes]
    ebiz.xxxx.lk [on port 80 for http protocol accessibility]
    This LBR IP & name must be resolve and respond on DNS network
    Server Farm detail for LBR Setup
    Following detail will be use for configuring the LBR:
    LBR IP and Name :
    IP : 172.25.45.x [should be on same 172.25.45 subnet of ap1ebs & ap2ebs nodes]
    ebiz.xxxx.lk [on port 80 for http protocol accessibility]
    This LBR IP & name must be resolve and respond on DNS network
    Server Farm Detail for LBR setup:
    Server 1 (EBS App1 Node, ap1ebs):
    IP : 172.25.45.19
    Server Name: ap1ebs.xxxx.lk [ap1ebs hostname is an example, actual hostname will be use]
    Protocol: http
    Port: 8000
    Server 2 (EBS App2 Node, ap2ebs):
    IP : 172.25.45.20
    Server Name: ap2ebs.xxxx.lk [ap2ebs hostname is an example, actual hostname will be use]
    Protocol: http
    Port: 8000
    Since my client needs to access URL ebiz.xxxx.lk which should be resolved by IP 172.25.45.21 (virtual IP) via http (80) before they deploy the app on the two servers I just ran web service on both servers (Linux) and was trying to access http://172.25.45.21 it was working fine and gave me index.html page. Now after my client has deployed the application then when he tries to access the page http://172.25.45.21 he cannot see his main login page. But still my testing web servers are there on both servers when I type http://172.25.45.21 it will get index.html page, but not my client web login page. What can I do for this ?
    Following are my latest config :
    probe http Get-Method
      description Check to url access /OA_HTML/OAInfo.jsp
      interval 10
      faildetect 2
      passdetect interval 30
      request method get url /OA_HTML/OAInfo.jsp
      expect status 200 200
    probe udp http-8000-iRDMI
      description IRDMI (HTTP - 8000)
      port 8000
    probe http http-probe
      description HTTP Probes
      interval 10
      faildetect 2
      passdetect interval 30
      passdetect count 2
      request method get url /index.html
      expect status 200 200
    probe https https-probe
      description HTTPS traffic
      interval 10
      faildetect 2
      passdetect interval 30
      passdetect count 2
      ssl version all
      request method get url /index.html
    probe icmp icmp-probe
      description ICMP PROBE FOR TO CHECK ICMP SERVICE
    rserver host ebsapp1
      description ebsapp1.xxxx.lk
      ip address 172.25.45.19
      conn-limit max 4000000 min 4000000
      probe icmp-probe
      probe http-probe
      inservice
    rserver host ebsapp2
      description ebsapp2.xxxx.lk
      ip address 172.25.45.20
      conn-limit max 4000000 min 4000000
      probe icmp-probe
      probe http-probe
      inservice
    serverfarm host ebsppsvrfarm
      description ebsapp server farm
      failaction purge
      predictor response app-req-to-resp samples 4
      probe http-probe
      probe icmp-probe
      inband-health check log 5 reset 500
      retcode 404 404 check log 1 reset 3
      rserver ebsapp1 80
        conn-limit max 4000000 min 4000000
        probe icmp-probe
        inservice
      rserver ebsapp2 80
        conn-limit max 4000000 min 4000000
        probe icmp-probe
        inservice
    sticky http-cookie jsessionid HTTP-COOKIE
      cookie insert browser-expire
      replicate sticky
      serverfarm ebsppsvrfarm
    class-map type http loadbalance match-any default-compression-exclusion-mime-type
      description DM generated classmap for default LB compression exclusion mime types.
      2 match http url .*gif
      3 match http url .*css
      4 match http url .*js
      5 match http url .*class
      6 match http url .*jar
      7 match http url .*cab
      8 match http url .*txt
      9 match http url .*ps
      10 match http url .*vbs
      11 match http url .*xsl
      12 match http url .*xml
      13 match http url .*pdf
      14 match http url .*swf
      15 match http url .*jpg
      16 match http url .*jpeg
      17 match http url .*jpe
      18 match http url .*png
    class-map match-all ebsapp-vip
      2 match virtual-address 172.25.45.21 tcp eq www
    class-map type management match-any remote_access
      2 match protocol xml-https any
      3 match protocol icmp any
      4 match protocol telnet any
      5 match protocol ssh any
      6 match protocol http any
      7 match protocol https any
      8 match protocol snmp any
    policy-map type management first-match remote_mgmt_allow_policy
      class remote_access
        permit
    policy-map type loadbalance first-match ebsapp-vip-l7slb
      class default-compression-exclusion-mime-type
        serverfarm ebsppsvrfarm
      class class-default
        compress default-method deflate
        sticky-serverfarm HTTP-COOKIE
    policy-map multi-match int455
      class ebsapp-vip
        loadbalance vip inservice
        loadbalance policy ebsapp-vip-l7slb
        loadbalance vip icmp-reply active
        nat dynamic 1 vlan 455
    interface vlan 455
      ip address 172.25.45.36 255.255.255.0
      peer ip address 172.25.45.35 255.255.255.0
      access-group input ALL
      nat-pool 1 172.25.45.22 172.25.45.22 netmask 255.255.255.0 pat
      service-policy input remote_mgmt_allow_policy
      service-policy input int455
      no shutdown
    ft interface vlan 999
      ip address 10.1.1.1 255.255.255.0
      peer ip address 10.1.1.2 255.255.255.0
      no shutdown
    ft peer 1
      heartbeat interval 300
      heartbeat count 10
      ft-interface vlan 999
    ft group 1
      peer 1
      no preempt
      priority 110
      associate-context Admin
      inservice
    ip route 0.0.0.0 0.0.0.0 172.25.45.1
    Hope you will reply me soon
    Thanks....!
    -Amal-

  • F4 help for a field in item table in a view cluster

    Hi All,
    I have a view cluster with two tables. I should provide F4 help for field in item table based on the selected row of the header table.
    F4 should have only values fetched based on the header table entry.
    How can i achieve this. Please suggest.
    Regards,
    Kusuma K.

    Hi amar_war,
    use screen painter to edit the screen layout where the field VKBUR is on. Double click the field to get the details screen. Here you can attach a search help to the field.
    H_TVBUR has no more parameters, it will display a list of al sales offices with description text
    H_TVBVK will also show the sales group, if VKGRP is a screen field, it will restrict the result list automatically on matching VKGRP.
    H_TVKBZ has also VKORG VTWEG SPART - it will take those values from screen if filled.
    If you are not satisfied with the results, you can easily create your own search help using own selection method and own search help exits if required.
    Try standard search help for standard field first.
    You can create a copy of VKBUR as ZVKBUR and attach search help there, you can create your own screen structure and attach the search help there - 50 ways to get your search help...
    Regards,
    Clemens

  • Helps for WLP9.2 in a cluster:  deploy errors, propagation errors

    I wrote a document for our portal project's production support that details how to fix several recuring problems with WLP9.2 (GA) in cluster. I'm sharing this with the community as a help.
    Email me if you'd like updates or have additional scenarios / fixes or you'd like the word doc version.
    Curt Smith, Atlanta WLP consultant, [email protected], put HELP WLP as a part of the subject please.
    1     WLP 9.2 Environment
    The cluster environment where these symptoms are frequently seen is described by the following:
    1.     Solaris 9 on 4 cpu sparc boxes. 16Gb Ram. The VM is given ?Xmx1034m.
    2.     Bea private patch: RSGT, fixed cluster deployments.
    3.     Bea private patch: BG74, fixed session affinity, changed the session tracking cookie name back to the standard name: JSESSIONID.
    4.     IBM UDB (DB2) using the Bea DB2 driver. FYI: the problems described below I don?t feel have any relationship to the DB used for a portal.
    2     Common problems and their fixes
    2.1     Failed deploy (Install) of a new portal application via the cluster console.
    Problem symptoms:
    After clicking Finish (or) Commit the console displays that there where errors. All errors require this procedure.
    Fix steps:
    1.     Shut down the whole cluster. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Admin server:
    a.     Remove file: <domain>/config/config.lok
    b.     Edit file: <domain>/config/config.xml
    c.     Remove all elements of: <app-deployment> ? </app-deployment>
    d.     Start the Admin server
    e.     Make sure it comes up in the running state. Use the console: servers ? admin ? control - resume if needed.
    3.     Both Managed servers:
    a.     cd <domain>/config
    b.     rm ?rf *
    This forces the re-down load of the cleaned up config.xml.
    c.     cd <domain>/servers/<managed_name>
    d.     rm ?rf tmp cache stage data
    This cleans up stale or jammed sideways deploys. Be sure to delete all three directorys: tmp, cache, stage and data.
    e.     Start each managed server.
    f.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    You can now use the console to Install your applications.
    2.2     The portal throws framework / container exceptions on one managed instance.
    Problem symptoms:
    If you see exceptions from the classloader re a framework class not found, serialization failure or error etc. In general the symptom is that the container is not stable, running correctly, not making sense or your application works on one managed instance but not on the other.
    Fix steps:
    1.     Shut down the problem managed instance. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Perform these clean up steps on one or both managed instances:
    a.     cd <domain>/config
    b.     rm ?rf *
    This forces the re-down load of the cleaned up config.xml.
    c.     cd <domain>/servers/<managed_name>
    d.     rm ?rf tmp cache stage data
    This cleans up stale or jammed sideways deploys. Be sure to delete all three directorys: tmp, cache, stage and data.
    e.     Start each managed server.
    f.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    3.     The libraries and applications should auto deploy as the managed instance comes up. Once the managed instance goes into the running state, or you Resume into the running state. Your application should be accessible. Sometimes it takes a few seconds after going into the running state for all applications to be instantiated.
    2.3     Content propagation fails on the commit step.
    Problem symptoms:
    In the log of the managed instance you specified in the propagation ant script you?ll see exceptions regarding not being able to create or instantiate a dynamic delegated role.
    There is an underlying bug / robustness issue with WLP9.2 (GA) where periodically you can?t create delegated roles either with the PortalAdmin or via the propagation utility.
    Important issue:
    This procedure was supplied by Bea which will remove from the internal LDAP and the portal DB your custom / created roles. This will leave your cluster in the new installation state with just the default users and roles: weblogic and portaladmin. The implications are that you?ll have to boot your cluster with console user: weblogic / weblogic. You can then add back your secure console user/password but you?ll have to do this over and over as propagations fail. The observed failure rate is once every 2-3 weeks if you do propagations daily.
    Note:
    The following assumes that you left the default console user weblogic password to be the default password of: weblogic. The following procedure deletes the local LDAP but leaves the rows in the DB.users table including the SHA-1 hashed passwords. The following procedure should still work if you changed the password for weblogic, but it probably won?t work if you try to substitute your secure console user/pw because there will be no delegated authorization roles mapping to your custom console user. You might experiment with this scenario.
    Fix steps:
    1.     Shut down the whole cluster. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Admin server:
    a.     Remove directory: <domain>/servers/AdminServer/data/ldap
    b.     Run this SQL script after you edit it for your schema:
    delete from yourschema.P13N_DELEGATED_HIERARCHY;
    delete from yourschema.P13N_ENTITLEMENT_POLICY;
    delete from yourschema.P13N_ENTITLEMENT_RESOURCE;
    delete from yourschema.P13N_ENTITLEMENT_ROLE;
    delete from yourschema.P13N_ENTITLEMENT_APPLICATION;
    commit;
    c.     Start the Admin server
    d.     Make sure it comes up in the running state. Use the console: servers ? admin ? control - resume if needed.
    3.     Both Managed servers:
    a.     cd <domain>/servers/<managed_name>
    b.     rm ?rf data
    This forces the re-down load of the LDAP directory.
    c.     Start each managed server.
    d.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    2.4     The enterprise portal DB fails and needs to be restored OR switch DB instances
    Restoring a portal DB or switching existing DBs are similar scenarios. The issues that you?ll face with WLP9.2 since it now uses a JDBCauthenticator to authenticate and authorize the console / boot user you need to first be able to connect to the DB before the admin and managed instances can boot. If you haven?t properly encrypted the DB user?s password in the <domain>/config/jdbc/*.xml files, then you?ll not be able to boot the admin server since you won?t be able to create a JDBC connection to the DB. The boot messages are not clear as to what the failure is.
    You?ll need to know an Admin role user and password that?s in the DB you?re wanting to connect to, to put into boot.properties and on the managed instances in their boot.properties or startManaged scripts. Don?t forget that the managed instances have local credentials which is new for 9.2. They are in the startManaged script in clear text or a local boot.properties.
    Note:
    The passwords in the DB are SHA-1 hashed and there is no SHA-1 hash generator tool so you can?t change a password via SQL, but you can move the password from one DB to another. This is possible because the domain encryption salt is not used to generate the SHA-1 string. As it turns out, the SHA-1 string is compatible with all 9.2 cluster domains. IE the domain DES3 salt has nothing to do with password verification. The same password SHA-1 string taken from different domains or even same domain but different users will be different, this is just a randomization put into the algorithm yet every domain will be able to validate the given password against the DB?s SHA-1 string. Because of this, I?ve not had any problem moving DB instances between clusters, especially if I?ve given up on security and use weblogic/weblogic as the console user and this user / pw is in every DB.
    Steps:
    The assumption for restoring a DB is that the DB has been restored but it?s an older version and doesn?t have the console user/pw that is in boot.properties. At this point swaping a DB is the same as restoring an old version of the portal DB.
    1.     Edit the console user and password as clear text into: <domain>/servers/AdminServers/security/boot.properties
    This is where you may give in and use weblogic/weblogic.
    2.     Set the correct DB access password encryption in the jdbc/*.xml files.
    a.     cd <domain>/bin
    b.     . ./setDomainEnv.sh or if you?re on windows just run: setDomainEnv.cmd
    c.     java weblogic.security.Encrypt <the_db_password>
    d.     Edit the returned string into every <domain>/config/jdbc/*.xml
    e.     Make sure the *.xml files point to the correct DB host, port, schema, DB name, DB user.
    3.     Start the Admin server. I should come up. If it doesn?t it has to be not being able to create a connection to the DB, which depends on the *.xml having the correct user and DES-3 encrypted password.
    4.     Edit the new console user/password on the managed instance bin/startManaged script or the local boot.properties.
    5.     Start the managed instances.
    2.5     Install patches on a host that does not have internet access
    The short description is to run smart update on a host that does have internet access. Fetch out of the <domain>/utils/bsu/cache_dir     directory the downloaded patch jar and xml. Manually apply the patch to your non-internet accessible hosts.

    I wrote a document for our portal project's production support that details how to fix several recuring problems with WLP9.2 (GA) in cluster. I'm sharing this with the community as a help.
    Email me if you'd like updates or have additional scenarios / fixes or you'd like the word doc version.
    Curt Smith, Atlanta WLP consultant, [email protected], put HELP WLP as a part of the subject please.
    1     WLP 9.2 Environment
    The cluster environment where these symptoms are frequently seen is described by the following:
    1.     Solaris 9 on 4 cpu sparc boxes. 16Gb Ram. The VM is given ?Xmx1034m.
    2.     Bea private patch: RSGT, fixed cluster deployments.
    3.     Bea private patch: BG74, fixed session affinity, changed the session tracking cookie name back to the standard name: JSESSIONID.
    4.     IBM UDB (DB2) using the Bea DB2 driver. FYI: the problems described below I don?t feel have any relationship to the DB used for a portal.
    2     Common problems and their fixes
    2.1     Failed deploy (Install) of a new portal application via the cluster console.
    Problem symptoms:
    After clicking Finish (or) Commit the console displays that there where errors. All errors require this procedure.
    Fix steps:
    1.     Shut down the whole cluster. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Admin server:
    a.     Remove file: <domain>/config/config.lok
    b.     Edit file: <domain>/config/config.xml
    c.     Remove all elements of: <app-deployment> ? </app-deployment>
    d.     Start the Admin server
    e.     Make sure it comes up in the running state. Use the console: servers ? admin ? control - resume if needed.
    3.     Both Managed servers:
    a.     cd <domain>/config
    b.     rm ?rf *
    This forces the re-down load of the cleaned up config.xml.
    c.     cd <domain>/servers/<managed_name>
    d.     rm ?rf tmp cache stage data
    This cleans up stale or jammed sideways deploys. Be sure to delete all three directorys: tmp, cache, stage and data.
    e.     Start each managed server.
    f.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    You can now use the console to Install your applications.
    2.2     The portal throws framework / container exceptions on one managed instance.
    Problem symptoms:
    If you see exceptions from the classloader re a framework class not found, serialization failure or error etc. In general the symptom is that the container is not stable, running correctly, not making sense or your application works on one managed instance but not on the other.
    Fix steps:
    1.     Shut down the problem managed instance. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Perform these clean up steps on one or both managed instances:
    a.     cd <domain>/config
    b.     rm ?rf *
    This forces the re-down load of the cleaned up config.xml.
    c.     cd <domain>/servers/<managed_name>
    d.     rm ?rf tmp cache stage data
    This cleans up stale or jammed sideways deploys. Be sure to delete all three directorys: tmp, cache, stage and data.
    e.     Start each managed server.
    f.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    3.     The libraries and applications should auto deploy as the managed instance comes up. Once the managed instance goes into the running state, or you Resume into the running state. Your application should be accessible. Sometimes it takes a few seconds after going into the running state for all applications to be instantiated.
    2.3     Content propagation fails on the commit step.
    Problem symptoms:
    In the log of the managed instance you specified in the propagation ant script you?ll see exceptions regarding not being able to create or instantiate a dynamic delegated role.
    There is an underlying bug / robustness issue with WLP9.2 (GA) where periodically you can?t create delegated roles either with the PortalAdmin or via the propagation utility.
    Important issue:
    This procedure was supplied by Bea which will remove from the internal LDAP and the portal DB your custom / created roles. This will leave your cluster in the new installation state with just the default users and roles: weblogic and portaladmin. The implications are that you?ll have to boot your cluster with console user: weblogic / weblogic. You can then add back your secure console user/password but you?ll have to do this over and over as propagations fail. The observed failure rate is once every 2-3 weeks if you do propagations daily.
    Note:
    The following assumes that you left the default console user weblogic password to be the default password of: weblogic. The following procedure deletes the local LDAP but leaves the rows in the DB.users table including the SHA-1 hashed passwords. The following procedure should still work if you changed the password for weblogic, but it probably won?t work if you try to substitute your secure console user/pw because there will be no delegated authorization roles mapping to your custom console user. You might experiment with this scenario.
    Fix steps:
    1.     Shut down the whole cluster. Using Bea?s shutdown script takes too long, or doesn?t work if the Admin is down or hung.
    a.     Find the PID to kill: lsof ?C | grep <listen port number>
    b.     kill <pid>
    c.     Run kill again and if the pid is still running then do: kill -9 <pid>
    2.     Admin server:
    a.     Remove directory: <domain>/servers/AdminServer/data/ldap
    b.     Run this SQL script after you edit it for your schema:
    delete from yourschema.P13N_DELEGATED_HIERARCHY;
    delete from yourschema.P13N_ENTITLEMENT_POLICY;
    delete from yourschema.P13N_ENTITLEMENT_RESOURCE;
    delete from yourschema.P13N_ENTITLEMENT_ROLE;
    delete from yourschema.P13N_ENTITLEMENT_APPLICATION;
    commit;
    c.     Start the Admin server
    d.     Make sure it comes up in the running state. Use the console: servers ? admin ? control - resume if needed.
    3.     Both Managed servers:
    a.     cd <domain>/servers/<managed_name>
    b.     rm ?rf data
    This forces the re-down load of the LDAP directory.
    c.     Start each managed server.
    d.     Make sure the managed instance comes up in the running state and not admin. Go to server-<instance>-control-Resume to set the run state to running.
    2.4     The enterprise portal DB fails and needs to be restored OR switch DB instances
    Restoring a portal DB or switching existing DBs are similar scenarios. The issues that you?ll face with WLP9.2 since it now uses a JDBCauthenticator to authenticate and authorize the console / boot user you need to first be able to connect to the DB before the admin and managed instances can boot. If you haven?t properly encrypted the DB user?s password in the <domain>/config/jdbc/*.xml files, then you?ll not be able to boot the admin server since you won?t be able to create a JDBC connection to the DB. The boot messages are not clear as to what the failure is.
    You?ll need to know an Admin role user and password that?s in the DB you?re wanting to connect to, to put into boot.properties and on the managed instances in their boot.properties or startManaged scripts. Don?t forget that the managed instances have local credentials which is new for 9.2. They are in the startManaged script in clear text or a local boot.properties.
    Note:
    The passwords in the DB are SHA-1 hashed and there is no SHA-1 hash generator tool so you can?t change a password via SQL, but you can move the password from one DB to another. This is possible because the domain encryption salt is not used to generate the SHA-1 string. As it turns out, the SHA-1 string is compatible with all 9.2 cluster domains. IE the domain DES3 salt has nothing to do with password verification. The same password SHA-1 string taken from different domains or even same domain but different users will be different, this is just a randomization put into the algorithm yet every domain will be able to validate the given password against the DB?s SHA-1 string. Because of this, I?ve not had any problem moving DB instances between clusters, especially if I?ve given up on security and use weblogic/weblogic as the console user and this user / pw is in every DB.
    Steps:
    The assumption for restoring a DB is that the DB has been restored but it?s an older version and doesn?t have the console user/pw that is in boot.properties. At this point swaping a DB is the same as restoring an old version of the portal DB.
    1.     Edit the console user and password as clear text into: <domain>/servers/AdminServers/security/boot.properties
    This is where you may give in and use weblogic/weblogic.
    2.     Set the correct DB access password encryption in the jdbc/*.xml files.
    a.     cd <domain>/bin
    b.     . ./setDomainEnv.sh or if you?re on windows just run: setDomainEnv.cmd
    c.     java weblogic.security.Encrypt <the_db_password>
    d.     Edit the returned string into every <domain>/config/jdbc/*.xml
    e.     Make sure the *.xml files point to the correct DB host, port, schema, DB name, DB user.
    3.     Start the Admin server. I should come up. If it doesn?t it has to be not being able to create a connection to the DB, which depends on the *.xml having the correct user and DES-3 encrypted password.
    4.     Edit the new console user/password on the managed instance bin/startManaged script or the local boot.properties.
    5.     Start the managed instances.
    2.5     Install patches on a host that does not have internet access
    The short description is to run smart update on a host that does have internet access. Fetch out of the <domain>/utils/bsu/cache_dir     directory the downloaded patch jar and xml. Manually apply the patch to your non-internet accessible hosts.

  • Can anyone help me?Creating the shadow copies in the file server cluster ,there are some errors occured, OS version is WSS 2012

    I construct a failover cluster(file server,AP module) for sharing files by WSS 2012,and I want to use the shadow copies to backup my data,but when making  a shadow copies on the volume which  is added to the cluster(not the CSV,just add
    it to the cluster and use it to share files,it plays the role of file server),there are some errors occured, these errors result in the shadow copies failed,error likes the following pictures:
    1: the disk F is added to the cluster,first I make the shadow copies by click the right key of mouse on the disk F,and chouse the configeration shadow copies,and click the settings, then click the schedule , and I wait just a few seconds, the error is appeared,like
    this picture 1, the wait operation timed out,and then ,
    I click the schedule button once again,a different error occured,like the following picture," the object already exists",if i don't set the schedule at first ,use the default shedule,and click the enable button also the same  error must 
    accure
    but the only diffrence is that, a shadow copy time point is created, also ,you can make the shadow copies by click " create now", that is creating the shadow copies manually. Although it can succesfully make the shadow copies, but when I select
    a time point to revert, error  occured, "A volume that contains operating system files or resides on a cluster shared disk cannot be reverted" 
    In a word,all the errors above make the shadow copies by schedulling failed,except making the shadow copies manually,but what makes me confused is that I have ever maked the shadow copies succesfully by schedul a policy,I don't know what makes it succesful,
    it's small probability, most of time ,it's failed.No matter what kind of situation, revert must be failed.
    I'm sorry for my pool english , it's the first time for me getting help in forum by english ,I don't know if I descripe my question clearly, also ,other method like the link
    http://technet.microsoft.com/en-us/library/cc784118(v=ws.10).aspx I have tried,but the same errors occured.Can anyone tell me How can I make the shadow copies in File Server
    cluster(AP module)?And I make a mistake in operating? Looking forward for your reply.Thanks!

    Hi,
    Please check the following 2 places:
    HKEY_LOCAL_MACHINE\Cluster\Tasks
    C:\Windows\System32\Tasks
    First please compare permission settings of the folder C:\Windows\System32\Tasks with a working computer. Correct permission settings if anything wrong. Specifically, confirm you current account do have permission on this folder.
    As it said "object already exists", find the schedules you created before, backup and delete all these schedules in both registry key and folder.
    Then test to create a new schedule to see if issue still exists.
    Meanwhile what kind of storage device you are using? The issue could occur on specific storage device, so test to enable shadow copy on a local disk to see if that will work.
    TechNet Subscriber Support in forum |If you have any feedback on our support, please contact [email protected]
    Thank you for you reply.On the local volume,all of these errors will not occur, but the volume in the file server cluser.There is no value in HKEY_LOCAL_MACHINE\Cluster\Tasks. On local volume, everything goes well about the shadow copy, so I do not
    think something is wrong about the permission settings of the folder C:\Windows\System32\Tasks.Storage device  is a SAN,we use RAID6 and provide the LUNs to the NAS engine, and the make the volume on these LUNs, Is Angthing wrong? Hope for you
    reply~~

  • Need Help Setting Up Cluster

    I'm a grad student in a lab that does a lot of bioinformatics computing. I am a bit more computer oriented than some of the others in the lab, and the PI (head professor) and another lab worker have turned to me to see if I can help them.
    They have told me that they followed the directions in the Apple manual for setting up the High Performance Computing and that they've gotten stuck at the same point after several attempts. They called a mac tech, but he only got them part of the way through and may not have used the correct settings on the changes he made.
    From what they've told me there is an external domain name that is used to reach the head node, which is also acting as a router, controller, open directory master, etc for all the compute nodes, and that the compute nodes are supposed to be on a private network with names in the form of stuff.cluster.private.
    First thing: I don't think the DNS settings are correct as typing in the recommended console commands to test the DNS setup (in the manual) resulted in a few weird errors.
    On the controller server, command "host cluster" returns the following error:
    Host cluster not found: 3(NXDOMAIN), though host cluster.private works, and the reverse host 192.168.0.1 works too.
    hostname returned the external name, which is inquiry.ccam.(my university's domain).edu
    On the node server that we have running, most of these commands result in an error:
    host: '.private' is not in legal name syntax (empty label)
    I'm not sure what this means, but it doesn't seem to be able to look up anything within the private domain.
    Secondarily, When a node is connected to an open directory master and working properly, should it's indicatory light be on or off? Right now, ours is off, but it is likely that the DNS problems are screwing up the open directory.
    Please help.

    I almost fixed the problem I was having before by adding the proper domain to the search option in system preferences -> network. However I still have an issue.
    I cannot get open directory/kerberos to work and as a result I cannot set up xgrid.
    Could this be related to the fact that the ip of the system is attached to an external hostname (which is causing the hostname that might needed by local services to be overridden). I cannot change the fact that the external hostname is going to point to the system, but I was thinking that a possible solution would be to add a server layer above the head node of the grid to act as a router and accept this IP address / hostname override and prevent it from mucking up the hostname of the head node. Is this a viable idea and is it even necessary?
    (This external domain name is a web-accessible site that is supposed to link in to a web service that will be provided by at least one of the servers on this grid setup, I'm not sure on the settings for that (it seems to work currently), but I'll worry about that when we get the rest of the system running.

  • Help with a Blind Configuration of a G5 Cluster node

    So I bought 2 G5 Cluster Nodes to dedicate some audiovisual processes to them. My only other mac computer is a Core 2 Duo Macbook Pro.
    Using Pacifist, I was able to do a clean install of Mac OSX onto the internal drive by putting it into an external enclosure.
    Now here is my problem: The cluster nodes have no videocard.
    I plan on using them through the OSX Screen Sharing function, when they will be conencted to the network, but I don't know how to do the initial configuration of Mac OS X on them, since I can not boot from a system using the Apple Partition Map on my Macbook pro, and the Cluster node will not boot from the GIUD partition scheme.
    Can anyone please help me?
    Thanks,
    Chuck

    Assuming you're running Mac OS X Server on the cluster node, just boot the server normally - it will run a special first-time-boot process that sets up a network listener.
    You can then install the Server Admin tools on your MacBook Pro and run Server Assistant. Server Assistant will look out over the network and find the new servers, then give you the opportunity to configure them remotely (assign account data, IP address, etc.).
    (note you can also do this as part of the initial install process - boot the server from the Install DVD and run the entire OS installation and configuration remotely via Server Assistant)
    Note: If you're not running Mac OS X Server on the cluster nodes then the above doesn't apply

Maybe you are looking for

  • What options do I have if I want to set up a dual screen?

    I have a Thinkpad T400 with a ATI Radeon HD3470 card. I am running the latest Arch x86_64 with Fluxbox and the 9.4 Catalyst drivers. What is the easiest solution to set up a dual screen? Is there a way to do this without editing xorg.conf? Not that e

  • "Safari Can't Find the Server" is chronic in Safari under Snow Leopard...

    Since moving to Snow Leopard (or perhaps it is just an updated Safari in SL and otherwise not related to SL??) when my Mac starts up from sleep with Safari being already open, I typically find a blank message with "Safari Can't Find the Server"... An

  • Error while adding members to a group

    Hello, We have iAS 9.0.2.1 running on windows. I am getting the following error while trying to add a member to a group created earlier : Request URI:/oiddas/cabo/jsps/frameRedirect.jsp Exception: java.io.FileNotFoundException: d:\ora9ias\apache\apac

  • IPhone announcment from AT&T

    AT&T announced today that they will start selling the iphone 3G contract free for $600-$700 depending on the model. You're still tied to AT&T as a carrier, and as of right now, the offer is limited to current AT&T customers. You also still have to si

  • IMac G5 17" won't boot with new RAM

    I tried to install 2 GB RAM into an iMac G5 17" - the computer won't boot at all. The sticks are fine (DDR PC3200 - tried them in a different computer), I reset the PRAM, all to no avail. With the original 256 MB stick, it boots fine, but with the ne