Cluster fragmentation

What happens when part of my cluster cannot communicate with the other part?
For example, what if I have two machine in a cluster and at some point the two machines can longer communicate with each other but are otherwise connected to the network and running fine.
Will they both assume the other machine has left the cluster and form two independent clusters? This can lead to problems as an update made by one machine will not be reflected in the other. Basically each clusters will become out of sync with the db as the other cluster makes updates.
Or, will one machine somehow become the sole member of “the cluster�? and the other will not be in “the cluster�? at all. If this is the case, how does the machine that has left the cluster behave? Do all cache access methods throw exceptions until it is able to rejoin the cluster or what?

Hi Rohan,
The two islands will start looking for each otherthe
next timethe application code calls into the
Coherence API (e.g. CacheFactory.getCache(...)).When
they will see each other again depends on howflaky
your server/network/switch is.I have observed that when the split occurs, if the
one server that is effectively in an island by itself
does not make calls into the Coherence API, it
remains disconnected even though the other servers in
the cluster continue to make their own calls into the
API.Exactly, there is no need for this node to be in the cluster if it is not using the cache (i.e. accessing the API).
Also, I have seen this warning:
2005-06-23 09:18:47,270 INFO [STDOUT] 2005-06-23
09:18:47.270 Tangosol Coherence 2.5/290 <Warning>
(thread=Cluster, member=3): The member formerly known
as Member(Id=1, Timestamp=Thu Jun 23 09:18:46 UTC
2005, Address=A.B.C.139, Port=8088, MachineId=2955)
has been forcefully evicted from the cluster, but
continues to emit a cluster heartbeat; henceforth,
the member will be shunned and its messages will be
ignored.
Does this mean that the shunned Member 1 will never
be able to rejoin the cluster, since its messages
will be ignored?The member will attempt to rejoin the cluster once it has (1) shut down all its Coherence services and (2) the application calls into the Coherence API again.
Log messages on Member 1:
2005-06-23 09:18:47,390 INFO [STDOUT] 2005-06-23
09:18:47.389 Tangosol Coherence 2.5/290 <Error>
(thread=Cluster, member=1): This senior Member(Id=1,
Timestamp=Thu Jun 23 08:53:10 UTC 2005,
Address=A.B.C.139, Port=8088, MachineId=2955) appears
to have been disconnected from other nodes due to a
long period of inactivity
and the seniority has been assumed by the
Member(Id=2, Timestamp=Thu Jun 23 08:53:17 UTC 2005,
Address=A.B.C.140, Port=8088, MachineId=2956);
stopping cluster.
2005-06-23 09:21:47,908 INFO [STDOUT] 2005-06-23
09:21:47.908 Tangosol Coherence 2.5/290 <Info>
(thread=Thread-48, member=1): Restarting NamedCache:
ruleCache
2005-06-23 09:21:47.908 Tangosol Coherence 2.5/290
<Info> (thread=Thread-48, member=1): Restarting
Service: DistributedCache
2005-06-23 09:21:47.908 Tangosol Coherence 2.5/290
<Info> (thread=Thread-48, member=n/a): Restarting
cluster
Member 1 did appear to be operating as normal after
this ... do these log messages mean it really was
part of the cluster again or did it just think it
was, even though it was being shunned?Correct, it is now part of the cluster again. As I stated above a shunned member will attempt to rejoin the cluster.
This is the level of fault-tolerance and reliability that is built into Coherence from the start. However, I would still suggest fixing the flaky server/network/switch.
Later,
Rob Misek
Tangosol, Inc.
Message was edited by: rmisek

Similar Messages

  • Cannot restart SAP NW04s j2ee : open cluster failed

    Hi,
    I can't start SAP j2ee, the dispatcher is in yellow light in SAPMMC.
    this is my sapstartsrv.log file:
    trc file: "sapstartsrv.log", trc level: 0, release: "700"
    pid        2028
    Thu Dec 18 10:32:26 2008
    SAP HA Trace: Build in SAP Microsoft Cluster library '700, patch 175, changelist 1007059' initialized
    Initializing SAPControl Webservice
    SapSSLInit failed => https support disabled
    Starting WebService Named Pipe thread
    Starting WebService thread
    Webservice named pipe thread started, listening on port
    .\pipe\sapcontrol_00
    Webservice thread started, listening on port 50013
    LJBW\bwdadm is starting SAP System at 2008/12/18 10:45:49
    SAP HA Trace: FindClusterResource: OpenCluster failed: 1753 [sapwinha.cpp, line 212]
    SAP HA Trace: SAP_HA_FindSAPInstance returns: SAP_HA_FAIL [sapwinha.cpp, line 907]
    LJBW\bwdadm is stopping SAP System at 2008/12/18 11:28:24
    SAP HA Trace: FindClusterResource: OpenCluster failed: 1753 [sapwinha.cpp, line 212]
    SAP HA Trace: SAP_HA_FindSAPInstance returns: SAP_HA_FAIL [sapwinha.cpp, line 907]
    LJBW\bwdadm is starting SAP System at 2008/12/18 11:32:13
    Many thanks
    George

    it says failed to update service environment from user environment.
    here is the dev_w0 file.
    trc file: "dev_w0", trc level: 1, release: "700"
    ACTIVE TRACE LEVEL           1
    ACTIVE TRACE COMPONENTS      all, MJ

    B Thu Dec 18 11:32:24 2008
    B  create_con (con_name=R/3)
    B  Loading DB library 'E:\usr\sap\BWD\DVEBMGS00\exe\dbmssslib.dll' ...
    B  Library 'E:\usr\sap\BWD\DVEBMGS00\exe\dbmssslib.dll' loaded
    B  Version of 'E:\usr\sap\BWD\DVEBMGS00\exe\dbmssslib.dll' is "700.08", patchlevel (0.171)
    B  New connection 0 created
    M sysno      00
    M sid        BWD
    M systemid   560 (PC with Windows NT)
    M relno      7000
    M patchlevel 0
    M patchno    175
    M intno      20050900
    M make:      multithreaded, ASCII, optimized
    M pid        4224
    M
    M  kernel runs with dp version 241(ext=110) (@(#) DPLIB-INT-VERSION-241)
    M  length of sys_adm_ext is 364 bytes
    M  ***LOG Q0Q=> tskh_init, WPStart (Workproc 0 4224) [dpxxdisp.c   1323]
    I  MtxInit: 30000 0 0
    M  DpSysAdmExtCreate: ABAP is active
    M  DpSysAdmExtCreate: VMC (JAVA VM in WP) is not active
    M  DpShMCreate: sizeof(wp_adm)          43512     (1176)
    M  DpShMCreate: sizeof(tm_adm)          3786880     (18840)
    M  DpShMCreate: sizeof(wp_ca_adm)          18048     (60)
    M  DpShMCreate: sizeof(appc_ca_adm)     6000     (60)
    M  DpCommTableSize: max/headSize/ftSize/tableSize=500/8/528040/528048
    M  DpShMCreate: sizeof(comm_adm)          528048     (1048)
    M  DpSlockTableSize: max/headSize/ftSize/fiSize/tableSize=0/0/0/0/0
    M  DpShMCreate: sizeof(slock_adm)          0     (96)
    M  DpFileTableSize: max/headSize/ftSize/tableSize=0/0/0/0
    M  DpShMCreate: sizeof(file_adm)          0     (72)
    M  DpShMCreate: sizeof(vmc_adm)          0     (1296)
    M  DpShMCreate: sizeof(wall_adm)          (22440/34344/56/100)
    M  DpShMCreate: sizeof(gw_adm)     48
    M  DpShMCreate: SHM_DP_ADM_KEY          (addr: 05510040, size: 4446512)
    M  DpShMCreate: allocated sys_adm at 05510040
    M  DpShMCreate: allocated wp_adm at 05511BA8
    M  DpShMCreate: allocated tm_adm_list at 0551C5A0
    M  DpShMCreate: allocated tm_adm at 0551C5D0
    M  DpShMCreate: allocated wp_ca_adm at 058B8E50
    M  DpShMCreate: allocated appc_ca_adm at 058BD4D0
    M  DpShMCreate: allocated comm_adm at 058BEC40
    M  DpShMCreate: system runs without slock table
    M  DpShMCreate: system runs without file table
    M  DpShMCreate: allocated vmc_adm_list at 0593FAF0
    M  DpShMCreate: allocated gw_adm at 0593FB30
    M  DpShMCreate: system runs without vmc_adm
    M  DpShMCreate: allocated ca_info at 0593FB60
    M  DpShMCreate: allocated wall_adm at 0593FB68
    M  rdisp/queue_size_check_value :  -> off
    M  ThTaskStatus: rdisp/reset_online_during_debug 0
    X  EmInit: MmSetImplementation( 2 ).
    X  MM global diagnostic options set: 0
    X  <ES> client 0 initializing ....
    X  Using implementation view
    X  <EsNT> Using memory model view.
    M  <EsNT> Memory Reset disabled as NT default
    X  ES initialized.

    M Thu Dec 18 11:32:25 2008
    M  ThInit: running on host bwdev

    M Thu Dec 18 11:32:26 2008
    M  calling db_connect ...
    C  Thread ID:5228
    C  Thank You for using the SLOLEDB-interface
    C  Using dynamic link library 'E:\usr\sap\BWD\DVEBMGS00\exe\dbmssslib.dll'
    C  dbmssslib.dll patch info
    C    patchlevel   0
    C    patchno      171
    C    patchcomment ODBC DBSL Error handling in DSQL batches (1233665)
    C  Local connection used on BWDEV to named instance: np:BWDEV\BWD
    C  CopyLocalParameters: dbuser is 'bwd'
    C  Using Provider SQLNCLI
    C  OpenOledbConnection: MARS property was set successfully.
    C  Provider Release:9.00.3042.00
    C  Using Provider SQLNCLI
    C  OpenOledbConnection: MARS property was set successfully.
    C  Cache sizes: header 52 bytes, 20000 names (26880000 bytes), 500 dynamic statements (2728000 bytes), total 29608052 bytes
    C  Using shared procedure name cache BWDEV_BWD_BWDBWD_BWD_MEM initialized by another process.
    C  Connected to db server : [BWDEV\BWD] server_used : [np:BWDEV\BWD], dbname: BWD, dbuser: bwd
    C  pn_id:BWDEV_BWD_BWDBWD_BWD
    C  Using MARS (on sql 9.0)
    B  Connection 0 opened (DBSL handle 0)
    B  Wp  Hdl ConName          ConId     ConState     TX  PRM RCT TIM MAX OPT Date     Time   DBHost         
    B  000 000 R/3              000000000 ACTIVE       NO  YES NO  000 255 255 20081218 113226 BWDEV\BWD      
    C  The IRow interface is supported by this OLEDB provider
    M  db_connect o.k.
    M  ICT: exclude compression: .zip,.cs,.rar,.arj,.z,.gz,.tar,.lzh,.cab,.hqx,.ace,.jar,.ear,.war,.css,.pdf,.js,.gzip,.uue,.bz2,.iso,.sda,.sar,.gif

    I Thu Dec 18 11:32:28 2008
    I  MtxInit: 0 0 0
    M  SHM_PRES_BUF               (addr: 0B550040, size: 4400128)
    M  SHM_ROLL_AREA          (addr: 9C490040, size: 61865984)
    M  SHM_PAGING_AREA          (addr: 0B990040, size: 33554432)
    M  SHM_ROLL_ADM               (addr: 0D9A0040, size: 618548)
    M  SHM_PAGING_ADM          (addr: 0DA40040, size: 525344)
    M  ThCreateNoBuffer          allocated 324144 bytes for 1000 entries at 0DAD0040
    M  ThCreateNoBuffer          index size: 3000 elems
    M  ThCreateVBAdm          allocated 7424 bytes (50 server) at 09590040
    X  EmInit: MmSetImplementation( 2 ).
    X  MM global diagnostic options set: 0
    X  <ES> client 0 initializing ....
    X  Using implementation view
    X  ES initialized.
    B  dbntab: NTAB buffers attached
    B  dbntab: Buffer FTAB(hash header)  (addr: 0DBB0088, size: 456)
    B  dbntab: Buffer FTAB(anchor array) (addr: 0DBB0250, size: 159884)
    B  dbntab: Buffer FTAB(item array)   (addr: 0DBD72E0, size: 719172)
    B  dbntab: Buffer FTAB(data area)    (addr: 0DC86C28, size: 30724096)
    B  dbntab: Buffer IREC(hash header)  (addr: 10770088, size: 456)
    B  dbntab: Buffer IREC(anchor array) (addr: 10770250, size: 159884)
    B  dbntab: Buffer IREC(item array)   (addr: 107972E0, size: 179784)
    B  dbntab: Buffer IREC(data area)    (addr: 107C3128, size: 6144000)
    B  dbntab: Buffer STAB(hash header)  (addr: 0F9E0088, size: 456)
    B  dbntab: Buffer STAB(anchor array) (addr: 0F9E0250, size: 159884)
    B  dbntab: Buffer STAB(item array)   (addr: 0FA072E0, size: 179784)
    B  dbntab: Buffer STAB(data area)    (addr: 0FA33128, size: 3072000)
    B  dbntab: Buffer TTAB(hash header)  (addr: 10DA0088, size: 4304)
    B  dbntab: Buffer TTAB(anchor array) (addr: 10DA1158, size: 159884)
    B  dbntab: Buffer TTAB(item array)   (addr: 10DC81E8, size: 479448)
    B  dbntab: Buffer TTAB(data area)    (addr: 10E3D2C0, size: 3755676)
    B  db_con_shm_ini:  WP_ID = 0, WP_CNT = 37, CON_ID = -1
    B  dbstat: TABSTAT buffer attached (addr: 119415D0)
    B  dbtbxbuf: Buffer TABL  (addr: 125B00C8, size: 30000128, end: 1424C4C8)
    B  dbtbxbuf: Buffer TABLP (addr: 142500C8, size: 10240000, end: 14C140C8)
    B  dbexpbuf: Buffer EIBUF (addr: 14C200D0, size: 4194304, end: 150200D0)
    B  dbexpbuf: Buffer ESM   (addr: 150300D0, size: 4194304, end: 154300D0)
    B  dbexpbuf: Buffer CUA   (addr: 111E00D0, size: 3072000, end: 114CE0D0)
    B  dbexpbuf: Buffer OTR   (addr: 154400D0, size: 4194304, end: 158400D0)
    B  dbcalbuf: Buffer CALE  (addr: 0FD40040, size: 500000, end: 0FDBA160)
    M  CCMS: AlInitGlobals : alert/use_sema_lock = TRUE.
    S  *** init spool environment
    S  initialize debug system
    T  Stack direction is downwards.
    T  debug control: prepare exclude for printer trace
    T  new memory block 0B45B140
    S  spool kernel/ddic check: Ok
    S  using table TSP02FX for frontend printing
    S  2 spool work process(es) found
    S  frontend print via spool service enabled
    S  printer list size is 150
    S  printer type list size is 50
    S  queue size (profile)   = 300
    S  hostspool list size = 3000
    S  option list size is 30
    S      found processing queue enabled
    S  found spool memory service RSPO-RCLOCKS at 1D630098
    S  doing lock recovery
    S  setting server cache root
    S  found spool memory service RSPO-SERVERCACHE at 1D6308E8
    S    using messages for server info
    S  size of spec char cache entry: 165020 bytes (timeout 100 sec)
    S  size of open spool request entry: 1216 bytes
    S  immediate print option for implicitely closed spool requests is disabled

    A Thu Dec 18 11:32:29 2008

    A  -PXA--
    A  PXA INITIALIZATION
    A  PXA: Locked PXA-Semaphore.
    A  System page size: 4kb, total admin_size: 34956kb, dir_size: 17352kb.
    A  Attached to PXA (address 4D080040, size 600000K, 2 fragments of 282524K )
    A  abap/pxa = shared protect gen_remote
    A  PXA INITIALIZATION FINISHED
    A  -PXA--


    A Thu Dec 18 11:32:30 2008
    A  ABAP ShmAdm attached (addr=BACA6000 leng=20955136 end=BC0A2000)
    A  >> Shm MMADM area (addr=BB025338 leng=134016 end=BB045EB8)
    A  >> Shm MMDAT area (addr=BB046000 leng=17154048 end=BC0A2000)
    A  RFC Destination> destination bwdev_BWD_00 host bwdev system BWD systnr 0 (bwdev_BWD_00)
    A  RFC Options> H=bwdev,S=00,d=1,
    A  RFC FRFC> fallback activ but this is not a central instance.
    A   
    A  RFC rfc/signon_error_log = -1
    A  RFC rfc/dump_connection_info = 0
    A  RFC rfc/dump_client_info = 0
    A  RFC rfc/cp_convert/ignore_error = 1
    A  RFC rfc/cp_convert/conversion_char = 23
    A  RFC rfc/wan_compress/threshold = 251
    A  RFC rfc/recorder_pcs not set, use defaule value: 1
    A  RFC rfc/delta_trc_level not set, use default value: 0
    A  RFC rfc/no_uuid_check not set, use default value: 0
    A  RFC rfc/bc_ignore_thcmaccp_retcode not set, use default value: 0
    A  RFC Method> initialize RemObjDriver for ABAP Objects
    M  ThrCreateShObjects          allocated 76458 bytes at 0FF30040
    N  SsfSapSecin: putenv(SECUDIR=E:\usr\sap\BWD\DVEBMGS00\sec): ok

    N  =================================================
    N  === SSF INITIALIZATION:
    N  ===...SSF Security Toolkit name SAPSECULIB .
    N  ===...SSF trace level is 0 .
    N  ===...SSF library is E:\usr\sap\BWD\DVEBMGS00\exe\sapsecu.dll .
    N  ===...SSF hash algorithm is SHA1 .
    N  ===...SSF symmetric encryption algorithm is DES-CBC .
    N  ===...completed with return code 5.
    N  =================================================
    B  dbtran INFO (init_connection '<DEFAULT>' [MSSQL:700.08]):
    B   max_blocking_factor =  50,  max_in_blocking_factor      = 255,
    B   min_blocking_factor =   5,  min_in_blocking_factor      =  10,
    B   prefer_union_all    =   1,  prefer_join                 =   1,
    B   prefer_fix_blocking =   0,  prefer_in_itab_opt          =   0,
    B   convert AVG         =   1,  alias table FUPD            =   0,
    B   escape_as_literal   =   0,  opt GE LE to BETWEEN        =   0,
    B   select *            =0x00,  character encoding          =SBCS / []:X,
    B   use_hints           = abap->1, dbif->0x1, upto->0, rule_in->0,
    B                         rule_fae->0, concat_fae->0, concat_fae_or->0
    N  MskiInitLogonTicketCacheHandle: Logon Ticket cache pointer retrieved from shared memory.
    N  MskiInitLogonTicketCacheHandle: Workprocess runs with Logon Ticket cache.
    M  JrfcVmcRegisterNativesDriver o.k.
    W  =================================================
    W  === ipl_Init() called
    W    ITS Plugin: Path dw_gui
    W    ITS Plugin: Description ITS Plugin - ITS rendering DLL
    W    ITS Plugin: sizeof(SAP_UC) 1
    W    ITS Plugin: Release: 700, [7000.0.175.20050900]
    W    ITS Plugin: Int.version, [33]
    W    ITS Plugin: Feature set: [17]
    W    ===... Calling itsp_Init in external dll ===>
    W  === ipl_Init() returns 0, ITSPE_OK: OK
    W  =================================================
    N  VSI: WP init in ABAP VM completed with rc=0
    E  Enqueue Info: rdisp/wp_no_enq=1, rdisp/enqname=<empty>, assume bwdev_BWD_00
    E  Enqueue Info: enque/disable_replication = 2
    E  Enqueue Info: replication disabled
    E  EnqCcInitialize: local lock table initialization o.k.
    E  EnqId_SuppressIpc: local EnqId initialization o.k.
    E  EnqCcInitialize: local enqueue client init o.k.
    M  MBUF info for hooks: MS component UP
    M  ThSetEnqName: set enqname by server list
    M  ThISetEnqname: enq name = >bwdev_BWD_00                            <

    E  *************** EnqId_EN_ActionAtMsUpHook ***************
    E  Hook on upcoming Ms (with EnqSrv), get auth EnqId and check it locally

    E  *************** ObjShMem_CheckAuthoritativeEnqId ***************
    E  Checking authoritative EnqId from EnqSrv into ObjShMem
    E  ObjShMem_CheckAuthoritativeEnqId: ObjShMem ...
    E  EnqId.EnqTabCreaTime    = -999
    E  EnqId.RandomNumber      = -999
    E  ReqOrd.TimeInSecs       = -999
    E  ReqOrd.ReqNumberThisSec = -999
    E  ObjShMem_CheckAuthoritativeEnqId: ObjShMem ...
    E  EnqId.EnqTabCreaTime    = -999
    E  EnqId.RandomNumber      = -999
    E  ReqOrd.TimeInSecs       = -999
    E  ReqOrd.ReqNumberThisSec = -999
    E  ObjShMem_CheckAuthoritativeEnqId: EnqId is initial in ShMem
    E  ObjShMem_CheckAuthoritativeEnqId: Overwrite incoming auth EnqId, continue
    E  EnqId inscribed into initial ObjShMem: (ObjShMem_CheckAuthoritativeEnqId)
    E  -SHMEM--
    E  EnqId:          EnqTabCreaTime/RandomNumber    = 18.12.2008 11:32:30  1229571150 / 4224
    E  ReqOrd at Srv:  TimeInSecs/ReqNumberThisSec    = 18.12.2008 11:32:30  1229571150 / 2
    E  ReqOrd at Cli:  TimeInSecs/ReqNumberThisSec    = 18.12.2008 11:32:30  1229571150 / 1
    E  Status:         STATUS_OK
    E  -
    M  ThActivateServer: state = STARTING
    L  BtcSysStartRaise: Begin
    L  Raise event SAP_SYSTEM_START with parameter <bwdev_BWD_00        >
    L  BtcSysStartRaise: End

    M Thu Dec 18 11:32:31 2008
    M  SosICreateNewAnchorArray: sos_search_anchor_semantics = 1

    C Thu Dec 18 11:32:33 2008
    C  The IRow interface is supported by this OLEDB provider

    B Thu Dec 18 11:32:36 2008
    B  table logging switched off for all clients

    M Thu Dec 18 11:32:37 2008
    M  SecAudit(RsauShmInit): WP attached to existing shared memory.
    M  SecAudit(RsauShmInit): addr of SCSA........... = 05300040
    M  SecAudit(RsauShmInit): addr of RSAUSHM........ = 05300490
    M  SecAudit(RsauShmInit): addr of RSAUSLOTINFO... = 053004C8
    M  SecAudit(RsauShmInit): addr of RSAUSLOTS...... = 053004D4

  • Web auth supporting fragmented SSL&TLS packets in 7.0.116?

    Dear collegues and Cisco experts.
    I hope anyone of you can reply if this is supported on thew current platform (WLC5508 sw rel 7.0.116)
    I have not been able to reproduce this myself, but some problems have been reported after mid january, when KB2585542 might be the culprit.
    Is the internal webauthentication portal in above platform able to handle this, or is s/w upgrade inevitable?
    Does WLC guest portal w high cipeher option handle the Fragmentation  of SSL/TLS application records, as described in the following RFCs:
    TLS 1.0: http://www.ietf.org/rfc/rfc2246.txt paragraph 6.2.1
    SSL 3.0: http://www.ietf.org/rfc/rfc6101.txt paragraph 5.2.1
    Environment
    Windows XP Professional SP3 clients with Internet Explorer 8
    Wireless lan controller cluster with redundant(2) webauthentication anchors (all AIR-CT5508-XXX-K9 sw rel 7.0.116)
    Microsoft statement
    "After installing MS12-006, you may experience authentication failure or loss of connectivity to some HTTPS servers. This issue occurs because this security update changes the way that records are sent to HTTPS servers. To address an information disclosure vulnerability, SChannel now implements certain ciphers used in SSL 3.0 and TLS 1.0 in a more secure fashion. The updated behaviour is fully compliant with the RFCs, but it is something that has generally not been used on the internet before
    There are two fixes involved: the SChannel fix makes the new behaviour available, and a fix included in the December Cumulative Update makes Internet Explorer request that more secure option. The behaviour will change only if both are present"
    Sincere regards
    Mats Nilson
       AIR-CT5508-100-K9

    Debug says your client is being requested to anchor but not moving passed that....
    So from the looks of things, this has nothing to do with DHCP. Instead, you are either trying to Anchor your clients to a non-existant anchor, or something is royally hosed with mobility.
    I'd suggest debug client   and "debug mobility handoff enable"  from this same WLC as well what ever other WLC your client is being sent to....

  • Precompile web application in cluster

    We are running 6.0 sp2 and are trying to determine how to have a web
              application precompile when deployed to a cluster.
              We have seen messages that there is an issue with 6.0 where you can only
              precompile if the web application is in exploded format.
              We have a scenario with 1 admin server servicing a cluster with 2 servers in
              the cluster. The web application is being deployed to the cluster.
              If we set the precompile option, the managed servers fail to deploy the
              application because of an error with precompiling.
              This seems to make sense since when the application is deployed, it is
              deployed as a war file.
              Does anyone have any advice as to how to deploy an application to a cluster
              and take advantage of precompiling?
              Is there any documentation that addresses this?
              

    Eric Gordon wrote in message <3b39278e$[email protected]>...
              >We are running 6.0 sp2 and are trying to determine how to have a web
              >application precompile when deployed to a cluster.
              >We have seen messages that there is an issue with 6.0 where you can only
              >precompile if the web application is in exploded format.
              This is precompiling at build time. You have to hardcode the name of the
              server into the path of the precompiled files.
              >We have a scenario with 1 admin server servicing a cluster with 2 servers
              in
              >the cluster. The web application is being deployed to the cluster.
              >If we set the precompile option, the managed servers fail to deploy the
              >application because of an error with precompiling.
              I assume you are you use the precompile option in the web.xml? If so the
              problem may be that not every .jsp file can compile on its own. I have this
              problem and am solving it by changing the extension of include files from
              .jsp to .inc. Otherwise you get these fragment .jsp files that won't
              precompile because they are never going to be used stand alone.
              The only other reason I can see which could affect your ability to
              precompile on the managed servers is if you depend on startWebLogic
              settings, like extra classpath that prevents precompiling.
              Its too bad the weblogic server doesn't give detailed output when
              precompiling a war file.
              

  • Fragmentation in tablespace in oracle 9i

    Hi All,
    We have oracle 9.2.0.8 installed in solaris 9, database size is 1.5 TB, we are looking for the tablespaces which are suffering from the fragmentation.
    We have locally manged tablesapaces in the database. so can you please tell me, is there any tablespace fragmentation in the locally managed tablespace in oracle 9i, and if it is there than how to relove the situation.?
    thanks in advance

    If you mean by fragmentation, a number of free space areas in the data files - sometimes called honeycombing, this script will show them to you.
    SELECT  'free space'      owner,
            ' '               object,
            ' '               obj_type,
            f.file_name,
            s.file_id,
            s.block_id,
            s.blocks,
            s.bytes/1048576   mbytes
      FROM  dba_free_space s,
            dba_data_files f
    WHERE  s.file_id = &file_id
       AND  s.file_id = f.file_id
    UNION
    SELECT  owner,
            segment_name,
            DECODE(segment_type, 'TABLE',          'T',
                                 'INDEX',          'I',
                                 'ROLLBACK',       'RB',
                                 'CACHE',          'CH',
                                 'CLUSTER',        'CL',
                                 'LOBINDEX',       'LI',
                                 'LOBSEGMENT',     'LS',
                                 'TEMPORARY',      'TY',
                                 'NESTED TABLE',   'NT',
                                 'TYPE2 UNDO',     'U2',
                                 'TABLE PARTITION','TP',
                                 'INDEX PARTITION','IP', '?'),
            f.file_name,
            s.file_id,
            s.block_id,
            s.blocks,
            s.bytes/1048576
      FROM  dba_extents s,
            dba_data_files f
    WHERE  s.file_id = &file_id
       AND  s.file_id = f.file_id
    ORDER
        BY  file_id,
            block_id
    However, when you look at the report you will see that in a locally managed tablespace, all of the free space 'chunks' are all 8 blocks, or a multiple of 8 blocks. This means that all of these free space chunks are able to be (re)used. And, they will be used. There is no need to do anything unless there is a pressing need to give disk back to the OS.

  • Does OCFS2 file system get fragmented

    We are running Production & Testing RAC databases on Oracle 9.2.0.8 RAC on Red Hat 4.0 using OCFS2 for the cluster file system.
    Every week we refresh our Test database by deleting the datafiles and cloning our Standby database to the Test database. The copying of the datafiles from the Standby mount points to the Test database mount points (same server), seems to be taking longer each time we do this.
    My question is : can the OCFS2 file system become fragmented over time from the constant deletion & copying of the datafiles and if so is there a way to defragment it.
    Thanks
    John

    Hi,
    I think it will get fragment if you constant deletion & copying of the datafiles on ocfs2.You can set the preferable block size and cluster size on the basis of actual applications,which can reduce the file fragments.
    Regards
    Terry

  • Cluster ratio & Defragmentation

    Scenario:
    Before:-Cluster ratio:1, Index cache=300MB
    After:- After running a calc scipt (max calc dim are used) cluster ratio-0.68 & index cache has increased.
    I am running the calc script as a part of test and I am not loading any data. Data is aggregated to Higher level and still I am running the calc scipt for aggregation.
    According to debag:-
    Fragmentation is unused disk space. Fragmentation is created when Essbase
    writes a data block to a new location on disk and leaves unused space in the former location of the data block.
    Block size increases because data from a data load or calculation is appended to the blocks; the blocks must therefore be written to the end of a data file.
    What does the above statement "calculation is appended to the blocks" mean? even though my higher level blocks are present.
    Does this mean that when I run a calculation my block size increases?
    If yes why does it increase?
    If yes, the next time when I run the calc script why my cluster ratio is still at 0.68 and why doesn't that decrease?
    NOTE: After fragmentation, all blocks are marked as dirty. Does this has anything to do with the decrease in the cluster ratio after rrunnnig the calc script (all blocks are marked as clean)? I run the calc script again (blocks are marked as clean) but still it calculates all the blocks becaouse intelligent calc is off. now after this calc also does it append the calculated value to the blocks?
    What is the logic behind this?
    I am not using intelligent calculation but my clearupdatestatus is set to after.

    Yes, defragmentation (restructuring) does throw away old (free for reuse) blocks.
    You must differentiate a little bit. dirty blocks != clean blocks != blocks available for reuse.
    Have a look at the documentation for restructuring, that's the time essbase runs through all blocks and tries to allign them back in best order for retrieval and throwing away the blocks available for reuse. That's also the reason essbase should have at least twice as much space left on the drive as your cube has in size.
    1- I keep on calculating again and again so the new blocks gets added again
    and again even though my blocks are clean increaing my data file Yes for intelligent calculation turned off AND certain conditions which influence the stepping of a calculation (commited mode, parallel calc, cache size changes between calcs ..)
    (or)
    2 - Is it like my blocks are clean so it doesn't add any new blocks for my further > calculation?Again yes :-) for intelligent calculation turned on AND certain conditions which influence the stepping of the calculation are NOT present.
    If the block is relocated, the former position becomes free for reuse
    Does this means, this block has both teh old and new values which increases
    the size of the block and it needs to be relocated as the existing space is not
    sufficient for this block?Quite so, but not one block is holding new and old data, each block is holding its own data. A new block is created for the new data, the old block remains untouched besides flagging it for being available for reuse.
    The flagging for reuse does not mean it is really lost. As it could be reused by a block which does fit in. But I do not know if reuse need an exact match or is a less than equal comparison. In the later case (which I would guess is the one used) small gaps would still be present as not the whole space for the block would be used up.

  • Unix Cluster Services provided with 10G

    The information that I have read with regards to Oracle 10G (for Unix customers) suggests that 10G now includes volume management, file systems and the cluster services.
    Within earlier release of Oracle DB and Apps, the above components would have been provided by software vendors (e.g. Veritas) or the hardware vendors themselves (Sun, IBM etc..).
    The provision of an Oracle 10G architecture to customers without the requirement to procure additional "clusterware" services will clearly appeal to customers on tight budgets or who or currently focused on ROI. However, what about customers who are focused on performance and the highest levels of availability?
    Looking at the file system provided by Oracle (OSFS?), does this file system include journaling capabilities (i.e. improve fsck performance under unplanned system crash conditions). In addition, does the ORacle file system allow storage check points to be taken at a file system level? Finally, is there any performance data available comparing the Oracle file system directly with say Vxfs (either vanila product or with direct I/O) or Sun's UFS?
    Thanks in advance.
    Nathan

    In Oracle 10g there are two features that provide files system type functions. The first is Oracle Clustered File System (OCFS). It has been built to support clustered files for RAC deployments on NT and Linux. As I am not an expert on OCFS, I will not attempt to address your specific OCFS questions at this time.
    The second feature is Automatic Storage Management (ASM) which is an integrated file system and volume manager purpose built for the Oracle data files. It is a tool for the DBA to manage database storage and to increase their productivity. It speeds up the process of database layout and greatly simplifies the process of extending the storage configuration for a database. However, it is not a general purpose file system nor a general purpose clustered file system as it does not support flat files, trace files or binaries at present. ASM supports both single instance databases on a single node as well as on a clustered set of nodes. It also supports RAC.
    As for performance and availability, ASM has been doing very well in stress testing and beta testing. It performs as well as other file system and volume manager environments and in most cases ASM does better than those other environments. As a non ASM database will typically becomes fragmented and needs to have storage capacity added to it, the performance of that database will either decline or will require vast amounts of DBA and SA time to keep its performance optimal. With the database running on ASM managed disks, the performance will stay optimal with minimal maintenance from DBA and SA. ASM will rebalance the data across new storage and by way of its distribution of the data across the entire storage pool, there will be no fragmentation or imbalance of the workload. With ASM the manual task of I/O performance tuning will not be required.
    We do not yet have any published benchmarks at this time but hope to be able to publish a number which compares ASM to other volume mgmt and file systems environments at two points in time; one is the set up of a new config (best case) and the other is 6 months later when both environments have had the same fragmentation and storage configuration changes. That benchmark will need to track the time spent managing those two databases over that 6 months of as well. The challenge is that no such benchmark exists today.

  • Essbase Fragmentation Question - Average Fragmentation Quotient

    According to the DBA guide, if we are using the 'Average Fragmentation Quotient' to evaluate whether the level likely will cause performance problems. We should defrag if the quotient is above 30%. (our db is over 2G)
    How do I read the below DBSTATS reults for that db? is 'Average Fragmentation Quotient' 43.797% or .437977 %? I've had differant answers from consultants on this.
    Average Cluster Ratio: 1
    Average Fragmentation Quotient: 0.4379771
    Free Space Recovery Needed: Yes
    Estimated Bytes of Recoverable Free Space
    Thank you!

    My guess that it is saying 43.79%, which suggests that there is fragmentation, however your Average cluster ratio is 1, indicating no fragmentation....
    when I say guess, I mean my understanding.
    Edited by: name_and_address_withheld on Jan 10, 2013 7:45 AM

  • Cluster db

    HI
    we setup a jumbo frame in our cluster the db is running , how to check is the setting is correct
    is there is any document avaialabe for this jubmo frame
    One document from oracle says if you use the follwoing command it will ping upto the 8972,
    If i use above 8972 it should fragmented as per document
    ping -M do -s 8973 node02-priv
    in my case even if i put above 9000 it is pinging i can ping upto 16000
    pls share if anyone come across this issue
    rds

    Here are a couple of support articles you may want to check out.
    Oracle® Database on Unix AIX®,HP-UX®,Linux®,Mac OS® X,Solaris®,Tru64 Unix® Operating Systems Installation and Configuration Requirements Quick Reference (8.0.5 to 11.1) #169706.1
    Linux IA64 example of allocating 48GB SGA using hugepages #397568.1
    Search in the Oracle-L archives on "huge pages" and Bobak. Mark has posted a couple of times on the use of this option.
    HTH -- Mark D Powell --

  • 'Read Multicast Msg Fragment'

    I have two weblogic instances on the same Sun box. I run an XSL/XSLT transformation
    on one instance, it takes 15 seconds. I run the exact same XSL/XSLT transformation
    on the other instance, it takes 14 minutes. This is running the same job on both
    instances at the same time. Both instances have available RAM and available threads
    while this job is running.
    One difference I see when comparing the default execute queues for these two weblogic
    instances is that the one that seems to hang has a 'Read Multicast Msg Fragment'
    request listed in the execute queue and the other does not. This appears to be
    associated with this in a thread dump:
    "ExecuteThread: '10' for queue: 'default'" daemon prio=5 tid=0x303248 nid=0x18
    runnable [0xe2a81000..0xe2a819d8]
    at java.net.PlainDatagramSocketImpl.receive(Native Method)
    at java.net.DatagramSocket.receive(DatagramSocket.java:387)
    at weblogic.cluster.FragmentSocket.receive(FragmentSocket.java:161)
    at weblogic.cluster.MulticastManager.execute(MulticastManager.java:336)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    What is this? Could it have something to do with my performance problem?
    Any insight appreciated.
    David Holk

    Are two instances in a cluster. The thread you point out is receiving
    multicast messages. heartbeats, jndi replication, and other internal
    communication happens on multicast.
    I doubt this is the problem.
    Duing the 14min duration please take 3-4 threaddumps at 30 second interval
    and attach it here. That should show whats happening in the server.
    sree
    "David Holk" <dholk@*remove*criticalresourcetech.com> wrote in message
    news:40b4ca58$1@mktnews1...
    >
    I have two weblogic instances on the same Sun box. I run an XSL/XSLTtransformation
    on one instance, it takes 15 seconds. I run the exact same XSL/XSLTtransformation
    on the other instance, it takes 14 minutes. This is running the same jobon both
    instances at the same time. Both instances have available RAM andavailable threads
    while this job is running.
    One difference I see when comparing the default execute queues for thesetwo weblogic
    instances is that the one that seems to hang has a 'Read Multicast MsgFragment'
    request listed in the execute queue and the other does not. This appearsto be
    associated with this in a thread dump:
    "ExecuteThread: '10' for queue: 'default'" daemon prio=5 tid=0x303248nid=0x18
    runnable [0xe2a81000..0xe2a819d8]
    at java.net.PlainDatagramSocketImpl.receive(Native Method)
    at java.net.DatagramSocket.receive(DatagramSocket.java:387)
    atweblogic.cluster.FragmentSocket.receive(FragmentSocket.java:161)
    atweblogic.cluster.MulticastManager.execute(MulticastManager.java:336)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:153)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:134)
    What is this? Could it have something to do with my performance problem?
    Any insight appreciated.
    David Holk

  • How to create an intensity waveform graph cluster with t_0 and dt ?

    Hi all,
    I would like to know whether it is possible to create an intensity waveform like you can do with a 1-d waveform (with "build waveform") so that you get a cluster with the waveform array, the t_0, the dtand the attributes. 
    If not I would like to know the following:  I use references to cluster typedefs to update my controls and indicaters on the front panel. Now if I use a property node for the intensity graph to set the offset and multiplier on the x-scale, the x-scale on the graphs on the sub-VI work perfectly, however not on the real front panel, probably since these get updated through a reference. Does anyone have a clue how to fix this?
    Regards, Pieter

    You are only writing the "value" of the type definition via the property node. This does not include properties such as offset and multiplier.
    On a sidenote, you are using way too much code for most operations.  
    For example, the to-I32 can be placed on the cluster, so only one instance is needed.
    Also property nodes are resizeable, so only one instance is needed.
    There are also Rube Goldberg constructs, such as ">0  *AND* TRUE", which is the same as a simple ">0"
    Overall, you are really fragmenting memory by constantly building and then resizing arrays to keep them at a max size of 2880. This can cause performance problems due to the constant need of reallocations. It might be better to use fixed size arrays and do things "in place".
    Message Edited by altenbach on 03-19-2009 09:57 AM
    LabVIEW Champion . Do more with less code and in less time .
    Attachments:
    OneIsEnough.png ‏8 KB
    CombineProperties.png ‏3 KB

  • Fragment a table in DDBMS

    how can I fragment a relation horizontaly and where these changes made n the schema or in sql query. I am working in project using oracle 10g ,sql ,and java.

    If by "fragment" you mean doing what is typically referred to as "sharding" and storing different subsets of the table on different nodes, no, you can't.
    Unlike how many databases do clustering with a "shared nothing" architecture, a RAC cluster involves multiple instances (software processes running on each node) to access a single database (set of data files) on a shared storage system. Sharding a table in this context doesn't make sense-- every instance is going to have access to the entire table. Sharding makes sense in other approaches to clustering where the storage system is not shared across multiple nodes.
    If you are trying to ensure that different nodes don't conflict with each other when requesting particular blocks, or at least to minimize that contention, you can potentially create different services that run on different nodes and have different users connect to those services. So, for example, if you create a different service for different geographic regions, configured those services to run on particular nodes, configured users in those geographic regions to connect to the service appropriate for their region and the data in a table is naturally separated by region, you would end up with each node preferentially caching blocks that have data associated with their particular region and you would have relatively few cache fusion requests where one node asks for a cached block from a different node.
    If you are trying to improve query performance, you can use partitioning instead of sharding the data (you can do this without using RAC as well). This allows Oracle to store different subsets of the data for a table in different physical segments so your queries can hit individual partitions rather than the entire table to retrieve the data they're after.
    Justin

  • Can Fragmentation Is Reduced while inserting data

    Hi All
    While testing few scenarios i faced one issue when i am not able to create fragmentation in table. This code usually works with me and with second insert i get 90 - 98% fragmentation.
    Today I was inserting 1 Lakh/ 100 Thousand records in single chunk. I repeated this process no. of times and fragmentation changes like
    1st Insert - 15 %
    2nd Insert- 4%
    3rd Insert - 8%
    4th Insert - 5.6%
    5th Insert - 1.5%
    6th Insert - 0.7%
    So is it possible that Inserting data might decrease Avg fragmentation and how?.
    Code i used is 
    create database BKP_Test
    GO
    USE BKP_Test
    GO
    CREATE TABLE Diff_Table
    (ID INT,
    FirstName VARCHAR(100),
    LastName VARCHAR(100),
    City VARCHAR(100))
    GO
    CREATE CLUSTERED INDEX PK_Diff_Table
    ON Diff_Table(ID)
    GO
    /**************Run below insert twice**********/
    INSERT INTO Diff_Table (ID,FirstName,LastName,City)
    SELECT TOP 100000 ROW_NUMBER() OVER (ORDER BY a.name) RowID,
    'Saurabh',
    CASE WHEN ROW_NUMBER() OVER (ORDER BY a.name)%2 = 1 THEN 'SINHA'
    ELSE 'Sumit' END,
    CASE
    WHEN ROW_NUMBER() OVER (ORDER BY a.name)%1000 = 1 THEN 'New Delhi'
    WHEN ROW_NUMBER() OVER (ORDER BY a.name)%10 = 1 THEN 'Mumbai'
    WHEN ROW_NUMBER() OVER (ORDER BY a.name)%10 = 5 THEN 'Benglaru'
    WHEN ROW_NUMBER() OVER (ORDER BY a.name)%10 = 3 THEN 'Kolkata'
    ELSE 'Bhopal' END
    FROM sys.all_objects a
    CROSS JOIN sys.all_objects b
    GO
    SELECT OBJECT_NAME(i.OBJECT_ID) AS TableName,
    i.name AS IndexName,
    indexstats.avg_fragmentation_in_percent
    FROM sys.dm_db_index_physical_stats(DB_ID(), OBJECT_ID('Diff_Table'), NULL, NULL, 'LIMITED') indexstats
    INNER JOIN sys.indexes i ON i.OBJECT_ID = indexstats.OBJECT_ID
    AND i.index_id = indexstats.index_id
    Thanks Saurabh Sinha http://saurabhsinhainblogs.blogspot.in/ Please click the Mark as answer button and vote as helpful if this reply solves your problem

    Hi,
    You can bring fragmentation using below code
    --Create Database Command
    CREATE DATABASE [IndexLogging]
    ON PRIMARY
    ( NAME = N'IndexLogging_dat',
    FILENAME = N'D:\Program Files\Microsoft SQL Server\MSSQL11.MSSQLSERVER\MSSQL\DATA\IndexLogging_data.mdf' ,
    SIZE = 51200KB ,
    MAXSIZE = 51200KB ,
    FILEGROWTH = 15 %)
    LOG ON
    ( NAME = N'IndexLogging_log',
    FILENAME = N'D:\Program Files\Microsoft SQL Server\MSSQL11.MSSQLSERVER\MSSQL\DATA\IndexLogging_log.ldf' ,
    SIZE = 10240KB ,
    MAXSIZE = 25600KB ,
    FILEGROWTH = 5120KB );
    GO
    USE [IndexLogging]
    GO
    --Create filler table
    CREATE TABLE [DBO].[FILLER]
    C1 INT IDENTITY,
    C2 CHAR(50) DEFAULT 'SOME DATA'
    GO
    --Insert default records into filler table
    INSERT INTO [DBO].[FILLER] DEFAULT VALUES;
    GO 10000
    --Create table named Logging
    CREATE TABLE [dbo].[Logging]
    [Id] [int] IDENTITY(1,1) NOT NULL,
    [c1] [char](50) DEFAULT 'SOME MORE DATA' NOT NULL
    --Create Unique cluster Index
    CREATE UNIQUE CLUSTERED INDEX IX_LOGGING ON [DBO].[LOGGING] (ID)
    GO
    --Insert default records
    INSERT INTO DBO.Logging DEFAULT VALUES;
    GO 10000
    --check fragmentation for the table Logging
    SELECT
    [avg_fragmentation_in_percent]
    FROM sys.dm_db_index_physical_stats (
    DB_ID (N'IndexLogging'), OBJECT_ID (N'LOGGING'), 1, NULL, 'LIMITED');
    GO
    You can see fragmentation is almost 4 %. Actually it is 3.79746835
    DROP TABLE DBO.FILLER;
    GO
    --Shrinking data file to bring fragmentation
    DBCC SHRINKFILE (IndexLogging_dat, 3)
    GO
    SELECT
    [avg_fragmentation_in_percent]
    FROM sys.dm_db_index_physical_stats (
    DB_ID (N'IndexLogging'), OBJECT_ID (N'LOGGING'), 1, NULL, 'LIMITED');
    --Now you would see fragmentation
    Is this you require or something else ?
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Cluster Monitoring showing 'n/a' for most columns for some servers

    Hi. The Cluster Monitoring tab shows 'n/a' for the Primaries, Secondary Distributions, Fragments Sent/Received and Lost Multicast messages for three of four servers in the cluster.
              However, there are no errors in the log files, even when Multi-cast debug is enabled.
              Any ideas?
              Thanks for any help.

    have you done the multicast test ?, please check the multicast ip given to make sure that no one used that ip and port.

Maybe you are looking for

  • Captivate 7 Click box and bottons not able to open an URL on a new window

    I inserted a button and a click box in CP7 to open an URL on a new window. The URL opens but in a New Tab. There is an issue that the URL cannot be opened on a NEW WINDOW, I am using IE and Chrome. Do you have any ideas about this issue and how to tr

  • How can I delete credit card information

    How do I delete my credit card information.

  • SAP QUERY does not execute in other client

    Hi, I am working on SAP QUERY on the Quality Assurance and Cust & Dev Client and i am facing a strange thing when i tried to execute the Query from SQ01 ,there i found itis not able to get executed but when i generated the report ,after that it get s

  • Normal Indexes created for Primary/Unique Keys in Oracle8/8i

    I remeber prior to Oracle8, when we create a Primary/Unique Key constraints, an unique index will be created by default. But I am noticing now in 8/8i when we create either Primary or Unique key, only normal/simple index is created. We are not able t

  • Original files in two locations

    I had to reinstall Windows XP, and in doing so my computer "name" changed (from my name to "administrator"). When I reinstalled Lightroom (my data files were back on the computer and were immediately available in the program), many of the images have