Why is flashback needed in Failover scenario?

Hello. Why do we need to turn flashback ON on primary and standby databases for fail-over to work?

IMHO Failover doesn't need flashback to be enabled.
This link lists the prerequisites for failover Role Transitions .
In addition to that you should check if standby is running in maximum protection mode and have to switch it over to maximum performance till the transition completes.
Could you post the error messages you came across?
CSM

Similar Messages

  • Why we dont need sender agreement and comm channel in IDOC to file scenario

    Hi,
    there are 2 queries:
    first query :
    Scenario is "Idoc to file transfer"
    why we dont need sender agreement and communication channel for IDOC to file ,while we do need
    both of these for "File to RFC "
    Answer : which i got is because IDOC sits at ABAP stack but same is valid for RFC also.
    Then why For "IDOC to file" we dont need that.
    Second query:
    Also when i did scenario for File1-RFC-File2 scenario using BPM i have created a interface
    between XI-RFC synchronous to send data which we got from file1 in XI to RFC importing parameter
    and getting data from RFC back to File2.
    Here in IDOC to file scenario why we dont need any interfface for IDOC-XI Asynchronous.
    Regards,
    Mayank
    Edited by: mayank verdia on Jan 13, 2009 11:49 AM
    Edited by: mayank verdia on Jan 13, 2009 11:50 AM

    Dear All,
    For file to idoc scenario is there any possibility to get line items details or xml details i.e the segments and its related field details using reference id, transaction id or interface name or message id in SAP PI 7.0. I know we need to click each and every message in sxi_monitor and look for details.
    For SAP(R3 System) I can create a report and set the job for specific time period so automatically it throws the details(like reference no, document date, invoice no from) in ftp path as .csv file. The same ftp path is maintained in program.
    I wanted to check FTP--->PI postings and I have set the job at r3 system it is working fine and Im monitoring it too.
    Now the end to end scenario is FTP--->PI--->ECC(R3 system). Please help.Many Thanks.

  • Need test documents for RAC failover Scenarios

    Hello friends...
    By the end of this week i have to produce sum test documents for RAC and Database server including Sun Cluster Failover Scenarios.
    Can sumone guide me to a link where i can get enough help.
    I have already managed to get enough information.. but i want to see to it that i cover most of the topics.
    Thanks, Regards
    Monu Koshy

    Please check the following links.
    http://download-uk.oracle.com/docs/cd/B19306_01/rac.102/b14197/toc.htm
    http://download-uk.oracle.com/docs/cd/B19306_01/install.102/b14205/toc.htm
    -aijaz

  • Uplink failover scenarios - The correct behavior

    /* Style Definitions */
    table.MsoNormalTable
    {mso-style-name:"Table Normal";
    mso-tstyle-rowband-size:0;
    mso-tstyle-colband-size:0;
    mso-style-noshow:yes;
    mso-style-priority:99;
    mso-style-qformat:yes;
    mso-style-parent:"";
    mso-padding-alt:0in 5.4pt 0in 5.4pt;
    mso-para-margin:0in;
    mso-para-margin-bottom:.0001pt;
    mso-pagination:widow-orphan;
    font-size:11.0pt;
    font-family:"Calibri","sans-serif";
    mso-ascii-font-family:Calibri;
    mso-ascii-theme-font:minor-latin;
    mso-hansi-font-family:Calibri;
    mso-hansi-theme-font:minor-latin;}
    Hello Dears,
    I’m somehow confused about the failover scenarios related to the uplinks and the Fabric Interconnect (FI) switches, as we have a lot of failover points either in the vNIC , FEX , FI or uplinks.
    I have some questions and I hope that someone can clear this confusion:
    A-     Fabric Interconnect failover
    1-      As I understand when I create a vNIC , it can be configured to use FI failover , which means if FI A is down , or the uplink from the FEX to the FI is down , so using the same vNIC it will failover to the other FI via the second FEX ( is that correct , and is that the first stage of the failover ?).
    2-      This vNIC will be seen by the OS as 1 NIC and it will not feel or detect anything about the failover done , is that correct ?
    3-      Assume that I have 2 vNICs for the same server (metal blade with no ESX or vmware), and I have configured 2 vNICs to work as team (by the OS), does that mean that if primary FI or FEX is down , so using the vNIC1 it will failover to the 2nd FI, and for any reason the 2nd vNIC is down (for example if uplink is down), so it will go to the 2nd vNIC using the teaming ?
    B-      FEX failover
    1-      As I understand the blade server uses the uplink from the FEX to the FI based on their location in the chassis, so what if this link is down, does that mean FI failover will trigger, or it will be assigned to another uplink ( from the FEX to the FI)
    C-      Fabric Interconnect Uplink failover
    1-      Using static pin LAN group, the vNIC is associated with an uplink, what is the action if this uplink is down ? will the vNIC:
    a.       Brought down , as per the Network Control policy applied , and in this case the OS will go for the second vNIC
    b.      FI failover to the second FI , the OS will not detect anything.
    c.       The FI A will re-pin the vNIC to another uplink on the same FI with no failover
    I found all theses 3 scenarios in a different documents and posts, I did not have the chance it to test it yet, so it will be great if anyone tested it and can explain.
    Finally I need to know if the correct scenarios from the above will be applied to the vHBA or it has another methodology.
    Thanks in advance for your support.
    Moamen

    Moamen
    A few things about Fabric Failover (FF)  to keep in mind before I try to address your questions.
    FF is only supported on the M71KR and the M81KR.
    FF is only applicable/supported in End Host Mode of Operation and applies only to ethernet traffic. For FC traffic one has to use multipathing software (the way FC failover has worked always). In End Host mode, anything along the path (adapter port, FEX-IOM link, uplinks) fails and FF is initiated for ethernet traffic *by the adapter*.
    FF is an event which is triggered by vNIC down i.e a vNIC is triggered down and the adapter initiates the failover i.e it sends a message to the other fabric to activate the backup veth (switchport) and the FI sends our gARPs for the MAC as part of it. As it is adapter driven, this is why FF is only available on a few adapters i.e for now the firmware for which is done by Cisco.
    For the M71KR (menlo's) the firmware on the Menlo chip is made by Cisco. The Oplin and FC parts of the card, and Intel/Emulex/Qlogic control that.
    The M81KR is made by Cisco exclusively for UCS and hence the firmware on that is done by us.
    Now to your questions -
    >1-      As I understand when I create a vNIC , it can be configured to use FI failover , which means if FI A is down , or the uplink from the FEX to the >FI is down , so using the same vNIC it will failover to the other FI via the second FEX ( is that correct , and is that the first stage of the failover ?).
    Yes
    > 2-      This vNIC will be seen by the OS as 1 NIC and it will not feel or detect anything about the failover done , is that correct ?
    Yes
    >3-      Assume that I have 2 vNICs for the same server (metal blade with no ESX or vmware), and I have configured 2 vNICs to work as team (by the >OS), does that mean that if primary FI or FEX is down , so using the vNIC1 it will failover to the 2nd FI, and for any reason the 2nd vNIC is down (for >example if uplink is down), so it will go to the 2nd vNIC using the teaming ?
    Instead of FF vNICs you can use NIC teaming. You bond the two vNICs which created a bond interface and you specify an IP on it.
    With NIC teaming you will not have the vNICs (in the Service Profile) as FF. So the FF will not kick in and the vNIC will be down for the teaming software to see on a fabric failure etc for the teaming driver to come into effect.
    > B-      FEX failover
    > 1-      As I understand the blade server uses the uplink from the FEX to the FI based on their location in the chassis, so what if this link is down, > >does that mean FI failover will trigger, or it will be assigned to another uplink ( from the FEX to the FI)
    Yes, we use static pinning between the adapters and the IOM uplinks which depends on the number of links.
    For example, if you have 2 links between IOM-FI.
    Link 1 - Blades 1,3,5,7
    Link 2 - Blades 2,4,6,8
    If Link 1 fails, Blade 1,3,5,7 move to the other IOM.
    i.e it will not failover to the other links on the same IOM-FI i.e it is no a port-channel.
    The vNIC down event will be triggered. If FF is initiated depends on the setting (above explanation).
    > C-      Fabric Interconnect Uplink failover
    > 1-      Using static pin LAN group, the vNIC is associated with an uplink, what is the action if this uplink is down ? will the vNIC:
    > a.       Brought down , as per the Network Control policy applied , and in this case the OS will go for the second vNIC
    If you are using static pin group, Yes.
    If you are not using static pin groups, the same FI will map it to another available uplink.
    Why? Because by defining static pinning you are purposely defining the uplink/subscription ratio etc and you don't want that vNIC to go to any other uplink. Both fabrics are active at any given time.
    > b.      FI failover to the second FI , the OS will not detect anything.
    Yes.
    > c.       The FI A will re-pin the vNIC to another uplink on the same FI with no failover
    For dynamic pinning yes. For static pinning NO as above.
    >I found all theses 3 scenarios in a different documents and posts, I did not have the chance it to test it yet, so it will be great if anyone tested it and >can explain.
    I would still highly recommend testing it. Maybe its me but I don't believe anything till I have tried it.
    > Finally I need to know if the correct scenarios from the above will be applied to the vHBA or it has another methodology.
    Multipathing driver as I mentioned before.
    FF *only* applies to ethernet.
    Thanks
    --Manish

  • Fault message types: Why do we need them?

    Hi guys,
    I have a bpm scenario created, tested and working.
    This bpm starts with a async receive, a Sync Send to a R3 system by an ABAP proxy and a final async send to the original system.
    I'm contemplating exception handling and other type of errors. Meaning, if the abap proxy dumps I can raise an exception and map it on the XI environment. The thing is, on  sxmb_moni I get the exception handler, with the exception texts, id's, etc... from the R3 being mapped to the XI fault message type.
    But the problem is I can't get to map the text from the fault message type into any field from the response of the original system.
    I've searched the forum and I've seen that this area is surrounded by doubts and questions...
    But in the end I'm forced to place the question, why do I need the fault message types? What use can I give them, besides of the cute little thunder on the sxmb_moni?
    Thanks for your attention guys

    Hi Krishna,
    I've read the SAP description about Fault Message Types but I'm still struggling on how to use them correctly.
    Why do we need them if I can't do nothing with it on the BPM side? After the mapping of the FMT, I'd like to acess it so that I can send a description back to the original system...The thing is, in BPM you can't acess the FMT, you can only create a new BRANCH (on a block step) for it...
    I've tried to acess the description text on trace which comes with the message, but it seems that the trace only serves for posting info, and not for getting any...
    The Mapping Runtime Constants have some information where you can read some descriptions, but nothing related to the errors....You can read the Receiver service, sender, interface_namespace, but nothing from the ERROR tag...
    Is there a way to get the description? Or anything related to the error tag?
    Thanks in advance

  • Solaris 10 Cluster 3.2 with  2 zones in a failover scenario

    Hi
    Looking for the best way to set up things for the following scenario
    I have 2 M5000 servers with internal storage and a 6140 array for shared storage
    I need to create 2 zones on each in a failover scenario (active /standby)
    On Server1 3 out of 4 cpus for Oracle Database Server 11g and 1 out of 4 cpus for Oracle Application Server
    On Server2 3 out of 4 cpus for Oracle Application Server and 1 out of 4 cpus for Oracle Database Server 11g
    Database files will be placed on the shared storage. In case of failure of Server1 Oracle Database will fail over to Server2 and in case Server2 is down Oracle Application Server will fail to Server1.
    Would a zone cluster using clzonecluster be better?if yes how can i achieve the difference in cpu power in case of failure.
    where is best to keep the zone root path on the internal storage or on the shared storage?
    What about the swap space for both zones?
    Better use exclusive ips or shared will be fine?
    Will it be better to have sparse zone installation for the zone or do the whole thing?
    What is the best way to achieve the cpu assignments needed and how much should be left for the global zone?
    Thanks in advance
    vangelis

    Hi Vangelis,
    Building a cluster, requires some planning and understanding the concepts.
    A good start would be reading some of the documents linked to in this url: http://docs.sun.com/app/docs/doc/819-2969/gcbkf?a=view
    Regards,
    Davy

  • Why do we need to rebalance disks manually when we shutdown the storage

    Dear all,
    Here is my scenario.
    We have extended RAC setup. Assuming everything is going fine and you have following status:
    SQL> select MOUNT_STATUS,HEADER_STATUS,MODE_STATUS,STATE,NAME,FAILGROUP,PATH from v$asm_disk;
    MOUNT_S HEADER_STATU MODE_ST STATE NAME FAILGROUP PATH
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote2
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote3
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote1
    CACHED MEMBER ONLINE NORMAL DATADG_0003 DATAFG2 /dev/rdisk/asm_data1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0001 FLASHFG2 /dev/rdisk/asm_arch1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0000 FLASHFG1 /dev/rdisk/asm_arch1a
    CACHED MEMBER ONLINE NORMAL DATADG_0000 DATAFG1 /dev/rdisk/asm_data1a
    Now you shutdown one of the storages and have following status:
    $ cd /u01/app/oracle/crs/log/hisdbs01/
    $ ls
    admin alerthisdbs01.log client crsd cssd evmd racg
    $ tail -f alerthisdbs01.log
    2008-06-14 17:00:05.710
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 17:00:05.710
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 17:01:05.720
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 17:01:05.720
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 17:02:09.241
    [cssd(2757)]CRS-1605:CSSD voting file is online: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 18:14:37.660
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 18:14:37.660
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    SQL> select MOUNT_STATUS,HEADER_STATUS,MODE_STATUS,STATE,NAME,FAILGROUP,PATH from v$asm_disk;
    MOUNT_S HEADER_STATU MODE_ST STATE NAME FAILGROUP PATH
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote2
    CLOSED UNKNOWN ONLINE NORMAL /dev/rdisk/asm_data1a
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote3
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote1
    CACHED MEMBER ONLINE NORMAL DATADG_0003 DATAFG2 /dev/rdisk/asm_data1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0001 FLASHFG2 /dev/rdisk/asm_arch1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0000 FLASHFG1 /dev/rdisk/asm_arch1a
    MISSING CANDIDATE OFFLINE FORCING DATADG_0000
    8 rows selected.
    sql> select MOUNT_STATUS,HEADER_STATUS,MODE_STATUS,STATE,NAME,FAILGROUP,PATH from v$asm_disk;
    MOUNT_S HEADER_STATU MODE_ST STATE NAME FAILGROUP PATH
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote3
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote1
    CLOSED UNKNOWN ONLINE NORMAL /dev/rdisk/asm_data1a
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote2
    CACHED MEMBER ONLINE NORMAL FLASHDG_0000 FLASHFG1 /dev/rdisk/asm_arch1a
    CACHED MEMBER ONLINE NORMAL DATADG_0003 DATAFG2 /dev/rdisk/asm_data1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0001 FLASHFG2 /dev/rdisk/asm_arch1b
    MISSING CANDIDATE OFFLINE HUNG DATADG_0000
    8 rows selected.
    After you have verified that the state is "HUNG", you startup the storage
    ========================
    $ cd /u01/app/oracle/crs/log/hisdbs01/
    $ ls
    admin alerthisdbs01.log client crsd cssd evmd racg
    $ tail -f alerthisdbs01.log
    2008-06-14 18:29:37.810
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 18:29:37.810
    [cssd(2757)]CRS-1604:CSSD voting file is offline: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    2008-06-14 18:30:41.241
    [cssd(2757)]CRS-1605:CSSD voting file is online: /dev/rdisk/crs_vote1. Details in /u01/app/oracle/crs/log/hisdbs01/cssd/ocssd.log.
    SQL> select MOUNT_STATUS,HEADER_STATUS,MODE_STATUS,STATE,NAME,FAILGROUP,PATH from v$asm_disk;
    MOUNT_S HEADER_STATU MODE_ST STATE NAME FAILGROUP PATH
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote2
    CLOSED MEMBER ONLINE NORMAL /dev/rdisk/asm_data1a
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote3
    CLOSED FOREIGN ONLINE NORMAL /dev/rdisk/crs_vote1
    CACHED MEMBER ONLINE NORMAL DATADG_0003 DATAFG2 /dev/rdisk/asm_data1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0001 FLASHFG2 /dev/rdisk/asm_arch1b
    CACHED MEMBER ONLINE NORMAL FLASHDG_0000 FLASHFG1 /dev/rdisk/asm_arch1a
    MISSING CANDIDATE OFFLINE HUNG DATADG_0000
    8 rows selected.
    Now my question is WHY do we need to rebalance the asm_data1a MANUALLY because as per above scenario it needs rebalancing. Should not ASM do rebalancing itself?
    Suggest solutions please.
    Thanks a lot. Waiting for your expert opinions.
    br,
    Anjum

    I believe you are using Normal redundancy in this case. When you are making the disk unavailable to Oracle, Oracle tries to drop the disk from diskgroup i.e why you see status as "forcing". After that Oracle recognizes that this disk is required to provide redundancy. At this time the disk goes in Hung state. so when you start the storage, then you have to issue rebalance manually. I am not sure why the design is like that. But I believe it could be due to the fact that disk rebalance takes place only when you add/drop disk. In this case you are not adding a disk but actually making it available at OS level. This will not trigger a rebalance.
    Amit
    http://askdba.org
    http://askdba.org/weblog

  • Why do we need query rewrite enabled for a function-based index?

    Oracle 9i
    ========
    I have searched a few sites but could not find any content on it. The question is why do we need to implement query rewrite enabled when we are trying out a function-based index?
    Thanks in advance.

    You don't, that's a legacy requirement from the early days of function based indexes in Oracle 8i. Here's a quick example running under 9.2.0.6
    drop table t1;
    create table t1 as
    select
    from
         all_objects
    where
         rownum <= 30000
    create or replace function pl_func(i_vc     varchar2)
    return varchar2
    deterministic
    as
    begin
         return soundex(i_vc);
    end;
    -- set the worst case scenario
    alter session set query_rewrite_enabled = false;
    alter session set query_rewrite_integrity = enforced;
    create index t1_i1 on t1(pl_func(object_name));
    execute dbms_stats.gather_table_stats(user, 't1')
    set autotrace traceonly explain
    select
         object_name
    from t1
    where pl_func(object_name) = 'T513'
    set autotrace offResults (after set feedback off)
    SQL> @temp
    Execution Plan
    Plan hash value: 1429545322
    | Id  | Operation                   | Name  | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT            |       |    27 |   675 |    10   (0)| 00:00:01 |
    |   1 |  TABLE ACCESS BY INDEX ROWID| T1    |    27 |   675 |    10   (0)| 00:00:01 |
    |*  2 |   INDEX RANGE SCAN          | T1_I1 |    27 |       |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       2 - access("TEST_USER"."PL_FUNC"("OBJECT_NAME")='T513')
    SQL> spool offRegards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk

  • Why do we need to have a observer run on the different computer

    Hello team,
    Why do we need to have a observer run on the different computer and how to install & confgure it on that different computer.?pls help
    5.5.6 Managing the Observer
    The observer is integrated in the DGMGRL client-side component and runs on a different computer from the primary or standby databases and from the computer where you manage the broker configuration. The observer continuously monitors the fast-start failover environment to ensure the primary database is available (described in Section 5.5.2.1). The observer's main purpose is to enhance high availability and lights out computing by reducing the human intervention required by the manual failover process that can add minutes or hours to downtime.
    Thanks

    854393 wrote:
    Thanks Aman,
    Do we need a same db binary & OS version for starting the observer on the different computer ?? here the primary & standby are liunx.
    Thanks in advance..
    Regards,it does not have to be on the same platform or operating system on which the databases reside, and it does not need Oracle EE and an instance (which means no extra license).
    To configure check this.
    http://gjilevski.wordpress.com/2010/03/06/configuring-10g-data-guard-broker-and-observer-for-failover-and-switchover/

  • Why do we need to upload meta data through IDX2

    Hi Experts,
    Why do we need to uplaod the meta data through idx2 ,if we won't do that what will be the impact on oue scenario.
    Regards

    The IDOC msg comes as a data stream from ECC to PI. PI uses the metadata in IDX2 to convert the IDOC msg to IDOC-XML msg...
    http://help.sap.com/saphelp_nw04/helpdata/EN/ab/bdb13b00ae793be10000000a11402f/content.htm
    If you don't load IDX2, When the 1st IDOC is received, the metadata is automatically loaded.
    For subsequent changes in ECC, (If new Z-fields are added in ECC), they are not reflected in PI, if you don't perform IDX2. Hence, IDOC msg to IDOC-XML convertion will have some problems.. Data will get into wrong fields... It is recommended to Perform IDX2 to aviod such issues..
    Regards,
    Siva Maranani

  • Why do you need a logical plant?

    Hi Guys
    Please explain me with an example, Why do I need a logical plant in SAP. If i decide to have one, can I have a logical plant without a valuation and then on moving the stock to the original plant it should get valuated. I raise this discussion my implentation is affected because of wrong valuation of stock. Please express your confident views regarding.
    Regards
    Pradeep/ICF

    Hi,
    Not sure why you need a logical plant. Normally dummy plants are made to facilitate material movements to cater complex scenarios. In OMS2 you can omit any Material type with regards to the plant as "Value updating" .
    Regards
    Binoy

  • Service Registry: why do i need two bindings (wsdlDeployment and endPoint)

    based on the technical note of Clemens Utschig "Enabling a Dynamic, Reusable SOA with Oracle BPEL Process Manager and Oracle Service Registry" i have one question:
    why do i need two bindings (wsdlDeployment and endPoint) in my service in the service registry. if i only define the wsdlDeployment if get this messages:
    <2006-12-07 09:39:02,559> <FATAL> <default.collaxa.cube.ws> no endpoint found - nothing to adjust
    <2006-12-07 09:39:02,559> <WARN> <default.collaxa.cube.ws> Could not lookup base endpoint!
    the problem of defining two bindings is that the url to my partner is redundant.

    Hi Clemens,
    The UDDI Version 3.0.2 Specification (http://www.uddi.org/pubs/uddi_v3.htm) say this:
    B.1.2 Using the "wsdlDeployment" value
    Instead of directly providing the network address in the accessPoint, it is occasionally useful or necessary to provide this information through indirect means. One common scenario for such a behavior is when the accessPoint is embedded within a WSDL file. In such a scenario, the UDDI accessPoint contains the address of the WSDL file, and the client then must retrieve the WSDL file and extract the end point address from the WSDL file itself.
    In this case, decorating the UDDI accessPoint with a useType="wsdlDeployment" is appropriate. A sample of such behavior is as follows:
    <bindingTemplate bindingKey="uddi:example.org:catalog">
       <description xml:lang="en">
           Browse catalog Web service
       </description>
       <accessPoint useType="wsdlDeployment">
           http://www.example.org/CatalogWebService/catalog.wsdl
       </accessPoint>
       <categoryBag>
           <keyedReference keyName="uddi-org:types:wsdl"
           keyValue="wsdlDeployment"
           tModelKey="uddi:uddi.org:categorization:types"/>
       </categoryBag>
    </bindingTemplate>In the example above, a client would be able to parse the result of the bindingTemplate and determine the end point of the Web service within the WSDL file discovered in the accessPoint element. Note that the bindingTemplate has also been categorized with the "wsdlDeployment" value from the uddi.org:categorization:types scheme so that it can be discovered through a find_binding API call.
    1) This means for me that providing one binding with useType=wsdlDeployment should be enough.
    2) Sadly the "Business Service Control" of the Service Registry (Publish new WSDL service) does not create the correct structure needed by the bpel-engine. The "Publish new WSDL service" wizard just creates one service with one binding with useType=other (http). So if i want to reuse this published service from bpel (as a service consumer) i have to switch to the "Service Registry Admin Control" and create a second binding with useType=wsdlDeployment. But i don't want the developers to get inside the "Service Registry Admin Control"! They should only work with the "Business Service Control".
    All this makes the usage of the service-registry very unhandy and reduce the acceptance for developers because of unnecessary manual steps and broken security principles.
    My wish for this issue is that the "Publish new WSDL service" inside the "Business Service Control" of the Service Registry would additionlly create this needed binding with useType=wsdlDeployment.
    Best Regards,
    Harald Reinmueller

  • Why there is need to define processing code (we41/we42) in Idocs?

    Hi All,
      Why there is need to define the processing codes when we already have defined the processing function for the IDOC type & message type combinations via we57.

    Hi,
    http://help.sap.com/saphelp_nw04/helpdata/en/dc/6b804043d711d1893e0000e8323c4f/content.htm
    The inbound function module needs to be linked to the message type and the mess type needs to b linked to the appropriate process code at the partner oprofile level before  any scenario is enabled.
    Regards,
    Gayathri

  • AlwaysOn Failover Scenarios

    Hi,
    I have implemented AlwaysOn feature between two standalone SQL Server instances hosted on two Clustered nodes in two different subnets (multisite clustering with Node and File Share Majority quorum). I have configured AlwaysOn for
    automatic failover between the primary and the only secondary replica. There are two databases in it. The implementation did went successful.
    Now, I before going live, I wanted to test the failover scenarios. First one was to manually failover between the nodes from SQL Server side, as well as from Failover Cluster Manager consoles. Both went perfect.
    But, the second test was to stop the SQL Server service, and see the result. When I stopped the primary, the resource group in the Cluster Manager failed, and the databases were also not connecting. I expected the Availability group should have
    failed over to the secondary node, and the db's should have been up and running.
    Did I miss something in the implementation part above, or it is an expected behavior of AlwaysON? If yes, then what does automatic failover imply?
    Thanks  & Regards

    Hi,
    I have implemented AlwaysOn feature between two standalone SQL Server instances hosted on two Clustered nodes in two different subnets (multisite clustering with Node and File Share Majority quorum). I have configured AlwaysOn for
    automatic failover between the primary and the only secondary replica. There are two databases in it. The implementation did went successful.
    Now, I before going live, I wanted to test the failover scenarios. First one was to manually failover between the nodes from SQL Server side, as well as from Failover Cluster Manager consoles. Both went perfect.
    But, the second test was to stop the SQL Server service, and see the result. When I stopped the primary, the resource group in the Cluster Manager failed, and the databases were also not connecting. I expected the Availability group should have
    failed over to the secondary node, and the db's should have been up and running.
    Did I miss something in the implementation part above, or it is an expected behavior of AlwaysON? If yes, then what does automatic failover imply?
    Thanks  & Regards
    upon re-reading the question.. i think we need some clarity..what do you mean failover  node..you do not fail over over the node,, you fail over the services running on the node but in your case, you said you have  stand alone sql instances - which
    is required for always on.. so, by fail over node - you mean you took node OFFLINE - either take the sql service OFFLINE and/or entire NODE offline... what did the Always ON status show when you connected to the sql instance running on another node. it should
    be now primary.
    when you take first node offline or one sql service offline, the sql service will not fail over because is not sql cluster, it will fail over only the database that are set up for always on. not even all the database,just the ones set up for it... 
    in other words, it will failover the "always on" service...which takes care of the AG databases on the node to determine which is priiary/secondary
    you dashboard on the secondary instance - AG dashboard should tell you the status,,,
    Hope it Helps!!

  • Failover scenarios for AlwaysOn

    What are the failover scenarios/situations in which the failover happens when the Databases are configured with ALways-On?
    In our situation windows failover cluster server node 1 fails over to node 2 but the databases are still pointing to node 1.
    thanks

    Hi,
    When failure occurs, whether an availability group will failover immediately depends on both
    the failover mode and the availability mode of the replica.
    Please check below articles for more information.
    http://msdn.microsoft.com/en-us/library/hh213151.aspx#Overview
    After a failover, client applications that need to access the primary databases must connect to the new primary replica. Also, if the new secondary replica is configured to allow read-only access, read-only client applications can connect to it. For information
    about how clients connect to an availability group, see
    Availability Group Listeners, Client Connectivity, and Application Failover (SQL Server).
    Hope the information helps.
    Tracy Cai
    TechNet Community Support

Maybe you are looking for