ONE SOURCE TWO DESTINATIONS

Hello
        my scenario is source JMS adapter, i want to  send this source file to two destinations, one is Database other one is JMS target adapter. can anyone  explain wht are the steps required.
Thanks

Hey
u can use conditional receiver determination for this,
if u have some conditions based upon which you want to branchc off the message then give that condition over there,if u simply want to send every message to both the receivers then just give the business system name for both the receivers,you do need 2 interface determination and 2 receiver determination for this
/people/shabarish.vijayakumar/blog/2005/08/03/xpath-to-show-the-path-multiple-receivers
Thanx
Aamir
Message was edited by:
        Aamir Suhail

Similar Messages

  • SLT - Splitting one source table into two tables in the destination

    Hi,
    I am wondering if we can split content of one source table into two different tables in the destination (HANA DB in my case) with SLT based on the codified mapping rules?
    We have the VBAK table in ERP which has the header information about various business objects (quote, sales order, invoice, outbound delivery to name a few). I want this to be replicated into tables specific to business object (like VBAK_QUOT, VBAK_SO, VBAK_INV, etc) based on document type column.
    There is one way to do it as far as i know - have multiple configurations and replicate to different schema. But we might have to be content with 4 different config at the max.
    Any help here will be highly appreciated
    Regards,
    Sesh

    Please take a look at these links related to your query.
    http://stackoverflow.com/questions/5145637/querying-data-by-joining-two-tables-in-two-database-on-different-servers
    http://stackoverflow.com/questions/7037228/joining-two-tables-together-in-one-database

  • Getting "IPSEC(epa_des_crypt): decrypted packet failed SA identity check" messages on packets from only one of two far-end sources sharing the same tunnel, the other source works fine. What exactly does this error mean?

    One computer at COMPANY-A is attempting to communicate with two
    computers located at COMPANY-B, via an IPsec tunnel between the
    two companies.
    All communications are via TCP protocol.
    All devices present public IP addresses to one another, although they
    may have RFC 1918 addresses on other interfaces, and NAT may be in use
    on the COMPANY-B side.  (NAT is not being used on the COMPANY-A side.)
    The players:(Note: first three octets have been changed for security reasons)
    COMPANY-A computer      1.2.3.161
    COMPANY-A router        1.2.3.8 (also IPsec peer)
    COMPANY-A has 1.2.3.0/24 with no subnetting.
    COMPANY-B router        4.5.6.228 (also IPsec peer)
    COMPANY-B computer #1   4.5.7.94 (this one has no issues)
    COMPANY-B computer #2   4.5.7.29 (this one fails)
    COMPANY-B has 4.5.6.0/23 subnetted in various ways.
    COMPANY-B also has 9.10.11.0/24, but it is not involved in the issue.
    What works:
    The COMPANY-A computer 1.2.3.161 can communicate via the single IPsec
    tunnel to COMPANY-B computer #1 4.5.7.94 without problems.
    The "show crypto session detail" command shows Inbound/Outbound packets
    flowing in the dec'ed and enc'ed positions.
    What doesn't:
    When the COMPANY-A computer 1.2.3.161 attempts to communicate
    via the single IPsec tunnel with the COMPANY-B computer #2 4.5.7.29,
    the COMPANY-A router eventually reports five of these messages:
    Oct  9 15:24:54.327: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    Oct  9 15:24:57.327: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    Oct  9 15:25:03.327: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    Oct  9 15:25:15.328: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    Oct  9 15:25:39.329: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    Oct  9 15:26:27.328: IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    and the "show crypto session detail" shows inbound packets being dropped.
    The COMPANY-A computer that opens the TCP connection never gets past the
    SYN_SENT phase of the TCP connection whan trying to communicate with the
    COMPANY-B computer #2, and the repeated error messages are the retries of
    the SYN packet.
    On the COMPANY-A side, this IPsec configuration has been set up on a 3745,
    a 3725, and some 76xx routers were tried, all with similar behavior,
    with packets from one far-end computer passing fine, and packets from
    another far-end computer in the same netblock passing through the same
    IPsec tunnel failing with the "failed SA identity" error.
    The COMPANY-A computer directs all packets headed to COMPANY-B via the
    COMPANY-A router at 1.2.3.8 with this set of route settings:
    netstat -r -n
    Kernel IP routing table
    Destination     Gateway         Genmask         Flags   MSS Window  irtt Iface
    4.5.7.0         1.2.3.8         255.255.255.0   UG        0 0          0 eth3
    1.2.3.8.0       0.0.0.0         255.255.255.0   U         0 0          0 eth3
    10.1.0.0        0.0.0.0         255.255.240.0   U         0 0          0 eth0
    169.254.0.0     0.0.0.0         255.255.0.0     U         0 0          0 eth3
    10.0.0.0        10.1.1.1        255.0.0.0       UG        0 0          0 eth0
    0.0.0.0         1.2.3.1         0.0.0.0         UG        0 0          0 eth3
    The first route line shown is selected for access to both COMPANY-B computers.
    The COMPANY-A router (IPsec tunnel endpoint, 1.2.3.8) has this
    configuration:
    crypto isakmp policy 10
    encr 3des
    authentication pre-share
    group 2
    lifetime 28800
    crypto isakmp key XXXXXXXXXXXXXXXXXXXXXXX address 4.5.6.228
    crypto ipsec security-association lifetime seconds 86400
    crypto ipsec transform-set COMPANY-B01 esp-3des esp-sha-hmac
    crypto map COMPANY-BMAP1 10 ipsec-isakmp
    description COMPANY-B VPN
    set peer 4.5.6.228
    set transform-set COMPANY-B01
    set pfs group2
    match address 190
    interface FastEthernet0/0
    ip address 1.2.3.8 255.255.255.0
    no ip redirects
    ip virtual-reassembly
    duplex auto
    speed auto
    no cdp enable
    crypto map COMPANY-BMAP1
    ip forward-protocol nd
    ip route 0.0.0.0 0.0.0.0 1.2.3.1
    ip route 10.0.0.0 255.0.0.0 10.1.1.1
    ip route 1.2.3.8.0 255.255.255.0 FastEthernet0/0
    access-list 190 permit ip host 1.2.3.161 4.5.7.0 0.0.0.255
    access-list 190 permit ip host 1.2.3.161 9.10.11.0 0.0.0.255
    bridge 1 protocol ieee
    One of the routers tried had this IOS/hardware configuration:
    Cisco IOS Software, 3700 Software (C3725-ADVIPSERVICESK9-M), Version 12.4(25c),
    RELEASE SOFTWARE (fc2)
    isco 3725 (R7000) processor (revision 0.1) with 115712K/15360K bytes of memory.
    Processor board ID XXXXXXXXXXXXXXX
    R7000 CPU at 240MHz, Implementation 39, Rev 3.3, 256KB L2 Cache
    2 FastEthernet interfaces
    4 ATM interfaces
    DRAM configuration is 64 bits wide with parity disabled.
    55K bytes of NVRAM.
    31296K bytes of ATA System CompactFlash (Read/Write)
    250368K bytes of ATA Slot0 CompactFlash (Read/Write)
    Configuration register is 0x2102
    #show crypto sess
    Crypto session current status
    Interface: FastEthernet0/0
    Session status: UP-ACTIVE
    Peer: 4.5.6.228 port 500
      IKE SA: local 1.2.3.8/500 remote 4.5.6.228/500 Active
      IPSEC FLOW: permit ip host 1.2.3.161 4.5.7.0/255.255.255.0
            Active SAs: 2, origin: crypto map
      IPSEC FLOW: permit ip host 1.2.3.161 9.10.11.0/255.255.255.0
            Active SAs: 0, origin: crypto map
    #show crypto sess det
    Crypto session current status
    Code: C - IKE Configuration mode, D - Dead Peer Detection
    K - Keepalives, N - NAT-traversal, X - IKE Extended Authentication
    Interface: FastEthernet0/0
    Session status: UP-ACTIVE
    Peer: 4.5.6.228 port 500 fvrf: (none) ivrf: (none)
          Phase1_id: 4.5.6.228
          Desc: (none)
      IKE SA: local 1.2.3.8/500 remote 4.5.6.228/500 Active
              Capabilities:(none) connid:1 lifetime:06:26:27
      IPSEC FLOW: permit ip host 1.2.3.161 4.5.7.0/255.255.255.0
            Active SAs: 2, origin: crypto map
            Inbound:  #pkts dec'ed 651 drop 16 life (KB/Sec) 4496182/23178
            Outbound: #pkts enc'ed 574 drop 2 life (KB/Sec) 4496279/23178
      IPSEC FLOW: permit ip host 1.2.3.161 9.10.11.0/255.255.255.0
            Active SAs: 0, origin: crypto map
            Inbound:  #pkts dec'ed 0 drop 0 life (KB/Sec) 0/0
            Outbound: #pkts enc'ed 0 drop 0 life (KB/Sec) 0/0
    The COMPANY-B device on their end of the IPsec VPN is a Juniper SSG1000
    Version 6.1 (ScreenOS)
    We only have a limited view into the Juniper device configuration.
    What we were allowed to see was:
    COMPANY-B-ROUTER(M)-> sh config | incl COMPANY-A
    set address "Untrust" "oss-COMPANY-A-1.2.3.161" 1.2.3.161 255.255.255.255
    set ike gateway "COMPANY-A-1-GW" address 1.2.3.8 Main outgoing-interface "ethernet2/1" preshare xxxxxxxxxxxxxxxxxxxxxx  proposal "pre-g2-3des-sha"
    set vpn "COMPANY-A-1-IKE" gateway "COMPANY-A-1-GW" no-replay tunnel idletime 0 proposal "g2-esp-3des-sha-28800"
    set policy id 2539 from "Untrust" to "Trust"  "oss-COMPANY-A-1.2.3.161" "9.10.11.0-24" "ANY" tunnel vpn "COMPANY-A-1-IKE" id 0x309a pair-policy 2500
    set policy id 2500 from "Trust" to "Untrust"  "9.10.11.0-24" "oss-COMPANY-A-1.2.3.161" "ANY" tunnel vpn "COMPANY-A-1-IKE" id 0x309a pair-policy 2539
    set policy id 2541 from "Trust" to "Untrust"  "4.5.7.0-24" "oss-COMPANY-A-1.2.3.161" "ANY" tunnel vpn "COMPANY-A-1-IKE" id 0x309b pair-policy 2540
    set policy id 2540 from "Untrust" to "Trust"  "oss-COMPANY-A-1.2.3.161" "4.5.7.0-24" "ANY" tunnel vpn "COMPANY-A-1-IKE" id 0x309b pair-policy 2541
    COMPANY-B-ROUTER(M)->
    I suspect that this curious issue is due to a configuration setting on the
    Juniper device, but neither party has seen this error before.  COMPANY-B
    operates thousands of IPsec VPNs and they report that this is a new error
    for them too.  The behavior that allows traffic from one IP address to
    work and traffic from another to end up getting this error is also unique.
    As only the Cisco side emits any error message at all, this is the only
    clue we have as to what is going on, even if this isn't actually an IOS
    problem.
    What we are looking for is a description of exactly what the Cisco
    IOS error message:
    IPSEC(epa_des_crypt): decrypted packet failed SA identity check
    is complaining about, and if there are any known causes of the behavior
    described that occur when running IPsec between Cisco IOS and a Juniper
    SSG device.  Google reports many other incidents of the same error
    message (but not the "I like that IP address but hate this one" behavior),
    and not just with a Juniper device on the COMPANY-B end, but for those cases,
    not one was found where the solution was described.
    It is hoped that with a better explanation of the error message
    and any known issues with Juniper configuration settings causing
    this error, we can have COMPANY-B make adjustments to their device.
    Or, if there is a setting change needed on the COMPANY-A router,
    that can also be implemented.
    Thanks in advance for your time in reading this, and any ideas.

    Hello Harish,
    It is believed that:
    COMPANY-B computer #1   4.5.7.94 (this one has no issues)
    COMPANY-B computer #2   4.5.7.29 (this one fails)
    both have at least two network interfaces, one with a public IP address
    (which we are supposedly conversing with) and one with a RFC 1918 type
    address.   COMPANY-B is reluctant to disclose details of their network or
    servers setup, so this is not 100% certain.
    Because of that uncertainty, it occurred to me that perhaps COMPANY-B
    computer #2 might be incorrectly routing via the RFC 1918 interface.
    In theory, such packets should have been blocked by the access-list on both
    COMPANY-A router, and should not have even made it into the IPsec VPN
    if the Juniper access settings work as it appears they should.  So I turned up
    debugging on COMPANY-A router so that I could see the encrypted and
    decrypted packet hex dumps.
    I then hand-disassembled the decoded ACK packet IP header received just
    prior to the "decrypted packet failed SA check" error being emitted and
    found the expected source and destination IP addresses (4.5.7.29 and 1.2.3.161),
    in the unecapsulated packet.  I also found the expected port numbers of the TCP
    conversation that was trying to be established in the TCP header.  So, it
    looks like COMPANY-B computer #2 is emitting the packets out the right
    interface.
    The IP packet header of the encrypted packet showed the IP addresses of the
    two routers at each terminus of the IPsec VPN, but since I don't know what triggers
    the "SA check" error message or what it is complaining about, I don't know what
    other clues to look for in the packet dumps.
    As to your second question, "can you check whether both encapsulation and
    decapsulation happening in 'show crypto ipsec sa'",   the enc'ed/dec'ed
    counters were both going up by the correct quantities.  When communicating
    with the uncooperative COMPANY-B computer #2, you would also see the
    received Drop increment for each packet decrypted.  When communicating
    with the working COMPANY-B computer #1, the Drop counters would not
    increment, and the enc'ed/dec'ed would both increment.
    #show crypto sess det
    Crypto session current status
    Code: C - IKE Configuration mode, D - Dead Peer Detection
    K - Keepalives, N - NAT-traversal, X - IKE Extended Authentication
    Interface: FastEthernet0/0
    Session status: UP-ACTIVE
    Peer: 4.5.6.228 port 500 fvrf: (none) ivrf: (none)
          Phase1_id: 4.5.6.228
          Desc: (none)
      IKE SA: local 1.2.3.8/500 remote 4.5.6.228/500 Active
              Capabilities:(none) connid:1 lifetime:07:59:54
      IPSEC FLOW: permit ip host 1.2.3.161 4.5.7.0/255.255.255.0
            Active SAs: 2, origin: crypto map
            Inbound:  #pkts dec'ed 376 drop 5 life (KB/Sec) 4458308/28784
            Outbound: #pkts enc'ed 401 drop 3 life (KB/Sec) 4458308/28784
    Attempt a TCP communication to COMPANY-B computer #2...
    show crypto sess det
    Crypto session current status
    Code: C - IKE Configuration mode, D - Dead Peer Detection
    K - Keepalives, N - NAT-traversal, X - IKE Extended Authentication
    Interface: FastEthernet0/0
    Session status: UP-ACTIVE
    Peer: 4.5.6.228 port 500 fvrf: (none) ivrf: (none)
          Phase1_id: 4.5.6.228
          Desc: (none)
      IKE SA: local 1.2.3.8/500 remote 4.5.6.228/500 Active
              Capabilities:(none) connid:1 lifetime:07:59:23
      IPSEC FLOW: permit ip host 1.2.3.161 4.5.7.0/255.255.255.0
            Active SAs: 2, origin: crypto map
            Inbound:  #pkts dec'ed 376 drop 6 life (KB/Sec) 4458307/28753
            Outbound: #pkts enc'ed 402 drop 3 life (KB/Sec) 4458307/28753
    Note Inbound "drop" changed from 5 to 6.  (I didn't let it sit for all
    the retries.)
    #show crypto ipsec sa
    interface: FastEthernet0/0
        Crypto map tag: COMPANY-BMAP1, local addr 1.2.3.8
       protected vrf: (none)
       local  ident (addr/mask/prot/port): (1.2.3.161/255.255.255.255/0/0)
       remote ident (addr/mask/prot/port): (4.5.7.0/255.255.255.0/0/0)
       current_peer 4.5.6.228 port 500
         PERMIT, flags={origin_is_acl,}
        #pkts encaps: 402, #pkts encrypt: 402, #pkts digest: 402
        #pkts decaps: 376, #pkts decrypt: 376, #pkts verify: 376
        #pkts compressed: 0, #pkts decompressed: 0
        #pkts not compressed: 0, #pkts compr. failed: 0
        #pkts not decompressed: 0, #pkts decompress failed: 0
        #send errors 3, #recv errors 6
         local crypto endpt.: 1.2.3.8, remote crypto endpt.: 4.5.6.228
         path mtu 1500, ip mtu 1500, ip mtu idb FastEthernet0/0
         current outbound spi: 0xDF2CC59C(3744253340)
      inbound esp sas:
          spi: 0xD9D2EBBB(3654478779)
            transform: esp-3des esp-sha-hmac ,
            in use settings ={Tunnel, }
            conn id: 2004, flow_id: SW:4, crypto map: COMPANY-BMAP1
            sa timing: remaining key lifetime (k/sec): (4458307/28600)
            IV size: 8 bytes
            replay detection support: Y
            Status: ACTIVE
         inbound ah sas:
         inbound pcp sas:
         outbound esp sas:
          spi: 0xDF2CC59C(3744253340)
            transform: esp-3des esp-sha-hmac ,
            in use settings ={Tunnel, }
            conn id: 2003, flow_id: SW:3, crypto map: COMPANY-BMAP1
            sa timing: remaining key lifetime (k/sec): (4458307/28600)
            IV size: 8 bytes
            replay detection support: Y
            Status: ACTIVE
         outbound ah sas:
         outbound pcp sas:
    The "send" errors appear to be related to the tunnel reverting to a
    DOWN state after periods of inactivity, and you appear to get one
    each time the tunnel has to be re-negotiated and returned to
    an ACTIVE state.  There is no relationship between Send errors
    incrementing and working/non-working TCP conversations to the
    two COMPANY-B servers.
    Thanks for pondering this very odd behavior.

  • How do i create one source file that feeds two users on the same pro book?

    how do i create one source file/folder of photos that feeds two users on the same pro book?

    Well you can do exactly the same with a folder of photos, just put in the Users/Shared folder.
    Iphoto is a mite more complex than this. It's a database and so there's a bit more involved.
    For iPhoto 09 (version 8.0.2) and later:
    What you mean by 'share'.
    If you want the other user to be able to see the pics, but not add to, change or alter your library, then enable Sharing in your iPhoto (Preferences -> Sharing), leave iPhoto running and use Fast User Switching to open the other account. In that account, enable 'Look For Shared Libraries'. Your Library will appear in the other source pane.
    Any user can drag a pic from the Shared Library to their own in the iPhoto Window.
    Remember iPhoto must be running in both accounts for this to work.
    If you want the other user to have the same access to the library as you: to be able to add, edit, organise, keyword etc.
    Quit iPhoto in both accounts. Move the Library to the Users / Shared Folder
    (You can also use an external HD set to ignore permissions, a Disk Image or even partition your Hard Disk.)
    In each account in turn: Double click on the Library to open it. (You may be asked to repair the Library Permissions.) From that point on, this will be the default library location. Both accounts will have full access to the library, in fact, both accounts will 'own' it.
    However, there is a catch with this system and it is a significant one. iPhoto is not a multi-user app., it does not have the code to negotiate two users simultaneously writing to the database, and trying will cause db corruption. So only one user at a time, and back up, back up back up.

  • One target table is loading from two different source but same columns but one source is in a database and other is in a flat file.

    Hope you all are doing good.
    I have a business issue to be implemented in ODI 11G. Here it is. I am trying to load a target table from two sources having same coulmn names. But one source is in file format and other is in Oracle Database.
    This is what i think i will create two mappings in the same interface using Union between the sources. But i am not sure how the interface would connect to different logical architecture to connect to two different sources.
    Thanks,
    SM

    You are on the right track, this can all be done in a single interface. Do the following
    1) Pull your file data model into the source designer and and your target table model to the target pane.
    2) Map all the relevant columns
    3) In the source designer create a new dataset and choose the UNION join type (this will create a separate tab in the source designer pane)
    4) Select the new dataset tab in the source designer pane and pull your source oracle table data model into the source designer. Map all the relevant columns to the target
    5) Make sure that your staging location is defined on a relational technology i.e. in this case the target would be an ideal candidate as that is where ODI will stage the data from both file and oracle source and perform the UNION before loading to the target
    If you want to look at some pretty screenshots showing the steps above take a look at http://odiexperts.com/11g-oracle-data-integrator-part-611g-union-minus-intersect/

  • Connect two BW systems to one source system

    Hi Experts!
    I would like to know if I can set up two BW systems connect to one source system.  We copied and built new BW system (BW2) from existing BW system (BW1).  We need to have both system connect to one source system.  When I execut Restore from BW2 - RSA1 - Source System, I'm getting a following prompt:
    "Connection cannot be used. The connection BB is used in the DEV400 source system as a connection BW1400 to BW. Do you want to delete this connection in source system".
    If I select "Delete" then it replace exising connection between BW1 - DEV and creates new connection BW2 - DEV.  If I go back to BW1 and do restore then same thing happens as BW2 connection removed.
    Is there anyway we can set up two BW systems connection to one source system with restore?  Connection works fine since each one can connection with no issues.  Just when I do restore, it removes other connection.
    Thanks.

    Check if it helps:
    Loading two different SAP BW system from one SAP R/3

  • Two BW systems Can use one source sytem?

    Hi Experts,
    We are planning to share one source sytem and its all master and LO data sources to share for Two BW systems, is it possible to share like this, we want only the full loads even for LO data sources can we load to BW sytems, sharing like this, Please clarify
    Thanks in advance

    Hi,
    Deinately it is possible.For more information , refer:
    https://forums.sdn.sap.com/click.jspa?searchID=1527332&messageID=3118492
    With rgds,
    Anil Kumar Sharma .P

  • One thermocouple sourcing two inputs?

    Is it possible to use one thermocouple to source two different input modules?  If so, how should I go about doing this?

    Basically you read the TC EMF eg. voltage and if you have an external cold junction with knows temperature (say maybe an icepoint) you can say you have a pair of copper wires with a voltage source (µV with an internal resistance) .
    You can attach two voltage meters to one source, however in this case your voltage is small and your meters might/will influence each other. (Bias currents, internal resistor, thermal error voltages by the connectors etc). To avoid that, one way would be TC+icepoint+buffer amplifier + 2meters. But you will add the buffer as an error source. I don't know if you can purchase buffers for this application (even if I already have seen one custom build), and dealing with µV isn't impossible, just need careful design.  
    At least it can be done, but it's not simple.
    How about using one TC to current (20mA) converter (industrial standard) and connect two current meters?
    Why do you wan't to do such a thing?
    Greetings from Germany
    Henrik
    LV since v3.1
    “ground” is a convenient fantasy
    '˙˙˙˙uıɐƃɐ lɐıp puɐ °06 ǝuoɥd ɹnoʎ uɹnʇ ǝsɐǝld 'ʎɹɐuıƃɐɯı sı pǝlɐıp ǝʌɐɥ noʎ ɹǝqɯnu ǝɥʇ'

  • SELECT * cannot be used in an INSERT INTO query when the source or destination table contains a multivalued field

    Hi,
    I am using Access 2013 and I have the following VBA code, 
    strSQL = "INSERT INTO Master SELECT * from Master WHERE ID = 1"
     DoCmd.RunSQL (strSQL)
    when the SQL statement is run, I got this error.
    SELECT * cannot be used in an INSERT INTO query when the source or destination table contains a multivalued field
    Any suggestion on how to get around this?
    Please advice and your help would be greatly appreciated!

    Rather than modelling the many-to-many relationship type by means of a multi-valued field, do so by the conventional means of modelling the relationship type by a table which resolves it into two one-to-many relationship types.  You give no indication
    of what is being modelled here, so let's assume a generic model where there is a many-to-many relationship type between Masters and Slaves, for which you'd have the following tables:
    Masters
    ....MasterID  (PK)
    ....Master
    Slaves
    ....SlaveID  (PK)
    ....Slave
    and to model the relationship type:
    SlaveMastership
    ....SlaveID  (FK)
    ....MasterID  (FK)
    The primary key of the last is a composite one of the two foreign keys SlaveID and MasterID.
    You appear to be trying to insert duplicates of a subset of rows from the same table.  With the above structure, to do this you would firstly have to insert rows into the referenced table Masters for all columns bar the key, which, presuming this to be
    an autonumber column, would be assigned new values automatically.  To map these new rows to the same rows in Slaves as the original subset you would then need to insert rows into SlaveMastership with the same SlaveID values as those in Slaves referenced
    by those rows in Slavemastership which referenced the keys of the original subset of rows from Masters, and the MasterID values of the rows inserted in the first insert operation.  This would require joins to be made between the original and the new subsets
    of rows in two instances of Masters on other columns which constitute a candidate key of Masters, so that the rows from SlaveMastership can be identified.
    You'll find examples of these sort of insert operations in DecomposerDemo.zip in my public databases folder at:
    https://onedrive.live.com/?cid=44CC60D7FEA42912&id=44CC60D7FEA42912!169
    If you have difficulty opening the link copy its text (NB, not the link location) and paste it into your browser's address bar.
    In this little demo file non-normalized data from Excel is decomposed into a set of normalized tables.  Unlike your situation this does not involve duplication of rows into the same table, but the methodology for the insertion of rows into a table which
    models a many-to-many relationship type is broadly the same.
    The fact that you have this requirement to duplicate a subset of rows into the same table, however, does make me wonder about the validity of the underlying logical model.  I think it would help us if you could describe in detail just what in real world
    terms is being modelled by this table, and the purpose of the insert operation which you are attempting.
    Ken Sheridan, Stafford, England

  • SSIS 2008 or 2012 Data Task: How to conditionally choose one of two flatfile columns on a per-row basis

    I searched the textbooks and searchable forum posts for something related to my question, but could not find it. I'm hoping when after I describe my question that it might sound familiar to someone.
    First, the context: another department using a black-box database product stopped use of one column and used a different column instead. However, the data in the flat-file export that I need to use for my own database must contain the entire
    table. If it was a clean binary switch, that would be great: I could split the rows into two groups based on date and deal with it that way. Instead, there was a period of overlap while users slowly got the hang of the practice.  This means that I have
    two columns where a row value is either in one column
    or the other.
    Now, the question: I currently use SSIS to import this flatfile into an SQL Server table and use a T-SQL script in an SQL Agent Job to handle the problem.
    However, I was hoping I could do this on the SSIS side while importing the data so that it appears into the SQL Server destination table nice and clean.
    This is a HIGHLY SIMPLIFIED view of my system, using dummy data and dummy names. The T-SQL script describes in essence my source and destination table and what I would like to do on the SSIS side. If a screenshot of my incomplete data tab will help, I can
    do it, but at this point I'm providing what I hope presents the background without presenting distractions.  
    Thank you for any hints or ideas.
    use Learning_Curve;
    go
    if OBJECT_ID('dbo.Relation_A','U') is not null
    drop table dbo.Relation_A;
    create table dbo.Relation_A (
    Rel_A_ID int primary key,
    Rel_A_Value varchar(20) not null,
    Rel_B_Value varchar(20) not null
    insert into dbo.Relation_A values (1,'Unknown','Measured'),(2,'Measured','N/A'),(3,'Unknown','Measured'),
    (4,'Measured','Unknown'),(5,'N/A','Measured'),(6,'Unknown','Measured');
    if OBJECT_ID('dbo.Relation_D','U') is not null
    drop table dbo.Relation_D;
    create table dbo.Relation_D (
    Rel_D_ID int primary key,
    Rel_D_Value varchar(20) not null
    insert into dbo.Relation_D
    select
    a.Rel_A_ID as Rel_D_ID,
    case
    when a.Rel_A_Value = 'N/A' or a.Rel_A_Value='Unknown'
    then a.Rel_B_Value
    else a.Rel_A_Value
    end as Rel_D_Value
    from dbo.Relation_A a;
    go

    You could have posted question in new thread.
    Anyway, you can achieve it by using below tasks shown in attached picture.
    - Merge Join will join by using first ID column
    - Derived Column will have below expressions
    [Column 1] == "" || ISNULL([Column 1]) ? [AnotherColumn 1] : [Column 1]
    [Column 2] == "" || ISNULL([Column 2]) ? [AnotherColumn 1] : [Column 2]
    Cheers,
    Vaibhav Chaudhari
    [MCTS],
    [MCP]

  • Learning to use lookup transformation - match value in one of two columns

    Im a little new to using lookup transforms.  Been reading a bit, but using two columns is throwing me off.
    Have a dataflow with the source having a column I want to use to lookup in a table in another DB.  The match could be in one of two columns in that table.  
    If a match is found, I need a different value from the lookup table added to the source flow that is going into the destination.
    How can I match on one of two columns in the lookup table, and use a totally different column added to the final input going into the destination ?  
    In the Lookup configuration, I can set the lookup operation to replacing the value in the source, but I want EITHER value match from the two columns from the lookup table to go into the value that will go into the destination.
    With TSQL - I would use a case statement and compare the source column to BOTH of the lookup table columns.

    You can handle this within single lookup using below query
    SELECT CASE WHEN <condn1> THEN Field1 ELSE Field2 END AS lookupField,
    CASE WHEN <condn1> THEN Field3 ELSE Field4 END AS MatchField
    FROM Table2
    Then just do single relationship with LookupField and select MatchField to be aded to output
    the condition will specify your fields value condition based on which you determine which column to be matched against
    Please Mark This As Answer if it solved your issue
    Please Mark This As Helpful if it helps to solve your issue
    Visakh
    My MSDN Page
    My Personal Blog
    My Facebook Page

  • After installing SSMS 2014 I can no longer add a source or destination in SSIS VS2010

    I really don't know what is happening but I have been using VS2010 and SSMS 2014 for a couple of months now with no issue. Yesterday when I opened an SSIS package that I had created in VS2010 I was no longer able to add a source or destination to any of
    the data flows. I can copy an existing one and change the properties but when I try to drag it from the tool bow I get the following error -
    TITLE: Microsoft Visual Studio
    The component could not be added to the Data Flow task.
    Could not initialize the component. There is a potential problem in the ProvideComponentProperties method.
    ADDITIONAL INFORMATION:
    Error at "WORKFLOW NAME" [SSIS.Pipeline]: Component ", clsid {GUID}" could not be created and returned error code 0x80070005 "Access is denied.". Make sure that the component is registered correctly.
    Error at "WORKFLOW NAME" [ [128]]: The component is missing, not registered, not upgradeable, or missing required interfaces. The contact information for this component is "".
    Exception from HRESULT: 0xC0048021 (Microsoft.SqlServer.DTSPipelineWrap)
    Program Location:
       at Microsoft.SqlServer.Dts.Pipeline.Wrapper.CManagedComponentWrapperClass.ProvideComponentProperties()
       at Microsoft.DataTransformationServices.Design.PipelineTaskDesigner.AddNewComponent(String clsid, Boolean throwOnError, Boolean select)
    NOTE:I replaced the workflow name and the GUID because I don't think it is important for the post.
    When I search for the GUID in regedit it leads me to an entry that has the folder DTSInfo and the value of
    Microsoft.DataTransformationServices.DataFlowUI.OleDbSourceUI, Microsoft.DataTransformationServices.DataFlowUI, Version=12.0.0.0, Culture=neutral, PublicKeyToken=********** 'This was not the original value, just hiding it for the post.
    for UITypeName
    No one here could figure out what the exact issue and how to fix it so my pc was rebuilt. Everything was working just fine until I installed SSMS 2014. Now I am right back where I started.
    I really need to know what is going on because I wasted an entire day on trying to figure out this issue and I have already tried reinstalling/repairing VS2010. Removing SSMS 2014. Registering OleDbSrc.dll but get an error when trying to do that.

    Somebody on my team had this issue. If I recall correctly you need to re-install SSDT 2012
    Arthur My Blog

  • Source and destination clarification needed

    Hi Guys,
    We are using SRM as an add-on component to ECC 6.0. We are on SRM Server 5.5. In the defination of backend system SAP asked us to create 4 entries.
    1. ONECLNTERP : RFC tick is on
    2. ONECLNTEBP:
    3. ONECLNTSUS:
    4. XXXCLNTYYY: Local tick is on.
    So this means 'ONECLNTERP' is my backend system i.e. ECC 6.0 and 'XXXCLNTYYY' is my local system.
    Now while defining backend system for Product category there is one 'Source System and there is 'Target system'. Source System: System from which master data is replicated. So this should be my SRM system (as Product category is a part of SRM) or this should be ECC (as My product categories are copy of material group from ECC).
    Target system is the system into which follow-on document of SC is transfered i.e. Backend ECC system.
    Am I correct to say the above. I'm confused about which system is source and which system is destination.
    While defining the Account Assignement category and defining the PR document type I'm facing the same problem as I've to define the source system.
    Can anybody clarify me??
    Thanks
    Debashish

    Hello Deb,
    I have not worked on ECC 6.0.
    But logically speaking the
    source system - has to be yr ECC as SRM would be taking the ref of matl masters from there only
    Target system - would be yr same ECC again if you are using only one backend (and I think you are)
    (if you are using multiple backend and you want followon document in diff backend then here the SID of that backend would come)
    BR
    Dinesh

  • One Request - Two Subject Area

    Hi,
    Is it possible to use two different SA for one request. I will be having a flag from table this flag will have my SA name it can be SA1 or SA2.
    Based on this flag i need to change SA in the request.
    thanks advance
    Vazir.

    Hi Chris,
    thanks for your reply,
    currently we are using the same method as you said.
    In our dashboard what I will see is any one request at a time.
    But the problem is both the requests are hitting source, then BI server decides which request to display or hide based on Guided navigation results.
    What I want is if my flag is SA1 then my query should fire only to SA1's source and viceversa
    So thats why raised like One request - Two Subject Area.

  • Copy Transferrules from one source system to another

    Hi everybody,
    we're uploading data from nearly 25 sourcesystems (including separate systems and systems with more than one client to extract from) via flat-file. The source-system exports data to a shared file directory and the bw-system reads it from there. Though this results in only <b>one</b> transferrule for all source-systems, it keeps generating problems every time upload is due.
    Somebody had the idea to switch to 'standard' R/3->BW method of data staging, using generic extractors and so on. But this will result in <b>25</b> transferrules which, I'm afraid, have to be set up and maintained seperately. As there's quite some coding in the rules (the transfertructure in fact has 241 fields), flaws are for sure!
    We know the trick of exporting a transferrule, changing the source-system-mapping and re-importing the same transport-request. That can be used to copy to one or two other systems, but not for 25.
    Does anybody know another way to copy a transferrule from one source system to multiple other source systems?
    Looking forward to your answers
    Thanks and Regards
    Robert

    Hi Ralph,
    I don't like that methode. Importing the same transport 15 times, each time changing the source-system mapping and making sure, that source-system gets the changes seems to be a huge source of errors, which will be hard to find.
    What do you think of the following idea to solve the issue a totally different way:
    All transfer-rules directly connected to the source-systems are stripped of all Coding, so there is a plain 1 by 1 transformation including 0logsys.
    The InfoSources attached all lead to one ODS-Object, which has the only purpose to collect the data from the source-system without any transformation. So some fields a filled by a system and other fields maybe not.
    The coding specific for source-systems is moved to transfer- / update-rules from that ODS-Object to the original ODS-Objects / Cubes and the "collecting" ODS-Object is emptied after every loading process.
    This will result in a single point of change. It will be quite a huge program (maybe in startroutine) but it's the one and only point, where changes and corrections have to be made.
    Regards
    Robert

Maybe you are looking for