ECC connecting to two different PI systems for proxy

Hi All,
We currently have a PI 7.1 system to which many client proxies are configured in the SAP ECC system.
Now that we have installed a new PI 7.3 system we would like to make all new integrations in PI 7.3. Note that the ECC system would remain the same.
(The idea is to migrate interfaces from PI 7.1 to PI 7.3 in a future date)
In the same ECC system we need to configure client proxies to the newly installed PI 7.3.
In SXMB_ADM we can only define one integration server for ECC system. Hence if we want to configure proxy to another PI system, we need to overwrite it.
Since our current ECC system is 6.0, the workaround given by SAP through transaction SAOMANAGER for ECC 7.0 systemsdoes not apply.
Kindly mention the alternative solutions/approaches available to configure ECC proxies to two separate PI systems.
Thanks,
John

Hi John
Have you checked this blog?
Connect SAP Application System to multiple SAP PI Integration Server - Variant n°1
Regards,
Giuseppe

Similar Messages

  • I am trying to connect a 24" cinema display with mini display port cable to a 24" imac with a displayport - is there a cable for this as these two connections are two different sizes

    I am trying to connect a 24" cinema display with mini display port cable to a 24" imac with a displayport - is there a cable for this as these two connections are two different sizes

    Hi Gilly,
    I have no proof, but I think it will not work with an adapter, only Mini Display Port or Thunderbolt.

  • Data copy from infocube to infocube in two different BW systems

    Dear All,
    If i have a scenario like i have an infocube for sales in two different bw systems, and if i want to send the data from one infocube to another infocube what would be the strategy we have to follow. Is it that i need to send it as a flat file or is there any other mechanism with which i can send the data into the other infocube.
    Yours answers would be rewarded.
    Regards
    Vijay

    hi Vijay,
    no, you have no need to send it as flat file if you have these 2 bw connected,
    you can use datamart scenario, where one bw act as source system for the other bw, we have such discussion several times, you can check the steps
    Loading data from cube1 to cube2
    Loading data from one cube to another cube.
    hope this helps.

  • How to use two different ojdbc14.jar for two web application.

    Hi,
    I have two web application running in same tomcat, I need to use the two different ojdbc14.jar for two application, now both are taking the jars from tomcat common/lib directory, I tried copying the new ojdbc14.jar in web-inf/lib folder of one application, but it is not working.
    Could you please let me know whether this will take the jar from tomcat by befault or from web-inf, and a solution how to proceed with this.
    Thanks in advance.

    Yes, I tried removing the jars from common/lib, but as the connection string is mentioned inside the server.xml it is showing db connection error while trying to connect to the database

  • JDBC connection to two different machines from the same program

    I want to use JDBC to connect to two different databases that are on two different MySQL installations on two different machines.
    One of the installations is MySQL 4.1.12 and the other is 4.1.14-nt
    For some reason, all my statements seem to be going only to one machine, even though I am very clearly set it up to go to two different machines.
    Does anyone know whether connecting to two different installations from one program presents any unusual problems?
    Thanks.

    Unless there is some horrible defect in the MySQL JDBC driver (which I doubt!), there should be no problem and like the other poster says, please post snippits of your code so we can see what your doing...

  • 'Howto' merge (consolidate) two different BW-Systems in one?

    Hi,
    actual we have two different BW System (BW-FI/CO and BW-CRM - both Version 3.5) which are installed on an own systems (historic grown).
    Now we want reduce our administration and hardware costs and our intention is to merge (consolidate) our two BW-Systems in one - using standard SAP transport method.
    Does anyone have any experience with merge of two and more BW Systems in one?
    Does anyone have a 'Howto Paper' about this topic?
    Thanks in advance,
    Ismet

    Hi Arun,
    thx a lot for your quick answer. My question was more 'merge/consolidate' of two BW-Systems as a 'System Copy'. I am trying to give you more information about our procedure:
    - we have two different BW-Systems - BW-System A and BW-System B
    - both BW-System have same Version and Support Package
    - we want to overtake ALL BW-Objects (InfoProvider, InfoObjects, Variables, etc......) from System B into System A
    - for this procedure we want use STMS (SAP Standard)
    - in second step we plan to transfer physical data from System B in to System A with Data Mart procedure (generate Export-Datsource in System B)
    Do you have (or anybody) any experience about this procedure?
    Thanks a lot and Best Regards.
    Ismet

  • Btrfs with different file systems for different mount points?

    Hey,
    I finally bought a SSD, and I want to format it to f2fs (and fat to boot on UEFI) and install Arch on it, and in my old HDD I intend to have /home and /var and  try btrfs on them, but I saw in Arch wiki that btrfs "Cannot use different file systems for different mount points.", it means I cannot have a / in f2fs and a /home in btrfs? What can I do? Better use XFS, ZFS or ext4 (I want the faster one)?
    Thanks in advance and sorry my english.

    pedrofleck wrote:Gosh, what was I thinking, thank you! (I still have a doubt: is btrfs the best option?)
    Just a few weeks ago many of us were worrying about massive data loss due to a bug introduced in kernel 3.17 that caused corruption when using snapshots. Because btrfs is under heavy developement, this sort of thing can be expected. That said, I have my entire system running with btrfs. I have 4 volumes, two raid1, a raid0 and a jbod. I also run rsync to an ext4 partition and ntfs. Furthermore I make offline backups as well.
    If you use btrfs make sure you have backups and make sure you are ready to use them. Also, make sure you check sum your backups. rsync has the option to use checksums in place of access times to determine what to sync.

  • Differentiating Master Data from two different source systems

    Friends,
    i have used standard InfoObjects that provide master data for two InfoCubes which take data from two different source systems. Now the some of the master data is identical in both the source systems (example: 10 stands for "Industrial" in one whereas 10 stands for "Agricultural" in the other). What do i do so that the system(BW) differentiates the two.
    Thanks in advance for all the help.
    Mike

    I tried to include the 0SOURSYSTEM in the compounding of this InfoObject (for master data) but it gives me a list of other objects that use this InfoObject as a reference characteristic and also says that this InfoObject is used in a ODS and data needs to be emptied in the ODS before activating  this InfoObject.  Please let me know if there is any way out of it.
    Thanks
    Mike

  • Is it possible to connect to two different clusters?

    Hi,
    I was wondering if it was possible to connect to two different clusters from the same process?
    What I would like to do is to able to get some cache on the first cluster and others on a second cluster. The cache-config DTD doesn't seem to forbid the definitions of several remote-cache-scheme. So I should be able to map all "c1-*" caches to the first cluster and all "c1-*" caches to a second cluster...
    Is it possible?

    Thanks,
    Yes I'm connection via extend.
    Could you give me an example of such configuration? I've made a test from a C++ client with the following client config:
    ====================================================================
    <?xml version="1.0"?>
    <!DOCTYPE cache-config SYSTEM "cache-config.dtd">
    <cache-config>
    <caching-scheme-mapping>
    <cache-mapping>
    <cache-name>c1-*</cache-name>
    <scheme-name>extend-1</scheme-name>
    </cache-mapping>
    <cache-mapping>
    <cache-name>c2-*</cache-name>
    <scheme-name>extend-2</scheme-name>
    </cache-mapping>
    </caching-scheme-mapping>
    <caching-schemes>
    <near-scheme>
    <scheme-name>near-c1</scheme-name>
    <back-scheme>
    <remote-cache-scheme>
    <scheme-ref>extend-1</scheme-ref>
    </remote-cache-scheme>
    </back-scheme>
    <invalidation-strategy>auto</invalidation-strategy>
    </near-scheme>
    <near-scheme>
    <scheme-name>near-c2</scheme-name>
    <back-scheme>
    <remote-cache-scheme>
    <scheme-ref>extend-2</scheme-ref>
    </remote-cache-scheme>
    </back-scheme>
    <invalidation-strategy>auto</invalidation-strategy>
    </near-scheme>
    <remote-cache-scheme>
    <scheme-name>extend-dist</scheme-name>
    <service-name>ExtendTcpCacheService</service-name>
    <initiator-config>
    <tcp-initiator>
    <remote-addresses>
    <socket-address>
    <address system-property="tangosol.coherence.proxy.address">ldnpsm020001413</address>
    <port system-property="tangosol.coherence.proxy.port">4321</port>
    </socket-address>
    </remote-addresses>
    <connect-timeout>10s</connect-timeout>
    </tcp-initiator>
    <outgoing-message-handler>
    <request-timeout>5s</request-timeout>
    </outgoing-message-handler>
    <serializer>
    <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
    </serializer>
    </initiator-config>
    </remote-cache-scheme>
    <remote-cache-scheme>
    <scheme-name>extend-paul</scheme-name>
    <service-name>ExtendTcpCacheService</service-name>
    <initiator-config>
    <tcp-initiator>
    <remote-addresses>
    <socket-address>
    <address system-property="tangosol.coherence.proxy.address">ldnpsm020001412</address>
    <port system-property="tangosol.coherence.proxy.port">4131</port>
    </socket-address>
    </remote-addresses>
    <connect-timeout>10s</connect-timeout>
    </tcp-initiator>
    <outgoing-message-handler>
    <request-timeout>5s</request-timeout>
    </outgoing-message-handler>
    <serializer>
    <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
    </serializer>
    </initiator-config>
    </remote-cache-scheme>
    </caching-schemes>
    </cache-config>
    ========================================================================
    if I do a CacheFactory::getCache("c1-Test"), the log shows a connection to the 1st cluster
    if I then do a CacheFactory::getCache("c2-Test"), the log shows a connection to the 1st cluster and not to the second cluster
    What did I do wrong?

  • 11g Multiple Rpd on one BI server connecting to two different data marts

    Hi
    Did anyone tried and successfully implemented multiple rpds on one bi server. I have similar requirement in one of the clients place.
    1. Install OBIEE 11g with vertical scaling( reason client doesnt want to spend extra money on additional server licenses) and would like to utilize the powerful existing servers.
    2. Install 2 rpds on One BI Server and connect to two different datamarts.
    3. Create two PS and provide two links to have two different reporting environments.
    Appreciate with any pointers. I am very clear on Vertical Scaling and bumping up the Bi , PS and Java Hosts components. I saw a post with successful implementation of the same functionility in 10g and nothing in 11g. I already reviewed ritmanmeed blogs but did not give a complete picture of successfull implementation.
    Thanks

    Hi,
    Yes, it is possible to have two different .rpds on one BI Server but then you need to set up two different Presentation Servers.
    Refer the below threads for more information.
    Re: multiples RPD in one server
    Re: Multiple RPD & Multiple Presentation service instance on single BI Server
    https://forums.oracle.com/forums/thread.jspa?messageID=9189677&#9189677
    Can i put two rpd files in nqsconfig init file in BI Admin
    Hope this help's
    Thanks
    satya

  • How to create two different implementation class for a Control

    Hi,
    I am a newbie to beehieve. I want to know that is it possible to create two different 'Impl' classes for a Control. If yes then how do we instantiate them ? How can we chain them (something like calling one 'Impl' from the other one? Thanks in Advance!!!
    Regards,
    Abhishek

    You are sure you are in the right forum?
    This is the JDeveloper and ADF forum...
    Timo

  • Internal order - Error - Different logical system for order

    Hello all,
    I try to send IDoc (INTERNAL_ORDER) from system A to system B.
    I have done:
    - BD64 / WE20 / WE21
    But after having customizing I have got following error:
    Different logical system for order xxxxxx
    Message no. 2075ALE001
    Diagnosis
    Internal order xxxxxxx already exists in system B. Its master system B is different from the master system A of the new internal order to be distributed with the same order number xxxxxxx.
    System Response
    Internal order xxxxxxx with master system A cannot be distributed in system B.
    Procedure
    Check customizing for ALE for internal orders.
    How to solve this?
    My aim is to maintain internal order master data in test system.
    Thanks for your help.
    David

    I have found something.
    If I change in table AUFK, logical from empty to A (master) and I run IDOC then IDOC has been correctly integrated.
    Then: how to define "master"?

  • TS1398 I am unable to connect to two different wifi at different locations having same name but different passwords

    I am unable to connect to two different wi if at different locations having same user name but different password one is at home other at work

    I was able to do a full reset on the router (pushing the tiny button near the power input). Then ran through the setup process and it started working on my device.

  • Why two different explain plan for same objects?

    Believe or not there are two different databases, one for processing and one for reporting, plan is show different for same query. Table structure and indexes are same. It's 11G
    Thanks
    Good explain plan .. works fine..
    Plan
    SELECT STATEMENT  ALL_ROWSCost: 12,775  Bytes: 184  Cardinality: 1                                                                        
         27 SORT UNIQUE  Cost: 12,775  Bytes: 184  Cardinality: 1                                                                   
              26 NESTED LOOPS                                                              
                   24 NESTED LOOPS  Cost: 12,774  Bytes: 184  Cardinality: 1                                                         
                        22 HASH JOIN  Cost: 12,772  Bytes: 178  Cardinality: 1                                                    
                             20 NESTED LOOPS SEMI  Cost: 30  Bytes: 166  Cardinality: 1                                               
                                  17 NESTED LOOPS  Cost: 19  Bytes: 140  Cardinality: 1                                          
                                       14 NESTED LOOPS OUTER  Cost: 16  Bytes: 84  Cardinality: 1                                     
                                            11 VIEW DSSADM. Cost: 14  Bytes: 37  Cardinality: 1                                
                                                 10 NESTED LOOPS                           
                                                      8 NESTED LOOPS  Cost: 14  Bytes: 103  Cardinality: 1                      
                                                           6 NESTED LOOPS  Cost: 13  Bytes: 87  Cardinality: 1                 
                                                                3 INLIST ITERATOR            
                                                                     2 TABLE ACCESS BY INDEX ROWID TABLE DSSODS.DRV_PS_JOB_FAMILY_TBL Cost: 10  Bytes: 51  Cardinality: 1       
                                                                          1 INDEX RANGE SCAN INDEX DSSODS.DRV_PS_JOB_FAMILY_TBL_CL_SETID Cost: 9  Cardinality: 1 
                                                                5 TABLE ACCESS BY INDEX ROWID TABLE DSSADM.DIM_JOBCODE Cost: 3  Bytes: 36  Cardinality: 1            
                                                                     4 INDEX RANGE SCAN INDEX DSSADM.STAN_JB_FN_IDX Cost: 2  Cardinality: 1       
                                                           7 INDEX UNIQUE SCAN INDEX (UNIQUE) DSSODS.DRV_PS_JOBCODE_TBL_SEQ_KEY_RPT Cost: 0  Cardinality: 1                 
                                                      9 TABLE ACCESS BY INDEX ROWID TABLE DSSODS.DRV_PS_JOBCODE_TBL_RPT Cost: 1  Bytes: 16  Cardinality: 1                      
                                            13 TABLE ACCESS BY INDEX ROWID TABLE DSSODS.DRV_PSXLATITEM_RPT Cost: 2  Bytes: 47  Cardinality: 1                                
                                                 12 INDEX RANGE SCAN INDEX DSSODS.PK_DRV_RIXLATITEM_RPT Cost: 1  Cardinality: 1                           
                                       16 TABLE ACCESS BY INDEX ROWID TABLE DSSADM.DIM_JOBCODE Cost: 3  Bytes: 56  Cardinality: 1                                     
                                            15 INDEX RANGE SCAN INDEX DSSADM.DIM_JOBCODE_EXPDT1 Cost: 2  Cardinality: 1                                
                                  19 TABLE ACCESS BY INDEX ROWID TABLE DSSODS.DRV_PS_JOB_RPT Cost: 11  Bytes: 438,906  Cardinality: 16,881                                          
                                       18 INDEX RANGE SCAN INDEX DSSODS.DRV_PS_JOB_JOBCODE_RPT Cost: 2  Cardinality: 8                                     
                             21 INDEX FAST FULL SCAN INDEX (UNIQUE) DSSADM.Z_PK_JOBCODE_PROMPT_TBL Cost: 12,699  Bytes: 66,790,236  Cardinality: 5,565,853                                               
                        23 INDEX RANGE SCAN INDEX DSSADM.DIM_PERSON_EMPL_RCD_SEQ_KEY Cost: 1  Cardinality: 1                                                    
                   25 TABLE ACCESS BY INDEX ROWID TABLE DSSADM.DIM_PERSON_EMPL_RCD Cost: 2  Bytes: 6  Cardinality: 1                                                         This bad plan ... show merge join cartesian and full table ..
    Plan
    SELECT STATEMENT  ALL_ROWSCost: 3,585  Bytes: 237  Cardinality: 1                                                              
         26 SORT UNIQUE  Cost: 3,585  Bytes: 237  Cardinality: 1                                                         
              25 NESTED LOOPS SEMI  Cost: 3,584  Bytes: 237  Cardinality: 1                                                    
                   22 NESTED LOOPS  Cost: 3,573  Bytes: 211  Cardinality: 1                                               
                        20 MERGE JOIN CARTESIAN  Cost: 2,864  Bytes: 70,446  Cardinality: 354                                          
                             17 NESTED LOOPS                                     
                                  15 NESTED LOOPS  Cost: 51  Bytes: 191  Cardinality: 1                                
                                       13 NESTED LOOPS OUTER  Cost: 50  Bytes: 180  Cardinality: 1                           
                                            10 HASH JOIN  Cost: 48  Bytes: 133  Cardinality: 1                      
                                                 6 NESTED LOOPS                 
                                                      4 NESTED LOOPS  Cost: 38  Bytes: 656  Cardinality: 8            
                                                           2 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DIM_JOBCODE Cost: 14  Bytes: 448  Cardinality: 8       
                                                                1 INDEX RANGE SCAN INDEX REPORT2.STAN_PROM_JB_IDX Cost: 6  Cardinality: 95 
                                                           3 INDEX RANGE SCAN INDEX REPORT2.SETID_JC_IDX Cost: 2  Cardinality: 1       
                                                      5 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DIM_JOBCODE Cost: 3  Bytes: 26  Cardinality: 1            
                                                 9 INLIST ITERATOR                 
                                                      8 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DRV_PS_JOB_FAMILY_TBL Cost: 10  Bytes: 51  Cardinality: 1            
                                                           7 INDEX RANGE SCAN INDEX REPORT2.DRV_PS_JOB_FAMILY_TBL_CL_SETID Cost: 9  Cardinality: 1       
                                            12 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DRV_PSXLATITEM_RPT Cost: 2  Bytes: 47  Cardinality: 1                      
                                                 11 INDEX RANGE SCAN INDEX REPORT2.PK_DRV_RIXLATITEM_RPT Cost: 1  Cardinality: 1                 
                                       14 INDEX UNIQUE SCAN INDEX (UNIQUE) REPORT2.DRV_PS_JOBCODE_TBL_SEQ_KEY_RPT Cost: 0  Cardinality: 1                           
                                  16 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DRV_PS_JOBCODE_TBL_RPT Cost: 1  Bytes: 11  Cardinality: 1                                
                             19 BUFFER SORT  Cost: 2,863  Bytes: 4,295,552  Cardinality: 536,944                                     
                                  18 TABLE ACCESS FULL TABLE REPORT2.DIM_PERSON_EMPL_RCD Cost: 2,813  Bytes: 4,295,552  Cardinality: 536,944                                
                        21 INDEX RANGE SCAN INDEX (UNIQUE) REPORT2.Z_PK_JOBCODE_PROMPT_TBL Cost: 2  Bytes: 12  Cardinality: 1                                          
                   24 TABLE ACCESS BY INDEX ROWID TABLE REPORT2.DRV_PS_JOB_RPT Cost: 11  Bytes: 1,349,920  Cardinality: 51,920                                               
                        23 INDEX RANGE SCAN INDEX REPORT2.DRV_PS_JOB_JOBCODE_RPT Cost: 2  Cardinality: 8                                          

    user550024 wrote:
    I am really surprise that the stat for good sql are little old. I just computed the states of bad sql so they are uptodate..
    There is something terribly wrong..Not necessarily. Just using the default stats collection I've seen a few cases of things suddenly going wrong. As the data increases, it gets closer to an edge case where the inadequacy of the statistics convinces the optimizer to do a wrong plan. To fix, I could just go into dbconsole, set the stats back to a time when they worked, and locked them. In most cases it's definitely better to figure out what is really going on, though, to give the optimizer better information to work with. Aside from the value of learning how to do it, for some cases it's not so simple. Also, many think the default settings of the database statistic collection may be wrong in general (in 10.2.x, at least). So much depends on your application and data that you can't make too many generalizations. You have to look at the evidence and figure it out. There is still a steep learning curve for the tools to look at the evidence. People are here to help with that.
    Most of the time it works better than a dumb rule based optimizer, but at the cost of a few situations where people are smarter than computers. It's taken a lot of years to get to this point.

  • Can we have two different logon screen for the Portal

    Can we have two different logon screens for the Portal?
    We should split the Logon application in two different applications:
    Login # 1
    should be read a DataSource.XML reading the Active Directory (Microsoft), to read-only the profiles from the AD, and block some IP address range.
    Login # 2
    should be read a DataSource.XML reading the Portal user only
    Is this possible?

    Hello Developer,
    Did you find an answer to this problem. I also have a similar issue at hand and am wondering if anybody has a solution for this.
    Thanks, Akhilesh

Maybe you are looking for