Multiple clustering integrating with AD

I have a customer that I am doing a design for that wants to integrate with Ad for CCM. No problem. BUT there are going to be maybe 15 sites or so that will each have their own cluster of CCMs. The AD structure that they will be integrating to is and enterprise AD. So is it possible to integrate multiple CCM clusters with the same AD, and let each cluster see the whole domain as its directory? Any help would be apprepciated.
Thanks,
Chris

Please be advised on the user accounts that are commonly used by applications, like the "ac" user for Attendant Console, "jtapi", "rmjtapi" user for IPCC Express etc. Since it is a single directory, you cannot have the same user id and hence you would need to create different user id's for these applications in each cluster to avoid any chances of data getting overwritten. For example, create user id "jtapi_cluster1" for cluster1, "jtapi_cluster2" for cluster2 etc.
Regards,
Anup

Similar Messages

  • Multiple clusters definition with the same httpd.conf

    I want to define two different weblogic clusters for the same virtual host based on the URL. I've tried this
    <Location /tocluster1 >
    WebLogicCluster machine1:8090,machine2:8090
    SetHandler weblogic-handler
    </Location >
    <Location /tocluster2 >
    WebLogicCluster machine3:9090, machine4:9090
    SetHandler weblogic-handler
    </Location >
    It doesn't work : requests are mixed up between both clusters.
    Is it possible to achieve using WL6.1SP5 plugin and apache 2.0.47 ?

    I had a similar issue and got it fixed after applying the latest version of the
    proxy plugin. If you already got the latest plugin then it is fine. Otherwise
    let me know, I can send it over to you. You may also open a case with BEA and
    get it.
    Thanks,
    Jiten
    "lauren" <[email protected]> wrote:
    >
    I am having the same issue with Weblogic 8.1 plugin on Apache 2.0.46.
    So it seems
    like the issue has been fixed yet. Do I have to use only one WeblogicCluster
    for each Apache Virtual host? Does anyone have a better solution? Thanks
    Lauren
    "Bjørn Bjerkeli" <[email protected]> wrote:
    WebLogic 5.1 supports only one config per VirtualHost, as specifiedin
    docs, and
    from 6.1 and up multiple configs are supported, also specified in docs.
    I would strongly recommend to use the 6.1-approach when proxying tomultiple
    clusters,
    even if the "old" style works too. I have done some testing that yielded
    high
    error ratios, apache bridge failures and READ_TIMEOUTS when using mod_proxy
    as
    a intermediate to proxy from one VirtualHost to antother containingthe
    cluster-definition
    /Bjørn Bjerkeli, Zenior AS
    "Yakov Broido" <[email protected]> wrote:
    Hey,
    I had the same problem and it didnt work.
    eventually I had to use multiple VirtualHosts.
    It seems that its a bug and the WebLogicCluster directive(and some
    others)
    are global for each vhost.
    This bug is going to be fixed in next ServicePack of WLS 7.0 but Idonno
    what about 6.1
    "Philippe Peeters" <[email protected]> wrote in message
    news:40165b9b$[email protected]..
    I want to define two different weblogic clusters for the same virtualhost
    based on the URL. I've tried this
    <Location /tocluster1 >
    WebLogicCluster machine1:8090,machine2:8090
    SetHandler weblogic-handler
    </Location >
    <Location /tocluster2 >
    WebLogicCluster machine3:9090, machine4:9090
    SetHandler weblogic-handler
    </Location >
    It doesn't work : requests are mixed up between both clusters.
    Is it possible to achieve using WL6.1SP5 plugin and apache 2.0.47

  • Load-balancing issues with iPlanet and multiple clusters

    We're in performance test of a large-scale clustered deployment based on WLS 5.1sp10.
    Due to scalability/functionality issues, some of which we've seen firsthand and
    some of which we've been informed of by associates as well as BEA representatives,
    we've chosen to implement multiple clusters with a maximum of three nodes each.
    These clusters will be fronted by a web server tier consisting of iPlanet servers
    using the proxy plugin.
    Due to hardware constraints (both in test and in production), however, we've configured
    the iPlanet servers to route across the multiple clusters. In our test environment,
    for instance, we've got a single iPlanet server routing across two 3-node clusters,
    and the configuration in obj.conf is as follows:
    <Object name="application" ppath="*/application">
    Service fn="wl-proxy" \
    WebLogicCluster="clusterA_1:9990,clusterB_1:9991,clusterA_2:9990,clusterB_2:9991,clusterA_3:9990,
    clusterB_3:9991" \
    CookieName="ApplicationSession"
    </Object>
    Our issue is that the load-balancing doesn't appear to work across the clusters.
    We're seeing one cluster get about 90% of the load, while the other receives
    only 10%.
    So, the question (finally!) is: Is this configuration correct (i.e., will it
    work according to the logic of the proxy plugin), and is it appropriate for this
    situation? Are there other alternative approaches that anyone can recommend?
    Thanks in advance,
    cramer

    I use weblogic6.1 with sp2+windows 2000.I develop a web application and deploy
    it to cluster.Through HttpClusterServlets proxy of weblogic I found that a server
    in cluster almost get 95% of requests but another only get 5% of requests.Why???
    I don't set any special parameter.And the weight of the two clustered server is
    equal.I use round-robin arithmetic.
    Thanks!
    "cramer" <[email protected]> wrote:
    >
    We're in performance test of a large-scale clustered deployment based
    on WLS 5.1sp10.
    Due to scalability/functionality issues, some of which we've seen firsthand
    and
    some of which we've been informed of by associates as well as BEA representatives,
    we've chosen to implement multiple clusters with a maximum of three nodes
    each.
    These clusters will be fronted by a web server tier consisting of iPlanet
    servers
    using the proxy plugin.
    Due to hardware constraints (both in test and in production), however,
    we've configured
    the iPlanet servers to route across the multiple clusters. In our test
    environment,
    for instance, we've got a single iPlanet server routing across two 3-node
    clusters,
    and the configuration in obj.conf is as follows:
    <Object name="application" ppath="*/application">
    Service fn="wl-proxy" \
    WebLogicCluster="clusterA_1:9990,clusterB_1:9991,clusterA_2:9990,clusterB_2:9991,clusterA_3:9990,
    clusterB_3:9991" \
    CookieName="ApplicationSession"
    </Object>
    Our issue is that the load-balancing doesn't appear to work across the
    clusters.
    We're seeing one cluster get about 90% of the load, while the other
    receives
    only 10%.
    So, the question (finally!) is: Is this configuration correct (i.e.,
    will it
    work according to the logic of the proxy plugin), and is it appropriate
    for this
    situation? Are there other alternative approaches that anyone can recommend?
    Thanks in advance,
    cramer

  • Multiple Clusters with same computers

    Is there a way to set up multiple clusters with the same computers.
    For example
    Cluster A has computers 1, 2
    Cluster B has computers 1, 2, 3, 4, 5
    The reason behind this is that computers 1 and 2 are always available, but 3, 4, 5 are used during the day. I'd like to have a day cluster and a night cluster that I can just select as needed.

    Hi Jake, I was playing with this type of thing when FCS1 first came out using some old powerbooks.
    Yes you can kind of do this using UNMANAGED SERVICES but I will stand corrected for COMPRESSOR3.app.
    In FCS2, compressor 3, simply if you use *MANAGED SERVICES* of the service nodes available through a defined a specific cluster (you would have used appleqmaster utility.app) than I believe those service nodes are dedicated to that cluster.
    Sure you can make 'INSTANCES" of compressor for example and the 'renderer" instances dedicated to particular cluster using this as an example... let say...
    HOST1 = MACPRO8CORE with 4 instances of compressor (as C1,C2,C3,C4 via vi 4 virtual clusters) and 8 instances of renderer (R1,R2,..R7 & R8).
    HOST2 = MACBOOKPRO with 1 instance of compressor (as C1 & C2 with 2 virtual clusters) and 2 instances of renderer (R1,R2).
    HOST3 = PowerbookG4 with 1 instance of compressor (as C1) and 1 instances of renderer (R1).
    only as an example
    ClusterA: Host1[C1,C2C3, R1,R2,R4,R5), Host2[C2]
    ClusterB: Host1[C4, R6,R3,R6,R7), Host2[C2,R1R2], HOST3{C1,R1]
    In fact I just tried it...
    However as I thought for managed services be assured that you cannot SHARE a service node with two or more clusters. I will stand corrected.
    Messy but seems to work but useless with the powerbook you'd agree
    However depending on your workflow and commecrial needs (say for busness prioriries for a specific client) for best results use *UNMANAGED SERVICES* ....
    Simply hook everything up over GBE, with all the usual tweaks such as "NEVER COPY SOURCE" and MOUNTING all source targets and compressor work files over HFS on the GBE subnet (and many other tweaks) and treat the set up as a huge bucket.
    Use priority on the batch submission. It's not too smart but at least you have some manipulation over the queues.
    The resource manager in QMASTER is not so smart. So despite a service node going idle it does not seem to redirect work from wone node to another for load balancing..
    I have only tried this with SEGMENTED TRANSCODING.
    Rendering (Shake) works great and seem simpler. My time is with multipass segented transcoding where I want H.264 from DVCPROHD and dont want to wait all day for it, especially where I have tweaked the timing in compressor a bit.
    Try it out.
    BTW as many contest on this forum, QMASTER/COMPRESSOR can be a bugger to fix if it plays up. and it has for me as well.
    post your results.
    HTH.
    w

  • One SRM integrated with multiple Backends

    Hi All
    Can One SRM be integrated with multiple Backends?
    If so any insights, documentation, pros-cons that can you all can share?
    Does the version of SRM or ECC matter?
    SRM version will be 7.0 and ECC 6 EP3.
    Thanks,
    Gamad

    Hi Gamad,
    multiple backend systems for one SRM is normal and works fine.
    You just need to connect the systems, replicate the materials/vendors you would like to use and maintain the SRM customizing:
    1) Add the further backend system(s):
    -> Supplier Relationship Management -> SRM Server -> Technical Basic Settings -> Define Backend Systems
    2) Define the scenario/follow-on document creation for the product categories
    -> Supplier Relationship Management -> SRM Server -> Technical Basic Settings  -> Define Backend System for Product Category
    3) Define the scenario/follow-on document creation for the product categories
    -> Supplier Relationship Management -> SRM Server -> Cross-Application Basic Settings -> Define Objects in Backend System (Purch. Reqs, Reservations, Purch. Orders)
    Regards,
    Peter

  • CUC 8.5 UM integration with multiple AD Forest

    Hi Guys,
    I need to do a unity connection 8.5 UM integration with exchange 2003 and 2010 in two different AD Forest. AD 2003/exchange 2003 is in one Forest (Account Forest) with 2003 mailboxes on a corporate network AND AD 2008/exchange 2010 is in second Forest (Resource Forest) with 2010 mailboxes in the CLOUD. There is also a trust relationship between 2 AD Forest. As per unity connection 8.x SRND, we must create separate unified messaging services account (AD account) for each Forest.
    Given the above scenario, I think we need 2 AD accounts, one on AD 2003 (for exchange 2003 mailboxes) and second one on AD 2008 (for exchange 2010 mailboxes). However, in this configuration if mailboxes are moved from 2003 to 2010 we will have to manually disassociate the unity connection subscriber mailbox with unified messaging service account (AD 2003 account) and re-associate it with unified messaging service account (AD 2008 account), is that correct? or is there a way to automate this if the mailboxes are moved to 2010 unity connection automatically detects the change?
    Any pointers in the right direction would be much appreciated.
    Thanks

    Can any one comment on this please? Perhaps TAC engineers might be able to provide some insight?
    Thanks

  • Session keep Alive when EBS is integrated with OIF acting as SP

    We have an E-Business Suite R12(12.1.3) Application which is integrated with Oracle Access Manager 11.1.1.5. For authentication the Oracle Access Manager delegates the authentication request to Oracle Identity Federation acting as a Service Provider (SP). There is a home-grown SSO which acts as the Identity provider by Federating it with OIF using SAML 2.0 protocol.
    The integration is successful and EBS is able to authenticate using this third party SSO.
    Now there are multiple applications which are integrated with this third party SSO and the user can access any of these applications including EBS.
    Considering a scenario where a user is using EBS Application say for more than 30 mins. Now when he tries to login to another application the challenge screen is thrown though he can continue to work on EBS without any issue. The reason being that the session timeout on the IDP side is set to 30 mins.
    To overcome this issue the solution that has been thought of using a javascript for each page load on the application which will update the SSO token with the latest UTC time on the IDP server. This has been successfully done for all other applications except EBS.
    So my question is how we can call this java script for each page load for EBS?
    Thanks & Regards
    Sarbashis

    It appears that OAM is removing the url parameters for the p_done_url sent to it by OIF.
         302 Redirect to -> Location: https://sso.mycompany.com/oam/server/logout?p_done_url=https%3A%2F%2Fsso.mycompany.com%2Ffed%2Fuser%2Fsloosso%3Fid%3Dosso%26type%3D3&invokeOSFSLogout=false
         Result: 200. Set-Cookie OAM-ID to expired.
    Form Submits: https://sso.mycompany.com/fed/user/sloosso
    Result: 500 Error
    Notice the p_done_url value in the redirect has extra parameters which when the Form is submitted are lost.
    If I manually enter the url: https://sso.mycompany.com/fed/user/sloosso?id=osso&type=3&invokeOSFSLogout=false
    Then I am successfully logged out.
    This seems almost like a bug in OAM...

  • Multiple clusters on a single machine

    Hi everyone. I have a question about running 2 clusters on a same
              machine.
              Here's the scenario. We have 8 WebLogic server (4.5.1 SP4) instances
              running on
              2 servers (Sun Enterprise 5500/Solaris 2.62) right now for one of our
              eCommerce
              application.
              Now we are getting ready to deploy a 2nd application and we want it to
              have its
              own cluster. So we are hoping to run 4 more WebLogic servers on each of
              the 2
              machines. So machine #1 would have 4 WebLogic servers for cluster A and
              4
              WebLogic servers for cluster B. And machine #2 will have 4 WebLogic
              servers for
              cluster A and 4 WebLogic servers for cluster B.
              All of these servers have unique IP address and each cluster has a
              different
              name and multicast IP address. Both clusters are using in-memory
              replication.
              On both machines, each of the cluster member will share the same global
              weblogic.properties file. The cluster specific portion resides in the
              cluster
              directory.
              Is anyone currently running such a configuration?? Is there something
              we should
              worry about?? Since both of these servers are in their own isolate
              network
              segment, is there any danger of multicast broadcast bleeding into the
              different
              cluster?? Thanks in advance.
              

    Hi Michael and thanks.. No, this is 2 systems on the same network and not two
              independent networks. Here's what I found that's very strange. I started
              server 1 ->4 for the first cluster and everything worked fine. Now I started
              server 5 on the first machine that's in a different cluster and has a totally
              different multicast IP address and yet, server 1 of the first cluster recorded
              server 5 joining the cluster. Server 2, 3 and 4 of the first cluster did not
              have RJVM 'Creating connection' message in there.
              On server 5's startup, server 1 decided to make server 5 it's secondary pair and
              I get the following error as server 5 does not have the same code for cluster1.
              After that error, I started getting tons of
              weblogic.cluster.replication.BadStatusException error in the logs for server 1
              listed below. I'm sure running multiple clusters off an enterprise level
              server is something people want to do. Is this a known issue?? Is this a bug
              in the server or the VM?
              Our environment: Sun 5500 Enterprise servers running Solaris 2.62, WebLogic
              4.5.1 SP4.
              Sun Jan 09 15:19:40 CST 2000:<I> <RJVM> Creating connection to
              stingray05.xx.com/10.x.x.x t3:7001 http:7001 t3s:7002 https:7002
              -1353087187406741800
              Sun Jan 09 15:19:41 CST 2000:<E> <MulticastSocket> Multicast socket receive
              error:
              Sun Jan 09 15:19:41 CST 2000:<E> <MulticastSocket>
              java.lang.ClassNotFoundException: Could not instantiate new instance of
              [com.strong.brokerage.server.CustomerSessionBeanHomeImpl_WLStub]
              Either the class is missing from classpath, the class does not have a public
              constructor, or the constructor raised an error; the actual exception was a :
              java.lang.NoClassDefFoundError: com.strong.brokerage.common.util.RegistrationCtx
              at java.lang.Throwable.<init>(Compiled Code)
              at java.lang.Exception.<init>(Compiled Code)
              at java.lang.ClassNotFoundException.<init>(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readLeftover(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectBody(Compiled Code)
              at weblogic.common.internal.WLObjectInputStreamBase.readObject(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectWL(Compiled Code)
              at weblogic.service.ReplicaAggregate.readExternal(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readPublicSerializable(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readLeftover(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectBody(Compiled Code)
              at weblogic.common.internal.WLObjectInputStreamBase.readObject(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectWL(Compiled Code)
              at weblogic.cluster.BasicServiceOffer.readObject(Compiled Code)
              at weblogic.cluster.AnnotatedServiceOffer.readObject(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readPublicSerializable(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readLeftover(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectBody(Compiled Code)
              at weblogic.common.internal.WLObjectInputStreamBase.readObject(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectWL(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readArrayList(Compiled Code)
              at weblogic.cluster.Announcement.readObject(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readPublicSerializable(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readLeftover(Compiled Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectBody(Compiled Code)
              at weblogic.common.internal.WLObjectInputStreamBase.readObject(Compiled
              Code)
              at
              weblogic.common.internal.WLObjectInputStreamBase.readObjectWL(Compiled Code)
              at weblogic.cluster.TMSocket.execute(Compiled Code)
              at weblogic.t3.srvr.ExecuteThread.run(Compiled Code)
              Sun Jan 09 15:24:42 CST 2000:<E> <HTTP> Servlet request terminiated with Error:
              Sun Jan 09 15:24:42 CST 2000:<E> <HTTP>
              weblogic.cluster.replication.BadStatusException: update found
              -5050872662681587012 but it is not the secondary
              at weblogic.rmi.extensions.BasicRequest.sendReceive(Compiled Code)
              at
              weblogic.cluster.replication.ReplicationManager_WLStub.update(Compiled Code)
              at
              weblogic.cluster.replication.ReplicationManager.updateSecondary(Compiled Code)
              at weblogic.servlet.internal.session.ReplicatedSession.sync(Compiled
              Code)
              at
              weblogic.servlet.internal.session.ReplicatedSessionContext.sync(Compiled Code)
              at weblogic.servlet.internal.ServletRequestImpl.syncSession(Compiled
              Code)
              at weblogic.servlet.internal.ServletStubImpl.invokeServlet(Compiled
              Code)
              at weblogic.servlet.internal.ServletContextImpl.invokeServlet(Compiled
              Code)
              at weblogic.servlet.internal.ServletContextImpl.invokeServlet(Compiled
              Code)
              at weblogic.socket.MuxableSocketHTTP.invokeServlet(Compiled Code)
              at weblogic.socket.MuxableSocketHTTP.execute(Compiled Code)
              at weblogic.t3.srvr.ExecuteThread.run(Compiled Code)
              --------------- nested within: ------------------
              weblogic.utils.NestedError: Tried to update secondary, but it thought it was the
              primary
              - with nested exception:
              [weblogic.cluster.replication.BadStatusException: update found
              -5050872662681587012 but it is not the secondary]
              at java.lang.Throwable.<init>(Compiled Code)
              at java.lang.Error.<init>(Compiled Code)
              at weblogic.utils.NestedError.<init>(Compiled Code)
              at weblogic.servlet.internal.session.ReplicatedSession.sync(Compiled
              Code)
              at
              weblogic.servlet.internal.session.ReplicatedSessionContext.sync(Compiled Code)
              at weblogic.servlet.internal.ServletRequestImpl.syncSession(Compiled
              Code)
              at weblogic.servlet.internal.ServletStubImpl.invokeServlet(Compiled
              Code)
              at weblogic.servlet.internal.ServletContextImpl.invokeServlet(Compiled
              Code)
              at weblogic.servlet.internal.ServletContextImpl.invokeServlet(Compiled
              Code)
              at weblogic.socket.MuxableSocketHTTP.invokeServlet(Compiled Code)
              at weblogic.socket.MuxableSocketHTTP.execute(Compiled Code)
              at weblogic.t3.srvr.ExecuteThread.run(Compiled Code)
              Michael Girdley wrote:
              > Vinny,
              >
              > It sounds like the real hard problem here is a system administration and
              > configuration issue. If you have two independent networks with independent
              > systems, then you should not have a problem running two clusters across two
              > machines. You just need to make sure that your routing for multicast is
              > appropriately configured so that they don't confuse the two different
              > WebLogic server clusters.
              >
              > I hope this helps. I might also recommend contacting our sales organization
              > who can put you in touch with a systems engineer who is a professional at
              > making these types are recommendations.
              >
              > Thanks!
              >
              > --
              > Michael Girdley
              > BEA Product Manager
              >
              > Vinny Carpenter <[email protected]> wrote in message
              > news:[email protected]...
              > > Hi everyone. I have a question about running 2 clusters on a same
              > > machine.
              > > Here's the scenario. We have 8 WebLogic server (4.5.1 SP4) instances
              > > running on
              > > 2 servers (Sun Enterprise 5500/Solaris 2.62) right now for one of our
              > > eCommerce
              > > application.
              > >
              > > Now we are getting ready to deploy a 2nd application and we want it to
              > > have its
              > > own cluster. So we are hoping to run 4 more WebLogic servers on each of
              > > the 2
              > > machines. So machine #1 would have 4 WebLogic servers for cluster A and
              > > 4
              > > WebLogic servers for cluster B. And machine #2 will have 4 WebLogic
              > > servers for
              > > cluster A and 4 WebLogic servers for cluster B.
              > >
              > > All of these servers have unique IP address and each cluster has a
              > > different
              > > name and multicast IP address. Both clusters are using in-memory
              > > replication.
              > > On both machines, each of the cluster member will share the same global
              > > weblogic.properties file. The cluster specific portion resides in the
              > > cluster
              > > directory.
              > >
              > > Is anyone currently running such a configuration?? Is there something
              > > we should
              > > worry about?? Since both of these servers are in their own isolate
              > > network
              > > segment, is there any danger of multicast broadcast bleeding into the
              > > different
              > > cluster?? Thanks in advance.
              > >
              > >
              > >
              > >
              

  • Cross co. code controlling area - CO Real time integration with FI

    Hi all of you,
    We are in to New GL, and implementing one controlling area for all the company codes (cross company code cost accounting) with group currency  "30". And my client requires the CO - FI Real time integration to take the segment reporting / profit center (new) FI reporting.
    My basic doubt is, when we are implementing one CO area, we taking all the company codes as cost center groupings and creating the overhead cost centers. When doing Repost CO line items in CO, how system arrives the company code since we are not referring the FI document or Company code, how system will create the FI document in the company code.
    Do we have any separate transaction to do the repost CO line items to hit the CO - FI real time integration?
    Please give me your valuable suggestions to use the CO-FI real time integration with cross controlling area assigned to multiple company codes.
    Thanks & Regards
    Ramki

    Thanks for your valuable reply Mr. Christian.
    So you mean to say that, when we are maintaining the cost center groupings (per company code), same will be maintained for profit center and will be assigned to the individual company codes, right.
    In CO - FI real time integration , I need to select the objects - company code and segment. Whenever any repost happened in the CO module, system will create one FI document referring to the real time objects. We have more than 12 company codes located in different countries. In this way, system is fulfilling the cross company code postings in CO and will generate the FI documents, right.
    Please give me your valuable suggestion and when I am looking the note in Market place (91055), it is showing the CAD Interface.
    Regards
    Ramki
    Edited by: Ramki on Feb 10, 2010 5:29 PM

  • Sourcing integration with ECC using PI (percentage rebate scenario)

    Hi Experts,
    Sourcing - OP 7
    ECC - configurations done for rebate conditions as mentioned in standard configuratin guide
    PI - SP02 patch 0
    We are publishing percentage rebates in Master agreement from sourcing to ECC to create Online Agreement using standard predelivered content for PI from SAP, however ECC returns error in RFC response mentioning
    No currency code found for ISO code P1 in field HEADER_CONDITION/CURRENCY_ISO
    I checked PI content & I see for % rebates, field HEADER_CONDITION/CURRENCY_ISO is always populated in PI with P1.
    What are the next steps to be done in ECC or in PI to overcome the mentioned error.
    Any pointers will be appreciated, Thanks.

    Hello Jeet,
    I would suggest to check first that in both your SAP ERP and SAP Sourcing an ISO Code P1 for percentage exists. Further I would check if a rebate with an absolute amount works fine (also create multiple amounts and check if a scale is created in ERP, just to ensure that it works in general). E.g. check if the price condition/condition type exists and is configured correctly in ERP.
    If that all looks good you might want to check the ERP transaction BBP_ES_ANALYZE. I recently created a SDN blog about some tips and tricks, it also explains this transaction, please see ERP integration with SAP Sourcing – tips and tricks Maybe the error message(s) in this transaction will help you to better understand what is going on.
    Hopes that will help you
    Markus

  • WLC not integrating with Radius Server

    Hello world,
    I have the following situation:
    One WLC 2000 Series (software version 7.0.230.0) with multiple SSID`s, one is with 802.1x integrated with a Radius Server.
    Everything worked fine until fiew days ago, when users were unable to logon via they`re certificates on Windows XP.
    The infrastracture didn`t suffer modifications.
    What i have checked: Radius certification isn`t expired, client certification isn`t expired, the password between controller and Radius is correct.
    There are no ACL`s between the WLC and the remote Server. I can ping the devices, other SSIDs on the same controller (wpa/psk) are working correct.
    The AP`s are 1242.
    I have tried deleting the SSID, configure it back. The OS on Windows Server is  2003 Standard. The AP`s are configured H-Reap.
    I have increased the Server Timeout from Radius Authentication Servers from 2 to 30 sec.
    The message logs recived on WLC Trap Logs:
    RADIUS server X.X.X.X:1812 failed to respond to request (ID 161) for client xx.xx.xx.xx.xx.xx/ user 'unknown'
    The message from the debug dot1x aaa enable:
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_CALLING_STATION_ID(31) index=1
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_CALLED_STATION_ID(30) index=2
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_NAS_PORT(5) index=3
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_INT_CISCO_AUDIT_SESSION_ID(7) index=4
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_NAS_IP_ADDRESS(4) index=5
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_NAS_IDENTIFIER(32) index=6
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_VAP_ID(1) index=7
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_SERVICE_TYPE(6) index=8
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_FRAMED_MTU(12) index=9
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_NAS_PORT_TYPE(61) index=10
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_EAP_MESSAGE(79) index=11
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_RAD_STATE(24) index=12
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Adding AAA_ATT_MESS_AUTH(80) index=13
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df AAA EAP Packet created request = 0x1cff348c.. !!!!
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df Sending EAP Attribute (code=2, length=6, id=10) for mobile xx.xx.xx.xx.xx.xx.
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00000000: 02 0a 00 06 0d 00                                 ......
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.318: 00:15:e9:33:75:df [BE-req] Sending auth request to 'RADIUS' (proto 0x140001)
    *radiusTransportThread: Mar 06 09:37:07.328: 00:15:e9:33:75:df [BE-resp] AAA response 'Interim Response'
    *radiusTransportThread: Mar 06 09:37:07.328: 00:15:e9:33:75:df [BE-resp] Returning AAA response
    *radiusTransportThread: Mar 06 09:37:07.328: 00:15:e9:33:75:df AAA Message 'Interim Response' received for mobile xx.xx.xx.xx.xx.xx.
    *Dot1x_NW_MsgTask_7: Mar 06 09:37:07.329: 00:15:e9:33:75:df Skipping AVP (0/27) for mobile xx.xx.xx.xx.xx.xx.
    The messages on Windows 2003 Standard:
    User Y was denied access.
    Fully-Qualified-User-Name = xx.domain.com/Users_T/user
    NAS-IP-Address = X.X>X.X
    NAS-Identifier = Cisco_
    Called-Station-Identifier = ---------------------
    Calling-Station-Identifier = ---------------------
    Client-Friendly-Name = ---------------------
    Client-IP-Address = ---------------------
    NAS-Port-Type = Wireless - IEEE 802.11
    NAS-Port = 1
    Proxy-Policy-Name = Use Windows authentication for all users
    Authentication-Provider = Windows
    Authentication-Server = <undetermined>
    Policy-Name = Wireless Policy
    Authentication-Type = EAP
    EAP-Type = Smart Card or other certificate
    Reason-Code = 262
    Reason = The supplied message is incomplete.  The signature was not verified.User Y was denied access.
    Fully-Qualified-User-Name = xx.domain.com/Users_T/user
    NAS-IP-Address = X.X>X.X
    NAS-Identifier = Cisco_
    Called-Station-Identifier = ---------------------
    Calling-Station-Identifier = ---------------------
    Client-Friendly-Name = ---------------------
    Client-IP-Address = ---------------------
    NAS-Port-Type = Wireless - IEEE 802.11
    NAS-Port = 1
    Proxy-Policy-Name = Use Windows authentication for all users
    Authentication-Provider = Windows
    Authentication-Server = <undetermined>
    Policy-Name = Wireless Policy
    Authentication-Type = EAP
    EAP-Type = Smart Card or other certificate
    Reason-Code = 262
    Reason = The supplied message is incomplete.  The signature was not verified.
    Can anyone help why i cannot log the users via 802.1x ?

    Okay that is good..... this is what I would do next.  I would create a test ssid that uses PEAP MSchapv2 and create a new policy in IAS that is basic.  Allow 802.1x wireless and user group only and see if you can reconfigure one of the XP machines for PEAP.  Can you also post a screen shot of your polices (connection and network) so we can review it. 

  • Integration with IP and other 3rd party procurement apps

    I don't know about the rest of you, but one of the primary hurdles that we have to get past with target customers is that most of the big ones already have procurement systems that produce POs. Ideally, they'd like to post those POs as either an RFQ/Auction or directly as a PO for a particular supplier. This is particularly true for big manufacturers who initially want to bring their OEM suppliers and contract manufacturers into an on-line procurement environment and then expand that to auctions and RFQs later. The scenario refers primarily to those trading partners that do not have their own back-end systems and EAI solutions.
    We think we have come up with a workaround by taking the EDI produced out of the buyer's system into our Hub, transforming it into our cannonical form, then turning the data into an OEX OAG XML auction document, uploading it through a dummy account for that buyer, once the bid is placed by the supplier, automatically accepting, and generating a final PO as a PO Akn back to the buyer's system. It is a little hokey, but it is the best we've been able come up with. (Of course, once the PO message from the buyer's system is sent, the buyer wants nothing to do with it after that. They just want to receive a PO Akn back).
    Any ideas?
    David Luper
    [email protected]
    415-325-5028

    David,
    It sounds like you are hitting on two of our critical issues. Specifically, our issues are the following:
    1. We have customers with existing procurement systems containing requisitioning/workflow/approval functionality. They want to continue using this functionality while using our Exchange for sourcing/bidding/tabulation/etc. This gives us the requirement for integration to these procurement systems (iP, SAP, Peoplesoft, etc). We need the ability to load to the Exchange all the parameters for an Auction (open date, close date, type of auction, etc), line items, and potentially even suppliers. We envision that this document load generated from the requisition system could even be flagged to open the auction immediately without the buyer ever having to go to the Exchange. The buyer would then go to the Exchange to see bids and to do the award process. Once the bid was awarded, this would generate a document back to the requisition system for issuance of a PO by the procurement system.
    Our second issue is the ability to simply issue a manual PO. This could be done within the Exchange and/or a procurement system. Right now we are missing significant transaction volume due to the inability for a buyer to go issue a direct PO to a vendor on the Exchange without going through the auction or catalog process. If the goods the buyer wants to buy are under their threshold for required bidding and are not in the catalog, they go around the Exchange with a paper PO. This eliminates all the benefits of integration and our chance to charge for this transaction
    Both of these issues are extremely critical for us. Dee, what can you tell us about the integration capabilities with procurement/requisition systems or the manual PO functionality? We have heard multiple conflicting reports about the availability of integration with iP.
    Thanks!
    John Alexander
    Way2Bid
    [EMAIL][email protected][EMAIL]
    417.447.5042 (Voice)

  • OAM 11g integration with Kerberos on cluster with load-balanced virtualhost

    Hello!
    I need to make a Kerberos integration with OAM.
    I find following notes about OAM 11g: WNA Configuration for HA Clusters [ID 1365888.1] (https://support.oracle.com/epmos/faces/ui/km/SearchDocDisplay.jspx?_afrLoop=223640518878014&type=DOCUMENT&id=1365888.1&displayIndex=1&_afrWindowMode=0&_adf.ctrl-state=14ehvbh4z2_61).
    "In an OAM Clustered environment, the OAM Principal for WNA must be the same on all tiers i.e. the load-balanced virtualhost for the OAM cluster.
    Therefore each OAM managed server will reference the same keytab file, generated for Principal HTTP/<virtualhost.domain>, and the keytab file will be in the same location on all OAM managed servers.
    For example: ${DOMAIN_HOME}/domains/${DOMAIN_NAME}/config/fmwconfig/oam/<keytab filename>.
    After copying the keytab file to the same directory on all OAM managed server machines, complete the configuration of the Kerberos authentication module in OAM Administration Console (/oamconsole).
    The AdminServer will ensure that the oam-config.xml file on all OAM managed server tiers in the cluster is updated with this configuration."
    The question is; When I generate oam.keytab with following command, What is the name of the server that I will must put in the command? Virtualhost (load-balanced), Node1 or Node2?
    ktpass -princ HTTP/<servername>@DOMAIN -pass XXXXXXX mapuser DOMAIN\user -out oam.keytab.
    Thanks in advance and best regards!
    PS: Sorry if my english is not clear.

    David,
    Your Principal name should be the SSO LB URL.(ie :sso.mycomany.com)
    ktpass -princ HTTP/sso.mycomany.com@DOMAIN -pass XXXXXXX mapuser DOMAIN\user -out oam.keytab.
    Also make sure sso.mycomany.com has a reverse DNS configured correctly.
    you can check using dig command
    ping sso.mycomany.com
    What ever the ip-address
    dig -x <IP-ADDRESS>
    Check in the reverse DNS section there should be 1 record.
    ;; ANSWER SECTION:
    1.1.1.1.in-addr.arpa. 3600 IN PTR sso.mycomany.com.
    Let me know if you have more questions.
    Thanks
    Saurabh

  • Single Switch for multiple clusters

    We currently have multiple independent RAC clusters, each with a low cost switch for the interconnect.
    For example server1 and server2 is the development RAC. Server3 and server4 are the test RAC
    and server5 and server6 are the production rac.
    Can a single 24 port switch handle the interconnect traffic for these 3 separate RACs?
    We will have a second switch for redundancy, but can the single switch handle the traffic for
    3 separate RAC clusters?
    Sincerely
    Dan Clark

    Can a single 24 port switch handle the interconnect traffic for these 3 separate RACs?That depends on the number of nodes in each cluster, and the redundancy of the Interconnect.
    If you are running a single cable from the HCA on the node into the switch, your switch can support 24 nodes in total. If you are running dual connections, then only 12.
    Performance.. well, I doubt that it will be a problem as a 24 port switch is designed to handle 24 connections. As the Interconnect needs to be fast, this switch is not a normal switch, but one that supports higher speeds and bigger pipes like Gig ethernet or Infiniband for example.
    On the latter you can use Sockets Direct Protocol (SDP) or SCSI RDMA Protocol (SRP) - which is significantly faster than UDP (which is considered the faster protocol itself in the IP suite).
    In your shoes, my biggest concern would be redundancy. If that switch fails for whatever reason, 3 clusters are down. And that is IMO unacceptable. I would instead consider either a more expensive switch with inbuilt redundancy (dual CPUs, PSUs, etc), or look at using 2 switches (which is a saving of 1 switch for 3 clusters).
    Remember that RAC is only as good as the h/w architecture. If that is not adequate ito redundancy (or performance), it is clearly going to be shown by RAC itself.

  • Biztalk 2013 integration with CRM 2015

    Hi,
    I am working on a BizTalk 2013 integration with CRM 2015 on premise server and currently developing a POC to see how it works. There are quite a few good articles on BizTalk-CRM integration and I am referring them while building this POC. However I thought
    of taking expet’s advice for this. Here are my requirements as of now.
    Search CRM contacts by Fname, Lname etc.. And definitely not by Contact ID which is GUID. In this case BizTalk can expect responses like Single contact, Multiple contacts, Contact Not found or soap fault.
    Create a single contact record in CRM from BizTalk.
    Update Contact details in CRM like multiple addresses for a single contact.
    So we have 2 approaches to integrate CRM, using untyped soap endpoint or CRM SDK.
    I did an orchestration today with soap endpoint for “Retrieve” operation bypassing a GUID and can see the response returned. But how to pass the fname, lname etc. instead of GUID as there are no fields to populate in the soap Retrieve schema?
    Is it wise decision to use untyped soap UI instead of CRM SDK proxy by keeping the above 3 requirements in mind?
    Using SOAP endpoint, how can achieve the requirement 1 that returns multiple contacts?
    I have seen articles using Query and Fetch to get multiple contacts and I really don’t understand it. Is it the only way to get multiple contacts?
    Thanks!
    JB

    Hi Abhishek,
    While exploring the CRM REST capabilities i found the below content on MSDN that says the REST endpoint works only for getting the contact by uniqueID and not by firstname and lastname etc...
    https://msdn.microsoft.com/en-us/library/gg328090.aspx#BKMK_RetrievingRecords
    Retrieving records
    HTTP GET is used when you are retrieving records. When a unique identifier for
    a specific record is provided, only that record will be retrieved. Otherwise, any system query options that are defined will be applied and up to 50 records will be retrieved that match any system query option filters.
    We can't use a GUID as it is name search from customer website and GUID available at that stage.
    So if it works only with GUID then we have to use either CRM SDK or SOAP enpoint right?
    Cheers
    JB

Maybe you are looking for