Best practice multi-org, MW, SOA, Siebel authen with static or dynamic url

All,
My client integration lead had a question about the current best practices for multi-org structures with MW, SOA and Siebel. Internally the client contact is being pressured to get dynamic urls for authentication (for each area and new addition…currently exposed web services include Acct, Payment, Contact, etc… currently 60-70 services). However, he would like to stay with his current process for web service integration and just add pos id, user id, org id, etc. in the message string that is passed.
Please let me know what you think and why so I can pass this information along.

Hi even we too struck up with the same kind of issue.Please let me know if you got any solution for this.Your help is highly Appreciated.
Thanks,
Ravi Kasu.
[email protected]

Similar Messages

  • Best-practice on versioning a soa suite-application

    Hi everyone,
    I recently organised a seminar for customers concerning the Soa Suite Stack and one of the interesting questions asked that day was a versioning-question.
    Let's say we've build a bpel/esb application interacting with different external and internal webservices and we've deployed this application to our production environment and we need to change an internal web service.
    How can we add versioning to this heterogenous system in a consistent way? I know you can tell esb which version of the bpel process it needs to use, but what about the custom and external webservices that we've integrated with?
    My 2 cents: You need to add a versioning-tag to you custom web services that you need to manage yourself, and use this versioning tag in the services your integrating with.
    Could somebody point me out what best practice is, or what Oracle's development team is working out regarding versioning systems for SOA-applications?

    Marc,
    Are you saying versioning isn't supported in ESB now? I thought that ESB already uses the versioning tag from bpel when you're integrating bpel and esb?
    During the development of my demo I've seen that the version-tag was used when invoking the bpel process through a soa service.

  • Best practice for managing a Windows 7 deployment with both 32-bit and 64-bit?

    What is the best practice for creating and organizing deployment shares in MDT for a Windows 7 deployment that has mostly 32-bit computers, but a few 64-bit computers as well? Is it better to create a single deployment share for Windows 7 and include both
    versions, or is it better to create two separate deployment shares? And what about 32-bit and 64-bit versions of applications?
    I'm currently leaning towards creating two separate deployment shares, just so that I don't have to keep typing (x86) and (x64) for every application I import, as well as making it easier when choosing applications in the Lite Touch installation. But I know
    each deployment share has the option to create both an x86 and x64 boot image, so that's why I am confused. 

    Supporting two task sequences is way easier than supporting two shares. Two shares means two boot media, or maintaining a method of directing the user to one or the other. Everything needs to be imported or configured twice. Not to mention doubling storage
    space. MDT is designed to have multiple task sequences, why wouldn't you use them?
    Supporting multiple task sequences can be a pain, but not bad once you get a system. Supporting app installs intelligently is a large part of that. We have one folder per app install, with a wrapper vbscript that handles OS detection. If there are separate
    binaries, they are placed in x86 and x64 subfolders. Everything runs from one folder via the same command, "cscript install.vbs". So, import once, assign once, and forget it. Its the same install package we use for Altiris, and we'll be using a Powershell
    version of it when we fully migrate to SCCM.
    Others handle x86 and x64 apps separately, and use the MDT app details to select what platform the app is meant for. I've done that, but we have a template for the vbscript wrapper and its a standard process, I believe its easier. YMMV.
    Once you get your apps into MDT, create bundles. Core build bundle, core deploy bundle, Laptop deploy bundle, etcetera. Now you don't have to assign twenty apps to both task sequences, just one bundle. When you replace one app in the bundle, all TS'es are
    updated automatically. Its kind of the same mentality as active directory. Users, groups and resources = apps, bundles and task sequences.
    If you have separate build and deploy shares in your lab, great. If not, separate your apps into build and deploy folders in your lab MDT share. Use a selection profile to upload only your deploy side to production. In fact I separate everything (except
    drivers) into Build and deploy folders on my lab server. Don't mix build and deploy, and don't mix Lab/QA and production. I also keep a "Retired" folder. When I replace an app, TS, OS, etcetera, I move it to the retired folder and append "RETIRED - " to the
    front of it  so I can instantly spot it if it happens to show up somewhere it shouldn't.
    To me, the biggest "weakness" of MDT is its flexibility. There's literally a dozen different ways to do everything, and there's no fences to keep you on the path. If you don't create some sort of organization for yourself, its very easy to get lost as things
    get complicated. Tossing everything into one giant bucket will have you pulling your hair out.

  • Best Practice to Setup an application to work with both oracle and db2 db

    Hi,
    We have an application that currently supports both oracle and db2 databases . It is currently using JPA with eclipselink as backend mechanism and we want to move to ADFBc as our backend . So what is best practice to do this?
    I came across an old post in https://groups.google.com/forum/#!topic/adf-methodology/UlJZSTu14Io that states to create two different model projects to support oracle and db2 .
    Is this still the standard ? Is there a way that we could work around rather than creating multiple projects?
    How do i get the view controller to work with multiple model projects if this is the case?
    Thank you.

    Thanks for the response .
    The problem is i would have the same schema on both the databases (both in oracle and db2) .
    I don't see a scenario where i want both the application model projects at the same time . What i meant is , Application will either be deployed with Oracle DB or DB2 but not with both on a production environment .
    So is there a way where i just change the connection parameters alone rather having two different models ?
    If i have to use two model projects ,  would it be possible say i built the view controller for Oracle , and use it for DB2 if i make sure the BC objects names are similar between model projects and switch just the model jar based on the deployment environment ?
    -Sam

  • Best practice for including additional DLLs/data files with plug-in

    Hi,
    Let's say I'm writing a plug-in which calls code in additional DLLs, and I want to ship these DLLs as part of the plug-in.  I'd like to know what is considered "best practice" in terms of whether this is ok  (assuming of course that the un-installer is set up to remove them correctly), and if so, where is the best place to put the DLLs.
    Is it considered ok at all to ship additional DLLs, or should I try and statically link everything?
    If it's ok to ship additional DLLs, should I install them in the same folder as the plug-in DLL (e.g. the .8BF or whatever), in a subfolder of the plug-in folder or somewhere else?
    (I have the same question about shipping additional files too, such as data or resource files.)
    Thanks
                             -Matthew

    Brother wrote:
    I probably should have posed this question here before I delved into writing Java to get data for reports, but better late than never.
    Our ERP is written in COBOL. We have a third party ODBC which allows us to access data using a version of SQL. I have several Java sources compiled in my database that access the data and return something relevant. The Java sources are written in a procedural style rather than taking advantage of object oriented programming with attributes and methods.
    OO is a choice not a mandate. Using Java in a procedural way is certainly not ideal but given that it is existing code I would look more into whether is well written procedural code rather than looking at the lack of OO.
    Now that I am becoming more comfortable with the Java language, I would greatly appreciate any feedback as to best practices for incorporating Java into my database.
    My guess is that it would be helpful to model the ERP "tables" with Java classes that would have attributes, which correspond to the fields, and methods to return the attributes in an appropriate way. Does that sound reasonable? If so, is there a way to automate the task of modeling the tables? If not reasonable, what would you recommend?Normally you create a data model driven by business need. You then implement using whatever means seem expedient in terms of other business constraints to closely model that data model.
    It is often the case that there is a strong correlation between data models and tables but certainly in my experience it is rare when there are not other needs driven by the data model (such as how foreign keys and link tables are implemented and used.)

  • What is the best practice for creating master pages and styles with translated text?

    I format translated text all the time for my company. I want to create a set of master pages and styles for each language and then import those styles into future translated documents. That way, the formatting can be done quickly and easily.
    What are the best practices for doing this? As a company this has been tried in the past, but without success. I'd like to know what other people are doing in this regard.
    Thank you!

    I create a master template that is usually void of content, with the exception I define as many of the paragraph styles I believe can/will be used with examples of their use in the body of the document--a style guide for that client. When beginning a new document for that client, I import those styles from the paragraph styles panel.
    Exception to this is when in a rush I begin documentation first, then begin new work. Then in the new work, I still pull in those defined paragraph and or object styles via their panels into the new work.
    There are times I need new styles. If they have broader applicability than a one-off instance or publication, then I open the style template for that client and import that style(s) from the publication containing the new style(s) and create example paragraphs and usage instructions.
    Take care, Mike

  • JDev 11g: Best Practice Multi Field Validation / Reset in a Dialog Control

    I'm using JDev 11g and want to do the following (excuse my newbie learning curve):
    1. Ability for the user to have a "dialog" form that they can edit their user information such as name, password, email, etc.
    2. When user clicks OK want the system to validate that password fields are the same (there are two of them).
    3. When user clicks Cancel I want the form to reset and the dialog to close.
    Sounds easy right? However, I've seen several ways to implement and none of them seem straight forward and don't appear to work like I would want so instead spending the next day coding want to know if there's an easy solution.
    For # 2:
    The examples on the web use a phase listener for multi-form validation which when using a small dialog in a larger arena seems like overkill since the entire view will get my event right? Is there a similar thing for subform or subview?
    Another example uses a listener for value change, however that doesn't work to well when trying to validate two fields match.
    For # 3:
    A standard OK Cancel dialog does not auto reset a form within it (it leaves the text - even invalid text in the field). Adding my own button for reset doesn't close the form and adding a reset listener doesn't close the form either. Is there a simple way and I'm just overlooking it? Should I use the standard buttons and somehow assign the cancel to reset the form or define my own and somehow close the popup? There's not too much documentation I can find about using Javascript to modify objects such as closing a dialog.
    Anyone have the best approach that should be followed with less headache downstream and better portability to other forms with more complex validation? Any examples I missed that shows this working in a dialog?
    Thank you,
    Kris

    Hi,
    I don't understand #2, so ignoring it
    #1 You can set the immediate property on the two password fields. In a value change listener you compare the two provided values and if they match, you continue processing. Otherwise you set an error message and call response complete. Note that as much as we like to provide everything declarative, coding is something that we cannot neglect completely in a programming environment.
    #3 You can intercept the dialog close event and in this event determine which button was pressed. If it was the cancel button, get the current row from the underlying iterator and call refresh() passing one of the arguments you find when typing "Row." e.g DB refreshWithDBForget.
    Frank

  • Best Practice to save record in a table with appropriate trigger

    Hi,
    At the same time, five users are inserting client information in a client table through Oracle form. Client table has the following fields:
    CLIENT ID
    CLIENT NAME
    CLIENT ADDRESS
    CLIENT ID is generating automatically by calling a procedure. In this procedure I am using MAX function to get maximum value of CLIENT ID. After that, I store newly generated CLIENT ID in a DATA BLOCK item say :MASTER.CLIENT ID.
    The problem is that all five users do get same MAX value(suppose 40) of CLIENT ID at the same time, and oracle form will surely throw an exception at the time of insertion the record in client table. CLIENT ID is PK and it is a member of MASTER DATA BLOCK.
    I hope all above will clearly illustrate the problem, further please guide, can PRE-INSERT trigger may handle this problem efficiently ? , if so, then how ? ...
    Thanks,

    Hello,
    Welcome to the forum!
    CLIENT ID is generating automatically by calling a procedure.So, in which trigger you are calling that procedure?
    After that, I store newly generated CLIENT ID in a DATA BLOCK item say :MASTER.CLIENT ID. I would guess that you are using the code generation procedure in the WHEN-CREATE-RECORD trigger block-level. Because this trigger normally intialize the value for new record insertion. If no then specify.
    further please guide, can PRE-INSERT trigger may handle this problem efficiently ? ,Yes, PRE-INSERT will work without any problem.
    if so, then how ? ... Because, PRE-INSERT will fetch the MAX number from the table at the time of save not at the time of creation. So, suppose five user will enter record at the same time but when they will save then PRE-INSERT will get the max sequence number. And it will never create problem. Because in five user's insertion sure there will be some time difference. So, max number can be get easily without any problem.
    -Ammad

  • Best practice on creating a managable questionnaire/survey with JSF

    Hello all,
    As a part of a bigger project we need to build a questionnaire with managable questions. This means that clients can add/modify/delete questions as they wish. The database side is figured out and designed, but what im really thinking is how to collect the answers. The questions are populated from the database and the answers will be given either trough radiobuttons or textfields.
    The thing i cant figure out is how to get the the answers of questions to the bean so they can be processed? It cant be said how many questions there will be so creating properties in to the bean beforehand wont work.
    We are using jsf,hibernate and icefaces.
    Thanks in advance.. this is really bugging me

    Shoul have searched the forums a bit better. Solution found at http://forums.sun.com/thread.jspa?forumID=427&threadID=5117301 with some modifications.

  • Any "Best Practice" regarding use of zfs in LDOM with zones

    I have 3 different networks and I want to create a guest-domain for each of the three networks on the same control domain.
    Inside each guest-domain, I want to create 3 zones.
    To make it easy to handle growth and also make the zones more portable, I want to create a zpool inside each guest domain and then a zfs for each zoneroot.
    By doing this I will be able to handle growth by adding vdisks to the zpool(in the guest domain) and also to migrate individual zones by using zfs send/receive.
    In the "LDoms Community Cookbook", I found a description on how to use zfs clone in the control domain to decrease deploy time of new guest domains:
    " You can use ZFS to very efficiently, easily and quickly, take a copy of a previously prepared "golden" boot disk for one domain and redeploy multiple copies of that image as a pre-installed boot disk for other domains."
    I can see clear advantages in using zfs in both the control domain and the guest domain, but what is the downside?
    I ends up with a kind of nested zfs where I create a zpool inside a zpool, the first in the control domain and the second inside a guest domain.
    How is zfs caching handled, will I end up with a solution with performance problems and a lot of I/O overhead?
    Kindest,
    Tor

    I'm not familiar with the Sybase agent code and you are correct, only 15.0.3 seems to be supported. I think we'd need a little more debug information to determine if there was a workaround. May be switching on *.info messages in syslogd.conf might get some more useful hints (no guarantee).
    Unfortunately, I can't comment on if, or when, Sybase 15.5.x might be supported.
    Regards,
    Tim
    ---

  • SOA 11g  Composite Deployment across multiple Instances: Best Practice

    Hi,
    We have a requirement where we need to deploy the composite acrocss mutilple instances(DEV,TEST,Production) without JDEV.
    We are using SOA11.1.3.3(cluster) and linux OS.
    Pls suggest me what is the best practice to deploy the SOA composite.
    Thanks,
    AB

    Why there are different ways to deploy the composite in different environment? Depending upon the environment, it's business importance increases and hence access to developers get more restricted and hence there are many ways for deploying. If you are developing an application, you would not like to export a SAR and then login to EM console to deploy it. For a developer it is always very convenient to use IDE itself for deployment and hence JDev is preferably used for Dev instances for deployment.
    Once development finishes, developer will check in the artifacts into version control system and if you want to deploy to a test instance then you have to check out the stable version, compile it and package it and then only you can deploy. Hence for test instances, ANT is the preferable mode for deployment. (Remember that developers may not be able to connect to Test environment directly and hence deployment from JDev is not possible)
    Once a configuration gets tested in Test env, it's SAR should be checked in into version control system, so that it would not be required to recompile and repackaging of it. This will also make sure that any artifact which is not tested on Test instance, must not go to higher instances (like UAT/PROD). Now in Pre-Prod/UAT/Prod, you may simply access the EM console and deploy the SAR which is already tested in Test instance. Remember that there will be very limited access to such critical environments and hence using Role based access of EM it would be very easy to deploy the resources. Moreover, it is more secure mode of deployment because only those users which have appropriate priviledge, will be able to deploy, and it would also be easier to track the changes.
    What is the proc and cons if we use only one way to deploy the composite across the Instances...As such there is no major pros and cons. You may use EM for all environments but it may be a little discomfort for developers/deployers of test environment. You may also use ANT for all the environments but it is not suggested until and unless you have a very good and secure process in place for deployments.
    Regards,
    Anuj

  • SAP best practice on scheduling agreements

    Hello all,
    Our client is an automotive supplier and currently in the SAP environment we use spot POs for indirect purchasing (thru SRM) and scheduling agreement for direct procurement. The scheduling agreements have multiple lines and we try to process as much invoices as ERS as possible.
    Currently some of the business units are proposing to have single line items in the SA rather than the multiple ones. The reason being huge no of SA lines being faxed and also supposing it to be a SAP suggested best practice to have ONLY one line item per SA.
    I am not sure of any such best practice. Could you please help me with an single Vs multi line scheduling agreements pros and cons. Also what is the best practice that are seen in the automotive sector or being suggested by SAP.
    Any SAP official documentation in the public domain will also help.
    Thanks

    As per my experience, we have used both the ways, both have its pro's and cons.
    In one instance we may feel that it is easy to handle one line per one SA and in that way we will create many SA's and in ither instance we may fell having one SA for multiple lines.
    SAP does not give any best approach...or else SAP woul dhave provided the option for only one line.
    As per my experience the Structuring of a SA should be based on the way you procure from the supplier  / Vendor.
    We have one supplier who supplies 800+ materials, we can create one SA with 800 lines... But we have created multiple SA's based on the product category whcih is a bit easy to identify whcoih material goes into which SA and the Vendor is also happy as he knows that for one product there is a specific SA.
    Hope this helps you in resolving your issue.,

  • DNS best practice question

    Hello,
    we currently have an issue regarding DNS in a multiple Domain Forest.
    first of all, in the forest there are 5 Domains (names changed):
    dom1.domain.org
    sub.dom1.domain.org
    dom2.domain.org
    dom1.url.de
    dom.de
    As you see, a forest full of Domains not matching ;-)
    We also have multiple sites, and as per network requirements, replication is made trough Domain: dom1.domain.org
    All other Domains replicate only with this one.
    The DNS is currently set up as follows:
    Each Domain Controller holds its own domain as primary AD integrated Domain in DNS (Domain wide repl.).
    All others are set up as Forest Wide AD integrated Stubs.
    Each startup we get Event 4515 on the DCs, that a Zone is available twice.
    So, I have to troubleshoot this infrastructure now.
    Can you tell me, what is best practice here to set up DNS correctly with less replication traffic as possible?
    Best regards

    By Default DNS Zone replication Scope is Domainwide but except _MSDCS Zone . _MSDCS Zone replication should be forest wide. In addition Replication Scope can be decide as per your business requirment.
    Regards~Biswajit
    Disclaimer: This posting is provided & with no warranties or guarantees and confers no rights.
    MCP 2003,MCSA 2003, MCSA:M 2003, CCNA, MCTS, Enterprise Admin
    MY BLOG
    Domain Controllers inventory-Quest Powershell
    Generate Report for Bulk Servers-LastBootUpTime,SerialNumber,InstallDate
    Generate a Report for installed Hotfix for Bulk Servers

  • Best Practice for setting systems up in SMSY

    Good afternoon - I want to cleanup our SMSY information and I am looking for some best practice advice on this. We started with an ERP 6.0 dual-stack system. So I created a logical component Z_ECC under "SAP ERP" --> "SAP ECC Server" and I assigned all of my various instances (Dev, QA, Train, Prod) to this logical component. We then applied Enhancement Package 4 to these systems. I see under logical components there is an entry for "SAP ERP ENHANCE PACKAGE". Now that we are on EhP4, should I create a different logical component for my ERP 6.0 EhP4 systems? I see in logical components under "SAP ERP ENHANCE PACKAGE" there are entries for the different products that can be updated to EhP4, such as "ABAP Technology for ERP EHP4", "Central Applications", ... "Utilities/Waste&Recycl./Telco". If I am supposed to change the logical component to something based on EhP4, which should I choose?
    The reason that this is important is that when I go to Maintenance Optimizer, I need to ensure that my version information is correct so that I am presented with all of the available patches for the parts that I have installed.
    My Solution Manager system is 7.01 SPS 26. The ERP systems are ECC 6.0 EhP4 SPS 7.
    Any assistance is appreciated!
    Regards,
    Blair Towe

    Hello Blair,
    In this case you have to assign products EHP 4 for ERP 6 and SAP ERP 6 for your system in SMSY.
    You will then have 2 entries in SMSY, one under each product, the main instance for EHP 4 for ERP 6 must be central applications and the one for SAP ERP 6 is SAP ECC SERVER.
    This way your system should be correctly configured to use the MOPZ.
    Unfortunately I'm not aware of a guide explaining these details.
    Some times the System Landscape guide at service.sap.com/diagnostics can be very useful. See also note 987835.
    Hope it can help.
    Regards,
    Daniel.
    Edited by: Daniel Nicol on May 24, 2011 10:36 PM

  • Best practices for loading swf's

    Greetings,
    Using CS5 AS2
    I'm creating a website in flash (all the files will be in one directory/folder on SharePoint) and want to make sure that what seems to be working fine is best practice.
    I have an index.swf with many buttons which will take the user to landing pages/content/other swfs. On these different buttons I have the script...
    on (release) {loadMovieNum("name.swf", 0);}                I could also do just {loadMovie("name.swf", 0);} ??
    The movie transitions nicely to name.swf and on this page I have a button that returns the user to the index.swf...
    on (release) {loadMovieNum("index.swf", 0);}   Things move back to index.swf nicely and user can chose to go to another landing page.
    It looks like I'm on the right track, bc nothing is going awry? but want to check. Am I following best practices for moving from one swf to another within a website?
    Thanks for help or confirmation!!

    loading into _level0 (and you should use loadMovieNum, not loadMovie, when loading into a level) undermines some of the benefits of a flash application:  all the assets must load with each display change and the user sees a flash instead of appearing to transition seamlessly from one display to the next.

Maybe you are looking for