ODP + Client Version Management Best Practices

I am working with a client who is using the Oracle Developer Tools for Visual Studio to develop their application. The IT folks deploy a new server with the latest version of the Oracle client (i.e. 11.1.0.7.0) and the developers are using the latest ODT (i.e. 11.1.0.7.20). As such the publisher policy never gets a chance to add any value. Because of the unusual versioning scheme we always end up having to copy the Oracle.DataAccess.dll from the Oracle client into the deployment folder and add an assembly binding redirect. If we do not we get the following exception:
The provider is not compatible with the version of Oracle client
As you might expect this is VERY annoying to deal with as it adds complexity to our deployment process. We use TNS names to connect and those files are stored in the client\network\admin folder as well. Can someone please help me clear up what the best practice is to deploy an application developed using ODT/ODP.NET so that we don't have to hack the Oracle.DataAccess.dll every time?
Thanks,
Colin

Thank you for your answer.
So you are basically proposing to work with the vms in the dev folder and publish the desired version to the test folder. And the test folder is out of version control in this scenario if I understood you correctly (like simple data storage)?
And how would you suggest promoting reports to the prod system? Simply by promoting the
desired version from dev folder directly to prod? This would probably lead to inconsistence because we would need to promote from dev system to test and dev to prod instead of promoting a straight line from dev over test to prod. Furthermore it would not solve the problem of the promoting result itself (A new folder called dev will be generated in prod but the report gets promoted to the prod folder if there was no report before).
Thank you for the link. I came across this page just a few days ago and found also lots
of other tutorials and papers describing the basic promoting process. The promoting process in general is clear to me but I wondered if it is possible to change some parameters to  prevent folder renaming for example.
Regards
Lars

Similar Messages

  • Working with version management and promotion management best practices BO 4.1

    Hi Experts
    I wondered if anybody knows if there is a document or something about best practices to work with the version management and promotion management in BO 4.1?
    Our Environment includes two servers. The first one is our development and test server. The second server is our prod system.
    Now on the dev server we have basically two folders called dev and test. We control access to them with a right system based on the folder structure.
    My question now is how you would work in this scenario (third server is not an option). Main target is to have as few reports as possible. Therefore we try to work with the version management system and only have one version of each report in the dev folder of the cms. But this is where problems start. Sometimes the newest version is not the version we want to publish to the test folder or even prod server.
    How would you publish the report to the other folder? Make a copy of the concerned report (transport to the same system via promotion management is not possible). Also how would you use the version management in regards to the folder structure? Only use version management in dev folder and export reports to test folder (out of vms control) or also use vms in test folder and how would that work?
    Further more I’d be interested in learning best practices with promotion management. I found out that the promotion of a report that doesn’t exist in prod doesn’t make any problems. But as soon as an older version already exists there is only partial success and the prod folder gets renamed to “test”.
    Any suggestions on how to handle these problems?
    Thank you and regards
    Lars

    Thank you for your answer.
    So you are basically proposing to work with the vms in the dev folder and publish the desired version to the test folder. And the test folder is out of version control in this scenario if I understood you correctly (like simple data storage)?
    And how would you suggest promoting reports to the prod system? Simply by promoting the
    desired version from dev folder directly to prod? This would probably lead to inconsistence because we would need to promote from dev system to test and dev to prod instead of promoting a straight line from dev over test to prod. Furthermore it would not solve the problem of the promoting result itself (A new folder called dev will be generated in prod but the report gets promoted to the prod folder if there was no report before).
    Thank you for the link. I came across this page just a few days ago and found also lots
    of other tutorials and papers describing the basic promoting process. The promoting process in general is clear to me but I wondered if it is possible to change some parameters to  prevent folder renaming for example.
    Regards
    Lars

  • Solution Manager best practices about environments

    Hello,
    we intend to use Solution Manager 4.0.
    My question : I wonder whether we need to have a single instance a SM (production) or whether we need to have multiple instances (one development SM where developments and customizing will be performed and one production SM populated with transport requests coming from the development SM) ?
    What are the best practices ?
    Thank you.
    Regards,
    Fabrice

    Dear Fabrice,
    In principle you donot need to have 2 instances of Solution Manager. 1 Instance is sufficient enough for monitoring all the Satellite Systems.
    However if you intending to have Customized ABAP on Solution Manager then it might be a good idea to do so in an different client in the same instance keeping the client as an development client.
    Most of the customizing in Solution Manager is non transportable hence it should be directly done in the productive client.
    Hope this answers your queries.
    Regards
    Amit

  • What are project management best practices?

    I created a test project in Premiere Elements 12, and saved it in a directory named "Michaels Posters".   Then I archived the project to this directory and it created a "Copied_My\ new\ video\ project1" directory with all of the media files.  Then I added a video clip to the project, archived it again, and it created the "Copied_My\ new\ video\ project1_001" folder below.
    My first real project will be a video highlights video of my 4 years old for 2013.  This will involve editing the same project several nights a week, for maybe a couple of months.  This would result in numerous "Copied_My\ new\ video\ project1_NNN" directories being created, assuming I archive the project each night.
    So what is the best practices for managing a larger project like this, and avoid using a lot of disk space for the same project?
    Michaels\ Posters/
    ├── Adobe\ Premiere\ Elements\ Preview\ Files
    │   └── My\ new\ video\ project1.PRV
    ├── Copied_My\ new\ video\ project1
    │   ├── Adobe\ Premiere\ Elements\ Preview\ Files
    │   ├── Encoded\ Files
    │   └── Layouts
    ├── Copied_My\ new\ video\ project1_001
    │   └── Adobe\ Premiere\ Elements\ Preview\ Files
    ├── Encoded\ Files
    │   └── My\ new\ video\ project1.prel
    ├── Layouts
    └── Media\ Cache\ Files

    I do work with the LAST archived project file, which contains ALL necessary resources to edit the video.  But then if I add video clips to the project, these newly added clips are NOT in the archived project, so I archive it again.
    The more I think about it, the more I like this workflow.  One disadvantage as you said is duplicate videos and resource files.  But a couple of advantages I like are:
    1. You can revert to a previous version if there are any issues with a newer version, e.g., project corruption.
    2. You can open the archived project ANYWHERE, and all video and resource files are available.
    In terms of a larger project containing dozens of individual clips like my upcoming 2013 video highlights video of my 4  year old, I'll delete older archived projects as I go, and save maybe a couple of previous archived projects, in case I want to revert to these projects.
    If you are familiar with the lack of project management iMovie, then you will know why I am elated to be using Premiere Elements 12, and being able to manage projects at all!
    Thanks again for your help, I'm looking forward to starting my next video project.

  • Multiple room management -- best practice -- server side http api update?

    Hi Folks, 
    Some of the forum postings on multiple room management are over year old now.  I have student/tutor chat application which has been in the wild for 5 months now and appears to be working well.  There is a single tutor per room, multiple chats and soon to be a whiteboard per student, which is shared with the tutor in a tabbed UI. 
    It is now time to fill out the multiple tutor functionality, which I considered and researched when building, but did not come to any conclusions.   I'm leaning towards a server side implementation.  Is there an impending update to the http api?
    Here is what I understand to be the flow:
    1) server side management of who is accessing the room
    2) load balance and manage the room access 1 time user and owner session from the server side
    3) for my implementation, a tutor will need to login to the room, in order for it to be available
    4) Any reconnection would in turn need to be managed by the server side, and is really a special case of room load balancing.
    My fear is that at some point I'm going to need access to the number of students in the room or similar and this is not available, so that I'll need client functionality, which will need update the server side manager.
    As well, I'm concerned about delays on the server side access to which might create race conditions in a re-connect situation.  User attempts to reconnect, but server side manager thinks that the user is already connected.
    Surely this simple room management has been built, does anyone have any wisdom they can impart?  Is there any best practice guidance or any samples?
    Thanks,
    Doug

    Hi Raff, Thanks a ton for the response.
    I wasn't clear on what I was calling load balancing.  What I mean by this is room assignment for student clients.  We have one tutor per room.  There are multiple students per room, but each is in their own one-on-one chat with the tutor.
    I'm very much struggling with where to do the room assignment / room managemnt, on the server side or on the client side (if that is even possible).  In my testing it is taking upwards of 10 seconds minimum to get a list of rooms (4 virtually empty rooms) and to query the users in a single room (also a minimum of users/nodes in the queried room).   If after this point, I 'redirect' the student to the least full room, then the student incurs the cost of creating a new session and logging into the room.  As well I intend to do a bit of xml parsing, and other processing, so that 10 seconds is likely to grow.
    Would I see better performance trying to do this in the client?
    As far as the server side, at what point does a room go to 'not-active'?
    When I'm querying the roomList, I am considered one of the 'OWNER' users in the UserLists.  At what point can it be safe to assume that I have left the room? 
    Is there documentation on the meaning and lifecycle of the different status codes?  not-active,  not-running, and ok?  Are there others?
    How much staleness can I expect from the server-side queries?
    As far as feature set, the only thing that comes to mind is xpath and or wild card support for getNode() but i think this was mentioned in other posts.
    Regarding the reconnection issues, I am timing out the student after inactivity, and this is probably by and large the bulk of my reconnect use cases.  This and any logout interaction from the student presents a use case where I  may want reassign the student return to the same room as before.  I can envision scenarios of a preferred tutor if available etc.  In this case, I'll need to know list of rooms.  In terms of reconnection failover, this is not not a LCCS / FMS issue.
    Thanks again for responding.

  • How to make Forms less dependent of client PC ? (Best Practice?)

    Hello,
    We are on Forms 10g (and I also noticed the same pb with previous versions)
    We built quite big applications that use Webutil and jInitiator (not Java plugin)
    The application is used via Web only.
    At the deploiement of the applications, we experienced many difficulties due to specific configuration of Internet Explorer (or another browser, Firefox) of each one person of the company
    I suppose this situation has also been experienced by others...
    Do you have any suggestion of best practices for making such application less dependent of the client PC configuration ?
    Thanks in advance,
    Olivier

    Try to use http, i don't thing direct open the socket can pass throught the socket if your proxy haven't forward those ports~
    Hi All,
    I have problems with my applet working through
    gh proxy.I am using a client applet which makes a
    socket connection to a Java application running on the
    same pc as the web server.
    Everything works fine when I am directly
    ly connected.However it does't work when connected
    through a proxy? I am using object o/p & i/p streams
    for the exchange of data b/w client applet & the
    server java application.How do I overcome this
    problem.
    Currently I am using WinGate 3.0.2 as my proxy(I
    (I did't see any firewaal setting there) but I intend
    to make applet work through any proxy for that
    matter.Using a higher level protocol such as UDP would
    help in this regard?i think UDP/TCP is not the matter for that!
    This is urgent.PLease reply ASAP...
    best rgds,
    prithvi

  • SRM EBP User management - best practice followed for your customer.

    Hello All,
    What are the best practices followed on SRM User manageemnt  followed for your customers.
    (1)When employee/ buyer  leave the organisation  ? what actions you do ? do you lock the users?
    (2) If any thing interested share your experiences.
    (3) What exactly customer expects from SRM systems on SRM user management?
    (4) SAP audit/ customer audit practice on USER management ?
    Any piece of information on your experiece/ best practice  is appreciated.
    regards
    Muthu

    Thanks Peter .
    it is happening only in SRM right.
    Is any work around for this issue.
    In future SRM any planing to take care of this.
    in ECC i can delete the user whenever the user moves .
    All SRM customers will be very happy if SRM gives some workaround for this issue.
    Every customer wants to reduce cost .
    How can I find what are the opening documents for this user in one shot ?
    thanks for answering this question.
    I have seen our Eden Kelly report helps for shopping cart and other BO.
    You are doing good job on our SRM  WIKI innovative topics and discussons. I appreciate.
    why i am raising this concern is that one user left the organisation and again we want to edit the data whch entered by the left user . system will not allow us to do after deleting the user.
    so we are approaching SAP for ths help.
    It is very difficult to convice the customers on this issues.
    br
    muthu

  • Hotfix Management | Best Practices | WCS | J2EE environment

    Hi All,
    Trying to exploit some best practices around hotfix management in a J2EE environment. After some struggle, we managed to handle the tracking of individual hotfixes using one of our home-grown tools. However, the issue remains on how to manage the 'automated' build of these hotfixes, rather than doing the same manually, as we are currently doing.
    Suppose we need to hotfix a particular jar file on a production environment, I would need to understand how to build 'just' that particular jar. I understand we can label the related code (which in this case could be just a few java files). Suppose this jar contains 10 files out of which 2 files need to be hotfixed. The challenge is to come up with a build script which builds -
    - ONLY this jar
    - the jar with 8 old files and 2 new files.
    - the jar using whatever dependent jars are required.
    - the hotfix build script needs to be generic enough to handle the hotfix build of any jar in the system.
    Pointers, more in line with a WCS environment would be very much appreciated!
    Regards,
    Mrinal Mukherjee

    Moderator Action:
    This post has been moved from the SysAdmin Build & Release Engineering forum
    To the Java EE SDK forum, hopefully for closer topic alignment.
    @)O.P.
    I don't think device driver build/release engineering is what you were intending.
    Additionally, your partial post that was accidentally created as a duplicate to this one
    has been removed before it confuses anyone.

  • X2100M2 Embedded LIghts Out Manager best practice

    Hi guys,
    I'm in worry about the best practice of configuring network interface on a X2100M2 Solaris 5.10 for the Embedded LIghts Out Manager. Hope you can help. I haven't find any documents of it which explain the best practice.
    Here is the situation :
    I've have 4 network interfaces but I only need two of them. So I decide to use the bge0 and the bge1 interfaces.
    bge0 is the server interface with an IP with .157
    bge1 is the ELOM interface with an IP with .156
    In the past, it was the reverse : bge0 was the ELOM with .156 and bge1 was the network server int. with .157
    Could you please guys let me know what is the best practice? Does the int.0 must be the server one? Is it possible to have network problem with this kind of configuration?
    Thanks
    Cheers,

    hi guys,
    No one have a clue? I've got some dukeDolars to offer...
    Tks

  • Operating system image build and management best practices?

    how do we create gold images for servers/desktops,
    Best practices image management,
    How do we control changes?
    How do we prevent unauthorized changes (installation of software)?
    What tools we can use for above.

    I use MDT 2013 Lite Touch to create my images
    http://www.gerryhampsoncm.blogspot.ie/2014/03/create-customised-reference-image-with.html
    You should use in-built ConfigMgr Role Based Access Control to manage images afterwards (look at the Operating System Deployment Manager role).
    Gerry Hampson | Blog:
    www.gerryhampsoncm.blogspot.ie | LinkedIn:
    Gerry Hampson | Twitter:
    @gerryhampson

  • OWB Change Management/Version Control Best Practice

    Hi
    I am about to start developing a data warehouse using OWB 10g R2, and I've been doing quite a lot of research into the various deployment/change management/version control techniques that can be used, but am still unsure which is the best to use.
    We will have 2-3 developers working on the project, and will be deploying from Development, to Test, to Production (each will have a separate repository). We want to be able to easily identify changes made between 1 release and the next to have a greater degree of control and awareness of what goes into each release. We also wish to use a source control system to track changes (we'll probably use SVN, but I don't think that the actual SCS tool makes a big difference to our decision at this point).
    The options available (that I'm aware of), are:
    1. Full MDL export/import.
    2. Snapshot MDL export/import.
    3. Manual coding of everything using OMB Plus.
    I am loath to use the full MDL export/import functionality since it will be difficult, if not impossible, to identify easily the changes made between 1 release and the next.
    The snapshot MDL export/import functionality is a little better at comparing releases, but it's still difficult to see exactly what has changed between 1 version and the next - particularly when a change to a transformation has been made. It also doesn't cope that well with tracking individually made changes to different components of the model.
    The manual coding using OMB Plus seems like the best option at the moment, though I keep thinking "What's the point of using a GUI tool, if I'm just going to code everything in scripts anyway?".
    I know that you can create OMB Plus code generation scripts to create your 'creation' scripts, but the code generation of the Alteration scripts seems that it would be more complicated than just writing the Alteration scripts manually.
    Any thoughts anyone out there has would be much appreciated.
    Thanks
    Liffey

    Well, you can also do per-object MDL exports and then manage those in your version control system. With a proper directory structure it would be fairly simple to code an OMB+ Script that scans a release directory tree and imports the objects one by one. I have done this before, although if you are using OWB as the primary metadata location for database objects then you have to come up with some way to manage object dependency order issues.
    The nice thing about this sort of system is that a patch can be easily shipped with only those objects that need to be updated.
    And if you force developers to put object-level MDL into your version control system then your system should also have pretty reporting on what objects were changed for a release and why.
    At my current job we do full exports of the project MDL and have a deployment script that drops the pre-existing deployed version of the project before importing and deploying the new version, which also works quite well - although as you note the tracking of what has changed in a release then needs to be carefully managed elsewhere. But we don't deploy any of our physical database objects through OWB. Those are deployed from Designer, and our patch script applies all physical changes first before we replace the mappings from the OWB project. We don't even bother synching the project metadata for tables / views / etc. at deployment. If the OWB project's metadata for database objects is not in sync with Designer, then we wind up with deployment errors. But on the whole it works pretty well.

  • Software Change Management best practices

    Hi All,
    I am curious how most people here are managing the changes that they make to their SAP code that is running on the IBM i.
    For example, how do you perform version control? Multiple release management? Task management? How do you manage approvals for promotion? How do you comply with SOX "separation of duties" (i.e. someone different has to promote to production).
    Thanks for any insight here!
    Joe.

    Hi All,
    I am curious how most people here are managing the changes that they make to their SAP code that is running on the IBM i.
    For example, how do you perform version control? Multiple release management? Task management? How do you manage approvals for promotion? How do you comply with SOX "separation of duties" (i.e. someone different has to promote to production).
    Thanks for any insight here!
    Joe.

  • MCX to Configure Client Dock Applications best Practice?

    I want to use WGM preferences to manage which applications show up on a user's dock when they login. The server does not have the same applications installed as the client machines thus do not show up as an available option to add using WGM.
    _The Question:_ *Is is bizarre to install client applications on the server so it may be available to to add to a user's dock preference in WGM?*
    Currently I'm using changes I've made to the default user template as a work around to getting the dock configured to our needs. Not my favorite way to administer 6 labs, but functional right now. If it's okay, I was thinking about installing all the software we use on our client machines on the server.
    Is there a better way of managing the dock for clients in a multiple lab environment?

    Run the WGM application from a workstation that has the same applications installed in the same locations as on your client workstations. Do not make changes to the user templates on the server or the workstations. When setting the preferences for the user's dock, do not select the checkbox, 'Merge with User's Dock', as that combines the customized settings you create with the dock settings from the user template.

  • CM (Configuration Managment) Best Practices for DBs?

    Hi:
    I've got a question on a topic I haven't seen discussed before. I have a project with application code and PL/SQL code and an Oracle 10g database that has triggers, constraints, etc. For a Java or C# project the concept of CM and a central code repository works, largely because the CM tool is able to enforce limits on who can do what to a given resource over time. Developers MUST check out the source files before modifying and adding back to the CM tree.
    Database objects are more on an honor system. Even if you put the code for the tables, triggers, constraints, packages etc. under CM, nothing prevents a developer or multiple developers from modifying those objects in the development database and never updating the create scripts that are in the CM tree. So given that there are thousands of systems out there with Java/C# or whatever front ends and PL/SQL back ends in databases that have all sorts of DB objects in them, what are the best CM practices? How does a developer know the package he's about to modify is the last one that got pushed to the production server?
    Constraints are problematic in several unique ways. They can be created in several places by different syntax (in the table create command; as a stand alone command after the table exist; named or system generated name if defined in the table definition). Things like SQL Developer and Toad can create code trees that contain all the scripts needed to build a database from scratch and a top level build script to run them all, but what's the best structure for that? Do table create scripts JUST define columns and separate scripts create all constraints?
    I like to think someone has crossed this particular bridge before and has some answers to my questions. Any takers?
    Thanks.

    Gaff wrote:
    No. With non-DB code, it isn't voluntary. The developer needs to check a file in or out before modifying it and the CM system enforces either that only one person at a time can edit the file or if multiple ones can, the merging of the results. Once the PL/SQL package is in a database, a developer can change it and NOT change the underlying file. I don't agree that there is a difference. All source control systems I have seen allow a copy to be taken or the source to be viewed without checking it out. I have a checked out source file on my PC. I copy it. I check in my changes, I make changes to the copy. I have made a change without changing the checked in source. Just like I could in the database. Should this change make it into the final build? No, and I can't complain when it doesn't, just like the unauthorized change directly in the database is overwritten and lost. I see no difference.
    Making changes to database objects directly in the database is a terrible practice, you can enforce the same lack of tolerance for it as you would for any other developer circumventing the source control and deployment process in any other language.
    - You tell them not to do it.
    - You overwrite their changes if they do do it.
    - You fire them if they continue to do it.
    >
    I think the solution is that by default DB developers have read-only access to the objects. When they identify the pieces that need to change to fix a given bug, some 3rd party (Software Lead? DBA?) grants them CRUD on the identified objects and verifies that their modified code makes it into the CM tree when the bug is fixed.The developer can make their own changes in the Dev database. Changes to the Test database prior to production deployment can only be made by the DBA using scripts from the source control system.
    This works flawlessly.

  • Azure Account Management Best Practices

    Hello,
    Hoping someone will be able to help me or point me in the right direction. First, a general overview of what I'm trying to do:
    I am trying to utilize the IaaS portion of Azure to setup my application on a Windows VM.  Each customer will have their own individual virtual machine configured specifically for them, running my application that they can remote into and use.
    My thought was that I would create a separate Subscription in Azure for each customer. That way, I could manage the billing information more easily. Within each Subscription would be all of the Azure resources for that customer: VMs, Storage Accounts, Backup
    Vaults, Virtual Networks, etc. Since there is no limit to the number of subscriptions per account, this seems like it would also scale well.
    My question is...is this the best way to do this? Since these customers will really have nothing in common besides using my application, I think creating separate Subscriptions with separate Virtual Networks would be the best way to keep them completely
    isolated from one another. Does anyone see any problem with this approach or know or a better way to do this? Is there a guide or blog post somewhere that details a similar scenario and how to approach it?
    Thanks in advance

    Hi,
    Yes, you are right.
    Best practise would be to create new accounts for your employees usage.
    But there are many other ways to do this apart from the one you says.
    As the services grow the best way to track is by creating new accounts and deleting the ones when his employees leave the organization.
    Please have a look into these articles:https://technet.microsoft.com/en-us/magazine/gg607453.aspx
    https://msdn.microsoft.com/en-in/library/azure/hh531793.aspx
    Regards,
    Azam khan

Maybe you are looking for

  • How can I use aperture after receiving the full adobe suite? It won't allow me to use it.

    How can I use aperture after receiving the full adobe suite? It won't allow me to use it. I receive a dialogue box which states previews are not allowed in  aperture and the application is shut down.

  • STPO's --- load date is not appearing in the item detail

    Hi SAP Expert's, We have a issue with inter company STPO, the Qty of mtrl is80 Kg. we received 10 kg on 1st feb. In item detail --Delivery schedule tab it showing the delivery date is 1st feb and goods issue is 5 th feb. Same like they delivery 10 kg

  • How do you stop the printer printing a huge unwanted pile of pictures?

    We've all done it - you have selected a full folder of 200 or 300 pictures to view and have accidentally sent them all to the (very expensive to run) bubble jet printer.  The printer start spewing out prints.... how on earth do you stop it immediatel

  • Group by caluse before where clause

    Hi all, Declare cursor c1 is select ARc_Month as mnth,count(*) as nof from dbarc group by ARc_Month where ARc_Month like '%02%'; e1 varchar2(1000); cnt number(10); begin e1:='1'; for crec in c1 loop cnt:= c1%rowcount; dbms_output.put_line(cnt||'-----

  • Database error found in SM21 logs

    Hi, we are facing problem with too many system logs as : Database error: TemSe->RTAB-S/G(8)->1 for table TCPSBUILD key. Found no information on google. If anyone might have come accross this problem or know the solution plz reply.. 00:17:39 DIA 000