VERSION CONTROL ACTIVATION

Sir,
how to activate version control as well as how many changes are allowed once the version is activated
regards
amey

hi,
It is activated here:
SPRO >> MM >> Puchasing >> Version management >> Set up Version management for external Puchasing docs..
You can also define the Reason of change here:
SPRO >> MM >> Puchasing >> Version management >> Dfeine reason of change
SPRO >> MM >> Puchasing >> Version management >> Set up change displays
For PO's you don't define the version relevent fields...
If you wanna print out, then you have to do additional settings for printout...
Regards
Priyanka.P

Similar Messages

  • Regarding version control

    Hi,
    Could u please help me out in giving some ideas of version control in sap?
    First let me give some example as follows:
    First if i develop something in the developement server then later when i transfer to the QA server and later to production server then is there any change in version.
    Please give me a details of this issue????
    Thanks,
    Batista....

    hi priya,
    Version Control
    Version control is a mechanism that helps maintaining the revision history of a development resource and tracking the changes done to it. It defines a set of constraints on how a development resource can be changed. A development resource that complies with the constraints defined by the version control is called a versioned resource. When a versioned resource is modified or deleted, a new version is created for the resource. A unique sequence number is associated with each version of the resource created in a particular workspace. This sequence number identifies the order in which the versions were created in that workspace. The DTR graphically represents the relationship between the different versions of a versioned resource in the form of a version graph.
    For the representation of version graphs, this document follows the conventions shown in this figure.
    The figure shows the meaning of the symbols in the version graph.
    The following changes are tracked by the version control mechanism of the DTR:
    ·        Addition of the resource to the repository
    ·        Modification of the resource in the repository
    ·        Deletion of the resource from the repository
    In all the above cases, a new version of the resource is created.
    Production Delivery
    Packaging
    To deliver your product, you have first to package it. There are different packages you can use for shipping your product to your customers:
    ●      Software Component Archives (SCAs) – this is the standard way to deliver software for the SAP NetWeaver platform.
    ●      Software Deployment Archives (SDAs) – for top-level applications you can deliver only the executable part of the software. You can directly deploy the SDA file.
    ●      Public Part Archives (PPAs) in Development Component Interface Archives (DCIAs) – for reusable components (Java EE server libraries, Web Dynpro components, Visual Composer components and so on). You can deliver only the metadata of the components. DCIA can be included in SCA file too.
    How to do that?
    Using the command line tool provided with the SAP NetWeaver Composition Environment you can:
    ●      package a collection of components into an SCA including only the deployable archives. This is required if you do not want others to reuse the delivered components.
    ●      package a collection of components into an SCA including the deployable archives and the corresponding interface archives. This allows customers to develop against these components. Those customers can directly import the SCA into their own SAP NetWeaver Development Infrastructure (NWDI) or into an SAP NetWeaver Developer Studio local installation.
    ●      package the public parts of a component together with the required metadata into a DCIA (and further into an SCA).
    ●      include source code into an SCA.
    ●      unpack a deliverable archive and drop the result into an existing version control system for example, or directly import them into an existing Design Time Repository (DTR).
    Delivery of Source Code for Further Customization
    In addition, you can deliver source code to your customers to allow further customizing or add-on development. The deliverable archive may contain sources for:
    ●      individual development components (DCs).
    ●      a collection of development components, for example a whole software component (SC).
    Example
    A customer can add a new source compartment to an existing configuration, and then locate that compartment in the file system where it is accessible by the version control system in charge. Then he or she extracts the sources with the command line tool to the compartments root directory and refreshes the configuration in the SAP NetWeaver Developer Studio. The compartment tree is populated with components from the archive. Afterwards, the customer may put those components under version control. Deliverables that contain only individual components may be treated accordingly.
    This mechanism may also be used for other purposes, for example for setting up a simple backup and restore mechanism for components in Developer Studio, or sharing DC sources without having a central version control system: a developer may pack a compartment and store the resulting SCA on a central share or backup system. Another developer may take that SCA and import it.
    Limitations
    Note the following limitations connected with this kind of source code delivery:
    ●      There is no support for handling conflicts when different actors in a delivery chain develop independently in the same source code. You cannot prevent the customer from modifying delivered sources. When you ship a new version of the sources, there is no special support for updating and no support for merging the update with modifications done by the customer. You and the customer have to agree on a process how those conflicts are handled. For example, the customer can decide not to import the update you deliver directly into the active development line, but to unpack the delivered sources to some unconnected sandbox system and perform the required merges manually.
    ●      When you deliver source code to customers, it is important that you also deliver the required libraries and generators that are needed to build these sources. For example, it may be necessary to ship some archive compartments that contain used components.
    ●      There is no support for delivering deletions in a new version. If a source file was deleted, the customer has to manually ensure that the file is also deleted in the Developer Studio or source code management system.
    ●      If a customer prefers to work with the SAP NetWeaver Development Infrastructure (NWDI), this customer cannot directly import the source delivery package into the NWDI landscape. Between NWDI landscapes at different places, sources usually are exchanged through a more sophisticated export format that contains not only the pure source code, but also the versioning meta information of the exporting DTRs. This ensures that the importing repository can detect conflicts that arise due to modifications. If this versioning information is not available, the only way to import source deliveries is to unpack them to a file system and manually put them under version control with the Design Time Repository perspective of the Developer Studio. In case of an update, the customer would have to check out all affected files, merge them with the new versions from the source delivery, and finally check them in as a new version.
    More information: Composition Environment Command Line Tool
    see this url
    http://www8.sap.com/businessmaps/0134713B1D6046C59DE21DD54E908318.htm
    thanks
    karthik
    reward me if usefull

  • Version Control for adobe forms

    Dear All,
    We have a situation where we already go live with version 0 of our adobe form. Now, we have an enhancement to the form i.e.: some new functionalities and new field. Previously we have tried to transport the new changes to the production without any version control and those processes that have been already started prior to the new changes encounter error when the user continues the process.
    I am wondering how do the version control works and any documentation on how to configure it?
    Thanks in advance
    Regards,
    Bryan

    Hi Brian,
    Here is some information that I found in the IMG on Create ISR Scenario...
    Create ISR Scenario
    Use
    In this IMG activity, you create an ISR scenario that has a one-to-one relationship with a form scenario. To be able to use a form scenario in a process, one ISR scenario must exist for each form scenario. The ISR scenario and the form scenario must be linked with each other. You make this setting in Customizing for HR Administrative Services in the IMG activity Link ISR Scenario with Form Scenario.
    In the form scenario, you define primarily the basic set of form fields and their processing through the backend services. In the ISR scenario, you specify the definition of the user interface. You also specify which form is used for the display and how the layout of this interactive form is designed. You use interactive forms based on Adobe software to create and process the forms.
    ISR scenarios and form scenarios are version dependent. The version numbers of the ISR scenario are assigned automatically. Note that a form scenario must have exactly the same version as the linked ISR scenario. For this reason, you should always create a new version in the ISR scenario first and then use the same version number when you create a version in the form scenario manually.
    Note
    If an ISR scenario or form scenario (with an existing version) has already been used in a productive process, you should not change the configuration. If you want to make changes to a process or an ISR scenario or form scenario, you should always create a new version, and only ever use that new version in the future. In this way, processes that have been started can be concluded with the old version and new processes can be started simultaneously with the new version.
    This is from the Create Form Scenario documentation...
    Create version
    Form scenarios are version dependent, which means that there is at least one version of each form scenario. Versions are linked with processes. Since processes can vary, you must also be able to adjust the associated scenarios. To be able to provide different forms for process variants, you create versions.
    You can still process and change an existing version at a later point in time. Once a version has been used to execute a process, you should not make any more changes to this version; instead, you should create a new version.
    The form scenario and the (linked) ISR scenario are both version dependent. They must always have exactly the same version numbers. Note that the version number of the ISR scenario is generated and cannot be entered manually. When you create a new version in the form scenario, you therefore have to use the version number generated in the ISR scenario.
    If you have already made extensive Customizing settings for the form scenario and want to create a new version based on the settings, you should use the IMG activity Manage Form Scenario.
    Hope this helps...
    Cheers,
    Kevin

  • OWB Process Flow - How is the best  version control tool ??

    HI all,
    I just start work with OWB and I have a question to know how is the best way to do something.
    Imagine the scenario below:
    If I have 2 or more requests for example:
    Request 1: Create a Dimension City.
    Request 2: Create a Dimension Products.
    I Have ONE process flow and i need put my changes inside. This is my problem.
    In my scenario I don't know what Request goes to Prod First.
    If I put the Request 1 and Request 2 in my PROCESS FLOW, maybe I need change is someone decide change MY REQUEST PRIORITY.
    There is something in OWB to "control the version or changes" ?? For a mapping I export the MDL and commit on SVN, but I dont know haw can i do with the process flow.
    Something to agree multiples peoples work in different mappings and a SAME PROCESS FLOW ??
    What is best way to work with process flow and version control.
    What are the best practices when it comes to version control?
    Thanks.

    Amit,
    Are you really doing this in 10.1.3.x and not 11g?
    At any rate, I don't see how #2 and #3 relate whatsoever to your choice of a version control system. OK, maybe in #2 if there is some "maintenance" activity to be done against the version control server. Subversion is the open source alternative that you listed there and is pretty commonly used. If your company is already using one of the mentioned tools, why change? About the only thing I'd mention is to advise you NOT to use CVS for well documented reasons (JDev does support it) - if you would have picked CVS otherwise, choose Subversion. As far as question #1 - I've only used Subversion (well, I did use CVS for a while) with JDeveloper, so I can say it was "effective enough for me." In 10.1.3.x, I also used the external svn tools for doing lots of things like merging and so forth; in 11g, the support is much much better.
    Best,
    John

  • OpenScript: Which files to put under Version Control?

    Hi,
    According to JB's description of [how to enable Subversion|https://forums.oracle.com/forums/thread.jspa?messageID=10677356] for OpenScript, I have now an active version control.
    The next question is now: which files I should put under version control? Obviously, the script.java is a good decision, but there are many more, such as java, properties, but also specific folders like "recorded data" and "results" folder.
    Which once I should put under version control, any recommendation?
    Thanks a lot in advance!
    Volker

    Thanks for your replies,
    it's nice to know that I was going the right way.. :)
    But how do you handle these little bugs, that I mentioned in my first message? Do you check-out all files in the project before work? Or they are TeamSystem - specific?
    Best wishes, Serg.

  • Version Control of PL/SQL programs in Database

    Is there a version control feature available in Oracle v9.x ???
    I am trying to implement version control on PL/SQL programs(Packages/Functions/Procedures). I should be able to rollback to old version and keep the system running if the latest ones failed. This should be done automatically without bringing the database down or recompiling the PL/SQL programs and also users do not need to reconnect(users might be caught in a LOCK and might get their sessions killed by DBA...)
    Ex: I have heard that in .NET, u can have more than one version of a DLL and have only one version active. If we want to go back to the old version of the DLL, u don't need to recompile the DLL to make it active. I am looking something similar to this....
    I have thought of several ways like creating a small repository table for my PL/SQL programs and store the PL/SQL code in the repository and based on the situation, compile only that program (which might be a OLD or a NEW version) and so forth... But this does not satisfy the requirement of rolling back to the old version without recompiling.(RENAMING a PL/SQL program feature doesn't seem to be implemented yet...)
    I don't want to use Designer just for this purpose...
    Any ideas..
    Thanks,
    Purush

    Are you dealing with code that's being called remotely (i.e. via a database link)?
    No
    I'd be concerned about the concept of rolling back code changes in production on a regular basis-- that would seem to indicate problems that ought to be addressed in development and QA. Rolling back code in production seems like it ought to be a rather painful process, if only because it indicates a massive failure elsewhere.
    This is not on a regular basis at all. Our Database applications are tied with lots of programs which directly control the robots and machines. Certain machines and robots needs to be working all the time and any downtime will cost time and money. To make sure that the implementation goes into production smoothly(without shutting down machines and robots) and then into maintainance mode, we are looking for some kind of source control to control the implementation and make sure to revert back (without shutting down machines and robots) if there are major issues.(There are certain things here which cannot be tested outside of a shop floor due to the physical and other constraints.)
    I have thought of otherways like a compile flag (in the DB.. Ex: a packaged variable) to set before compiling and reset after compiling. The programs on the shopfloor will always read this flag and check buffer(time taken to do this will have to be considered) before calling a DB txns and if the flag is set, buffer the txns and and move on to the next task the machine should do. The next time it call a txn. if the flag is reset, it checks the buffer and if buffer exists, execute the buffer txns first.. and then proceed to actual txn. The things that bothers me is time taken to compile the huge package and the no. of txns getting buffered and the overall txn time.
    I am trying to come up with some kind of solution for this issue if possible.....
    Thanks,
    Purush

  • Version control and deployment strategies

    Hi,
    I was looking for input from the community on general
    strategies for using version control and managing deployments to
    test/stg/production.
    Currently, I am using subversion to track my source code, and
    using the standard flex builder build routine to produce my binary
    output and test. My output is stored locally on a shared vmware
    drive, so that it can be served up with a linux vm running apache
    (this is not dissimilar to just local testing)
    Now I'm getting ready to deploy to some remote testing server
    though, so I'm trying to think of the best way to go about it. I
    would like to tag my code in svn with a release tag, as is my
    practice in other platforms. Should I also store the bin folder in
    svn? Should I check in the resulting binary code independently in a
    separate repository/directory and then tag it there? Should I
    create a new build target to deploy directly to my testing server?
    The issue with the tagging approach seems to be that if I
    want to rebuild the code or redeploy it for any reason, I would
    have to checkout the tagged code in a separate directory, import it
    as a new project, rebuild and then redeploy.
    If I checked in the tested binaries into a separate
    repository/folder, I could always just do an svn export for
    deployment, but I'm not sure if that would cause any weird issues,
    and it seems a bit wasteful. I suppose I could build from the tag
    and zip up the resulting release and just make it available via
    normal download, but it seems that I would likely then have lots of
    working directory checkouts as flex builder projects for each tag
    or release, just so that I could rebuild from them....doesn't seem
    very elegant.
    I'm very interested in hearing any feedback on this. How do
    you do it?
    thanks,
    Cliff

    flex with ant is still not very very popular
    combination,since FB does all for you anyways, but I have to say
    ant is lot more flexible, especialy if you combine it with FB, ant
    can do pretty much anything, ...
    here is link about flex's ant tasks
    http://livedocs.adobe.com/flex/3/html/help.html?content=anttasks_1.html
    most of the projects I have done in Flex were with ant,
    here is general approach:
    for internal testing I let flex builder build and deploy
    within integrated tomcat,this is also where I do debugging, FB is
    pretty good about that.
    then I have following targets :
    build target - builds optimized version of flex app, only
    using library classes that are needed by project, also using that
    to feed my Modules building tasks so that they exclude all class
    references (like Button, Tabnavigation etc... ) from their compiled
    units,
    something like this (this is my old fb2 example) i don't have
    fb3 example handy right now :
    quote:
    <mxmlc file="${basedir}/main.mxml"
    debug="false"
    optimize="true"
    output="${dir.web}/main.swf"
    show-binding-warnings="false"
    show-actionscript-warnings="false"
    link-report="${basedir}/docs/my-links.xml"
    use-network="true">
    <load-config
    filename="/configdir/modified/flex-config.xml"/>
    <source-path path-element="${FLEX_HOME}/frameworks"/>
    <compiler.library-path dir="." append="true">
    <include name="lib" />
    </compiler.library-path>
    <compiler.source-path path-element=""/>
    <compiler.source-path path-element="src"/>
    <metadata description="some app">
    <contributor name="John Doe" />
    <contributor name="Apple Orangino" />
    </metadata>
    </mxmlc>
    <!-- compile module mymodule-->
    <mxmlc file="${basedir}/mymodule.mxml"
    debug="false"
    optimize="true"
    output="${dir.web}/mymodule.swf"
    show-binding-warnings="false"
    show-actionscript-warnings="false"
    load-externs="${basedir}/docs/my-links.xml"
    use-network="true" >
    <load-config
    filename="/configdir/modified/flex-config.xml"/>
    <source-path path-element="${FLEX_HOME}/frameworks"/>
    <compiler.source-path path-element=""/>
    <compiler.source-path path-element="src"/>
    <metadata description="some app">
    <contributor name="John Doe" />
    <contributor name="Apple Orangino" />
    </metadata>
    </mxmlc>
    tag -target tags release based on parameter of latest tag
    plus number increment I configure in properties file.
    Utility targets :
    classpath target - builds classpath string for compc task.
    commit target : commits source code, before building.
    resources target - copies all resources files to build
    directory,
    deploy-local target- deploy to local Integration server
    deploy-remote target deploy to remote uat server.
    test -target - run test cases over classes and generate
    report.
    and of course all famous asDoc target :)
    good thing is that you can create "ant builder" under project
    properties and chain your targets with flexbuilder's build
    commands,
    you can also easily integrate with build servers ( I use
    hudson)
    here is example :
    http://hudson.amostudio.com/
    MR hudson checks out code for you, builds it using ant
    targets you tell it to use, and reports to you, its pretty cool and
    very handy to always have active build proccess over codebase, of
    course in some cases its overkill, but most of the times, MR hudson
    is good to have.
    unfortunately all my ant files are for external clients I
    cant disclose them, but I can write blog about some general (apples
    and oranges) example, hhm that's actually good idea :) I can shake
    off some stress as well :)
    thanks for the idea :)
    hth
    regards
    levan

  • VERSION CONTROL FOR PO

    Sir,
    I have activated version control as well as fields relevant for po but it is not reflecting
    ply help me how can i reconfigure it
    regards
    amey

    Hello Amey,
    Pl. activate version control for PO in SPRO. Steps are as follows
    SPRO/MM/Purchasing/Version Management/Set Up Version Management for External Purchasing Documents then Add new entry for the Purchase order document type and respective Purch. organisation and put a tick in the text box.
    This would solve your problem.
    With regards.
    Sanjay

  • Raptor and Version Control

    Here are some discussion points for use of Raptor with version control.
    Assumptions/Terminology.
    Roles:
    Developer, Tester, Production DBA.
    Each developer has their own database or schema to work with.
    Database code is PL/SQL and DDL
    Client code is everything else including sql scripts.
    VC stands for generic version control. I may slip into CVS speak, but ideally this is product neutral.
    Scenario 1: New developer on existing project.
    1 Check out sources
    Obviously need to define a VC repository connection in Raptor
    Probably also need to define filesystem directory to check out to. Conceivably raptor could do this under the covers but if the source code is a mixture of database code and client code this isn't possible.
    2 Run sources against database to create schema objects
    This would probably be a manual job within raptor but a nice feature would be a tool to work out the dependencies and run everything in the right order.
    3 Edit-save-compile-edit objects in the database
    At this stage the database code is out of sync with the filesystem and the VC. Since the definitive version is in VC the filesystem is irrelevant so it would make sense to save the object back to the filesystem when you click the save/compile button.
    If VC isn't active this is obviously a bad idea since the filesystem version is the backup.
    4 Create new object in database.
    The new object needs to be immediately saved to the filesystem and added to VC(but not committed). The create dialog should ask for a filename as well as object name.
    5 Commit changes to VC
    What gets committed? All the database code in the schema of the current connection? All the database code which is known to CVS (which should include new ones created at step 4)? What about client code? Is there a need for a project to which objects can be added manually or automatically.
    There may be clashes between this developer's changes and anothers.
    Need to be able to compare current database with current cvs and merge in changes before commiting. The merged changes need to be immediately applied to the database.
    6 Whoops forgot something, edit object on filesystem but don't check in.
    7 Next day, try to edit same object in the database.
    We now have a 3-way discrepancy between vc, filesystem and database. Raptor needs to allow diff/merge between database and filesystem as well as between database and VC.
    Scenario 2: New Project
    1 Developer creates some database code
    edit-save-compile is entirely within the database at this stage and no VC is involved. At some point the developer has to define a vc repository and directory. Raptor needs to be able to create a new repository or module (cvs import for example) either at this stage or the first time code is saved.
    There could be an option to make VC compulsory so that whenever a developer creates an object, the relevant VC actions are taken.
    2 This scenario merges into scenario 1.
    Scenario 3: Tester begins new testing cycle
    Scenario 4: DBA implements new code in production database.
    1     User checks out latest code from CVS
    2     User runs code against database to create schema objects.
    This is very straight forward. I am not even sure that this should involve raptor, it should all be done by scripts in sqlplus. The one raptor involvement could be reconcilation of the objects in the test database with the VC repository.
    A few other thoughts.
    I haven't thought about multi-schema development, ie making changes to more than one schema at a time. Should a VC repository be associated with a particular connection?
    I haven't thought about managing 'create table' vs 'alter table'. A tool which generated an alter table from the different between db and vc or between 2 vc revisions would be useful.
    All checkout/export/diff actions should be against labels/tags/branches as well as the tip.
    Filenames and formats.
    There should be preferences to allow the user to specify file extensions and other naming conventions (eg upper/lower case). I use the following for ddl files
    table_name.tbl -- create table
    table_name.con -- constraints for table
    table_name.idx -- indexes for table
    view_name.vue
    sequence_name.seq
    Table DDL
    Currently table ddl as generated from raptor includes storage details. Not all of this is relevant in the development process and could cause spurious differences in version control.
    I also like my constrains in a separate file from the table ddl.
    PL/SQL Templates
    The ability to define basic package and procedure templates with VC keywords etc in place. New | Procedure etc would use these templates
    Message was edited by:
    smitjb

    Thanks for the detailed use cases. We can use this message to track Version control feature requests.
    We do want to provide this support in future releases.
    Donal

  • SD quotation managed through version control

    Hello all,
    How can I active the version control for a SD quotation? What I need is, check the changes made in a SD quotation using the version control.
    Thnks in advanced.
    RB

    In our current J2EE environment an ant script builds a war and copies it to a staging file server, then there is a java app that deploys the war to weblogic with the weblogic scripting extensions. We are thinking of modifying this app to also deploy the PTE files that any portlets in the war depend on. Day one a jsp portlet requires a single session preference. The next day it needs a login token to use the PRC. How do you coordinate this? For us it will probably be done with custom scripting.
    A developer modifies the jsp and checks that in along with a revised remote service pte file which has the new settings. The ant script will then build the war and package it into a tar with the pte file and finally drop it on the staging server. From there our existing java deployment web app can extract the tar, deploy the war and run a migration import on ALUI. Hopefully all this can happen in a transactional context so that that if the import fails, the new portlet is not deployed.
    I don't want to run into the case where a new portlet is out on the QA box and the portlet service is misconfigured and isn't supplying the correct headers to the portlet tier.

  • Configuration version control software

    I know some people like to put their /etc under version control with CVS or hg or something, to make it easier to recover if they misconfigure something. I think this is kind of overkill, though. For one, you don't really need "distributed" or "centralized" version control, you just need to control the version. And for another, version-controlling the entire /etc doesn't really make sense considering that you're usually only working with a few files at a time, which you would generally control separately. So, I came up with the idea for a VCS specific to managing your system configuration. I'm going with the name Tessen for now since it sounds cool.
    The two basic concepts involved here are applications and snapshots. An "application" refers to a particular set of config files managed by Tessen. For example, the pacman application would own pacman.conf, makepkg.conf, and pacman.d (a directory). Ownership in this context means that it will automatically track changes to the contents of the files and folders. (Each application has an active snapshot, which is basically "the last one you saved or reverted to.")
    A "snapshot" is how an application's config files looked at any given time. A snapshot has a number (since it's not distributed, a simple autoincrementing integer will work), a timestamp (for obvious reasons), an action (to indicate exactly what caused this snapshot to be taken - for the user manually saving it would be "save", if it was installed by a package manager, "install", etc.), a note (if you feel like it - a good sysadmin would always have one), and the files' contents. (You can also add additional notes to snapshots after the fact, but this won't affect the note that was saved with the snapshot in the first place.)
    Anyway, the workflow would run something like this:
    "tessen newapp pacman pacman.conf makepkg.conf pacman.d". Tessen creates a new application named "pacman" that owns pacman.conf, makepkg.conf, and the directory pacman.d. (Tessen would automatically save a snapshot at this point consisting of the current contents of said files. That would be snapshot #1.)
    Make changes to the configuration, like using aria2c as the transfer command.
    "tessen save pacman -n 'set XferCmd to aria2c'". Tessen will take a snapshot of all the files owned by pacman and save that, making it #2.
    DOOM! You forgot that you don't have aria2 actually installed, and now that XferCmd is aria2c you can't install it! (This is just a contrived example.)
    "tessen revert pacman". Since you're at #2 but haven't made any changes, it will switch you back to #1. (If you had made changes, it would reset to #2 instead.) All necessary files will be reverted.
    Now you can "pacman -S aria2" and make sure it works this time.
    You could edit it manually to set it back to aria2 (which if you saved it would make it #3) but it's easier just to "tessen revert pacman 2". revert jumps back one step by default, but you can tell it to revert to anything. Then the next change you make would be #3.
    One major thing about the design I haven't quite decided would be how ownership is stored, and what to do if you change the ownership and try to revert to an older snapshot. That would be tricky to figure out. Still, what do you think about this approach?

    ewaller wrote:I think I'm missing something :-/
    I know the OP dismissed git and hg at the outset -- but why?  It seems to me they are ideal for just this sort of thing.  On this machine, I do have an Hg archive living under /etc.  Being Hg, it is simple to push that off to other systems and keep them synced (or not).
    Perhaps tessen could wrap git or Hg?
    The reasons I had for not using a traditional DVCS (as much as I love them) include:
    A VCS in all of /etc would manage your configuration as a whole. Tessen lets you manage each application separately without awkwardness.
    Also, there is stuff in /etc that isn't really configuration (I'm looking at you, /etc/bash_completion!). You could get around this by simply not "hg add"ing those files, but it does make it trickier. (Though in an ideal world, each app would store its configuration in a separate directory and things like the completion scripts would be in /share or /usr/share.)
    Finally, if you're using a DVCS so you can sync configuration, this would actually be better. For example, if you're trying to share the Apache and CouchDB configurations between your boxen, you have no way to keep out any other files you might have added (like, say, fstab).
    Really, it's mostly just that the concepts behind Tessen map more clearly to the problem than those in hg or git. At least in my mind. Although using a DVCS as the storage backend would be interesting...
    fukawi2 wrote:What language are you planning on writing this in?
    Probably Python. (My original concept would use SQLite as the database.)

  • How do I fix project after "Remove From Version Control" corrupted it?

    I am using RoboHelp 9.0.1 and installed both Tortoise SVN 1.6.9 and latest PushOK SVNSCC then added my large RoboHelp project to SVN. I was able to check in and out files from SVN but had several issues with it:
    1) Super super slow. Working with folders or any renames would take 10 seconds per file and up to 1 hour if needed to refresh the root folder.
    2) I could not perform some actions at all, such as delete, rename, or move folders. I kept getting COM errors.
    I therefore decided that working with SVN and RoboHelp is not practical, at least not on my VPN so I decided to disconnect the project from source control and just work locally. The only option that I saw that sounded like it would do that was the "Remove from Version Control". This started a process that lasted for several hours. At the end of it, I now have several significant issues:
    1) The order of the files and folders in my Project Manager is completely wrong now. I have almost 1000 topics and reordering all of them is not possible.
    2) The Table of Contents, Glossary, and Index files appear empty. They had content before.
    2) A couple of the Single Source Layouts I had created are completely missing.
    3) Many, but not all, of the folders have tons of files with the extension ending in "_temp_removed_by_svn"
    4) Many, but not all, of the files are actually gone from SVN so I can't recover a clean image. There was no warning that this command would actually delete the files from SVN (I thought it would just remove the version control connection).
    5) Who know what other issue exist that I haven't seen.
    Any idea how I can fix this?
    Thanks in advance,
    Dan

    Are the "_temp_removed_by_svn" files in your local folder or SVN? Let us know how you get on with the new project. It sounds like something is wrong with SVN. Can you use the SVN Log command to see whether there is a different version you can restore. This might also give you an indication of what might have caused the problem. You could try deleting your CPD file. It gets rebuilt it is isn't there anyway. This file can become bloated and it is good practice to delete it when it gets close to 2mb in size. Your project is fairly large and has a lot of folders and may affect performance. Have you considered splitting them and merging the output? I know you probably don't want to consider this right now, but I think it may be a better long term solution.
      The RoboColum(n)
      @robocolumn
      Colum McAndrew

  • Can I have a library of PDF docs with version control? Can it cope with version nos. embedded in the file name?

    I manage a set of documents which are edited in Word but "published" as PDFs (using Word 2010's save to PDF capability).
    I want to create a library for them on SharePoint (my company has SharePoint Online via its Office 365 subscription).
    I'm pretty much a SharePoint novice but even I can see it's easy to upload the documents to a simple library. The things that are giving me a headache are:
    Can I tell SharePoint what the version number of the uploaded PDF document is? The version number as understood by the library needs to match the version number written into the document (where it is called a revision number and increments in whole numbers
    starting from zero).
    How do I handle replacing the uploaded PDF documents with new versions? If they were Word documents I could edit them by opening them from the SharePoint library, checking them out if necessary, and SharePoint would handle version control.
    But since the PDFs are generated from editable masters (Word documents) which are NOT on SharePoint I would need to edit the local Word document on my PC then generate a PDF version then upload it to replace the existing PDF document in the
    library. Is it easy to upload a new document over the top of an old in a SharePoint library?
    Hoping someone can give me some answers.
    Regards,
    Bruce Officer

    hi Bruce,
    1. It sounds like what you need is to set the starting version number since your revision number increments in whole numbers, it would match up to SharePoint once the starting version number is set. You can potentially create a new custom field in the
    library to manually track the version of the uploaded PDF document, but this might not match up with SharePoint's own version number and could get confusing. Another possibility is to upload dummy versions of the PDF document until the SP version
    matches with the revision version and then delete these dummy versions.
    2. When you upload the PDF document again into the library, it should prompt you to see if you want to replace the existing. If you proceed with the upload, it should replace and increment the SharePoint version number.
    Please Mark Answered if my reply solves your problem. Thanks!
    Jeff Thai
    Technical Solutions Architect, AvePoint
    http://www.AvePoint.com

  • Oracle Service Bus Configurations version control and deployment automation

    Hi,
    Currently we have OSB10gR3 installed and we use the web based sbconsole to create projects and proxy services. It's all working well and good!!
    We are at the state where we need to think about source control, migration of artifacts from dev to test and to prod.
    I'm looking for pointers to version control the artifacts of OSB projects, what could be version controlled (no binaries) and how do we extract those artifacts?
    How do we customize those artifacts while migrating to different environments in an automated fashion?
    Please point me to best practices and gotchas that we should be aware of while dealing with deploying OSB proejcts from test to Prod.
    Thanks in Advance!!

    After reading the threads mentioned by Deba, I'm able to get this all worked out with SBConsole itself. Experts, please review my approach below and let me know if I have overlooked anything.
    Simple advantage I see in using sbconsole is that it requires less maintenance, i.e. avoids rolling out another IDE (Eclipse -Workshop plug-in) to IT developers and at the same time provides the functionality that we are looking for. Currently, JDeveloper is our primary IDE, so we thought it's best to wait till OSB development gets integrated into JDev.
    This is the deployment workflow which worked for us,
    Developer:
    1)Develops Proxy services using SBconsole in Dev environment.
    2)Creates sbconfig.jar by using the export functionality available in System Administration link in sbconsole
    3)Checks-in ALL the files present in the above jar into version control under the proxy service project name
    4) Creates customization file using the customization file link in System Administration and modifies the value for each environment i.e. creates two files test_customfille, prod_customfile
    5)Checks-in the customization files into version control under the same proxy service project
    Promotion to Test and Prod
    1) From the source control, service proxy is built (actually the jar file of all files including the custom file is created)
    2) SCP proxy_sbconfig.jar file to Test or Prod box
    3) Follow the steps mentioned in Auto deploy of ALSB/OSB artifacts - Proxy, WSDL and webservices...
    4) Depending on the server, test or prod, pick the right customization file and deploy using ANT.
    Thanks!!

  • Oracle 10g R2 - Metadata Issue With Version Controlled Resources

    We are trying to utilize the Oracle 10g R2 feature that allows user-defined metadata to be attached to XML DB resources. However, we are encountering errors when we try to append/attach metadata to a version-controlled resource (VCR). The following steps were followed:
    1. Register a new schema to be used for metadata (XMLSCHEMA.registerSchema).
    2. Add a new resource to to the XML DB and make it version controlled (DBMS_XDB.createResource and DBMS_XDB_VERSION.makeVersioned).
    3. Checkout the newly created version-controlled resource (DBMS_XDB_VERSION.checkout)
    4. Add metadata to the version-controlled resource (DBMS_XDB.appendResourceMetadata)
    However, after step 4, the errors shown below are encountered:
    ========================================================
    begin
    ERROR at line 1:
    ORA-00600: internal error code, arguments: [qmxStrCopy: INLOB 2], [], [], [],
    ORA-06512: at "XDB.DBMS_XDB", line 890
    ORA-06512: at line 2
    ========================================================
    The steps followed for adding metadata to a resource were taken from Chapter 26 of the "Oracle XML DB Developer's Guide 10G Release 2" (http://download-east.oracle.com/docs/cd/B19306_01/appdev.102/b14259/xdb_repos_meta.htm#sthref2302).
    Is is possible to append and maintain metadata to an XML DB version controlled resource (VCR)? If so, what are the correct steps to do this? Note that if this is tried with a resource that is not version controlled, then no errors are encountered.
    Thanks,
    Steve

    It appears that this issue has been solved in the next release of the database
    QL> spool testcase.log
    SQL> --
    SQL> connect &1/&2
    Connected.
    SQL> --
    SQL> var schemaURL varchar2(256)
    SQL> var schemaPath varchar2(256)
    SQL> var resourcePath varchar2(256)
    SQL> --
    SQL> begin
      2    :schemaURL := 'metadata.xsd';
      3    :schemaPath := '/public/metadata.xsd';
      4    :resourcePath := '/public/metadataTest.txt';
      5  end;
      6  /
    PL/SQL procedure successfully completed.
    SQL> call dbms_xmlSchema.deleteSchema(:schemaURL,4)
      2  /
    Call completed.
    SQL> declare
      2    res boolean;
      3    xmlSchema xmlType := xmlType(
      4  '<?xml version="1.0" encoding="UTF-8"?>
      5  <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns:xdb="http://xmlns.oracle.com/xdb" elementFormDefault="qualified" attributeFormDefault="unqualified" xdb:storeVarrayAsTable="true">
      6          <xs:element name="root" type="rootType" xdb:defaultTable="METADATA_TABLE"/>
      7          <xs:complexType name="rootType" xdb:SQLType="METADATA_ROOT_T">
      8                  <xs:sequence>
      9                          <xs:element name="child" type="childType"/>
    10                  </xs:sequence>
    11          </xs:complexType>
    12          <xs:complexType name="childType" xdb:SQLType="METADATA_CHILD_T">
    13                                  <xs:sequence>
    14                                          <xs:element name="Child1" type="xs:string"/>
    15                                          <xs:element name="Child2" type="xs:string"/>
    16                                  </xs:sequence>
    17          </xs:complexType>
    18  </xs:schema>
    19  ');
    20  begin
    21    if (dbms_xdb.existsResource(:schemaPath)) then
    22      dbms_xdb.deleteResource(:schemaPath);
    23    end if;
    24    res := dbms_xdb.createResource(:schemaPath,xmlSchema);
    25  end;
    26  /
    PL/SQL procedure successfully completed.
    SQL> begin
      2    dbms_xmlschema.registerSchema
      3    (
      4      :schemaURL,
      5      xdbURIType(:schemaPath).getClob(),
      6      TRUE,TRUE,FALSE,TRUE,
      7      enableHierarchy=>DBMS_XMLSCHEMA.ENABLE_HIERARCHY_RESMETADATA
      8    );
      9  end;
    10  /
    PL/SQL procedure successfully completed.
    SQL> declare
      2    res boolean;
      3  begin
      4    if dbms_xdb.existsResource(:resourcePath) then
      5      dbms_xdb.deleteResource(:resourcePath,4);
      6    end if;
      7    res := dbms_xdb.createResource(:resourcePath,'Mary Had a Little Lamb');
      8  end;
      9  /
    PL/SQL procedure successfully completed.
    SQL> commit
      2  /
    Commit complete.
    SQL> select xdbUriType(:resourcePath).getClob()
      2  from dual
      3  /
    XDBURITYPE(:RESOURCEPATH).GETCLOB()
    Mary Had a Little Lamb
    SQL> declare
      2    resid raw(16);
      3  begin
      4    resid := dbms_xdb_version.makeVersioned(:resourcePath);
      5  end;
      6  /
    PL/SQL procedure successfully completed.
    SQL> commit
      2  /
    Commit complete.
    SQL> begin
      2    dbms_xdb_version.checkout(:resourcePath);
      3  end;
      4  /
    PL/SQL procedure successfully completed.
    SQL> commit
      2  /
    Commit complete.
    SQL> set long 10000 pages 0 lines 150
    SQL> --
    SQL> select r.res.getClobVal()
      2    from resource_view r
      3   where equals_path(res,:resourcePath) = 1
      4  /
    <Resource xmlns="http://xmlns.oracle.com/xdb/XDBResource.xsd" Hidden="false" Inv
    alid="false" VersionID="1" ActivityID="0" Container="false" CustomRslv="false" V
    ersionHistory="false" StickyRef="true">
      <CreationDate>2006-06-10T05:01:25.355009</CreationDate>
      <ModificationDate>2006-06-10T05:01:25.355009</ModificationDate>
      <DisplayName>metadataTest.txt</DisplayName>
      <Language>en-US</Language>
      <CharacterSet>UTF-8</CharacterSet>
      <ContentType>text/plain</ContentType>
      <RefCount>1</RefCount>
      <ACL>
        <acl description="Public:All privileges to PUBLIC" xmlns="http://xmlns.oracl
    e.com/xdb/acl.xsd" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:sch
    emaLocation="http://xmlns.oracle.com/xdb/acl.xsd                           http:
    //xmlns.oracle.com/xdb/acl.xsd">
          <ace>
            <grant>true</grant>
            <principal>PUBLIC</principal>
            <privilege>
              <all/>
            </privilege>
          </ace>
        </acl>
      </ACL>
      <Owner>SCOTT</Owner>
      <Creator>SCOTT</Creator>
      <LastModifier>SCOTT</LastModifier>
      <SchemaElement>http://xmlns.oracle.com/xdb/XDBSchema.xsd#binary</SchemaElement
    >
      <Contents>
        <text>Mary Had a Little Lamb</text>
      </Contents>
      <VCRUID>15DE79F263F7CC00E040578C2A0656F8</VCRUID>
    </Resource>
    SQL> begin
      2    dbms_xdb.appendResourceMetadata(:resourcePath,xmltype(
      3  '<root xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
      4         xsi:noNamespaceSchemaLocation="metadata.xsd">
      5      <child>
      6        <Child1>AAA</Child1>
      7        <Child2>BBB</Child2>
      8      </child>
      9  </root>'));
    10  end;
    11  /
    PL/SQL procedure successfully completed.
    SQL> commit
      2  /
    Commit complete.
    SQL> select r.res.getClobVal()
      2    from resource_view r
      3   where equals_path(res,:resourcePath) = 1
      4  /
    <Resource xmlns="http://xmlns.oracle.com/xdb/XDBResource.xsd" Hidden="false" Inv
    alid="false" VersionID="0" ActivityID="0" Container="false" CustomRslv="false" V
    ersionHistory="false" StickyRef="true">
      <CreationDate>2006-06-10T05:01:25.355009</CreationDate>
      <ModificationDate>2006-06-10T05:01:25.558913</ModificationDate>
      <DisplayName>metadataTest.txt</DisplayName>
      <Language>en-US</Language>
      <CharacterSet>UTF-8</CharacterSet>
      <ContentType>text/plain</ContentType>
      <RefCount>1</RefCount>
      <ACL>
        <acl description="Public:All privileges to PUBLIC" xmlns="http://xmlns.oracl
    e.com/xdb/acl.xsd" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:sch
    emaLocation="http://xmlns.oracle.com/xdb/acl.xsd                           http:
    //xmlns.oracle.com/xdb/acl.xsd">
          <ace>
            <grant>true</grant>
            <principal>PUBLIC</principal>
            <privilege>
              <all/>
            </privilege>
          </ace>
        </acl>
      </ACL>
      <Owner>SCOTT</Owner>
      <Creator>SCOTT</Creator>
      <LastModifier>SCOTT</LastModifier>
      <SchemaElement>http://xmlns.oracle.com/xdb/XDBSchema.xsd#text</SchemaElement>
      <Contents>
        <text>Mary Had a Little Lamb</text>
      </Contents>
      <VCRUID>15DE79F263F7CC00E040578C2A0656F8</VCRUID>
      <Parents>15DE79F263F8CC00E040578C2A0656F8</Parents>
      <root xmlns="" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:noNam
    espaceSchemaLocation="metadata.xsd">
        <child>
          <Child1>AAA</Child1>
          <Child2>BBB</Child2>
        </child>
      </root>
    </Resource>
    SQL> quit
    Disconnected from Oracle Database 11g Enterprise Edition Release 11.1.0.0.0 - Alpha
    With the Partitioning and Data Mining options
    I've filed bug 5313655 for the 10.2.x behavoir. If this is important to you you might want to consider signing up for the forthcoming beta program for the next release of the database

Maybe you are looking for

  • Print message when there are no pages to show in output

    Hi, I need to give a custom message like "Report did not produce any output due to .... " in the BI publisher report output instead of adobe giving the error 'there are no pages" in the pdf output. Any inputs ? Thanks&regards, Mukthi

  • Photoshop.exe Application Error

    Open image in CS4, after a couple of minutes a dialog box opens, Photoshop.exe - Application Error, how to fix please?

  • How do i transfer downloaded alerts from itunes to sounds

    Hi, I have seen other posts from people so reckon this must be a bug that needs fixing. what's the point of selling things to customers and not allowing us to use them as they are intended. Money back please apple.

  • Creating text file

    I use Flash cs4 , i need to know how to store and load text file into actionscript .

  • MRP-Vendor Monitoring Report

    Hello, There is a need of vendor monitoring report from my client. Here is the detail of the report need: Its a report ,that will be used by MRP controller and expidetators to track vendor for the critical items These critical items are the one ,whic