Best practices: search directories and project hierarchies

Hello,
I'm wondering what conventional wisdom says about structuring a TestStand project.
When I develop a LabVIEW project, this is the general structure I use:
Root folder\
My LV Project.lvproj
Main.vi
Functionality1\
Sub1a.vi
Sub1b.vi
Functionality2\
Sub2a.vi
Sub2b.vi
For my first TestStand project, I used this structure to separate my TestStand bits from my LabVIEW bits:
Root folder\
My TS Workspace.tsw
My TS Project.tpj
Main.seq
Sub1.seq
Sub2.seq
LabVIEW Backend\
My LV Project.lvproj
Test1.vi
Test2.vi
Unfortunately, every time I add an Action or Test and select my .lvproj file, TestStand tells me "The file you selected does not reside in any of the search directories."
So far, I've opted to "Use a relative path for the file you selected" every time, but this is not an optimal workflow. "Add the directory that contains the file you selected to the list of search directories" seems to be the pain-free option, but from what I understand, search directories are system-wide; I only want to add this directory to this TestStand project.
What are your thoughts? How should I manage my projects? Would these choices affect what I need to do to deploy the software on my customer's PC?
Thanks!

I recommend specifying your code module files (e.g. lvproj and vi files) relative to your sequence files. That way you can change your root folder without affecting anything and you don't have to add any search directories.
For this to work best, the files need to be in a directory underneath the sequence file in which it is specified or in the same directory as the sequence file.
If the code module is directly in the same directory as the sequence file, you will not be prompted to use the relative path, it will automatically do so. If it's not you will be prompted for a relative path, but that should be the default and you can just click the ok button. If it's above the sequence file's directory then the file browse dialog is no help for specifying relative to the sequence file because it will not do paths with "..". You can manually type paths with .. in them though.
Alternatively, using a search directory might be reasonable in your use case if most of your code modules are in the same LabVIEW Backend directory. You are correct though that there is not currently a per project or per sequence file way to specify search directories. One alternative I might recommend is that you have a single search directory (i.e. "LabVIEW Backend"), do NOT make it recursive, and add subdirectories for each of your teststand projects for which you want different code modules and specify your code modules with a more complete relative path such as: <subdir_under_searchdir>\myvi.vi
Hope this helps,
-Doug

Similar Messages

  • How can i get best practice for SD and MM

    Please, can any body tell me how can i get best practices for SD and MM for functional approach?
    Thanks
    Utpal

    Hello Utpal,
    I am really surprised, in just 10 minutes you searched that site and found it not useful. <b>Check out my previous reply "you will not find screen shot in this but you can add it in this"</b>
    You will not find readymade document, you need to add this as per your requirement.
    btw, the following link gives you some more link for new SAP guys, this will be helpful. <b>Check out HOW to BASIC transaction</b>
    New to Materials Management / Warehouse Management?
    Hope this helps.
    Regards
    Arif Mansuri

  • Best Practice for new BI project

    we are about to start new BI project.
    what is the best practice to start new BI project
    1. start only with FI as poc
    2. more than one module does not matter if the project was planned and scoped well
    can you give me link to best practice in starting BI projects?
    Regards

    Hi,
    Please go through following links.
    Business Intelligence : Getting Started with Business Intelligence ,Reporting, Analysis, and Planning , Data Warehousing, BI Accelerator Embedded and Process-Centric BI
    https://www.sdn.sap.com/irj/sdn/nw-bi
    The Home of SAP Business Warehouse (BW)
    http://www.erpgenie.com/sap/sapfunc/bw.htm
    SAP Business Information Warehouse
    http://help.sap.com/saphelp_nw04/helpdata/en/b2/e50138fede083de10000009b38f8cf/content.htm
    Business Intelligence : Programming in BW
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/bi/programminginBW&
    BW and Portals 2005
    https://www.sdn.sap.com/irj/sdn/bi-and-portals2005
    SAP Business Warehouse (BW) Overview
    http://gleez.com/sap/bw/overview
    Business Intelligence : Steps to get started with SAP BW
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/bi/stepstogetstartedwithSAPBW&
    SAP Business Information Warehouse Scenarios
    http://help.sap.com/bp_biv335/BI_EN/html/Bw.htm
    SAP BW Learning Guide
    http://searchsap.techtarget.com/general/0,295582,sid21_gci1077480,00.html
    SAP BW Business Warehouse - Introduction
    http://www.thespot4sap.com/Articles/SAP_BW_Introduction.asp
    Business Content Frontend Design Guidelines (NW04)
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/bda556e9-0c01-0010-83b0-d519d6deb9e9
    How To…Create and Maintain UI Patterns of BI Content
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/ad214fe9-0c01-0010-4291-a629e5ba5f90
    SAP NetWeaver BI Integrated Planning for Finance
    http://www.sap-press.de/katalog/buecher/htmlleseproben/gp/htmlprobID-113
    Info object,infocube,infosource,datasource,commn structure,extract structure ..etc..
    http://www.erpgenie.com/sapgenie/docs/MySAP%20BW%20Cookbook%20Vol%201.pdf
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/01b9395c-0e01-0010-6786-c4ee5e5d2154
    BI Performance Tuning
    FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap
    Business Intelligence Journal Improving Query Performance in Data Warehouses
    http://www.tdwi.org/Publications/BIJournal/display.aspx?ID=7891
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    SAP Business Intelligence Accelerator : A High - Performance Analytic Engine for SAP Ne tWeaver Business Intelligence
    http://www.sap.com/platform/netweaver/pdf/BWP_AR_IDC_BI_Accelerator.pdf
    BI Performance Audit
    http://www.xtivia.com/downloads/Xtivia_BIT_Performance%20Audit.pdf
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/10564d5c-cf00-2a10-7b87-c94e38267742
    Enhancements in SAP BW
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/59069d90-0201-0010-fd81-d5e11994d8b5
    Customer Enhancements in SAP NetWeaver BI (Exits, BAdIs and ABAP™ in the SAP NetWeaver BI Back End)
    http://www.sap.com/community/pub/showdetail.epx?itemID=5257
    BW 3.x and 3.5 "How To" Guides List
    https://www.sdn.sap.com/irj/sdn/docs?rid=/webcontent/uuid/ee14e25d-0501-0010-11ad-8eb2861a7ec0 [original link is broken]
    Assign ponits if it helps
    Thanks & Regards
    santo

  • Best Practice for Planning and BI

    What's the best practice for Planning and BI infrastructure - set up combined on one box or separate? What are the factors to consider?
    Thanks in advance..

    There is no way that question could be answered with the information that has been provided.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Best Practice regarding using and implementing the pref.txt file

    Hi All,
    I would like to start a post regarding what is Best Practice in using and implementing the pref.txt file. We have reached a stage where we are about to go live with Discoverer Viewer, and I am interested to know what others have encountered or done to with their pref.txt file and viewer look and feel..
    Have any of you been able to add additional lines into the file, please share ;-)
    Look forward to your replies.
    Lance

    Hi Lance
    Wow, what a question and the simple answer is - it depends. It depends on whether you want to do the query predictor, whether you want to increase the timeouts for users and lists of values, whether you want to have the Plus available items and Selected items panes displayed by default, and so on.
    Typically, most organizations go with the defaults with the exception that you might want to consider turning off the query predictor. That predictor is usually a pain in the neck and most companies turn it off, thus increasing query performance.
    Do you have a copy of my Discoverer 10g Handbook? If so, take a look at pages 785 to 799 where I discuss in detail all of the preferences and their impact.
    I hope this helps
    Best wishes
    Michael Armstrong-Smith
    URL: http://learndiscoverer.com
    Blog: http://learndiscoverer.blogspot.com

  • Best practice for Plan and actual data

    Hello, what is the best practice for Plan and actual data?  should they both be in the same app or different?
    Thanks.

    Hi Zack,
    It will be easier for you to maintain the data in a single application. Every application needs to have the category dimension, mandatorily. So, you can use this dimension to maintain the actual and plan data.
    Hope this helps.

  • SAP Business One Best-Practice System Setup and Sizing

    <b>SAP Business One Best-Practice System Setup and Sizing</b>
    Get recommendations from SAP and hardware specialists on system setup and sizing
    SAP Business One is a single, affordable, and easy-to-implement solution that integrates the entire business across financials, sales, customers, and operations. With SAP Business One, small businesses can streamline their operations, get instant and complete information, and accelerate profitable growth. SAP Business One is designed for companies with less than 100 employees, less than $75 million in annual revenue, and between 1 and 30 system users, referred to as the SAP Business One sweet spot. The sweet spot covers various industries and micro-verticals which have different requirements when it comes to the use of SAP Business One.
    One of the initial steps during the installation and implementation of SAP Business One is the definition of the system landscape and architecture. Numerous factors affect the system landscape that needs to be created to efficiently run SAP Business One.
    The <a href="http://wiki.sdn.sap.com/wiki/display/B1/BestPractiseSystemSetupand+Sizing">SAP Business One Best-Practice System Setup and Sizing Wiki</a> provides recommendations on how to size and configure the system landscape and architecture for SAP Business One based on best practices.

    For such high volume licenses, you may contact the SAP Local Product Experts.
    You may get their contact info from this site
    [https://websmp209.sap-ag.de/~sapidb/011000358700001455542004#India]

  • DNS best practices for hub and spoke AD Architecture?

    I have an Active Directory Forest with a forest root such as joe.co and the root domain of the same name, and root DNS servers (Domain Controllers) dns1.joe.co and dns2.joe.co
    I have child domains with names in the form region1.joe.com, region2.joe.co and so on, with dns servers dns1.region1.joe.co and so on.
    Each region has distribute offices that may have a DC in them, servers named in the form dns1branch1.region1.joe.co
    Over all my DNS tests out okay, but I want to get the general guidelines for setting up new DCs correct.
    Configuration:
    Root DC/DNS server dns1.joe.co adapter settings points DNS to itself, then two other root domain DNS/DCs dns2.joe.co and dns3.joe.co.
    The other root domain DNS/DCs adapter settings point to root server dns1.joe.co and then to itself dns2.joe.co, and then 127.0.0.1
    The regional domains have a root dns server dns1.region1.joe.co with adapter that that points to root server dns1.joe.co then to itself.
    The additional region domain DNS/DCs adapter settings point to dns1.region1.joe.co then to itself then to dn1.joe.co
    What would you do to correct this topology (and settings) or improve it?
    Thanks in advance
    just david

    Hi,
    According to your description, my understanding is that you need suggestion about your DNS topology.
    In theory, there is no obvious problem. Except for the namespace and server plaining for DNS, zone is also needed to consideration. If you place DNS server on each domain and subdomain, confirm that if the traffic browsed by DNS will affect the network performance.
    Besides, fault tolerance and security are also necessary.
    We usually recommend that:
    DC with DNS should point to another DNS server as primary and itself as secondary or tertiary. It should not point to self as primary due to various DNS islanding and performance issues that can occur. And when referencing a DNS server on itself, a DNS client
    should always use a loopback address and not a real IP address. detailed information you may reference:
    What is Microsoft's best practice for where and how many DNS servers exist? What about for configuring DNS client settings on DC’s and members?
    http://blogs.technet.com/b/askds/archive/2010/07/17/friday-mail-sack-saturday-edition.aspx#dnsbest
    How To Split and Migrate Child Domain DNS Records To a Dedicated DNS Zone
    http://blogs.technet.com/b/askpfeplat/archive/2013/12/02/how-to-split-and-migrate-child-domain-dns-records-to-a-dedicated-dns-zone.aspx
    Best Regards,
    Eve Wang
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Best practices for setting up projects

    We recently adopted using Captivate for our WBT modules.
    As a former Flash and Director user, I can say it’s
    fast and does some great things. Doesn’t play so nice with
    others on different occasions, but I’m learning. This forum
    has been a great source for search and read on specific topics.
    I’m trying to understand best practices for using this
    product. We’ve had some problems with file size and
    incorporating audio and video into our projects. Fortunately, the
    forum has helped a lot with that. What I haven’t found a lot
    of information on is good or better ways to set up individual
    files, use multiple files and publish projects. We’ve decided
    to go the route of putting standalones on our Intranet. My gut says
    yuck, but for our situation I have yet to find a better way.
    My question for discussion, then is: what are some best
    practices for setting up individual files, using multiple files and
    publishing projects? Any references or input on this would be
    appreciated.

    Hi,
    Here are some of my suggestions:
    1) Set up a style guide for all your standard slides. Eg.
    Title slide, Index slide, chapter slide, end slide, screen capture,
    non-screen capture, quizzes etc. This makes life a lot easier.
    2) Create your own buttons and captions. The standard ones
    are pretty ordinary, and it's hard to get a slick looking style
    happening with the standard captions. They are pretty easy to
    create (search for add print button to learn how to create
    buttons). There should instructions on how to customise captions
    somewhere on this forum. Customising means that you can also use
    words, symbols, colours unique to your organisation.
    3) Google elearning providers. Most use captivate and will
    allow you to open samples or temporarily view selected modules.
    This will give you great insight on what not to do and some good
    ideas on what works well.
    4) Timings: Using the above research, I got others to
    complete the sample modules to get a feel for timings. The results
    were clear, 10 mins good, 15 mins okay, 20 mins kind of okay, 30
    mins bad, bad, bad. It's truly better to have a learner complete
    2-3 short modules in 30 mins than one big monster. The other
    benefit is that shorter files equal smaller size.
    5) Narration: It's best to narrate each slide individually
    (particularly for screen capture slides). You are more likely to
    get it right on the first take, it's easier to edit and you don't
    have to re-record the whole thing if you need to update it in
    future. To get a slicker effect, use at least two voices: one male,
    one female and use slightly different accents.
    6) Screen capture slides: If you are recording filling out
    long window based databse pages where the compulsory fields are
    marked (eg. with a red asterisk) - you don't need to show how to
    fill out every field. It's much easier for the learner (and you) to
    show how to fill out the first few fields, then fade the screen
    capture out, fade the end of the form in with the instructions on
    what to do next. This will reduce your file size. In one of my
    forms, this meant the removal of about 18 slides!
    7) Auto captions: they are verbose (eg. 'Click on Print
    Button' instead of 'Click Print'; 'Select the Print Preview item'
    instead of 'Select Print Preview'). You have to edit them.
    8) PC training syntax: Buttons and hyperlinks should normally
    be 'click'; selections from drop down boxes or file lists are
    normally 'select': Captivate sometimes mixes them up. Instructions
    should always be written in the correct order: eg. Good: Click
    'File', Select 'Print Preview'; Bad: Select 'Print Preview' from
    the 'File Menu'. Button names, hyperlinks, selections are normally
    written in bold
    9) Instruction syntax: should always be written in an active
    voice: eg. 'Click Options to open the printer menu' instead of
    'When the Options button is clicked on, the printer menu will open'
    10) Break all modules into chapters. Frame each chapter with
    a chapter slide. It's also a good idea to show the Index page
    before each chapter slide with a progress indicator (I use an
    animated arrow to flash next to the name of the next chapter), I
    use a start button rather a 'next' button for the start of each
    chapter. You should always have a module overview with the purpose
    of the course and a summary slide which states what was covered and
    they have complete the module.
    11) Put a transparent click button somewhere on each slide.
    Set the properties of the click box to take the learner back to the
    start of the current chapter by pressing F2. This allows them to
    jump back to the start of their chapter at any time. You can also
    do a similar thing on the index pages which jumps them to another
    chapter.
    12) Recording video capture: best to do it at normal speed
    and be concious of where your mouse is. Minimise your clicks. Most
    people (until they start working with captivate) are sloppy with
    their mouse and you end up with lots of unnecessarily slides that
    you have to delete out. The speed will default to how you recorded
    it and this will reduce the amount of time you spend on changing
    timings.
    13) Captions: My rule of thumb is minimum of 4 seconds - and
    longer depending on the amount of words. Eg. Click 'Print Preview'
    is 4 seconds, a paragraph is longer. If you creating knowledge
    based modules, make the timing long (eg. 2-3 minutes) and put in a
    next button so that the learner can click when they are ready.
    Also, narration means the slides will normally be slightly longer.
    14) Be creative: Capitvate is desk bound. There are some
    learners that just don't respond no matter how interactive
    Captivate can be. Incorporate non-captivate and desk free
    activities. Eg. As part of our OHS module, there is an activity
    where the learner has to print off the floor plan, and then wander
    around the floor marking on th emap key items such as: fire exits;
    first aid kit, broom and mop cupboard, stationary cupboard, etc.
    Good luck!

  • Best Practices for Packaging and Deploying Server-Specific Configurations

    We have some server-specific properties that vary for each server. We'd
    like to have these properties collected together in their own properties
    file (either .properties or .xml is fine).
    What is the best-practices way to package and deploy an application (as an
    ear file), where each server needs some specific properties?
    We'd kind of like to have the server-specific properties file be stored
    external to the ear on the server itself, so that the production folks can
    configure each server's properties at the server. But it appears that an
    application can't access a file external to the ear, or at least we can't
    figure out the magic to do it. If there is a way to do this, please let me
    know how.
    Or do we have to build a unique ear for each server? This is possible, of
    course, but we'd prefer to build one deployment package (ear), and then
    ship that off to each server that is already configured for its specific
    environment. We have some audit requirements where we need to ensure that
    an ear that has been tested by QA is the very same ear that has been
    deployed, but if we have to build one for each server, this is not
    possible.
    Any help or pointers would be most appreciated. If this is an old issue,
    my apologies, would you please point me to any previous material to read?
    I didn't see anything after searching through this group's archives.
    Thanks much in advance,
    Paul
    Paul Hodgetts -- Principal Consultant
    Agile Logic -- www.agilelogic.com
    Consulting, Coaching, Training -- On-Site & Out-Sourced Development
    Java, J2EE, C++, OOA/D -- Agile Methods/XP/Scrum, Use Cases, UI/IA

    The one draw back to this is you have to go all the way back to ant and the
    build system to make changes. You really want these env variables to be
    late binding.
    cheers
    mbg
    "Sai S Prasad" <[email protected]> wrote in message
    news:[email protected]...
    >
    Paul,
    I have a similar situation in our project and I don't create ear filesspecific
    to the environment. I do the following:
    1) Create .properties file for every environment with the same attributename
    but different values in it. For example, I have phoneix.properties.NT,phoenix.properties.DEV,
    phoenix.properties.QA, phoenix.properties.PROD.
    2) Use Ant to compile, package and deploy the ear file
    I have a .bat file in NT and .sh for Solaris that in turn calls theant.bat or
    ant.sh respectively. For the wrapper batch file or shell script, you canpass
    the name of the environment. The wrapper batch file will copy theappropriate
    properties file to "phonenix.properties". In the ant build.xml, I alwaysrefer
    to phonenix.properties which is available all the time depending on theenvironment.
    >
    It works great and I can't think of any other flexible way. Hope thathelps.
    >
    Paul Hodgetts <[email protected]> wrote:
    We have some server-specific properties that vary for each server. We'd
    like to have these properties collected together in their own properties
    file (either .properties or .xml is fine).
    What is the best-practices way to package and deploy an application (as
    an
    ear file), where each server needs some specific properties?
    We'd kind of like to have the server-specific properties file be stored
    external to the ear on the server itself, so that the production folks
    can
    configure each server's properties at the server. But it appears that
    an
    application can't access a file external to the ear, or at least we can't
    figure out the magic to do it. If there is a way to do this, please
    let me
    know how.
    Or do we have to build a unique ear for each server? This is possible,
    of
    course, but we'd prefer to build one deployment package (ear), and then
    ship that off to each server that is already configured for its specific
    environment. We have some audit requirements where we need to ensure
    that
    an ear that has been tested by QA is the very same ear that has been
    deployed, but if we have to build one for each server, this is not
    possible.
    Any help or pointers would be most appreciated. If this is an old issue,
    my apologies, would you please point me to any previous material to read?
    I didn't see anything after searching through this group's archives.
    Thanks much in advance,
    Paul
    Paul Hodgetts -- Principal Consultant
    Agile Logic -- www.agilelogic.com
    Consulting, Coaching, Training -- On-Site & Out-Sourced Development
    Java, J2EE, C++, OOA/D -- Agile Methods/XP/Scrum, Use Cases, UI/IA

  • What are best practice for packaging and deploying j2EE apps to iAS?

    We've been running a set of J2EE applications on a pair of iAS SP1b for about a year and it has been quite stable.
    Recently however we have had a number of LDAP issues, particularly when registering and unregistering applications (registering ear files sometimes fails 1st time but may work 2nd time). Also We've noticed very occasionally that old versions of classes sometimes find their way onto our machines.
    What is considered to be best practice in terms of packaging and deployment, specifically:
    1) Packaging - using the deployTool that comes with iAS6 SP1b to package is a big manual task, especially when you have 200+ jsp files. Are people out there using this or are they scripting it with a build tool such as Ant?
    2) Deploying an existing application to multiple iAS's. Are you guys unregistering old application then reregistering new application? Are you shutting down iAS whilst doing the deployment?
    3) Deploying ear files can take 5 to 10 mins, is this normal?
    4) In a clustered scenario where HTTPSession is shared what are the consequences of doing deployments to data stored in session?
    thanks in asvance for your replies
    Owen

    You may want to consider upgrading your application server environment to a newer service pack. There are numerous enhancements involving the deployment tool and run time layout of your application that make clear where you're application is loading its files from.
    If you've at a long running application server environment, with lots of deployments under your belt, you might start to notice slow downs in deployment and kjs start time. Generally this is due to garbage collecting in your iAS registry.
    You can do several things to resolve this. The most complete solution is to reinstall the application server. This will guarantee a clean ldap registry. Of course you've got to restablish your configurations and redeploy your applications. When done, backup your application server install space with the application server and directory server off. You can use this backup to return to a known configuation at some future time.
    For the second method: <B>BE CAREFUL - BACKUP FIRST</B>
    There is a more exhaustive solution that involves examining your deployed components to determine the active GUIDS. You then search the NameTrans section of the registry searching for Applogic Servlet *, and Bean * entries that represent your previously deployed components but are represented in the set of deployed GUIDs. Record these older GUIDs, remove them from ClassImp and ClassDef. Finally remove the older entries from NameTrans.
    Best practices for deployment depend on your particular environmental needs. Many people utilize ANT as a build tool. In later versions of the application server, complete ANT scripts are included that address compiling, assembly and deployment. Ant 1.4 includes iAS specific targets and general J2EE targets. There are iAS specific targets that can be utilized with the 1.3 version. Specialized build targets are not required however to deploy to iAS.
    Newer versions of the deployment tool allow you to specify that JSPs are not to be registered automatically. This can be significant if deployment times lag. Registered JSP's however benefit more fully from the services that iAS offers.
    2) In general it is better to undeploy then redeploy. However, if you know that you're not changing GUIDs, recreating an existing application with new GUIDs, or removing registered components, you may avoid the undeploy phase.
    If you shut down the KJS processes during deployment you can eliminate some addition workload on the LDAP server which really gets pounded during deployment. This is because the KJS processes detect changes and do registry loads to repopulate their caches. This can happen many times during a deployment and does not provide any benefit.
    3) Deploying can be a lengthy process. There have been improvements in that performance from service pack to service pack but unfortunately you wont see dramatic drops in deployment times.
    One thing you can do to reduce deployment times is to understand the type of deployment. If you have not manipulated your deployment descriptors in any way, then there is no need to deploy. Simply drop your newer bits in to the run time space of the application server. In later service packs this means exploding the package (ear,war, or jar) in to the appropriate subdirectory of the APPS directory.
    4) If you've changed the classes of objects that have been placed in HTTPSession, you may find that you can no longer utilize those objects. For that reason, it is suggested that objects placed in session be kept as simple as possible in order to minimize this effect. In general however, is not a good idea to change a web application during the life span of a session.

  • Kernel: PANIC! -- best practice for backup and recovery when modifying system?

    I installed NVidia drivers on my OL6.6 system at home and something went bad with one of the libraries.  On reboot, the kernel would panic and I couldn't get back into the system to fix anything.  I ended up re-installing the OS to recovery my system. 
    What would be some best practices for backing up the system when making a change and then recovering if this happens again?
    Would LVM snapshots be a good option?  Can I recovery a snapshot from a rescue boot?
    EX: File system snapshots with LVM | Ars Technica -- scroll down to the section discussing LVM.
    Any pointers to documentation would be welcome as well.  I'm just not sure what to do to revert the kernel or the system when installing something goes bad like this.
    Thanks for your attention.

    There is often a common misconception: A snapshot is not a backup. A snapshot and the original it was taken from initially share the same data blocks. LVM snapshot is a general purpose solution which can be used, for example, to quickly create a snapshot prior to a system upgrade, then if you are satisfied with the result, you would delete the snapshot.
    The advantage of a snapshot is that it can be used for a live filesystem or volume while changes are written to the snapshot volume. Hence it's called "copy on write (COW), or copy on change if you want. This is necessary for system integrity to have a consistent data status of all data at a certain point in time and to allow changes happening, for example to perform a filesystem backup. A snapshot is no substitute for a disaster recovery in case you loose your storage media. A snapshot only takes seconds, and initially does not copy or backup any data, unless data changes. It is therefore important to delete the snapshot if no longer required, in order to prevent duplication of data and restore file system performance.
    LVM was never a great thing under Linux and can cause serious I/O performance bottlenecks. If snapshot or COW technology suits your purpose, I suggest you look into Btrfs, which is a modern filesystem built into the latest Oracle UEK kernel. Btrfs employs the idea of subvolumes and is much more efficient that LVM because it can operate on files or directories while LVM is doing the whole logical volume.
    Keep in mind however, you cannot use LVM or Btrfs with the boot partition, because the Grub boot loader, which loads the Linux kernel, cannot deal with LVM or BTRFS before loading the Linux kernel (catch22).
    I think the following is an interesting and fun to read introduction explaining basic concepts:
    http://events.linuxfoundation.org/sites/events/files/slides/Btrfs_1.pdf

  • Validation Best Practice for an javafx project

    Hello,
    I am investigating how to validate a JavaFX project and specifically where to display the validation results with appropriate icons and text. Do you have any thoughts on best practice for this? Are there any features in JavaFX to do this? Are there any third-party library that does this?
    Thanks

    See:
    Re: Forms and validations - here's some of my ideas, what are yours? "Thread: Forms and validations - here's some of my ideas, what are yours?"
    And messages in this mailing list thread:
    http://mail.openjdk.java.net/pipermail/openjfx-dev/2012-February/000717.html "Validate Me"

  • Best Practice for unimplemented OTC project transports?

    Hello,
    We implemented and went live with portions of ERP modules COPA, MM, GL(new), and Consolidations 2 years ago as a Phase 1 implementation.  Phase 2, which began directly afterwards, included CRM, OTC, and BW.  The problem that has come up is that due to numerous implementation issues, we have not gone live with Phase 2 and it is still undetermined when/if we will implement at least OTC in ERP.  We have a 3-system (Dev, QA, Prod) ERP landscape and are running into issues due to inconsistencies between QA and Prod.  All the transports related to the Phase 2 project were created in Dev and moved into QA, but due to the status of the project were never moved into Production.  We recently had an issue with moving changes to our operating concern through the landscape, due to the existing inconsistencies of the operating concern configurations between Dev, QA, and Prod which led to us not being able to re-activate the Operating Concern in Prod until we moved in additional transports that were tied to Phase 2.
    My question is this...What would be the best practice/approach to resolving the inconsistencies in our ERP landscape to assure that we have accurate QA testing of our Phase 1 implementation, but also trying not to lose the existing Phase 2 development if we decide to implement OTC in the future?  I'm considering the below options:
    A)  Move all Phase 2 requests
    - Refresh QA (via system copy of Prod)
    - Move all Phase 2 transports that were originally moved into QA into the refreshed system and test existing Phase 1 business processes to determine risk of moving into Prod
    - Move all Phase 2 transports into Prod in order to maintain 3-system consistency
    B)  2 system consistency
    - Refresh QA (via system copy of Prod)
    - Leave all Phase 2 transports in import queue for QA, and maintain Prod/QA consistency only
    - OTC implementation can be realized with moving Phase 2 transports through QA at some point in the future
    C)  "Reset button"
    - Refresh QA (via system copy of Prod)
    - Refresh Dev (via system copy of Prod) - I'm not sure what technical considerations would need to be made around the development system's role as the origin of rep and dictionary objects and how this can be maintain in a system copy?
    - This would wipe out all Phase 2 development
    I would greatly appreciate anyone's guidance on our options given our current scenario.
    thanks,
    John

    I would suggest to go with the option A though it seems more work with this option. But this is the only option which can effectively work in the long run. With option B, you will not completely eliminate the problem and with option C, you will loose all your phase 2 work, which will be big waste of the efforts.
    Additional advantage with option A is that whenever your organization decides to go-live with Phase 2 work then minimal regression testing will be required as most of your work will be already tested and verified. Regression testing and remediation is significant work whevever a solution is introduced in a working environment.

  • Best Practice Re: Workflows and Workflow Components

    Hello,
    As many know, CUP does not allow you to delete workflows and workflow components if they are referenced in a request somewhere. The solution I'm seeing here on the forums is to run a Request Delete Script from SAP. This is all well and good in a Sandbox environment, but we would not like to delete request information in Productoin.
    So my question is: what is best practice for managing workflows and their components in Production? As time goes on, our workflows will change, our components might change, we might need to add a stage here or there or switch an initiator, etc etc. However, if I create a new workflow or component, I can't delete the old one. So essentially there's potential to have a lot of junk workflow/component data sitting out there because it's referenced in an old request somewhere.
    Does anyone have any recommendations on how to manage this? Ex: If you change a workflow to add another stage, what are you doing with the old workflow since it's useless?
    Let me know if I need to clarify further.
    Thanks!!
    Jes

    f l,
    I'm not sure deleting keys from the registry is ever a best practice, however Xcelsius has listings in:
    HKEY_CURRENT_USER > Software > Business Objects > Xcelsius
    HKEY_LOCAL_MACHINE > SOFTWARE > Business Objects > Suite 12.0 > Xcelsius
    The current user folder holds temporary settings, such as how you've modified your interface.
    The local machine folder holds more important information.
    As always, it's recommended that you backup the registry and/or create a restore point before modifying or deleting any keys.
    As for directories, the only directory Xcelsius uses is the one you install to.  It also places some install logs in the temp directory, but they have no effect on the application.

Maybe you are looking for

  • Where can i get an apps to install photoshop 5.5 on windows 8

    anyone know where to get this apps thanks

  • Score Box

    I have a Dynamic Score box that does not display the correct score.  This is the code I have if (man.hitTest(circle)){   circle._x=5;   score=score + 1   scoreDisplay.text=score Can you help

  • Garnishment Payment Method

    HR would like to perform a garnishment bank wire to the state. the state vendor's master data is set up with a payment methos of "T". However, after third party is run the line items indicate a payment method of "C." Each line item is then changed to

  • [CDA+ISE] Multi-Homed (Wired/Wireless) Workstations only partially ID'd in CDA

    CDA Patch 2 ISE 1.2 Patch 5 CDA with AD and ISE integration is currently working, however for machines with BOTH a Wired and Wireless connection, CDA is auto-unmapping the wired connection, but leaving the wireless in place. ISE shows two valid sessi

  • ITunes Plus - why do we have to upgrade everything?

    I see iTunes is giving us no choice about which tracks/albums that we upgrade -- it's all or nothing. The "plus" tracks definitely sound better, but to upgrade my entire purchased library would cost me $600 in upgrade fees (assuming all of my 1,400+