Best practice for making form read-only at certain steps in process?

I have a process with a single form and 4 Assign Task operations.  During Assign Task operations 1 and 3, the form is edited.  During Assign Task operation 2 and 4, it should be read-only (it is reviewed and acknowledged, but not changed).   There is no single criteria in the form data that indicates whether a form should be read-only or not -- this is only signaled by the step in the process.  What is the best way to manage the read-only state?
Is there a way to make the form read-only via a setting in Workbench, or should I do this via a script in the form?
Thanks!
Toby

I have a process with a single form and 4 Assign Task operations.  During Assign Task operations 1 and 3, the form is edited.  During Assign Task operation 2 and 4, it should be read-only (it is reviewed and acknowledged, but not changed).   There is no single criteria in the form data that indicates whether a form should be read-only or not -- this is only signaled by the step in the process.  What is the best way to manage the read-only state?
Is there a way to make the form read-only via a setting in Workbench, or should I do this via a script in the form?
Thanks!
Toby

Similar Messages

  • Best practices for deploying forms in a 'cluster'?

    Anyone know of any public docs that discuss typical best practices for
    - forms deployment;
    - forms apps management and version control; and/or
    - deploying (and keeping) the .frm/frx in sync when using multiple forms servers in a HA or load balancing envrionment?

    Hi adil,                      
    Based on your description, you want to know the best practices for search service in a SharePoint farm.
    Different farms have different search topologies, for the best search performance, I recommend that you follow the guidance for small, medium, and large farms.
    The article is about the guidance for different farms. 
    Search service can run with other services in the same server, if condition permits and you want to have better performance for search service and other services including BI performance, you can deploy search service in dedicated server.
    If condition permits, I recommend combining a query component with a front-end Web server to avoid putting crawl components and query components on the same serve.
    In your SharePoint farm, you can deploy the query components in a WFE server and the crawl components in an application server.
    The articles below describe the best practices for enterprise search.
    https://technet.microsoft.com/en-us/library/cc850696(v=office.14).aspx
    https://technet.microsoft.com/en-us/library/cc560988(v=office.14).aspx
    Best regards      
    Sara Fan
    TechNet Community Support

  • SAP Best Practices for Data Migration :repositories only on MS SQL Server ?

    Hi,
    I'm implementing the "SAP Best Practices for Data Migration" (see https://websmp109.sap-ag.de/bp-datamigration).
    As part of the installation you have to install MS SQL Server Express Edition. The installation guide contains detailed steps to do this. All repositories for Data Services should be running on SQL Server, according to the installation guide.
    The customer I'm working for now does not want to use SQL Server, but DB2, as company standard.
    So I use DB2 for the local and profiler repositories.
    I notice however that the web application http://localhost:8080/MigrationServices does not support DB2.The only database type you can select in the configuration area is MS SQL Server.
    Is this a limitation, a by design ?

    Hans,
    The current release of SAP Best Practices for Data Migration, v1.32, supports only MS SQL Server.  The intent when developing the DM content was to quickly set up a temporary, standardized data migration environment, using tools that are available to everyone.  SQL Server Express was chosen to host the repositories, because it is easy to set up and can be downloaded for free.  Sone users have successfully deployed the content on Oracle XE, but as you have found, the MigrationServices web application works only with SQL Server.
    The next release, including the web app, will support SQL Server and Oracle, but not DB2.
    Paul

  • SQL Server 2012 - Wat Is The Best Solution For Creating a Read Only Replicated/AlwaysOn Database

    Hi there I was wondering if someone may have a best recommendation for the following requirement I have with regards setting up a third database server for reporting?
    Current Setup
    SQL Server 2012 Enterprise setup at two sites (Site A & Site B).
    Configured to use AlwaysOn Availability groups for HA and DR.
    Installed on Windows 2012 Servers.
    This is all working and failover works fine and no issues. So…
    Requirement
    A third server needs to be added for the purpose of reporting, to be located on another site (Site C) possibly in another domain. This server needs to have a replicated read only copy of the live database from site A or Site B, whichever is in use. The Site
    C reporting database should be as up-to-date to the Site A or Site B database as possible – preferably within a few seconds anyway….
    Solution - What I believe are available to me
    I believe I can use AlwaysOn and create a ReadOnly replica for the Site C. If so do I assume Site C needs to have the Enterprise version of SQL server i.e. to match Site A & Site B?
    Using log shipping which if I am correct means the Site C does not need to be an Enterprise version.
    Any help on the best solution for this would be greatly appreciated.
    Thanks, Steve

    for always on - all nodes should be part of one windows cluster..if there site C is on different domain - I do not think it works.
    Logshipping works --as long as the sql on site C is is same or higher version(sql 2012 or above).  you can only do read only.
    IMHo, if you can make site C in the same domain then, Always is better solution else log shipping
    also, if your database has enterprise level features such as - partitonin, data compression -- you cannot restore the database on lower editions- so you need to have enterprise edition.
    Hope it Helps!!

  • Best Practice for making material obselete

    Hai Experts,
    Could any one please advice me what s the best practice to make a material obselete. Is there any best practice or incase if it is not there what would be the diff ways it cant be made obselete.
    Please advice me on this.
    Thanking you in advance
    Regards,
    Gopalakrishnan.S

    Archiving is a process, rather than a transaction. you have to take care that you fulfill local laws and retain your data for x years for internal and external auditors.
    archiving requires customizing, you have to tell SAP where your archive is and what name it has and how big it can be, and when (definition of retention period) a document can be archived.
    you have to talk with your business how long they want the data keep in the production system (this depends on how long they need to run reports on this data)
    you have to define who can access the archived data and how this data is accessed.
    Who runs archiving?
    Tough question. Data owner is the business, so they should run.
    But archiving just buckets fragments your tables.
    so many companies have a person who is fulltime responsible for archiving and does it for all business units and organisations.
    Archiving is nothing what can be done "on the fly", you will encounter all kind of errors, processes that are not designed well will certainly create a lot problems in archiving and need to be reworked.
    I had an archving project with 50 days project time, and could develope guidelines to archive about 30 different objects. there are 100 more objects possible, and we have still not archived master data like material, vendors and customers because of so many dependencies and object that need to be archived prior to those objects.

  • Best practice for making a report of 10,000 to 20,000 rows(OBIEE 10.3.4.1)

    My Scenario is like this:*
    Hi i have 2 fact tables fact1 and fact 2 and four dimension tables D1,D2,D3 ,D4 & D1.1 ,D1.2 the relations in the data model is like this :
    NOTE: D1.1 and D1.2 are derived from D1 So D1 might be snow Flake.
    [( D1.. 1:M..> Fact 1 , D1.. 1:M..> Fact 2 ), (D2.. 1:M..> Fact 1 , D2.. 1:M..> Fact 2 ), ( D3.. 1: M.> Fact 1 , D3.. 1:M..> Fact 2 ),( D4.. 1:M..> Fact 1 , D4 ... 1:M..> Fact 2 )]
    Now from D1 there is a child level like this: [D1 --(1:M)..> D1.1 and from D1.1.. 1:M..> D1.2.. 1:M..> D4]
    Please help me in modeling these for making a report of 10,000 rows and also let me know for which tables do i need to enable cache?
    PS: There shouldn't be performance issue so please help me in modeling this.
    Thanks in Advance for the Experts who are helping me for a while.

    Shudn't be much problem with just these many rows...
    Model something like this only Re: URGENT MODELING SNOW FLAKE SCHEMA
    There are various ways of handling performance issues if any in OBIEE.
    Go for caching strategy for complete warehouse. Make sure to purge it after every data load..If you have aggr calculations at higher level then you can also go for aggregated tables in OBIEE for better performance.
    http://www.rittmanmead.com/2007/10/using-the-obiee-aggregate-persistence-wizard/
    Hope this is clear...Go ahead with actual implementation and lets us know incase you encounter any major issues.
    Cheers

  • Best practices for making the end result web help printable

    Hi all, using TCS3 Win 7 64 bit.  All patched and up to date.
    I was wondering what the best practices are for the following scenario:
    I am authoring in Frame, link by reference into RH.
    I use Frame to generate PDFs and RH to generate webhelp.
    I have tons of conditional text which ultimately produce four separate versions of PDFs as well as online help - I handle these codes in FM and pull them into RH.
    I use a css on all pages of my RH to make it 'look' right.
    We now need to add the ability for end users to print the webhelp - outside of just CTRL+P because a)that cuts off the larger images and b)it doesn't show header, footer, logo, date, etc. (stuff that is in the master pages in FM).
    My thought is doing the following:
    Adding four sentences (one for each condition) in the FM book on the first page. Each one would be coded for audience A, B, C, or D (each of which require separate PDFs) as well as coded with ONLINE so that they don't show up in my printed PDFs that I generate out of Frame. Once the PDFs are generated, I would add a hyperlink in RH (manually) to each sentence and link the associated PDF (this seems to add the PDF file to the baggage files in RH). Then when I generate my RH webhelp, it would show the link, with the PDF, correctly based on the condition of the user looking at the help.
    My questions are as follows:
    1- This seems more complicated than it needs to be. Is it?
    2- I would have to manually update every single hyperlink each time I update my FM book, because I am single sourcing out of Frame and I am unable (as far as I can tell) to link a PDF within the frame doc. I update the entire book (over 1500 pages) once every 6 weeks so while this wouldn't be a common occurrence it will happen regularly, and it would be manual (as far as I can tell)?
    3- Eventually, I would have countless PDFs inside RH. I assume this will eventually impact performance. So this also doesn't seem ideal?
    If anyone has thoughts/suggestions on a simpler way or better way to do this, I'd certainly appreciate it. I have watched the Adobe TV tutorial on adding a master page but that seems to remove the ability to use a css across all my topics and it also requires the manual addition of a manual hyperlink to the PDF file, so that is what I am proposing above, anyway (not sure the benefit, therefore).
    Thanks in advance,
    Adriana

    Anything other than CTRL + P is going to create a lot of work so perhaps I can comment on what you see as drawbacks to that.
    a)that cuts off the larger images and b)it doesn't show header, footer,
    logo, date, etc. (stuff that is in the master pages in FM).
    Larger images.
    I simply make a point of keeping my image sizes down to a size that works. It's not a problem for me but that doesn't mean it will work for you. Here all I am doing is suggesting you review how big a problem that would be.
    Master Page Details
    I have to preface this with the statement that I don't work with FM. The details you refer to print when they are in RoboHelp master pages. Perhaps one of the FM users here can comment on how to get FM master pages to come through.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • Best practice for making changes to Oracle apps business views and BAs/fold

    HI
    The oracle BI solution comes with pre-defined Business Views- database views and Business Areas and folders. If we want to customize those database views or BAs and folders what will be the best practice in order to avoid losing it during any upgrades.
    For ex Oracle out-of box Order Management BA that we are using heavily needs some additional fields to be added to Order Header and Order Lines folders and we also want to add some custom folders to this BA.
    If we do the changes to the database views behind this BA would they be lost during the upgrade or do we have to copy(duplicate) those views, updated them and create a custom BA and folders against those views.
    Thanks

    Hi,
    If you are adding new folders then just add them to the Oracle Business Area. The business area is just a collection of folders. If the business area was changed in an upgrade the new folder would not be deleted.
    If you want to add fields to the existing folders/views then you have 2 options. Add the field to the defining base view (these are the views beginning OEBV and OEFV) and then regenerate the business views. This may be overwritten if the view is upgrade but this is unlikely.
    Alternatively, copy the view to create a new version and then map the old folder to the new view and refresh. You may need to re-map the folder if the folder is upgraded, but at least you have a single folder used by both Oracle and custom reports.
    Rod West

  • Best Practices for Multiple Forms-Reports Instances (WebLogic) on Win2008R2

    Hello all,
    I’ve succeeded in creating two instances of Forms/Report (FR) in WebLogic and am looking at about 5 or 6 FR instances on a one Windows Server 2008R2 box. I understand each instances will have its own folder (FR_Inst1) structure under the Middleware folder. Currently I have two instances configured each with its own Home (FR_Home1 & FR_Home2) and Domain (FR_Domain1 & FR_Domain2). Both the of the separate FP applications function correctly.
    Can multiple similar type instances share a single Home (FR_Home)? Can multiple similar type instances share a single Domain outside of the WebLogic Domain, such as FR_Domain?
    Thanks,
    Ron

    Thanx for the reply! I read through the documents and they are very good at explaining how to install the different components individually. I still can't find much on installing them together. I hope it's not just going to be a trial and error thing.
    So far I've installed done the following successfully:
    Installed 10.3.5 weblogic
    Forms and Reports 11g on top of 10.3.5
    I've created an additional managed server for our ADF applications.
    My next step is upgrading the JSF to 2.x. I would have to stage patches 12917525 and 12979653. I'm afraid it will break the forms and reports though. Any ideas?

  • Best practice for making database connection to Forms 10 apps?

    Hi
    To upgrade our Forms applications we are moving from version 3 to 10.
    Our old system runs Forms applications and the connection to the database is based on the individual user. This means that any tables or views used require that the user has specific access granted to them. We have a bespoke system to manage this which generates scripts (GRANT statements) based on lists of tables and users and their appropriate access.
    I have concerns that managing the table access for thousands of individual users in the Forms 10 environment is going to be technically difficult, especially with RADs to consider. Is it feasible to generate and frequently refresh RAD scripts to maintain the current list of users and their permissions?
    I am trying to decide if it is better to:
    A) Connect with the same database user (such as "APP_USER") which has access to everything
    or
    B) Connect with individual usernames/passwords
    Currently, the individual user database passwords are generated weekly and users have means to obtain them (once signed in) rather than setting and remembering them. Some views refer to the Oracle system parameter "USER" to decide what data is returned so this functionality would need to be preserved.
    Any help is greatly appreciated, especially if you can tell me if option A or B is how you connect at your site.

    Thanks for the advice so far.
    It would appear that connecting with individual usernames is not a fundamental error, which I was concerned about.
    Will it still be necessary to create and refresh RAD scripts, or is this only an issue when using OID? We have OID here already because we have a website using Oracle Portal. The sign-on process for this connects to Active Directory for authentication.
    I do not like the idea of having to schedule a refresh of RAD scripts, perhaps 3 times a day, just to keep it current. I do not think the RADs are expected to change as frequently as this, but perhaps other forum members have experience of this?

  • Best practice for large form input data.

    I'm developing an HIE system with FLEX. I'm looking for a strategy, and management layout for pretty close to 20-30 TextInput/Combobox/Grid controls.
    What is the best way to present this in FLEX without a lot of cluter, and a given panel being way too large and unweildy.
    The options that I have come up with so far.
    1) Use lots of Tabs combined with lots of accordions, and split panes.
    2) Use popup windows
    3) Use panels that appear, capture the data, then, make the panel go away.
    I'm shying away from the popup windows, as, that strategy ALWAYS result in performance issues.
    Any help is greatly appreciated.
    Thanks.

    In general the Flex navigator containers are the way to go. ViewStack is probably the most versatile, though TabNavigator and Accordion are good. It all depends on your assumed workflow.
    If this post answers your question or helps, please mark it as such.

  • Best practice for making a mirror of the system disk

    I've got an HP ML115 machine that I'm installing Solaris 10 on. It has four disks: a 160GB (boot drive), a 300GB spare, and two 1.5TB drives. I've made a mirrored pool of the two 1.5TB drives. I'd like to create a mirror of the 160GB boot drive (I realise I'll lose some space). However, I'm unsure of how to do this. Attempting to create a mirror (zpool create -f bootmirror c0t0d0 c0t1d0) tells me that the partitions are mounted.
    Any ideas on what the best way to achieve this is, please?

    What filesystem type is root currently UFS or ZFS.
    If its currently UFS, you will have to use SVM for mirroring. A UFS root can't be converted to a ZFS root.
    If its already a ZFS root just not mirrored, it should be simple. Just add the disk to the pool to "promote" it to mirrored.
    You do that with a zpool attach command.

  • Best practices for making space on hard drive?

    My relatively trusty ol' 466 mhz G4 is bogging down. I am showing 9.77 GB out of 28.6 capacity. I have the Adobe CS programs and do a lot of Photoshop work. I decided to clean house and dump as much stuff as I can, including the older programs (Adobe Design Suite, Virtual PC, etc.) I don't use these programs anymore, and have the original CD's, so assume there is no reason to take up space with them. My questions are:
    1- Is there a general rule for optimal functioning of a Mac in terms of the % of HD space that should be available?
    2- What is the difference between regular 'empty trash' and 'secure empty trash'?
    3- I bought the LaCie for backing up the hard drives and files. Am using iMSafe, but am a little spooked by the warmings that "removing obsolete files from the destination could result in data loss". Any suggestions?
    PowerMac G4 Digital Audio 466 mhz   Mac OS X (10.3.9)   896 MB SDRAM, 30GB HD, 75 GB additional int. HD, and a 120 GB LaCie external HD

    10-15% free HD space. So having 25-30% free should not pose a problem?
    That should be fine for now, but you know how quickly space can disappear, and the less space you have the slower the system will become.
    Paranoid that something trashed will be found? Or that something trashed will be lost, then needed?
    Something trashed will be found. If you're using your computer in an everyday way and you delete something you later decide you really want, even using the most sophisticated data recovery programs you can only find stuff that hasn't already been written over.
    How does a backup program (such as iMSafe) decide which files are "obsolete"? I hardly know myself! So how do I know if it is OK to trust the program to decide?
    Personally, I don't like the fact that it does this, but I believe it is simply comparing the original with the copy. If you've deleted the original, then it wants to know if the backup file is now obsolete.
    I would suggest you consider something easier to use. Have a look at ChronoSync and see what you think.

  • [Ask] Best practice for making pools with CAM.

    Dear all,
    there's a few things I want to ask:
    Condition:
    - I have 48 disk with 300GB each and want to make some pool that contains /ctl , /archive , /data , /backup , /redo
    - i use software common array manager 6.9
    - based on the information I got, when use more than 1 pool, when the data is processed is faster then if only make 1 pool.
    Question:
    - which/what is the best practise when the data is spread out in several pool? my target is iops < 4ms
    - How many pool do I make?
    - How many hard drive is set in 1 pool?
    Thank you in advance.
    Rey
    Edited by: 952840 on Jan 3, 2013 11:41 PM

    Hi,
    If you are adding new folders then just add them to the Oracle Business Area. The business area is just a collection of folders. If the business area was changed in an upgrade the new folder would not be deleted.
    If you want to add fields to the existing folders/views then you have 2 options. Add the field to the defining base view (these are the views beginning OEBV and OEFV) and then regenerate the business views. This may be overwritten if the view is upgrade but this is unlikely.
    Alternatively, copy the view to create a new version and then map the old folder to the new view and refresh. You may need to re-map the folder if the folder is upgraded, but at least you have a single folder used by both Oracle and custom reports.
    Rod West

  • Best practices for Flash forms

    This question was posted in response to the following article: http://help.adobe.com/en_US/ColdFusion/9.0/Developing/WSc3ff6d0ea77859461172e0811cbec10372 -7fed.html

    Hi,
    I can say the "Recompliation" is nothing but Refreshing the
    Browser window where the SWF is again generated which is considered
    to be the expensive process...
    So to overcome this, The "CFC's" are come into play which
    eventually communicates with the CF Server and returning the
    results back to the UI behind the scenes..

Maybe you are looking for