What is the best practice for PXI controller, connect to the company network and install antivirus? Special Subnet?

I need your suggestions and common practices. 

Hello TomMex,
Thanks for posting. If what you are looking for are suggestions for how to use your PXI controller in regards to some of the issues you mentioned, then here are my suggestions. For networking purposes, you can consider your PXI controller the same as any other computer; you should be able to connect it to your network just fine and it will be able to see other computers and devices that are on the same subnet. Antivirus software in general should be fine for your system until you want to install new NI software, at which point you may want to disable it to avoid issues during installation. Does this answer your question? Let me know, thanks!
Regards,
Joe S.

Similar Messages

  • Best practice for how to access a set of wsdl and xsd files

    I've recently beeing poking around with the Oracle ESB, which requires a bunch of wsdl and xsd files from HOME/bpel/system/xmllib. What is the best practice for including these files in a BPEL project? It seems like a bad idea to copy all these files into every project that uses the ESB, especially if there are quite a few consumers of the bus. Is there a way I can reference this directory from the project so that the files can just stay in a common place for all the projects that use them?
    Bret

    Hi,
    I created a project (JDeveloper) with local xsd-files and tried to delete and recreate them in the structure pane with references to a version on the application server. After reopening the project I deployed it successfully to the bpel server. The process is working fine, but in the structure pane there is no information about any of the xsds anymore and the payload in the variables there is an exception (problem building schema).
    How does bpel know where to look for the xsd-files and how does the mapping still work?
    This cannot be the way to do it correctly. Do I have a chance to rework an existing project or do I have to rebuild it from scratch in order to have all the references right?
    Thanks for any clue.
    Bette

  • What are some best practices for Effective Sequences on the PS job record?

    Hello all,
    I am currently working on an implementation of PeopleSoft 9.0, and our team has come up against a debate about how to handle effective sequences on the job record. We want to fully grasp and figure out what is the best way to leverage this feature from a functional point of view. I consider it to be a process-related topic, and that we should establish rules for the sequence in which multiple actions are inserted into the job record with a same effective date. I think we then have to train our HR and Payroll staff on how to correctly sequence these transactions.
    My questions therefore are as follows:
    1. Do you agree with how I see it? If not, why, and what is a better way to look at it?
    2. Is there any way PeopleSoft can be leveraged to automate the sequencing of actions if we establish a rule base?
    3. Are there best practice examples or default behavior in PeopleSoft for how we ought to set up our rules about effective sequencing?
    All input is appreciated. Thanks!

    As you probably know by now, many PeopleSoft configuration/data (not transaction) tables are effective dated. This allows you to associate a dated transaction on one day with a specific configuration description, etc for that date and a different configuration description, etc on a different transaction with a different date. Effective dates are part of the key structure of effective dated configuration data. Because effective date is usually the last part of the key structure, it is not possible to maintain history for effective dated values when data for those configuration values changes multiple times in the same day. This is where effective sequences enter the scene. Effective sequences allow you to maintain history regarding changes in configuration data when there are multiple changes in a single day. You don't really choose how to handle effective sequencing. If you have multiple changes to a single setup/configuration record on a single day and that record has an effective sequence, then your only decision is whether or not to maintain that history by adding a new effective sequenced row or updating the existing row. Logic within the PeopleSoft delivered application will either use the last effective sequence for a given day, or the sequence that is stored on the transaction. The value used by the transaction depends on whether the transaction also stores the effective sequence. You don't have to make any implementation design decisions to make this happen. You also don't determine what values or how to sequence transactions. Sequencing is automatic. Each new row for a given effective date gets the next available sequence number. If there is only one row for an effective date, then that transaction will have a sequence number of 0 (zero).

  • What is a best practice for managing a large amount of ever-changing hyperlinks?

    I am moving an 80+ page printed catalog online. We need to add hyperlinks to our Learning Management System courses to each reference of a class - there are 100s of them. I'm having difficulty understanding what the best practice is for consistent results when I need to go back and edit (which we will have to do regularly).
    These seem like my options:
    Link the actual text - sometimes when I go back to edit the link I can't find it in InDesign but can see it's there when I open up the PDF in Acrobat
    Draw an invisible box over the text and link it - this seems to work better but seems like an extra step
    Do all of the linking in Acrobat
    Am I missing anything?
    Here is the document in case anyone wants to see it so far. For the links that are in there, I used a combination of adding the links in InDesign then perfecting them using Acrobat (removing additional links or correcting others that I couldn't see in InDesign). This part of the process gives me anxiety each month we have to make edits. Nothing seems consistent. Maybe I'm missing something obvious?

    what exatly needs to be edited, the hyperlink or content or?

  • What is tthe Best practice for Variant List, Add, Edit and Display Forms?

    Requirement:
    I have single list.  The list has a large number of columns and a large number of items (lets say 20,000).
    I want to show users a different view of the list based on clicking on a different left-hand navigation option.
    Lets say I have four types of users:  Sales, Manufacturing, Shipping and Finance. I would like to have four options in the left-hand navigation.
    All of them would be pointing at the same list, BUT, I want each of them to have a customer list form.  The only difference between the custom list forms would be:
    Each would have its own set of views, and hence its own default view.
    Each would have its own New, Edit and Display Forms.  The only difference between the forms in one variant list and another would be: The order of the columns and which columns are modifiable.
    I would like to achieve this in SharePoint Designer in such a way that the "users" could still add/modify views and could even modify the forms from the SharePoint Menu.  BTW, I don't want to use InfoPath for obvious reasons.
    What is the best approach to meeting this requirement?  I have at least 20 sites and 70 lists overall that need variant forms made.
    HELP!!
    Savin
    BTW We are using SharePoint 2013 and I selected the wrong forum *sigh*.  But I think its probably the same answer.
    Cheers, Savin Smith

    Hi,
    I understand that you want to have different forms based on different view.
    Per my knowledge, there are no out of box method to achieve it.
    As a workaround, you can add the JavaScript code to the different view page.
    For example, to open different new form based on different view, you can get the windows.location, and then judge the view, then change the onclick event of the "New item" button.
    For more information, you can refer to:
    http://css-tricks.com/snippets/javascript/get-url-and-url-parts-in-javascript/
    http://samsharepoint.wordpress.com/2013/05/01/change-the-default-sharepoint-ok-and-cancel-button/
    Thanks,
     Linda
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Linda Li
    TechNet Community Support

  • What is SAP best practice for SU24 "no check" indicators

    Hi Experts,
    Let's say during testing if we find a t-code needs some authorization objects for end to end execution, and those objects are maintained as "CHECK" "NO" in USOBX_C.
    Please suggest the best practice...
    Should we change the proposal in USOBX_C to "CHECK" "YES" and populate specific values in USOBT_C or we can insert those authorization objects manually in the roles without changing proposals??
    As far as I know, it makes upgrade activities more difficult when there is more customization in these tables....However manual insert of auth objects are not impacted during upgrade and also any SU24 maintenance of custom t-codes.
    Thanks,
    Deb
    Edited by: Julius Bussche on Feb 8, 2012 10:31 AM
    Subject title made more meaingful.

    Normally there is some thought which went into the "no check" flag so you should put some thought into it before testing it to turn it on. It might force you to grant access for that transaction context but have the implication that in other transactions the user can access too much again.
    There are however some authorization objects which were added with support packs with this "no check" in the hope of adding the authority-checks into the code to make them possible, but deliver them as backward compatible with existing roles. This is a work-around for not being able to deliver it as deactivated globally in transaction AUTH_SWITCH_OBJECTS.
    You can find the candidates by sorting the + 12k entries in USOBX_C by object name and finding those which were dealt with in a very liberal way. Check the docs and OSS notes for them.
    Other than that I can only say that a best practice which I believe in is to remove some of the odd things in SU24 immediately after the installation. This means that you later (and as required) only need to add proposals and checks. That is much less error-prone than removing proposals and checks again!
    Cheers,
    Julius

  • What are considered 'best practices' for creating a responsive design, specifically mobile vs. desktop viewing?

    I've stumbled upon a lot of work arounds but I'm trying to decide which is the best way to go.
    I'm creating an image intensive, portfolio style, single page website with text & button based 'calls to action.'
    It features a full width image slider, so resolution is important on desktops, and loading speed important on mobile.
    The built in 'responsive scaling' feature is nice but it has a few major issues.
    - different screen sizes and aspect ratios can make some elements considerably too large or too small, especially text.
    - one version of a file is used for all applications and viewing conditions so I need to compromise quality on desktop and speed on mobile rather than optimize for both.
    - it looks like **** in Safari, for some reason.
    The Edge Commons work around would seem to add to the loading issue while also creating the frustration of seeing the percentage based elements rearranging themselves.
    I'd thought of trying to use media queries to load different animations, but I'm not sure this is doable.

    As for absolute cropping, I agree with your stance. Until the layout is fixed, preserving your ability to easily manipulate photo size and positioning is key.
    Some clever image management methods have been described in the discussion forums, and one that appealed most to me was the use of duplicate linked image folders. Having a high-res (CMYK) folder and a low-res (RGB) folder to switch between for different output enables you to use both to your advantage. Use the low-res images for layout, for internal proofing, and for EPUB/online PDF/HTML output. Then it's simply a quick switch to the high-res image folder for print purposes. You can easily prepare the alternate collection of images with a Photoshop batch convert script or with the Photoshop Image Processor. Save your presets!

  • Best practice for making changes to Oracle apps business views and BAs/fold

    HI
    The oracle BI solution comes with pre-defined Business Views- database views and Business Areas and folders. If we want to customize those database views or BAs and folders what will be the best practice in order to avoid losing it during any upgrades.
    For ex Oracle out-of box Order Management BA that we are using heavily needs some additional fields to be added to Order Header and Order Lines folders and we also want to add some custom folders to this BA.
    If we do the changes to the database views behind this BA would they be lost during the upgrade or do we have to copy(duplicate) those views, updated them and create a custom BA and folders against those views.
    Thanks

    Hi,
    If you are adding new folders then just add them to the Oracle Business Area. The business area is just a collection of folders. If the business area was changed in an upgrade the new folder would not be deleted.
    If you want to add fields to the existing folders/views then you have 2 options. Add the field to the defining base view (these are the views beginning OEBV and OEFV) and then regenerate the business views. This may be overwritten if the view is upgrade but this is unlikely.
    Alternatively, copy the view to create a new version and then map the old folder to the new view and refresh. You may need to re-map the folder if the folder is upgraded, but at least you have a single folder used by both Oracle and custom reports.
    Rod West

  • Best Practices for pulling data out of Documentum (w/Crystal and Universe)

    I was curious to see if we had any best practices to see if we can pull data out of Documentum.
    Thanks
    Ian S

    None, use what works.

  • What is a best practice for counting number of payments in Payroll Run?

    I'm trying to get a feel for what people are using as a means for counting number of payments being created during a payroll run.  I would like to get a total number of checks, direct deposits, wires and zero net checks.  My initial thought is to use the Pre-DME number "Selected and Evaluated Persons".  This number seems to always match the number of persons "Selected" and "Evaluated" in posting to FI/CO.
    Since many companies are restricting use of SE16 and access to many of the BT tables is locked down, I would like to use something that the end user has access to and can easily use.  I don't believe the spools from RFFOUS_C and T are appropriate since it would have overflow pages (2 page REM Statements) in that number and thus would not be an accurate reflection of total number of payments.
    I would appreciate your insight and thoughts on this topic.
    Thanks!

    Hello Jennie,
    This is what we do every payroll:
    1. After running live payroll and before exiting our users run wage type reporter /559 to get the payroll net pay. They also run Postings to FICO simulation run to check payroll net pay GL account. This match all the time. So this is the first pass.
    2. They finish all subsequent activities in PRD run upto RFFOUS_T.
    3. We have developed a custom report which almost looks like REGUH which gives the net pay details by every payment method and amount to be deposited or paid. We match this report from the output we received from step 1.
    4. Run postings to FICO live and check the payroll net pay GL account for final validation.
    The steps might sound tedious but of a great use.
    Arti

  • What is Considered Best Practice for Error Handling in Components?

    What would be the best approach to handling errors in custom (possibly composite) components?
    I see three different possibilities:
    Having the error handling funciton within the component. The problem is that this is not adaptable to the consumers error handling setup (i.e. logging to database, redirecting to a pretty error screen, etc.).
    Passing in an error function as a parameter. The problem here is how to have the externally defined error function interact with external elements that are not defined within the component, i.e. passing the error or fault object to a logging class, or something similar, of which the component knows nothing about.
    Referencing the external error handling methods directly from within the component. (Of course that's shabby programming, just listing as a technical option.)
    How do you solve this?
    My goal is to have a versatile custom component that can utilize any given consumer's error handler.
    Thanks,
    Mike

    or define a customized event and dispatch it when error occur, then you can just listen to it outside

  • What is a best practice for our end users submitting their completed forms?

    I've created an application with fillable fields using Adobe 10. I'm trying to figure out the best way to get the info back. I really don't want them emailing it as it will have sensitive data. I've seen others on this forum referring to using "servers" but I can't seem to find any additional info on this solution. Any advice?

    If the data is sensitive, email is an absolute no-no.
    In addition to being on a server - that is, a web server - you need to have a secure connection.
    That in turn is a web address starting https not just http.
    On the web server, an experienced programmer installs a "script".
    When the form filler clicks SUBMIT, the information from the form is sent to the script.
    The script, in turn, being a program, does whatever the programmer thinks, such as
    - put it in a database
    - check the data and reply immediately
    - email it (very bad idea in this case)
    Typically it's in a database, and you - the person in charge - visits a different script on the
    same server, to get stuff out of the database and process the information.
    I underline experienced. This is no job for a beginner or amateur, or even someone who
    has never worked with web scripts. Bad scripts are the main reason that web sites get
    hacked and details stolen, or your site defaced or otherwise bought into disrepute.

  • What is current best practice for backing up a hard disc?

    With the availability of various cloud applications, I would like to know if backing up an hard disc is now best done to the cloud?  For example, should you buy space in Amazon.com, or somewhere else on the Internet,  and then backup your hard disc to that space? Could you backup an entire harddisc to iCloud?  Your knowledge and experience is greatly appreciated.  Thank you.

    iCloud doesn't provide storage except for pictures
    and your music, and only under limited circumstances.
    Merchant online backup services such as Carbonite
    and others exist for this reason. And you can sync
    your files with/to Dropbox. Many users keep a local
    copy of their files on  Time Machine or other hard
    drives in addition to remote backups (in case of a
    fire or other event compromising the local solution).
    Backup software or cloning programs are useful in
    all the above. You have to work at backup a bit -
    it's like insurance, a bit of expense now to save a
    lot of heartache later.Good to see you're determined
    to forestall potential disaster - too many lack it.
    Good luck, Tom

  • Best Practice for custom controller

    Hi Experts,
    In my current project we have one DC with single component and 12 views (4 business functionality each has 3 view flow ). Now complete logic is in component controller itself and planning to divide business logic into 4 custom controller based on business functionality.
    Please guide me how i need to manger context variable and approach from below?
    1. Maintain context in each custom controller and map to views. (Cust Cont ->Views) not maintained in comp Controller
    2. Maintain all context in component controller and from there custom controller and from there views. (Comp Cont -> Cust cont  -> Views)
    3. Maintain context in custom controller and map to component controller and from there mapping to View. (Cust Cont -> Comp cont  -> Views)
    Regards,
    Satya.

    HI,
    If you are going to use Custom controller, always try to bind all related functionality and context in respective custom contollers.
    Else there is no meaning in using that.When your application will expand in future this will help you.
    Also i will not advice to separate if it is not important.Because if you are creating a new custom controller that much additional work will raise.
    Thanks ,
    Ram

  • Best practice for making database connection to Forms 10 apps?

    Hi
    To upgrade our Forms applications we are moving from version 3 to 10.
    Our old system runs Forms applications and the connection to the database is based on the individual user. This means that any tables or views used require that the user has specific access granted to them. We have a bespoke system to manage this which generates scripts (GRANT statements) based on lists of tables and users and their appropriate access.
    I have concerns that managing the table access for thousands of individual users in the Forms 10 environment is going to be technically difficult, especially with RADs to consider. Is it feasible to generate and frequently refresh RAD scripts to maintain the current list of users and their permissions?
    I am trying to decide if it is better to:
    A) Connect with the same database user (such as "APP_USER") which has access to everything
    or
    B) Connect with individual usernames/passwords
    Currently, the individual user database passwords are generated weekly and users have means to obtain them (once signed in) rather than setting and remembering them. Some views refer to the Oracle system parameter "USER" to decide what data is returned so this functionality would need to be preserved.
    Any help is greatly appreciated, especially if you can tell me if option A or B is how you connect at your site.

    Thanks for the advice so far.
    It would appear that connecting with individual usernames is not a fundamental error, which I was concerned about.
    Will it still be necessary to create and refresh RAD scripts, or is this only an issue when using OID? We have OID here already because we have a website using Oracle Portal. The sign-on process for this connects to Active Directory for authentication.
    I do not like the idea of having to schedule a refresh of RAD scripts, perhaps 3 times a day, just to keep it current. I do not think the RADs are expected to change as frequently as this, but perhaps other forum members have experience of this?

Maybe you are looking for