Tips and best practices for translating C into LabVIEW? SERIOUS newbie...

I need to translate a C function into LabVIEW.  This will be my *first* LabVIEW project.  I've been reading some tutorials, and I'm still struggling to get my brain out of "C/C++ mode" and learn the LabVIEW paradigms.
Structurally, the function that I need to translate gets called from a while-loop and performs a bunch of mathematical calculations. 
The basic layout is something like this (this obviously isn't the actual code, it just illustrates the general flow control and techniques that it uses).
struct Params
// About 20 int and float parameters
int CalculateMetrics(Params *pParams,
float input1, float input2 [etc])
int errorCode = 0;
float metric1;
float metric2;
float metric3;
// Do some math like:
metric1 = input1 * (pParams->someParam - 5);
metric2 = metric1 + (input2 / pParams->someOtherParam);
// Tons more simple math
// A couple for-loops
if (metric1 < metric2)
// manipulate metric1 somehow
else
// set some kind of error code
errorCode = ...;
if (!errorCode)
metric3 = metric1 + pow(metric2, 3);
// More math...
// etc...
  // update some external global metrics variables  
return errorCode;
I'm still too green to understand whether or not a function like this can translate cleanly from C to LabVIEW, or whether the LabVIEW version will have significant structural differences. 
Are there any general tips or "best practices" for this kind of task?
Here are some more specific questions:
Most of the LabVIEW examples that I've seen (at least at the beginner level) seem to heavily rely on using the front panel controls  to provide inputs to functions.  How do I build a VI where the input arguments(input1, input2, etc) come as numbers, and aren't tied to dials or buttons on the front panel?
The structure of the C function seems to rely heavily on the use of stack variables like metric1 and metric2 in order to perform calculations.  It seems like creating temporary "stack" variables in LabVIEW is possible, but frowned upon.  Is it possible to keep this general structure in the LabVIEW VI without making the code a mess?
Thanks guys!

There's already a couple of good answers, but to add to #1:
You're clearly looking for a typical C-function. Any VI that doesn't require front panel opening (user interaction) can be such a function.
If the front panel is never opened the controls are merely used to send data to the VI, much like (identical to) the declaration of a C-function. The indicators can/will be return values.
Which controls and indicators are used to sending data in and out of a VI is almost too easy; Click the icon of the front panel (top right) and show connector, click which control/indicator goes where. Done. That's your functions declaration.
Basically one function is one VI, although you might want to split it even further, dont create 3k*3k pixel diagrams.
Depending on the amount of calculations done in your If-Thens they might be sub vi's of their own.
/Y
LabVIEW 8.2 - 2014
"Only dead fish swim downstream" - "My life for Kudos!" - "Dumb people repeat old mistakes - smart ones create new ones."
G# - Free award winning reference based OOP for LV

Similar Messages

  • Can anyone recommend tips and best practices for FrameMaker-to-RoboHelp migration ?

    Hi. I'm planning a migration from FM (unstructured) to RH. I'd appreciate any tips and best practices for the migration process. (Note that at the moment I plan to import the FM documents into, not link to them from, RH.)
    For example, my current FM files are presently not optimally "chunked", so that autoconverting FM file sections (based on, say, Header 1 paragraph layout) won't always result in an optimal topic set. I'm thinking of going through the FM docs and inserting dummy paragraphs with a tag somethike like "topic_break", placed in more appropriate locations that the existing headers. Then, during import to RH, I'd use the topic_break paragraph to demark the topics. Is this a good technique? Beyond paragraph-based import delineation, do you know of any guidelines for redrafting FM chapter file content into RH topics?
    Also, are there any considerations/gotchas in the areas of text review workflow, multiple authoring, etc. after the migration? (I've not managed an ongoing RH doc project before, so any advice would be greatly appreciated.
    Thanks in advance!
    -Kurt
    BTW, the main reason for the migration: Info is presently scattered in various (and way to many) PDF files. There's no global index. I'd like to make a RoboHelp HTML interface (probably WebHelp layout) so it can be a one-stop documentation shop for users.

    Jeff
    Fm may produce better output for your requirements but for many what Rh produces works just fine. My recent finding re Word converting images to JPG before import will mean a better experience for many.
    Once Rh is set up, and it's not difficult, for many its printed documents will do the job. I would say try it and then judge.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • Best practices for translating FrameMaker content?

    I'm interested in learning from Adobe or experienced users about best practices to follow (or references worth reading) about translating FrameMaker publications.

    Haven't done this for a while, but I believe the first recommendation used to be "work with a reputable translation agency who will accept, translate and return .mif files." fwiw, I still remember a run-in with one agency that claimed they did but didn't … they passed our .mif files on to translators who used Word =8-} The result in our case was that all the autonumbering in a long, complex regulatory document failed, because Word had slipped a fine-space character in before every colon in the French translation.
    A separate issue is whether it's worth moving to structured Frame, which is said to make translation management easier as your document evolves.
    There'll be lots of people who can give you more complete, more up-to-date answers, though.

  • BASELINE PACKAGE - V1-V2.603 and Best Practices for Pharmaceuticals

    Hi All,
             > Recently we have upgraded to EHP4 stack 4
             > I am trying to install BASELINE PACKAGE - V1-V2.603 I chose that because its localized for India.
             > When i am trying to install  BP-ERP05 603V7 and BP-INSTASS 600V7 its asking for BBPCRM 600, BBPCRM
                700 and SAP_APPL 600, SAP_APPL 603.
             > Why does the installer ask for lower versions? ofcourse i have read that if the version is Higher than the one mentioned
                in the "Quick Guide to Installing the SAP Best Practices Baseline Package (IN)" will not work.
             > But do we have any BaseLine Packages Specific to EHP4?
             > If not, then could any one tell me where and how to download and install the BBPCRM as an add-on.
             > I only need that Software component because I strongly feel that all the interdependencies are linked with this one only.
    Any help and suggestions are welcome.
    Regards,
    Antony Chaitanya.

    Hi Sunny,
                    Thanks very much for your response.
    The major problem that i am having is I some how did not include the BBPCRM software component or the related software components for CRM at the time of upgrade is my best guess.
    So, the latest add ons BP-CRM60 600V3 and BP-CRM70 700V1 or the add ons BP-ERP05 600VD/ 603V7 along with that BP-INSTASS 600V7 are not able to install because the prerequisite are not meeting. Only BP-INSTASS 600V1 got installed.
    Can any one tell me how and what to do now.... plzzz???
    Things like how to get the CRM related s/w components installed or if there is a work around to get the Baseline Packages (IN) activated with out the CRM.
    Regards,
    Antony Chaitanya.

  • UI Defaults Import problems and best practice for using UI defaults

    Hi,
    I'm hoping some of you experts can tell me how to Import the UI Defaults from Schema A and Workspace A to Schema B and Workspace B. My goal is to fluidly share the UI defaults with another developer as we march forward creating new versions of the application (inside new Workspaces with new schema names). My longer term goal was to create a "library" of UI defaults that other future developers could easily access as well.
    The basic setup I have been using is that each developer gets their own schema and their own workspace. When a new version of the database design is made, I keep the older versions of the workspaces and schemas intact and then create a new workspace and new schema for each developer (just two of us right now). Then we Export apps, respectively, from our old workspaces and Import apps into our new workspaces. I also create a new db tablespace and new db user account and map to new workspace as well. This way I always know what db design version we are using against what Apex app version and can, conversely, delete tablespaces, user accounts, workspaces and Apex schemas in one swallow without worrying about getting the wrong version of something accidentally deleted. 9It may seem like I am doing a lot of extra work, but my sanity is still (relatively) intact.)
    I keep getting
    "ORA-20001: GET_BLOCK Error. ORA-20001: Execution of the statement was unsuccessful. ORA-20001: You do not have access to the schema that you are importing. Import failed. &lt;pre&gt;begin wwv_flow_api.g_id_offset := 0; wwv_flow_hint.g_schema := 'MY_OTHER_SCHEMA'; wwv_flow_hint.check_schema_privs; end;&lt;/pre&gt;"
    when I try importing it.
    We want to share our UI defaults that we each set up in our respective workspaces and schemas. We do have some overlap of tables (which may be a no no I suppose). We also each have differently named (test) apps.
    So, like app Import/Export, I thought I could just Export my app (along with UI) from Schema A and Workspace A and simply import it into Schema B and Workspace B and then in the new Workspace have access to the UI. I tried editing the UI *.SQL using the app ID and new schema name but it still balks. It seems like there is an internal table ID code specific to each table that is particular to the table and the workspace it was created in.
    I can go to a specific workspace, create a sample UI, Export it's app, and Import the same app (new ID) into same workspace and the UI shows up fine.
    So, what good is the UI defaults if you can't carry them forward to the next version of your app that is in a new workspace? It seems like we are forced into creating the equivalent of one workspace, say the "Library", and both of us with separate schemas sign into here. But then how do I get the UIs from the "Library" to our "Beta" workspace, and similarly from there to 'Production' without recreating UIs each time? And, if I prefer a sample set of records, say only 100 test recs, and the other developer prefers a full set, say 30,000 recs, and we have to use the same workspace, since the workspace can only be associated with one db tablespace, how can I load both 100 and 30,000 recs into same area? (and since we can't, that is another reason we wanted to use separate workspaces and pass our respective UIs back and forth and finally into the shared "Library" area.)
    Well, thank for reading through this and hopefully it makes sense.

    OK Scott, let me try to clarify here:
    Regardless of what reasons I have, my immediate goal is to sucessfully Export the UI Defaults from Schema A and Workspace A and Import into Schema B and Workspace B.
    So, logging in as Schema A into Workspace A, I go to my app, then Export/Import then "Export User Interface Defaults". There is only one schema available to choose from in the "Schema" dropdown select list. I export the UI and save it to a file.
    Next I logout and then login as Schema B into Workspace B. I go to Shared Components UI Defaults and display my tables. Most of these tables (about 30 of them) have the same exact name as the tables used by the other schema. UI Defaults do not exist.
    I choose "Import" in the Tasks list, File Type "User Interface Defaults" and pick file I just exported. File is successfully Imported. Now I'm on screen for installing the UI Defaults. When I try to Import the UI it brings me to "Confirm Replace UI Defaults" and lists the OLD workspace schema (not the new one I expected)as being removed and replaced. I push ahead, tell it to do the replace and get the "You do not have access to the schema that you are importing". So I guess I'm doing something wrong.

  • Tabular Modeling. What is the best practice for importing data into VS to limit the records in the designer?

    Should I wrap the queries in a procedure with a @StartDate and @EndDate and create a test partition to pass a small date range? 
    Or can i use the Table properties screen to put the command there and will it run and not be affected or affect the partitions?This would be nice if this SQL statement on this screen was independent of the partitions and I could just leave it with the the
    command text = EXEC TransactionDetail '2014-01-01', '2014-05-31' Especially since if you have many tables that load based on a date range. i would not want to jump in and change that query on all of them.
    Is there a a way to have a parameter in the project so all tables would get the same @startDate and @EndDate so I could change it in one place?
    And I am not stuck to these questions\options, If there is a better way to mass change the queries to run a subset of data for the designer I'd like to hear it.
    Thank You,
    Phil

    Hi Phil,
    According to your description, you are looking for the best way to control the rows that are loaded into a table, right?
    When importing data to a table of tabular model, we can apply filters to control the rows that are loaded into a table. After you have imported the data, you cannot delete individual rows. However, you can apply custom filters to control the way that
    rows are displayed. Rows that do not meet the filtering criteria are hidden. For the detail information about it, please refer to the link below.Filter Data in a Table (SSAS Tabular)
    If I have anything misunderstand, please point it out.
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here
    Charlie Liao
    TechNet Community Support

  • How do I prepare my iMac for sale? And best practice for transferring media to a new system.

    This is my first time posting in a forum, so I will apologize in advance if I'm doing something wrong. I've tried to search for a very clear answer to my questions in the existing discussions, but have not had much luck.
    I ordered a Macbook Pro today and would like to sell my current iMac to supplement some of the cost. I purchased my iMac in March of 2011 with OS X Snow Leopard preinstalled. I've since updated to Mountain Lion. When my computer came in it did not have OS DVDs, so I am not sure exactly how to take it back to factory default and remove ALL of my information safely in order to prepare it for sale. I'm afraid to tell it to format if I am going to need a boot disc/drive to reinstall the OS. I've been trying to research all day and have learned that I will not be able to sell it with the current OS, only the one that came with the computer at the time of purchase (to the best of my understanding).
    Secondary to all of this, I have a question about backing up info and porting it to my new system. I have run the Time Machine backup to my external hard drive. However, I DO NOT want to port it to my new MBP as an image of my past machine. I did that when I got my iMac from my old MBP and my iMac thinks it is a MBP (MBP EFI/SMC Firmware Update Icons in my system utilities folder as well as identified as a Macbook Pro on my network). I've never been able to get all of these things to go away and I feel like it's chewing up hard drive space for no purpose whatsoever. I also worry that it could create system issues that I'm just not savy enough to catch before they are beyond control. That being said, I'd like to begin fresh with my new MBP. What is the best method of bringing my photos from iPhoto and music/movies/books/etc from iTunes to my new system? My word documents are easy enough...I keep 3 flash drives updated with them at all times.
    Thanks in advance!

    Whoever said you can't sell you iMac with the current Mac OS X is wrong, I believe.
    Unfortunately, it is not wrong.
    Any OS purchased/obtained at the app store - as is the case here - is not transferable. The license is tied forever to the Apple ID used to obtain it. The SLA states that the seller needs to erase the drive and reinstall the original system.
    @ Kamikazwe:
    You can call Apple and give them your serial number and they will send you copies of your original disks at a nominal charge.
    You do need to use your original Snow Leopard disk, boot from it, erase the drive, and reinstall Snow Leopard. The buyer of the machine will need to purchase his/her own copy of Lion or Mountain Lion with their own Apple ID or they will never be able to reinstall or update the OS.
    SLA Excerpt:
    3. Transfer.
    A. If you obtained the Apple Software preinstalled on Apple-branded hardware, you may make a one-time permanent transfer of all of your license rights to the Apple Software (in its original form as provided by Apple) to another party, provided that:  the Apple Software is transferred
    together with your Apple-branded hardware; (ii) the transfer must include all of the Apple Software, including all its component parts, printed materials and this License; (iii) you do not retain any copies of the Apple Software, full or partial, including copies stored on a computer or
    other storage device; and (iv) the party receiving the Apple Software accepts the terms andcconditions of this License. For purposes of this License, if Apple provides an update (e.g., version 10.8 to 10.8.1) to the Apple Software, the update is considered part of the Apple
    Software and may not be transferred separately from the pre-update version of the Apple Software.
    B. If you obtained your license to the Apple Software from the Mac App Store, it is not transferable. If you sell your Apple-branded hardware to a third party, you must remove the Apple Software from the Apple-branded hardware before doing so, and you may restore your system to the version of the Apple operating system software that originally came with your Apple hardware (the “Original Apple OS”) and permanently transfer the Original Apple OS together with your Apple hardware, provided that:  the transfer must include all of the Original
    Apple OS, including all its component parts, printed materials and its license; (ii) you do not retain any copies of the Original Apple OS, full or partial, including copies stored on a computer or other storage device; and (iii) the party receiving the Original Apple OS reads and agrees to
    accept the terms and conditions of the Original Apple OS license.

  • Best Practices for AD Import into FIM MV

     I'm working on a project which will require AD users to be imported into FIM and then be shown in the FIM portal. I know I could import users from a specific OU, but what I'm not sure about is how FIM and the MV keeps track of users.
    For example:
    I configure "john.doe" in OU "London" and then import the account into the MV.
    john.doe is moved into the Hong Kong OU and I run a delta sync - can FIM track this change?
    Would I be better off using an anchor on objectsid to deal with user account moves and users changing names\logons?
    Thanks

    When you are configuring AD MA you are selecting OUs in which you are managing object through MA. As long as you will keep user in selected OUs FIM will "see" this user and will import changes on object.
    FIM is using objectGUID as anchor internally (I assume this - never looked into agent code :) ) but presents DN as an anchor for you. You can't change this. When user will be moved between OUs you will see it as rename operation and you will see it as DN
    change.
    Tomek Onyszko, memberOf Predica FIM Team (http://www.predica.pl), IdAM knowledge provider @ http://blog.predica.pl

  • What is the guideline and/or best practice for EMC setup on ASM?

    We are going to use EMC CX4-480 for ASM storage on RAC. What is the guideline and best practice for EMC setup on ASM?
    Thanks for the advice!

    Probably a poor choice of words. Sorry.
    So basically, I have gotten further, but I just noticed related problem.
    I'm using the WebServices(WS) 1.0. I insert an account, then, on a separate WS call, I insert my contacts for the account. I include the AccountID, and a user defined key from the Account when creating the Contact.
    When I look at my Contact on the CRMOD web page, it shows the appropriate links back to the Account. But when I look at my Account on the CRMOD web page, it shows no Contacts.
    So when I say workflow or Best Practice, I was hoping for guidance on how to properly write my code to accomplish all of the necessary steps. As in this is how you insert an account with a contact(s) and it updates the appropriate IDs so that it shows up properly on the CRMOD web pages.
    Based on the above, it looks like I need to, as the next step, take the ContactID and update the Account with it so that their is a bi-directional link.
    I'm thinking there is a better way in doing this.
    Here is my psuedocode:
    AccountInsert()
    AccountID = NewAcctRec
    ContactInsert(NewAcctRec)
    ContactID = NewContRec
    AccountUpdate(NewContRec)
    Thanks,

  • Best practice for a deplomyent (EAR containing WAR/EJB) in a productive environment

    Hi there,
    I'm looking for some hints regarding to the best practice deployment in a productive
    environment (currently we are not using a WLS-cluster);
    We are using ANT for buildung, packaging and (dynamic) deployment (via weblogic.Deployer)
    on the development environment and this works fine (in the meantime);
    For my point of view, I would like to prefere this kind of Deploment not only
    for the development, also for the productive system.
    But I found some hints in some books, and this guys prefere the static deployment
    for the p-system.
    My question now:
    Could anybody provide me with some links to some whitepapers regarding best practice
    for a deployment into a p-system ??
    What is your experiance with the new two-phase-deploment coming up with WLS 7.0
    Is it really a good idea to use the static deployment (what is the advantage of
    this kind of deployment ???
    THX in advanced
    -Martin

    Hi Siva,
    What best practise are you looking for ? If you can be specific on your question we could provide appropriate response.
    From my basis experience some of the best practices.
    1) Productive landscape should have high availability to business. For this you may setup DR or HA or both.
    2) It should have backup configured for which restore has been already tested
    3) It should have all the monitoring setup viz application, OS and DB
    4) Productive client should not be modifiable
    5) Users in Production landscape should have appropriate authorization based on SOD. There should not be any SOD conflicts
    6) Transport to Production should be highly controlled. Any transport to Production should be moved only with appropriate Change Board approvals.
    7) Relevant Database and OS security parameters should be tested before golive and enabled
    8) Pre-Golive , Post Golive should have been performed on Production system
    9) EWA should be configured atleast for Production system
    10) Production system availability using DR should have been tested
    Hope this helps.
    Regards,
    Deepak Kori

  • In the Begining it's Flat Files - Best Practice for Getting Flat File Data

    I probably should have posed this question here before I delved into writing Java to get data for reports, but better late than never.
    Our ERP is written in COBOL. We have a third party ODBC which allows us to access data using a version of SQL. I have several Java sources compiled in my database that access the data and return something relevant. The Java sources are written in a procedural style rather than taking advantage of object oriented programming with attributes and methods.
    Now that I am becoming more comfortable with the Java language, I would greatly appreciate any feedback as to best practices for incorporating Java into my database.
    My guess is that it would be helpful to model the ERP "tables" with Java classes that would have attributes, which correspond to the fields, and methods to return the attributes in an appropriate way. Does that sound reasonable? If so, is there a way to automate the task of modeling the tables? If not reasonable, what would you recommend?
    Thanks,
    Gregory

    Brother wrote:
    I probably should have posed this question here before I delved into writing Java to get data for reports, but better late than never.
    Our ERP is written in COBOL. We have a third party ODBC which allows us to access data using a version of SQL. I have several Java sources compiled in my database that access the data and return something relevant. The Java sources are written in a procedural style rather than taking advantage of object oriented programming with attributes and methods.
    OO is a choice not a mandate. Using Java in a procedural way is certainly not ideal but given that it is existing code I would look more into whether is well written procedural code rather than looking at the lack of OO.
    Now that I am becoming more comfortable with the Java language, I would greatly appreciate any feedback as to best practices for incorporating Java into my database.
    My guess is that it would be helpful to model the ERP "tables" with Java classes that would have attributes, which correspond to the fields, and methods to return the attributes in an appropriate way. Does that sound reasonable? If so, is there a way to automate the task of modeling the tables? If not reasonable, what would you recommend?Normally you create a data model driven by business need. You then implement using whatever means seem expedient in terms of other business constraints to closely model that data model.
    It is often the case that there is a strong correlation between data models and tables but certainly in my experience it is rare when there are not other needs driven by the data model (such as how foreign keys and link tables are implemented and used.)

  • Best practice for including additional DLLs/data files with plug-in

    Hi,
    Let's say I'm writing a plug-in which calls code in additional DLLs, and I want to ship these DLLs as part of the plug-in.  I'd like to know what is considered "best practice" in terms of whether this is ok  (assuming of course that the un-installer is set up to remove them correctly), and if so, where is the best place to put the DLLs.
    Is it considered ok at all to ship additional DLLs, or should I try and statically link everything?
    If it's ok to ship additional DLLs, should I install them in the same folder as the plug-in DLL (e.g. the .8BF or whatever), in a subfolder of the plug-in folder or somewhere else?
    (I have the same question about shipping additional files too, such as data or resource files.)
    Thanks
                             -Matthew

    Brother wrote:
    I probably should have posed this question here before I delved into writing Java to get data for reports, but better late than never.
    Our ERP is written in COBOL. We have a third party ODBC which allows us to access data using a version of SQL. I have several Java sources compiled in my database that access the data and return something relevant. The Java sources are written in a procedural style rather than taking advantage of object oriented programming with attributes and methods.
    OO is a choice not a mandate. Using Java in a procedural way is certainly not ideal but given that it is existing code I would look more into whether is well written procedural code rather than looking at the lack of OO.
    Now that I am becoming more comfortable with the Java language, I would greatly appreciate any feedback as to best practices for incorporating Java into my database.
    My guess is that it would be helpful to model the ERP "tables" with Java classes that would have attributes, which correspond to the fields, and methods to return the attributes in an appropriate way. Does that sound reasonable? If so, is there a way to automate the task of modeling the tables? If not reasonable, what would you recommend?Normally you create a data model driven by business need. You then implement using whatever means seem expedient in terms of other business constraints to closely model that data model.
    It is often the case that there is a strong correlation between data models and tables but certainly in my experience it is rare when there are not other needs driven by the data model (such as how foreign keys and link tables are implemented and used.)

  • Best practice for responsive projects

    Does anyone have tips on best practices for responsive project?
    I understand that 3 different layouts can be created. What happens if a learner is not using one of the 3 devices that were set up in a responsive project, and their screen size is different from any of those

    Jay,
    Dr. Pooja Jaisingh offered very valuable tips for good practice in responsive design last week in her webinar. 'Do's and Don'ts of creating Responsive Projects with Captivate 8'. I don't see the recording yet On Demand, but keep an eye on it.
    Did you test a responsive project with F11 (Preview in Browser)? You will be able to change the resolution of the browser window and see that the content, if well designed (you can have absolute positioning, size as well) will move, shrink to adapt. The break points (3 devices) allow you to make more invasive changes at those points: dragging some objects out of the stage in the scratch area because they take up too much space for phones is one example. Or replacing a big screenshot with many details by a zoomed in detail screenshot for the mobile breakpoint. That is my way of explaining, responsive is not just have the three layouts for devices, it is also adapting between those breakpoints.

  • Best practice for application help for a custom screen?

    Hi,
    The system is Netweaver 7.0 SP 15 with e-recruiting .
    We have some custom SAP GUI transactions and have written Word documents with screen prints and explanations. I would like to make the procedure document accessible from the custom transaction or at least provide custom help text that includes a link to the full documents.
    Can anyone help me out with options and best practices for providing customized application help for custom SAP GUI transactions?
    Thanks,
    Margaret

    Hello Margaret,
    sorry I though you might be still in a design or proof of concept phase where the decision for the technology is still adjustable.
    If the implementation is already done things change of course. The standard in-system documentation is surely not fitting your needs as including screenshots won't work well.
    I would solve the task the following way:
    I'd make a web or pdf document out of the word document and put it on a web ressource - as you run e-recruiting you have probably the possibility for that.
    I would then just put a button into the transaction an open a web container to show the document.
    I am not sure if this solution really qualifies as "best practise" but SAP does the same if you call the Help for application in the help menue. This is implemented in function module SAPGUIHC_OPEN_HELP_CENTER. I'd just copy it, throw out what I do not need and hard code the url to call.
    Perhaps someone could offer a better solution but I think this works a t least without exxagerated costs.
    Kind Regards
    Roman

  • What is the best practice for creating master pages and styles with translated text?

    I format translated text all the time for my company. I want to create a set of master pages and styles for each language and then import those styles into future translated documents. That way, the formatting can be done quickly and easily.
    What are the best practices for doing this? As a company this has been tried in the past, but without success. I'd like to know what other people are doing in this regard.
    Thank you!

    I create a master template that is usually void of content, with the exception I define as many of the paragraph styles I believe can/will be used with examples of their use in the body of the document--a style guide for that client. When beginning a new document for that client, I import those styles from the paragraph styles panel.
    Exception to this is when in a rush I begin documentation first, then begin new work. Then in the new work, I still pull in those defined paragraph and or object styles via their panels into the new work.
    There are times I need new styles. If they have broader applicability than a one-off instance or publication, then I open the style template for that client and import that style(s) from the publication containing the new style(s) and create example paragraphs and usage instructions.
    Take care, Mike

Maybe you are looking for

  • Is there a way to open Canon 5d mkii RAW files with Elements 7?

    I keep getting error messages. I have looked at the updates but wonder if Elements 7 is too old?

  • (still) Unable to install Flash Player update (process hang)

    I haven't been able to install updates for Flash Player for over a year now. Ever since the website changed to the web-install package, i've been unable to update from the download page (http://get.adobe.com/flashplayer/). The file is downloaded succ

  • Downloaded app to wrong computer

    I recently received an email from Apple that I had been gifted an app. Without realizing the consequences, I downloaded the app to my Work computer (WinXP Pro). But all of my Apps and Music are stored on my Home computer. Now if I try to sync to my W

  • Implicit conversion working then not working (ORA-017222)

    I have a table like following: One of the SQL of my application does the following: SELECT C1 FROM MyTable WHERE C1 =

  • How to verify a genuine earpod?

    How Can I make sure that I am buying a original earpod when it's been bought from a retail shop or non Apple store? Are there any serial number verification process? Or are there any special features or packaging techniques(Like the way the earpod ca