Tabbing and Data Entry Performance With Forms

Hello,
I am using Adobe Reader 11.0.2 on Windows XP. I've found that there is a lag between the time I hit the tab key to move between form fields and when my cursor actually moves between the form fields. I've also found that there is a lag between the time I type information into a form field and when the information actually appears in the form field. This is especially true with numbers. There have been times when I type numbers and only the first few numbers appear. I waited 30 seconds for the information to appear and it didn't, but as soon as I hit the right arrow on my keyboard, the numbers appear with my cursor at the end.
I've tested the PDF in other readers to determine if it was the file, but I have found that other readers do not have this issue.
Anyone have any thoughts on this?
Thanks,
Will

Hi Sally,
Yes you can define it at infoprovider level as well in planning application,If  you are  using IP then go infoprovider tab in RSPLAN Tcode,there you will see CR tab in lower panel - where you can specify characteristic relatioship and how to derive.
Similarly you can use in BPS as well.
Can refer to the link below for more details for IP:-
http://help.sap.com/saphelp_nw70/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
For BPS:-
http://help.sap.com/saphelp_sem320bw/helpdata/en/ea/39a64281c611d5b3e50050dadfb23f/content.htm

Similar Messages

  • SEM BCS - Block Posting and Data Entry

    Hi Everybody,
    I have a scenario to share with you:
    I do not want to make any postings other that automatic roll up of P&L into rollup accounts like Retained earnings. What will be the effect If I enable the functionality "Block Posting and Data Entry" for these accounts?
    Please share your views.
    Thanks
    RJ

    According to my understanding and experience, this option is useful to refrain the local users from posting on the opening movement type => this opening position should be be calculated by the system during the carry forward only. The exception is the period of the system initialization, i-e the first closing period in BCS, where the opening MT can be impacted by anyone, on all Posting Levels.
    This assumption also means that the central users (i-e users who perform the "top" consolidation activities) shall not be impacted by this option : they are thus allowed to modify this MT during cons group change activities (PL02, 12, 22) and top conso adjustements (PL30), whatever the period.

  • Best performance with forms 4.5 than 6i ?!!!

    Hi,
    Since i'd upgrade my app with forms 6i, i find that when i make a query, this is more slow with 6i than 4.5, with the same db(8i).
    Does anyone else had the same problem?
    What could i do to resolve the problem?
    Thanks.
    null

    See my earlier posting 'Speeding load time of WEB Forms50'. I
    have had no takers yet on a solution to this. There may be some
    hints for you there.
    Good luck on this one - Forms over the WEB seems to be very
    processor and memory hungry. We would like to support about 200
    users on a VERY simple search screen, but we need about 2Gbyte
    of physical memory on the server to get any amount of useful
    speed out of it. Each Forms session takes up about 10Mbyte on
    the server - as soon as this starts swapping out to disk then
    your run-time performance goes right downhill. The earlier-
    mentioned posting gives some hints on speeding up the startup
    time, and it does work.
    Regards,
    Jason Judge
    Natasha Murillo (guest) wrote:
    : How I can Get better performance with Forms 4.5 and 6.0 on the
    : WEB???
    : It is very slow and my customers are not very happy with
    that...
    : Please help me...
    null

  • Transport data entries along with table

    How can we transport data entries along with table?
    Do I need to declare of specific type??
    Thank you,
    AP

    Hi,
    If you have created a table maintenance..
    You can use SM30..
    Press maintain button..
    In the menu..Table view -> transport...Then select the entries include them in the CR.
    If the transport menu is not enabled..
    Go to SE11..
    In the table maintenance generator...Choose the radio button which says "Standard recording routine" and then save..
    Now the transport menu option will be enabled..
    Thanks,
    Naren

  • Data entry sheet with combined info objects?

    Hi,
    we are using Bex workbook as data entry sheet for planning function. It is working fine, however, I have 0cust_sales as parameter. As we know, 0cust_sales is combined with sales org and distribution channel.
    The problem:
    1) requirement is to plan sales org DE21, distribution channel 01, all customer
       input parameters: DE21/01/#
    2) requirement is to plan DE21, all distribution channel, customer 4711
      input parameters: DE21/#/4711
    my question:
    1) why do I have to type the parameter for customer '#' in requirement No.1?
    2) why do I have to type all the combined keys even I have sales org and distribution channel also in the data entry sheet?
    3) are these normal functions? if yes, what is the workaound / solution for this?
    your feedback is highly appreciated!
    Regards,
    Sally

    Hi Sally,
    Yes you can define it at infoprovider level as well in planning application,If  you are  using IP then go infoprovider tab in RSPLAN Tcode,there you will see CR tab in lower panel - where you can specify characteristic relatioship and how to derive.
    Similarly you can use in BPS as well.
    Can refer to the link below for more details for IP:-
    http://help.sap.com/saphelp_nw70/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
    For BPS:-
    http://help.sap.com/saphelp_sem320bw/helpdata/en/ea/39a64281c611d5b3e50050dadfb23f/content.htm

  • QUERY PERFORMANCE AND DATA LOADING PERFORMANCE ISSUES

    WHAT ARE  QUERY PERFORMANCE ISSUES WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES...PLZ URGENT
    WHAT ARE DATALOADING PERFORMANCE ISSUES  WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES PLZ URGENT
    WILL REWARD FULL POINT S
    REGARDS
    GURU

    BW Back end
    Some Tips -
    1)Identify long-running extraction processes on the source system. Extraction processes are performed by several extraction jobs running on the source system. The run-time of these jobs affects the performance. Use transaction code SM37 — Background Processing Job Management — to analyze the run-times of these jobs. If the run-time of data collection jobs lasts for several hours, schedule these jobs to run more frequently. This way, less data is written into update tables for each run and extraction performance increases.
    2)Identify high run-times for ABAP code, especially for user exits. The quality of any custom ABAP programs used in data extraction affects the extraction performance. Use transaction code SE30 — ABAP/4 Run-time Analysis — and then run the analysis for the transaction code RSA3 — Extractor Checker. The system then records the activities of the extraction program so you can review them to identify time-consuming activities. Eliminate those long-running activities or substitute them with alternative program logic.
    3)Identify expensive SQL statements. If database run-time is high for extraction jobs, use transaction code ST05 — Performance Trace. On this screen, select ALEREMOTE user and then select SQL trace to record the SQL statements. Identify the time-consuming sections from the results. If the data-selection times are high on a particular SQL statement, index the DataSource tables to increase the performance of selection (see no. 6 below). While using ST05, make sure that no other extraction job is running with ALEREMOTE user.
    4)Balance loads by distributing processes onto different servers if possible. If your site uses more than one BW application server, distribute the extraction processes to different servers using transaction code SM59 — Maintain RFC Destination. Load balancing is possible only if the extraction program allows the option
    5)Set optimum parameters for data-packet size. Packet size affects the number of data requests to the database. Set the data-packet size to optimum values for an efficient data-extraction mechanism. To find the optimum value, start with a packet size in the range of 50,000 to 100,000 and gradually increase it. At some point, you will reach the threshold at which increasing packet size further does not provide any performance increase. To set the packet size, use transaction code SBIW — BW IMG Menu — on the source system. To set the data load parameters for flat-file uploads, use transaction code RSCUSTV6 in BW.
    6)Build indexes on DataSource tables based on selection criteria. Indexing DataSource tables improves the extraction performance, because it reduces the read times of those tables.
    7)Execute collection jobs in parallel. Like the Business Content extractors, generic extractors have a number of collection jobs to retrieve relevant data from DataSource tables. Scheduling these collection jobs to run in parallel reduces the total extraction time, and they can be scheduled via transaction code SM37 in the source system.
    8). Break up your data selections for InfoPackages and schedule the portions to run in parallel. This parallel upload mechanism sends different portions of the data to BW at the same time, and as a result the total upload time is reduced. You can schedule InfoPackages in the Administrator Workbench.
    You can upload data from a data target (InfoCube and ODS) to another data target within the BW system. While uploading, you can schedule more than one InfoPackage with different selection options in each one. For example, fiscal year or fiscal year period can be used as selection options. Avoid using parallel uploads for high volumes of data if hardware resources are constrained. Each InfoPacket uses one background process (if scheduled to run in the background) or dialog process (if scheduled to run online) of the application server, and too many processes could overwhelm a slow server.
    9). Building secondary indexes on the tables for the selection fields optimizes these tables for reading, reducing extraction time. If your selection fields are not key fields on the table, primary indexes are not much of a help when accessing data. In this case it is better to create secondary indexes with selection fields on the associated table using ABAP Dictionary to improve better selection performance.
    10)Analyze upload times to the PSA and identify long-running uploads. When you extract the data using PSA method, data is written into PSA tables in the BW system. If your data is on the order of tens of millions, consider partitioning these PSA tables for better performance, but pay attention to the partition sizes. Partitioning PSA tables improves data-load performance because it's faster to insert data into smaller database tables. Partitioning also provides increased performance for maintenance of PSA tables — for example, you can delete a portion of data faster. You can set the size of each partition in the PSA parameters screen, in transaction code SPRO or RSCUSTV6, so that BW creates a new partition automatically when a threshold value is reached.
    11)Debug any routines in the transfer and update rules and eliminate single selects from the routines. Using single selects in custom ABAP routines for selecting data from database tables reduces performance considerably. It is better to use buffers and array operations. When you use buffers or array operations, the system reads data from the database tables and stores it in the memory for manipulation, improving performance. If you do not use buffers or array operations, the whole reading process is performed on the database with many table accesses, and performance deteriorates. Also, extensive use of library transformations in the ABAP code reduces performance; since these transformations are not compiled in advance, they are carried out during run-time.
    12)Before uploading a high volume of transaction data into InfoCubes, activate the number-range buffer for dimension IDs. The number-range buffer is a parameter that identifies the number of sequential dimension IDs stored in the memory. If you increase the number range before high-volume data upload, you reduce the number of reads from the dimension tables and hence increase the upload performance. Do not forget to set the number-range values back to their original values after the upload. Use transaction code SNRO to maintain the number range buffer values for InfoCubes.
    13)Drop the indexes before uploading high-volume data into InfoCubes. Regenerate them after the upload. Indexes on InfoCubes are optimized for reading data from the InfoCubes. If the indexes exist during the upload, BW reads the indexes and tries to insert the records according to the indexes, resulting in poor upload performance. You can automate the dropping and regeneration of the indexes through InfoPackage scheduling. You can drop indexes in the Manage InfoCube screen in the Administrator Workbench.
    14)IDoc (intermediate document) archiving improves the extraction and loading performance and can be applied on both BW and R/3 systems. In addition to IDoc archiving, data archiving is available for InfoCubes and ODS objects.
    Hope it Helps
    Chetan
    @CP..

  • I have problems restoring tabs and saving pin apps with Firefox 5 on Mac OS X version 10.6.7

    I bought a Macbook Pro a few months ago and since installing Mozilla Firefox on my system, I have never been able to restore previous sessions after closing Firefox. I currently use Firefox 5 and I've tried pinning open tabs as apps tabs but those too disappear when I restart Firefox.
    I use a few add-ons including DownThemAll, Force-TLS, Rikaichan and Rikaichan's Japanese-English Dictionary File.

    App (pinned) tabs and Tab Groups (Panorama) are stored as part of the session data in the file sessionstore.js in the Firefox profile folder.
    * http://kb.mozillazine.org/Profile_folder_-_Firefox Firefox
    * http://kb.mozillazine.org/sessionstore.js
    * http://kb.mozillazine.org/Session_Restore
    Make sure that you do not use [[Clear Recent History]] to clear the "Browsing History" when Firefox is closed because that prevails and prevents Firefox from opening tabs from the previous session.
    * https://support.mozilla.com/kb/Clear+Recent+History
    It is possible that there is a problem with the files sessionstore.js and sessionstore.bak in the Firefox Profile Folder.
    Delete the files sessionstore.js and sessionstore.bak in the Firefox Profile Folder.
    * Help > Troubleshooting Information > Profile Directory: Open Containing Folder
    * http://kb.mozillazine.org/Profile_folder_-_Firefox
    * http://kb.mozillazine.org/sessionstore.js
    If you see files sessionstore-##.js with a number in the left part of the name like sessionstore-1.js then delete those as well.<br />
    Deleting sessionstore.js will cause App Tabs and Tab Groups and all open tabs to get lost, so you will have to create them again (make a note or bookmark them).
    See:
    * http://kb.mozillazine.org/Session_Restore

  • Import tab delimited data in livecycle designed form

    Hello,
    Is it possible to import tab delimited data in livecycle designed xdp form?
    Also will it work when we user right enable form so that user can import their data in PDf form?
    Thanks
    Manoj

    Hi Paul,
    This is what I get when in use the function importData
    NotAllowedError: Security settings prevent access to this property or method.
    Doc.importXFAData:-1:XFA:topmostSubform[0]:Receipt[0]:Button8[0]:click
    so what settings are needed and how should my file be formatted. doe sthis function work for XMl or text files or both?
    Now when I use load xml function I get this error
    Xml parsing error: syntax error (error code 2), line 1, column 1 of file
    the file contents are:
    <?xml version="1.0" encoding="ISO-8859-1"?>
          <Receipt>
               <cfien>This is the new string</cfien>
          </Receipt>
    where receipt is my form name and cfien is my field name.
    Can you tell me how to make an XMl file to load the data or give a sample file. Do I need to have SOM expressions in field names?
    Thanks
    Manoj

  • Is data analysis possible with Forms Central?

    Does FormsCentral have an internal feature that you turn on, wherein it gives you an average rating of specific collected data? For instance, if I am doing a survey, does Forms Central have the capacity to give me a statistics report of the collected answers to one question?

    >
    Olga Trofileeva wrote:
    > I have some questions. Is it possible to build Enterprise Data Warehouse (EDW) with Business Objects (BO) software? Or it is preferable to use SAP BW for this? What is the OLAP data model of BO?
    >
    > Thanks,
    > Olga
    To answer your first question: yes, you can construct an Enterprise Data Warehouse using the Business Objects product stack.
    BO Data Integrator XI (now bundled with Data Quality in the Data Services XI offering) is very well suited to design enterprise-strength ETL processes to load, transform and cleanse your source data. When you aren't building a very complex solution, you may even create your data models from within Data Integrator XI although I would normally prefer to use an external tool such as Sybase PowerDesigner, Embarcadero ER/Studio or ERWin (if all else fails).
    The Business Objects XI suite allows your end-users to access the data in your EDW environment using whatever product or combination of products is most suited for your needs - Web Intelligence, Crystal Reports, Voyager, Xcelsius, Explorer.
    The choice to use SAP BW or an external (RBMS based) Enterprise Data Warehouse completely depends on your specific requirements, data sources, infrastructure, etc. Your BI Strategy and Roadmap should give you a clear insight.
    Attempting to implement an enterprise-wide BI solution without a proper business-driven strategy would not be recommendable if you want your BI solution to deliver tangible value to your stakeholders and deliver on any ROI promises made to generate business buy-in.

  • Two issues: activation of transfer rules and data load performance

    hi,
    I have two problems I face very often and would like to get some more info on that topics:
    1. Transfer rules activation. I just finished transport my cubes, ETL etc. on productive system and start filling cubes with data. Very often during data load it occurs that transfer rules need to be activated even if I transport them active and (I think) did not do anything after transportation. Then I again create transfer rules transports on dev, transport changes on prod and have to execute data load again.
    It is very annoying. What do you suggest to do with this problem? Activate all transfer rules again before executing process chain?
    2. Differences between dev and prod systems in data load time.
    On dev system (copy of production made about 8 months ago) I have checked how long it takes me to extract data from source system and it was about 0,5h for 50000 records but when I executed load on production it was 2h for 200000 records, so it was twice slower than dev!
    I thought it will be at least so fast as dev system. What can influence on data load performance and how I can predict it?
    Regards,
    Andrzej

    Aksik
    1 How freequently this activation problem occurs. If it is one time replicate the datasource and activate thetransfer structure( But in general as you know activation of transfer structure should be done automatically after transport of the object)
    2 One thing for difference of time is environmental as you know in production system so many jobs will run at the same time so obiously system performance will be slow compare to Dev System. In your case both the systems are performing equally. You said in dev system for 50000 records half an hour and in production 200000 records 2hrs so records are more in Production system and it took longer time. If it is really causing problem then you have to do some performance activities.
    Hope this helps
    Thnaks
    Sat

  • OIM: How to get the last operation and data from a child form action

    Dear OIM guru,
    I have a regular Process form, a child form and a process definition. When there is an insert/delete into child form a trigger is raised and the process task gets called. The process task does whatever it needs to do on the target system. This is all fine. However, after the task completes, lets say successfully, I want to send out an email to the user saying, for example, group 1 has been added to you or just got deleted.
    I am not able to figure out how to get that last operation and child form data. I am sure you came across this and if so please let me know how to do this.
    Thanks

    So, Can I add a similar process task with the same exact condition one more time? I thought I couldn't use the same condition again. For example, when a group is deleted can I use the process data->child data->group Id with old value check box checked on another process task?

  • T520 - 42435gg / Sound stutter and slow Graphic performance with Intel Rapid Storage AHCI Driver

    Hi everybody,
    I have serious Problems with my 42435gg
    Any time I install the Intel Storage AHCI Driver (I've tried plenty of different versions) which is suggested by System Update I experience a horrible Sound stutter and slow Graphic performance in Windows 7 64-Bit.
    The funny thing in this case: If the external e-sata port is connected the problems do not occur. If the port is unused again, the stutter begins immediately.
    The only thing I can do is using the Windows internal Storage Driver with which I am not able to use my DVD recorder for example.
    The device was sent to lenovo for hardware testing with no result. It was sent back without any repairing.
    Anybody experience on this?
    Kind regards,
    Daniel

    Did you try the 11.5 RST beta? Load up DPClat and see if DPC conditions are favorable.
    What are you using to check graphics performance?
    W520: i7-2720QM, Q2000M at 1080/688/1376, 21GB RAM, 500GB + 750GB HDD, FHD screen
    X61T: L7500, 3GB RAM, 500GB HDD, XGA screen, Ultrabase
    Y3P: 5Y70, 8GB RAM, 256GB SSD, QHD+ screen

  • Report Preview Tab and data loading

    I have 2 questions:
    When I first started developing a report, when I opened the file I had both the design and preview tabs displayed. It seems after I downloaded and viewed 1 of the sample reports, the preview tab disappeared (tho, it's not obvious the 2 are related). I checked the box under Options/Layout/Display Preview Panel, but it did not seem to help. Is there something else I missed?
    The other question, and this may have an effect on my 1st question, is that when I open my report file, the database records are not automatically loaded (thus, nothing to preview at that point). Is there a way to set an automatic log on to the database so that the load time happens at the beginning and not when I actually want to preview results?
    Thanks in advance....Steph

    Don,
    Thanks for the response. I tried what you suggested and it does not seem to work all the time. There may be something else that I need to consider.
    I was wondering...when I do the "Save Data with Report", does it save ONLY the data? Or does it save some intelligent connection back to the database?  Also, is this function the same as the one listed under Options/Reporting? I selected the one on the File menu (as you suggest), but it doesn't seem to change the selection on the Options/Reporting window.
    Apologies for my attempt at a "2-fer-1". I only did that because I believed the 2 might be related...maybe not. I'll repost as a separate issue/concern.
      Cheers...Steph

  • Copy access data into oracle with form builder 9i

    hello
    i need to copy/import data and tables from an MS access database into oracle 9i database. the situation is each time i login to my forms applicatin i will need to do that. any idea pls...
    the access db is frequently updated. so when i login into my forms application it needs to be reimport the data from access db. can any1 pls help me?

    To my opinion, there are 3 potential approaches :
    1. rewrite the application, that data are stored directly in the oracle database
    or the program writes flat files somewhere to the file system which are able getting
    imported from the target instance on daily basis or job driven;
    2. The "Access-driven approach" : link oracle tables into the MS-Access application
    using an ODBC-connection and build a macro in Access which is exporting the stuff to
    the target instance; in order to do so, the Oracle Client and ODBC-Drivers needs to
    get installed on the PC where MS-Access is started to start the exportation process
    (which is probably yours).
    You may then create tables in the target instance and link them from within Access
    using ODBC as the datasource
    3. The "Oracle driven approach". explained in detail here, never tried it, but maybe
    worth a look
    http://www.orafaq.com/node/60
    Basically, install the Oracle Software on a PC, configure an ODBC-Connection and start a
    Listener there;
    The target instance (which may no run on a UNIX system) is now able to commuincate
    with the local running listener on your PC which itself is instantiating a connection to
    the Access "Database" using the ODBC-Connection.
    In this way, your PC acts as a database server from the Oracle Net point of view.
    In theory, this could even work, depending on the ability to connect from the
    instance server directly to the given IP of your PC (Security ?? Hmmmz)

  • Does CHAR data type work with Forms 11g

    We have an application in Forms 11g (32bit - version 11.1.2.0.0). DB is a Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production.
    When our Forms calls a stored procedure which has parameters defined using column name type (i.e. like p_customer_code customers.code%TYPE ) and if the column data type is CHAR we get a run time error when executing the package (ORA-06502: PL/SQL: numeric or value error string).
    When we change the data type of the tables column to varchar2, it works??
    Problem is we have dozens of tables with CHAR data type columns and we can't do all the changes manually. For one thing, if they are part of a primary or unique key or even foreign key, we can't change the data type of the column without dropping the constraint.
    What is the solution to this pls??

    When our Forms calls a stored procedure which has parameters defined using column name type (i.e. like p_customer_code customers.code%TYPE ) and if the column data type is CHAR we get a run time error when executing the package (ORA-06502: PL/SQL: numeric or value error string).this might also be related to compile settings, check your NLS settings (NLS_LANG, NLS_LENGTH_SEMANTICS) both times when you compile forms AND your stored procedures.
    As for the CHAR datatype: CHAR internally is treated like a blank padded varchar:
    http://asktom.oracle.com/pls/apex/f?p=100:11:0::::P11_QUESTION_ID:1542606219593
    Speaking of that it might be time to use varchar2 instead your char columns. You'd have to drop and recreate your constraints, but with the help of the data dictionary (+user_tables+, user_tab_cols, user_cons_columns, user_constraints) and dbms_metadata this shouldn't be that hard.
    Also you'd of course need to recompile all your forms after you are finished, but I guess you are aware of that ;)
    cheers

Maybe you are looking for