How to avoid data loss when an action is perfomed ....

hi,
I am using a dynamic Tab. Each tab contains a seperate jsp page.( the jsp page is included for the corresponding tab). Each page can contain more than 25 fields. The problem is for example i will select some check box in the first tab and i will go the second tab and i will do some insert operations , when i came back to the first tab , the checkbox which i had selected or the text what i had entered should be there. If it can be solved by using AJAX, pls guide me.
Tools i am using : jsp, struts.
Looking forward to hear to solve this problem.

hi....
c through that when u eturn to tab.... set the form values to the page
i mean if u r using a form bean for ur jsp. use name name atrribute of the sturts html tag. and give the form bean name to the name attribute i hope this would solve your problem
thaks
with rgards
shekhar

Similar Messages

  • How to avoid data repetation when using select statements with innerjoin

    how to avoid data repetation when using select statements with innerjoin.
    thanks in advance,
    satheesh

    you can use a query like this...
      SELECT DISTINCT
             frg~prc_group1                  "Product Group 1
             frg~prc_group2                  "Product Group 2
             frg~prc_group3                  "Product Group 3
             frg~prc_group4                  "Product Group 4
             frg~prc_group5                  "Product Group 5
             prc~product_id                  "Product ID
             txt~short_text                  "Product Description
    UP TO 10 ROWS
    INTO TABLE l_i_data
    FROM
    Joining CRMM_PR_SALESG and
    COMM_PR_FRG_ROD
    crmm_pr_salesg AS frg
    INNER JOIN comm_pr_frg_rod AS prd
    ON frgfrg_guid = prdfragment_guid
    Joining COMM_PRODUCT and
    COMM_PR_FRG_ROD
    INNER JOIN comm_product AS prc
    ON prdproduct_guid = prcproduct_guid
    Joining COMM_PRSHTEXT and
    COMM_PR_FRG_ROD
    INNER JOIN comm_prshtext AS txt
    ON prdproduct_guid = txtproduct_guid
    WHERE frg~prc_group1 IN r_zprc_group1
       AND frg~prc_group2 IN r_zprc_group2
       AND frg~prc_group3 IN r_zprc_group3
       AND frg~prc_group4 IN r_zprc_group4
       AND frg~prc_group5 IN r_zprc_group5.
    reward it it helps
    Edited by: Apan Kumar Motilal on Jun 24, 2008 1:57 PM

  • How to avoid the dialogue when converting from context menu

    This is a follow-up to http://forums.adobe.com/message/2016146#443364 which was locked due to a bug.
    i would like to know how i can avoid the 'save as' dialogue when using the context menu to convert a word file to pdf using windows XP/acrobat 8/word 2007.
    so far, it always prompts me for the filename, which is annoying as it does not allow me to batch-convert several files and let the thing run its course.
    the solution provided by Steve in the other thread does not work - even if the plugin from word does not propt for a filename, it still does when triggered from explorer/context menu.
    so back to square one: how to avoid that dialogue when not opening word at all?
    cheers, thanks for any help. michael

    For a permanent change, START>PRINTERS>right click on Adobe PDF printer and select properties>General>Printing Preferences>Adobe PDF Settings. Under the settings tab, uncheck the box related to asking for a name. Pretty sure that is the location, but it may vary with version.

  • How to avoid launching Labview when invoking VI by Teststand

    Hi, friends.
    I'm a learner of teststand. I wrote some VIs by Labview. Now I want to invoke them in teststand environment. But when I run the sequence. Labview will be automatically launched. How to avoid launching Labview when I run the sequence?
    Thanks.
    Solved!
    Go to Solution.

    I'm not sure that I follow what you are saying..  But it sounds like when the VI is called, the LabVIEW development software is started and you want to avoid that.
    If you create an executable or dll and call that from within TestStand, it should not call the development environment.  You would need to have the LabVIEW run-time engine installed on the target PC. 
    It has been many years since I've used TestStand with LV, so my memory may not be accurate.  I think the above had worked for me.  However, it seems that I recall something about having the LV code in debug mode causing it to open the development environment.... but it's a distant memory.    You're not using breakpoints where the LV VI is being called, are you?
    R

  • How do I avoid data loss during setup?

    Hi All,
    Update: below is my original question/concern. But since posting, I've come across others posting about data loss at various stages of install and initial setup and syncing. None of those posts have been replied to, and I think many of us need some answers. So after reading one guys calendar got scrubbed, I checked mine and the same thing... I had synced through iTunes before upgrading, and all my freshly updated calendar events had been erased when I enabled iCloud. Honestly, what is it with Apple's obsession with deleting people's data when adding new equipment or services?
    So add to my question below, a how to for getting data from existing devices to cloud w/o having cloud erase files and data from devices...
    I'm just setting up. I want to go into this cloud thing slowly, starting with just calendar syncing.
    I go to settings: Documents and Data and click it off. I get a pop up witch says "All documens and Data stored in iCloud will be deleted from this iPad"
    What?
    This switch was in the on position upon install. Have my Documents and data already been scarfed to the cloud, where now if I turn off the feature I lose my files?
    And what about down the road if one wants to turn off the feature after it's been in use? Doe my iPad get scrubbed of my files?
    Am I learning that once my files get sent to the cloud, the server holds a copy which I can never scrub off iCloud?
    The way that warning reads, my only option for taking a device off the grid is to have my data scrubbed from that device.
    What am I missing? I want to control which of my data makes it to the cloud. This says they take it all up front and steal my copy if I try to get my device off the grid. No?
    Please help me understand how to control which of my files go to the cloud, and how not to have my files removed from my device if I chose to opt out of the docs and data sync feature?
    S.
    PS why would they ever make that a consequence of turning off the feature?

    Unless you have signed up for AT&T's international roaming for phone calls, I suggest putting the device into Airplane mode. That is what I did when I did the 3 day Bahamas Cruise. And to support what AmishCake said, the cruise lines charge for their wi-fi. Generally a daily 24-hour charge, or they will have a special for the length of the cruise, but it is pricey.

  • How to avoid clear record when tab pages changes

    Hi All
    I am using oracle forms 10g and db 10g.
    I have created a form with four tab pages. Namely "EXPENSE" , "AMOUNT_DETAILS", "SUPPLIER" , "ACCOUNT".
    When i enter a data in page 1 ie Expense and move to next page page2 "AMOUNT_DETAILS", and enters data in page3 "SUPPLIER" and when i come back to page1 "EXPENSE" and also Page2 "AMOUNT" the data get cleared in Tab pages. There is no data again i need to enter the data manually.
    Can any one suggest me how to avoid this clear data.
    Thanks & Regards
    Srikkanth

    Hi,
    Thanks once again for your quick response.
    I have checked it , i was working with oracle apps.Now i have entered all the datas in the four tab and press save button in the screen at that time also the data get cleared.Can you please tell is there any work around for this.
    regards
    Srikkanth

  • How to avoid shared locks when validating foreign keys?

    I have a table with a FK and I want to update a row in that table without being blocked by another transaction which is updating the parent row at the same time. Here is an example:
    CREATE TABLE dbo.ParentTable
    PARENT_ID int NOT NULL,
    VALUE varchar(128) NULL,
    CONSTRAINT PK_ParentTable PRIMARY KEY (PARENT_ID)
    GO
    CREATE TABLE dbo.ChildTable
    CHILD_ID int NOT NULL,
    PARENT_ID INT NULL,
    VALUE varchar(128) NULL,
    CONSTRAINT PK_ChildTable PRIMARY KEY (CHILD_ID),
    CONSTRAINT FK_ChildTable__ParentTable FOREIGN KEY (PARENT_ID)
    REFERENCES dbo.ParentTable (PARENT_ID)
    GO
    INSERT INTO ParentTable(PARENT_ID, VALUE)
    VALUES (1, 'Some value');
    INSERT INTO ChildTable(CHILD_ID, PARENT_ID, VALUE)
    VALUES (1, 1, 'Some value');
    GO
    Now I have 2 transactions running at the same time:
    The first transaction:
    SET TRANSACTION ISOLATION LEVEL READ UNCOMMITTED;BEGIN TRAN
    UPDATE ParentTable
    SET VALUE = 'Test'
    WHERE PARENT_ID = 1;
    The second transaction:
    SET TRANSACTION ISOLATION LEVEL READ UNCOMMITTED;BEGIN TRAN
    UPDATE ChildTable
    SET VALUE = 'Test',
    PARENT_ID = 1
    WHERE CHILD_ID = 1;
    If 'UPDATE ParentTable' statement runs a bit earlier, then 'UPDATE ChildTable' statement is blocked untill the first transaction is committed or rollbacked. It happens because SQL Server acquires shared locks when validating foreign keys, even
    if the transaction is using read uncommitted, read committed snapshot (read committed using row versioning) or snapshot isolation level. I cannot see why change in the ParentTable.VALUE should prevent me from updating ChildTable. Please note that ParentTable.PARENT_ID
    is not changed by the first transaction, which means that from FK's point of view whatevere is set to the ParentTable.VALUE is never a problem for referential integrity. So, such blocking behavior seems to me not logical. Furthermore, it contradicts to the
    MSDN:
    Transactions running at the READ UNCOMMITTED level do not issue shared locks to prevent other transactions from modifying data read by the current transaction. READ UNCOMMITTED transactions are also not blocked by exclusive locks that would prevent the
    current transaction from reading rows that have been modified but not committed by other transactions. 
    Does anybody know how to workaround the issue? In other words, are there any tricks to avoid shared locks when validating foreign keys? (Disabling FK is not an option.) Thank you.
    Alexey

    If you change the primary key of the parent table to be nonclustered, there is no blocking.
    Indeed, when I update ParentTable.VALUE, then:
    in case of PK_ParentTable is clustered, a particular row in the clustered index is locked (request_mode:X, resource_type: KEY)
    in case of PK_ParentTable is non-clustered, a particular physical row in the heap is locked (request_mode:X, resource_type: RID).
    and when I update ChildTable.PARENT_ID, then:
    in case of PK_ParentTable is clustered, this index is used to verify referential integrity:
    in case of PK_ParentTable is non-clustered, again this index is used to verify referential integrity, but this time it is not locked:
    It is important to note that in both cases SQL Server acquires shared locks when validating foreign keys. The principal difference is that in case of clustered PK_ParentTable the request is blocked, while for non-clustered index it is granted.
    Thank you, Erland for the idea and explanations.
    The only thing that upsets me is that this solution cannot be applied, because I don't want to convert PK_ParentTable from clustered to non-clustered just to avoid blocking issues. It is a pity that SQL Server is not smart enough to realize that:
    ParentTable.PARENT_ID is not changed and, as a result, should not be locked
    ChildTable.PARENT_ID is not actually changed either (old value in my example is 1 and the new value is also 1) and, as a result, there is no need at all for validating the foreign key.
    In fact, the problem I described is just a tip of the iceberg. The real challenge is that I have deadlocks because of the FK validation. In reality, the first transaction has an additional statement which updates ChildTable:
    SET TRANSACTION ISOLATION LEVEL READ UNCOMMITTED;
    BEGIN TRAN
    UPDATE ParentTable
    SET VALUE = 'Test'
    WHERE PARENT_ID = 1;
    UPDATE ChildTable
    SET VALUE = 'Test'
    WHERE PARENT_ID = 1;
    The result is famous message:
    Msg 1205, Level 13, State 51, Line xx
    Transaction (Process ID xx) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
    I know that if I change the order of the two statements, it will solve the deadlock issue. But let's imagine I cannot do it. What are the other options?
    Alexey

  • How to avoid data lost in Dataguard failover ?

    Hi gurus,
    We are implementing Dataguard with manual failover (Not with Fast Start feature)
    How to avoid/minimize lost of data if the primary server goes down before the last log is sent ?
    Thank you for your help,
    xtanto

    Most important the dataguard protection mode, you have 3 choices: Maximum Performance,Maximum Availability,Maximum Protection . If you are interested in no data loss choose Maximum Protection, that's the definition:
    Maximum Protection
    This protection mode ensures that zero data loss occurs if a primary database fails. To
    provide this level of protection, the redo data needed to recover a transaction must be
    written to both the online redo log and to at least one synchronized standby database
    before the transaction commits. To ensure that data loss cannot occur, the primary
    database will shut down, rather than continue processing transactions, if it cannot
    write its redo stream to at least one synchronized standby database.
    Because this data protection mode prioritizes data protection over primary database
    availability, Oracle recommends that a minimum of two standby databases be used to
    protect a primary database that runs in maximum protection mode to prevent a single
    standby database failure from causing the primary database to shut down.
    For details see the documentation for your unknown database version.
    Werner

  • How to get data back from an action ?

    Hello,
    Would it be possible to get data back from an action (out of the question EO_MESSAGE & ET_FAILED_KEY )?
    For example we got an order with order positions and we need a "function" to e.g. count all positions. Because of Performance the function should not be processed each time the order is changed, read or a position is added. Instead the function should be processed only if it was called explicitly.
    Is it possible to create a kind of action which is actually counting all entries and export the number of them?
    How to mark a parameter in is_parameters as exporting?
    Is this just done by (naming-) convention?
    What is the preferred way to have “methods” with returning/exporting values?
    Regards,
    Lorenz

    Hello Lorenz,
    As you have already figured out , the Action API provides you with only the messages and failed keys if any.
    Post action execution , you can always execute a retrieve or retrive by association , to get the latest buffer snapshot , which of course would include the changes that you have made in your action.
    If you want to ensure that users have explict control on execution of your "fucntion", then of course , you should model it as an action on the BO.
    The parameter is_parameters is an IMPORTING parameter. You CANNOT use it to export anything back from the action. For importing ,  you can of course have any structure to use as the is_paramaters , which you model as the action parameter structure which modelling your BO action.
    From an external entity the only way to interact with a BO is by consuming the BO services and you are bound by the BOBF standard interfaces. Any and all data you require needs to be modelled as node attributes ( persistent or transient ) and fetched using the RETRIEVE, RETRIEVE_BY_ASSOCIATION or QUERY services.
    Regards,
    Indranil.

  • Data Loss when a database crashes

    Hi Experts,
    I was asked the question of "how much data is lost when you pull the plug off an oracle database all of a sudden".. and my answer was "all the data in the buffers that have not been committed". We know that you can have committed data sitting in the redo logs (that have not been written to the datafiles) that the instance will use for recovery once the instance is restarted again; however, this got me thinking and asking the question of how much uncommitted data is actually sitting in memory that potentially can be lost if the instance goes down all of a sudden?
    With the use of sga_target and sga_max_size, the memory allocation for the cache buffer will vary from time to time.. So, is it possible to quantify the amount of lost data at all (in byts, kb, mb..etc)?
    For example if the sga is set to 1gb (sga_max_size=1000mb), check point every 15mins (as we can't predict/know how often the app commits).. assume a basic transaction size for any small to medium size database. Redo logs set to 50mb (even though this doesn't come into play at this stage).
    I would be really interested in your thoughts and ideas please.
    Thanks

    All Oracle Data Manipulation Language (DML) and Date Definition Language (DDL) statements must record an entry in the Redo log buffer before they are being executed.
    The Redo log buffer is:
    •     Part of the System Global Area (SGA)
    •     Operating in a circular fashion.
    •     Size in bytes determined by the LOG_BUFFER init parameter.
    Each Oracle instance has only one log writer process (LGWR). The log writer operates in the background and writes all records from the Redo log buffer to the Redo log files.
    Well, just to clarify, log writer is writing committed and uncommitted transactions from the redo log buffer to the log files more or less continuously - not just on commit (when the log buffer is 10mb full, 1/3 full, every 3 seconds or every commit - whichever comes first - those all trigger redo writes).
    The LGWR process writes:
    •     Every 3 seconds.
    •     Immediately when a transaction is committed.
    •     When the Redo log buffer is 1/3 full.
    •     When the database writer process (DBWR) signals.
    Crash and instance recovery involves the following:
    •     Roll-Forward
    The database applies the committed and uncommitted data in the current online redo log files.
    •     Roll-Backward
    The database removes the uncommitted transactions applied during a Roll-Forward.
    What also comes into play in the event of a crash is MTTR, for which there is a advisory utility as of Oracle 10g. Oracle recommends using the fast_start_mttr_target initialization parameter to control the duration of startup after instance failure.
    From what I understand, uncommitted transactions will be lost, or more precisely undone after an instance crash. That's why it is good practice to manually commit transactions, unless you plan to use SQL rollback. Btw, every DDL statement and exit from sqlplus implies an automatic commit.
    Edited by: Markus Waldorf on Sep 4, 2010 10:56 AM

  • How to avoid safety warning when opening a pdf from an link in another pdf?

    I make pdf in Acrobat with links to other pdf files av have made. When i try to click a link to open the pdf, i get the safety question if trust it or if i will block it. How to avoid that question all the time?

    Hello Kvalitetetsansvarlig
    Try adding the PDF to the trust location in Acrobat
    Edit -> Preferences-> Security (Enhanced), Click on "Add File", and browse to your file.
    Click on "OK" button and check.
    Regards,
    Deepak

  • How to populate date & time when user enter data for custom table in sm30

    Can anyone tell me How to populate system date & time when user enter data for custom table in sm30..
      Req is
      i have custom table and using sm30 user can enter data.
    after saving date i want to update date & time in table
    Pls let me know where to write the code?
    Thanks in Advance

    You have to write the code in EVENT 01 in SE54 transaction. Go to SE54, enter your Ztable name and in the menu 'Environment-->Events'. Press 'ENTER' to go past the popup message. In the next screen, click on 'New Entries'. In the first column, enter 01 and in the next column give some name for your routine(say UPDATE_USER_DATE_TIME). Then click on the souce code icon that appears in blue at the end of the row. In the code, you need logic like below.
    FORM update_user_date_time.
      DATA: f_index LIKE sy-tabix.
      DATA: BEGIN OF l_total.
              INCLUDE STRUCTURE zztable.
      INCLUDE  STRUCTURE vimtbflags.
      DATA  END OF l_total.
      DATA: s_record TYPE zztable.
      LOOP AT total INTO l_total.
        IF l_total-vim_action = aendern OR
           l_total-vim_action = neuer_eintrag.
          MOVE-CORRESPONDING l_total TO s_record.
          s_record-zz_user = sy-uname.
          s_record-zz_date = sy-datum.
          s_record-zz_time = sy-uzeit.
          READ TABLE extract WITH KEY l_total.
          IF sy-subrc EQ 0.
            f_index = sy-tabix.
          ELSE.
            CLEAR f_index.
          ENDIF.
          MOVE-CORRESPONDING s_record TO l_total.
          MODIFY total FROM l_total.
          CHECK f_index GT 0.
          MODIFY extract INDEX f_index FROM l_total.
        ENDIF.
      ENDLOOP.
    ENDFORM.                    " UPDATE_USER_DATE_TIME
    Here ZZTABLE is the Z table and ZZ_USER, ZZ_DATE, and ZZ_TIME are the fields that are updated.

  • ALV List Output: How to avoid data statistics from printing in Background

    Hi,
    I am running an ALV list display in Background mode and the output is showing the data statistics first and then the actual output of the ALV...
    I dont want the data statistics to be printed..
    How to avoid that.. is there any to do it...

    Hi Neela
    If you are using the good old REUSE_ALV_GRID_DISPLAY function module, you should be able to suppress the List Info if SY-BATCH is 'X' using the IS_PRINT parameter.
      MOVE 'X' TO is_print-no_print_listinfos .
    Presumably there is a similar parameter using classes.
    Cheers
    Lyal

  • How to avoid the prompt when verifing signature?

    I sign the doc with my cert.
    The cert is not trusted by Acrobat/Reader itself so when i try to verify the signature Acrobat/Reader will prompt the user cert is not trusted.
    My question is how to avoid this prompt through Acrobat SDK or how to add my cert to trusted cert list through Acrobat SDK?
    Thanks.

    Patrick Leckey wrote:
     You will also see the BER Decoding error when you have a self-signed peer-to-peer cert in your Windows Cert Store, even if it is not the signature you are using to sign.
    Start > Run > certmgr.msc
    Look for any certs that have "peer-to-peer" in the Intended Purpose column
    Remove that cert
    Restart Acrobat
    Give that a shot.
    Patrick,I have not found the Intended purpose of cert names "peer-to-peer".
    But I have just found the cert which will cause the signature failure.
    The key is the cert can not be removed.
    I'm very curious that why a cert has no relationship with the signature could cause the signature failure?
    Is it a BUG of Acrobat/Adobe Reader?
    If not how can i operate successfully?
    Thanks.

  • How to avoid opening photos when connecting a camera

    how can I avoid opening photos when connecting a camera?

    Connect the Camera.
    Launch Image Capture (in the Applications Folder)
    You can set what happens when you connect this device in the lower left of the Image Capture window.

Maybe you are looking for

  • JDBC Receiver Adapter FATAL ERROR

    Hi experts, I have a ccBPM to JDBC sync scenario. My Integration Process (IP_Workflow) has a Send Step (Mode = Synchronous, Interface = SIA_Update). This SIA_Update is abstract, synchronous and It´s message type has the following structure: <?xml ver

  • Downloading a Bianry File

    If suppose user has sent a Http request to get a binary file from the server, is it possible to send the required file throgh Http response in some way?

  • Time Machine/Time Capsule where is the data?

    Bought Time Capsule a month ago when I got a Macbook Air. First time with 2 computers. Today exploring and trying to connect an old hard drive and TimeCapsule has only 2 files, both in a data folder and each labeled with one of the two computers dot

  • Pre-requisite for SCWCD-5.0.

    what is pre-requisite for SCWCD-5.0. I have SCJP-1.2, am i elligible for SCWCD-5.0 Thanks

  • Rendering BSP page in CIC0 taking lot of time

    Hi All, A new bsp application is developed and is called from CIC0 (Interaction Center Winclient). The UI (bsp page) rendering is taking lot of time in SAPGUI. I'm unable to find the exact reason for this problem. Has any one of us experienced this p