Input Ready Queries in Workbook are taking lots of time to save

Hi Folks,
We have workbook which contains input ready queries. When user enter planning values and save the inputted values then it is taking more than 30 minutes to save the data to real time infocube. We have some VBA code on this workbook which is used for formatting and displaying pop-ups when user enters wrong values. Even though we have VBA it should not be taking 30 mins to save.
Please suggest what might have went wrong.
Thanks.
Regards,
Rik

Hi raj,
Thanks for your reply.
Please tell me with which BEx patch you have faced this issue.
We are on BI7.0 SP 19.
Thanks.
Regards,
Rik

Similar Messages

  • PO taking lot of time to save / generate.

    Hi Guys,
    I'm trying to create a new PO in my Quality ECC 6.0 system. The PO gets generated successfully and the Spool is also generated for the same, but the problem is that it was taking around 20 minutes earlier to save the PO ( Purchase order ).
    Now its taking 5 minutes which is also a lot of time as compared to milli seconds which it should take.
    There are 2 tables which remain locked till its saved and so we cannot proceed with the subsequent steps with the same PO
    Please guide me what can be the reason for this and the correct solution to the same.
    1. There is one BTC job in Stopped mode for 600118 seconds , could this be a reason.
    2. There were around 78 users logged on when I generated the PO, which includes RFC & DIA, Its not many users but could this be a reason?
    Please suggest at the earliest if possible how can I reduce this time and what can be the possible reasons for this High response time, my Integration Test is all held up cause of this
    Regards,
    Nitin

    Hi Sreenivas
    I checked the Transaction Profile -> Earlywatch.
    For transaction ME21n the total Response time was 3381s around 56 minutes.
    For my user the response time came out to be 647s around 10 minutes.
    Please tell me how shall i reduce the time,
    Regards,
    Nitin

  • FB32 Taking lots of time while save(like 1 hour and more)

    Hi All,
    While i run FB32 its used for customer credit limit When i changes the limit and click on save button its stuck for hours and when i check through basis guy it was stuck in VBPA table i dont know whats worng with this standrerd T-CODE if any one can help me?
    Tnx in Advance
    Reg.
    Lalit Paliwal

    Hi raj,
    Thanks for your reply.
    Please tell me with which BEx patch you have faced this issue.
    We are on BI7.0 SP 19.
    Thanks.
    Regards,
    Rik

  • (Urgent) form taking lots of time to load and fetch the data

    Hi
    I have very serious problem of perfomance. I have installed oracle portal and configured it properly.
    Now except form everything working perfectly. but only forms and links are taking lots of time to load
    as well as for fetching data. I try to tune my sga and aslo findout hits and miss but i didnt find any
    any problem for it. it is ok. can u help me pls. how can i make my form fast.
    My operation system is NT.
    version of database oracle 8.1.7
    Oracle Portal 3.0.8
    Thanks in advance
    raju parmar
    [email protected]
    null

    hi chetan
    Lots of Thanks.
    Now it working perfect.
    Raju
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Chetan Kashyap ([email protected]):
    The workaround is provided at: http://technet.oracle.com:89/ubb/Forum81/HTML/000395.html <HR></BLOCKQUOTE>
    null

  • Unposted values are not display in input-ready queries - BRAIN 485

    Hi experts,
    After installing the support package level from "7.00 SP16" to "7.00 SP21" we meet an issue in our input-ready queries.
    When we use the setting "Access Type for Result Values" according to master data in a query with time characteristics in columns, the system error BRAIN 485 occurs & no cells are not displayed.
    Any idea ?
    Thanks in advance,
    Laurent

    Hi raj,
    Thanks for your reply.
    Please tell me with which BEx patch you have faced this issue.
    We are on BI7.0 SP 19.
    Thanks.
    Regards,
    Rik

  • Convert to Formula using input ready queries

    Hi,
    Has anyone used converting to formula for input ready queries?  When I use this for some reason all the values disappear in the rows and columns (including characteristic and key figure values).  I would be interested to know if anyone has used this approach successfully.
    Many thanks,

    Also I want to convert this update routine 3.5 to work in 7
    fill the internal table "MONITOR", to make monitor entries
      IF COMM_STRUCTURE-costelmnt+2(1) = '7'.
        IF COMM_STRUCTURE-costelmnt+2(3) BETWEEN '700' AND '774'.
          RESULT = '70'.
        ELSEIF COMM_STRUCTURE-costelmnt+2(3) BETWEEN '775' AND '789'.
          RESULT = '75'.
        ELSEIF COMM_STRUCTURE-costelmnt+2(3) BETWEEN '790' AND '799'.
          RESULT = '79'.
        ENDIF.
      ELSE.
        CASE COMM_STRUCTURE-costelmnt+2(1).
          WHEN '1'.
            RESULT = '10'.
          WHEN '2'.
            RESULT = '20'.
          WHEN '3'.
            RESULT = '30'.
          WHEN '4'.
            RESULT = '40'.
          WHEN '5'.
            RESULT = '50'.
          WHEN '6'.
            RESULT = '60'.
          WHEN '8'.
            RESULT = '80'.
          WHEN OTHERS.
            RESULT = '00'.
        ENDCASE.
      ENDIF.
    if the returncode is not equal zero, the result will not be updated
      RETURNCODE = 0.
    if abort is not equal zero, the update process will be canceled
      ABORT = 0.
    The source is KSTAR(costelement)

  • Function Module Extraction from KONV Table taking lot of time for extractio

    Hi
    I have a requirement wherein i need to get records from KONV Table (Conditions (Transaction Data) ). i need the data corresponding to Application (KAPPL) = 'F'.
    For this i had written one function module but it is taking lot of time (@ 2.5 hrs) for fetching records as there are large number of records in KONV Table.
    I am pasting the Function Module code for reference.
    <b>kindly guide me as to how the extraction performance can be improved.</b>
    <b>Function Module Code:</b>
    FUNCTION ZBW_SHPMNT_COND.
    ""Local interface:
    *"  IMPORTING
    *"     VALUE(I_REQUNR) TYPE  SBIWA_S_INTERFACE-REQUNR
    *"     VALUE(I_ISOURCE) TYPE  SBIWA_S_INTERFACE-ISOURCE OPTIONAL
    *"     VALUE(I_MAXSIZE) TYPE  SBIWA_S_INTERFACE-MAXSIZE OPTIONAL
    *"     VALUE(I_INITFLAG) TYPE  SBIWA_S_INTERFACE-INITFLAG OPTIONAL
    *"     VALUE(I_UPDMODE) TYPE  SBIWA_S_INTERFACE-UPDMODE OPTIONAL
    *"     VALUE(I_DATAPAKID) TYPE  SBIWA_S_INTERFACE-DATAPAKID OPTIONAL
    *"     VALUE(I_PRIVATE_MODE) OPTIONAL
    *"     VALUE(I_CALLMODE) LIKE  ROARCHD200-CALLMODE OPTIONAL
    *"  TABLES
    *"      I_T_SELECT TYPE  SBIWA_T_SELECT OPTIONAL
    *"      I_T_FIELDS TYPE  SBIWA_T_FIELDS OPTIONAL
    *"      E_T_DATA STRUCTURE  ZBW_SHPMNT_COND OPTIONAL
    *"      E_T_SOURCE_STRUCTURE_NAME OPTIONAL
    *"  EXCEPTIONS
    *"      NO_MORE_DATA
    *"      ERROR_PASSED_TO_MESS_HANDLER
    The input parameter I_DATAPAKID is not supported yet !
      TABLES: KONV.
    Auxiliary Selection criteria structure
      DATA: l_s_select TYPE sbiwa_s_select.
    Maximum number of lines for DB table
      STATICS: l_maxsize TYPE sbiwa_s_interface-maxsize.
    Maximum number of lines for DB table
      STATICS: S_S_IF TYPE SRSC_S_IF_SIMPLE,
    counter
              S_COUNTER_DATAPAKID LIKE SY-TABIX,
    cursor
              S_CURSOR TYPE CURSOR.
    Select ranges
      RANGES: L_R_KNUMV  FOR KONV-KNUMV,
              L_R_KSCHL  FOR KONV-KSCHL,
              L_R_KDATU  FOR KONV-KDATU.
    Declaring internal tables
    DATA : I_KONV LIKE KONV OCCURS 0 WITH HEADER LINE.
      DATA : Begin of I_KONV occurs 0,
             MANDT LIKE konv-mandt,
             KNUMV LIKE konv-knumv,
             KPOSN LIKE konv-kposn,
             STUNR LIKE konv-stunr,
             ZAEHK LIKE konv-zaehk,
             KAPPL LIKE konv-kappl,
             KSCHL LIKE konv-kschl,
             KDATU LIKE konv-kdatu,
             KBETR LIKE konv-kbetr,
             WAERS LIKE konv-waers,
             END OF I_KONV.
    Initialization mode (first call by SAPI) or data transfer mode
    (following calls) ?
      IF i_initflag = sbiwa_c_flag_on.
    Initialization: check input parameters
                    buffer input parameters
                    prepare data selection
    The input parameter I_DATAPAKID is not supported yet !
    Invalid second initialization call -> error exit
        IF NOT g_flag_interface_initialized IS INITIAL.
          IF
            1 = 2.
            MESSAGE e008(r3).
          ENDIF.
          log_write 'E'                    "message type
                    'R3'                   "message class
                    '008'                  "message number
                    ' '                    "message variable 1
                    ' '.                   "message variable 2
          RAISE error_passed_to_mess_handler.
        ENDIF.
    Check InfoSource validity
        CASE i_isource.
          WHEN 'X'.
         WHEN 'Y'.
         WHEN 'Z'.
          WHEN OTHERS.
           IF 1 = 2. MESSAGE e009(r3). ENDIF.
           log_write 'E'                  "message type
                     'R3'                 "message class
                     '009'                "message number
                     i_isource            "message variable 1
                     ' '.                 "message variable 2
           RAISE error_passed_to_mess_handler.
        ENDCASE.
    Check for supported update mode
        CASE i_updmode.
    For full upload
          WHEN 'F'.
          WHEN 'D'.
          WHEN OTHERS.
           IF 1 = 2. MESSAGE e011(r3). ENDIF.
           log_write 'E'                  "message type
                     'R3'                 "message class
                     '011'                "message number
                     i_updmode            "message variable 1
                     ' '.                 "message variable 2
           RAISE error_passed_to_mess_handler.
        ENDCASE.
        APPEND LINES OF i_t_select TO g_t_select.
    Fill parameter buffer for data extraction calls
        g_s_interface-requnr    = i_requnr.
        g_s_interface-isource   = i_isource.
        g_s_interface-maxsize   = i_maxsize.
        g_s_interface-initflag  = i_initflag.
        g_s_interface-updmode   = i_updmode.
        g_s_interface-datapakid = i_datapakid.
        g_flag_interface_initialized = sbiwa_c_flag_on.
    Fill field list table for an optimized select statement
    (in case that there is no 1:1 relation between InfoSource fields
    and database table fields this may be far from beeing trivial)
        APPEND LINES OF i_t_fields TO g_t_fields.
    Interpretation of date selection for generic extraktion
       CALL FUNCTION 'RSA3_DATE_RANGE_CONVERT'
         TABLES
           i_t_select = g_t_select.
      ELSE.                 "Initialization mode or data extraction ?
       CASE g_s_interface-updmode.
         WHEN 'F' OR 'C' OR 'I'.
    First data package -> OPEN CURSOR
        IF g_counter_datapakid = 0.
       L_MAXSIZE = G_S_INTERFACE-MAXSIZE.
          LOOP AT g_t_select INTO l_s_select WHERE fieldnm = 'KNUMV'.
            MOVE-CORRESPONDING l_s_select TO l_r_knumv.
            APPEND l_r_knumv.
          ENDLOOP.
          LOOP AT g_t_select INTO l_s_select WHERE fieldnm = 'KSCHL'.
            MOVE-CORRESPONDING l_s_select TO l_r_kschl.
            APPEND l_r_kschl.
          ENDLOOP.
          Loop AT g_t_select INTO l_s_select WHERE fieldnm = 'KDATU'.
            MOVE-CORRESPONDING l_s_select TO l_r_kdatu.
            APPEND l_r_kdatu.
          ENDLOOP.
    *In case of full upload
    Fill field list table for an optimized select statement
    (in case that there is no 1:1 relation between InfoSource fields
    and database table fields this may be far from beeing trivial)
       APPEND LINES OF I_T_FIELDS TO S_S_IF-T_FIELDS.
          OPEN CURSOR G_CURSOR FOR
            SELECT MANDT
                   KNUMV
                   KPOSN
                   STUNR
                   ZAEHK
                   KAPPL
                   KSCHL
                   KDATU
                   KBETR
                   WAERS
            FROM   KONV
            WHERE KNUMV IN l_r_knumv
            AND   KSCHL IN l_r_kschl
            AND   KDATU IN l_r_kdatu
            AND   KAPPL EQ 'F'.
        ENDIF.
        Refresh I_KONV.
        FETCH NEXT CURSOR G_CURSOR
                   APPENDING CORRESPONDING FIELDS OF TABLE I_KONV
                   PACKAGE SIZE S_S_IF-MAXSIZE.
        IF SY-SUBRC <> 0.
          CLOSE CURSOR G_CURSOR.
          RAISE NO_MORE_DATA.
        ENDIF.
        LOOP AT I_KONV.
         IF I_KONV-KAPPL EQ 'F'.
          CLEAR :E_T_DATA.
          E_T_DATA-MANDT = I_KONV-MANDT.
          E_T_DATA-KNUMV = I_KONV-KNUMV.
          E_T_DATA-KPOSN = I_KONV-KPOSN.
          E_T_DATA-STUNR = I_KONV-STUNR.
          E_T_DATA-ZAEHK = I_KONV-ZAEHK.
          E_T_DATA-KAPPL = I_KONV-KAPPL.
          E_T_DATA-KSCHL = I_KONV-KSCHL.
          E_T_DATA-KDATU = I_KONV-KDATU.
          E_T_DATA-KBETR = I_KONV-KBETR.
          E_T_DATA-WAERS = I_KONV-WAERS.
          APPEND E_T_DATA.
       ENDIF.
        ENDLOOP.
        g_counter_datapakid = g_counter_datapakid + 1.
      ENDIF.
    ENDFUNCTION.
    Thanks in Advance
    Regards
    Swapnil.

    Hi,
    one option to investigate is to select the data with a condition on KNUMV (primary IDX).
    Since shipment costs are store in VFKP I would investigate if all your F condition records are used in this table (field VFKP-KNUMV).
    If this is the case then something like
    SELECT *
    FROM KONV
    WHERE KNUMV IN (SELECT DISTINCT KNUMV FROM VFKP)
    or
    SELECT DISTINCT KNUMV
    INTO CORRESPONDING FIELD OF <itab>
    FROM VFKP
    and then
    SELECT *
    FROM KONV
    FOR ALL ENTRIES IN <itab>
    WHERE...
    will definitively speed it up.
    hope this helps....
    Olivier

  • Select query is taking lot of time to fetch data.....

    Select query is taking lot of time to fetch data.
        SELECT algnum atanum  abdatu abzeit abname abenum bmatnr bmaktx bqdatu bqzeit bvlenr bnlenr bvltyp bvlber b~vlpla
               bnltyp bnlber bnlpla bvsola b~vorga INTO TABLE it_final FROM ltak AS a
                       INNER JOIN ltap AS b ON  btanum EQ atanum AND algnum EQ blgnum
                       WHERE a~lgnum = p_whno
                       AND a~tanum IN s_tono
                       AND a~bdatu IN s_tocd
                       AND a~bzeit IN s_bzeit
                       AND a~bname IN s_uname
                       AND a~betyp = 'P'
                       AND b~matnr IN s_mno
                       AND b~vorga <> 'ST'.
    Moderator message: Please Read before Posting in the Performance and Tuning Forum
    Edited by: Thomas Zloch on Mar 27, 2011 12:05 PM

    Hi Shiva,
    I am using two more select queries with the same manner ....
    here are the other two select query :
    ***************1************************
    SELECT * APPENDING CORRESPONDING FIELDS OF TABLE tbl_summary
        FROM ztftelpt LEFT JOIN ztfzberep
         ON  ztfzberep~gjahr = st_input-gjahr
         AND ztfzberep~poper = st_input-poper
         AND ztfzberepcntr  = ztftelptrprctr
        WHERE rldnr  = c_telstra_projects
          AND rrcty  = c_actual
          AND rvers  = c_ver_001
          AND rbukrs = st_input-bukrs
          AND racct  = st_input-saknr
          AND ryear  = st_input-gjahr
          and rzzlstar in r_lstar             
          AND rpmax  = c_max_period.
    and the second one is
    *************************2************************
      SELECT * APPENDING CORRESPONDING FIELDS OF TABLE tbl_summary
        FROM ztftelnt LEFT JOIN ztfzberep
         ON  ztfzberep~gjahr = st_input-gjahr
         AND ztfzberep~poper = st_input-poper
         AND ztfzberepcntr  = ztftelntrprctr
        WHERE rldnr  = c_telstra_networks
          AND rrcty  = c_actual
          AND rvers  = c_ver_001
          AND rbukrs = st_input-bukrs
          AND racct  = st_input-saknr
          AND ryear  = st_input-gjahr
          and rzzlstar in r_lstar                              
          AND rpmax  = c_max_period.
    for both the above table program is taking very less time .... although both the table used in above queries have similar amount of data. And i can not remove the APPENDING CORRESPONDING. because i have to append the data after fetching from the tables.  if i will not use it will delete all the data fetched earlier.
    Thanks on advanced......
    Sourabh

  • Partition switching taking lots of time

    I have a partitioned table. All the partitions of this table are mapped to a single filegroup.
    When I am trying to switch partition, it is taking lots of time more than 5 mins for a single partition.
    I can see that there is a file in the filegroup and its size is very high that is 234GB.
    Can partition switching take more time due to file size?
    or is there any other issue?
    I need help in finding out what is the problem.
    Thanks,

    So you are running ALTER TABLE SWITCH? Check for blocking. The operation requires an Sch-M lock which means that it is blocked any query that is running, including queries using NOLOCK.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Extractor taking lots of time.

    hi experts,
    we are using extractor 2LIS_13_VDKON to extract the data.
    but its taking lots of time.how can i solve this problem.
    any help appreciated.

    Hi,
    Have you done any enhancements(Append) to this DS.If so, make sure that code is performance optimistic.
    Basically the setup table for 2LIS_13_VDKON will has the lot of data. For example, it holds at n avarage 10 records for each billing item record.
    So also check system is uploading records to the New table of ODS or not. 
    With rgds,
    Anil Kumar Sharma .P

  • Web-I Report taking lots of time to refresh when filters changed

    In BO XI 3.1 Web-I report is taking lot of time to refresh when any changes are made in edit query mode on the filters.
    When the query is runned for the first time it runs well, but if we make any changes in the filters pane, it is taking lots of time to refresh, and when we cancel the query web-i server is not releasing the server resources (CPU, Memory)
    Did anyone face this kind of a problem, and resolved it.
    Please let me know your thoughts
    Thank you

    Hi,
    why do you need 100K rows in your reports? Is there a way to consolidate/aggregate your data on the database?
    Comparing sqlplus and BOBJ is a little bit unfair. Let me explain you why:
    sqlplus starts displaying data after the query returns the first results. I am sure that if you measure the time that sqlplus needs to display ALL rows, this should be way more than just 10 secs.
    On the other hand BOBJ will display something as soon as the first page can be created. Depending on the way your report is structured, it may be the case that the WebI server should first get ALL rows in order to be able to display the first page of your report. This is for example the case if you let the total number of pages being displayed or if you display total sums already at the header of your report or if you have added a chart based on your data also in the header of your report. Of course, after the first call the report will be cached (and thus fast to be displayed) until you change again the parameters values for your query.
    Since you do not display (or do you use this function in a variable or a formula, maybe?) the total number of pages in your report, I can only assume that you either trying to display total sums (or other kinds of aggregated values) or a chart in the first page of your report. Can you confirm this?
    Regards,
    Stratos

  • Browser is taking lot of time in loading forms

    Hi
    I am working on forms 10g..
    when i am running my form.... the browser is loading applet and jar files.
    After loading this it is taking lot of time to show forms...
    Even I am clearing all my temp files frequently...
    Why it is taking so much time??
    What I have to do so that it can load forms in very least time??
    Thanks

    After loading this it is taking lot of time to show forms...This could be caused by a resource issue on the Client PC. By resource I mean available RAM, JRE memory settings, etc. It could also be a conflict between the Client OS, installed JRE and the Browser. Please tell us what the Client OS, JRE and Browser are as well as how much RAM is installed on the client pc.
    Even I am clearing all my temp files frequently...Are you clearing the Browsers temporary internet files or the JRE's JAR cache? If the JRE's JAR cache then you are forcing the client to download the JAR each time you clear the JAR cache.
    The slowness can also be caused by how large your FORMS files are and what you are doing in the Pre-Form and When-New-Form-Instance triggers. If you are performing numerous calls to the database or calling SYNCHRONIZE too many time in any of the triggers that fire when a Form initially loads then this will cause your form to load slower.
    Hope this helps,
    Craig B-)
    If someone's response is helpful or correct, please mark it accordingly.

  • Delete taking lot of time

    Hi
    I have a delete statement..which is taking lot of time. If I select this scenerio only 500 records are coming. But delete is taking lot of time.
    Please advise.
    delete from whs_bi.TRACK_PLAY_EVENT a
    where a.time_stamp >=to_date('5/27/2013','mm/dd/yyyy')
    and a.time_stamp <to_date('5/28/2013','mm/dd/yyyy')Thanks in adv.
    KPR

    Lets check the wait events.
    Open 2 sessions, 1 for running the update command and another for monitoring wait events. The session in which you want to run UPDATE, find the SID of that session ( SELECT userenv('SID') from dual ).
    Now run the update in one session (of which we have already found the SID).
    Run following query in another session
    select w.sid sid,
           p.spid PID,
           w.event event,
           substr(s.username,1,10) username,
           substr(s.osuser, 1,10) osuser,
           w.state state,
           w.wait_time wait_time,
           w.seconds_in_wait wis,
           substr(w.p1text||' '||to_char(w.P1)||'-'||
                  w.p2text||' '||to_char(w.P2)||'-'||
                  w.p3text||' '||to_char(w.P3), 1, 45) P1_P2_P3_TEXT
    from v$session_wait w, v$session s, v$process p
    where s.sid=w.sid
      and p.addr  = s.paddr
      and w.event not in ('SQL*Net message from client', 'pipe get')
      and s.username is not null
      and s.sid = &your_SIDWhile UPDATE is running in another session, run above query in second session for 5-6 times, with gap of (say) 10 seconds. If you can give us the output of monitoring query (from all 5-6 runs), that might throw more light on whats going under the hood.

  • Taking lot of time for loading

    Hi,
      We are loading data from HR to BI. Connection between HR and BI has been done recently. When I am trying to load data from HR  to BI it taking lot of time--for example to load 5recs its taking 8hrs. Its same for every datasource. Should we change anything in settings to makes IDOCS work proper? Thanks

    You have to isolate the part that is taking the time.
    - Is R/3 extraction quick? (You can check with RSA3 and see how long does it take.)
    - If R/3 extraction is slow, then is the selection using an index? How many records are there in the table / view?
    - Is there a user exit? Is user exit slow?
    You can find out using monitor screen:
    - After R/3 extraction completed, how long did it take to insert into PSA?
    - How long did it take in the update rules?
    - How long did it take to activate?
    Once you isolate the problem area, post your findings here and someone will help you.

  • Reading, comparing and appending two files is taking lot of time

    public static void main(String args[])
         BufferedReader bSource = null;
         String sSource="";
         String sSourceData="";
         BufferedReader bDestination = null;
         String sDestination="";
         String sDestinationData="";
         try
              bSource = new BufferedReader(new FileReader("C:/TEMP/Source.map"));
              while((sSource = bSource.readLine()) != null)
                   sSourceData+=sSource;
              bDestination = new BufferedReader(new FileReader("D:/TEMP/Destination.map"));
              while((sDestination = bDestination.readLine()) != null)
                   sDestinationData+=sDestination;
         catch(IOException ioe)
              System.out.println(ioe.getMessage());
         finally
              try
                   bSource.close();
                   bDestination.close();
              catch(Exception e)
         if(sDestinationData.contains(sSourceData))
              fDestinationFlag=false;
         else
              fDestinationFlag=true;
         if(fDestinationFlag)
              BufferedReader br2 = null;
              BufferedWriter bw2 = null;
              try
                   br2 = new BufferedReader(new FileReader("C:/TEMP/Source.map"));     
                   bw2 = new BufferedWriter(new FileWriter("D:/TEMP/Destination.map",true));
                   while((line2 = br2.readLine()) != null)
                        bw2.write(line2);
                        bw2.newLine();
                        bw2.flush(); 
              catch(IOException ioe)
                   System.out.println(ioe.getMessage());
              finally
                   try
                        br2.close();
                   catch(Exception e)
                   try
                        bw2.close();
                   catch(Exception e)
    }I am comparing the contents of the two files Source.map (containing 3 lines) and Destination.map (containing 5000+ lines). The contents of the Source.map gets appended in the Destination.map only if it does not contain the same.
    Now my program is doing the necessary, but it is taking lot of time. It's taking 1 min 26 seconds.
    Could anyone please tell me what do I modify in may program to make it execute much faster.

    1) Just compare the byte content - no need to convert to Strings.
    2) Compare a block (say 256K bytes) at a time - no need to read the whole file into memory.
    3) Compare the file lengths first - if they are different then the content cannot be the same.
    4) When appending one file to the other, set the 'append' flag in the constructor of FileOutputStream to true.

Maybe you are looking for