NX 6125 too much time to complete POST. USB error

I have been having some issues over the last week or two
NX 6125 takes too much time to boot. I'm dual booting it , Windows XP and Linux Mint.
It takes more time to complete the POST, during POST, LEDs on all the USB devices attached to it will go ON & OFF.
then comes the boot menu. from that step onwards it works normally.
Once it boots into Windows, there is a pop up "USB device malfunctioned" which keeps on coming even if no devices are attached to USB.(it has an internal card reader and fingerprint reader module)
While using Linux, the kernel throws the error "  [ 2094.490070] hub 1-0:1.0: connect-debounce failed, port 2 disabled "
Checked the RAM with memtest and hdd with self test from BIOS, says OK.
The major issue is it can't boot from any other devices other than internal harddisk !
Previously I could boot Linux frm DVDs and bootable USB drives.
Windows XP cd when booted gives "NTLDR missing" error.
Kindly help me.

card reader is a usb device, not sure about the fingerprint reader.
one of them has malfunctioned. check the card reader slot for debris.
You may have to open up the  notebook and disconnect the card reader, see if the error stops.

Similar Messages

  • Sites Taking too much time to open and shows error

    hi, 
    i've setup sharepoint 2013 environement correctly and created a site collection everything was working fine but suddenly now when i am trying to open that site collection or central admin site it's taking too much time to open a page but most of the time
    does not open any page or central admin site and shows following error
    event i go to logs folder under 15 hive but nothing useful found please tell me why it takes about 10-12 minutes to open a site or any page and then shows above shown error. 

    This usually happens if you are low on hardware requirements.  Check whether your machine confirms with the required software and hardware requirements.
    https://technet.microsoft.com/en-us/library/cc262485.aspx
    http://sharepoint.stackexchange.com/questions/58370/minimum-real-world-system-requirements-for-sharepoint-2013
    Please remember to up-vote or mark the reply as answer if you find it helpful.

  • Hyperion Planning - Calculation script take too much time to complete

    Dear all,
    I'm using version 11.1.2.1, and i need your help to solve our problem
    I had 5 dimension:
    Entity : Parent Company with > 50 level 0 member
    Items: Capex-Asset Item with > 500 level 0 member
    Account: Capex Account with > 20 level 0 member and 1 member is Exchange_rate
    Currency Dimension: with 3 member: VND, Euro & USD
    And 1 dimension with 2 member: Input & Total
    - I enter data in Input member and with Total member i write a rule:
    Fix (@descendants("Parent Company",0),@descendants("Capex-Asset Item",0),@descendants("Capex Account",0),VND,Total)
    Jan = Jan->Input + (Jan->Input->USD * Jan->Input->USD->Exchange_rate) + (Jan->Input->Euro * Jan->Input->Euro->Exchange_rate);
    Dec= Dec->Input + (Dec->Input->USD * Dec->Input->USD->Exchange_rate) + (Dec->Input->Euro * Dec->Input->Euro->Exchange_rate);
    ENDFIX
    - After running this rule, the data didn't come, but when i enter manually data in 1 row and save -> calculated data came.
    - I think this issue occur because the rule can't create block, so i add 1 command in my rule : SET CREATENONMISSINGBLK ON;
    - The calculated data came after adding new command but it took more than 30 mins to complete.
    Please give me some advices or solutions to turning this rule to run more faster.
    Many thanks for your help....
    Regards,
    SonNT

    You have a block creation issue. The reason it worked after you entered a 1 is because that created the block. If you have a dense dimension on the left (Jan), it will never create the block (unless you have the createblockonequation setting ON). If you turn this setting on, it will take a REALLY long time to run.
    My suggestion to you is that you change the formula so you have a sparse member on the left side of the equation. If you are trying to copy, in some fashion, data from input to total, put total on the left side of the equation and change your fix to Input. You won't need to reference any period as it will run on all the periods because they aren't in the fix. You will basically be looping through the Input blocks, and for every block, you will put the data TO Total.
    Again, assuming you are trying to copy to Total
    Change
    Jan = Jan->Input + (Jan->Input->USD * Jan->Input->USD->Exchange_rate) + (Jan->Input->Euro * Jan->Input->Euro->Exchange_rate);
    To
    Input = Input + (Input->USD * Input->USD->Exchange_rate) + (Input->Euro * Input->Euro->Exchange_rate);
    Again, don't forget to change the fix from Total to Input
    Kyle Goodfriend
    http://www.in2hyperion.com
    h4. Please make sure to assign helpful/answered to responses if applicable - it rewards those who help and benefits the user community.

  • UTL_FILE reqeust taking too much time for completion

    Hi,
    We are running utl_file reqeust.
    the select statement fetching 6000 records, for this it taking 22 mins for completion.
    the code is
    CREATE OR REPLACE PACKAGE BODY "XXC"."XXC_MOD_IN_068_AP_TO_FIS_PKG"
    AS
    * Module Type : PL/SQL
    * Module Name : XXC_MOD_IN_068_AP_FIS_PKG
    * Description : This package is used for AP to Fiscal Interface.
    * Run Env. : SQL*Plus
    * Procedure Name Description
    * XXC_MOD_068_AP_PR XXC_MOD_068_AP_PR Procedure is used to insert transactions
    * into CSV OutPut File from Oracle Account Payables.
    * Calling Module: None
    * Modules Called: XXC_COMMON_INT_PK.INSERT_AUDIT
    * Module Number : MOD_IN_068
    * Known Bugs and Restrictions: none
    * History
    * =======
    * Version Name Date Description of Change
    * 0.1 Sanjeev Khurana 25-JULY-2011 Initial Creation.
    * 0.2 Rohit 09-DEC-2011 Updated header details for the file
    * 0.3 Amit Kulwal 28-AUG-2012 Updated the cursor query for incident 671520
    * 0.4 Swaraj Goud 20-Nov-2012 Updated as per the CR 671520
    | PACKAGE BODY
    -- Actual Code Start Here
    -- Procedure : XXC_MOD_068_AP_PR
    -- Description : XXC_MOD_068_AP_PR Procedure is used to insert transactions
    -- into CSV OUTPUT File from Oracle Account Payables.
    -- Parameters:
    -- Parm Name I/O Description
    -- p_errbuf OUT Error message.
    -- p_retcode OUT Error code. Returns 0 if no errors otherwise returns 1.
    -- p_start_date IN Start Date
    -- p_end_date IN End Date
    PROCEDURE xxc_mod_068_ap_pr (
    p_errbuf OUT VARCHAR2,
    p_retcode OUT NUMBER,
    p_start_date IN VARCHAR2,
    p_end_date IN VARCHAR2
    IS
    -- Define variables and assign default values
    l_sucess_count NUMBER := 0;
    l_error_count NUMBER := 0;
    -- Standard declaration
    l_source VARCHAR2 (10);
    l_target VARCHAR2 (10);
    lc_module_description VARCHAR2 (50)
    := 'MOD_IN_068 - AP to Fiscal';
    l_status CONSTANT VARCHAR2 (50) := 'NEW';
    p_status NUMBER;
    l_batch_id NUMBER;
    l_batch_id_next NUMBER
    := apps_common_out_batch_id_s1.NEXTVAL;
    l_proc_name VARCHAR2 (100) := 'XXC_MOD_IN_068';
    l_request_id NUMBER
    := fnd_global.conc_request_id;
    l_audit_master_id NUMBER := NULL;
    l_mod_code VARCHAR2 (100);
    l_log_type NUMBER := 1; --INFORMATION
    l_det_status_success NUMBER := 0; --SUCCESS
    l_det_status_inprocess NUMBER := 3; --INPROCESS
    l_det_status_rejected NUMBER := 4; --REJECTED
    l_det_status_err NUMBER := 3; --Error
    l_det_status_complete NUMBER := 9; --COMPLETE
    -- Standard who Columns
    l_created_by NUMBER := fnd_global.user_id;
    l_creation_date DATE := SYSDATE;
    l_last_update_date DATE := SYSDATE;
    l_last_update_login NUMBER := fnd_global.user_id;
    v_file UTL_FILE.file_type;
    l_location VARCHAR2 (150);
    l_archive_location VARCHAR2 (150);
    l_date VARCHAR2 (50);
    l_filename VARCHAR2 (50);
    l_open_mode VARCHAR2 (1) := 'W';
    --- l_max_linesize NUMBER := 32767;
    l_max_linesize VARCHAR2 (150); -- Updated 09-Nov-2012
    --Cursor is used to fetch valid records for the interface
    CURSOR get_ap_data_inv
    IS
    SELECT asp.segment1 supplier_ref,
    -- asp.vendor_name supplier_name,
    replace(asp.vendor_name, ',', ' ') supplier_name,
    --aia.invoice_num invoice_number,
    replace(aia.invoice_num, ',','') invoice_number,
    aia.invoice_date,
    aia.invoice_amount amount,
    aia.doc_sequence_value unique_id,
    aia.creation_date date_invoice_entered,
    apsa.due_date date_invoice_paid,
    aia.SOURCE user_id,
    aia.payment_status_flag,
    aia.invoice_type_lookup_code doc_type,
    --aia.description,
    replace(aia.description, ',' , ' ') description,
    apsa.gross_amount
    FROM ap_invoices_all aia,
    ap_suppliers asp,
    apps.ap_payment_schedules_all apsa
    -- apps.iby_payments_all iba,
    -- apps.iby_docs_payable_all dp
    WHERE aia.invoice_id = apsa.invoice_id
    AND aia.vendor_id = asp.vendor_id
    AND aia.org_id = apsa.org_id
    -- AND apsa.payment_status_flag != 'Y' -- commented for CR
    -- AND dp.payment_id = iba.payment_id(+)
    -- AND aia.invoice_id = dp.calling_app_doc_unique_ref2(+)
    -- AND apsa.due_date <= (SYSDATE + 1)
    AND TRUNC (aia.creation_date)
    BETWEEN NVL (fnd_date.canonical_to_date (p_start_date),
    TRUNC (aia.creation_date))
    AND NVL (fnd_date.canonical_to_date (p_end_date),
    TRUNC (aia.creation_date));
         TYPE xxc_tbl IS TABLE OF get_ap_data_inv%ROWTYPE;
    xxc_tbl1 xxc_tbl;
    BEGIN
    l_batch_id := apps_common_out_batch_id_s1.CURRVAL;
    xxc_common_int_pk.insert_audit (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_inprocess,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Process Starts',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    -- Get Module Code
    BEGIN
    SELECT TRIM (fval.flex_value),
    TRIM (SUBSTR (fval.description,
    1,
    INSTR (fval.description, ' -')
    INTO l_source,
    l_mod_code
    FROM fnd_flex_values_vl fval, fnd_flex_value_sets vset
    WHERE vset.flex_value_set_id = fval.flex_value_set_id
    AND vset.flex_value_set_name IN ('XXC_COMM_INT_CONFIG')
    AND fval.enabled_flag = 'Y'
    AND fval.description = lc_module_description;
    EXCEPTION
    WHEN OTHERS
    THEN
    xxc_common_int_pk.insert_audit (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_err,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Error Mode Code',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    raise_application_error (-20045, SQLERRM);
    END;
    --File Location Path for OutPut File
    BEGIN
    SELECT fnd_profile.VALUE ('XXC_MOD_IN_068_AP_OUTBOUND'),
    fnd_profile.VALUE ('XXC_MOD_IN_068_AP_ARCHIVE')
    INTO l_location,
    l_archive_location
    FROM DUAL;
    EXCEPTION
    WHEN OTHERS
    THEN
    xxc_common_int_pk.insert_audit
    (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_rejected,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Profile Value not found',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    END;
    BEGIN
    SELECT TO_CHAR (SYSDATE, 'YYYYMMDDhh24miss')
    INTO l_date
    FROM DUAL;
    EXCEPTION
    WHEN OTHERS
    THEN
    xxc_common_int_pk.insert_audit (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_rejected,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'status not found',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    END;
    l_filename := 'AP_Fiscal_' || l_date || '.csv';
    v_file :=
    UTL_FILE.fopen (LOCATION => l_location,
    filename => l_filename,
    open_mode => l_open_mode,
    max_linesize => l_max_linesize
    -- Changed as per Sarah's email on 9th Decemeber
    /* UTL_FILE.put_line (v_file,
    'SUPPLIER_REF'
    || ','
    || 'SUPPLIER_NAME'
    || ','
    || 'INVOICE_NUMBER'
    || ','
    || 'INVOICE_DATE'
    || ','
    || 'AMOUNT'
    || ','
    || 'UNIQUE_ID'
    || ','
    || 'DATE_INVOICE_ENTERED'
    || ','
    || 'DATE_INVOICE_PAID'
    || ','
    || 'USER_ID'
    || ','
    || 'PAYMENT_STATUS_FLAG'
    || ','
    || 'DOC_TYPE'
    || ','
    || 'DESCRIPTION'
    || ','
    || 'PAYMENT_AMOUNT'
    UTL_FILE.put_line (v_file,
    'SUPPLIERREF'
    || ','
    || 'SUPPLIERNAME'
    || ','
    || 'INVOICENUMBER'
    || ','
    || 'DATE'
    || ','
    || 'AMOUNT'
    || ','
    || 'UNIQUEID'
    || ','
    || 'DATEINVOICEENTERED'
    || ','
    || 'DATEINVOICEPAID'
    || ','
    || 'USERID'
    || ','
    || 'PAYMENTSTATUS'
    || ','
    || 'DOCTYPE'
    || ','
    || 'DESCRIPTION'
    || ','
    || 'PAYMENTAMOUNT');
    UTL_FILE.put_line (v_file,
    'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX'
    || ','
    || 'XX');
                             open get_ap_data_inv;
                             loop
                             fetch get_ap_data_inv bulk collect into xxc_tbl1 limit 6000;
    fnd_file.put_line(fnd_file.log, 'Cursor Count is : '||xxc_tbl1.count);
                             for i in xxc_tbl1.first .. xxc_tbl1.count
    --FOR cur_rec IN get_ap_data_inv
    LOOP
    BEGIN
    --Common package used for proper sequence for Record_id and Bath_id
    l_sucess_count := l_sucess_count + 1;
    --Insert into CSV file
    fnd_file.put_line (fnd_file.LOG, 'Before Utl file');
    UTL_FILE.put_line (v_file,
    xxc_tbl1(i).supplier_ref
    || ','
    || xxc_tbl1(i).supplier_name
    || ','
    || xxc_tbl1(i).invoice_number
    || ','
    || xxc_tbl1(i).invoice_date
    || ','
    || xxc_tbl1(i).amount
    || ','
    || xxc_tbl1(i).unique_id
    || ','
    || xxc_tbl1(i).date_invoice_entered
    || ','
    || xxc_tbl1(i).date_invoice_paid
    || ','
    || xxc_tbl1(i).user_id
    || ','
    || xxc_tbl1(i).payment_status_flag
    || ','
    || xxc_tbl1(i).doc_type
    || ','
    || xxc_tbl1(i).description
    || ','
    || xxc_tbl1(i).gross_amount);
    fnd_file.put_line (fnd_file.LOG,
    'Supplier Reference : ' || xxc_tbl1(i).supplier_ref);
    xxc_common_int_pk.insert_audit
    (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_complete,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Inserting records from AP to Fiscal Successfully',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    EXCEPTION
    WHEN OTHERS
    THEN
    l_error_count := l_error_count + 1;
    fnd_file.put_line (fnd_file.LOG,
    'Error While Inserting from AP to Fiscal '
    || SQLERRM);
    -- Create audit log for AP Inv records insert Exception
    --Insert into the Audit table XXC_COMM_AUDIT_DETAIL_LOG and XXC_COMM_AUDIT_MASTER_LOG
    xxc_common_int_pk.insert_audit
    (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_rejected,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Error While Inserting from AP to Fiscal',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    END;
    END LOOP;
    exit when get_ap_data_inv%NOTFOUND;
    end loop;
         close get_ap_data_inv;
    UTL_FILE.fclose (v_file);
    UTL_FILE.fcopy (l_location,
    l_filename,
    l_archive_location,
    l_filename
    -- Create audit log for Successfully processed records
    -- Procedure call to insert in Audit tables
    xxc_common_int_pk.insert_audit
    (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_complete,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Compeleted Sucessfully AP to Fiscal',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    --Insert into the Audit table XXC_COMM_AUDIT_DETAIL_LOG and XXC_COMM_AUDIT_MASTER_LOG for populating email drop table
    BEGIN
    SELECT transaction_status
    INTO p_status
    FROM xxc_comm_audit_master_log
    WHERE audit_master_id = l_audit_master_id;
    EXCEPTION
    WHEN OTHERS
    THEN
    xxc_common_int_pk.insert_audit (p_batch_id => l_batch_id,
    p_request_id => l_request_id,
    p_source_system => l_source,
    p_proc_name => l_proc_name,
    p_log_type => l_log_type,
    p_det_status => l_det_status_err,
    p_msg_code => NULL,
    p_entity => NULL,
    p_msg_desc => 'Status Error',
    p_mast_request_id => l_request_id,
    p_record_id => NULL,
    p_source => l_source,
    p_target => l_target,
    p_email => NULL,
    p_mod_code => l_mod_code,
    p_audit_master_id => l_audit_master_id
    END;
    IF p_status <> 0
    THEN
    xxc_comm_audit_log_pk.populate_email_drop_table (l_audit_master_id,
    l_batch_id);
    END IF;
    EXCEPTION
    WHEN UTL_FILE.invalid_path
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error (-20000, 'File location is invalid.');
    WHEN UTL_FILE.invalid_mode
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error (-20001,
    'The open_mode parameter in FOPEN is invalid.');
    WHEN UTL_FILE.invalid_filehandle
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error (-20002, 'File handle is invalid.');
    WHEN UTL_FILE.invalid_operation
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error
    (-20003,
    'File could not be opened or operated on as requested.');
    WHEN UTL_FILE.write_error
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error
    (-20005,
    'Operating system error occurred during the write operation.');
    WHEN UTL_FILE.file_open
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error
    (-20008,
    'The requested operation failed because the file is open.');
    WHEN UTL_FILE.invalid_maxlinesize
    THEN
    UTL_FILE.fclose (v_file);
    raise_application_error
    (-20009,
    'The MAX_LINESIZE value for FOPEN() is invalid; it should '
    || 'be within the range 1 to 32767.');
    COMMIT;
    ROLLBACK TO data_extract;
    WHEN OTHERS
    THEN
    raise_application_error (-20045, SQLERRM);
    UTL_FILE.fclose (v_file);
    END xxc_mod_068_ap_pr;
    END xxc_mod_in_068_ap_to_fis_pkg;
    Show Errors
    Iam implemented BULK collect concept in programe,can anyone please suggest how can I imporve performance....
    Thanks,
    Rakesh

      CREATE OR REPLACE PACKAGE BODY "XXC"."XXC_MOD_IN_068_AP_TO_FIS_PKG"
    AS
      PROCEDURE xxc_mod_068_ap_pr (
        p_errbuf                         OUT      VARCHAR2,
        p_retcode                        OUT      NUMBER,
        p_start_date                     IN       VARCHAR2,
        p_end_date                       IN       VARCHAR2
      IS
        -- Define variables and assign default values
        l_sucess_count                          NUMBER := 0;
        l_error_count                           NUMBER := 0;
        -- Standard declaration
        l_source                                VARCHAR2 (10);
        l_target                                VARCHAR2 (10);
        lc_module_description                   VARCHAR2 (50)
                                                     := 'MOD_IN_068 - AP to Fiscal';
        l_status                       CONSTANT VARCHAR2 (50) := 'NEW';
        p_status                                NUMBER;
        l_batch_id                              NUMBER;
        l_batch_id_next                         NUMBER
                                             := apps_common_out_batch_id_s1.NEXTVAL;
        l_proc_name                             VARCHAR2 (100) := 'XXC_MOD_IN_068';
        l_request_id                            NUMBER
                                                      := fnd_global.conc_request_id;
        l_audit_master_id                       NUMBER := NULL;
        l_mod_code                              VARCHAR2 (100);
        l_log_type                              NUMBER := 1;   --INFORMATION
        l_det_status_success                    NUMBER := 0;   --SUCCESS
        l_det_status_inprocess                  NUMBER := 3;   --INPROCESS
        l_det_status_rejected                   NUMBER := 4;   --REJECTED
        l_det_status_err                        NUMBER := 3;   --Error
        l_det_status_complete                   NUMBER := 9;   --COMPLETE
        -- Standard who Columns
        l_created_by                            NUMBER := fnd_global.user_id;
        l_creation_date                         DATE := SYSDATE;
        l_last_update_date                      DATE := SYSDATE;
        l_last_update_login                     NUMBER := fnd_global.user_id;
        v_file                                  UTL_FILE.file_type;
        l_location                              VARCHAR2 (150);
        l_archive_location                      VARCHAR2 (150);
        l_date                                  VARCHAR2 (50);
        l_filename                              VARCHAR2 (50);
        l_open_mode                             VARCHAR2 (1) := 'W';
    --- l_max_linesize                          NUMBER := 32767;
        l_max_linesize                          VARCHAR2 (150); -- Updated 09-Nov-2012
        --Cursor is used to fetch valid records for the interface
        CURSOR get_ap_data_inv
        IS
          SELECT       asp.segment1 supplier_ref,
                 -- asp.vendor_name supplier_name,
                 replace(asp.vendor_name, ',', ' ') supplier_name,
                 --aia.invoice_num invoice_number,
                 replace(aia.invoice_num, ',','') invoice_number,
                 aia.invoice_date,
                 aia.invoice_amount amount,
                 aia.doc_sequence_value unique_id,
                 aia.creation_date date_invoice_entered,
                 apsa.due_date date_invoice_paid,
                 aia.SOURCE user_id,
                 aia.payment_status_flag,
                 aia.invoice_type_lookup_code doc_type,
                 --aia.description,
                 replace(aia.description, ',' , ' ') description,
                 apsa.gross_amount
          FROM   ap_invoices_all aia,
                 ap_suppliers asp,
                 apps.ap_payment_schedules_all apsa
                -- apps.iby_payments_all iba,
            --     apps.iby_docs_payable_all dp
          WHERE  aia.invoice_id = apsa.invoice_id
          AND    aia.vendor_id  = asp.vendor_id
          AND    aia.org_id     = apsa.org_id
      --  AND    apsa.payment_status_flag != 'Y'   -- commented for CR
      --  AND    dp.payment_id = iba.payment_id(+)
      --  AND    aia.invoice_id = dp.calling_app_doc_unique_ref2(+)
      --  AND    apsa.due_date <= (SYSDATE + 1)
          AND    TRUNC (aia.creation_date)
                   BETWEEN NVL (fnd_date.canonical_to_date (p_start_date),
                                TRUNC (aia.creation_date))
                       AND NVL (fnd_date.canonical_to_date (p_end_date),
                                TRUNC (aia.creation_date));
                                     TYPE xxc_tbl IS TABLE OF get_ap_data_inv%ROWTYPE;
                                  xxc_tbl1 xxc_tbl;
      BEGIN
        l_batch_id                      := apps_common_out_batch_id_s1.CURRVAL;
        xxc_common_int_pk.insert_audit (p_batch_id                        => l_batch_id,
                                        p_request_id                      => l_request_id,
                                        p_source_system                   => l_source,
                                        p_proc_name                       => l_proc_name,
                                        p_log_type                        => l_log_type,
                                        p_det_status                      => l_det_status_inprocess,
                                        p_msg_code                        => NULL,
                                        p_entity                          => NULL,
                                        p_msg_desc                        => 'Process Starts',
                                        p_mast_request_id                 => l_request_id,
                                        p_record_id                       => NULL,
                                        p_source                          => l_source,
                                        p_target                          => l_target,
                                        p_email                           => NULL,
                                        p_mod_code                        => l_mod_code,
                                        p_audit_master_id                 => l_audit_master_id
    -- Get Module Code
        BEGIN
          SELECT TRIM (fval.flex_value),
                 TRIM (SUBSTR (fval.description,
                               1,
                               INSTR (fval.description, ' -')
          INTO   l_source,
                 l_mod_code
          FROM   fnd_flex_values_vl fval, fnd_flex_value_sets vset
          WHERE  vset.flex_value_set_id = fval.flex_value_set_id
          AND    vset.flex_value_set_name IN ('XXC_COMM_INT_CONFIG')
          AND    fval.enabled_flag = 'Y'
          AND    fval.description = lc_module_description;
        EXCEPTION
          WHEN OTHERS
          THEN
            xxc_common_int_pk.insert_audit (p_batch_id                        => l_batch_id,
                                            p_request_id                      => l_request_id,
                                            p_source_system                   => l_source,
                                            p_proc_name                       => l_proc_name,
                                            p_log_type                        => l_log_type,
                                            p_det_status                      => l_det_status_err,
                                            p_msg_code                        => NULL,
                                            p_entity                          => NULL,
                                            p_msg_desc                        => 'Error Mode Code',
                                            p_mast_request_id                 => l_request_id,
                                            p_record_id                       => NULL,
                                            p_source                          => l_source,
                                            p_target                          => l_target,
                                            p_email                           => NULL,
                                            p_mod_code                        => l_mod_code,
                                            p_audit_master_id                 => l_audit_master_id
            raise_application_error (-20045, SQLERRM);
        END;
    --File Location Path for OutPut File
        BEGIN
          SELECT fnd_profile.VALUE ('XXC_MOD_IN_068_AP_OUTBOUND'),
                 fnd_profile.VALUE ('XXC_MOD_IN_068_AP_ARCHIVE')
          INTO   l_location,
                 l_archive_location
          FROM   DUAL;
        EXCEPTION
          WHEN OTHERS
          THEN
            xxc_common_int_pk.insert_audit
                                          (p_batch_id                        => l_batch_id,
                                           p_request_id                      => l_request_id,
                                           p_source_system                   => l_source,
                                           p_proc_name                       => l_proc_name,
                                           p_log_type                        => l_log_type,
                                           p_det_status                      => l_det_status_rejected,
                                           p_msg_code                        => NULL,
                                           p_entity                          => NULL,
                                           p_msg_desc                        => 'Profile Value not found',
                                           p_mast_request_id                 => l_request_id,
                                           p_record_id                       => NULL,
                                           p_source                          => l_source,
                                           p_target                          => l_target,
                                           p_email                           => NULL,
                                           p_mod_code                        => l_mod_code,
                                           p_audit_master_id                 => l_audit_master_id
        END;
        BEGIN
          SELECT TO_CHAR (SYSDATE, 'YYYYMMDDhh24miss')
          INTO   l_date
          FROM   DUAL;
        EXCEPTION
          WHEN OTHERS
          THEN
            xxc_common_int_pk.insert_audit (p_batch_id                        => l_batch_id,
                                            p_request_id                      => l_request_id,
                                            p_source_system                   => l_source,
                                            p_proc_name                       => l_proc_name,
                                            p_log_type                        => l_log_type,
                                            p_det_status                      => l_det_status_rejected,
                                            p_msg_code                        => NULL,
                                            p_entity                          => NULL,
                                            p_msg_desc                        => 'status not found',
                                            p_mast_request_id                 => l_request_id,
                                            p_record_id                       => NULL,
                                            p_source                          => l_source,
                                            p_target                          => l_target,
                                            p_email                           => NULL,
                                            p_mod_code                        => l_mod_code,
                                            p_audit_master_id                 => l_audit_master_id
        END;
        l_filename                      := 'AP_Fiscal_' || l_date || '.csv';
        v_file                          :=
          UTL_FILE.fopen (LOCATION                          => l_location,
                          filename                          => l_filename,
                          open_mode                         => l_open_mode,
                          max_linesize                      => l_max_linesize
                           -- Changed as per Sarah's email on 9th Decemeber
        /* UTL_FILE.put_line (v_file,
                               'SUPPLIER_REF'
                            || ','
                            || 'SUPPLIER_NAME'
                            || ','
                            || 'INVOICE_NUMBER'
                            || ','
                            || 'INVOICE_DATE'
                            || ','
                            || 'AMOUNT'
                            || ','
                            || 'UNIQUE_ID'
                            || ','
                            || 'DATE_INVOICE_ENTERED'
                            || ','
                            || 'DATE_INVOICE_PAID'
                            || ','
                            || 'USER_ID'
                            || ','
                            || 'PAYMENT_STATUS_FLAG'
                            || ','
                            || 'DOC_TYPE'
                            || ','
                            || 'DESCRIPTION'
                            || ','
                            || 'PAYMENT_AMOUNT'
        UTL_FILE.put_line (v_file,
                              'SUPPLIERREF'
                           || ','
                           || 'SUPPLIERNAME'
                           || ','
                           || 'INVOICENUMBER'
                           || ','
                           || 'DATE'
                           || ','
                           || 'AMOUNT'
                           || ','
                           || 'UNIQUEID'
                           || ','
                           || 'DATEINVOICEENTERED'
                           || ','
                           || 'DATEINVOICEPAID'
                           || ','
                           || 'USERID'
                           || ','
                           || 'PAYMENTSTATUS'
                           || ','
                           || 'DOCTYPE'
                           || ','
                           || 'DESCRIPTION'
                           || ','
                           || 'PAYMENTAMOUNT');
        UTL_FILE.put_line (v_file,
                              'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX'
                           || ','
                           || 'XX');
                                open get_ap_data_inv;
                                loop
                                fetch get_ap_data_inv bulk collect into xxc_tbl1 limit 6000;
                 fnd_file.put_line(fnd_file.log, 'Cursor Count is : '||xxc_tbl1.count);
                                for i in xxc_tbl1.first .. xxc_tbl1.count
        --FOR cur_rec IN get_ap_data_inv
        LOOP
          BEGIN
            --Common package used for proper sequence for Record_id and Bath_id
            l_sucess_count                  := l_sucess_count + 1;
            --Insert into CSV file
            fnd_file.put_line (fnd_file.LOG, 'Before Utl file');
            UTL_FILE.put_line (v_file,
                                  xxc_tbl1(i).supplier_ref
                               || ','
                               || xxc_tbl1(i).supplier_name
                               || ','
                               || xxc_tbl1(i).invoice_number
                               || ','
                               || xxc_tbl1(i).invoice_date
                               || ','
                               || xxc_tbl1(i).amount
                               || ','
                               || xxc_tbl1(i).unique_id
                               || ','
                               ||  xxc_tbl1(i).date_invoice_entered
                               || ','
                               ||  xxc_tbl1(i).date_invoice_paid
                               || ','
                               ||  xxc_tbl1(i).user_id
                               || ','
                               ||  xxc_tbl1(i).payment_status_flag
                               || ','
                               ||  xxc_tbl1(i).doc_type
                               || ','
                               ||  xxc_tbl1(i).description
                               || ','
                               ||  xxc_tbl1(i).gross_amount);
            fnd_file.put_line (fnd_file.LOG,
                               'Supplier Reference : ' ||  xxc_tbl1(i).supplier_ref);
            xxc_common_int_pk.insert_audit
                  (p_batch_id                        => l_batch_id,
                   p_request_id                      => l_request_id,
                   p_source_system                   => l_source,
                   p_proc_name                       => l_proc_name,
                   p_log_type                        => l_log_type,
                   p_det_status                      => l_det_status_complete,
                   p_msg_code                        => NULL,
                   p_entity                          => NULL,
                   p_msg_desc                        => 'Inserting records from AP to Fiscal Successfully',
                   p_mast_request_id                 => l_request_id,
                   p_record_id                       => NULL,
                   p_source                          => l_source,
                   p_target                          => l_target,
                   p_email                           => NULL,
                   p_mod_code                        => l_mod_code,
                   p_audit_master_id                 => l_audit_master_id
          EXCEPTION
            WHEN OTHERS
            THEN
              l_error_count                   := l_error_count + 1;
              fnd_file.put_line (fnd_file.LOG,
                                    'Error While Inserting from AP to Fiscal '
                                 || SQLERRM);
                 -- Create audit log for AP Inv records insert Exception
              --Insert into the Audit table XXC_COMM_AUDIT_DETAIL_LOG and XXC_COMM_AUDIT_MASTER_LOG
              xxc_common_int_pk.insert_audit
                           (p_batch_id                        => l_batch_id,
                            p_request_id                      => l_request_id,
                            p_source_system                   => l_source,
                            p_proc_name                       => l_proc_name,
                            p_log_type                        => l_log_type,
                            p_det_status                      => l_det_status_rejected,
                            p_msg_code                        => NULL,
                            p_entity                          => NULL,
                            p_msg_desc                        => 'Error While Inserting from AP to Fiscal',
                            p_mast_request_id                 => l_request_id,
                            p_record_id                       => NULL,
                            p_source                          => l_source,
                            p_target                          => l_target,
                            p_email                           => NULL,
                            p_mod_code                        => l_mod_code,
                            p_audit_master_id                 => l_audit_master_id
          END;
        END LOOP;
         exit when get_ap_data_inv%NOTFOUND;
        end loop;
         close get_ap_data_inv;
        UTL_FILE.fclose (v_file);
        UTL_FILE.fcopy (l_location,
                        l_filename,
                        l_archive_location,
                        l_filename
          -- Create audit log for Successfully processed records
        -- Procedure call to insert in Audit tables
        xxc_common_int_pk.insert_audit
                               (p_batch_id                        => l_batch_id,
                                p_request_id                      => l_request_id,
                                p_source_system                   => l_source,
                                p_proc_name                       => l_proc_name,
                                p_log_type                        => l_log_type,
                                p_det_status                      => l_det_status_complete,
                                p_msg_code                        => NULL,
                                p_entity                          => NULL,
                                p_msg_desc                        => 'Compeleted Sucessfully AP to Fiscal',
                                p_mast_request_id                 => l_request_id,
                                p_record_id                       => NULL,
                                p_source                          => l_source,
                                p_target                          => l_target,
                                p_email                           => NULL,
                                p_mod_code                        => l_mod_code,
                                p_audit_master_id                 => l_audit_master_id
        --Insert into the Audit table XXC_COMM_AUDIT_DETAIL_LOG and XXC_COMM_AUDIT_MASTER_LOG for populating email drop table
        BEGIN
          SELECT transaction_status
          INTO   p_status
          FROM   xxc_comm_audit_master_log
          WHERE  audit_master_id = l_audit_master_id;
        EXCEPTION
          WHEN OTHERS
          THEN
            xxc_common_int_pk.insert_audit (p_batch_id                        => l_batch_id,
                                            p_request_id                      => l_request_id,
                                            p_source_system                   => l_source,
                                            p_proc_name                       => l_proc_name,
                                            p_log_type                        => l_log_type,
                                            p_det_status                      => l_det_status_err,
                                            p_msg_code                        => NULL,
                                            p_entity                          => NULL,
                                            p_msg_desc                        => 'Status Error',
                                            p_mast_request_id                 => l_request_id,
                                            p_record_id                       => NULL,
                                            p_source                          => l_source,
                                            p_target                          => l_target,
                                            p_email                           => NULL,
                                            p_mod_code                        => l_mod_code,
                                            p_audit_master_id                 => l_audit_master_id
        END;
        IF p_status <> 0
        THEN
          xxc_comm_audit_log_pk.populate_email_drop_table (l_audit_master_id,
                                                           l_batch_id);
        END IF;
      EXCEPTION
        WHEN UTL_FILE.invalid_path
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error (-20000, 'File location is invalid.');
        WHEN UTL_FILE.invalid_mode
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error (-20001,
                                   'The open_mode parameter in FOPEN is invalid.');
        WHEN UTL_FILE.invalid_filehandle
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error (-20002, 'File handle is invalid.');
        WHEN UTL_FILE.invalid_operation
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error
                           (-20003,
                            'File could not be opened or operated on as requested.');
        WHEN UTL_FILE.write_error
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error
                     (-20005,
                      'Operating system error occurred during the write operation.');
        WHEN UTL_FILE.file_open
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error
                        (-20008,
                         'The requested operation failed because the file is open.');
        WHEN UTL_FILE.invalid_maxlinesize
        THEN
          UTL_FILE.fclose (v_file);
          raise_application_error
                    (-20009,
                        'The MAX_LINESIZE value for FOPEN() is invalid; it should '
                     || 'be within the range 1 to 32767.');
          COMMIT;
          ROLLBACK TO data_extract;
        WHEN OTHERS
        THEN
          raise_application_error (-20045, SQLERRM);
          UTL_FILE.fclose (v_file);
      END xxc_mod_068_ap_pr;
    END xxc_mod_in_068_ap_to_fis_pkg;Please verify...

  • Execution script too much time than before

    Hi experts,
    We run tunning script on daily basis which analyze multiple table of oracle ebs 11i oracle 10g and we also have execute schema statistics,
    Issue is this it take too much time to complete some time more than 12 hrs. It looks like hang on gathereing statistics Inv and ont schema.
    Kindly give me solution
    e.g
    analyze table inv.mtl_transaction_types compute statistics;
    analyze table wip.wip_schedule_groups compute statistics;
    analyze table WSH.WSH_DELIVERY_DETAILS compute statistics;
    exec DBMS_STATS.GATHER_SCHEMA_STATS('APPS'); ---APPS, APPLSYS, INV,ONT,WSH, SYS, SYSTEM, GL,HR,AR,AP,CE,WIP,
    ---PO,MSC,PJI
    exec DBMS_STATS.GATHER_SCHEMA_STATS('APPLSYS');
    exec DBMS_STATS.GATHER_SCHEMA_STATS('INV');
    exec DBMS_STATS.GATHER_SCHEMA_STATS('ONT');
    exec DBMS_STATS.GATHER_SCHEMA_STATS('WSH');
    Now its taking more than 15 minut and not yet complete.
    Thanks

    but to gather statisics for some schema it takes too much time. First it take 3 hrs for whole script now it is taking more than 15 hrs and some scheam go to
    complete and some are still hang then we have to cancel this job.
    I also experiment to execute one by one schema at that time it is completed normally.Have you done any changes recently (i.e. upgrade, installed new patches, data load ..etc)?
    Have you tried to use different parameters? -- Definition of Parameters Used in Gather Schema Statistics Program [ID 556466.1]
    You could enable trace and generate the TKPROF file to find out why it takes that long to gather schema statistics.
    FAQ: Common Tracing Techniques within the Oracle Applications 11i/R12 [ID 296559.1]
    Enabling Concurrent Program Traces [ID 123222.1]
    How To Get Level 12 Trace And FND Debug File For Concurrent Programs [ID 726039.1]
    I would also suggest you log a SR as "Gather Schema Statistics" is a seeded concurrent program and Oracle Support should be able to help.
    Thanks,
    Hussein

  • Client import taking too much time

    hi all,
    i am importing a client , i it has complete copy table 19,803 of 19,803 but for last four hours its status is processing
    scc3
    Target Client           650
    Copy Type               Client Import Post-Proc
    Profile                 SAP_CUST
    Status                  Processing...
    User                    SAP*
    Start on                24.05.2009 / 15:08:03
    Last Entry on           24.05.2009 / 15:36:25
    Current Action:         Post Processing
    -  Last Exit Program    RGBCFL01
    Transport Requests
    - Client-Specific       PRDKT00004
    - Texts                 PRDKX00004
    Statistics for this Run
    - No. of Tables             19803 of     19803
    - Deleted Lines                 7
    - Copied Lines                  0
    sm50
    1 DIA 542           Running Yes             SAPLTHFB 650 SAP*     
    7 BGD 4172   Running Yes 11479  RGTBGD23 650 SAP* Sequential Read     D010INC
    sm66
    Server  No. Type PID Status  Reason Sem Start Error CPU Time   User Report   Action          Table
    prdsap_PRD_00  7  BTC 4172 Running   Yes    11711 SAP* RGTBGD23 Sequential Read D010INC
    plz guide me why it is taking too much time , while it has finished most of the things
    best regard
    Khan

    The import is in post processing. It digs through all the documents and adapts them to the new client. Most of the tables in the application area have a "MANDT" (= client) field which needs to be changed. Depending of the size of the client this can take a huge amount of time.
    You can try to improve the speed by updating the table statistics for table D010INC.
    Markus

  • SAP GUI taking too much time to open transactions

    Hi guys,
    i have done system copy from Production to Quality server completely.
    after that i started SAP on Quality server, it is taking too much time for opening SAP transactions (it is going in compilation mode).
    i started SGEN, but it is giving time_out errors. please help me on this issue.
    MY hardware details on quality server
    operating system : SuSE Linux 10 SP2
    Database : 10.2.0.2
    SAP : ECC 6.0  SR2
    RAM size : 8 GB
    Hard disk space : 500 GB
    swap space : 16 GB.
    regards
    Ramesh

    Hi,
    >i started SGEN, but it is giving time_out errors. please help me on this issue.
    You are supposed to run SGEN as a batch job and so, it should be possible  to get time out errors.
    I've seen Full SGEN lasting from 3 hours on high end systems up to 8 full days on PC hardware...
    Regards,
    Olivier

  • Itunes takes too much time to open and don´t recogneize my iphone

    itunes takes too much time to open and don´t recogneize my iphone
    I´ve alredy unistall and re install and didn´t work

    5-10 minutes is normal for me when opening iTunes. But then my library is stored on network drive which slows everything down.
    There has been many posts about problems with RAID and iTunes. Don't know if any of those relate.

  • PRS600 ereader taking too much time to load all the books

    Whenever I open My PRS600 from complete shutdown it takes too much time to load all the files which are present in Memory card. Is anyone else facing same issue?

    Hello,
    Recently, I have been having some serious issues when it comes to my Sony PRS-600. I am running on a Windows 7 64-bit, and an updated Reader Library 3.3
    The issue comes when transferring books to the e-reader from the library, and from the e-reader to a collection. The problem is that the software becomes intolerably slow while it's processing the command. The Reader Library software window grays out, displays (Not Working) and if clicked on it, shades to a white color and displays either "cancel operation" or "wait until program responds". If I do close the operation, it appears as if the e-reader doesn't follow the operation and still displays "Do not disconnect". Since I do not see any other way to disconnect (other than the eject option), I remove the USB plug which causes a bit more issues with the reader (such as removing all of my collections, for example!).
    But anyway, that's not the main issue here. The main issue is that the book transferring is really slow. I need to wait a couple of minutes (or even more) just for the software to process. Moving just 1 MB of data requires so much time as if it's 1 GB. Sometimes it's random and does it fast, and sometimes the application is better not to be dealt with at all while it's processing the command. If I would inspect My Computer, the simple loading of the e-reader storage icons and information would make the Windows Explorer to "crash" (e.g, close all windows and then reopen them). It just happens that in all randomness even the creation of a collection makes the software slow.
    So to recap: the reader software is slow when adding and moving books.
    I hope someone will help me resolve this annoyance.
    Thank you,
    KQ

  • [Solved] Dolphin taking too much time to respond

    Dolphin, while used by an ordinary user, takes too much time to respond, especially when firefox is running. The time it takes to initialize is also very lengthy. At times, it does not even open up and even if active, it hangs for quite sometime before being responsive again while trying to select a file or two.
    At the same time, it works rather faster when called as super user through command line using $sudo dolphin
    But then, the console displays lot of error messages as follows.
    sebinaj ~
    $ sudo dolphin
    Error: "/var/tmp/kdecache-sebinaj1Jz46I" is owned by uid 1002 instead of uid 0.
    Error: "/tmp/kde-sebinaj" is owned by uid 1002 instead of uid 0.
    sebinaj ~
    $ Error: "/tmp/ksocket-sebinaj" is owned by uid 1002 instead of uid 0.
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Invalid iterator."
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Invalid iterator."
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Unsupported operation (2)": "Invalid model"
    "/usr/bin/dolphin(3298)" Error in thread 140247744997200 : "Invalid iterator."
    Error: alias title requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#title, http://www.semanticdesktop.org/ontologies/2007/03/22/nco#title
    Error: alias comment requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#comment, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#comment
    Error: alias count requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#count, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#count
    Error: alias created requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#created, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#created
    Error: alias description requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#description, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#description
    Error: alias duration requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#duration, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#duration
    Error: alias encoding requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#encoding, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#encoding
    Error: alias role requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nco#role, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#role
    Error: alias url requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nco#url, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#url
    Error: alias version requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#version, http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#version
    Error: alias bitsPerSample requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#bitsPerSample, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#bitsPerSample
    Error: alias copyright requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#copyright, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#copyright
    Error: alias date requested by several properties: http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#date, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#date
    Error: alias dateTime requested by several properties: http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#dateTime, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#dateTime
    Error: alias geo requested by several properties: http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#geo, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#geo
    Error: alias height requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#height, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#height
    Error: alias width requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#width, http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#width
    Error: alias date requested by several properties: http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#date, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#date
    Error: alias fileOwner requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#fileOwner, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#fileOwner
    Error: alias language requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#language, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#language
    Error: alias length requested by several properties: http://www.semanticdesktop.org/ontologies/2007/05/10/nexif#length, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#length
    Error: alias publisher requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nco#publisher, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#publisher
    Error: alias title requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#title, http://www.semanticdesktop.org/ontologies/2007/05/10/nid3#title
    Error: alias contributor requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nco#contributor, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#contributor
    Error: alias created requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#created, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#created
    Error: alias creator requested by several properties: http://www.semanticdesktop.org/ontologies/2007/03/22/nco#creator, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#creator
    Error: alias description requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#description, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#description
    Error: alias identifier requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#identifier, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#identifier
    Error: alias lastModified requested by several properties: http://www.semanticdesktop.org/ontologies/2007/04/02/ncal#lastModified, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#lastModified
    Error: alias version requested by several properties: http://www.semanticdesktop.org/ontologies/2007/01/19/nie#version, http://www.semanticdesktop.org/ontologies/2007/08/15/nao#version
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#fileExtension' is not defined in any rdfs ontology database.
    WARNING: field 'http://strigi.sf.net/ontologies/0.9#debugParseError' is not defined in any rdfs ontology database.
    /usr/lib/strigi/strigiea_ics.so
    /usr/lib/strigi/strigiea_jpeg.so
    /usr/lib/strigi/strigiea_vcf.so
    /usr/lib/strigi/strigila_cpp.so
    /usr/lib/strigi/strigila_deb.so
    /usr/lib/strigi/strigila_diff.so
    /usr/lib/strigi/strigila_mobi.so
    /usr/lib/strigi/strigila_namespaceharvester.so
    /usr/lib/strigi/strigila_po.so
    /usr/lib/strigi/strigila_txt.so
    /usr/lib/strigi/strigila_xpm.so
    /usr/lib/strigi/strigita_au.so
    /usr/lib/strigi/strigita_audible.so
    /usr/lib/strigi/strigita_avi.so
    /usr/lib/strigi/strigita_dds.so
    /usr/lib/strigi/strigita_dvi.so
    /usr/lib/strigi/strigita_font.so
    /usr/lib/strigi/strigita_gif.so
    /usr/lib/strigi/strigita_ico.so
    /usr/lib/strigi/strigita_mp4.so
    /usr/lib/strigi/strigita_pcx.so
    /usr/lib/strigi/strigita_rgb.so
    /usr/lib/strigi/strigita_sid.so
    /usr/lib/strigi/strigita_ts.so
    /usr/lib/strigi/strigita_wav.so
    /usr/lib/strigi/strigita_xbm.so
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#usesNamespace' is not defined in any rdfs ontology database.
    WARNING: field 'translation.total' is not defined in any rdfs ontology database.
    WARNING: field 'translation.translated' is not defined in any rdfs ontology database.
    WARNING: field 'translation.untranslated' is not defined in any rdfs ontology database.
    WARNING: field 'translation.obsolete' is not defined in any rdfs ontology database.
    WARNING: field 'diff.stats.modify_file_count' is not defined in any rdfs ontology database.
    WARNING: field 'diff.first_modify_file' is not defined in any rdfs ontology database.
    WARNING: field 'content.format_subtype' is not defined in any rdfs ontology database.
    WARNING: field 'content.generator' is not defined in any rdfs ontology database.
    WARNING: field 'diff.stats.hunk_count' is not defined in any rdfs ontology database.
    WARNING: field 'diff.stats.insert_line_count' is not defined in any rdfs ontology database.
    WARNING: field 'diff.stats.modify_line_count' is not defined in any rdfs ontology database.
    WARNING: field 'diff.stats.delete_line_count' is not defined in any rdfs ontology database.
    WARNING: field 'translation.fuzzy' is not defined in any rdfs ontology database.
    WARNING: field 'translation.last_translator' is not defined in any rdfs ontology database.
    WARNING: field 'translation.translation_date' is not defined in any rdfs ontology database.
    WARNING: field 'translation.source_date' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/2007/03/22/nfo#colorCount' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#formatSubtype' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nfo#bitsPerSample' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#audioSampleDataType' is not defined in any rdfs ontology database.
    WARNING: field 'content.mime_type' is not defined in any rdfs ontology database.
    WARNING: field 'audio.title' is not defined in any rdfs ontology database.
    WARNING: field 'audio.artist' is not defined in any rdfs ontology database.
    WARNING: field 'todo.audio.narrator' is not defined in any rdfs ontology database.
    WARNING: field 'media.codec' is not defined in any rdfs ontology database.
    WARNING: field 'todo.audible.user_id' is not defined in any rdfs ontology database.
    WARNING: field 'todo.audible.user_alias' is not defined in any rdfs ontology database.
    WARNING: field 'audio.duration' is not defined in any rdfs ontology database.
    WARNING: field 'content.description' is not defined in any rdfs ontology database.
    WARNING: field 'content.copyright' is not defined in any rdfs ontology database.
    WARNING: field 'content.keyword' is not defined in any rdfs ontology database.
    WARNING: field 'content.creation_time' is not defined in any rdfs ontology database.
    WARNING: field 'content.maintainer' is not defined in any rdfs ontology database.
    WARNING: field 'content.ID' is not defined in any rdfs ontology database.
    WARNING: field 'audio.channel_count' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nfo#colorDepth' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#colorSpace' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#compressionAlgorithm' is not defined in any rdfs ontology database.
    WARNING: field 'font.family' is not defined in any rdfs ontology database.
    WARNING: field 'font.weight' is not defined in any rdfs ontology database.
    WARNING: field 'font.slant' is not defined in any rdfs ontology database.
    WARNING: field 'font.width' is not defined in any rdfs ontology database.
    WARNING: field 'font.spacing' is not defined in any rdfs ontology database.
    WARNING: field 'font.foundry' is not defined in any rdfs ontology database.
    WARNING: field 'content.version' is not defined in any rdfs ontology database.
    WARNING: field 'content.genre' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_trackNumber' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_discNumber' is not defined in any rdfs ontology database.
    WARNING: field 'content.author' is not defined in any rdfs ontology database.
    WARNING: field 'content.comment' is not defined in any rdfs ontology database.
    WARNING: field 'audio.album' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_audio.albumartist' is not defined in any rdfs ontology database.
    WARNING: field 'content.links' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_content.purchaser' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_content.purchasedate' is not defined in any rdfs ontology database.
    WARNING: field 'media.duration' is not defined in any rdfs ontology database.
    WARNING: field 'TODO_video.duration' is not defined in any rdfs ontology database.
    WARNING: field 'av.audio_codec' is not defined in any rdfs ontology database.
    WARNING: field 'av.video_codec' is not defined in any rdfs ontology database.
    WARNING: field 'content.thumbnail' is not defined in any rdfs ontology database.
    WARNING: field 'user.rating' is not defined in any rdfs ontology database.
    WARNING: field 'image.width' is not defined in any rdfs ontology database.
    WARNING: field 'image.height' is not defined in any rdfs ontology database.
    WARNING: field 'media.sample_rate' is not defined in any rdfs ontology database.
    WARNING: field 'media.sample_format' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#artist' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#albumTrackCount' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#musicAlbum' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#genre' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#composer' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#trackNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#setNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#performer' is not defined in any rdfs ontology database.
    WARNING: field 'http://www.semanticdesktop.org/ontologies/nmm#internationalStandardRecordingCode' is not defined in any rdfs ontology database.
    WARNING: field 'Product Id' is not defined in any rdfs ontology database.
    WARNING: field 'Events' is not defined in any rdfs ontology database.
    WARNING: field 'Journals' is not defined in any rdfs ontology database.
    WARNING: field 'Todos' is not defined in any rdfs ontology database.
    WARNING: field 'Todos Completed' is not defined in any rdfs ontology database.
    WARNING: field 'Todos Overdue' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#ccdWidth' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#focusDistance' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#targetQuality' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#givenName' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#familyName' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#emailAddress' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#homepageContactURL' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#contentComment' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#cellPhoneNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#homePhoneNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#workPhoneNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#faxPhoneNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#phoneNumber' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#homePostalAddress' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#workPostalAddress' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#postalAddress' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#honorificPrefix' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#honorificSuffix' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#subject' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#title' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#author' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#description' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#copyright' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#isContentEncrypted' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#contentKeyword' is not defined in any rdfs ontology database.
    WARNING: field 'http://freedesktop.org/standards/xesam/1.0/core#paragraphCount' is not defined in any rdfs ontology database.
    WARNING: field 'http://rdf.openmolecules.net/0.9#moleculeCount' is not defined in any rdfs ontology database.
    kDebugStream called after destruction (from void KDirWatchPrivate::removeEntry(KDirWatch*, KDirWatchPrivate::Entry*, KDirWatchPrivate::Entry*) file /home/phil/kdemod/core/kdelibs/src/kdelibs-4.3.3/kio/kio/kdirwatch.cpp line 901)
    Cancelled INotify (fd 9, 1) for "/home/sebinaj/.local/share"
    ^C
    I am using KDEmod + Arch
    Last edited by absolutevoid (2009-11-05 17:23:59)

    There's a large thread around about this Dolphin problem.
    Disabling Nepomuk in System Settings has proved to be the
    cure in many cases.
    Deej

  • Azure Data Sync Log Take Too Much Time to Load

    Hi,
    I've found that the Azure Sync Log inside the Portal take too much time to load if my sync group was created long time ago. i.e. for those new sync group with only a few logs, it will take around 5 seconds for loading. However, I've many sync group created
    more than 1 year, and it will take me up to 10 minutes for viewing the first page of the logs.
    It seems the Portal will download ALL the logs to my browser even I just want to view the most recent logs (maybe I just want to view the last 10 records, but it will download the whole 5000 pages?)
    It will be great if the Portal only fetch the data of the page I am currently viewing, instead of fetching all the logs.
    Is there any ways to speed up the log loading time?
    Michael Yung

    Hello,
    I suggest you post a feedback on
    Azure SQL Data Sync feedback forum.
    All of the feedback you share in the forum will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
    Regards,
    Fanny Liu
    Fanny Liu
    TechNet Community Support

  • Sun cluster 3.2 - resource hasstorageplus taking too much time to start

    I have a disk resource called "data" that takes too much time to startup when performing a switchover. Any idea what may control this ?
    Jan 28 20:28:01 hnmdb02 Cluster.Framework: [ID 801593 daemon.notice] stdout: becoming primary for data
    Jan 28 20:28:02 hnmdb02 Cluster.RGM.rgmd: [ID 350207 daemon.notice] 24 fe_rpc_command: cmd_type(enum):<3>:cmd=<null>:tag=<hnmdb.data.10>: Calling security_clnt_connect(..., host=<hnmdb02>, sec_type {0:WEAK, 1:STRONG, 2:DES} =<0>, ...)
    Jan 28 20:28:02 hnmdb02 Cluster.RGM.rgmd: [ID 316625 daemon.notice] Timeout monitoring on method tag <hnmdb.data.10> has been resumed.
    Jan 28 20:34:57 hnmdb02 Cluster.RGM.rgmd: [ID 515159 daemon.notice] method <hastorageplus_prenet_start> completed successfully for resource <data>, resource group <hnmdb>, node <hnmdb02>, time used: 23% of timeout <1800 seconds>

    heldermo wrote:
    I have a disk resource called "data" that takes too much time to startup when performing a switchover. Any idea what may control this ?I'm not sure how this is supposed to be related to Messaging Server. I suggest you ask your question in the Cluster forum:
    http://forums.sun.com/forum.jspa?forumID=842
    Regards,
    Shane.

  • SELECT query takes too much time! Y?

    Plz find my SELECT query below:
    select w~mandt
    wvbeln wposnr wmeins wmatnr wwerks wnetwr
    wkwmeng wvrkme wmatwa wcharg w~pstyv
    wposar wprodh wgrkor wantlf wkztlf wlprio
    wvstel wroute wumvkz wumvkn wabgru wuntto
    wawahr werdat werzet wfixmg wprctr wvpmat
    wvpwrk wmvgr1 wmvgr2 wmvgr3 wmvgr4 wmvgr5
    wbedae wcuobj w~mtvfp
    xetenr xwmeng xbmeng xettyp xwepos xabart
    x~edatu
    xtddat xmbdat xlddat xwadat xabruf xetart
    x~ezeit
    into table t_vbap
    from vbap as w
    inner join vbep as x on xvbeln = wvbeln and
    xposnr = wposnr and
    xmandt = wmandt
    where
    ( ( werdat > pre_dat ) and ( werdat <= w_date ) ) and
    ( ( ( erdat > pre_dat and erdat < p_syndt ) or
    ( erdat = p_syndt and erzet <= p_syntm ) ) ) and
    w~matnr in s_matnr and
    w~pstyv in s_itmcat and
    w~lfrel in s_lfrel and
    w~abgru = ' ' and
    w~kwmeng > 0 and
    w~mtvfp in w_mtvfp and
    x~ettyp in w_ettyp and
    x~bdart in s_req_tp and
    x~plart in s_pln_tp and
    x~etart in s_etart and
    x~abart in s_abart and
    ( ( xlifsp in s_lifsp ) or ( xlifsp = ' ' ) ).
    The problem: It takes too much time while executing this statement.
    Could anybody change this statement and help me out to reduce the DB Access time?
    Thx

    Ways of Performance Tuning
    1.     Selection Criteria
    2.     Select Statements
    •     Select Queries
    •     SQL Interface
    •     Aggregate Functions
    •     For all Entries
    Select Over more than one internal table
    Selection Criteria
    1.     Restrict the data to the selection criteria itself, rather than filtering it out using the ABAP code using CHECK statement. 
    2.     Select with selection list.
    SELECT * FROM SBOOK INTO SBOOK_WA.
      CHECK: SBOOK_WA-CARRID = 'LH' AND
             SBOOK_WA-CONNID = '0400'.
    ENDSELECT.
    The above code can be much more optimized by the code written below which avoids CHECK, selects with selection list
    SELECT  CARRID CONNID FLDATE BOOKID FROM SBOOK INTO TABLE T_SBOOK
      WHERE SBOOK_WA-CARRID = 'LH' AND
                  SBOOK_WA-CONNID = '0400'.
    Select Statements   Select Queries
    1.     Avoid nested selects
    SELECT * FROM EKKO INTO EKKO_WA.
      SELECT * FROM EKAN INTO EKAN_WA
          WHERE EBELN = EKKO_WA-EBELN.
      ENDSELECT.
    ENDSELECT.
    The above code can be much more optimized by the code written below.
    SELECT PF1 PF2 FF3 FF4 INTO TABLE ITAB
        FROM EKKO AS P INNER JOIN EKAN AS F
          ON PEBELN = FEBELN.
    Note: A simple SELECT loop is a single database access whose result is passed to the ABAP program line by line. Nested SELECT loops mean that the number of accesses in the inner loop is multiplied by the number of accesses in the outer loop. One should therefore use nested SELECT loops only if the selection in the outer loop contains very few lines or the outer loop is a SELECT SINGLE statement.
    2.     Select all the records in a single shot using into table clause of select statement rather than to use Append statements.
    SELECT * FROM SBOOK INTO SBOOK_WA.
      CHECK: SBOOK_WA-CARRID = 'LH' AND
             SBOOK_WA-CONNID = '0400'.
    ENDSELECT.
    The above code can be much more optimized by the code written below which avoids CHECK, selects with selection list and puts the data in one shot using into table
    SELECT  CARRID CONNID FLDATE BOOKID FROM SBOOK INTO TABLE T_SBOOK
      WHERE SBOOK_WA-CARRID = 'LH' AND
                  SBOOK_WA-CONNID = '0400'.
    3.     When a base table has multiple indices, the where clause should be in the order of the index, either a primary or a secondary index.
    To choose an index, the optimizer checks the field names specified in the where clause and then uses an index that has the same order of the fields. In certain scenarios, it is advisable to check whether a new index can speed up the performance of a program. This will come handy in programs that access data from the finance tables.
    4.     For testing existence, use Select.. Up to 1 rows statement instead of a Select-Endselect-loop with an Exit. 
    SELECT * FROM SBOOK INTO SBOOK_WA
      UP TO 1 ROWS
      WHERE CARRID = 'LH'.
    ENDSELECT.
    The above code is more optimized as compared to the code mentioned below for testing existence of a record.
    SELECT * FROM SBOOK INTO SBOOK_WA
        WHERE CARRID = 'LH'.
      EXIT.
    ENDSELECT.
    5.     Use Select Single if all primary key fields are supplied in the Where condition .
    If all primary key fields are supplied in the Where conditions you can even use Select Single.
    Select Single requires one communication with the database system, whereas Select-Endselect needs two.
    Select Statements SQL Interface
    1.     Use column updates instead of single-row updates
    to update your database tables.
    SELECT * FROM SFLIGHT INTO SFLIGHT_WA.
      SFLIGHT_WA-SEATSOCC =
        SFLIGHT_WA-SEATSOCC - 1.
      UPDATE SFLIGHT FROM SFLIGHT_WA.
    ENDSELECT.
    The above mentioned code can be more optimized by using the following code
    UPDATE SFLIGHT
           SET SEATSOCC = SEATSOCC - 1.
    2.     For all frequently used Select statements, try to use an index.
    SELECT * FROM SBOOK CLIENT SPECIFIED INTO SBOOK_WA
      WHERE CARRID = 'LH'
        AND CONNID = '0400'.
    ENDSELECT.
    The above mentioned code can be more optimized by using the following code
    SELECT * FROM SBOOK CLIENT SPECIFIED INTO SBOOK_WA
      WHERE MANDT IN ( SELECT MANDT FROM T000 )
        AND CARRID = 'LH'
        AND CONNID = '0400'.
    ENDSELECT.
    3.     Using buffered tables improves the performance considerably.
    Bypassing the buffer increases the network considerably
    SELECT SINGLE * FROM T100 INTO T100_WA
      BYPASSING BUFFER
      WHERE     SPRSL = 'D'
            AND ARBGB = '00'
            AND MSGNR = '999'.
    The above mentioned code can be more optimized by using the following code
    SELECT SINGLE * FROM T100  INTO T100_WA
      WHERE     SPRSL = 'D'
            AND ARBGB = '00'
            AND MSGNR = '999'.
    Select Statements  Aggregate Functions
    •     If you want to find the maximum, minimum, sum and average value or the count of a database column, use a select list with aggregate functions instead of computing the aggregates yourself.
    Some of the Aggregate functions allowed in SAP are  MAX, MIN, AVG, SUM, COUNT, COUNT( * )
    Consider the following extract.
                Maxno = 0.
                Select * from zflight where airln = ‘LF’ and cntry = ‘IN’.
                 Check zflight-fligh > maxno.
                 Maxno = zflight-fligh.
                Endselect.
    The  above mentioned code can be much more optimized by using the following code.
    Select max( fligh ) from zflight into maxno where airln = ‘LF’ and cntry = ‘IN’.
    Select Statements  For All Entries
    •     The for all entries creates a where clause, where all the entries in the driver table are combined with OR. If the number of entries in the driver table is larger than rsdb/max_blocking_factor, several similar SQL statements are executed to limit the length of the WHERE clause.
         The plus
    •     Large amount of data
    •     Mixing processing and reading of data
    •     Fast internal reprocessing of data
    •     Fast
         The Minus
    •     Difficult to program/understand
    •     Memory could be critical (use FREE or PACKAGE size)
    Points to be must considered FOR ALL ENTRIES
    •     Check that data is present in the driver table
    •     Sorting the driver table
    •     Removing duplicates from the driver table
    Consider the following piece of extract
              Loop at int_cntry.
      Select single * from zfligh into int_fligh
      where cntry = int_cntry-cntry.
      Append int_fligh.
                          Endloop.
    The above mentioned can be more optimized by using the following code.
    Sort int_cntry by cntry.
    Delete adjacent duplicates from int_cntry.
    If NOT int_cntry[] is INITIAL.
                Select * from zfligh appending table int_fligh
                For all entries in int_cntry
                Where cntry = int_cntry-cntry.
    Endif.
    Select Statements Select Over more than one Internal table
    1.     Its better to use a views instead of nested Select statements.
    SELECT * FROM DD01L INTO DD01L_WA
      WHERE DOMNAME LIKE 'CHAR%'
            AND AS4LOCAL = 'A'.
      SELECT SINGLE * FROM DD01T INTO DD01T_WA
        WHERE   DOMNAME    = DD01L_WA-DOMNAME
            AND AS4LOCAL   = 'A'
            AND AS4VERS    = DD01L_WA-AS4VERS
            AND DDLANGUAGE = SY-LANGU.
    ENDSELECT.
    The above code can be more optimized by extracting all the data from view DD01V_WA
    SELECT * FROM DD01V INTO  DD01V_WA
      WHERE DOMNAME LIKE 'CHAR%'
            AND DDLANGUAGE = SY-LANGU.
    ENDSELECT
    2.     To read data from several logically connected tables use a join instead of nested Select statements. Joins are preferred only if all the primary key are available in WHERE clause for the tables that are joined. If the primary keys are not provided in join the Joining of tables itself takes time.
    SELECT * FROM EKKO INTO EKKO_WA.
      SELECT * FROM EKAN INTO EKAN_WA
          WHERE EBELN = EKKO_WA-EBELN.
      ENDSELECT.
    ENDSELECT.
    The above code can be much more optimized by the code written below.
    SELECT PF1 PF2 FF3 FF4 INTO TABLE ITAB
        FROM EKKO AS P INNER JOIN EKAN AS F
          ON PEBELN = FEBELN.
    3.     Instead of using nested Select loops it is often better to use subqueries.
    SELECT * FROM SPFLI
      INTO TABLE T_SPFLI
      WHERE CITYFROM = 'FRANKFURT'
        AND CITYTO = 'NEW YORK'.
    SELECT * FROM SFLIGHT AS F
        INTO SFLIGHT_WA
        FOR ALL ENTRIES IN T_SPFLI
        WHERE SEATSOCC < F~SEATSMAX
          AND CARRID = T_SPFLI-CARRID
          AND CONNID = T_SPFLI-CONNID
          AND FLDATE BETWEEN '19990101' AND '19990331'.
    ENDSELECT.
    The above mentioned code can be even more optimized by using subqueries instead of for all entries.
    SELECT * FROM SFLIGHT AS F INTO SFLIGHT_WA
        WHERE SEATSOCC < F~SEATSMAX
          AND EXISTS ( SELECT * FROM SPFLI
                         WHERE CARRID = F~CARRID
                           AND CONNID = F~CONNID
                           AND CITYFROM = 'FRANKFURT'
                           AND CITYTO = 'NEW YORK' )
          AND FLDATE BETWEEN '19990101' AND '19990331'.
    ENDSELECT.
    1.     Table operations should be done using explicit work areas rather than via header lines.
    READ TABLE ITAB INTO WA WITH KEY K = 'X‘ BINARY SEARCH.
    IS MUCH FASTER THAN USING
    READ TABLE ITAB INTO WA WITH KEY K = 'X'.
    If TAB has n entries, linear search runs in O( n ) time, whereas binary search takes only O( log2( n ) ).
    2.     Always try to use binary search instead of linear search. But don’t forget to sort your internal table before that.
    READ TABLE ITAB INTO WA WITH KEY K = 'X'. IS FASTER THAN USING
    READ TABLE ITAB INTO WA WITH KEY (NAME) = 'X'.
    3.     A dynamic key access is slower than a static one, since the key specification must be evaluated at runtime.
    4.     A binary search using secondary index takes considerably less time.
    5.     LOOP ... WHERE is faster than LOOP/CHECK because LOOP ... WHERE evaluates the specified condition internally.
    LOOP AT ITAB INTO WA WHERE K = 'X'.
    ENDLOOP.
    The above code is much faster than using
    LOOP AT ITAB INTO WA.
      CHECK WA-K = 'X'.
    ENDLOOP.
    6.     Modifying selected components using “ MODIFY itab …TRANSPORTING f1 f2.. “ accelerates the task of updating  a line of an internal table.
    WA-DATE = SY-DATUM.
    MODIFY ITAB FROM WA INDEX 1 TRANSPORTING DATE.
    The above code is more optimized as compared to
    WA-DATE = SY-DATUM.
    MODIFY ITAB FROM WA INDEX 1.
    7.     Accessing the table entries directly in a "LOOP ... ASSIGNING ..." accelerates the task of updating a set of lines of an internal table considerably
    Modifying selected components only makes the program faster as compared to Modifying all lines completely.
    e.g,
    LOOP AT ITAB ASSIGNING <WA>.
      I = SY-TABIX MOD 2.
      IF I = 0.
        <WA>-FLAG = 'X'.
      ENDIF.
    ENDLOOP.
    The above code works faster as compared to
    LOOP AT ITAB INTO WA.
      I = SY-TABIX MOD 2.
      IF I = 0.
        WA-FLAG = 'X'.
        MODIFY ITAB FROM WA.
      ENDIF.
    ENDLOOP.
    8.    If collect semantics is required, it is always better to use to COLLECT rather than READ BINARY and then ADD.
    LOOP AT ITAB1 INTO WA1.
      READ TABLE ITAB2 INTO WA2 WITH KEY K = WA1-K BINARY SEARCH.
      IF SY-SUBRC = 0.
        ADD: WA1-VAL1 TO WA2-VAL1,
             WA1-VAL2 TO WA2-VAL2.
        MODIFY ITAB2 FROM WA2 INDEX SY-TABIX TRANSPORTING VAL1 VAL2.
      ELSE.
        INSERT WA1 INTO ITAB2 INDEX SY-TABIX.
      ENDIF.
    ENDLOOP.
    The above code uses BINARY SEARCH for collect semantics. READ BINARY runs in O( log2(n) ) time. The above piece of code can be more optimized by
    LOOP AT ITAB1 INTO WA.
      COLLECT WA INTO ITAB2.
    ENDLOOP.
    SORT ITAB2 BY K.
    COLLECT, however, uses a hash algorithm and is therefore independent
    of the number of entries (i.e. O(1)) .
    9.    "APPEND LINES OF itab1 TO itab2" accelerates the task of appending a table to another table considerably as compared to “ LOOP-APPEND-ENDLOOP.”
    APPEND LINES OF ITAB1 TO ITAB2.
    This is more optimized as compared to
    LOOP AT ITAB1 INTO WA.
      APPEND WA TO ITAB2.
    ENDLOOP.
    10.   “DELETE ADJACENT DUPLICATES“ accelerates the task of deleting duplicate entries considerably as compared to “ READ-LOOP-DELETE-ENDLOOP”.
    DELETE ADJACENT DUPLICATES FROM ITAB COMPARING K.
    This is much more optimized as compared to
    READ TABLE ITAB INDEX 1 INTO PREV_LINE.
    LOOP AT ITAB FROM 2 INTO WA.
      IF WA = PREV_LINE.
        DELETE ITAB.
      ELSE.
        PREV_LINE = WA.
      ENDIF.
    ENDLOOP.
    11.   "DELETE itab FROM ... TO ..." accelerates the task of deleting a sequence of lines considerably as compared to “  DO -DELETE-ENDDO”.
    DELETE ITAB FROM 450 TO 550.
    This is much more optimized as compared to
    DO 101 TIMES.
      DELETE ITAB INDEX 450.
    ENDDO.
    12.   Copying internal tables by using “ITAB2[ ] = ITAB1[ ]” as compared to “LOOP-APPEND-ENDLOOP”.
    ITAB2[] = ITAB1[].
    This is much more optimized as compared to
    REFRESH ITAB2.
    LOOP AT ITAB1 INTO WA.
      APPEND WA TO ITAB2.
    ENDLOOP.
    13.   Specify the sort key as restrictively as possible to run the program faster.
    “SORT ITAB BY K.” makes the program runs faster as compared to “SORT ITAB.”
    Internal Tables         contd…
    Hashed and Sorted tables
    1.     For single read access hashed tables are more optimized as compared to sorted tables.
    2.      For partial sequential access sorted tables are more optimized as compared to hashed tables
    Hashed And Sorted Tables
    Point # 1
    Consider the following example where HTAB is a hashed table and STAB is a sorted table
    DO 250 TIMES.
      N = 4 * SY-INDEX.
      READ TABLE HTAB INTO WA WITH TABLE KEY K = N.
      IF SY-SUBRC = 0.
      ENDIF.
    ENDDO.
    This runs faster for single read access as compared to the following same code for sorted table
    DO 250 TIMES.
      N = 4 * SY-INDEX.
      READ TABLE STAB INTO WA WITH TABLE KEY K = N.
      IF SY-SUBRC = 0.
      ENDIF.
    ENDDO.
    Point # 2
    Similarly for Partial Sequential access the STAB runs faster as compared to HTAB
    LOOP AT STAB INTO WA WHERE K = SUBKEY.
    ENDLOOP.
    This runs faster as compared to
    LOOP AT HTAB INTO WA WHERE K = SUBKEY.
    ENDLOOP.

  • PDPageDrawContentsToWindowEx takes too much time

    We are using an Acrobat plugin that renders the PDF file to a bitmap in memory.
    We are using Acrobat Professional X. But same problems also appear on Acrobat 9.
    We have received several problematic PDF files from our customers that is causing the call that renders the image -
    PDPageDrawContentsToWindowEx() to take unreasonably long time.
    My target resolutions is 600 dpi but I couldn't wait for the call to return, after more than 6 minutes I kill the process.
    The same PDFs render in Acrobat with slight delay (flickering and repainting) but in reasonable time.
    I have written for this problem on previous occasions (Aug 5 2010). Since then further problematic samples
    show that is linked somehow with transparency being present, but not on all PDFs with transparency.
    We have a fast computer so the problem is somewhere in the PDF analysis.
    Trying to optimize the file didn't help.
    Checking with Preflight for PDF syntax issues also didn't find anything.

    I'll have to check the headers, but I KNOW that we exposed it to plugins in A9 – it was necessary since we pulled the DrawToWindow call on the Mac (carbon vs. cocoa).
    What you are asking the SDK to do is going to be painful on large complex images.  Drawing into an HDC/Window adds SIGNIFICANT overhead to our rendering process, since we have to do all the work in our own "bit buffer" and then copy that buffer into the OS's provided HDC.  OUCH!  This is why  DrawToMemory is better.
    Additionally, if you have files with complex transparency AND you want OverprintPreview, that's also going to be VERY complex rendering pipeline – made WORSE by the need to end up in an HDC.   Forgetting separations for the moment, consider that we have to convert everything to CMYK (since you can only compute OP in CMYK), blend colors, then convert all of that to RGB.   And that's assuming SIMPLE transparency.  If you have multiple blending groups, soft masks, etc. then you just raised the bar even more!
    How long does it take Acrobat to open up the PDF and render it completely to screen?   For separations, how long does it take to do a "Flatten Transparency" operation?
    UpdateRect won't help because of the OP and then transparency flattening
    From: Adobe Forums <[email protected]<mailto:[email protected]>>
    Reply-To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>>
    Date: Mon, 5 Dec 2011 06:45:46 -0800
    To: Leonard Rosenthol <[email protected]<mailto:[email protected]>>
    Subject: PDPageDrawContentsToWindowEx takes too much time
    Re: PDPageDrawContentsToWindowEx takes too much time
    created by Nikolay Tasev<http://forums.adobe.com/people/Nikolay+Tasev> in Acrobat SDK - View the full discussion<http://forums.adobe.com/message/4064198#4064198

  • Finishing Backup takes too much time.

    Finishing backup takes too much time in my system , sometimes 30 minutes even 1 hour! and sometimes right after finishing backup(before the icon in the menu bar stops) it starts to take another backup( and usually 10MB or something small !) or the estimation is wrong .
    Sometimes I see my system that is taking backups hours and hours(just with small stops!)
    and the speed rate I of data transfer is reduced, while I set it to be on maximum speed in Time Capsule.

    And the log with Time-stamps from Console
    4/6/09 2:31:24 PM /System/Library/CoreServices/backupd[2141] Starting standard backup
    4/6/09 2:31:34 PM /System/Library/CoreServices/backupd[2141] Mounted network destination using URL: afp://[email protected]/Sina's%20Time%20Capsule
    4/6/09 2:31:34 PM /System/Library/CoreServices/backupd[2141] Backup destination mounted at path: /Volumes/Sina's Time Capsule
    4/6/09 2:31:39 PM /System/Library/CoreServices/backupd[2141] Disk image /Volumes/Sina's Time Capsule/Sina’s MacBook_0017f2347181.sparsebundle mounted at: /Volumes/Backup of Sina’s MacBook
    4/6/09 2:31:39 PM /System/Library/CoreServices/backupd[2141] Backing up to: /Volumes/Backup of Sina’s MacBook/Backups.backupdb
    4/6/09 2:39:45 PM /System/Library/CoreServices/backupd[2141] No pre-backup thinning needed: 862.6 MB requested (including padding), 205.56 GB available
    4/6/09 2:46:52 PM /System/Library/CoreServices/backupd[2141] Bulk setting Spotlight attributes failed.
    4/6/09 2:48:07 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 2:48:07 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 2:48:07 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 2:50:06 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 2:50:06 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 2:57:21 PM /System/Library/CoreServices/backupd[2141] Bulk setting Spotlight attributes failed.
    4/6/09 3:03:57 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 3:03:58 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 3:03:58 PM /System/Library/CoreServices/backupd[2141] Unable to rebuild path cache for source item. Partial source path:
    4/6/09 3:12:55 PM /System/Library/CoreServices/backupd[2141] Copied 29426 files (162.6 MB) from volume Macintosh Disk.
    4/6/09 3:17:36 PM /System/Library/CoreServices/backupd[2141] No pre-backup thinning needed: 678.9 MB requested (including padding), 205.56 GB available
    4/6/09 3:37:18 PM /System/Library/CoreServices/backupd[2141] Bulk setting Spotlight attributes failed.
    4/6/09 3:37:57 PM /System/Library/CoreServices/backupd[2141] Copied 2362 files (100.3 MB) from volume Macintosh Disk.
    4/6/09 3:42:00 PM /System/Library/CoreServices/backupd[2141] Starting post-backup thinning

Maybe you are looking for

  • Mail drops accounts on quit

    Hi I am having a real problem with Mail. Everytime I open the app, I get the "welcome to mail" setup dialogue. I proceed to set up my gmail account and it all works fine until I quit. On reopening the app I again get the "welcome to mail" setup scree

  • IMessage works in one network and not another...

    Since I installed it I couldn't get iMessage Beta to work on my Mac at home. When I used it at work today, it worked perfectly.  But now I'm back home and can't get it to work.  Factime isn't working either.  What's going on?

  • Tax value difference in FI document & report S_ALR_87012357

    Hi Team There is tax value difference in FI document and report S_ALR_87012357.why this difference come? In FI Document GST input value S$414.89 In Report GST input value : S$3,599.34 Why these difference coming ? Please suggest.. Thank you Regard, D

  • Hyper-V is missing in features on off windows 10 preview

    Hello, I've installed windows 10 preview. My computer enables virtualization and slat and everything it needs but why  there is no Hyper-V enable in feature on-off?  I want to check the windows phone 8.1 emulator on visual studio but it says there is

  • Windows 8.1 Run as different user from search function

    Hi There, After about a year of Windows 8 I was actually quite enjoying the ability to type a program name in start menu and then use right click to run as a different user. which is achieved via changing local security policy (i.e. running rsat tool