Long time for loading Planning books/data views

Hi
Could someone throw some pointers/solutions for extremely horrible time taken for a planning view (data view) to load and populate for certains selection profiles that already have large number of trasactional records. In my case it is specifically the run time dumps thrown for a popular combination (large number of transactions).
Urgent suggestions/solutions required. Pls. call 9923170825. India, if you are lazy enough to type it out. or just type in some key words, tcodes etc.
Thanks

I hope you dont have too many macros hogging the memory in interactive book. The other thing is to be on the latest LC build. Also try to have default unit of measures (keep the dataview settings for UoM blank).
Can you confirm if you build the TLB view brand new in 5.0 or did you do some migration of the books ? The first time opening of book takes longer due to some compilation of certain objects. in SM50, try to identify where the process takes longest - whether it is at client end, at the DB procedure of LC or at application level.

Similar Messages

  • Take long time for loading

    Hi Experts,
    One of my data target (Data comming from another ODS) is taking long time for loading, basically it takes below 10 times only, but today it is running from last 40 minesu2026
    In Status Tab it showing....
    Job termination in source system
    Diagnosis
    The background job for data selection in the source system has been terminated. It is very likely that a short dump has been logged in the source system
    Procedure
    Read the job log in the source system. Additional information is displayed here.
    To access the job log, use the monitor wizard (step-by-step analysis)  or the menu path <LS>Environment -> Job Overview -> In Source System
    Error correction:
    Follow the instructions in the job log messages.
    Can anyone please solve my problem.
    Thanks in advance
    David

    Hi Experts,
    Thanks for your answers, and my load failed when the data comes from one ODS to another ODS. find the below job log
    Job started
    Step 001 started (program SBIE0001, variant &0000000007169, user ID RCREMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 8ZPP_OP3
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 20006838008                             *
    abap/heap_area_total.......... 20006838008                             *
    abap/heaplimit................ 83886080                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 5000000                                 *
    ztta/roll_extension........... 4294967295                              *
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 1 in task 0002 (1 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:31:55, End = 06.09.2010 01:31:55
    Asynchronous transmission of info IDoc 3 in task 0003 (1 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 2 in task 0004 (2 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:00, End = 06.09.2010 01:32:00
    Asynchronous transmission of info IDoc 4 in task 0005 (2 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 3 in task 0006 (3 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:04, End = 06.09.2010 01:32:04
    Asynchronous transmission of info IDoc 5 in task 0007 (3 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 4 in task 0008 (4 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:08, End = 06.09.2010 01:32:08
    Asynchronous transmission of info IDoc 6 in task 0009 (4 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 5 in task 0010 (5 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:11, End = 06.09.2010 01:32:11
    Asynchronous transmission of info IDoc 7 in task 0011 (5 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Asynchronous send of data package 13 in task 0026 (6 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:01, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:44, End = 06.09.2010 01:32:45
    tRFC: Data Package = 8, TID = 0AEB465C00AE4C847CEA0070, Duration = 00:00:17,
    tRFC: Start = 06.09.2010 01:32:29, End = 06.09.2010 01:32:46
    Asynchronous transmission of info IDoc 15 in task 0027 (5 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 14 in task 0028 (6 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:48, End = 06.09.2010 01:32:48
    tRFC: Data Package = 9, TID = 0AEB465C00AE4C847CEF0071, Duration = 00:00:18,
    tRFC: Start = 06.09.2010 01:32:33, End = 06.09.2010 01:32:51
    Asynchronous transmission of info IDoc 16 in task 0029 (5 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 15 in task 0030 (6 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:52, End = 06.09.2010 01:32:52
    Asynchronous transmission of info IDoc 17 in task 0031 (6 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 16 in task 0032 (7 parallel tasks)
    tRFC: Data Package = 10, TID = 0AEB465C00684C847CF30070, Duration = 00:00:18,
    tRFC: Start = 06.09.2010 01:32:37, End = 06.09.2010 01:32:55
    tRFC: Data Package = 11, TID = 0AEB465C02E14C847CF70083, Duration = 00:00:17,
    tRFC: Start = 06.09.2010 01:32:42, End = 06.09.2010 01:32:59
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:32:56, End = 06.09.2010 01:32:56
    Asynchronous transmission of info IDoc 18 in task 0033 (5 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 17 in task 0034 (6 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:33:00, End = 06.09.2010 01:33:00
    tRFC: Data Package = 12, TID = 0AEB465C00AE4C847CFB0072, Duration = 00:00:16,
    tRFC: Start = 06.09.2010 01:32:46, End = 06.09.2010 01:33:02
    Asynchronous transmission of info IDoc 19 in task 0035 (5 parallel tasks)
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 100,000 records
    Result of customer enhancement: 100,000 records
    Asynchronous send of data package 18 in task 0036 (6 parallel tasks)
    tRFC: Data Package = 0, TID = , Duration = 00:00:00, ARFCSTATE =
    tRFC: Start = 06.09.2010 01:33:04, End = 06.09.2010 01:33:04
    Asynchronous transmission of info IDoc 20 in task 0037 (6 parallel tasks)
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Thanks
    Daivd
    Edited by: david Rathod on Sep 6, 2010 12:04 PM

  • SNP - Planning Book/Data View not showing up in /SAPAPO/SNP94

    Hi All:
    I created a planning Book out of a custom SNP planning area. While I could see the Planning Book in the list of Planning Book in the /SAPAPO/SDP94 - Interactive Supply Network Planning (all Books)   I could not see the custom planning book in /SAPAPO/SDP94 - Interactive Supply Network Planning (all Books) .  I could ony see the SAP supplied PB there.  Will SDP94 transaction display only the standard SNP Planning book?  How can I make sure that the custom PB is displaed in /SAPAPO/SDP94?
    pl. help
    Thanks
    Ryan

    Hi Narayanan,
    a couple of simple checks..
    can you please check that you have not set a filter critera in the shuffler in planning book/data view?
    you can also check if by mistake only a limited number of users are assigned to this planning book and data view ?
    else you may want to check note : , 327371
    Let me if this helps.
    Rgds, Sandeep

  • Rename & deletion of planning book & Data view

    Hi Experts,
    Is it possible to rename & deletion of planning book & Data view, if yes what is the procedure.
    Also suggest if hiding of same is possible

    Hello  jrd333
    You can delete planning book/data view via transaction /SAPAPO/SDP8B
    There you can choose PB/DV that needs corrections.
    Also you maybe need to deactivate your planning area to wich this PB is assigned.
    You can do this in transaction /SAPAPO/MSDP_ADMIN
    This things is related to custom Planning Books/Views.

  • Impdp taking long time for only few MBs data...

    Hi All,
    I have one query related to impdp. I have one expdp file and size is 47M. When I restore this dmp using impdp it will take more time. Also initially table_data loaded finsih very fast but then later on alter function/procedure/view taking a lot time almost 4 to 5 hrs.
    I have no idea why its taking long time... Earlier I can see one DB link is failed and it given error for "TNS name could not resolved" so I create DB link also before run impdp but the same result. Can any one suggest what could be the cause to taking long time for only 47MB data?
    Note - Both expdp and impdp database version is 11.2.0.3.0. If I import the same dmp file into 11.2.0.1.0 then its done in few mins.
    Thanks...

    Also Read
    Checklist For Slow Performance Of DataPump Export (expdp) And Import (impdp) [ID 453895.1]
    DataPump Import (IMPDP) is Very Slow at Object/System/Role Grants, Default Roles [ID 1267951.1]

  • Planning book: Data views

    Hello:
    I am defining a planning book and I want to have the same data views (total demands and total receipts) of standard pl.book 9ASNP94, data view SNP94(1), can anyone help me design this.
    My planning area is a copy of 9ASNP02.
    Thanks in advance.
    Mono

    hi Harish:
    I have create a planning area which is identical to 9ASNP02 for my pl.book except the unit of measure (EA for ST), and when I try as you said I get a message stating that <b>The planning area of planning books 9ASNP94 and Z_mono (my pl.book) is not identical</b>.
    Any other way?
    Thank you
    Mono

  • Input help takes longer times for the planning var in Portal

    Hi Gurus,
    I had a planning report with planning function copy historic demand and some planning filters forecast analysis ID.When i'm executing the report in portal by providing the direct values to   forecast analysis ID,but when i press F4 button the screen keep on refreshing.
    Regards,
    rambabu

    Dear Rambabu
    Have you maintained the selection criteria for the infoobject as Posted values or from master data, it will depend on this how fast the selection retrieves
    Regards
    Gajesh

  • Takes long time for my mac book pro t connect to the airport extreme

    The first time I go to email or safari it takes a few minutes to get connected

    Are you recovering Lion or later Time Machine backup with a Lion or later machine?
    If you are, then use this article:
    http://support.apple.com/kb/HT4718
    Other Time Machine restore options are here:
    http://support.apple.com/kb/ht1427
    Note, if recovering from a PowerPC Mac (iMac G5, PowerMac, Mac Mini G4, iBook, Powerbook), this tip is important:
    http://discussions.apple.com/docs/DOC-2295
    Otherwise use the recommended restore option with whatever backup software you used initially.

  • Best practices for loading apo planning book data to cube for reporting

    Hi,
    I would like to know whether there are any Best practices for loading apo planning book data to cube for reporting.
    I have seen 2 types of Design:
    1) The Planning Book Extractor data is Loaded first to the APO BW system within a Cube, and then get transferred to the Actual BW system. Reports are run from the Actual BW system cube.
    2) The Planning Book Extractor data is loaded directly to a cube within the Actual BW system.
    We do these data loads during evening hours once in a day.
    Rgds
    Gk

    Hi GK,
    What I have normally seen is:
    1) Data would be extracted from APO Planning Area to APO Cube (FOR BACKUP purpose). Weekly or monthly, depending on how much data change you expect, or how critical it is for business. Backups are mostly monthly for DP.
    2) Data extracted from APO planning area directly to DSO of staging layer in BW, and then to BW cubes, for reporting.
    For DP monthly, SNP daily
    You can also use the option 1 that you mentioned below. In this case, the APO cube is the backup cube, while the BW cube is the one that you could use for reporting, and this BW cube gets data from APO cube.
    Benefit in this case is that we have to extract data from Planning Area only once. So, planning area is available for jobs/users for more time. However, backup and reporting extraction are getting mixed in this case, so issues in the flow could impact both the backup and the reporting. We have used this scenario recently, and yet to see the full impact.
    Thanks - Pawan

  • Crystal reports for Eclipse (CR4E)  - taking long time to load

    HI
    We have used your Crystal reports for Eclipse (CR4E) but it is taking long time to load if the record count is more than 3 million
    Our architecture is we are using Oracle DB and web sphere application server both are running on SUN server,
    If we buy original version it will work fast?
    Or do you  have alternative solution for this, this one to implement one of our major client

    You have 3 million records, this is going to take time to process.  Also, you don't define what is meant by a long time.  Are we talking a couple of minutes, or something more like 20 minutes? 
    The real question is, do you really need the report to bring back 3 million records?  Could you do any filtering of the data to reduce the number of records being brought back?  Also, 3 million records could vary.  is each record a single field, or is it multiple.   The more fields, the larger the result set.
    Another thing you want to look at, is just how long is the query taking to run against the database.

  • Query taking long time for EXTRACTING the data more than 24 hours

    Hi ,
    Query taking long time for EXTRACTING the data more than 24 hours please find the query and explain plan details below even indexes avilable on table's goe's to FULL TABLE SCAN. please suggest me.......
    SQL> explain plan for select a.account_id,round(a.account_balance,2) account_balance,
    2 nvl(ah.invoice_id,ah.adjustment_id) transaction_id,
    to_char(ah.effective_start_date,'DD-MON-YYYY') transaction_date,
    to_char(nvl(i.payment_due_date,
    to_date('30-12-9999','dd-mm-yyyy')),'DD-MON-YYYY')
    due_date, ah.current_balance-ah.previous_balance amount,
    decode(ah.invoice_id,null,'A','I') transaction_type
    3 4 5 6 7 8 from account a,account_history ah,invoice i_+
    where a.account_id=ah.account_id
    and a.account_type_id=1000002
    and round(a.account_balance,2) > 0
    and (ah.invoice_id is not null or ah.adjustment_id is not null)
    and ah.CURRENT_BALANCE > ah.previous_balance
    and ah.invoice_id=i.invoice_id(+)
    AND a.account_balance > 0
    order by a.account_id,ah.effective_start_date desc; 9 10 11 12 13 14 15 16
    Explained.
    SQL> select * from table(dbms_xplan.display);
    PLAN_TABLE_OUTPUT
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)|
    | 0 | SELECT STATEMENT | | 544K| 30M| | 693K (20)|
    | 1 | SORT ORDER BY | | 544K| 30M| 75M| 693K (20)|
    |* 2 | HASH JOIN | | 544K| 30M| | 689K (20)|
    |* 3 | TABLE ACCESS FULL | ACCOUNT | 20080 | 294K| | 6220 (18)|
    |* 4 | HASH JOIN OUTER | | 131M| 5532M| 5155M| 678K (20)|
    |* 5 | TABLE ACCESS FULL| ACCOUNT_HISTORY | 131M| 3646M| | 197K (25)|
    | 6 | TABLE ACCESS FULL| INVOICE | 262M| 3758M| | 306K (18)|
    Predicate Information (identified by operation id):
    2 - access("A"."ACCOUNT_ID"="AH"."ACCOUNT_ID")
    3 - filter("A"."ACCOUNT_TYPE_ID"=1000002 AND "A"."ACCOUNT_BALANCE">0 AND
    ROUND("A"."ACCOUNT_BALANCE",2)>0)
    4 - access("AH"."INVOICE_ID"="I"."INVOICE_ID"(+))
    5 - filter("AH"."CURRENT_BALANCE">"AH"."PREVIOUS_BALANCE" AND ("AH"."INVOICE_ID"
    IS NOT NULL OR "AH"."ADJUSTMENT_ID" IS NOT NULL))
    22 rows selected.
    Index Details:+_
    SQL> select INDEX_OWNER,INDEX_NAME,COLUMN_NAME,TABLE_NAME from dba_ind_columns where
    2 table_name in ('INVOICE','ACCOUNT','ACCOUNT_HISTORY') order by 4;
    INDEX_OWNER INDEX_NAME COLUMN_NAME TABLE_NAME
    OPS$SVM_SRV4 P_ACCOUNT ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT CUSTOMER_NODE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_ACCOUNT_TYPE ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_PREVIOUS_INVOICE PREVIOUS_INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_LAST_MODIFIED_ACCOUNT LAST_MODIFIED ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE_ACCOUNT INVOICE_ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT SEQNR ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_INVOICE INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA CURRENT_BALANCE ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_LMOD LAST_MODIFIED ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_PAYMENT PAYMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADJUSTMENT ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_APPLIED_DT APPLIED_DATE ACCOUNT_HISTORY
    OPS$SVM_SRV4 P_INVOICE INVOICE_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE CUSTOMER_INVOICE_STR INVOICE
    OPS$SVM_SRV4 I_LAST_MODIFIED_INVOICE LAST_MODIFIED INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT ACCOUNT_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_BILL_RUN BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_INVOICE_TYPE INVOICE_TYPE_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_CUSTOMER_NODE CUSTOMER_NODE_ID INVOICE
    32 rows selected.
    Regards,
    Bathula
    Oracle-DBA

    I have some suggestions. But first, you realize that you have some redundant indexes, right? You have an index on account(account_name) and also account(account_name, account_id), and also account_history(invoice_id) and account_history(invoice_id, adjustment_id). No matter, I will suggest some new composite indexes.
    Also, you do not need two lines for these conditions:
    and round(a.account_balance, 2) > 0
    AND a.account_balance > 0
    You can just use: and a.account_balance >= 0.005
    So the formatted query isselect a.account_id,
           round(a.account_balance, 2) account_balance,
           nvl(ah.invoice_id, ah.adjustment_id) transaction_id,
           to_char(ah.effective_start_date, 'DD-MON-YYYY') transaction_date,
           to_char(nvl(i.payment_due_date, to_date('30-12-9999', 'dd-mm-yyyy')),
                   'DD-MON-YYYY') due_date,
           ah.current_balance - ah.previous_balance amount,
           decode(ah.invoice_id, null, 'A', 'I') transaction_type
      from account a, account_history ah, invoice i
    where a.account_id = ah.account_id
       and a.account_type_id = 1000002
       and (ah.invoice_id is not null or ah.adjustment_id is not null)
       and ah.CURRENT_BALANCE > ah.previous_balance
       and ah.invoice_id = i.invoice_id(+)
       AND a.account_balance >= .005
    order by a.account_id, ah.effective_start_date desc;You will probably want to select:
    1. From ACCOUNT first (your smaller table), for which you supply a literal on account_type_id. That should limit the accounts retrieved from ACCOUNT_HISTORY
    2. From ACCOUNT_HISTORY. We want to limit the records as much as possible on this table because of the outer join.
    3. INVOICE we want to access last because it seems to be least restricted, it is the biggest, and it has the outer join condition so it will manufacture rows to match as many rows as come back from account_history.
    Try the query above after creating the following composite indexes. The order of the columns is important:create index account_composite_i on account(account_type_id, account_balance, account_id);
    create index acct_history_comp_i on account_history(account_id, invoice_id, adjustment_id, current_balance, previous_balance, effective_start_date);
    create index invoice_composite_i on invoice(invoice_id, payment_due_date);All the columns used in the where clause will be indexed, in a logical order suited to the needs of the query. Plus each selected column is indexed as well so that we should not need to touch the tables at all to satisfy the query.
    Try the query after creating these indexes.
    A final suggestion is to try larger sort and hash area sizes and a manual workarea policy.alter session set workarea_size_policy = manual;
    alter session set sort_area_size = 2147483647;
    alter session set hash_area_size = 2147483647;

  • DTP running long time to load master data due to lock on SID table

    Hi,
    we are facing a problem while DTP loading master data into targets(Info Objects).
    It is taking long time to load data due to below error.
    18.03.2014 01:03:23 All records forwarded                                                                             RSM2          730          S
    18.03.2014 01:13:39 EXTRACTION OF DATAPACKAGE 000009                                                                  RSAR          051          S
    18.03.2014 01:13:40 All records forwarded                                                                             RSM2          730          S
    18.03.2014 01:24:50 EXTRACTION OF DATAPACKAGE 000010                                                                  RSAR          051          S
    18.03.2014 01:24:51 All records forwarded                                                                             RSM2          730          S
    18.03.2014 01:38:34 EXTRACTION OF DATAPACKAGE 000011                                                                  RSAR          051          S
    18.03.2014 01:38:41 All records forwarded                                                                             RSM2          730          S
    18.03.2014 01:52:10 EXTRACTION OF DATAPACKAGE 000012                                                                  RSAR          051          S
    18.03.2014 01:52:10 All records forwarded                                                                             RSM2          730          S
    18.03.2014 02:06:51 EXTRACTION OF DATAPACKAGE 000013                                                                  RSAR          051          S
    18.03.2014 02:06:53 All records forwarded                                                                             RSM2          730          S
    18.03.2014 02:22:32 EXTRACTION OF DATAPACKAGE 000014                                                                  RSAR          051          S
    18.03.2014 02:22:33 All records forwarded                                                                             RSM2          730          S
    18.03.2014 02:38:36 EXTRACTION OF DATAPACKAGE 000015                                                                  RSAR          051          S
    18.03.2014 02:38:37 All records forwarded                                                                             RSM2          730          S
    18.03.2014 02:55:25 EXTRACTION OF DATAPACKAGE 000016                                                                  RSAR          051          S
    18.03.2014 02:55:25 All records forwarded                                                                             RSM2          730          S
    18.03.2014 03:13:56 SQL: 18.03.2014 03:13:56 ALEREMOTE                                                                DBMAN         099          I
    18.03.2014 03:13:56 TRUNCATE TABLE "/BI0/0600000066"                                                                  DBMAN         099          I
    18.03.2014 03:13:56 SQL-END: 18.03.2014 03:13:56 00:00:00                                                             DBMAN         099          I
    18.03.2014 03:13:56 SQL: 18.03.2014 03:13:56 ALEREMOTE                                                                DBMAN         099          I
    18.03.2014 03:13:56  LOCK TABLE "/BI0/0600000066" IN EXCLUSIVE MODE                                                   DBMAN         099          I
    18.03.2014 03:13:56 NOWAI                                                                                             DBMAN         099          I
    18.03.2014 03:13:56 SQL-END: 18.03.2014 03:13:56 00:00:00                                                             DBMAN         099          I
    18.03.2014 03:13:58 EXTRACTION OF DATAPACKAGE 000017                                                                  RSAR          051          S
    18.03.2014 03:13:59 All records forwarded                                                                             RSM2          730          S
    18.03.2014 03:34:24 EXTRACTION OF DATAPACKAGE 000018                                                                  RSAR          051          S
    18.03.2014 03:34:25 All records forwarded                                                                             RSM2          730          S
    18.03.2014 03:56:03 EXTRACTION OF DATAPACKAGE 000019                                                                  RSAR          051          S
    18.03.2014 03:56:03 All records forwarded                                                                             RSM2          730          S
    18.03.2014 04:19:55 EXTRACTION OF DATAPACKAGE 000020                                                                  RSAR          051          S
    18.03.2014 04:19:56 All records forwarded                                                                             RSM2          730          S
    18.03.2014 04:44:07 SQL: 18.03.2014 04:44:07 ALEREMOTE                                                                DBMAN         099          I
    18.03.2014 04:44:07 TRUNCATE TABLE "/BI0/0600000068"                                                                  DBMAN         099          I
    18.03.2014 04:44:09 SQL-END: 18.03.2014 04:44:09 00:00:02                                                             DBMAN         099          I
    18.03.2014 04:44:09 SQL: 18.03.2014 04:44:09 ALEREMOTE                                                                DBMAN         099          I
    18.03.2014 04:44:09  LOCK TABLE "/BI0/0600000068" IN EXCLUSIVE MODE                                                   DBMAN         099          I
    18.03.2014 04:44:09 NOWAI                                                                                             DBMAN         099          I
    18.03.2014 04:44:09 SQL-END: 18.03.2014 04:44:09 00:00:00                                                             DBMAN         099          I
    18.03.2014 04:44:12 EXTRACTION OF DATAPACKAGE 000021                                                                  RSAR          051          S
    18.03.2014 04:44:13 All records forwarded                                                                             RSM2          730          S
    For this long running data load we didnt find any short dumps and no locks available on SM12 also.
    Please help on this.
    Thanks,
    Sankar.

    Can we know what kind of data source is this. attribute or text data source.
    what is data volume? its less or more?
    Actually we load direct access data thru dtp(No psa) into info objects for text data sources where have less data only.
    While running your dtp is any attribute change run for the same info object ot any transactional loads are happening?
    Is this happens on particular day or weekend or anydays?
    Thanks

  • Long time to load data from PSA to DSO -Sequential read RSBKDATA_V

    Hi ,
    It is taking long time to load data from PSA to DSO. It is doing Sequential read on RSBKDATA_V and table contents no data .
    we are at - SAPKW70105. It started since yesterday . There is no changes in system parameters.
    Please advice. 
    Thanks
    Nilesh

    Hi Nilesh,
    I guess the following SAP Note will help you in this situation.
    [1476842 - Performance: read RSBKDATA only when needed|https://websmp107.sap-ag.de/sap/support/notes/1476842]
    Just note that the reference to Support Packages is wrong. It is also included in SAP_BW 701 06.

  • Data form taking long time to load

    Hi All,
    My data form is taking a very long time to load. It has period dimension which is having 53 weeks. If I collase Period dimension then the dataform opens up quickly otherwise its taking a very long time (around 7 mins) to open up.
    What can be done to improve the speed?
    Thanks

    Abhishek wrote:
    Hope this helps
    Improving data form performance
    So do I :)
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Black screen with cursor blinking for a long time before loading windows 7

    Hey guys, i'm having an annoying problem with my macbook pro... i installed windows 7 using bootcamp and now i get a black screen with cursor blinking for a long time before loading windows... it *****, it takes about 3 minutes to get my notebook running
    The first time i installed windows seven i did not have this problem... i already deleted  the partition several times and installed windows again, but the black screen problem is still happening. i don't know what do
    You guys have any tips? anyone with the same problem?? this ***** a lot. thanks

    Hey Theoos..
    Yeah, my version is not.. the first time i installed it was ok, but after 30 days it was asking for a key... then i installed a cracked version, and the cursor started blinking on start up, and took like 2 minutes...
    I tested several cracks but with all of them happened the same.
    Thanks

Maybe you are looking for

  • Tabbed panels no longer showing up

    Hi folks, for quite some time I have used the widge Spry Tabbed Panels 2.0 in this page: www.arrowmark.co.nz/clients  and on this pagehttp://tatlerprime.co.nz/conference.html but the tabs no longer show up and all the content is laid out vertically d

  • How to populate View Criteria Item programatically

    Hi, I'm using Jdeveloper 11.1.2.3.0. I created a query panel with operator Between for EmployeeId attribute and used bind variables in the VC fromEmpId, toEmpId, and dropped the VC as a query panel on a jsf page. At run time, when i entered a value i

  • Using multiple sql query

    Im writing a jsp that pulls data from the database and loads it into dropdown-list looks like this: <select name="Source"> <% conn = DriverManager.getConnection(url, username, password); Statement stmt = conn.createStatement(ResultSet.TYPE_SCROLL_SEN

  • Flex2 and CF7.02 Remoting Errors

    For three days straight I have been trying to debug this blasted remoting issue based off of the helloworld quick start. I am getting the 405 Error when attempting remoting. If I goto directly to http://localhost/flex2gateway/ I get a 404 error. I ha

  • Display Distribution Lines, Invoice Number, and the PO Number

    In Toad, I'm trying to make a query that will display the invoice number and all distribution lines as well as if it's tied to a PO or not, and if so, what is the PO number - is this possible? This is what I've made so far and I don't think the outer