Delete setup table takes long time

Hi Experts,
I'm trying to use 2LIS_02_HDR data source to populate ODS 0BBP_PO, 0BBP_INV and 0BBP_CON as this is a basis for the spent analysis cube 0BBP_C01.
I'm pretty sure before i ran the setup i deleted the setup files. however when i loaded the data the number of docs shown in NPRT is 3429. however when i loaded the data to BI, it pulled in almost 10 million records for the datasource, which totally baffled me. Can anyone tell me why this is?
since this did not seem correct. i decided to do the initialization again so did a LBWG and this has been running for the last 2 days. Can anyone tell me why this is taking so long as well?
This is on the dev box and not even on production.
help please.
thanks
dilanke

Hi soma,
Thanks for you're reply but this is not a basis issue.
My main question is why only 10 Million records where pulled into BI when only 3500 docs were read into the setup table?
thanks
dilanke

Similar Messages

  • Help!!! Purchasing setup run takes long time

    Hi Experts,
    I'm pulling purchasing data into BI from R/3. I'm having a problem where the setup run takes an extremely long time. we have approximately 3 years of data on R/3. when i look at EKBE table, there are clost to 5 Million records.
    Basically when i run the scedule it took atleast 3 weeks and still did not complete. even when i run it only for 2008 it took 6 days but still didnt complete. so today i ran it for one month (Jan 2008) and its been running for the last 3 hours without completing.
    has anybody come up with an issue like this?
    I'm not sure if there is something wrong with the setup run as in im wondering if the job is hanging or something? is there anyway to figure this out?
    I'm stuck with this at the moment, ive looked around but to no avail.
    would really appreciate you're help as deadlines on this are looming.
    thanks,
    dilanke
    Points will be rewarded.

    Hi,
    5 million purchasing records should take max 6 hour and it seems that you have a problem with problem which is filling setup table.Did you check OSS notes? There should be something pointing this isssue.
    You can run NPRT transaction in order to see  how long it took in the previous jobs.
    Regards.

  • Delete Index in Process Chain Takes long time after SAP BI 7.0 SP 27

    After upgrading to SAP BI 7.0 SP 27 Delete index Process & Create index process in Process chain takes long time.
    For example : Delete index for 0SD_C03 takes around 55 minutes.
    Before SP upgrade it takes around 2 minutes to delete index from 0SD_C03.
    Regards
    Madhu P Menon

    Hi,
    Normally  index  creation or deletion can take long time in case  your database statistics are not updated properly, so can check  stat  after your data loading is completed and index generation is done,  Do creation of database statistics.
    Then try to recheck ...
    Regards,
    Satya

  • INSERT INTO TABLE using SELECT takes long time

    Hello Friends,
    --- Oracle version 10.2.0.4.0
    --- I am trying to insert around 2.5 lakhs records in a table using INSERT ..SELECT. The insert takes long time and seems to be hung.
    --- When i try to SELECT the query fetches the rows in 10 seconds.
    --- Any clue why it is taking so much time

    vishalrs wrote:
    Hello Friends,hello
    >
    >
    --- Oracle version 10.2.0.4.0
    alright
    --- I am trying to insert around 2.5 lakhs records in a table using INSERT ..SELECT. The insert takes long time and seems to be hung.
    I don't know how a lakh is, but it sounds like a lot...
    --- When i try to SELECT the query fetches the rows in 10 seconds.
    how did you test this? and did you fetch the last record, or just the first couple of hundred.
    --- Any clue why it is taking so much timeWithout seeing anything, it's impossible to tell the reason.
    Search the forum for "When your query takes too long"

  • CV04N takes long time to process select query on DRAT table

    Hello Team,
    While using CV04N to display DIR's, it takes long time to process select query on DRAT table. This query includes all the key fields. Any idea as to how to analyse this?
    Thanks and best regards,
    Bobby
    Moderator message: please read the sticky threads of this forum, there is a lot of information on what you can do.
    Edited by: Thomas Zloch on Feb 24, 2012

    Be aware that XP takes approx 1gb of your RAM leaving you with 1gb for whatever else is running. MS Outlook is also a memory hog.
    To check Virtual Memory Settings:
    Control Panel -> System
    System Properties -> Advanced Tab -> Performance Settings
    Performance Options -> Adavanced Tab - Virtual Memory section
    Virtual Memory -
    what are
    * Initial Size
    * Maximum Size
    In a presentation at one of the Hyperion conferences years ago, Mark Ostroff suggested that the initial be set to the same as Max. (Max is typically 2x physical RAM)
    These changes may provide some improvement.

  • SELECT statement takes long time

    Hi All,
    In the following code, if the T_QMIH-EQUNR contains blank or space values ,SELECT statement takes longer time to acess the data from OBJK table. If it T_QMIH-EQUNR contains values other than blank, performance is good and it fetches data very fast.
    Already we have indexes for EQUNR in OBJK table.
    Only for blank entries , it takes much time.Can anybody tell why it behaves for balnk entries?
    if not T_QMIH[] IS INITIAL.
            SORT T_QMIH BY EQUNR.
            REFRESH T_OBJK.
            SELECT EQUNR OBKNR
              FROM OBJK INTO TABLE T_OBJK
              FOR ALL ENTRIES IN T_QMIH
              WHERE OBJK~TASER = 'SER01' AND
             OBJK~EQUNR = T_QMIH-EQUNR.
    Thanks
    Ajay

    Hi
    You can use the field QMIH-QMNUM with OBJK-IHNUM
    in QMIH table, EQUNR is not primary key, it will have multiple entries
    so to improve the performance use one dummy internal table for QMIH  and sort it on EQUNR
    delete adjacent duplicates from d_qmih and use the same in for all entries
    this will improve the performance.
    Also use the fields in sequence of the index and primary keys also in select
    if not T_QMIH[] IS INITIAL.
    SORT T_QMIH BY EQUNR.
    REFRESH T_OBJK.
    SELECT EQUNR OBKNR
    FROM OBJK INTO TABLE T_OBJK
    FOR ALL ENTRIES IN T_QMIH
    WHERE  IHNUM =  T_QMIH-QMNUM
    OBJK~TASER = 'SER01' AND
    OBJK~EQUNR = T_QMIH-EQUNR.
    try this and let me know
    regards
    Shiva

  • Reactivating Aggregates takes long time

    Hi All,
    Last week, we were in the process of removing some data from an Infocube, which lied outside its retention period (still attempting to setup an archive)...
    Anyways, we would deactivate the aggregate on the cube, and then do a selective delete on several different date ranges.  After the deletes are completed, we would rebuild or fill the aggregate.  The first time it took around 4.5 hours to rebuild, which seems long.  Then after deactivating the aggregate again, and doing more deletes on the same cube the day after next, we tried rebuilding the aggregate once more.  This took around 7 hours to complete, with less data in the infocube.
    Any ideas???

    Hi,
    Several reasons to it. Something which I know of would be the following reasons.
    1. The aggregate which you are deactiviating and filling is having the date field in it along with another date / time field which takes longer time for filling once adjusted with some deletion.
    2. Check the size of the aggregates in terms of records being added everytime.
    3. Under Manage Aggregates there is somehting called as "Aggregate Tree" under Goto -> Aggregate Tree. Here check the hiearchy of the aggregate you are filling. Go in the given sequence. As that will always be faster, regardless of how the aggregates are placed in the manage tab.
    Hope these helps.
    Thanks,
    Pradip.

  • BPM Process chain takes long time to process

    We have BI7, Netweaver 2004s on Oracle and SUN Solaris
    There is a process chain (BPM) which pulls data from the CRM system into BW. The scheduled time to run this chain is 0034 hrs. This chain should ideally complete before / around 0830 Hrs. <b>Now the problem is that every alternate day this chain behaves normally and gets completed well before 0830 hrs but every alternate day this chain fails…</b> there are almost 40 chains running daily. Some are event triggered (dependent with each other) or some run in parallel. In this, (BPM) process chain, usually there are 5 requests with 3 Delta and 2 full uploads (Master Data). The delta uploads finishes in 30 minutes without any issues with very few record transfers. The first full upload is from 0034 hrs to approximately 0130 hrs and the 2nd upload is from 0130 hrs to 0230 hrs. Now if the 1st upload gets delayed then the people who are initiating these chains, stop the 2nd full upload and continue it after all the process chains are completed. Now this entire BPM process chain sometimes takes 17 -18 hrs to complete!!!!!
    No other loads in CRM or BW when these process chains are running
    CRM has background jobs to push IDOCS to BW which run every 2 minutes which runs successfully
    Yesterday this chain got completed successfully (well within stipulated time) with over 33,00,000 records transferred but sometimes it has failed to transfer even 12,00,000 records!!
    Attaching a zip file, please refer the “21 to 26 Analysis screen shot.doc” from the zip file
    Within the zip file, attaching “Normal timings of daily process chains.xls” – the name explains it….
    Also within the zip file refer “BPM Infoprovider and data source screen shot.doc” please refer this file as the infopackage (page 2) which was used in the process chain is not displayed later on in page number 6 BUT CHAIN GOT SUCESSFULLY COMPLETED
    We have analyzed:--
    1)     The PSA data for BPM process chain for past few days
    2)     The info providers for BPM process chain for past few days
    3)     The ODS entries for BPM process chain for past few days
    4)     The point of failure of BPM process chain for past few days
    5)     The overall performance of all the process chains for past few days
    6)     The number of requests in BW for this process chain
    7)     The load on CRM system for past few days when this process chain ran on BW system
    As per our analysis, there are couple of things which can be fixed in the BW system:--
    1)     The partner agreement (transaction WE20) defined for the partner LS/BP3CLNT475 mentions both message types RSSEND and RSINFO: -- collect IDOCs and pack size = 1 Since the pack size = 1 will generate 1 TRFC call per IDOC, it should be changed to 10 so that less number of TRFCs will be generated thus less overhead for the BW server resulting in the increase in performance
    2)     In the definition of destination for the concerned RFC in BW (SM59), the “Technical Setting” tab says the “Load balancing” option = “No”. We are planning to make it “Yes”
    But we believe that though these changes will bring some increase in performance, this is not the root cause of the abnormal behavior of this chain as this chain runs successfully on every alternate day with approximately the same amount of load in it.
    I was not able to attach the many screen shots or the info which I had gathered during my analysis. Please advice how do I attach these files
    Best Regards,

    Hi,
    Normally  index  creation or deletion can take long time in case  your database statistics are not updated properly, so can check  stat  after your data loading is completed and index generation is done,  Do creation of database statistics.
    Then try to recheck ...
    Regards,
    Satya

  • Enhancing the standard  Data source without deleting setup tables

    Hi all,
    I am in the the Support Project. My requirement is I want to Enhance u201C2LIS_13_VDITMu201D LO- data source with two fields without disturbing the delta.
    Please suggest me how I have to do this.
    As Per my Knowledge ,
    1. we have to delete setup tables
    2. Enhance the data source & re populate the setup tables.
    3. Delete the data in the cube & add the two new fields in the cube & repopulate the cube with new Initial .
    4. after that delta will be enabled through job control.
    But this process is not suitable for our requirement because delta was enabled long back it is going very smooth till date, I donu2019t want to disturb that process.
    So please suggest me is there any other procedure to do this.
    Thanks,
    Kiran Manyam

    Hi,
    If historical data (loaded earlier in to BW) are not required for the two enhance field, then it is not required to deleted the setup table and reload them to BI.
    In this case simply you can follow the following procedure.
    1. Enhance the fields, and update the transfor structure(to unhide these fields). In BI update the required data target with respective IO. and in exit populate the enhance fields. No need to disturb the delta
    2. Replicated the DS in BI and do the mappings in tranformation.
    Here the existing delta is working, and you will be populating the two fields in the exit only.
    Thanks,
    Jugal.

  • MVIEW refresh takes long time

    Materialized view takes long time to refresh but when i tried select & insert into table it's very fast.
    i executed SQL and it takes ust 1min ( total rows is 447 )
    but while i refresh the MVIEW it takes 1.5 hrs ( total rows is 447 )
    MVIEW configration :-
    CREATE MATERIALIZED VIEW EVAL.EVALSEARCH_PRV_LWC
    TABLESPACE EVAL_T_S_01
    NOCACHE
    NOLOGGING
    NOCOMPRESS
    NOPARALLEL
    BUILD DEFERRED
    REFRESH FORCE ON DEMAND
    WITH PRIMARY KEY
    Not sure why so much diffrence

    infant_raj wrote:
    Materialized view takes long time to refresh but when i tried select & insert into table it's very fast.
    i executed SQL and it takes ust 1min ( total rows is 447 )
    but while i refresh the MVIEW it takes 1.5 hrs ( total rows is 447 )A SELECT does a consistent read.
    A MV refresh does that and also writes database data.
    These are not the same thing and cannot be directly compared.
    So instead of pointing at the SELECT execution time and asking why the MV refresh is not as fast, look instead WHAT the refresh is doing and HOW it is doing that.
    Is the execution plan sane? What events are the top ones for the MV refresh? What are the wait states that contributes most to the processing time of the refresh?
    You cannot use the SELECT statement's execution time as a direct comparison metric. The work done by the refresh is more than the work done by the SELECT. You need to determine exactly what work is done by the refresh and whether that work is done in a reasonable time, and how other sessions are impacting the refresh (it could very well be blocked by another session).

  • Takes Long time for Data Loading.

    Hi All,
    Good Morning.. I am new to SDN.
    Currently i am using the datasource 0CRM_SRV_PROCESS_H and it contains 225 fields. Currently i am using around 40 fields in my report.
    Can i hide the remaining fields in the datasource level itself (TCODE : RSA6)
    Currently data loading takes more time to load the data from PSA to ODS (ODS 1).
    And also right now i am pulling some data from another ODS(ODS 2)(LookUP). It takes long time to update the data in Active data table of the ODS.
    Can you please suggest how to improve the performance of dataloading on this Case.
    Thanks & Regards,
    Siva.

    Hi....
    Yes...u can hide..........just Check the hide box for those fields.......R u in BI 7.0 or BW...........whatever ........is the no of records is huge?
    If so u can split the records and execute............I mean use the same IP...........just execute it with different selections.........
    Check in ST04............is there are any locks or lockwaits..........if so...........Go to SM37 >> Check whether any Long running job is there or not.........then check whether that job is progressing or not............double click on the Job >> From the Job details copy the PID..............go to ST04 .....expand the node............and check whether u r able to find that PID there or not.........
    Also check System log in SM21............and shortdumps in ST04........
    Now to improve performance...........u can try to increase the virtual memory or servers.........if possiblr........it will increase the number of work process..........since if many jobs run at a time .then there will be no free Work prrocesses to proceed........
    Regards,
    Debjani......

  • The 0co_om_opa_6 ip in the process chains takes long time to run

    Hi experts,
    The 0co_om_opa_6 ip in the process chains takes long time to run around 5 hours in production
    I have checked the note 382329,
    -> where the indexes 1 and 4 are active
    -> index 4 was not "Index does not exist in database system ORACLE"- i have assgined to " Indexes on all database systems and ran the delta load in development system, but guess there are not much data in dev it took 2-1/2 hrs to run as it was taking earlier. so didnt find much differnce in performance.
    As per the note Note 549552 - CO line item extractors: performance, i have checked in the table BWOM_SETTINGS these are the settings that are there in the ECC system.
    -> OLTPSOURCE -  is blank
       PARAM_NAME - OBJSELSIZE
       PARAM_VALUE- is blank
    -> OLTPSOURCE - is blank
       PARAM_NAME - NOTSSELECT
       PARAM_VALUE- is blank
    -> OLTPSOURCE- 0CO_OM_OPA_6
       PARAM_NAME - NOBLOCKING
       PARAM_VALUE- is blank.
    Could you please check if any other settings needs to be done .
    Also for the IP there is selction criteris for FISCALYEAR/PERIOD from 2004-2099, also an inti is done for the same period as a result it becoming difficult for me to load for a single year.
    Please suggest.

    The problem was the index 4 was not active in the database level..it was recommended by the SAP team to activate it in se14..however while doing so we face few issues se14 is a very sensitive transaction should be handled carefully ... it should be activate not created.
    The OBJSELSIZE in the table BWOM_SETTINGS has to be Marked 'X' to improve the quality as well as the indexe 4 should be activate at the abap level i.e in the table COEP -> INDEXES-> INDEX 4 -> Select the  u201Cindex on all database systemu201D in place of u201CNo database indexu201D, once it is activated in the table abap level you can activate the same indexes in the database level.
    Be very carefull while you execute it in se14 best is to use db02 to do the same , basis tend to make less mistake there.
    Thanks Hope this helps ..

  • Procedure takes long time to execute...

    Hi all
    i wrote the proxcedure but it takes long time to execute.
    The INterdata table contains 300 records.
    Here is the procedure:
    create or replace procedure inter_filter
    is
         /*v_sessionid interdata.sessionid%type;
         v_clientip interdata.clientip%type;
         v_userid interdata.userid%type;
         v_logindate interdata%type;
         v_createddate interdata%type;
         v_sourceurl interdata%type;
         v_destinationurl interdata%type;*/
         v_sessionid filter.sessionid%type;
         v_filterid filter.filterid%type;
         cursor c1 is
         select sessionid,clientip,browsertype,userid,logindate,createddate,sourceurl,destinationurl
         from interdata;
         cursor c2 is
         select sessionid,filterid
         from filter;
    begin
         open c2;
         loop
              fetch c2 into v_sessionid,v_filterid;
              for i in c1 loop
                   if i.sessionid = v_sessionid then
                        insert into filterdetail(filterdetailid,filterid,sourceurl,destinationurl,createddate)
                        values (filterdetail_seq.nextval,v_filterid,i.sourceurl,i.destinationurl,i.createddate);
                   else
                        insert into filter (filterid,sessionid,clientip,browsertype,userid,logindate,createddate)
                        values (filter_seq.nextval,i.sessionid,i.clientip,i.browsertype,i.userid,i.logindate,i.createddate);
                        insert into filterdetail(filterdetailid,filterid,sourceurl,destinationurl,createddate)
                        values (filterdetail_seq.nextval,filter_seq.currval,i.sourceurl,i.destinationurl,i.createddate);
                   end if;
              end loop;
         end loop;
         commit;
    end
    Please Help!
    Prathamesh

    i wrote the proxcedure but it takes long time to execute.Please define "long time". How long does it take? What are you expecting it to take?
    The INterdata table contains 300 records.But how many records are there in the FILTER table? As this is the one you are driving off this is going to determine the length of time it takes to complete. Also, this solution inserts every row in the INTERDATA table for each row in the FILTER table - in other words, if the FILTER table has twenty rows to start with you are going to end up with 6000 rows in FILTERDETAIL. No wonder it takes a long time. Is that want you want?
    Also of course, you are using PL/SQL cursors when you ought to be using set operations. Did you try the solution I posted in Re: Confusion in this  scenario>>>>>>> on this topic?
    Cheers, APC

  • It takes long time to load the parameter's value when running we run report

    Hi,
    It takes long time to load the parameter's value when running we run report. What could cause this? How to troubleshoot the behavior of the report? Could I use Profile and what events should i select?
    Thanks

    Hi jori5,
    Based on my understanding, after changing the parameter, the report render very slow, right?
    In Reporting Service, the total time to generate a report include TimeDataRetreval, TimeProcessing and TimeRendering. To analyze which section take much time, we can check the table Executionlog3 in the ReportServer database. For more information, please
    refer to this article:
    More tips to improve performance of SSRS reports.
    In your scenario, since you mention the query spends less time, the delay might happens during report processing and report rendering section. So you should check Executionlog3 to check which section costs most of time, then you can refer to this article
    to optimize your report:
    Troubleshooting Reports: Report Performance.
    If you have any question, please feel free to ask.
    Best regards,
    Qiuyun Yu

  • Report execution takes long time

    Dear all,
    we have a report which takes long time to exceute due to select statement.. here is the code..
    SELECT vkorg vtweg spart kunnr kunn2 AS division FROM knvp
        INTO CORRESPONDING FIELDS OF TABLE hier
        WHERE kunn2 IN s_kunnr
        AND   vkorg EQ '0001'
        AND   parvw EQ 'ZV'.
        l_parvw = 'WE'.
    SORT hier.
    * select all invoices within the specified invoice creation dates.
      CHECK NOT hier[] IS INITIAL.
      SELECT vbrk~vbeln vbrk~fkart vbrk~waerk vbrk~vkorg vbrk~vtweg vbrk~spart vbrk~knumv
             vbrk~konda vbrk~bzirk vbrk~pltyp vbrk~kunag vbrp~vbeln vbrp~aubel vbrp~posnr
             vbrp~fkimg vbrp~matnr vbrp~prctr vbpa~kunnr
             vbrp~pstyv vbrp~uepos
             vbrp~kvgr4  vbrp~ean11          
        INTO CORRESPONDING FIELDS OF TABLE it_bill
        FROM vbrk INNER JOIN vbrp ON vbrp~vbeln = vbrk~vbeln
                  INNER JOIN vbpa ON vbpa~vbeln = vbrk~vbeln
        FOR ALL entries IN hier
        WHERE (lt_syntax)
        AND   vbrk~vbeln IN s_vbeln
    *    AND   vbrk~erdat IN r_period
        AND   vbrk~fkdat IN r_period       
        AND   vbrk~rfbsk EQ 'C'
        AND   vbrk~vkorg EQ hier-vkorg
        AND   vbrk~vtweg EQ hier-vtweg
        AND   vbrk~spart EQ hier-spart.
    Can anyone say about how to reduce the execution time.?
    Edited by: Thomas Zloch on Sep 22, 2010 2:46 PM - please use code tags

    Hi
    first of all never use move corressponding.
    Rather you should declare a work area for table hier.
    select values into the work area and then append that workarea into the table hier.
    In case of for all entries include all the primary keys in the selection and for the keys which are of no use declare constants with initial valules like:
    'prmkey' is a primary field for table 'tab1' .
    Constants: field1 type tab1-prmkey value initial.
    and then in your where condition write.
    prmkey GE field1.
    I hope it is clear to you now.
    Thanks
    lalit Gupta

Maybe you are looking for

  • Time field in alv list

    Hi all , i want to display alv list with time field . if no data exists in that field its showing 00:00:00 , i have to show it blank. if i change time field as type c in my internal table its showing like this 121314, i need to display it 12:13:14 .

  • HT201237 Final Cut Pro X 10.0.8 not working under Yosemite

    I have Final Cut Pro X 10.0.8 and it stopped working under Yosemite. Are there any fixes or work arounds? Oddly enough, my copy of Final Cut Express 4.01 still works! MacBook Pro 13" (Late 2011) 12 GB RAM 180 GB Intel 530 SSD

  • NVision report through Report Book generating special characters

    Problem generating nVision Tabular report using Report Book. Reports brings special characters on two tabs after half way down the report. When we run the same report through two-tier mode it comes out clean. Tested the production and test layouts, t

  • Midi issues

    I'm using Logic Express 7 with latest upgrade, Roland D-20 keyboard, TD-20 Drum Set connected with a MidiSport 4X4. I can record midi with the D-20. I can even record multimidi tracks. But when I go to record with the TD-20 to lay out a drum midi tra

  • Safari starts automatically when restarting (Snow Leopard)

    Two days ago, Safari began to start automatically when restarting the system (iMac, Snow Leopard). I checked in Accounts>Login Item and it doesn't appear there. I also right-clicked the icon in the dock and the option to open when starting session wa