Data load taking so much time

Hello Everybody
i have enhanced the 0info_rec_attr datasource for few fields. I am getting right result for which i have done enhancement. But during infopackage run and DTP run it is taking so much time. when i run infopackage it is taking arround 12hrs for 4 lack records.
previously it use to take only 5mint to load same amount of data. Same thing is happening DTP also.
Kindly suggest why this is happeneing. Alhough data are coming in right way in infoprovider.
Regards
Saurabh

Hello Durgesh following is entire code. Kindly help me in optimization of tcode .
WHEN '0INFO_REC_ATTR'.
    CLEAR : t_inforec,t_inforec1,t_infodata.
    REFRESH:t_inforec,t_inforec1,t_infodata.
    LOOP AT i_t_data INTO wa_inforec.
      l_tabix = sy-tabix.
      SELECT infnr
             ekgrp
             ekorg
             loekz
             esokz
             werks
             waers
             norbm
             aplfz
             netpr
             peinh
             mwskz
             INTO CORRESPONDING FIELDS OF TABLE t_inforec1
             FROM eine
             WHERE infnr = wa_inforec-infnr.
      IF sy-subrc = 0.
        LOOP AT t_inforec1.
          MOVE-CORRESPONDING wa_inforec TO t_inforec.
          MOVE t_inforec1-ekgrp TO t_inforec-zzekgrp.
          MOVE t_inforec1-ekorg TO t_inforec-zzekorg.
          MOVE t_inforec1-esokz TO t_inforec-zzesokz.
          MOVE t_inforec1-werks TO t_inforec-zzwerks.
          MOVE t_inforec1-aplfz TO t_inforec-zzaplfz.
          MOVE t_inforec1-mwskz TO t_inforec-zzmwskz.
          MOVE t_inforec1-loekz TO t_inforec-zzdelflag_eine.
          MOVE t_inforec1-meins TO t_inforec-zzmeins.
          MOVE t_inforec1-netpr TO t_inforec-zznetpr.
          MOVE t_inforec1-peinh TO t_inforec-zzpeinh.
          MOVE t_inforec1-waers TO t_inforec-zzwaers.
          MOVE t_inforec1-norbm TO t_inforec-zznorbm.
         Delete flag from EINA
          SELECT SINGLE loekz meins
                 INTO (t_inforec-zzdelflag_eina,
                      t_inforec-zzmeins)
            FROM eina
            WHERE infnr = t_inforec1-infnr.
          APPEND t_inforec.
        ENDLOOP.
      ENDIF.
      APPEND LINES OF t_inforec TO t_infodata.
      CLEAR : t_inforec,t_inforec1.
      REFRESH:t_inforec,t_inforec1.
    ENDLOOP.
    DATA:
     l_datbi TYPE kodatbi,
     l_datab TYPE kodatab,
     l_knumh TYPE knumh,
     l_kopos TYPE kopos,
     l_kschl TYPE kscha,
     l_kbetr TYPE kbetr_kond,
     l_flifn TYPE flifn.
    LOOP AT t_infodata.
      SELECT SINGLE datbi datab knumh
      FROM a017
      INTO (l_datbi, l_datab, l_knumh)
      WHERE lifnr = t_infodata-lifnr AND
            matnr = t_infodata-matnr AND
            ekorg = t_infodata-zzekorg AND
            werks = t_infodata-zzwerks.
      t_infodata-zzdatbi = l_datbi.
      t_infodata-zzdatab = l_datab.
      t_infodata-zzknumh = l_knumh.
      MODIFY t_infodata INDEX sy-tabix.
      CLEAR: l_datbi, l_datab, l_knumh.
   ENDLOOP.
   LOOP AT t_infodata.
      SELECT SINGLE kopos kschl kbetr
        FROM konp
        INTO (l_kopos, l_kschl, l_kbetr)
        WHERE knumh = t_infodata-zzknumh.
      SELECT SINGLE flifn
        FROM eord
        INTO l_flifn
        WHERE matnr = t_infodata-matnr AND
            werks = t_infodata-zzwerks AND
            lifnr = t_infodata-lifnr.
      t_infodata-zzkopos = l_kopos.
      t_infodata-zzkschl = l_kschl.
      t_infodata-zzkbetr = l_kbetr.
      t_infodata-zzflifn = l_flifn.
      MODIFY t_infodata INDEX sy-tabix.
      CLEAR: l_kopos, l_kschl, l_kbetr.
    ENDLOOP.
    CLEAR : i_t_data.
    REFRESH:i_t_data.
    i_t_data[] = t_infodata[].
Regards
saurabh

Similar Messages

  • ODS to CUBE load taking too much time..

    Hi all ,
    we are loading the data from our ZODS to ZCUBE, but the data load is taking too much time , we haven't created any indexes , we alsotried by making infosource for the ODS but still tha same problem .. It is always showing 0 from 345674 records that is the records are not getting extracted from ODS .
    Can anybody help me in this regards , it is a bit urgent ..
    Thanks in advance.

    Hi,
    there are a few you can check. First you should check if this job hasn't ended in a dump with ST22.
    The next thing you can do, if the job doesn't end abnormaly, is to reduce the amount of records processed at the same time. Sometimes the system has trouble if the amount of records that it has to process is too large. Go to the InfoPackage -> DataS. Default Data Transfer -> Fill the maximum to 10% of de the default value. Try to run the load again.
    If the job still doesn't finish then you have to check wether or not there are any ABAP routines and/or formula involved in the update rule ? Maybe they running in a loop.
    regards,
    Raymond Baggen
    Uphantis bv

  • ODS to CUBE loading - taking too much time

    Hi Experts,
    I am loading data from R/3(4.7) to BW (3.5).
    I am loading with option --> PSA and then Data Target (ODS ).
    I have a selection criteria in Infopackage while loading from standard Datasource to ODS.
    It takes me 20 mins to load 300K records.
    But, from ODS to Infocube ( update method: Data Target Only), it is taking 8 hours.
    The data packet size in Infopackage is 20,000 ( same for ODS and Infocube).
    I also tried changing the data packet size, tried with full load , load with initialization,..
    I tried scheduling it as a background job too.
    I do not have any selection criteria in the infopackage from ODS to Cube.
    Please let me know how can I decrease this loading time from ODS to Infocube.

    Hi,
    To improve the data load performance
    1. If they are full loads then try to see if you make them delta loads.
    2. Check if there are complex routines/transformations being performed in any layer. In that case see if you can optimize those codes with the help of an abaper.
    3. Ensure that you are following the standard procedures in the chain like deleting Indices/secondary Indices before loading etc.
    4. Check whether the system processes are free when this load is running
    5. try making the load as parallel as possible if the load is happening serially. Remove PSA if not needed.
    6. When the load is not getiing processed due to huge volume of data, or more number of records per data packet, Please try the below option.
    1) Reduce the IDOC size to 8000 and number of data packets per IDOC as 10. This can be done in info package settings.
    2) Run the load only to PSA.
    3) Once the load is succesfull , then push the data to targets.
    In this way you can overcome this issue.
    Ensure the data packet sizing and also the number range buffering, PSA Partition size, upload sequence i.e, always load master data first, perform change run and then transaction data loads.
    Check this doc on BW data load perfomance optimization
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    BI Performance Tuning
    FAQ - The Future of SAP NetWeaver Business Intelligence in the Light of the NetWeaver BI&Business Objects Roadmap
    Thanks,
    JituK

  • Data load taking lot of time

    Hi All,
    I am trying to upload 2LIS_02_SCL data, I am trying to refresh the entire data using Init, it is taking lot of time in Prodcution. This is creating the problem for me for next Back ground Job which runs in the night for delta's. Please can anybody guide me how to speed up the load? This datasource is connected to 2 Infocubes and 1 ODS.
    Regards,
    Rajesh

    Hi Sal,
    I have done the same things as you said, my mistake is
    R/3:
    Locked all R/3 users.
    Deleted from LBWG.
    Filled setup tables using Document Date from 2006 to 2008 (Month wise i have filled).
    BW:
    I have cleaned entire data from 0PUR_C01 and 0PUR_C04.
    Loaded data at a time to 2 Cubes and 1 ODS using Init upload.
    It is started taking long time. Actual problem is load was not finished at the time of Daily load Process Chain starts(Night).
    I have cancelled the job. made available only delta to Process Chain.
    Now problem is escaled, again i started today same Init for 1 Cube only, again taking long time.
    Regards,
    Rajesh

  • Data load taking very long time between cube to cube

    Hi
    In our system the data loading is taking very long time between cube to cube using DTP in BI7.0,
    the maximum time consumption is happening at start of extraction step only, can anybody help in decreasing the start of extraction timing please
    Thanks
    Kiran

    Kindly little bit Elaborate your issue, Like how is mapping between two cubes, Is it One to one mapping or any Routine is there in Transformation. Any Filter/ Routine in DTP.  Also before loading data to Cube did you deleted Inedxes?
    Regards,
    Sushant

  • Taking too much time to load application

    Hi,
    I have deployed a j2ee application on oracle 10g version 10.1.2.0.2. But the application is taking too much time to load. After loading ,everything works fast.
    I have another 10g server (same version) in which the same application is loading very fast.
    When I checked the apache error logs found this :-
    [Thu Apr 26 09:17:31 2007] [warn] [client 10.1.20.9] oc4j_socket_recvfull timed out
    [Thu Apr 26 09:17:31 2007] [error] [client 10.1.20.9] [ecid: 89128867058,1] (4)Interrupted system call: MOD_OC4J_0038: Receiving data from oc4j exceeded the configured "Timeout" value and the error code is 4.
    [Thu Apr 26 09:17:31 2007] [error] [client 10.1.20.9] [ecid: 89128867058,1] MOD_OC4J_0054: Failed to call network routine to receive an ajp13 message from oc4j.
    [Thu Apr 26 09:17:31 2007] [error] [client 10.1.20.9] [ecid: 89128867058,1] MOD_OC4J_0033: Failed to receive an ajp13 message from oc4j.
    [Thu Apr 26 09:17:31 2007] [warn] [client 10.1.20.9] [ecid: 89128867058,1] MOD_OC4J_0078: Network connection errors happened to host: lawdb.keralalawsect.org and port: 12501 while receiving the first response from oc4j. This request is recoverable.
    [Thu Apr 26 09:17:31 2007] [error] [client 10.1.20.9] [ecid: 89128867058,1] MOD_OC4J_0121: Failed to service request with network worker: home_15 and it is not recoverable.
    [Thu Apr 26 09:17:31 2007] [error] [client 10.1.20.9] [ecid: 89128867058,1] MOD_OC4J_0013: Failed to call destination: home's service() to service the request.
    [Thu Apr 26 11:36:36 2007] [notice] FastCGI: process manager initialized (pid 21177)
    [Thu Apr 26 11:36:37 2007] [notice] Oracle-Application-Server-10g/10.1.2.0.2 Oracle-HTTP-Server configured -- resuming normal operations
    [Thu Apr 26 11:36:37 2007] [notice] Accept mutex: fcntl (Default: sysvsem)
    [Thu Apr 26 11:36:37 2007] [warn] long lost child came home! (pid 9124)
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0015: recv() returns 0. There has no message available to be received and oc4j has gracefully (orderly) closed the connection.
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0054: Failed to call network routine to receive an ajp13 message from oc4j.
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0033: Failed to receive an ajp13 message from oc4j.
    [Thu Apr 26 11:39:51 2007] [warn] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0078: Network connection errors happened to host: lawdb.keralalawsect.org and port: 12501 while receiving the first response from oc4j. This request is recoverable.
    [Thu Apr 26 11:39:51 2007] [warn] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0184: Failed to find an oc4j process for destination: home
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0145: There is no oc4j process (for destination: home) available to service request.
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0119: Failed to get an oc4j process for destination: home
    [Thu Apr 26 11:39:51 2007] [error] [client 10.1.20.9] [ecid: 80547835731,1] MOD_OC4J_0013: Failed to call destination: home's service() to service the request.
    [Thu Apr 26 11:46:33 2007] [notice] FastCGI: process manager initialized (pid 21726)
    [Thu Apr 26 11:46:34 2007] [notice] Oracle-Application-Server-10g/10.1.2.0.2 Oracle-HTTP-Server configured -- resuming normal operations
    [Thu Apr 26 11:46:34 2007] [notice] Accept mutex: fcntl (Default: sysvsem)
    [Thu Apr 26 11:46:34 2007] [warn] long lost child came home! (pid 21182)
    [Thu Apr 26 11:53:32 2007] [warn] [client 10.1.20.9] oc4j_socket_recvfull timed out
    [Thu Apr 26 11:53:32 2007] [error] [client 10.1.20.9] [ecid: 89138452752,1] (4)Interrupted system call: MOD_OC4J_0038: Receiving data from oc4j exceeded the configured "Timeout" value and the error code is 4.
    [Thu Apr 26 11:53:32 2007] [error] [client 10.1.20.9] [ecid: 89138452752,1] MOD_OC4J_0054: Failed to call network routine to receive an ajp13 message from oc4j.
    [Thu Apr 26 11:53:32 2007] [error] [client 10.1.20.9] [ecid: 89138452752,1] MOD_OC4J_0033: Failed to receive an ajp13 message from oc4j.
    [Thu Apr 26 11:53:32 2007] [warn] [client 10.1.20.9] [ecid: 89138452752,1] MOD_OC4J_0078: Network connection errors happened to host: lawdb.keralalawsect.org and port: 12501 while receiving the first response from oc4j. This request is recoverable.
    [Thu Apr 26 11:53:32 2007] [error] [client 10.1.20.9] [ecid: 89138452752,1] MOD_OC4J_0121: Failed to service request with network worker: home_15 and it is not recoverable.
    [Thu Apr 26 11:53:32 2007] [error] [client 10.1.20.9] [ecid: 89138452752,1] MOD_OC4J_0013: Failed to call destination: home's service() to service the request.
    Please HELP ME...

    Hi this is what the solution given by your link
    A.1.6 Connection Timeouts Through a Stateful Firewall Affect System Performance
    Problem
    To improve performance the mod_oc4j component in each Oracle HTTP Server process maintains open TCP connections to the AJP port within each OC4J instance it sends requests to.
    In situations where a firewall exists between OHS and OC4J, packages sent via AJP are rejected if the connections can be idle for periods in excess of the inactivity timeout of stateful firewalls.
    However, the AJP socket is not closed; as long as the socket remains open, the worker thread is tied to it and is never returned to the thread pool. OC4J will continue to create more threads, and will eventually exhaust system resources.
    Solution
    The OHS TCP connection must be kept "alive" to avoid firewall timeout issues. This can be accomplished using a combination of OC4J configuration parameters and Apache runtime properties.
    Set the following parameters in the httpd.conf or mod_oc4j.conf configuration files. Note that the value of Oc4jConnTimeout sets the length of inactivity, in seconds, before the session is considered inactive.
    Oc4jUserKeepalive on
    Oc4jConnTimeout 12000 (or a similar value)
    Also set the following AJP property at OC4J startup to enable OC4J to close AJP sockets in the event that a connection between OHS and OC4J is dropped due to a firewall timeout:
    ajp.keepalive=true
    For example:
    java -Dajp.keepalive=true -jar oc4j.jar
    Please tell me where or which file i should put the option
    java -Dajp.keepalive=true -jar oc4j.jar ??????/

  • BPC application is taking too much time to load

    Hi experts!
    I'm facing a very weird problem...
    We've developed a BPC application (app name: USM).
    This application is taking too much time to be loaded  in some computers (around 8 minutes to load).  Yes, in SOME computers.
    There is around 100.000 registers in the database and most coming from material master data.
    If I try to load this USM application in another computer, the process loads smoothly. The computer's hardware is all the same, the server is hyper estimated and everyone is in the same network.
    I talked to infrastructure departament and we made several tests. We run BPC on the server (loaded quickly), on several computers (some loads quickly, others don't), used wireless and cable connection (got all the same result) and checked communication between BW and BPC but it is ok.
    After all, I tried to load APSHEL application in the same enviroment and it loaded intantly. So, I guess it is something wrong with my application. But if was this, I suppose it should happen to all computers and not only with part of them.
    Have anybody ever seen something like this?
    Thank you in advance.
    Rubens
    Edited by: Rubens Massayuki Kumori on May 12, 2011 8:43 PM
    Edited by: Rubens Massayuki Kumori on May 12, 2011 8:46 PM

    Hi Rubens,
    I would try making a couple of test:
    1. I will install the client in a machine that is located in the same network segment, or try using a vpn that comunicates with the server bypassing all security devices, only to see if the network it's the problem.
    2. Making a full optimize of one application to see if maybe the problem it's related to the segmentation of the cubes (i don't think that this is the problem but give it a try).
    It is very wierd that in some computers happends and in others don't... also try to clean up the local cache of the applications in those computers that are giving to you bad performce and retry.
    hope it helps,

  • Full DTP taking too much time to load

    Hi All ,
    I am facing an issue where a DTP is taking too much time to load data from DSO to Cube via PC and also while manually running it.
    There are 6 such similar DTP's which load data for different countries(different DSO's and Cubes as source and target respectively) for last 7 days based on GI Date. All the DTP's are pulling almost same no. of records and finish within 25-30 min. But only one DTP takes around 3 hours. The problem started couple of days back.
    I have change the Parallel processes from 3->4->5 and packet size from 50,000->10,000->1.00.000 but no improvement. Also want to mention that all the source DSO's and target Cubes have the same structure. All the transformations have Field Routines and End Routines.
    Can you all please share some pointers which can help.
    Thanks
    Prateek

    HI Raman ,
    This is what I get when I check the report. Can this be causing issues as 2 rows have % >= 100
    ETVC0006           /BIC/DETVC00069     rows:      1.484    ratio:          0  %
    ETVC0006           /BIC/DETVC0006C     rows: 15.059.600    ratio:        103  %
    ETVC0006           /BIC/DETVC0006D     rows:        242    ratio:          0  %
    ETVC0006           /BIC/DETVC0006P     rows:         66    ratio:          0  %
    ETVC0006           /BIC/DETVC0006T     rows:        156    ratio:          0  %
    ETVC0006           /BIC/DETVC0006U     rows:          2    ratio:          0  %
    ETVC0006           /BIC/EETVC0006      rows: 14.680.700    ratio:        100  %
    ETVC0006           /BIC/FETVC0006      rows:          0    ratio:          0  %
    ETVC0007           rows: 13.939.200    density:              0,0  %

  • FDM 11.1.2.1 taking too much time to Export (Multi-load)

    Hi...
    There is something very odd here...
    I'm using FDQM 11.1.2.1 to multi-load a file. (300mb)
    So, the steps: Import (about 5min), Validade (about 5 min) - I have some explicit maps (about 2300).
    Export is my problem.... its taking more than 1 hour.
    I made a test using a Essbase rule to load this data using SQL with Join on FDM mapping table - TDATAMAP (because of the maps). So, in this way i spend about 3 minutes to load everything...
    The question is: why using FDM web client is taking to much time to Export data?
    Target: Planning App
    Thanks in advance
    -Cheers,
    Rafael

    Yes, you are exactly right. That is the correct order for the outline. Sorry, it's been a long time since I've used Essbase and my memory isn't what it used to be. Still, what keeps data load times to a minimum is to have the sparse fields in sort order so that the same data blocks don't have to be pulled into memory multiple times. The sparse fields should be sorted and in the same order as the outline. The best way to ensure that this is occurring is to have the sparse fields first, dense fields last.
    With dense dimensions sorted 1st, you can see how the sparse fields can get out of sort order, causing blocks to be revisited:
    D1A -> D2A -> S1A -> S2A -> S3A -> Data
    D1A -> D2A -> S1A -> S2A -> S3B -> Data
    D1B -> D2A -> S1A -> S2A -> S3A -> Data
    The 3rd record forces Essbase to re-visit the datablock that was already loaded by the 1st record.
    On the other hand, if sparse fields are first and sorted left to right, the first two records are loaded at the same time to the same block:
    S1A -> S2A -> S3A -> D1A -> D2A -> Data
    S1A -> S2A -> S3A -> D1B -> D2A -> Data
    S1A -> S2A -> S3B -> D1A -> D2A -> Data
    Sorry for the confusion. Hope this clarifies field sort order vs. outline order.

  • Ssrs Report taking to much time to load

    Hello all,
    I have a matrix report which was running nice before some days, but now taking too much time to load 
    i tested store procedure but it is ok to return same data in 5 seconds 
    and when i tested in ssrs BIDS side i set 60 sec for dataset timeout option, and it gets timeout 
    after i set 1000 seconds timeout but it is just rolling and rolling loading 
    when i check in executionlog of reportserver i found  something like ..
    TimeDataRetrieval     TimeProcessing     TimeRendering       Status                                          
            RowCount
    131951   323    
    95         rsHttpRuntimeClientDisconnectionError    1784
    Reply me fast please, Help must appreciated .. 
    Thanks ..
    Dilip Patil..

    Hi Dilip,
    According to your description, your report keeps loading symbol when rendering. Right?
    In this scenario, as you can see in the Excecution Log, it spends a lot of time on data retrieval. Based on the status information, it's the issue on the conncetion. Since you have tested the stored procedure, and it works properly. Please check if the credentials
    for data source has permisson to connect the database. You can also Test Connection when creating data source. As you mentioned, it uses Raido Frequency to connection. There might be large network traffic on transfering data. 
    Reference:
    Troubleshooting Reports: Report Performance
    Best Regards,
    Simon Hou

  • Creative Cloud is taking too much time to load and is not downloading the one month trial for Photoshop I just paid money for.

    Creative Cloud is taking too much time to load and is not downloading the one month trial for Photoshop I just paid money for.

    stop the download if it's stalled, and restart your download.

  • Load is taking too much time & going to short dump.

    Hi,
                We have around 1000 records to be loaded from r/3 . The  load is taking too much time and it is still running.
    When tried to manually update the records ,we are getting an error ' converting  the rate into indirect quote: CAD/USD( ex. rate type EURX). Did any one faced this type of error.
    Thanks,
    APR

    Hi,
    what are you trying to do? If you wanna load exchange rates this can be done via context menu of the appropriate source system ( transfer global settings/exchange rates)

  • Job is taking too much time during Delta loads

    hi
    when i tried to extract Delta records from R3 for Standard Extractor 0FI_GL_4 it is taking 46 mins while there are very less no. of delta records(193 records only).
    PFA the R3 Job log. the major time is taking in calling the Customer enhacement BW_BTE_CALL_BW204010_E.
    please let me know why this is taking too much time.
    06:10:16  4 LUWs confirmed and 4 LUWs to be deleted with FB RSC2_QOUT_CONFIRM_DATA
    06:56:46  Call up of customer enhancement BW_BTE_CALL_BW204010_E (BTE) with 193 records
    06:56:46  Result of customer enhancement: 193 records
    06:56:46  Call up of customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 193 records
    06:56:46  Result of customer enhancement: 193 records
    06:56:46  Asynchronous sending of data package 1 in task 0002 (1 parallel tasks)
    06:56:47  IDOC: InfoIDOC 2, IDOC no. 121289649, duration 00:00:00
    06:56:47  IDOC: Begin 09.05.2011 06:10:15, end 09.05.2011 06:10:15
    06:56:48  Asynchronous sending of InfoIDOCs 3 in task 0003 (1 parallel tasks)
    06:56:48  Through selection conditions, 0 records filtered out in total
    06:56:48  IDOC: InfoIDOC 3, IDOC no. 121289686, duration 00:00:00
    06:56:48  IDOC: Begin 09.05.2011 06:56:48, end 09.05.2011 06:56:48
    06:56:54  tRFC: Data package 1, TID = 3547D5D96D2C4DC7740F217E, duration 00:00:07, ARFCSTATE =
    06:56:54  tRFC: Begin 09.05.2011 06:56:47, end 09.05.2011 06:56:54
    06:56:55  Synchronous sending of InfoIDOCs 4 (0 parallel tasks)
    06:56:55  IDOC: InfoIDOC 4, IDOC no. 121289687, duration 00:00:00
    06:56:55  IDOC: Begin 09.05.2011 06:56:55, end 09.05.2011 06:56:55
    06:56:55  Job finished
    Regards
    Atul

    Hi Atul,
    Have you written any customer exit code . If yes check for the optimization for it .
    Kind Regards,
    Ashutosh Singh

  • Query taking too much time with dates??

    hello folks,
    I am trying pull some data using the date condition and for somereason its taking too much time to return the data
       and trunc(al.activity_date) = TRUNC (SYSDATE, 'DD') - 1     --If i use this its takes too much time
      and al.activity_date >= to_date('20101123 000000', 'YYYYMMDD HH24MISS')
       and al.activity_date <= to_date('20101123 235959', 'YYYYMMDD HH24MISS') -- If i use this it returns the data in a second. why is that?
    How do i get the previous day without using the hardcoded to_date('20101123 000000', 'YYYYMMDD HH24MISS'), if i need to retrieve it faster??

    Presumably you've got an index on activity_date.
    If you apply a function like TRUNC to activity_date, you can no longer use the index.
    Post execution plans to verify.
    and al.activity_date >= TRUNC (SYSDATE, 'DD') - 1
    and al.activity_date < TRUNC (SYSDATE, 'DD')

  • I am extracting the data from ECC To bw .but Data Loading taking long tim

    Hi All,
                     i am extracting the data from ECC To BI Syatem..but Data Loading Taking Long time. from last   6 hoursinfopackage is running.still it is showing yellow.Manually i made the red.and delete again i applied repeat of the last delta.but same proble is coming .in the status job is showing bckground job is not finished at source system.we requested to basis.basis people killed that job.again we schedule the chain also again same problem is coming.how can i solve this issue.
    Thanks ,
    chandu

    Hi,
    There are different places to track your job. Once your job is triggered in BW, you can track your load job where exactly it is taking more time and why. Follow below steps:
    1) After InfoPackage is triggered, then take the request number and go to source system to check your extraction job status.
    You can get the job status by taking the request number from BW and go to transaction SM37 in ECC. Then give the request number with begining '' and ending ''.  Also give '*' to user name.
    Job name:  REQ_XXXXXX
    User Name: *
    Check the job status whether job is completed or cancelled or short dump. If the job is still running check in SM66 whether you can see any process. If not accordingly you got to check in ST22 or SM21 in ECC. If the job is complete, then the same in BW side now.
    2) Check the data arrived in PSA, if not check whether Transfer routines or start routines are having bad SQL or code. Similarly in update rules.
    3) Once it is through in Source system (ECC), Transfer rules , Update Rules, then the next task is updating the data might some time take more time which might be based on some parameters ( Number of parallel process to update database ). Check whether updating the database is taking more time and may be you got to check with the DBA guy also.
    At all the times you should see minimum of atleast once process running all the time in SM66 till the time your job gets complete. If not you will see a log in ST22.
    Let me know if you still have questions.
    Assigning points is the only way of saying thanks in SDN.
    Thanks,
    Kumar.

Maybe you are looking for