Delete Index in Process Chain Takes long time after SAP BI 7.0 SP 27

After upgrading to SAP BI 7.0 SP 27 Delete index Process & Create index process in Process chain takes long time.
For example : Delete index for 0SD_C03 takes around 55 minutes.
Before SP upgrade it takes around 2 minutes to delete index from 0SD_C03.
Regards
Madhu P Menon

Hi,
Normally  index  creation or deletion can take long time in case  your database statistics are not updated properly, so can check  stat  after your data loading is completed and index generation is done,  Do creation of database statistics.
Then try to recheck ...
Regards,
Satya

Similar Messages

  • The 0co_om_opa_6 ip in the process chains takes long time to run

    Hi experts,
    The 0co_om_opa_6 ip in the process chains takes long time to run around 5 hours in production
    I have checked the note 382329,
    -> where the indexes 1 and 4 are active
    -> index 4 was not "Index does not exist in database system ORACLE"- i have assgined to " Indexes on all database systems and ran the delta load in development system, but guess there are not much data in dev it took 2-1/2 hrs to run as it was taking earlier. so didnt find much differnce in performance.
    As per the note Note 549552 - CO line item extractors: performance, i have checked in the table BWOM_SETTINGS these are the settings that are there in the ECC system.
    -> OLTPSOURCE -  is blank
       PARAM_NAME - OBJSELSIZE
       PARAM_VALUE- is blank
    -> OLTPSOURCE - is blank
       PARAM_NAME - NOTSSELECT
       PARAM_VALUE- is blank
    -> OLTPSOURCE- 0CO_OM_OPA_6
       PARAM_NAME - NOBLOCKING
       PARAM_VALUE- is blank.
    Could you please check if any other settings needs to be done .
    Also for the IP there is selction criteris for FISCALYEAR/PERIOD from 2004-2099, also an inti is done for the same period as a result it becoming difficult for me to load for a single year.
    Please suggest.

    The problem was the index 4 was not active in the database level..it was recommended by the SAP team to activate it in se14..however while doing so we face few issues se14 is a very sensitive transaction should be handled carefully ... it should be activate not created.
    The OBJSELSIZE in the table BWOM_SETTINGS has to be Marked 'X' to improve the quality as well as the indexe 4 should be activate at the abap level i.e in the table COEP -> INDEXES-> INDEX 4 -> Select the  u201Cindex on all database systemu201D in place of u201CNo database indexu201D, once it is activated in the table abap level you can activate the same indexes in the database level.
    Be very carefull while you execute it in se14 best is to use db02 to do the same , basis tend to make less mistake there.
    Thanks Hope this helps ..

  • BPM Process chain takes long time to process

    We have BI7, Netweaver 2004s on Oracle and SUN Solaris
    There is a process chain (BPM) which pulls data from the CRM system into BW. The scheduled time to run this chain is 0034 hrs. This chain should ideally complete before / around 0830 Hrs. <b>Now the problem is that every alternate day this chain behaves normally and gets completed well before 0830 hrs but every alternate day this chain fails…</b> there are almost 40 chains running daily. Some are event triggered (dependent with each other) or some run in parallel. In this, (BPM) process chain, usually there are 5 requests with 3 Delta and 2 full uploads (Master Data). The delta uploads finishes in 30 minutes without any issues with very few record transfers. The first full upload is from 0034 hrs to approximately 0130 hrs and the 2nd upload is from 0130 hrs to 0230 hrs. Now if the 1st upload gets delayed then the people who are initiating these chains, stop the 2nd full upload and continue it after all the process chains are completed. Now this entire BPM process chain sometimes takes 17 -18 hrs to complete!!!!!
    No other loads in CRM or BW when these process chains are running
    CRM has background jobs to push IDOCS to BW which run every 2 minutes which runs successfully
    Yesterday this chain got completed successfully (well within stipulated time) with over 33,00,000 records transferred but sometimes it has failed to transfer even 12,00,000 records!!
    Attaching a zip file, please refer the “21 to 26 Analysis screen shot.doc” from the zip file
    Within the zip file, attaching “Normal timings of daily process chains.xls” – the name explains it….
    Also within the zip file refer “BPM Infoprovider and data source screen shot.doc” please refer this file as the infopackage (page 2) which was used in the process chain is not displayed later on in page number 6 BUT CHAIN GOT SUCESSFULLY COMPLETED
    We have analyzed:--
    1)     The PSA data for BPM process chain for past few days
    2)     The info providers for BPM process chain for past few days
    3)     The ODS entries for BPM process chain for past few days
    4)     The point of failure of BPM process chain for past few days
    5)     The overall performance of all the process chains for past few days
    6)     The number of requests in BW for this process chain
    7)     The load on CRM system for past few days when this process chain ran on BW system
    As per our analysis, there are couple of things which can be fixed in the BW system:--
    1)     The partner agreement (transaction WE20) defined for the partner LS/BP3CLNT475 mentions both message types RSSEND and RSINFO: -- collect IDOCs and pack size = 1 Since the pack size = 1 will generate 1 TRFC call per IDOC, it should be changed to 10 so that less number of TRFCs will be generated thus less overhead for the BW server resulting in the increase in performance
    2)     In the definition of destination for the concerned RFC in BW (SM59), the “Technical Setting” tab says the “Load balancing” option = “No”. We are planning to make it “Yes”
    But we believe that though these changes will bring some increase in performance, this is not the root cause of the abnormal behavior of this chain as this chain runs successfully on every alternate day with approximately the same amount of load in it.
    I was not able to attach the many screen shots or the info which I had gathered during my analysis. Please advice how do I attach these files
    Best Regards,

    Hi,
    Normally  index  creation or deletion can take long time in case  your database statistics are not updated properly, so can check  stat  after your data loading is completed and index generation is done,  Do creation of database statistics.
    Then try to recheck ...
    Regards,
    Satya

  • DSO activation in Process Chains - takes long time

    Dear All,
    We have included DSO activation in Process Chains. This process takes a long time to execute. We manually cancel the corresponding process in Process Overview and repeat it from chain. The chain runs daily and this issue also occurs on a daily basis.
    Does anyone one of you have an idea of how to deal with this performance issue?
    Regards.

    Figure out in which class does your process fall in:
    Class A- High Priority
    Class B- Medium Priority
    Class C- Least Priority
    Background processes
    Class A jobs
    The number of work processes reserved for job class A is a subset of the number of background processes.
    You should only reserve work processes for job class A if it makes sense within your system organization. Work processes reserved for class A jobs are no longer available for job classes B or C.
    Set parallel processing for a specific BW process in the (variant) maintenance of the process
    Call the function for setting the parallel processes.
    You can call the function in the process variant maintenance of a process chain or in the process maintenance. The function call varies for the different BW processes.
    For example, in the data transfer process you call the function with Goto-->  Background Manager Settings.
    The Settings for Parallel Processing dialog box appears
      Under Number of Processes, define the maximum number of work processes that should be used to process the BW process.
    If you enter 1, the BW process is processed serially.
    If you enter a number greater than 1, the BW process is processed in parallel
    In the Parallel Processing group frame, make the relevant settings for parallel processing in the background:
    Enter a job class for defining the job priority.
    The job priority defines how the jobs are distributed among the available background work processes
      In the group frame Parallel Processing, you can define whether parallel processing should take place in dialog work processes or in background work processes for the processes ODSACTIVAT, ODSSID and ODSREQUDEL for the DataStore object
    Transport:
    The entries in tables RSBATCHPARALLEL and RSBATCHSERVER are written on a transport request of the Change and Transport System.
    Edited by: ram.pch on Oct 7, 2011 9:55 PM

  • What could be the reason for Crawl process to take long time or get in to a hung state.

    Hi All,
    What could be the reason for Crawl process to take long time or get in to a hung state? Is it something also related to the DB Server resources crunch? Does this lead to Index file corruption?
    What should be the process to be followed when index file is corrupted? How do we come to know about that?
    Thanks in Advance.

    "The crawl time depends on what you are crawling -- the number of items, the size of the items, the location of the items. If you have a lot of content that needs to be crawled, it will take much time".
    http://social.msdn.microsoft.com/Forums/sharepoint/en-US/f4cad578-f3bc-4822-b660-47ad27ce094a/sharepoint-2007-crawl-taking-long-time-to-complete?forum=sharepointgeneralprevious
    "The only clean and recommended way to recover from an index corruption is to completely rebuild the index on all the servers in the farm."
    http://blogs.technet.com/b/victorbutuza/archive/2008/11/11/event-id-4138-an-index-corruption-was-detected-in-component-shadowmerge-in-catalog-portal-content.aspx
    Whenever search index file got corrupted it will got the details to Event logs
    http://technet.microsoft.com/en-us/library/ff468695%28v=office.14%29.aspx
    My Blog- http://www.sharepoint-journey.com|
    If a post answers your question, please click Mark As Answer on that post and Vote as Helpful

  • Process Chain taking long time in loading data in infocube

    Dear Expert,
      We are loading data thru PC in AR cube it takes data frm
    PSA-> DSO->Activation->Index Deletion->DTP(load infocube)->IndexCreation->Create Aggregates.
    In Index creation everyday its taking long time around 9 to 10 hrs to create it
    when we go in RSRV and repair the infocube thr loading of data happens fast. We are doing it(RSRV) everyday. In DB02 we have seen dat 96% tablespace is used.
    Please tell permanent solution.
    Please suggest its BI Issue or Basis.
    Regards,
    Ankit

    Hi ,
    We are loading data thru PC in AR cube it takes data frm
    PSA-> DSO->Activation->Index Deletion->DTP(load infocube)->IndexCreation->Create Aggregates.
    In the above steps insted of Create Aggregates it should be Roll up Process of aggregates.
    You can ask the basis team to check the Table space in the transaction db02old/db02.
    Check if there is long running job in SM66/SM50 kill that job.
    check there should be enough Batch process to perform the steps.
    Hope this helps.
    "Assigning points is the ways to say thanks on SDN".
    Br
    Alok

  • Process chain taking long time to complete

    Hi,
    I am having following issues in the process chain daily loads
    1) The PSA deletion is taking very long time on fridays only (nearly 3 hrs). Other days it gets deleted in 1 hr max.
    2) Just to load 381 records via dtp from DSO to cube takes nearly 2 hrs(Delta load). No major routines are written in the transformation.
    How to analyse a process in process chain which does not fail, but takes very long time to complete. Is there any tool or transaction in BW which can help us to analyse why the process chain takes so long time to complete in different days. One day it completes in 8 hrs, another day it takes 12 hrs.
    None of the transactions, SM37, SM50, SLG1, LOGS etc are giving me any help to analyse this issue.
    If it fails we have error logs to check and analyse, but without failing how we can analyse and fix the delay and reduce the time of loading data. Please guide me.
    Thanks in advance.
    Vishwanath

    Hi.......
    1) This might be due to the poor performance of the system ..there wont be enoughwork processes avilabe in teh system .- What needs to be done if it does not have enough work processes available. Though I can find that lots of dialog processes were free during these jobs as well some Background processes were also free.?
    Look to increase Work Process.......................u can do two things..............U can cancel some Job...........which r not progressing...........or which r not very important for the time being..........
    Or if possible........u can increase number of servers..........
    Look When a Job will start...............first it will run in background.............u can start in in Dialog also..........but for Dialogue after some time ..load will go to time out.................now many Child jobs will be created for a single background jobs..............they may run in Dialogue..........u can monitor them through SM66............
    2) Check ST04 for lock waits and dead locks ..if the same lock persist for long then check with basis team.
    What needs to be conveyed to the basis team when these locks happen?
    generally these will be temporary Locks.............after some time locks will be realeased..............if they persist for a long time...........then u can contact Basis people..........Or if u can understand that which job is creating the lock.............like if u double click on the job >> click on Job details...............from there u will find the PID................copy that.............and check in ST04..........whether they matches or not..............now suppose it match.............and u know that job is not very important............then u can cancel the job.............
    Check also the table space availba in St04 . the usuage should nt be more than 90%
    What needs to be conveyed to the basis team if the usage is more than 90%.?
    U can ask them to increase the Table space........
    Check the sm37 >> delay column it should nt be high
    What needs to be done if delay column is high. Actually it is high for these jobs (PSA deletion and load from DSO to Cube visa DTP (Delta load)?
    If Delay is more.............it means ...........that there is no free work process...jobs r going to Release state............solution I had already given u.......
    Check the PID refeclting in lock waits in sm66 wether they are progressing or nt ..
    If they are not progressing what action needs to be taken?
    Actually...........suppose one Dialogue job is running............it may change..........ie  that Dialogue job will go to Stop status.........some other job will start............anyways if it does'nt progress............u can cacel the JOb.........
    Check OS07 for DB if the idle time is less then 20% then its a problem ..
    If the Idle time is less than 20% what action needs to be taken?
    U can contact the Basis people.......
    Check sm21 and RFC connection with other source systems in sm59 ..
    What needs to be checked in SM21 and SM59 specifically? What parameter i need to check?
    In SM21.............if red status is there............check the Log beside that..............may be it will be some Terminal disconnected.......
    In SM59...........clck on test connection..........
    if the problem is occuring due to only one source system then check the performance of that system ..
    How to check the performance of these systems? Any tools available in R/3 system to check its performance?
    No.....in this way u can check performance for all the source system...........go to SM59............double click on the desired source system>> click on test connection......
    If All these are persisting then its a performance problem and check with basis team ..
    Is there any special settings which needs to be maintained to achieve better performance of Process chain loads?
    U can Improve the performance of process chain by Parallel Processing..........ie split the loads by giving selections...........and execute thenm in parallel..........
    Regards,
    Debjani........

  • Process Chain takes alonf time to run

    Hello All,
    We have a process chain which loads data for texts, attributes etc.
    On certain days the process chain will run extremely fast and maybe take a couple of hours, other days the same process chain may take 24 hours to run.  It appears to get stuck when extracting the data from our source system.
    Is there any reason for these strange performance issues and does any body else have the same problem?  Is there anything that can be done about it?
    Thanks,
    Nick.

    Hi Nick,
    This error normally occurs whenever BW encounters error and is not able to classify them. There could be multiple reasons for the same
    Whenever we are loading the Master Data for the first time, it creates SID’s. If system is unable to create SID’s for the records in the Data packet, we can get this error message.
    If the Indexes of the cube are not deleted, then it may happen that the system may give the caller 70 error.
    Whenever we are trying to load the Transactional data which has master data as one of the Characteristics and the value does not exist in Master Data table we get this error. System can have difficultly in creating SIDs for the Master Data and also load the transactional data.
    If ODS activation is taking place and at the same time there is another ODS activation running parallel then in that case it may happen that the system may classify the error as caller 70. As there were no processes free for that ODS Activation.
    It also occurs whenever there is a Read/Write occurring in the Active Data Table of ODS. For example if activation is happening for an ODS and at the same time the data loading is also taking place to the same ODS, then system may classify the error as caller 70.
    It is a system error which can be seen under the “Status” tab in the Job Over View.
    Cheers
    Raj

  • Infopackage in process chain taking long time to run

    HI Experts,
    One of the element (Infopackage) in process chain(Daily process chain) is taking much longer time(6 hr and still runing) as generally it takes 15-20 mins to complete and the status is in yellow and still running in process monitor without giving any clear picture of  error. Manually we are making the staus red and we are updating from PSA and this time it is getting completed in time as specified.
    Flow from PSA to datatarget(DSO) in series
    For last one week we are facing same issue and i would like to mention that we don't have access to SM37,SM12 to look the logs
    and any locks.
    without this i need to investigate what is the root cause for the same.
    with regards,
    murali

    Hi,
    please find the job log.
    Date
    Time
    Message
    MsgID/No./Ty
    13.12.2010
    21:45:30
    Job started
    00
    516
    S
    13.12.2010
    21:45:30
    Step 001 started (program SBIE0001, variant &0000000109368, user ID RFCUSER)
    00
    550
    S
    13.12.2010
    21:45:30
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    R3
    413
    S
    13.12.2010
    21:45:30
    DATASOURCE = 2LIS_17_I3HDR
    R3
    299
    S
    13.12.2010
    21:45:30
    RLOGSYS    = PBACLNT200
    R3
    299
    S
    13.12.2010
    21:45:30
    REQUNR     = REQU_D9GSSV08BTS93Z6Y7XWP1ZEUJ
    R3
    299
    S
    13.12.2010
    21:45:30
    UPDMODE    = D
    R3
    299
    S
    13.12.2010
    21:45:30
    LANGUAGES  = *
    R3
    299
    S
    13.12.2010
    21:45:30
    R8
    048
    S
    13.12.2010
    21:45:30
             Current Values for Selected Profile Parameters               *
    R8
    049
    S
    13.12.2010
    21:45:30
    R8
    048
    S
    13.12.2010
    21:45:30
    abap/heap_area_nondia......... 0                                       *
    R8
    050
    S
    13.12.2010
    21:45:30
    abap/heap_area_total.......... 25500319744                             *
    R8
    050
    S
    13.12.2010
    21:45:30
    abap/heaplimit................ 40000000                                *
    R8
    050
    S
    13.12.2010
    21:45:30
    zcsa/installed_languages...... ED                                      *
    R8
    050
    S
    13.12.2010
    21:45:30
    zcsa/system_language.......... E                                       *
    R8
    050
    S
    13.12.2010
    21:45:30
    ztta/max_memreq_MB............ 2047                                    *
    R8
    050
    S
    13.12.2010
    21:45:30
    ztta/roll_area................ 3000320                                 *
    R8
    050
    S
    13.12.2010
    21:45:30
    ztta/roll_extension........... 2000000000                              *
    R8
    050
    S
    13.12.2010
    21:45:30
    R8
    048
    S
    13.12.2010
    21:45:31
    70 LUWs confirmed and 70 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    RSQU
    036
    S
    13.12.2010
    21:45:33
    Call customer enhancement BW_BTE_CALL_BW204010_E (BTE) with 9,895 records
    R3
    407
    S
    13.12.2010
    21:45:33
    Result of customer enhancement: 9,895 records
    R3
    408
    S
    13.12.2010
    21:45:33
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 9,895 records
    R3
    407
    S
    13.12.2010
    21:45:33
    Result of customer enhancement: 9,895 records
    R3
    408
    S
    13.12.2010
    21:45:33
    PSA=0 USING & STARTING SAPI SCHEDULER
    R3
    299
    S
    13.12.2010
    21:45:33
    Asynchronous send of data package 1 in task 0002 (1 parallel tasks)
    R3
    409
    S
    13.12.2010
    21:45:34
    IDOC: Info IDoc 2, IDoc No. 5136702, Duration 00:00:00
    R3
    088
    S
    13.12.2010
    21:45:34
    IDoc: Start = 13.12.2010 21:45:30, End = 13.12.2010 21:45:30
    R3
    089
    S
    13.12.2010
    21:45:35
    Asynchronous transmission of info IDoc 3 in task 0003 (1 parallel tasks)
    R3
    413
    S
    13.12.2010
    21:45:35
    Altogether, 0 records were filtered out through selection conditions
    RSQU
    037
    S
    13.12.2010
    21:45:35
    IDOC: Info IDoc 3, IDoc No. 5136703, Duration 00:00:00
    R3
    088
    S
    13.12.2010
    21:45:35
    IDoc: Start = 13.12.2010 21:45:35, End = 13.12.2010 21:45:35
    R3
    089
    S
    13.12.2010
    21:55:37
    tRFC: Data Package = 1, TID = 0AF0842B00C84D0694000F00, Duration = 00:10:03, ARFCSTATE = SYSFAIL
    R3
    038
    S
    13.12.2010
    21:55:37
    tRFC: Start = 13.12.2010 21:45:34, End = 13.12.2010 21:55:37
    R3
    039
    S
    13.12.2010
    21:55:37
    Synchronized transmission of info IDoc 4 (0 parallel tasks)
    R3
    414
    S
    13.12.2010
    21:55:37
    IDOC: Info IDoc 4, IDoc No. 5136717, Duration 00:00:00
    R3
    088
    S
    13.12.2010
    21:55:37
    IDoc: Start = 13.12.2010 21:55:37, End = 13.12.2010 21:55:37
    R3
    089
    S
    13.12.2010
    21:55:37
    Job finished
    00
    517
    S

  • Issue with Delete Index in process chain

    Hi All,
    I have a situation with Delete Index in my process chain.
    Delete Index step, before loading data to cube, is taking around 2Hr to complete but some times it completes within 5 mins.
    Anyone faces such issue? If yes, how can I fix it?
    Thanks and Regards,
    Alok

    Hi,
    If Statistics are not maintained for the cube then also it might casue the high run time in index deletion.
    if its urgent then, can u try doing it manually form performance tab....
    or once process chain step is complete, create statistics for the cube. Next run should take less time
    Please try.
    Regards,
    Gaurav

  • Process chain taking long time

    Hi,
        I developed a Process chain. In developemnt for Completion of the Update rules(Start routine) its taking 60m( for around 50,000 records).
    After transporting the Process chain to Production for Start routine completion (For 1,70,000 records) its taking 7hrs and sometimes its failing as the "Process is Over due" and sometimes due to "Short dump".
    How to handle this.........
    Thanks

    Hi js,
    Check the ST22... Any dumps
    Check out for the short dump in source system and try loading the delta again
    then
    I have got the problem when u have posted...but little later.Even I was about to post the issue.. Actually my issue was...
    My PC is like this
    1) Start Process
    2) Delete index
    3) Load IP
    4) delete overlapping request
    5) create Index
    6) ABAP program
    Process chain hangs at Overlapping requests process type but the status was yellow. And the previous step (3) was success. From there its not moving to next further steps.. But the Job in SM37 was canceled.
    I saw Ur post and saw the note 833168 and which matches exactly our issue.
    Like we run the PCs thru the TIDAL where due to some problems PC got triggered twice... I can see the two logs in the PC. (Log view) and two requests at RSMO.
    Everything was recovered again thru manual steps. This is the first time we encountered... the note helped us ... We now are very cautious about the Tidal runs...
    Now PC is running perfectly
    we have instructed the Supported team to make the necessary changes or fix this issues in TIDAL...
    i hope this is helpful to u
    bye now
    suresh

  • Process chain running long time

    Hi all..
    I have one process chain.that chain haveing very small data.when iam created time it runs nearly 20 to 30 mints.
    Now the same chain running nearly 3 hours..
    Why its taking very long time?How to reduce the chain running time.Any steps to reduce the chain running time..
    Thanks
    Ramya

    Hi,
    Check the load in the chain which is taking long time.
    Check out how much data is that load uploading.
    Regards,
    Rohini

  • Info Package in Process Chain taking long time

    Dear All,
    Info Package in Process Chain having 0 records is taking 1 hr 5 mins to complete. Time out time is set to 10 mins.
    Does any one of you know the reason why it is happening like this?
    Regards.

    When the info package status turns to green : final message is :
    No data available
    Diagnosis
    The data request was a full update.
    In this case, the corresponding table in the source system does not
    contain any data.
    System Response
    Info IDoc received with status 8.
    Procedure
    Check the data basis in the source system.
    Diagnosis
         The data request was a full update.
         o   In this case, the corresponding tab
                 contain any data.
         o   System Response
                 Info IDoc received with status
                 Procedure
    Check the data basis in the source system.
    Analysis
    ========
    No data available in the source system for this selection
    No further action necessary
    When the status was red, I checked BD87 and SM58 and there were no entries.
    The corresponding job in sm37 took 869 seconds to complete. Below was the job Log.
    Job started
    DATASOURCE = 0CO_PC_02
    Call customer enhancement BW_BTE_CALL_BW204010_E (BTE) with 0 records
    Result of customer enhancement: 0 records
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 0 records
    Result of customer enhancement: 0 records
    IDOC: Info IDoc 2, IDoc No. 1345319, Duration 00:00:00
    IDoc: Start = 24.10.2011 18:29:54, End = 24.10.2011 18:29:54
    Synchronized transmission of info IDoc 3 (0 parallel tasks)
    IDOC: Info IDoc 3, IDoc No. 1345330, Duration 00:00:00
    IDoc: Start = 24.10.2011 18:44:18, End = 24.10.2011 18:44:18
    Job finished.
    PLease help.
    Regards.

  • Master data Process chain running long time

    Dear SDN Team,
    PC Process Chain: General Master Data - time running is about 4.5 hours . how we can improve running time.
    There are 15-25 info objects are loading from this process chain.
    any steps to improve the performace
    thanks and kind regards,
    Lakshman Kumar G

    Hi,
    Go to the function module RSD_IOBJ_GET to find the object name of the dimension. Go to SE37 & proceed as follows.
      I_IOBJNM = 'Info Object Name'
      I_OBJVERS = 'A'
      I_BYPASS_BUFFER = 'X'
    and then Execute the FM.
    Double click on the "'E_S_VIOBJ'" and find the number in the field 'u201ENUMBRANR‟'
    and add  BIM to the number taken  from NUMBRANR.
    Goto SNRO T-Code and enter BIM+the number and edit.
    From Edit--> Set up Buffering --> Main memory, select the buffering check box and enter approx 500.
    This needs to be done for all the master and your dataload performance would improve.
    Regards,
    Mani

  • Query Prediction takes long time - After upgrade DB 9i to 10g

    Hi all, Thanks for all your help.
    we've got an issue in Discoverer, we are using Discoverer10g (10.1.2.2) with APPS and recently we upgraded Oracle DatBase from 9i to 10g.
    After Database upgrade, when we try to run reports in Discoverer plus taking long time for query prediction than used to be(double/triple), only for query prediction taking long time andthen takes for running query.
    Have anyone got this kind of issues seen before, could you share your ideas/thoughts that way i can ask DBA or sysadmin to change any settings at Discoverer server side
    Thanks in advance
    skat

    Hi skat
    Did you also upgrade your Discoverer from 9i to 10g or did you always have 10g?
    If you weren't always on 10g, take a look inside the EUL5_QPP_STATS table by running SELECT COUNT(*) FROM EUL5_QPP_STATS on both the old and new systems
    I suspect you may well find that there are far more records in the old system than the new one. What this table stores is the statistics for the queries that have been run before. Using those statistics is how Discoverer can estimate how long queries will take to run. If you have few statistics then for some time Discoverer will not know how long previous queries will take. Also, the statistics table used by 9i is incompatible with the one used by 10g so you can't just copy them over, just in case you were thinking about it.
    Personally, unless you absolutely rely on it, I would turn the query predictor off. You do this by editing your PREF.TXT (located on the middle tier server at $ORACLE_HOME\Discoverer|util) and change the value of QPPEnable to 0. AFter you have done this you need to run the Applypreferences script located in the same folder and then stop and start your Discoverer service. From that point on queries will no longer try to predict how long they will take and they will just start running.
    There is something else to check. Please run a query and look at the SQL. Do you by change see a database hint called NOREWRITE? If you do then this will also cause poor performance. Should you see this let me know and I will let you know how to override it.
    If you have always been on 10g and you have only upgraded your database it could be that you have not generated your database statistics for the tables that Discoverer is using. You will need to speak with your DBA to see about having the statistics generated. Without statistics, the query predictor will be very, very slow.
    Best wishes
    Michael

Maybe you are looking for

  • HT201210 What is error 28 (unable to restore ipad)?

    Ipad needs to be restored, but won't complete the restore.  Error message contains code 28 in parenthesis.  Any clues about what I could do?

  • Mail Dock icon "new mail" count wrong

    The Mail icon in my dock shows 11 new mails. I open Mail and there is no new mail. The icon then shows none, then a bit later it shows 11 again. If I get a real new mail it will show 12 new mails. The actual number of new mail adds to the phantom 11.

  • "The computer started using Windows installation media." Installing windows 10 technical preview on virtual machine

    The computer started using Windows installation media. Remove the installation media and restart your computer so that Windows starts normally. The, insert the installation media and restart the upgrade. (Do not select |custom[advanced]| to perform a

  • Free Ftp clients

    Hello, New to iWeb and uploaded my first pages yesterday with good results. However, I have two questions. 1) Seeing the page up and running there were a few things to be changed. Uploading the whole folder again with changes didn´t change a thing on

  • Report performace

    Hi, Which one will provide better performance when fetching data from multiple infotypes. 1) multiple  'HR_READ_INFOTYPE' 2) RP_PROVIDE_FROM_LAST or RP_PROVIDE_FROM_FIRST. 3) PROVIDE....ENDPROVIDE with HR_READ_INFOTYPE inside it any other method...pl