Job running for ever...

Hi Gurus,
I created a generic extractor and when execute it using RSA3 transaction in SRM, is OK.
When i execute the function module is also OK.
When i start a job (infopackage) using this extractor, the records are load in BI but the job still running...
I saw in SM5O that the work process is reading a ROIS table...
Why it it taking so long...it has only 20 records.
Appreciate your response.
Thanks,
rCm.

Hi,
When a BI load takes long time to finish.It may be hanged in tRFC.
Do the following steps.
1.Go to the long running load monitor (T-code: RSMO) screen and from menu at top navigate to u2018Environmentu2018 -> u2018Transact. RFCu2018 and then,
-> To u2018In the Data Warehouseu2018 , if the data source is in same SAP system as of the current load
-> To u2018In the source systemu2018 , if the data source exists in external SAP system (like R/3).
2.Now we will be in u2018Transactional RFCu2018 list selection screen. Here enter the input parameters. Put wild character asterisk u2018*u2019 in you need to get entire list of tRFCu2019s in the system.
3.Now the screen will list all the tRFCs currently scheduled to execute or held up. If a tRFC is held for long time, then we can manually push it by clicking on it from the list and then from menu go to u2018Editu2018 -> u2018Execute LUWu2018 or pressing F6.
Regards,
Shiva.

Similar Messages

  • Dbms_scheduler job neither succeeds nor errors its just keep running for ever

    Hi All,
    I am trying to run a shell script from plsql using the dbms_scheduler, job is getting created and it keeps running for ever without success or error.
    Can some please help me where I am doing the mistake.
    Thanks
    DB Version:
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bi
    PL/SQL Release 10.2.0.5.0 - Production
    Plsql Script:
    BEGIN
       DBMS_SCHEDULER.create_job (
          job_name              => 'SFTP_PAYMENTECH_BATCHID1',
          job_type              => 'EXECUTABLE',
          job_action            => '/app07/ebdev/ebdevappl/xxtpc/bin/test.sh',
          number_of_arguments   => 1,
          enabled               => FALSE,
          auto_drop             => TRUE,
          comments              => 'SFTP Batch File to Paymentech');
       DBMS_OUTPUT.put_line (
          'Job Created Successfully:' || 'SFTP PAYMENTECH BATCHID');
       fnd_file.put_line (
          fnd_file.output,
          'Job Created Successfully:' || 'SFTP PAYMENTECH BATCHID');
       COMMIT;
       DBMS_SCHEDULER.SET_JOB_ARGUMENT_VALUE (
          job_name            => 'SFTP_PAYMENTECH_BATCHID1',
          argument_position   => 1,
          argument_value      => '66667'                        /*v_printer_name*/
       DBMS_SCHEDULER.enable (name => 'SFTP_PAYMENTECH_BATCHID1');
    EXCEPTION
       WHEN OTHERS
       THEN
          fnd_file.put_line (fnd_file.output,
                             'Error while creating the Job:' || SQLERRM);
          DBMS_OUTPUT.put_line ('Error while creating the Job:' || SQLERRM);
    END;
    Shell Script which I am calling:
    #!/usr/bin/ksh
    FILENAME=$1
    PMTHOST=198.xx.xx.xx
    PMTUSER=xxxxx
    PMTPW=xxxxx
    a='apps'
    b='xxxxxx'
    c='EBDEV'
    INST=`echo $TWO_TASK | sed 's/_BALANCE//'`
    echo INSTANCE: $INST
    echo
    File_Path=$XXTPC_TOP/iby/out
    echo File Name: $FILENAME
    echo $PMTHOST
    echo $PMTUSER
    echo
    echo Date: `date`
    echo
    echo File System User: `whoami`
    echo
    echo Instance: $TWO_TASK
    echo
    echo File_Path: $File_Path
    echo
    echo PMT SFTP
    # Fetch file using MBATCHID as File Name
    cd $File_Path
    echo
    echo -----------------------------------------------
    echo
    echo Current File          :$FILENAME
    l_date_time=`date +%Y%m%d%H%M%S`
    echo SFTP Remittance File
    # sftp $PMTUSER@$PMTHOST << EOF
    lftp -u $PMTUSER,$PMTPW sftp://$PMTHOST << EOF
    lcd $File_Path
    cd test/945299
    put $FILENAME
    exit
    EOF
    #`sqlplus -s apps/tpcdev2013@EBDEV @try.sql $FILENAME`

    Have you tried running the script manually to confirm it isnt just hanging on input for the lftp ?
    You could add a -o <file> to the lftp command line which would output debug info into <file> so you could see.
    Creation of the job looks fine so as GregV said, what output do you have in ALL_SCHEDULER_RUNNING_JOBS and any ADDITIONAL_INFO in the ALL_SCHEDULER_JOB_RUN_DETAILS or in $ORACLE_HOME/scheduler/log that might indicate what the problem is ?
    If this answer was helpful, please mark as helpful answer. 

  • Background job running for a long time - trace is not capturing.

    hi guys,
    My background job runs for more thatn 4 hours.  some times. (normally it completes withing one hour)
    i tried to trace using st12 ....and executed this program. but not got complete trace report of 4 hours.
    what is the best way to trace long running programs like this.
    is there any alternative for this...???
    regards
    girish

    Giri,
    There is no need to trace a program for the full 4 hours. There is usually just one piece of rogue code causing the problem, and it can be identified quite quickly.
    If I were you I would just watch the job run via SM50, to see if it lingers on any database tables. Also check if the memory is filling up - this is the sign of a memory leak.
    You can also try stopping the program (via SM50 debugging) at random points, to see what piece of code it is stuck in.
    The issue should reveal itself fairly quickly.
    Tell us what you find!
    cheers
    Paul

  • Sql is running for ever

    We have one report which will process every night which is running good from couple of years. Today that report is running for ever. I extrated sql from that report and if I run that sql by giving * after select, it is giving result with in one minute. If I run the sql by selecting column names, it is running for ever. Our daily report will run by selecting few columns from three tables. We are using choose based optimizer.
    But when I run the same query by selecting the columns, with yesterday, I will get the output with in one minute. If I run the same query for today's date, getting that problem. But we had almost more than 10 thousand transactions over night. Not sure whether I need to run compute statistics on those tables which can fix the problem.
    Pls advice how to fix this. Our db is 9i.
    Thanks in advance.
    Hari Babu
    Edited by: user6367891 on Oct 8, 2009 12:55 PM

    Pl also see HOW TO: Post a SQL statement tuning request - template posting
    The information you have posted so far is insufficient to make any kind of diagnosis.
    HTH
    Srini

  • Query runs for ever - suggestions - ideas - tips

    Hello all,
    I have the following situation:
    1) Database 1: XE 10g Release 10.2.0.1.0 - Production
    2) Database 2: 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    3) Dblink from Database 1 pointing to Database 2
    4) Synonyms created withing Database 1, pointing to Database 2 packages and tables:
    - "mytable1" exists in Database 2 therefore a synonym "mytable1" is created in Database 1
    - "mytable2" exists in Database 2 therefore a synonym "mytable2" is created in Database 1
    - "mypackage" exists in Database 2 therefore a synonym "mypackage" is created in Database 1
    5) Of course select and execute rights are granted
    6) Query:
    SELECT field1
    FROM myshcema.mytable1
    WHERE field2 = myshcema.mypackage.myfunction ('PARAMVALUE'))
    AND EXISTS (SELECT 1 FROM myshcema.mytable2 WHERE field_in_tab2 = field_in_tab1)
    7) Problem.This query runs for ever
    8) Explain plan
    OPERATION - OBJECT_NAME - OPTIONS - COST
    SELECT STATEMENT - - - 17
    -> FILTER - - -
    -> -> FILTER - - - 3
    -> -> -> REMOTE - mytable1 - -
    -> -> REMOTE - mytable2 - - 1
    9) Now I get the value returned by the function and re-execute the previous query:
    - select myshcema.mypackage.myfunction ('PARAMVALUE')) from dual;
    - return value is *89787621*
    - re-executing the query:
    SELECT field1
    FROM myshcema.mytable1
    WHERE field2 = *89787621*
    AND EXISTS (SELECT 1 FROM myshcema.mytable2 WHERE field_in_tab2 = field_in_tab1)
    10) Explain plan
    OPERATION - OBJECT_NAME - OPTIONS - COST
    SELECT STATEMENT - - REMOTE - 3
    -> NESTED LOOPS - - SEMI - 3
    -> -> INDEX - MY_UK_NAME - RANGE SCAN - 3
    -> -> INDEX - MY_2ND_UK - UNIQUE SCAN - 0
    And the results are instantaneous
    Now my question: Can somebody direct me on whether a hint or something else will correct this situation? Due to restrictions that I cannot currently explain here, I am not able to change the code to first get the value and then execute the query.
    Thanks for helping
    G
    Edited by: G on May 17, 2011 9:07 AM
    Edited by: G on May 17, 2011 9:10 AM

    I would think the driving site hint woud be the frist thing I tried. If for some reason Oracle does not seem to be able to use the hint or if hints are heavily frowned on at your site then here is another approach to the problem.
    Code the query into a view and define the view on the remote database then issues a local query against the remote view.
    HTH -- Mark D Powell --

  • GRC AC 5.3 - Sync Mode background job running for days.

    We have the following performance issues on the GRC System.
    1- Please note that the sync mode background job has been running for DAYS.
    My question is if I  truncate table virsa_cc_prmvl;  (which stores all the analysis results), will my GRC system be effected.
    2- Secondly we have a job analysing roles, which causes the performance to go down. And due to we have a extremely high number of roles, we would like to know if there would be a way to improve the analysis for roles. Is there any parameter modification in GRC / J2EE/ backend related to roles?
    Details of our roles:
    - we have more than 60000 roles in our system
    - 75 % of the roles are not assigned to any user (is there a way to avoid these roles? these roles are not put in order so too tedious to choose roles individually or by range.
    3- We are running a job with four risks and it's taking more than two minutes per user (we've got 10.000 + users in total)
    How can I resolve this performance issue.
    4- What is the recommended size of the batch for user sincronization. Should I use 1000 or 10,000 (based on our above environment).
    Thanks,
    Haleem

    Hello, Abdul!
    How you have solved a problem?
    Thank you a lot.
    Best regards,
    Denis Gaykov

  • Job runs for a much longer time

    Hi Guys,
    I face a situation wherein a SM37 job keeps on running for a much longer time ~1.5 lacs seconds.Can you suggest the reason of this.
    Regards,
    Rajiv

    Lots of reasons - please as a more specific question.

  • Background Jobs running for long time.

    Dear All,
    I had done the Homogeneous System Copy on my ECC 6 QA System successfully by restoring the offline backup of Production system. My system is up & running but  I have some Performace issues in the system. I also observed that some jobs are running since long time & most commonly i got the below line in the BGD Trace "WP has reached abap/heaplimit = 40000000 bytes".
    I search for the notes but couldn't find anything.
    Please help me to resolve this issue.
    Below is the Trace :-
    trc file: "dev_w7", trc level: 1, release: "700"
    ACTIVE TRACE LEVEL           1
    ACTIVE TRACE COMPONENTS      all, MJ
    M sysno      12
    M sid        UBQ
    M systemid   370 (Solaris on SPARCV9 CPU)
    M relno      7000
    M patchlevel 0
    M patchno    52
    M intno      20050900
    M make:      single threaded, Unicode, 64 bit, optimized
    M pid        16335
    M
    M
    M Mon Nov 10 19:01:10 2008
    M  kernel runs with dp version 210000(ext=109000) (@(#) DPLIB-INT-VERSION-210000-UC)
    M  length of sys_adm_ext is 572 bytes
    M  ThInit: running on host UBSQAS
    M  calling db_connect ...
    C  Client NLS settings: AMERICAN_AMERICA.UTF8
    C  Logon as OPS$-user to get SAPSR3's password
    C  Connecting as /@UBQ on connection 0 (nls_hdl 0) ... (dbsl 700 240106)
    C  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    C    0 UTF8                                                      1 0x1067f1be0 0x1067f96a0 0x106809e38
    C  Attaching to DB Server UBQ (con_hdl=0,svchp=0x106809d68,svrhp=0x10680c0b8)
    C  Starting user session (con_hdl=0,svchp=0x106809d68,srvhp=0x10680c0b8,usrhp=0x1067f9eb8)
    C  Now '/@UBQ' is connected (con_hdl 0, nls_hdl 0).
    C  Got SAPSR3's password from OPS$-user
    C  Disconnecting from connection 0 ...
    C  Closing user session (con_hdl=0,svchp=0x106809d68,usrhp=0x1067f9eb8)
    C  Now I'm disconnected from ORACLE
    C  Connecting as SAPSR3/<pwd>@UBQ on connection 0 (nls_hdl 0) ... (dbsl 700 240106)
    C  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    C    0 UTF8                                                      1 0x1067f1be0 0x1067f96a0 0x106809e38
    C  Starting user session (con_hdl=0,svchp=0x106809d68,srvhp=0x10680c0b8,usrhp=0x1067f9eb8)
    C  Now 'SAPSR3/<pwd>@UBQ' is connected (con_hdl 0, nls_hdl 0).
    C  Database NLS settings: AMERICAN_AMERICA.UTF8
    C  Database instance UBQ is running on UBSQAS with ORACLE version 10.2.0.2.0 since 20081109
    B  Connection 0 opened (DBSL handle 0)
    B  Wp  Hdl ConName          ConId     ConState     TX  PRM RCT TIM MAX OPT Date     Time   DBHost
    B  000 000 R/3              000000000 ACTIVE       NO  YES NO  000 255 255 20081110 190110 UBSQAS
    M  db_connect o.k.
    M  ICT: exclude compression: .zip,.cs,.rar,.arj,.z,.gz,.tar,.lzh,.cab,.hqx,.ace,.jar,.ear,.war,.css,.pdf,.js,.gzip
    I  MtxInit: 7 0 0
    M  SHM_PRES_BUF###(addr: 0xffffffff6c400000, size: 4400000)
    M  SHM_ROLL_AREA##(addr: 0xfffffffe56000000, size: 134217728)
    M  SHM_PAGING_AREA##(addr: 0xfffffffe50000000, size: 67108864)
    M  SHM_ROLL_ADM###(addr: 0xffffffff7000a000, size: 2672386)
    M  SHM_PAGING_ADM##(addr: 0xffffffff77a00000, size: 656416)
    M  ThCreateNoBuffer##allocated 572152 bytes for 1000 entries at 0xfffffffe48004000
    M  ThCreateNoBuffer##index size: 3000 elems
    M  ThCreateVBAdm##allocated 12176 bytes (50 server) at 0xffffffff77900000
    X  EmInit: MmSetImplementation( 2 ).
    X  MM diagnostic options set: 0
    X  <ES> client 7 initializing ....
    X  Using implementation std
    X  ES initialized.
    B  db_con_shm_ini:  WP_ID = 7, WP_CNT = 12, CON_ID = -1
    B  dbtbxbuf: Buffer TABL  (addr: 0xfffffffe42000100, size: 30000000, end: 0xfffffffe43c9c480)
    B  dbtbxbuf: Buffer TABLP (addr: 0xffffffff6b800100, size: 10240000, end: 0xffffffff6c1c4100)
    B  dbexpbuf: Buffer EIBUF (addr: 0xffffffff65800108, size: 4194304, end: 0xffffffff65c00108)
    B  dbexpbuf: Buffer ESM   (addr: 0xffffffff65000108, size: 4194304, end: 0xffffffff65400108)
    B  dbexpbuf: Buffer CUA   (addr: 0xfffffffe4b01e108, size: 3072000, end: 0xfffffffe4b30c108)
    B  dbexpbuf: Buffer OTR   (addr: 0xffffffff64800108, size: 4194304, end: 0xffffffff64c00108)
    M  rdisp/reinitialize_code_page -> 0
    M  icm/accept_remote_trace_level -> 0
    M  rdisp/no_hooks_for_sqlbreak -> 0
    M  CCMS: AlInitGlobals : alert/use_sema_lock = TRUE.
    S  *** init spool environment
    S  initialize debug system
    T  Stack direction is downwards.
    T  debug control: prepare exclude for printer trace
    T  new memory block 0x106ab5bb0
    S  spool kernel/ddic check: Ok
    S  using table TSP02FX for frontend printing
    S  1 spool work process(es) found
    S  frontend print via spool service enabled
    S  printer list size is 150
    S  printer type list size is 50
    S  queue size (profile)   = 300
    S  hostspool list size = 3000
    S  option list size is 30
    S      found processing queue enabled
    S  found spool memory service RSPO-RCLOCKS at 0xfffffffe4b30e070
    S  doing lock recovery
    S  setting server cache root
    S  found spool memory service RSPO-SERVERCACHE at 0xfffffffe4b30e548
    S    using messages for server info
    S  size of spec char cache entry: 297032 bytes (timeout 100 sec)
    S  size of open spool request entry: 2168 bytes
    S  immediate print option for implicitely closed spool requests is disabled
    A Mon Nov 10 19:01:11 2008
    A
    A     PXA
    A  PXA INITIALIZATION
    A  PXA: Fragment Size too small: 73 MB, reducing # of fragments
    A  System page size: 8kb, total admin_size: 11048kb, dir_size: 10992kb.
    A  Attached to PXA (address 0xfffffffe20000000, size 300000K)
    A  abap/pxa = shared protect gen_remote
    A  PXA INITIALIZATION FINISHED
    A     PXA
    A
    A  ABAP ShmAdm attached (addr=0xfffffffe60590000 leng=20938752 end=0xfffffffe61988000)
    A  >> Shm MMADM area (addr=0xfffffffe6083edd0 leng=140480 end=0xfffffffe60861290)
    A  >> Shm MMDAT area (addr=0xfffffffe60862000 leng=17981440 end=0xfffffffe61988000)
    A  RFC rfc/signon_error_log = -1
    A  RFC rfc/dump_connection_info = 0
    A  RFC rfc/dump_client_info = 0
    A  RFC rfc/cp_convert/ignore_error = 1
    A  RFC rfc/cp_convert/conversion_char = 23
    A  RFC rfc/wan_compress/threshold = 251
    A  RFC rfc/recorder_pcs not set, use defaule value: 2
    A  RFC rfc/delta_trc_level not set, use default value: 0
    A  RFC rfc/no_uuid_check not set, use default value: 0
    A  RFC rfc/bc_ignore_thcmaccp_retcode not set, use default value: 0
    A  RFC Method> initialize RemObjDriver for ABAP Objects
    M  ThrCreateShObjects##allocated 12676 bytes at 0xffffffff77700000
    N  SsfSapSecin: putenv(SECUDIR=/usr/sap/UBQ/DVEBMGS12/sec): ok
    N
    N  =================================================
    N  === SSF INITIALIZATION:
    N  ===...SSF Security Toolkit name SAPSECULIB .
    N  ===...SSF trace level is 0 .
    N  ===...SSF library is /usr/sap/UBQ/DVEBMGS12/exe/libsapsecu.so .
    N  ===...SSF hash algorithm is SHA1 .
    N  ===...SSF symmetric encryption algorithm is DES-CBC .
    N  ===...sucessfully completed.
    N  =================================================
    N  =================================================
    N  MskiInitLogonTicketCacheHandle: Logon Ticket cache pointer retrieved from shared memory.
    N  MskiInitLogonTicketCacheHandle: Workprocess runs with Logon Ticket cache.
    M  JrfcVmcRegisterNativesDriver o.k.
    W  =================================================
    W  === ipl_Init() called
    W    ipl_Init(): wp task type is 4:BTC, skip initialization
    W  === ipl_Init() returns 0, ITSPE_OK: OK
    W  =================================================
    E  Replication is disabled
    E  EnqCcInitialize: local lock table initialization o.k.
    E  EnqId_SuppressIpc: local EnqId initialization o.k.
    E  EnqCcInitialize: local enqueue client init o.k.
    B
    B Mon Nov 10 19:30:28 2008
    B  dbtran INFO (init_connection '<DEFAULT>' [ORACLE:700.08]):
    B   max_blocking_factor =  15,  max_in_blocking_factor      =   5,
    B   min_blocking_factor =  10,  min_in_blocking_factor      =   5,
    B   prefer_union_all    =   0,  prefer_join                 =   0,
    B   prefer_fix_blocking =   0,  prefer_in_itab_opt          =   1,
    B   convert AVG         =   0,  alias table FUPD            =   0,
    B   escape_as_literal   =   1,  opt GE LE to BETWEEN        =   0,
    B   select *            =0x0f,  character encoding          = STD / <none>:-,
    B   use_hints           = abap->1, dbif->0x1, upto->2147483647, rule_in->0,
    B                         rule_fae->0, concat_fae->0, concat_fae_or->0
    M  SecAudit(RsauShmInit): WP attached to existing shared memory.
    M  SecAudit(RsauShmInit): addr of SCSA........... = 0xffffffff78500000
    M  SecAudit(RsauShmInit): addr of RSAUSHM........ = 0xffffffff78500768
    M  SecAudit(RsauShmInit): addr of RSAUSLOTINFO... = 0xffffffff785007a0
    M  SecAudit(RsauShmInit): addr of RSAUSLOTS...... = 0xffffffff785007ac
    A
    A Mon Nov 10 20:04:25 2008
    A  WP has reached abap/heaplimit = 40000000 bytes
    A
    A Tue Nov 11 02:25:41 2008
    A  generated report /1BCWDY/BSAEKROD3IX58HSBOTLJ==CP         with fingerprint 1078C98BF377F2618A7F4843DE257FC8   50556
    Tue Nov 11 02:36:40 2008
    actual OCI-type (1) of OUT-variable 0 differs from cached statement (96) ==> reparsing
    sc_p=0x10673a550,no=176,idc_p=0x106750c68,con=0,act=1,slen=149,smax=768,#vars=5,stmt=0x107ba9dc0,table=TADIR
    prep=1,lit=0,nsql=0,lobret=0,#exec=455,dbcnt=0,upsh_p=(nil),ocistmth_p=0x1073f3ec0
    IN : cols=5,rmax=1,xcnt=0,rpc=0,rowi=0,rtot=0,upto=4294967295,rsize=108,vmax=32,bound=1,iobuf_p=0x1068680e0,vda_p=0x1073f8960
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    OUT: cols=1,rmax=0,xcnt=0,rpc=0,rowi=0,rtot=0,upto=1,rsize=0,vmax=32,bound=1,iobuf_p=(nil),vda_p=0x1073f90f0
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    SELECT /*+ FIRST_ROWS  */ "OBJ_NAME" FROM "TADIR" WHERE "PGMID" = :A0 AND ( "OBJECT" = :A1 OR "OBJEC\
    T" = :A2 ) AND "OBJ_NAME" = :A3 AND ROWNUM <= :A4;
    actual OCI-type (96) of OUT-variable 0 differs from cached statement (1) ==> reparsing
    sc_p=0x10673a550,no=176,idc_p=0x106750c68,con=0,act=1,slen=149,smax=768,#vars=5,stmt=0x107ba9dc0,table=TADIR
    prep=1,lit=0,nsql=0,lobret=0,#exec=2,dbcnt=0,upsh_p=(nil),ocistmth_p=0x1073f3ec0
    IN : cols=5,rmax=1,xcnt=0,rpc=0,rowi=0,rtot=0,upto=4294967295,rsize=108,vmax=32,bound=1,iobuf_p=0x1068680e0,vda_p=0x1073f8960
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    OUT: cols=1,rmax=0,xcnt=0,rpc=0,rowi=0,rtot=0,upto=1,rsize=0,vmax=32,bound=1,iobuf_p=(nil),vda_p=0x1073f90f0
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    SELECT /*+ FIRST_ROWS  */ "OBJ_NAME" FROM "TADIR" WHERE "PGMID" = :A0 AND ( "OBJECT" = :A1 OR "OBJEC\
    T" = :A2 ) AND "OBJ_NAME" = :A3 AND ROWNUM <= :A4;
    Tue Nov 11 02:36:42 2008
    actual OCI-type (1) of OUT-variable 0 differs from cached statement (96) ==> reparsing
    sc_p=0x10673a550,no=176,idc_p=0x106750c68,con=0,act=1,slen=149,smax=768,#vars=5,stmt=0x107ba9dc0,table=TADIR
    prep=1,lit=0,nsql=0,lobret=0,#exec=3,dbcnt=0,upsh_p=(nil),ocistmth_p=0x1073f3ec0
    IN : cols=5,rmax=1,xcnt=0,rpc=0,rowi=0,rtot=0,upto=4294967295,rsize=108,vmax=32,bound=1,iobuf_p=0x1068680e0,vda_p=0x1073f8960
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    OUT: cols=1,rmax=0,xcnt=0,rpc=0,rowi=0,rtot=0,upto=1,rsize=0,vmax=32,bound=1,iobuf_p=(nil),vda_p=0x1073f90f0
          lobs=0,lmax=0,lpcnt=0,larr=(nil),lcurr_p=(nil),rret=0
    SELECT /*+ FIRST_ROWS  */ "OBJ_NAME" FROM "TADIR" WHERE "PGMID" = :A0 AND ( "OBJECT" = :A1 OR "OBJEC\
    T" = :A2 ) AND "OBJ_NAME" = :A3 AND ROWNUM <= :A4;
    Thanks,
    Shambo

    Hi Shambo
    have you checked the oracle parameters as per the sap note 830576,and also the backup which you have taken is a logical backup,in that case you have to run the update database statistics either from DB13 or BRTOOLS
    also check ST04 and also ST02 to do a  performance tuning
    I believe its more Database Parameter issue
    regards
    George Varghese

  • What are the areas impact query to run for ever/long?

    Hi All,
    When I have to talk about long running scripts or procedures I focus on DTA which suggests on Indexing and infact I think this is the main cause. What are the other areas which we need to consider for Long running queries and troubleshooting tools we need
    to use do you think?
    Ofcourse blocking impacts but am purely thinking at query perspective.
    Thanks
    Swapna

    What are the other areas which we need to consider for Long running queries and troubleshooting tools we need to use do you think?
    One of the reasons is blocking. Please take a look at these links:
    INF: Understanding and resolving SQL Server blocking problems
    Troubleshoot Slow-Running Queries In SQL Server
    T-SQL Articles
    T-SQL e-book by TechNet Wiki Community
    T-SQL blog

  • LMS 3.2 discovery process runs for ever

    Dear friends,
    I am having a problem with discovery on Ciscoworks LMS 3.2. The discovery process runs eternally and does not discover any devices.
    Here are the individual device versions:
    Name   Version   Size   Status   Expiration Date 
    1. CM 5.2 300 Purchased Never
    2. DFM 3.2 300 Purchased Never
    3. HUM 1.0 100 Evaluation Sun Jul 26 12:18:01 GMT+03:00 2009
    4. IPM 4.2 300 Purchased Never
    5. RME 4.3 300 Purchased Never
    6. VNM 1.0 Unlimited Purchased Never
    Please find enclosed the csdiscovery.log which indicates the following error
    1000 [main] DEBUG com.cisco.nm.csdiscovery.CSDiscoveryManager  - FATAL: Exception occured in the start discovery Exception occured during start of discovery in CS.
    Reason : failed to initialize DiscoveryController
    I am also enclosing the CSDiscovery-config.xml under the config directory and the specific job as well.
    ngdiscovery.log shows 0 bytes. I am not sure if it is to be populated only when we initiate some debugging?
    Thanks a lot
    Gautam

    It seems you're missing some critical files.  Under NMSROOT/conf/csdiscovery, you should have the following zip files:
    SharedDiscoveryAUS.zipSharedDiscoveryCluster.zipSharedDiscoveryGeneral.zipSharedDiscoveryIGX.zip
    The casusers group should have full control on all four files.  If these files really are missing, you will need to contact TAC to get new ones, or you will need to reinstall Common Services.

  • Materialised view runs for ever ...

    Hi
    I'd appreciate if someone could shine some light on an issue I'm having with a materialised view.
    I inherited this when I came into my current project, so before you ask me why tables are not partitioned or why a MV is being used for such a scenario, the answer is I don't know!!
    Anyway ... I'm using "Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - Prod" running on SUSE ...
    There are a number of normalised tables two of which contain in the region of 400million rows each and tend to grow at the rate of 27000 rows per day.
    There's a materialised view (for the purposes of providing fast interactive reporting) that is setup as a normal join between 4 tables; the above two and the other two are just basic dictionary type lookups each containing around 100 or so rows.
    SELECT .... FROM LargeTable1, LargeTable2, Lookup3, Lookup4 WHERE Lookup3.ID=LargeTable1.LID and LargeTable2.ID=LargeTable1.ID and LargeTable2.LID=Lookup4.ID ;
    Running a FULL refresh on the MV failes with a 'ora-01555 snapshot too old' error after about 5 days!! Then takes around the same amount of time to rollback. Can't do a FAST refresh, I think because logs and/or mv indecies are out of sync.
    Table space is 63% full and undo space is set to auto-extend.
    I've tried /*+RULE */ but that doesn't seem to make a huge difference.
    Why does it take so long? Why does it fail? What can I do to speed it up?
    Thanks in advance for your assistance.
    Edited by: 883758 on Sep 7, 2011 2:16 PM

    You should approach this as a SQL tuning exercise, without the ability to change the query itself.
    Are table statistics up to date? Look at the execution plan – does it look reasonable for you? Are there any indexes that can help? (I am thinking covering indexes on the large tables …)
    Iordan Iotzov
    http://iiotzov.wordpress.com/

  • Query runs for ever

    I have three summary tables which are yearly,monthly and daily. The yearly table has year and some aggreagted facts. The monthly table has year,month and some aggregated facts while day table has year,month and day and some aggregated facts.
    I ran a Query on yearly table and monthly table with a join on year for these fields
    Year(year from yearly table)
    month and some aggregated facts (monthly table). This query runs forever. But if I pull all these fields from monthly table it runs fast. Can I join on year to pull the data. I dont have any keys in these tables.

    Query from different tables:
    SELECT
    PMAC.ACT_ALL_SUM_YEAR.DEPT_AIRPORT,
    PMAC.ACT_ALL_SUM_YEAR.ARR_AIRPORT,
    (PMAC.ACT_ALL_SUM_YEAR.DATE_YEAR),
    PMAC.ACT_ALL_SUM_MONTH.AVG_AIRB,
    (PMAC.ACT_ALL_SUM_MONTH.DATE_MONTH)
    FROM
    PMAC.ACT_ALL_SUM_YEAR,
    PMAC.ACT_ALL_SUM_MONTH,
    PMAC.ACT_ALL_SUM_DAY
    WHERE
    ( PMAC.ACT_ALL_SUM_YEAR.DATE_YEAR=PMAC.ACT_ALL_SUM_DAY.DATE_YEAR )
    AND ( PMAC.ACT_ALL_SUM_DAY.DATE_YEAR=PMAC.ACT_ALL_SUM_MONTH.DATE_YEAR )
    AND
    PMAC.ACT_ALL_SUM_YEAR.DEPT_AIRPORT = 'ATL'
    AND
    PMAC.ACT_ALL_SUM_YEAR.ARR_AIRPORT = 'JFK'
    AND
    (PMAC.ACT_ALL_SUM_YEAR.DATE_YEAR) In ( '2001','2002','2003' )
    )Query to pull data from single table:
    SELECT
      PMAC.ACT_ALL_SUM_MONTH.AVG_AIRB,
      (PMAC.ACT_ALL_SUM_MONTH.DATE_MONTH),
      PMAC.ACT_ALL_SUM_MONTH.ARR_AIRPORT,
      PMAC.ACT_ALL_SUM_MONTH.DEPT_AIRPORT,
      (PMAC.ACT_ALL_SUM_MONTH.DATE_YEAR)
    FROM
      PMAC.ACT_ALL_SUM_MONTH
    WHERE
       PMAC.ACT_ALL_SUM_MONTH.DEPT_AIRPORT  =  'ATL'
       AND
       PMAC.ACT_ALL_SUM_MONTH.ARR_AIRPORT  =  'JFK'
       AND
       (PMAC.ACT_ALL_SUM_MONTH.DATE_YEAR)  In  ( '2001','2002','2003'  )
      )These are two queries which I was working on. The first query takes too much time to run and it runs fast on the second query. How can I optimise the first query. Can we include any keys in these tables to make the query faster.
    Edited by: user9518444 on Dec 3, 2008 7:34 AM
    Edited by: user9518444 on Dec 3, 2008 7:38 AM

  • Activation Job running Forever

    Dear HANA Development Team,
    I created an Analytic view in the SAP HANA Sandbox system and the activation job is running for ever and the view is not getting activated.
    Could you please help me in resolving the issue?
    Thanks,
    Goutham

    i have same problem here. i tried
    sql>DBMS_SCHEDULER.STOP_JOB ('REFRESHFROMMETALINK', TRUE);
    where TRUE is FORCE STOP_JOB, but it doesn't seem to find the job class or job.
    Perhaps the jobs scheduled by OEM are administered somewhere else.
    Can anyone help with this ?
    10.1.0.4 SEO on w32

  • Background Job  running long time.

    Hello All,
    One of customer background job running for long time more than 11 hours, but usually this jobs get completed within 4 hrs.
    I have checked in WP trace file, it shows "WP has reached abap/heaplimit=40000000".
    This is the Value I can see in RZ11. Now it is not possible to change the value.
    Frequently facing this problem, please guide me to solve this issue.
    Regards
    Vinay.

    Hi
    First of all, abap/heaplimit is the limit of memory usage by a WP.
    Follwing is the documentation from help.sap.com
    The value of the parameter should be between 10000000 (10 MB) and 2000000000 (2GB), the recommended default setting is 40000000 (40 MB).
    The objective is to have the least number of work process restarts as possible, without a swap space bottleneck occurring.  The heap memory allocated by the work processes has to be released again.
    As shown in the graphic, the value of abap/heaplimit should be smaller than abap/heap_area_dia or abap/heap_area_nondia, so that the dialog step that is running can still be executed. This prevents the work process from working against the operating systemu2019s swap space limit if there is a programmed termination of the work process.
    So,Do the above checks. And importantly check the memory utilizations at the time of running your job and check whether any other job was also running at the sametime.
    Increasing the parametes is not a good idea.. If you continuously face this issue then only think of parametes. As u said ,this problem is not happenng too often.
    Hope this is useful

  • How do I set a Phase Listener to only run for a specific page?

    How do I set a Phase Listener to only run for a specific page?
    I understand that a Phase Listener runs for ever page in the application, however I want a phase listener to only run for one page at the beginning of the application.
    I discovered I can't use a Filter, because the JSF Application instance always returns null.
    Thus, I thought a Phase Listener would be a good approach.
    Thoughts?
    Thanks,
    --Todd                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

    I have moved forward with checking the requested URI...
    (shivers)
    Thanks,
    --Todd                                                                                                                                                                                   

Maybe you are looking for

  • HP Officejet Pro 8620 e-all-in-one printer & dual band router

    Hi, I'm hoping someone will be able to confirm whether this printer is compatible with my router which is a dual band 2.4 GHz and 5 GHz. I am not able to connect it to the internet. The box says that the printer is compatible only with 2.4 GHz router

  • BOM Explosion in Delivery (VL01N)

    Hi, I have business requirement in BOM Explosion in sales.  At the time of sales order (basically this is created when receive the PO from the customer), BOM is not finalised in the system and this is finalised before producing the material & even sa

  • I want a refund for my app purchase.

    Hello, I download a app call "iblack-unwanted call" for my iphone 5.  It doesn't work and I want my money back.  I want to block a caller from texting and calling.  I saw the reviews and it show that it was positive.  But once I download it, it look

  • Pro*C CONNECT or RELEASE taking 20+ seconds

    The basic problems are <li> the EXEC SQL ROLLBACK WORK RELEASE; statement takes a long time (I've seen 25 seconds) <li> the EXEC SQL CONNECT :usrname; statement also takes a long time (I've seen 8 seconds) What are the things I should check? h1. Deta

  • READ_TEXT showing problem in smart-forms

    Hi All, I am using the FM READ_TEXT in Smartforms.I have a text like M & M.When passed in the FM i get the output as M<(>&<)>M. Symbol & is displayed as junk. I debugged the FM and i can find it is getting converted at if rt_header-tdtexttype is init