Dump during archive job runs -  TSV_TNEW_PAGE_ALLOC_FAILED (snote 1017000)

Hi All,
    I'm currently having problems with the deletion/archiving process on one of my XI boxes. I has been
configure to use the Switch Procedure and, the error happens during the TABLE_SWITCH procedure.
    Due to this error, I have 1 dump per day and the other jobs related with this task are failing.
    The dump says:
           TSV_TNEW_PAGE_ALLOC_FAILED
    and I have found note 1017000 for this issue and sucesfully applied.
    According to the note, next step should be:
             "To stop the archiving process, start the deletion process for the failed
              archiving runs manually directly from the archive administration."
    by I'm not able to do that. I go to XMS_SATA and with object name BC_XMB try to run the "Delete" action
    but can not continue as there is no archive file to select.
    Anyone with the same issue?
    Thanks in advance?

Hi Prateek,
   Thanks for your help but I don see how is this related to XI. I read about IS-OIL and IS-MINE. I already found a note that helps with this issue but I'm stuck with one of the steps and can cancel the copy process.
    Regards,
        Encinas

Similar Messages

  • Performance Optimization during the Data Archiving Initial runs

    Dear Experts,
    We are running archiving jobs (Write Job & Preprocessing Jobs) which are running for nearly 56hrs. It is writting and analyzing nearly 4 million records.
    Any suggestions to improve the performance for running these jobs to complete in less time. Its ISU Archiving.
    I have gone through Many SAP Notes, but i'm unable to find related to improving performance during the archival runs. These are initial runs.
    Thank You,
    Mahesh C

    You should check the sql statements that are being executed by these runs.. Possibly by adding additional indexes you can speedup the process.
    Kind regards,
    Mark

  • Archiving Job. Running and no file..

    Hello all!
    For a some cubes (not for all!!!) I have problem.
    Archiving job write to log
    DROP VIEW "/BI0/0302712378"                                                 
    SQL-END: 15.01.2007 12:09:29 00:00:00                                       
    Selection conditions for the archiving request were checked successfully    
    Archiving session 000012 is being created                                   
    And it is all... Job running up to 48 hours and no files created...
    Other cube have not problem:
    28.12.2006 12:04:49 SQL-END: 28.12.2006 12:04:49 00:00:00                                   
    28.12.2006 12:04:49 Selection conditions for the archiving request were checked successfully
    28.12.2006 12:04:50 Archiving session 000010 is being created                               
    28.12.2006 12:04:52 Path: /Archive/                                                         
    28.12.2006 12:04:52 Name for new archive file: archive_BW_1228_0.120452.GOLOD
    28.12.2006 17:09:09 Name for new archive file: archive_BW_1228_0.170909.GOLOD        
    Help me.
    Thanks.

    Hello,
    Please ask in the
    Visual Basic .Net Language forum on MSDN if you want a program.
    If you are creating a script or batch file, I'd ask in
    The Official Scripting Guys Forum!
    Neither place is a good place to go to if you want someone else to do the work for you.
    If you want hints or suggestions to improve what you already have, then the above suggestions are valid.
    Karl
    When you see answers and helpful posts, please click Vote As Helpful, Propose As Answer, and/or Mark As Answer.
    My Blog: Unlock PowerShell
    My Book:
    Windows PowerShell 2.0 Bible
    My E-mail: -join ('6F6C646B61726C406F75746C6F6F6B2E636F6D'-split'(?<=\G.{2})'|%{if($_){[char][int]"0x$_"}})

  • Spool get lost while job runs in background

    Hi
    We wrote a custom program where in we use two custom bapi's which were copied from original bapi's. When we run this interface program in backgroud, we have lost the spool. But when we run the test with limited data in foreground, it works fine. What are the possible solutions to correct this issue. Is there anyway we can force this spool to stay even though any BAPI fails in the middle or fail to provide proper error?
    Does this cause because of any memory leak? or insufficient memory? But there are no short dumps... in system.
    If we direct to print the spool during background job execution, it comes back with spool object not found error.
    Any help would be highly appreciated.
    Thanks
    Pandiri

    What exactly do you mean by: we have lost the spool? Did you check in SM37 (job overview), and the spool icon is missing, or are you looking at SP01 and you don't see any output / spool requests?
    When running program in background, are you providing the correct print parameters?
    When you run it in foreground, do you create a spool request during the report?
    Is report aborting when running in backgroud, because than you won't get any output / spool request, just a message stating that program was aborted / failed somewhere along the  line?
    Need some additional info before we can make an educated guess....

  • Error occurred during CTM planning run

    Hi folks,
    Appreciate your co-operations!
    I am facing the problem while running the CTM with the profile DEMO2.
    CTM Planning Run gives one error and alert.
    Error: Error occurred during CTM planning run
    Technical Data
    Message type__________ A (Cancel)
    Message class_________ /SAPAPO/CTM1 (CTM: Messgaes)
    Message number________ 401
    Problem class_________ 1 (very important)
    Number________________ 1
    Environment Information
    CTM Action____________ G
    Message type__________ A
    Alert: Internal error has occurred (<!> Segmentation fault)
    Technical Data
    Message type__________ E (Error)
    Message class_________ /SAPAPO/CTM1 (CTM: Messgaes)
    Message number________ 571
    Message variable 1____ <!> Segmentation fault
    Number________________ 1
    Environment Information
    CTM Action____________ G
    Message type__________ C
    Log file display
    <i> 04:37:59 optsvr_main.cpp(1363) 'SuperVisor' => Commandline : 4 respected parameters ...
    Args:
    m0001006
    sapgw04
    28812935
    IDX=1
    <i> 04:37:59 optsvr_main.cpp(645) 'SuperVisor'  * SAP APO CTM Engine [CTM/ctmsvr]
    <i> 04:37:59 optsvr_main.cpp(646) 'SuperVisor'  * Copyright u00A9 SAP AG 1993-2009
    <i> 04:37:59 optsvr_main.cpp(647) 'SuperVisor'  *
    <i> 04:37:59 optsvr_main.cpp(648) 'SuperVisor'  * Version        : 7.0_REL SP05, 407661, Nov 25 2009 22:59:47
    <i> 04:37:59 optsvr_main.cpp(649) 'SuperVisor'  * Platform       : ntamd64/x64
    <i> 04:37:59 optsvr_main.cpp(650) 'SuperVisor'  * Interface      : 2.0
    <i> 04:37:59 optsvr_main.cpp(651) 'SuperVisor'  * Build date     : Nov 25 2009 22:59:47 [1259186387]
    <i> 04:37:59 optsvr_main.cpp(652) 'SuperVisor'  * Build machine  : PWDFM163
    <i> 04:37:59 optsvr_main.cpp(653) 'SuperVisor'  * Latest change  : 407661
    <i> 04:37:59 optsvr_main.cpp(654) 'SuperVisor'  * NW release     : 7100.0.3300.0
    <i> 04:37:59 optsvr_main.cpp(655) 'SuperVisor'  * Perforce branch: 7.0_REL
    <i> 04:37:59 optsvr_main.cpp(656) 'SuperVisor'  *
    <i> 04:37:59 optsvr_main.cpp(676) 'SuperVisor'  * Hostname       : m0001006
    <i> 04:37:59 optsvr_main.cpp(677) 'SuperVisor'  * OS version     : 5.2.3790 (WinServer2003, NTAMD64) SP2.0 (Service Pack 2), SERVER ENTERPRISE TERMINAL SINGLEUSERTS
    <i> 04:37:59 optsvr_main.cpp(678) 'SuperVisor'  * PID            : 6768
    <i> 04:37:59 optsvr_main.cpp(683) 'SuperVisor'  * CWD            : D:\usr\sap\SC6\DVEBMGS04\log
    <i> 04:37:59 optsvr_main.cpp(684) 'SuperVisor'  *
    <i> 04:37:59 core_sysinfo.cpp(453) 'SuperVisor' * free disk space: 190433 MB
    <i> 04:37:59 core_sysinfo.cpp(454) 'SuperVisor' *
    <i> 04:37:59 core_sysinfo.cpp(409) 'SuperVisor' * Memory information:
    <i> 04:37:59 core_sysinfo.cpp(409) 'SuperVisor' *   physical memory: 10238 MB total, 6511 MB available [63% free]
    <i> 04:37:59 core_sysinfo.cpp(409) 'SuperVisor' *   page file      : 73212 MB total, 60889 MB available [83% free]
    <i> 04:37:59 core_sysinfo.cpp(409) 'SuperVisor' *   virtual memory : 8388607 MB total, 8388499 MB available [99% free]
    <i> 04:37:59 optsvr_main.cpp(693) 'SuperVisor'  *
    <i> 04:37:59 optsvr_main.cpp(783) 'SuperVisor' * running in invoke mode
    <i> 04:37:59 optsvr_rfcconnection.cpp(871) 'MsgMgr' <RFC> RfcPing(RFC_HANDLE=1) received in thread#6912
    <i> 04:37:59 optsvr_rfcconnection.cpp(692) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_PARAM_SET' for sending of parameters/options
    <i> 04:37:59 optsvr_rfcconnection.cpp(703) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_PARAM_GET' for receiving of parameters/options
    <i> 04:37:59 optsvr_rfcconnection.cpp(712) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_PROGRESS' for progress informations
    <i> 04:37:59 optsvr_rfcconnection.cpp(721) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_MESSAGE' for messages
    <i> 04:37:59 optsvr_rfcconnection.cpp(730) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_RESULT' for (intermediate) result informations
    <i> 04:37:59 optsvr_rfcconnection.cpp(739) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_SYSINFO' for system informations
    <i> 04:37:59 optsvr_rfcconnection.cpp(748) 'MsgMgr' <RfcConnection> using function module 'RCCF_COMM_PERFINFO' for performance informations
    <i> 04:37:59 optsvr_rfcconnection.cpp(1269) 'MsgMgr' <RFC> skipping empty profile value [GENERAL] sPROFILE_CUST_ID
    <i> 04:37:59 optsvr_rfcconnection.cpp(1835) 'MsgMgr'
    Sender/Receiver RFC_HANDLE#1:
    <RFC> * RFC connection attributes:
      Own Host    : m0001006
      Partner Host: m0001006
      Destination : OPTSERVER_CTM01
      Program Name: SAPLRCC_COMM_ENGINE
      SystemNr    :              04       SystemId    : SC6
      Client      :             700       User        : MBATCHA    
      Language    :               E       ISO Language: EN
      CodePage    :            1100       Partner CP  : 1100
      Kernel Rel. :            701        Partner Rel.: 701
      Own Release :            711        CPIC ConvId : 28812935
      Own Type    :               E       PartnerType : 3
      Trace       :                       RFC Role    : S
    <RFC> * RFC statistic information:
      number of calls        : 7
      number of received data: 10569
      number of sent data    : 1349
      overall reading time   : 9073
      overall writing time   : 162
    <i> 04:37:59 optsvr_main.cpp(1110) 'SuperVisor' * Starting MainScript ...
    <i> 04:37:59 optsvr_main.cpp(1445) 'SuperVisor'
    ***************************** OPTSVR - OPTIONS ***************************** *
    [CTM_PROFILE]
    nCTMENGINEPACKAGESIZE = 500
    sCOMPONENT = SCM
    sCTMLOGFILE = ctm.DEMO2.0000_0001.20091201043758.log
    sCTMLOGFLAG = 0
    sCTMPROFILE = DEMO2
    sRELEASE = 700
    [general]
    bUNICODE = true
    nSLOT_MAXIMUM = 1
    nSLOT_MINIMUM = 1
    nSLOT_RESERVED = 1
    sAPO_RELEASE = 700
    sAPPLICATION = CTM
    sExeDir = d:\apoopt\ctm\bin
    sExeName = ctmsvr.exe
    sHOST = m0001006
    sInvokeMode = invoke
    sLANGU = E
    sMANDT = 700
    sPRODUCT_NAME = APO
    sPRODUCT_PATCHLEVEL = 0001
    sPRODUCT_RELEASE = 700
    sPROFILE = DEMO2
    sSESSION = tju5Bmz21}6WVG0Sn6pv3W
    sSYSTEM = SC6
    sUNAME = MBATCHA
    [init]
    sSECTION0001 = INIT
    sSECTION0002 = GENERAL
    sSECTION0003 = PASSPORT
    sSECTION0004 = CTM_PROFILE
    [PASSPORT]
    bIS_REMOTE = false
    nACTION_TYPE = 1
    nSERVICE = 1
    sACTION = /SAPAPO/CTMB
    sPRE_SYSID = SC6
    sSYSID = SC6
    sTRANSID = 2205DEDE7A5BF16DA07D001CC46CF90E
    sUSERID = MBATCHA
    ************************** OPTSVR OPTIONS - END **************************** *
    <i> 04:37:59 core_msgmgr.cpp(440) 'MsgMgr' * Sending progress number 802 to OutputInterface from []
    <i> 04:37:59 core_supervisor.cpp(728) 'SuperVisor' <M> Invoking module 'CTMModelGenerator' [6]->download
    <i> 04:37:59 core_msgmgr.cpp(440) 'MsgMgr' * Sending progress number 806 to OutputInterface from [MG]
    <i> 04:37:59 ctm_modelgen.cpp(166) 'CTMModelGenerator' ======================================================================
    <i> 04:37:59 ctm_modelgen.cpp(167) 'CTMModelGenerator' MG::download
    <i> 04:37:59 core_msgmgr.cpp(1110) 'MsgMgr' renaming tracefile
    <i> 04:37:59 core_msgmgr.cpp(1111) 'MsgMgr' old name: optsvr_trace20091201_043759_1a70.log
    <i> 04:37:59 core_msgmgr.cpp(1112) 'MsgMgr' new name: ctm.DEMO2.20091201_043759_1a70.log
    logfile reopened : Tue Dec 01 04:37:59 2009
    logfile name     : ctm.DEMO2.20091201_043759_1a70.log
    <i> 04:37:59 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_STATUS_SET
    <i> 04:37:59 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_PRDAT_RFC_READ
    <i> 04:37:59 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_PLPAR_RFC_READ
    <i> 04:37:59 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_LOC_RFC_READ
    <i> 04:37:59 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_PPM_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_TRANS_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_RES_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_SSTCK_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_CAL_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_PLPER_RFC_READ
    <i> 04:38:02 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_INCMD_RFC_READ
    <i> 04:38:03 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_STATUS_SET
    <i> 04:38:03 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_DEM_RFC_READ
    <i> 04:38:04 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_SUP_RFC_READ
    <i> 04:38:04 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_UCMAP_RFC_READ
    <i> 04:38:04 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_STATUS_SET
    <i> 04:38:04 core_msgmgr.cpp(440) 'MsgMgr' * Sending progress number 810 to OutputInterface from [MG]
    <i> 04:38:04 ctm_modelgen.cpp(735) 'CTMModelGenerator' MG::download done
    <i> 04:38:04 ctm_modelgen.cpp(736) 'CTMModelGenerator' ======================================================================
    <i> 04:38:04 core_supervisor.cpp(750) 'SuperVisor' <M> Returning from module 'CTMModelGenerator' [6]->download = success [ctx size : 1]
    <i> 04:38:04 core_supervisor.cpp(692) 'SuperVisor' <SCR> Starting script 'CTM Solve' with 9.22337e+012 seconds left
    <i> 04:38:04 core_supervisor.cpp(692) 'SuperVisor' <SCR> Starting script 'CTM Match' with 9.22337e+012 seconds left
    <i> 04:38:04 ctm_executionmanager.cpp(102) 'SuperVisor' ======================================================================
    <i> 04:38:04 ctm_executionmanager.cpp(103) 'SuperVisor' statistics:
    <i> 04:38:04 ctm_executionmanager.cpp(104) 'SuperVisor' number of demands: 7
    <i> 04:38:04 ctm_executionmanager.cpp(105) 'SuperVisor' ======================================================================
    <i> 04:38:04 ctm_executionmanager.cpp(107) 'SuperVisor' ======================================================================
    <i> 04:38:04 ctm_executionmanager.cpp(108) 'SuperVisor' parameters:
    <i> 04:38:04 ctm_executionmanager.cpp(118) 'SuperVisor' time continuous planning
    <i> 04:38:04 ctm_executionmanager.cpp(125) 'SuperVisor' backward scheduling
    <i> 04:38:04 ctm_executionmanager.cpp(184) 'SuperVisor' CBCLP enabled
    <i> 04:38:04 ctm_executionmanager.cpp(457) 'SuperVisor' ======================================================================
    <i> 04:38:04 core_supervisor.cpp(728) 'SuperVisor' <M> Invoking module 'CtmEngine' [7]->run
    <i> 04:38:04 ctm_executionmanager.cpp(523) 'SuperVisor' ======================================================================
    <i> 04:38:04 ctm_executionmanager.cpp(524) 'SuperVisor' EM::execute for packet 1
    <i> 04:38:04 ctm_executionmanager.cpp(1570) 'SuperVisor' EM::execute for packet 1 done
    <i> 04:38:04 ctm_executionmanager.cpp(1571) 'SuperVisor' ======================================================================
    <i> 04:38:04 core_supervisor.cpp(750) 'SuperVisor' <M> Returning from module 'CtmEngine' [7]->run = success [ctx size : 1]
    <i> 04:38:04 core_supervisor.cpp(728) 'SuperVisor' <M> Invoking module 'CTMModelGenerator' [6]->upload
    <i> 04:38:04 ctm_modelgen.cpp(1097) 'CTMModelGenerator' ======================================================================
    <i> 04:38:04 ctm_modelgen.cpp(1098) 'CTMModelGenerator' MG::upload of packet 1
    <e> 04:38:05 ctmsvr_script.cpp(229) 'SuperVisor' <!> STRING EXCEPTION : <!> Segmentation fault
    <i> 04:38:05 rfc_connection.cpp(599) 'MsgMgr' <rfc> calling function module /SAPAPO/CTM_INT_STATUS_SET
    <i> 04:38:05 optsvr_main.cpp(1166) 'MsgMgr' Current check values:
    [CHECK_EQUAL]
    [CHECK_UPPERBOUND]
    nPEAK_MEMORY_NTAMD64 = 45344
    [CHECK_LOWERBOUND]
    <i> 04:38:05 optsvr_main.cpp(1209) 'MsgMgr' Performance values:
    bSuccess     false
    nCPU_TIME     0
    nPEAK_MEMORY     45344
    nPEAK_VIRTUAL_BYTES     141844
    nREAL_TIME     6
    tracefile     ctm.DEMO2.20091201_043759_1a70.log
    <i> 04:38:05 optsvr_main.cpp(1235) 'MsgMgr' Performance Monitor values:
    ENGINE_VERSION     7.0_REL SP05, 407661, Nov 25 2009 22:59:47
    nCPU_TIME     0
    nHD_FREESPACE     190433
    nPEAK_MEMORY     45344
    nPEAK_VIRTUAL_BYTES     141844
    nREAL_TIME     6
    <i> 04:38:05 optsvr_dsr.cpp(96) 'MsgMgr' <writeDSRdata> tracing not active => no DSR written
    <i> 04:38:05 optsvr_main.cpp(1256) 'SuperVisor'
    Finished->FAILED ...
    <i> 04:38:05 core_memmgr.cpp(564) 'MsgMgr'   transferring memory of heap 6912 to main heap
    <i> 04:38:05 core_memmgr.cpp(606) 'MsgMgr'   finished transfer of heap 6912
    <i> 04:38:05 optsvr_rfcconnection.cpp(1835) 'MsgMgr'
    Sender/Receiver RFC_HANDLE#1:
    <RFC> * RFC connection attributes:
      Own Host    : m0001006
      Partner Host: m0001006
      Destination : OPTSERVER_CTM01
      Program Name: SAPLRCC_COMM_ENGINE
      SystemNr    :              04       SystemId    : SC6
      Client      :             700       User        : MBATCHA    
      Language    :               E       ISO Language: EN
      CodePage    :            1100       Partner CP  : 1100
      Kernel Rel. :            701        Partner Rel.: 701
      Own Release :            711        CPIC ConvId : 28812935
      Own Type    :               E       PartnerType : 3
      Trace       :                       RFC Role    : S
    <RFC> * RFC statistic information:
      number of calls        : 116
      number of received data: 420457
      number of sent data    : 39262
      overall reading time   : 5.30093e+006
      overall writing time   : 3831
    <i> 04:38:05 optsvr_main.cpp(1332) 'MsgMgr'
    OptimizeServer says  GOOD BYE
    Please help me to resolve this issue.
    Thanks & Regards,
    Khadar

    Hi Khadar,
    1) The information you have provided is the CTM optimiser log.
    Run the job in background and in  sm37 and click on job log &
    analyse the exact error happened.  In case if you are not able
    to do, please provide the error log.
    2) Check the livecache is stable in its operations when the job
    runs (check with basis team)
    3) Run consistency check for master data before CTM run
    4) Check for any struck queues and clear those and rerun
    5) If you feel more inconsistencies in system, run livecache
    consistency and rerun CTM run
    Regards
    R. Senthil Mareeswaran.

  • SQL server agent job running as Agent Service Account whose service account does not have r/w access but is still able to write?

    Hi. I am newer to SQL server security and am reviewing some of our SQL server's configuration to make sure the services are running under accounts with least privilege.  I have a SQL server 2012 instance whose Agent service is configured to run
    under an AD user account named 'SQLServices'.  The jobs on this server are configured to run as 'SQL server agent service account', which means they should execute as user 'SQLServices'.  The jobs are set up to execute SSIS packages which read and
    write to a database on the same server where the agent job is scheduled and SSIS package installed (all on same server).  The jobs are currently executing without error and are reading writing data correctly.  Upon close examination, it turns out the
    SQLServices account is not assigned to the 'sysadmin' role and had no users mapped to any databases on this server.  How are these jobs working?  I verified in profiler that the login name indeed is 'SqlServices'.  I also verified
    that SQLServices login has no database access by remote-ing onto the server and trying to log into the DB, and access was denied as expected.  According to the literature, the Agent service needs to be a member of 'sysadmin role' but I am reading
    some cases where that is not necessarily the case.  So this is not so concerning.  What is concerning is that the login 'SQLServices' had no access to the databases on that server yet it is reading and writing to the databases as if it does. 
    The only thing I can think of is maybe jobs run as 'SQL server agent service account' on the same server as the databases it r/w to somehow has some kind of default access.  What am I missing here?  Any input would be helpful.

    After 2 days on this forum I found the answer to my own question.  In retrospect, I should have posted this under 'SQL Server Security', but I didn't know it existed.
    The 2 threads below explain that Sql agent actually runs using SID (service) NT SERVICE\SQLSERVERAGENT if you chose that when you installed.  This will automatically create an associated login NT SERVICE\SQLSERVERAGENT in SQL server with sqladmin
    role.  This is the login that Agent uses to connect to the local instance of SQL server.  If you changed to domain account to run the service during install or after using config manager, basically NT SERVICE\SQLSERVERAGENT is still
    used to connect to your local instance behind the scenes (even though you will still see your domain user as account), and the domain account is used to reach outside the server. 
    https://social.msdn.microsoft.com/Forums/sqlserver/en-US/9e6bb2de-8fd0-45de-ab02-d59bbe05f72e/servicedatabase-accounts-nt-servicemssqlserver-nt-servicesqlserveragent-what-are-they-for
    https://social.technet.microsoft.com/Forums/sqlserver/en-US/b83a52fd-fe11-4c28-a27b-88be8ae79f2a/how-do-i-change-sql-server-agent-service-account-to-nt-servicesqlserveragent?forum=sqlsecurity

  • Dump during BAPI_ENTRYSHEET_CREATE calling from Portal

    Hi,
    I am getting dump during creation of Service Entry Sheet using the 'BAPI_ENTRYSHEET_CREATE' when triggering it from Portal end. But if I directly call the FM from ECC6.0 with the same passing values it gives me the success meassage. The dump only occurs when I am triggering it from Portal via another custom RFC funtion module.
       I have checked in ST22 and the message text is like ....." Exception condition "GUI_NO_MODE_AVAILABLE" raised."
    Also have checked the SAP Note1271652 and have suggested to apply kernel patch 184. But kernel patch 185 already been applied as our basis team have confirmed. Can you please help me out of this situaion.
    Thanks & Regards,
    Apurba

    Hi,
    I came across this issue and have blogged a solution that gets round it for our purposes. This may help:
    http://community.altiusconsulting.com/blogs/adamadshead/archive/2009/09/24/bex-analyzer-s-file-upload-bug-when-integrated-with-sap-portal.aspx
    Regards,
    Adam

  • Dump during creation of service entry sheet (ML81N)

    Dump during creation of service entry sheet (ML81N)
    Hi,
    ML81N>Select PO>Enter>goto entry sheet create (empty page)>service entry sheet displayed with internal no. assigned >click tab "service selection">Input PO No.--> enter (here we get Dump ABAP RUNTIME ERROR "TABLE_INVALID_INDEX").
    This is happening only for some POs. What can be the problem?. Is this has something to do with SAP database ???
    - rahul

    Hi,
    it could have the cause that the note 1113220 was missing from your system or that data are inconsistent in your PO.
    If you implemented 1113220 and you still get the dump, please check in transaction ST22 the run time error long text to the dump:
    Parts:
    - How to correct the error?
    - Information on where terminated
    - Source Code Extract -> if you set a breakpoint before the line where the error occured, you could debug the process.
    - Active Calls/Events: if you have an own program displayed here, please check first your program wether it has caused the issue.
    If these do not help to find out the problem cause, please open a SAP ticket and include all of the relevant information regarding the dump.
    Regards,
    Edit

  • ERRORImport Monitor jobs: running 0, waiting 1, completed 17, failed 1,

    Hello,
    I am installing ECC 6.0 on database Maxdb7.6. window 2003 service pack 2 processor AMD 3400 RAM 2 GB,
    I am getting the following error at the ERRORImport Monitor jobs: running 0, waiting 1, completed 17, failed 1, total 19.
    WARNING 2010-10-29 22:40:15
    Execution of the command "C:\j2sdk1.4.2_09\bin\java.exe -classpath migmon.jar -showversion com.sap.inst.migmon.imp.ImportMonitor -dbType ADA -importDirs D:\Ides\exportdb\db1\EXP1;D:\Ides\exportdb\db2\EXP2;D:\Ides\exportdb\db3\EXP3;D:\Ides\exportdb\db4\EXP4;D:\Ides\exportdb\db5\EXP5;D:\Ides\exportdb\db6\EXP6;D:\Ides\exportdb\db6\EXP7;D:\Ides\exportdb\db6\EXP8;D:\Ides\exportdb\db6\EXP9;D:\Ides\exportdb\db6\EXP10;D:\Ides\exportdb\db6\EXP11 -installDir "C:\Program Files\sapinst_instdir\ERP\SYSTEM\ADA\CENTRAL\AS" -orderBy "" -r3loadExe E:\usr\sap\IOL\SYS\exe\uc\NTI386\R3load.exe -tskFiles yes -extFiles no -dbCodepage 4103 -jobNum 3 -monitorTimeout 30 -loadArgs " -nolog -c 0" -trace all -sapinst" finished with return code 103. Output: java version "1.4.2_09"Java(TM) 2 Runtime Environment, Standard Edition (build 1.4.2_09-b05)Java HotSpot(TM) Client VM (build 1.4.2_09-b05, mixed mode)Import Monitor jobs: running 1, waiting 1, completed 17, failed 0, total 19.Loading of 'SAPSSEXC' import package: ERRORImport Monitor jobs: running 0, waiting 1, completed 17, failed 1, total 19.
    ERROR 2010-10-29 22:40:15
    CJS-30022  Program 'Migration Monitor' exits with error code 103. For details see log file(s) import_monitor.java.log, import_monitor.log.
    ERROR 2010-10-29 22:40:15
    FCO-00011  The step runMigrationMonitor with step key |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|1|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_ABAP_Import_Dialog|ind|ind|ind|ind|5|0|NW_ABAP_Import|ind|ind|ind|ind|0|0|runMigrationMonitor was executed with status ERROR .
    import_monitor.java.log
    java version "1.4.2_09"
    Java(TM) 2 Runtime Environment, Standard Edition (build 1.4.2_09-b05)
    Java HotSpot(TM) Client VM (build 1.4.2_09-b05, mixed mode)
    Import Monitor jobs: running 1, waiting 1, completed 17, failed 0, total 19.
    Loading of 'SAPSSEXC' import package: ERROR
    Import Monitor jobs: running 0, waiting 1, completed 17, failed 1, total 19.
    PLZ help me
    Thanks&regards
    Vijay kumar
    Moved from ABAP, General
    Edited by: kishan P on Nov 2, 2010 10:17 AM

    ERROR: 2010-10-29 22:07:26 com.sap.inst.migmon.LoadTask run
    Loading of 'SAPSSEXC' import package is interrupted with R3load error.
    Process 'E:\usr\sap\IOL\SYS\exe\uc\NTI386\R3load.exe -i SAPSSEXC.cmd -dbcodepage 4103 -l SAPSSEXC.log -nolog -c 0' exited with return code 2.
    For mode details see 'SAPSSEXC.log' file.
    Standard error output:
    sapparam: sapargv( argc, argv) has not been called.
    sapparam(1c): No Profile used.
    sapparam: SAPSYSTEMNAME neither in Profile nor in Commandline
    WARNING: 2010-10-29 22:07:56
    Cannot start import of packages with views because not all import packages with tables are loaded successfully.
    WARNING: 2010-10-29 22:07:56
    1 error(s) during processing of packages.
    INFO: 2010-10-29 22:07:56
    Import Monitor is stopped.
    INFO: 2010-10-29 22:39:45
    Import Monitor is started.
    CONFIG: 2010-10-29 22:39:45
    Application options:
    dbCodepage=4103
    dbType=ADA
    extFiles=no
    importDirs=D:\Ides\exportdb\db1\EXP1;D:\Ides\exportdb\db2\EXP2;D:\Ides\exportdb\db3\EXP3;D:\Ides\exportdb\db4\EXP4;D:\Ides\exportdb\db5\EXP5;D:\Ides\exportdb\db6\EXP6;D:\Ides\exportdb\db6\EXP7;D:\Ides\exportdb\db6\EXP8;D:\Ides\exportdb\db6\EXP9;D:\Ides\exportdb\db6\EXP10;D:\Ides\exportdb\db6\EXP11
    installDir=C:\Program Files\sapinst_instdir\ERP\SYSTEM\ADA\CENTRAL\AS
    jobNum=3
    loadArgs= -nolog -c 0
    monitorTimeout=30
    orderBy=
    r3loadExe=E:\usr\sap\IOL\SYS\exe\uc\NTI386\R3load.exe
    sapinst=
    trace=all
    tskFiles=yes
    CONFIG: 2010-10-29 22:39:45
    List of packages with table structure: 'SAP0000'.
    CONFIG: 2010-10-29 22:39:45
    List of packages with views: 'SAPVIEW'.
    TRACE: 2010-10-29 22:39:45 com.sap.inst.migmon.imp.ImportStandardTask preCreate
    Parse of 'C:\Program Files\sapinst_instdir\ERP\SYSTEM\ADA\CENTRAL\AS\DDLADA.TPL' template file is started.
    INFO: 2010-10-29 22:39:45 com.sap.inst.migmon.imp.ImportStandardTask preCreate
    Parse of 'C:\Program Files\sapinst_instdir\ERP\SYSTEM\ADA\CENTRAL\AS\DDLADA.TPL' template file is successfully completed.
    Primary key creation: before load.
    Index creation: after load.
    INFO: 2010-10-29 22:39:45
    Data codepage 4103 is determined using TOC file 'D:\Ides\exportdb\db1\EXP1\DATA\SAPAPPL0.TOC' for package 'SAPAPPL0'.
    TRACE: 2010-10-29 22:39:45 com.sap.inst.migmon.LoadTask run
    Loading of 'SAPSSEXC' import package is started.
    TRACE: 2010-10-29 22:39:45 com.sap.inst.migmon.LoadTask processPackage
    Loading of 'SAPSSEXC' import package into database:
    E:\usr\sap\IOL\SYS\exe\uc\NTI386\R3load.exe -i SAPSSEXC.cmd -dbcodepage 4103 -l SAPSSEXC.log -nolog -c 0
    ERROR: 2010-10-29 22:39:46 com.sap.inst.migmon.LoadTask run
    Loading of 'SAPSSEXC' import package is interrupted with R3load error.
    Process 'E:\usr\sap\IOL\SYS\exe\uc\NTI386\R3load.exe -i SAPSSEXC.cmd -dbcodepage 4103 -l SAPSSEXC.log -nolog -c 0' exited with return code 2.
    For mode details see 'SAPSSEXC.log' file.
    Standard error output:
    sapparam: sapargv( argc, argv) has not been called.
    sapparam(1c): No Profile used.
    sapparam: SAPSYSTEMNAME neither in Profile nor in Commandline
    WARNING: 2010-10-29 22:40:15
    Cannot start import of packages with views because not all import packages with tables are loaded successfully.
    WARNING: 2010-10-29 22:40:15
    1 error(s) during processing of packages.
    INFO: 2010-10-29 22:40:15
    Import Monitor is stopped.

  • Long runnning IDOC Archiving jobs

    Hi Techies,
    Long runnning IDOC Archiving jobs: ARV_IDOC_WRI* are being observed in PRD system with Avg runtime > 150,000 sec.
    Job Details:
    Main Program: RSEXARCA
    Parameter: IDOC ALL MONTH
    These jobs are also resource consuming & extensively using CPU & Memory.
    For info: These jobs are getting finished but only after long runtime.
    Some SQL stmts observed while this job-execution:
    SELECT
      "OBJKEY" , "OBJTYPE" , "LOGSYS" , "ROLETYPE" , "ROLEID" , "UTCTIME"
    FROM
      "SRRELROLES"
    WHERE
      "CLIENT" = :A0 AND "OBJKEY" = :A1 AND "OBJTYPE" = :A2#
    SELECT
      "MANDT" , "DOCNUM" , "COUNTER" , "PAGENO" , "TIMESTMP" , "PAGELG" , "VARDATA"
    FROM
      "EDI40"
    WHERE
      "MANDT" = :A0 AND "DOCNUM" = :A1
    ORDER BY
      "MANDT" , "DOCNUM" , "COUNTER" , "PAGENO"#
    SELECT
      "DOCNUM" , "DOCREL" , "STATUS"
    FROM
      "EDIDC"
    WHERE
      "MANDT" = :A0 AND "UPDDAT" >= :A1 AND "UPDDAT" BETWEEN :A2 AND :A3 AND "DIRECT" BETWEEN :A4 AND :A5 AND
      "STATUS" = :A6 AND NOT "MESTYP" IN ( :A7 , :A8 , :A9 , :A10 , :A11 , :A12 , :A13 , :A14 , :A15 , :A16 ,
      :A17 , :A18 , :A19 , :A20 , :A21 , :A22 , :A23 , :A24 , :A25 , :A26 , :A27 , :A28 , :A29 , :A30 , :A31 )
    Please suggest as to how the job runtime can be reduced.
    Awaiting for your reply.

    Hi,
    Since the variant used is fetching data for the entire month, you could try splitting the jobs and have multiple runs instead.
    You can also try to schedule the jobs when there is minimum system load, incase its not already done.
    As far as the program is concerned, if this is standard program, there is little scope for tuning.
    You can always check for related index quality and statements with high CPU costs to better understand why the job is running for such a long duration.

  • Automate Archiving Jobs

    Hi
    We wanted to run the Archive jobs every year by Company Code and Object wise.
    We have around 7 sales companies  and 10 objects so we have to run 70 jobs Every year. So we wanted to automate this process so that we dodnt need to eneter the texts etc and schedule the jobs every year.
    But these are the problems which we are facing:
    1.If we wanted to Archive SD_VBAK object for the ABC Sales org , then we need to enter some text( some thing like "ABC Slaes org 1999 Year Sales docs")  in the varaint(this can done using table TVARVC(By using Dynamic variants).
    2. But after the jobs, we need to change the Spool paramaters,where in we will select the output device,Storage Mode as " Archive " only,     and Text same as which I used earlier in step .1.
    But how can we make this Text field to get the value from TVARVC or somewhere ? Sicne its Popo-up box.(This spool text will be the Archove File name ,which will be displayed when we are reading the Archives.)
    Can somebody help to advice what are ways to Automate the Archive jobs( so that users doenst need to run these 70 jobs manually every year by entering Text and spool text etc..?
    We are running on SAP 4.7 version.
    Thankyou

    SARA is the tcode for archive objects

  • Scheduling of Archive jobs

    Hi,
    i really need to apologize here, but i did not found an appropriate forum to ask this, neither a forum for ADK archiving development kit nor a forum about XBP (intercepted jobs). If there is a froum for asking questions like this you can slap me for that
    The archiving object is configured to schedule the ARVSTO jobs automatically which we want to keep to spread the load over a certain time and not to have a bunch of archiving jobs always running at a specific time.
    In this system we do not have a connection to a central job scheduling tool due to license costs.
    The problem is that sometimes ARVSTO jobs are scheduled exactly at the time where our IXOS serving is going down for 15 minutes for maintanence.
    I'm now looking for ANY possibility to avoid that ARVSTO jobs are scheduled between 4 and 5 o'clock. I hat a look at OSS notes 458670 and 604496, with those notes i can avoid that the jobs are starting automatically but i need an external job scheduler or i need to start them manually. I dont have a job scheduler and i dont want to do this manually.
    I had a look at report RSARCHD to do a modifaction but as far as i can see this one is only scheduling the ARVDEL jobs.
    So does anyone have an idea to solve this? Even if it is a modification, or en enhancement, a BADI or en exit, i dont care.
    Thanks in advance.

    Create one procedure and write all your logic which you want to schedure.
    Please execute the below block to set the job.
    Please change ### with Your PROCEDURE NAME WITH PARAMETER.
    'SYSDATE+1/1440' = Indicate that your job execute every minute..
    DECLARE
    X NUMBER;
    BEGIN
    SYS.DBMS_JOB.SUBMIT
    ( job => X
    ,what => '###'
    ,next_date => to_date('17/02/2010 12:54:32','dd/mm/yyyy hh24:mi:ss')
    ,interval => 'SYSDATE+1/1440 '
    ,no_parse => TRUE
    SYS.DBMS_OUTPUT.PUT_LINE('Job Number is: ' || to_char(x));
    END;
    commit;
    Edited by: user10594896 on Feb 16, 2010 11:28 PM

  • Get Idoc Types Processed by Jobs Running Programs RBDAPP01 or RBDMANI2

    Hi Everyone! I would just like to ask if there is a way I can identify all background jobs which are processing a specific idoc type in SAP?
    For example, I want to get the Jobs that are processing inbound idoc type WMMBXY using program RBDAPP01 or RBDMANI2.
    Is there a way to extract this information?
    Thanks in advance!

    Vincent, we have a conservative approach RE those programs that re-process IDOCs in pending or error status : we do not define those jobs with open selections, instead we are specific in terms of message type and partner where applicable.  We also set a limit in the number or attempts on a failed IDOC by setting a 'rolling start date'  (like current date - 7).  Besides audit considerations about lingering processes,  you also need to take into account the impact of this process on the tale size. Each time the job runs and the IDOC fails, a status record will be added to the EDIDS table for each of those IDOC.  If you do not pay attention, you may end up with large amounts of data. This is not only going to unplease your BASIS and DB administrator but eventually put in jeopardy your archiving process (wchich could lead to a system crash !!).
    We usually aslo complement those jobs with a status change step (moving from error to no further processing) ; the date parameter can be set to something like anythiing with a creation date > current date + 3 months
    To come back to your question : SM37 using the program name as parameter will help you find the associated jobs, but you might need to look into the step(s) and associated variant to get the messages those jobs / steps are processing

  • Regarding R/3 down time during the INIT Run

    Hi All,
    Please let me know the methods to reduce to R/3 Down time during the INIT Run.
    What happends if the tables( Setup / Update) are not locked in R/3 during the INIT Run.
    Your help would be appreciated.
    Thanks,
    Vijay

    Hi,
    To conclude..
    In Direct delta:
    note that no documents can be posted during delta initialization procedure from the start of the recompilation run in R/3 (setup tables filling job) until all records have been successfully updated in BW: every document posted in the meantime is irrecoverably lost.
    In Queued delta:
    The document postings (relevant for the involved application) can be opened again as soon as the execution of the recompilation run (or runs, if several and running in parallel) ends, that is when setup tables are filled, and a delta init request is posted in BW, because the system is able to collect new document data during the delta init uploading too (with a deeply felt recommendation: remember to avoid update collective run before all delta init requests have been successfully updated in your BW!).
    don’t forget that, regardless of the update method selected, it’s ALWAYS necessary to stop any document postings (for the relevant application) during setup tables recompilation run !
    Hope its clear.
    reward if helpful.
    shylaja.

  • Problem with Sales Order CIF IM-job running longer

    We have CIF integratiom Model generation(RIMODGEN) and activation(RIMODAC2) jobs
    running everyday in sequence.
    One of the jobs is for Sales Order Generation
    and activation. The activation job used to take 30-40 seconds earlier(since last
    1 year) but for the last month the job has been taking a lot of time to
    complete. And that too intermittently. On some days the job completes in 30-40
    seconds but on some it takes more than 3000 seconds(and still doesnt complete-
    eventually we mark it successful manually to let successor jobs to start).
    When we analysed the log for this job transfer in SLG1 we found out that
    during activation of this Integration Model, not only Sales Orders pertaining to
    new materials(that would have been added within the day between last IM
    generation job and current IM generation job) were getting transferred. But
    Sales orders pertaining to Materials(which have been active since a number of
    days) were also getting transferred. Ideally, any changes/new Sales orders
    pertaining to these older materials(which were already in an active integration
    model) should have been transferred real time and should not have waited to be
    transferred during the IM activation job.
    We have also noticed that for
    some Sales Olders(for which the materials were already active for long) and
    there has been NO change in the Sales Orders are also getting transferred
    throguh our IM activation job.
    The mesage that we see in the log for
    these Sales Order is shown below:
    Source system:XXXXXX user: RSP100
    transaction CFM2 T_SLS 0001119 function module: /SAPAPO/CIF_XXXXX (As far as I
    understand CFM2 T_SLS* prefix is for Sales Order transfer happen during IM
    activation job)
    "Customer Requirement N BM 0001223200 00010 0001: Method
    changed from N to G"
    "Customer Requirement 0001223200: Requirement adopted
    in APO"

    Is the full name of the directory where you have your source code Documents and Settings as in Win2k or WinXP. In that case, move the source code to a different directory(you might like to create a separate directory). Note that the directory name should not have spaces in it. Try and see what you get.

Maybe you are looking for

  • JMS To proxy Runtime error

    Hello,       I am doing a JMS to Proxy scenario, so in mapping at run time i am getting an error as: Fatal Error: com.sap.engine.lib.xml.parser.ParserException: XMLParser: No data allowed here: (hex) 30, 30, 20(:main:, row:1, col:3) com.sap.aii.utilx

  • Disable log-in feature for Photoshop Elements 12

    I really do not want to be connected to Adobe in any way, shape or form while using their product. I detest and loathe this company.  The day another company produces a way to do accurate frequency seperation for retouching, I will be gone for good.

  • Can't get Win2k/NT/XP service working

    I am trying to create a service and have followed the example here: http://zone.ni.com/devzone/conceptd.nsf/webmain/A9B2FA96B851394586256B26006ECB98?opendocument This question may actually be more related to VI server's however, I'm not sure. I have

  • Appletalk, what is it and how does it work?

    Can someone explain how Appletalk works? I have a printer at work that can use Appletalk,but the zones and all that, I dont understand. Thanks Nana

  • Pacman won't let me update

    Hi I get the following errors when trying pacman -Suy... strangely my man pacman command has also stopped working, I am just wondering if there is a force option I should/ could use? Here is the output [root@sam tmp]# pacman -Syu :: Synchronising pac