TSV_TNEW_PAGE_ALLOC_FAILED with BAPI_AR_ACC_GETOPENITEMS

hi all,
I am using BAPI_AR_ACC_GETOPENITEMS to get open tems, process and update to a Ztable.The structure in the BAPI is having 111 Fields. Of which I am using only 16 fields. I think we cannot restrict the selection of fields only to particular fileds. When I am executing the program with same variant it is giving following runtime errors:
Sometimes TSV_TNEW_PAGE_ALLOC_FAILED ( at Append Lineitems) and sometimes TIME_OUT.
Most of the times I am getting first error.
1)My first question is Why I am getting two errors whrn I execute with same variant?
2)How to avoid TSV_TNEW_PAGE_ALLOC_FAILED
3) How to avoid TIME_OUT error?
I have searched in SDN for TSV_TNEW_PAGE_ALLOC_FAILED. There are many posts but I could not find a solution which supports my requirement?
Can anyone please help.
Regards,
Murthy.

(1) It seems that based on the selection criteria a huge amount of data is selected. Since there are other users in the system as well as other processes running at some point your report just can't grab any more memory because all memory is already used by your report and other users. You get different dumps because of different system resources available.
If the DB is not very busy it returns the results much faster => you run out of memory
If the DB is very busy you reach the limit for an online transaction and get the TIME_OUT dump
(2) Either BASIS need to add more memory to the application server you report is running on or you need to restrict the selection by running two or three smaller sets rather than one big one
(3) Just schedule the job in bacjground, there is no time out limit for background jobs - However if the results are represented in an ALV grid or other user interaction is necessary this option doens't work.
Hope that helps,
Michael

Similar Messages

  • TSV_TNEW_PAGE_ALLOC_FAILED  with internal table declared with occurs 0

    HI guys,
    when the internal table is declared as occurs 0, the dump TSV_TNEW_PAGE_ALLOC_FAILED is encountered, when changing the declaration into standard table, th dump disappears. Why is this so?
    Thanks!

    There are three type of tables: Standard, sorted and hashed tables (see [here|http://help.sap.com/abapdocu_70/en/ABAPDATA_ITAB.htm]). When you define a table using the keyword OCCURS you're still defining a standard table.
    Your exception indicates that you're running out of memory. The difference in the declarations that you mention shouldn't cause that. Main difference between the two declaration versions is probably that your table defined via OCCURS has a header line, whereas your other table doesn't (unless you declared it explicitly as WITH HEADER LINE).
    I don't think though that with that little information anybody could explain the short dump. Actually it sounds rather odd that the change you mention should cause the dump (or make it go away). So I suspect you probably have to post further details. You might want to check the shortdump yourself and place a breakpoint at this place and run it for each version. Maybe you can see some difference...

  • Dump TSV_TNEW_PAGE_ALLOC_FAILED with memory empty

    Hi!
    I'm with problem. My Z prog was giving dump TSV_TNEW_PAGE_ALLOC_FAILED but memory is empty.
    I'm trying append from internal table A to internal table B. I just has 1000.000 of records. Another internal table and structures used in program are clear.
    Why it's occur? And how can i solved it?

    Hi friend!
    Did you look this blog?
    /people/rajeev.p/blog/2010/07/31/top-10-abap-dumps
    "2) TSV_TNEW_PAGE_ALLOC_FAILED
    The error TSV_NEW_PAGE_ALLOC_FAILED means that more memory was requested by the system because the program needed to expand an internal table, but not is available. When Extended Memory is completely used up, the process will go into PRIV mode and it will starts using Heap Memory in case of Windows or vice versa in case of Unix. Once this enters into PRIV mode, no other user will be able to use the corresponding work process. If there is enough memory for it to finish, you will not see the error.
    Please refer the following SAP notes:
    SAP Note 649327 - Analysis of memory consumption.
    SAP Note 20527 - Runtime error TSV_TNEW_PAGE_ALLOC_FAILED
    SAP Note 185185 - Application: Analysis of memory bottlenecks
    SAP Note 369726 - TSV_TNEW_PAGE_ALLOC_FAILED "
    Maybe basis team have to expand memory config for internal tables.
    Best regards!
    Rodrigo Paisante

  • Strange TSV_TNEW_PAGE_ALLOC_FAILED dump...

    Hi guys!
    I have this weird problem:
    I have a program that does some updates to standard tables. The program runs ok, but if I comment the "update" lines, the execution ends with this dump: TSV_TNEW_PAGE_ALLOC_FAILED
    With the UPDATE doesn't cancel, without the update DOES cancel...
    Do you have any idea?????
    Thanks in advance!
    Bet

    Hi,
    probably there is a loop with reading data in an internal table:
    - without update the process does not terminate at all or nearly does not 
    - with update the process terminates soon.
    With kind R´regards
    Walter Habich

  • SAP 64bit Linux application server (Linux on Power) memory config issues

    Hello Everyone,
    Let me give some background. I have been tasked with replacing our AIX6.1 SAP application server with new Linux on Power servers (SLES11SP3)
    Our current production BW application servers (AIX) are running fine given their current setup. We are fine implementing the same size of the AIX systems on the new Linux systems, but we now have more room to grow in the fur, however when I tried to implement the Linux server using the same memory sizing I am getting runtime errors: "TSV_TNEW_PAGE_ALLOC_FAILED" with a specific user created program and process chain. The AIX servers do not run into this issue running the exact same program. When this process runs, it utilizes 3 BTC process for the program.
    Using the note 941735, I have tried to correctly setup the system using STD memory implementation, however I have a feeling that I am missing something or setup something incorrectly which is causing the above issue. Looking in ST02 here is the SAP buffer parameters for two of the systems. Perhaps someone might notice something that I am not see or point me in the right direction. Please let me know if more parameters from the profile are needed. Thanks!
    AIX system (existing):
    ztta/roll_area              6500000    Byte Roll area per workprocess (total)
    ztta/roll_first             1          Byte First amount of roll area used in a dialog WP
    ztta/short_area             3200000    Byte Short area per workprocess
    rdisp/ROLL_MAXFS            32768      8 kB Maximum roll file size
    rdisp/ROLL_SHM              16384      8 kB Part of roll file in shared memory
    rdisp/PG_SHM                16384      8 kB Part of paging file in shared memory
    rdisp/PG_LOCAL              150        8 kB Paging buffer per workprocess
    em/initial_size_MB          2027       MB   Initial size of extended memory
    em/blocksize_KB             4096       kB   Size of one extended memory block
    em/address_space_MB         4092       MB   Address space reserved for ext. mem. (NT only)
    em/global_area_MB           256        MB   Extended global memory
    ztta/roll_extension         2000000000 Byte Max. extended mem. per session (external mode)
    abap/heap_area_dia          2000000000 Byte Max. heap memory for dialog workprocesses
    abap/heap_area_nondia       2000000000 Byte Max. heap memory for non-dialog workprocesses
    abap/heap_area_total        2000000000 Byte Max. usable heap memory
    abap/heaplimit              250000000  Byte Workprocess restart limit of heap memory
    abap/use_paging             0               Paging for flat tables used (1) or not (0)
    abap/shared_objects_size_MB 20         MB   Shared objects
    Linux system (new):
    ztta/roll_area              6500352    Byte Roll area per workprocess (total)
    ztta/roll_first             1024       Byte First amount of roll area used in a dialog WP
    ztta/short_area             3200000    Byte Short area per workprocess
    rdisp/ROLL_MAXFS            131072     8 kB Maximum roll file size
    rdisp/ROLL_SHM              32768      8 kB Part of roll file in shared memory
    rdisp/PG_SHM                16384      8 kB Part of paging file in shared memory
    rdisp/PG_LOCAL              150        8 kB Paging buffer per workprocess
    em/initial_size_MB          4092       MB   Initial size of extended memory
    em/blocksize_KB             4096       kB   Size of one extended memory block
    em/address_space_MB         4096       MB   Address space reserved for ext. mem. (NT only)
    em/global_area_MB           256        MB   Extended global memory
    ztta/roll_extension         2000683008 Byte Max. extended mem. per session (external mode)
    abap/heap_area_dia          2000683008 Byte Max. heap memory for dialog workprocesses
    abap/heap_area_nondia       2000683008 Byte Max. heap memory for non-dialog workprocesses
    abap/heap_area_total        2000683008 Byte Max. usable heap memory
    abap/heaplimit              250609664  Byte Workprocess restart limit of heap memory
    abap/use_paging             0               Paging for flat tables used (1) or not (0)
    abap/shared_objects_size_MB 20         MB   Shared objects

    Yes sapconf has been installed. sapconf however only sets the kernel settings.
    I was able to resolve the issue by having Zero Admin tool using 'es/implementation = map' and setting the PHYS_MEMSIZE to the desired amount. Then commenting out all the values that are defaulted in the 32bit version of the memory management guide for linux and let it size many of the values. Then I simply copied those from st02. I updated the profile with those values to override default for  'es/implementation = std' and commented out PHYS_MEMSIZE as it is not needed. In a nut shell, the following parameters were updated.
    es/implementation = std
    em/address_space_MB = 4096
    em/blocksize_KB = 4096
    em/initial_size_MB = 14320
    em/max_size_MB = 14326
    rdisp/PG_MAXFS = 131072
    rdisp/PG_SHM = 16384
    rdisp/ROLL_MAXFS = 131072
    rdisp/ROLL_SHM = 131072
    ztta/roll_area = 6500352
    ztta/roll_extension = 4000000000
    ztta/roll_extension_dia = 4000000000
    ztta/roll_extension_nondia = 4000000000
    ztta/roll_first = 1

  • CTM Run failing with TSV_TNEW_PAGE_ALLOC_FAILED runtime error

    Hi,
    While doing the CTM run,system is giving the  memory dump with run time error TSV_TNEW_PAGE_ALLOC_FAILED.
    I understand this is because memory constraint and possible recommendation to either increase the Memory or reduce the master data selections for planning but would like to know if some master data discrepancy can cause the infinite run of CTM and hence memory dump like product at location being set as Inhouse production but no PPM maintained for it.
    Any suggestions on this willl reward points.
    Thanks,
    Manish
    Edited by: Manish Thukral on Jun 24, 2008 9:34 PM

    Marcelo,
    Thanks for you reponse.Can you please share some more information on what explanation profile is and where it is set so that can I check.
    Manish

  • Delta load failing with dump TSV_TNEW_PAGE_ALLOC_FAILED

    Hi,
    When we are loading BW statistics delta load, the load is failing with error "Job in source system terminated --> Request is set to red." and its giving short dump with "TSV_TNEW_PAGE_ALLOC_FAILED".
    I have already gone through Forum and changed package size also. But could not solve the problem.
    We are using 3.5 version, Could you please post your thought ASAP.
    Regards
    Satya

    Hi there,
    Yes that dump is lack of memory...
    So probably your package has too many records to be processed, especially if you have code in start routine or end routine or something like that, the process for each package might take too long and fails with out of memory.
    Try in fact change the package size to a smaller size, so fewer records are processed in each package.
    Diogo.

  • Job CRM_TERRMAN_PROC_REL failed with error TSV_TNEW_PAGE_ALLOC_FAILED

    Hi Experts
    I have scheduled the CRM_TERRMAN_PROC_REL using INITIAL run to update the territory on accounts. In between INITAL run got failed/cancelled due to error "Internal session terminated with a run time error TSV_TNEW_PAGE_ALLOC_FAILED"..
    I have re-run the job using INITIAL again. Again Job ended up with same error.
    Is it ok to run job multiple time in INITIAL mode when Job gets failed in between?
    What could be the reason for the error Internal session terminated with a run time error TSV_TNEW_PAGE_ALLOC_FAILED"..?
    Thanks
    Chand

    Check if below sap note valid for your release.
    1933212 - Dump TSV_TNEW_PAGE_ALLOC_FAILED when running CRM_TERRMAN_PROC_REL
    Refer:
    1732395 - Memory allocation dump during territory relation processing
    2054266 - CRM Territory Relationship Build Program Performance issue

  • Some SAP Jobs cancel with error TSV_TNEW_PAGE_ALLOC_FAILS

    Some of background jobs started cancelling since 10 days ago.
    Here are info from ABAP Dump:
    The amount of storage space (in bytes) filled at termination time was:
    Roll area...................... 6221248
    Extended memory (EM)........... 50278176
    Assigned memory (HEAP)......... 799483568
    Short area..................... " "
    Paging area.................... 24576
    Maximum address space.......... 18446744063552822256
    Below are my current setting:
    Roll, extended and heap memory EXTM
    ztta/roll_area        6500000    Byte
    ztta/roll_first       1000000    Byte
    ztta/short_area       3200000    Byte
    rdisp/ROLL_SHM        16384      8 kB
    rdisp/PG_SHM          8192       8 kB
    rdisp/PG_LOCAL        150        8 kB
    em/initial_size_MB    2048       MB
    em/blocksize_KB       4096       kB
    em/address_space_MB   4092       MB
    ztta/roll_extension   50000000   Byte
    abap/heap_area_dia    1800000000 Byte
    abap/heap_area_nondia 800000000  Byte
    abap/heap_area_total  1000000000 Byte
    abap/heaplimit        20000000   Byte
    abap/use_paging       0
    Here are more info
    ECC6 SPS13
    Oracle 10.2.0.2
    Solaris 5.9
    Can some please let me know what will be correct parameter setting?
    I have played with them a little bit, but still I am getting the error.
    Regards,
    Ray

    Hi,
    Please check the link this may resolve your issue.
    Re: dump TSV_TNEW_PAGE_ALLOC_FAILED
    Regards
    K.R

  • Unable to generate WebService Client with Exported (BAPI) WSDL file from XI

    Hi there,
    I, trying to setup a simple Web Service environment where I'll be exposing a BAPI as Web Service via XI.
    So these are the steps I followed:
    <b>1)</b> first I imported the BAPI meta data into my SWC in the repository. <b>2)</b> Second I defined my interrface etc.. <b>3)</b> Third I exported the generated BAPI WSDL file. <b>4)</b> Fourth, I tried to generate a Java proxy/client for my Web Service using a Java IDE (IBM WebSphere Application Developer and XML Spy).
    Now the problem is that Im not able to generate the Web Service client using the exported WSDL file, the WSDL seems not to be valid...From the generated errors in the IDE, I can see that the problem has to do somewhere/somehow with the WSDL Binding element, which input parameters do not match the input parameters of the port type.
    This is a small part of the WSDL file as generated by XI:
         <wsdl:message name="BAPI_AR_ACC_GETOPENITEMS.Input">
              <wsdl:part name="parameters" element="s0:BAPI_AR_ACC_GETOPENITEMS" xmlns:wsdl="http://schemas.xmlsoap.org/wsdl/"/>
         </wsdl:message>
         <wsdl:message name="BAPI_AR_ACC_GETOPENITEMS.Output">
              <wsdl:part name="parameters" element="s0:BAPI_AR_ACC_GETOPENITEMS.Response" xmlns:wsdl="http://schemas.xmlsoap.org/wsdl/"/>
         </wsdl:message>
         <wsdl:portType name="BAPI_AR_ACC_GETOPENITEMS.PortType">
              <wsdl:operation name="BAPI_AR_ACC_GETOPENITEMS">
                   <wsdl:input message="s0:BAPI_AR_ACC_GETOPENITEMS.Input"/>
                   <wsdl:output message="s0:BAPI_AR_ACC_GETOPENITEMS.Output"/>
              </wsdl:operation>
         </wsdl:portType>
         <wsdl:binding name="BAPI_AR_ACC_GETOPENITEMS.SAPBinding" type="s0:BAPI_AR_ACC_GETOPENITEMS.PortType" xmlns:wsdl="http://schemas.xmlsoap.org/wsdl/">
              <s0:binding style="rfc"/>
    Has anyone faced a situation like this before? Any clues where to look? I appreciate any feedback, thanks!
    Rob.
    Message was edited by: rob viana

    Hi Rob!
    I'm facing the same problem - how did you solve your issue then? I'm getting following error during generation on .NET side:
    ============================================================================================
    Error: There was an error processing 'http://server/wsdl/MIAMI_ACCDOCPost.wsdl'.
      - The document at the url http://server/wsdl/MIAMI_ACCDOCPost.wsdl
    was not recognized as a known document type.
    The error message from each known type may help you fix the problem:
    - Report from 'DISCO Document' is 'Discovery document at the URL http://server/wsdl/MIAMI_ACCDOCPost.wsdl could not be found.'.
      - The document format is not recognized (the content type is 'content/unknown').
    - Report from 'WSDL Document' is 'There is an error in XML document (2, 70138).'.
      - Value cannot be null.
    Parameter name: name
    - Report from 'XML Schema' is 'The root element of a W3C XML Schema should be <schema> and its namespace should be 'http://www.w3.org/2001/XMLSchema'.'.
    ============================================================================================
    kr Martin

  • Clearing: how to get open items with net amount?

    Hi
    I need to obtain the open items for clearing <b>with the net amount</b> as it shows up in the transaction FB05 after considering any discounts that may apply. The BAPI BAPI_AR_ACC_GETOPENITEMS does not calculate the net amount.
    I see that the data is calculated in SAPMF05B and saved in the internal table POSTAB which has the structure RFOPS. Basically i need to obtain the data of this internal table.
    I was thinking of exporting the table in some user exit, but didn't find any suitable exit yet.
    Any ideas anyone...?
    Thanks for any hint.
    Andreas

    Hi All,
    Thanks for the reply..
    ==============================================================
    Basically create one ckf and two rkfs for each aging bucket
    ie ckf 0-30 days = rkf 0-30 days open + rkf 0-30 days subsequently cleared
    Because you need to restate the AP report to a key date - some items have changed their status since that key date but were open at the key date (ie we call those subsequently cleared)
    so rkf 0-30 days open is
    posting date <= key date
    item status = O
    net due date between variable date (cmod which reads key date) offset 0 and variable date (cmod which reads key date) offset -30
    rkf 0-30 days subsequently cleared is
    posting date <= key date
    cleared date > variable date (which reads cmod of key date)
    net due date between variable date (cmod which reads key date) offset 0 and variable date (cmod which reads key date) offset -30
    ========================================================================
    In the above two cases, while creating restricted key figures how to design this;
    net due date between variable date (cmod which reads key date) offset 0 and variable date (cmod which reads key date) offset -30
    And also after creation of two rkfs i created a ckf i.e ckf = rkf1+rkf2.
    Now i dragged this ckf into columns and when i checked the query it is giving error that key date variable is used more than once. ( Bcz we have used key date variable both in rkf1 and rkf2 for getting posting date <= to key date )
    What is the solution for this ?
    Regards,
    Rao.

  • Short Dump       TSV_TNEW_PAGE_ALLOC_FAILED

    Hi All,
    I am facing the short dump "TSV_TNEW_PAGE_ALLOC_FAILED" problem in my PRD system.
    Please find ST22 log and suggest the solution:
    Runtime Errors         TSV_TNEW_PAGE_ALLOC_FAILED
    Date and Time          18.11.2009 12:12:09
    Short text
         No more storage space available for extending an internal table.
    What happened?
         You attempted to extend an internal table, but the required space was
         not available.
    What can you do?
         Note which actions and input led to the error.
         For further help in handling the problem, contact your SAP administrator
         You can use the ABAP dump analysis transaction ST22 to view and manage
         termination messages, in particular for long term reference.
         Try to find out (e.g. by targetted data selection) whether the
         transaction will run with less main memory.
         If there is a temporary bottleneck, execute the transaction again.
         If the error persists, ask your system administrator to check the
         following profile parameters:
         o  ztta/roll_area            (1.000.000 - 15.000.000)
                Classic roll area per user and internal mode
                usual amount of roll area per user and internal mode
    o  ztta/roll_extension       (10.000.000 - 500.000.000)
            Amount of memory per user in extended memory (EM)
    o  abap/heap_area_total      (100.000.000 - 1.500.000.000)
            Amount of memory (malloc) for all users of an application
            server. If several background processes are running on
            one server, temporary bottlenecks may occur.
            Of course, the amount of memory (in bytes) must also be
            available on the machine (main memory or file system swap).
            Caution:
            The operating system must be set up so that there is also
            enough memory for each process. Usually, the maximum address
            space is too small.
            Ask your hardware manufacturer or your competence center
            about this.
            In this case, consult your hardware vendor
    abap/heap_area_dia:        (10.000.000 - 1.000.000.000)
            Restriction of memory allocated to the heap with malloc
            for each dialog process.
    Parameters for background processes:
    abap/heap_area_nondia:        (10.000.000 - 1.000.000.000)
            Restriction of memory allocated to the heap with malloc
            for each background process.
    Other memory-relevant parameters are:
    em/initial_size_MB:         (35-1200)
            Extended memory area from which all users of an
            application server can satisfy their memory requirement.
    or analysis
    The internal table "\FUNCTION-POOL=EL40\DATA=GL_NODETAB[]" could not be further
      extended. To enable
    Error handling, the table had to be delete before this log was written.
    As a result, the table is displayed further down or, if you branch to
    the ABAP Debugger, with 0 rows.
    At the time of the termination, the following data was determined for
    the relevant internal table:
    Memory location: "Session memory"
    Row width: 2160
    Number of rows: 1782088
    Allocated rows: 1782088
    Newly requested rows: 4 (in 1 blocks)
    to correct the error
    The amount of storage space (in bytes) filled at termination time was:
    Roll area...................... 4419712
    Extended memory (EM)........... 2002743520
    Assigned memory (HEAP)......... 2000049152
    Short area..................... " "
    Paging area.................... 32768
    Maximum address space.......... " "
    If the error occures in a non-modified SAP program, you may be able to
    find an interim solution in an SAP Note.
    If you have access to SAP Notes, carry out a search with the following
    keywords:
    "TSV_TNEW_PAGE_ALLOC_FAILED" " "
    "SAPLEL40" or "LEL40U11"
    "ISU_ELWEG_HIERARCHY_BUILD"
    Please help me out to resolve the issue.
    Regards,
    Nitin Sharma

    hI cHANDRU,
    tHANKS FOR YOUR RESPONSE:
    pLEASE FIND BELOW MENTIONED DETALS:
    Operating system..... "Windows NT"
    Release.............. "5.2"
    Hardware type........ "8x AMD64 Level"
    Character length.... 16 Bits
    Pointer length....... 64 Bits
    Work process number.. 0
    Shortdump setting.... "full"
    Database server... "SVPSAPECP01"
    Database type..... "MSSQL"
    Database name..... "ECP"
    Database user ID.. "ecp"
    Char.set.... "C"
    SAP kernel....... 700
    created (date)... "Nov 18 2008 22:53:36"
    create on........ "NT 5.2 3790 Service Pack 1 x86 MS VC++ 14.00"
    Database version. "SQL_Server_8.00 "
    Patch level. 185
    Patch text.. " "

  • Internal table Memory Issue Exception TSV_TNEW_PAGE_ALLOC_FAILED

    Hi experts,
    I am working on a conversiojn programme. This programme is dealing with 4 input files.
    Each of these files is having more than 50,000 records. I am reading the corresponding application server files to fill
    the internal tables related to these files.
    The files are being read properly and internal tables are being filled.
    However when i try to assign any of these 4 internal tables to other temproray internal tables in programme(requirement)
    i get a dump  TSV_TNEW_PAGE_ALLOC_FAILED.
    The dump is related to memory issue.
    I think The memory available in the programme at this point is not sufficient for table assignment.
    Please suggest any alternatives where i can save any memory .
    Changig of basis setting is not an option.
    Regards,
    Abhishek Kokate

    Hi Kiran,
    I am not agree with you , I am agree with Hermann.
    While writting file you restrict the record max 5,000 to 10,000 records and process don't store the mutch data into internal table.
    After every used refresh the internal table, Declare table where necessary.
    But you can try to avoid the copy cost.
    Rgds
    Ravi Lanjewar

  • Production error - Urgent -  TSV_TNEW_PAGE_ALLOC_FAILED

    Hi All,
    Im getting dumps in production system with
    Runtime Error:          TSV_TNEW_PAGE_ALLOC_FAILED. and when checked the trace, it says as Data type "MCVBAPB" was found in a newer version than required.
    Any pointers to this wil be helpful
    thanks
    Booma

    the TSV_TNEW_PAGE_ALLOC_FAILED dump has naught to do with the version of MCVBAPB. the dump clearly states that there was no more paging area available to run the report/transaction that caused the dump. this might either be due to an too big selection or a lack of memory. Consult your basis-people with that one.
    the version of MCVBAPB: MCVBAPB is a structure belonging to LIS (Logistics Information System) so obviously someone changed that structure. check in t-code SE11, MCVBAPB the runtime and the database object of the structure (in case it is active).

  • Run time error : TSV_TNEW_PAGE_ALLOC_FAILED in st22

    Hi All,
    we are getting runtime error : TSV_TNEW_PAGE_ALLOC_FAILED in production system. for this report rsppfprocess we have batch jobs so may be its running all the action definition and coming up with consuming more memory ? we already checked with basis but they are saying its rsppfprocess program issue. can you please guide me how to solve this issue.
    Jimmi.

    Hi Jimmi,
    The error TSV_NEW_PAGE_ALLOC_FAILED means that more memory was requested
    because the program needed to expand an internal table, but is not available.
    When Extended Memory is used up, the process will go into PRIV mode
    as it starts using Heap Memory  (or vice-versa).  No other user will be able to use this work process while it is in PRIV mode.
    If there is enough heap for it to finish, you will not see the error TSV_NEW_PAGE_ALLOC_FAILED and the work process will be freed.
    This seems to be a problem with loading of shared memory.
    Check value of parameter abap/shared_objects_size_MB in RZ11.
    Set the parameter to minimum 250 and increase it when necessary, as per SAP note 1281896. You must start the system newly so that the change becomes effective.
    Also take a look at SAP Note 1166259.
    Gervase

Maybe you are looking for

  • 11gR2 : Failed to start EVM while running root.sh in Node1

    DB verision:11GR2 Platform : Solaris 5.10 (Amd64) I've reached the last step in the installation in Node1 # ./root.sh .............all messages say succeeded till here............ CRS-2676: Start of 'ora.crsd' on 'hudson1' succeeded CRS-2672: Attempt

  • Problem while creating a sale order against a purchase order

    functional description 1 Create a module-pool program for material management and do following:- a)     Create a screen to display all purchase Order 2 same screen should have a tree for all the available PO -done till this part 4 Select a PO from tr

  • Brand new iMac 24" crashes and burns playing games in bootcamp

    If this continues I guess I will have to call support but for those that are interested here is the story: I have a 24" iMac 3gig and I have installed bootcamp - although I am not convinced as to how well this worked with the driver install as I have

  • Is external hard drive necessary for FCE?

    I recently purchased brand new imac with Final Cut Express. It hasn't even arrived yet!!! Is an external fire wire HD necessary? I plan on editing home movies with approx. 60 Minutes of raw footage at a time and then putting on DVD. The imac comes wi

  • Dreamweaver CS4 not working under MacOS 10.9

    on starting DW CS4 it asks for Java Runtime engine 6, but Mavericks has Java v.7 installed. Since Java 6 was part of the OS of previous systems, I cannot install it seperately, besides it would probably cause conflicts with OS 10.9. Do I have to tras