Limiting number of records to be written in application server

Hi Experts,
I have a requirement where i have to limit he file size being writen to server to 500000 records per file. if the number of records to be written to file is 1 million then we ned to have 2 files of 500000 records in server.
The file need to take data from year 2008 to onwards.
Any sugguestion how to proceed?
Points will be assigned for useful answers.
Kevin

Hi,
go through the link below it may help you
http://help.sap.com/saphelp_nw70/helpdata/en/69/c24e034ba111d189750000e8322d00/content.htm
Regards,
Marasa.

Similar Messages

  • Number of records in given path of application server

    How
    to find number of records in given path of
    application server

    I think that this almost works.
    report zrich_0001.
    data: begin of itab occurs 0,
          rec(1000) type c,
          end of itab.
    data: wa(1000) type c.
    data: number_of_files type i.
    data: number_of_records type i.
    data: tot_size type i.
    data: tot_number_of_records  type i.
    data: p_file type localfile.
    data: ifiles type table of  salfldir with header line.
    data: file type string.
    parameters: p_path type salfile-longname
                        default '/usr/sap/TST/DVEBMGS01/data/'.
    call function 'RZL_READ_DIR_LOCAL'
         exporting
              name           = p_path
         tables
              file_tbl       = ifiles
         exceptions
              argument_error = 1
              not_found      = 2
              others         = 3.
    loop at ifiles.
      clear itab. refresh itab.
      concatenate p_path ifiles-name into file.
      catch system-exceptions dataset_read_error  = 1.
        open dataset file for input in text mode.
        if sy-subrc = 0.
          do.
            read dataset file into itab.
            if sy-subrc <> 0.
              exit.
            endif.
            append itab.
          enddo.
        endif.
        close dataset file.
      endcatch.
      if sy-subrc = 0.
        number_of_files = number_of_files + 1.
        describe table itab lines number_of_records .
        tot_number_of_records
            = tot_number_of_records  + number_of_records .
        tot_size = tot_size + ifiles-size.
        write:/ ifiles-name, ifiles-size,  number_of_records.
      endif.
    endloop.
    Regards,
    Rich Heilman

  • How many Maximum number of work process supported by one application server

    Friends,I have one confusion about the WP number:
    how many maximum WP supported by one applicaiton server?
    the appication server ENV is ECC6.0aix5.3oracle 10g on p595,if the supported maximum WP number is determined by number of CPU?if yes,what ratio between WPs and CPUs,thanks
    Edited by: qz sun on Apr 8, 2008 10:47 AM

    Hi Sun,
    Please click on below link and you will get the solution.
    http://help.sap.com/saphelp_erp2004/helpdata/en/19/da1c3d2b058c4ce10000000a114084/content.htm
    Regards,
    Anil

  • Process huge number of records in XI

    Hello
    I am having a simple scenario where I am sending details from SAP to legacy system on weekly basis. I am having SAP ABAP proxy as a sender and file as a receiver.
    My problem is when the data is huge (around 600,000 records) the its giving RFC timeout error on SAP ECC side. I tried to sending limited number of records per submit to XI but it creates multiple files on receiving FTP server.
    Instead of that can I use collect pattern of BPM to collect messages in BPM? Will it not be an overhead using BPM in this case as the data is tooo huge?
    If someone has already tackeled this problem pl respond.
    Thanks in advance.
    Regards
    Rajeev

    Hi Rajeev,
    dont use BPM for solving that, BPM itself has still a lot performance problems.
    If possible, split the 600.000 at R/3 side into several proxy requests. If required, use modus "add lines" of the file adapter to collect the messages again. You should think about to send the messages "Exacly Once In Order" to avoid processing more than one message at the same time.
    Regards,
    Udo

  • NUMBER OF SESSIONS CONNECTED TO AN APPLICATION SERVER

    how i can find out number of database sessions connected to an application server
    thanks and regards
    Ajith

    Current Sessions is the current total number of open sessions for each application. Yes, you can use JMX or WLST within a script.
    The WebAppComponentRuntimeMBean mbean of your application provides this information through OpenSessionsCurrentCount attribute.
    For instance, in WLST enter:
    wls:/my_domain/serverRuntime/ApplicationRuntimes/my_application/ComponentRuntimes/my_app> cmo.getOpenSessionsCurrentCount()
    For JMX I suggest to you see [http://download.oracle.com/docs/cd/E13222_01/wls/docs91/wlsmbeanref/core/index.html]
    Edited by: Betino on 22/10/2009 11:26 AM

  • Is a subquery in a BO report limited to a max number of records???

    Here's my problem:
    I recieved an excel sheet with 700 records of customers from a client who wants me to to create a report with specific data for these customers in my Business Objects universe (BO6.5 on SQL Server).
    So I created a dataprovider with query 1, i.e. the requested data of customers. Then I created a second dataprovider, query 2, based on 'personal files', i.e. the excel sheet. In query 1 I added to the conditions that each customer should be in (sub)query 2 (CustomerId In list of the query result ('query2.CustomerId').
    the syntax I have used for this seems OK.
    However, I recieve the following error: "Too many selected values (LOV0001)". I know this error has to do with parameter MAX_INLIST_VALUES, which is limited by default to 99 and can be extended to 256 max. But I thought it refers to the max number of items in lists of values.
    When I limit the number of records in the excel sheet to 99 the result is perfect (proof that I got the syntax right!). I can upgrade the parameter to 256, and can split the excel sheet into three, but that will not be useful when next time my client sends me 10.000 customer records.
    Can I make reports in BO which use subqueries that result in more than 256 records at all? (hardly imaginable).
    What is the best way to do this?
    Thanks in advance!

    Hi Lucas,
    Following is the information regarding the issue you are getting and might help you to resolve the issue.
    ADAPT00519195- Too many selected values (LOV0001) - Select Query Result operand
    For XIR2 Fixed Details-Rejected as this is by design
    I have found that this is a limitation by design and when the values exceed 18000 we get this error in BO.
    There is no fix for this issue, as itu2019s by design. The product always behaved in this manner.
    Also an ER (ADAPT00754295) for this issue has already been raised.
    Unfortunately, we cannot confirm if and when this Enhancement Request will be taken on by the developers.
    A dedicated team reviews all ERs on a regular basis for technical and commercial feasibility and whether or not the functionality is consistent with our product direction. Unfortunately we cannot presently advise on a timeframe for the inclusion of any ER to our product suite.
    The product group will then review the request and determine whether or not the functionality/feature will be included in a future release.
    Currently I can only suggest that you check the release notes in the ReadMe documents of future service packs, as it will be listed there once the ER has been included
    The only workaround which I can suggest for now is:
    Workaround 1:
    Test the issue by keep the value of MAX_Inlist_values parameter to 256 on designer level.
    Workaround 2:
    The best solution is to combine 'n' queries via a UNION. You should first highlight the first 99 or so entries from the LOV list box and then combine this query with a second one that selects the remaining LOV choices.
    Using UNION between queries; which is the only possible workaround
    Please do let me know if you have any queries related to the same.
    Regards,
    Sarbhjeet Kaur

  • MDM Limitation Maximum number of records ?

    Hello,
    we have requirement to implement SAP MDM. Customer has 20 million entries for its customer data. Can MDM support this? I heard maximum MDM can handle only 6 millions entries, above that it cannot work.
    Please confirm.

    Hi Dony,
    I think you need to do sizing as per number of records.
    https://websmp109.sap-ag.de/~sapidb/011000358700001921872008E
    You need SAP Service Market place SUID and Password for open above link (MDM sizing guide).
    I think you can check with SAP if you still have doubts regarding the number of records.
    Please find similar thread as given below:
    Does MDM 7.1 have any limitations re: number of records per repository?
    Hope this helps..
    Regards,
    Mandeep Saini

  • Limitting the number of records in query

    I like to limit the number of records retrieved in the query. I set Query All Records block property to NO, Query Array Size to the number I want to retrieve. Before doing that, I set form Preference - Array Processing checked. The result is form always retrieves all records meets the search condition. Could someone tell me what's wrong with it?

    I do not understand why and what exactly you want to do. The query array size defines the number records your form should fetch at a time, i.e. if you have 100 records in the database and you set the array size to 20, forms will perform 5 fetch cycles to retrieve all records. However, this is transparent to the user and he will see all 100 records. This property is used for performance tuning when working with a large volume of records.

  • Limit the number of record to download to excel from a report

    I am just wondering if there is a way to limit the number of rows that I want to download to excel from a report in apex, right now when I click download to excel it downloads all records. If I am displaying 15 records on report page I want the ability to download only those records.
    Any help with that would be appreciated.
    Kind Regards,
    Sofia.

    Sofia,
    The same report query runs for downloading the data into excel, I don't think you can limit the number of records in download.
    You can achieve it using custom code, like on click of download excel redirect to another page and then restrict the data as per your need, or you can use the custom procedure to download the limited data.
    Denes' utility to download into excel.
    http://htmldb.oracle.com/pls/otn/f?p=31517:108:1476564836494581:::RP,::
    Regards,
    Manish

  • How to handle a number of records in stored procedure?

    i wanna handle a number of records in a stored procedure one by one.
    what should i do?
    can any one give me some sample about the following question?
    Q:
    tb_main,tb_attach are two tables.
    i want to create a procedure to write off a record in tb_main. before do that, i want to write off all records in tb_attach which is related with the record to be written off in tb_main. and a procedure named pr_write_off_attach for writing off a record in tb_attach has been created already.
    what should i do?
    help!!
    null

    Dg.dataProvider.length is the number of records in the ArrayCollection
    Dg.rowCount is the number of visible rows.
    Alex Harui
    Flex SDK Developer
    Adobe Systems Inc.
    Blog: http://blogs.adobe.com/aharui

  • Number of records in Oracle 11G XE

    Hi.
    I'm establishing a database regime for my phd and am looking for a suitable database system.
    I'm looking for a simple data storage without the need for any complex queries or application connected to the database.
    In terms of numbers (if thats possible to say) would the XE be suitable for approx 14 mill records (in 15 columns with an average of 10 char pr column) plus an increase in number of records of approx. 100.000 each year for the next 5 years.
    Hope that was comprehansable.
    Best regards
    Nicolai

    would the XE be suitable for approx 14 mill records (in 15 columns with an average of 10 char pr column) plus an increase in number of records of approx. 100.000 each year for the next 5 years.Oracle Database Express Edition 11g Release 2 has known limitations. See Licensing Restrictions... but it probably will be able to fulfill your requirement:
    14 mill + 500k makes 14.5 mill in total.
    Simple maths would be:
    14.5 mill * 15 * 10 result in approx. 2 GB of data. Add some overhead for indexes and other things you may or may not need to store aside from your main data pool the 11 GB allowed to be used in 11.2 XE should be sufficient.
    On the other hand, you may have some restrictions on performance, since memory and cpu usage are limited as well, so processing large amounts of data at once (in a single query) might result in poor response times.
    -Udo

  • Max number of records in MDM workflow

    Hi All
    Need urgent recommendations.
    We have a scenario where we need to launch a workflow upon import of records. The challenge is source file contains 80k records and its always a FULL load( on daily basis) in MDM. Do we have any limitation in MDM workflow for the max number of records? Will there be significant performance issues if we have a workflow with such huge number of records in MDM?
    Please share your inputs.
    Thanks-Ravi

    Hi Ravi,
    Yes it can cause performance overhead and you will also have to optimise MDIS parametrs for this.
    Regarding WF i think normally it is 100 records per WF.I think you can set a particular threshold for records after which the WF will autolaunch.
    It is difficult to say what optimum number of records should be fed in Max Records per WF so I would suggest having a test run of including 100/1000 records per WF.Import Manager guide say there are several performance implications of importing records in a WF,so it is better to try for different ranges.
    Thanks,
    Ravi

  • Max number of records in an internal table

    Hi,
    Can any one tell me what is the Max Number of records we can get into an internal table.
    if you have any link of sap help on this please FWD.
    thanks in Adv.
    Regards,
    Lakshmikanth.T.V

    Hi lakshmikanth,
    Internal Tables as Dynamic Data Objects
    Internal tables are always completely specified regarding row type, key and access type. However, the number of lines is not fixed. Thus internal tables are dynamic data objects, since they can contain any number of lines of a particular type. The only restriction on the number of lines an internal table may contain are the limits of your system installation. The maximum memory that can be occupied by an internal table (including its internal administration) is 2 gigabytes. A more realistic figure is up to 500 megabytes. An additional restriction for hashed tables is that they may not contain more than 2 million entries. The line types of internal tables can be any ABAP data types - elementary, structured, or internal tables. The individual lines of an internal table are called table lines or table entries. Each component of a structured line is called a column in the internal table.
    regards,
    keerthi.

  • Max number of records in a cube & architectural issues

    Hi,
    Sorry if my question was already done but i can't find the same question with the research button (may be i have not the good words for search, I'm not english).
    I am on a BIG BIG IP project. The forecast volume of planned records is about 1.000.000.000 records a year. So we choose to split records toward many cubes. 
    1) Is there a maximum number of record supported by a cube to be planned? We planned to put 100.000.000 records maximum on a cube to be planned. Is it too much?
    2) If I make 100 cubes (one for each organizational entity) with 10.000.000 records per cube and if I "plug" a planning layout on these 100 cubes with a multiprovider, will IP:
    spread time to search the good cube to write in (due to the selected entity) in the 100 cubes (too much time !), or,
    search directly in the good cube (thanks to user exit that match cube with selected entity), so the response time will be about the same as the one for 1 layout plugged on 1 cube of 10.000.000 records?
    Thanks a lot, and sorry for my english language level
    Georges

    Hi Georges,
    Having too many records in the cube should not be very detrimental for performance of planning application as long as you can ensure that the data volume that you fetch at one go is restricted to reasonable limits, using restrictions in filters (the more restrictive the better). Take care of this while modelling both your planning functions/sequences and input-ready queries.
    I understand that you'll need to create a multiprovider for reporting purposes, but if you dont need the data from more than one cube for planning purpose, it will be better to create the aggregation level (and rest of the planning model) on top of individual cube. In case, you want to use the same planning function/input queries for multiple cubes (which will probably be the case), you can create the aggregation levels on the multiprovider but make sure you restrict the characteristic 'infoprovider' properly in the filter restrictions to avoid the function reading unnecessary data from many cubes.
    Hope this helps.

  • Applescript - how can I count the number of records in a Word data source?

    I have written a script in AppleScript to open Word and then perform a mail merge from a data source and everything is working fine.
    One thing I want to do but am struggling with is to return the number of records in the mail merge.
    I am trying things like:
    get last record of datasource of mydatasource
    and also haven't been able to find anything in the AppleScript 2004 Word Reference Guide.
    Can anyone help as I have been searching the web and trying things all day and I am sure it can't be difficult.
    Thank you.

    I don't think there is an automatic way to get a line count, but I don't have any problem just counting the lines on a page. Maybe some more information about what you are trying to do and why you can't do it would help someone to come up with a helpful suggestion.

Maybe you are looking for

  • APP not picking correct numbers

    Hi, I assigned one number range (eg: 100 to 200) for two document types ZP and ZV. While running the payment program the system is picking the numbers differently. Eg: when i run payment run on 01.04.2007 the system is picking number ranges 100,101,1

  • Comment blocks in labview?

    is there any way to make a comment block, i.e. /* comments.... */ in labview? I suppose you could enclose your code in a case or while loop that never executes- but what if you are trying to comment out code that doesn't work? Seems like this would b

  • Using Java 5, how to develop customizable web-pages  like iGoogle ?

    Is it possible to develop customizable web-pages like iGoogle does, using Java 5 ( simply using JSPs, servlets, JSFs, AJAX etc. and without using portals) ? A simple example explaining how to provide 3-4 header graphics options ( like igoogle does) w

  • Cookies, Zotero and Google Scholar

    Looks like there is some issue with Safari and Google Scholar and cookies. I've been unable to use Google Scholar with Zotero--a reference manager--as have other people I know. When I try to save a setting in Google Scholar, it does not do so, even t

  • Transactional replication from SQL Server 2014 to Sybase 12.5

    Hi All, Is it possible to set up replication between SQL Server 2014 and Sybase 12.5 in which SQL server would be the publisher and Sybase as  subscriber. Regards.. aa