Max. number of records in a package

Hi,
I was asked an interview question of 'what are the maximum number of records a BW package can have?'
Can some one please answer?
Thanks
Sarah

Hi
Maximum size of a data packet in kByte                            20000
U can find this in Scheduler(Maintain Info package) screen. In this screen in the menu go to Scheduler ---> DataS.Default Data transfer
regards
kiran

Similar Messages

  • Max number of records in MDM workflow

    Hi All
    Need urgent recommendations.
    We have a scenario where we need to launch a workflow upon import of records. The challenge is source file contains 80k records and its always a FULL load( on daily basis) in MDM. Do we have any limitation in MDM workflow for the max number of records? Will there be significant performance issues if we have a workflow with such huge number of records in MDM?
    Please share your inputs.
    Thanks-Ravi

    Hi Ravi,
    Yes it can cause performance overhead and you will also have to optimise MDIS parametrs for this.
    Regarding WF i think normally it is 100 records per WF.I think you can set a particular threshold for records after which the WF will autolaunch.
    It is difficult to say what optimum number of records should be fed in Max Records per WF so I would suggest having a test run of including 100/1000 records per WF.Import Manager guide say there are several performance implications of importing records in a WF,so it is better to try for different ranges.
    Thanks,
    Ravi

  • Max number of records in an internal table

    Hi,
    Can any one tell me what is the Max Number of records we can get into an internal table.
    if you have any link of sap help on this please FWD.
    thanks in Adv.
    Regards,
    Lakshmikanth.T.V

    Hi lakshmikanth,
    Internal Tables as Dynamic Data Objects
    Internal tables are always completely specified regarding row type, key and access type. However, the number of lines is not fixed. Thus internal tables are dynamic data objects, since they can contain any number of lines of a particular type. The only restriction on the number of lines an internal table may contain are the limits of your system installation. The maximum memory that can be occupied by an internal table (including its internal administration) is 2 gigabytes. A more realistic figure is up to 500 megabytes. An additional restriction for hashed tables is that they may not contain more than 2 million entries. The line types of internal tables can be any ABAP data types - elementary, structured, or internal tables. The individual lines of an internal table are called table lines or table entries. Each component of a structured line is called a column in the internal table.
    regards,
    keerthi.

  • What is the max number of records in table

    Hello Friends,
    am using oracle 11g .
    How many records we can store in a table or what is the maximum size of the table . On what factors it depends.
    If the number of records are ever growing , what is the best possible solutiion ?
    thanks/kumar

    There is a limit based on the limit of the ROWID.
    You may find this limit in Oracle documentation.
    From database version 9.0 it is virtually unlimited for us, as it is hardly likely to reach the max value of the ROWID with data we can store now and with the actual speed of our computers (reported to our lifetime).

  • Is a subquery in a BO report limited to a max number of records???

    Here's my problem:
    I recieved an excel sheet with 700 records of customers from a client who wants me to to create a report with specific data for these customers in my Business Objects universe (BO6.5 on SQL Server).
    So I created a dataprovider with query 1, i.e. the requested data of customers. Then I created a second dataprovider, query 2, based on 'personal files', i.e. the excel sheet. In query 1 I added to the conditions that each customer should be in (sub)query 2 (CustomerId In list of the query result ('query2.CustomerId').
    the syntax I have used for this seems OK.
    However, I recieve the following error: "Too many selected values (LOV0001)". I know this error has to do with parameter MAX_INLIST_VALUES, which is limited by default to 99 and can be extended to 256 max. But I thought it refers to the max number of items in lists of values.
    When I limit the number of records in the excel sheet to 99 the result is perfect (proof that I got the syntax right!). I can upgrade the parameter to 256, and can split the excel sheet into three, but that will not be useful when next time my client sends me 10.000 customer records.
    Can I make reports in BO which use subqueries that result in more than 256 records at all? (hardly imaginable).
    What is the best way to do this?
    Thanks in advance!

    Hi Lucas,
    Following is the information regarding the issue you are getting and might help you to resolve the issue.
    ADAPT00519195- Too many selected values (LOV0001) - Select Query Result operand
    For XIR2 Fixed Details-Rejected as this is by design
    I have found that this is a limitation by design and when the values exceed 18000 we get this error in BO.
    There is no fix for this issue, as itu2019s by design. The product always behaved in this manner.
    Also an ER (ADAPT00754295) for this issue has already been raised.
    Unfortunately, we cannot confirm if and when this Enhancement Request will be taken on by the developers.
    A dedicated team reviews all ERs on a regular basis for technical and commercial feasibility and whether or not the functionality is consistent with our product direction. Unfortunately we cannot presently advise on a timeframe for the inclusion of any ER to our product suite.
    The product group will then review the request and determine whether or not the functionality/feature will be included in a future release.
    Currently I can only suggest that you check the release notes in the ReadMe documents of future service packs, as it will be listed there once the ER has been included
    The only workaround which I can suggest for now is:
    Workaround 1:
    Test the issue by keep the value of MAX_Inlist_values parameter to 256 on designer level.
    Workaround 2:
    The best solution is to combine 'n' queries via a UNION. You should first highlight the first 99 or so entries from the LOV list box and then combine this query with a second one that selects the remaining LOV choices.
    Using UNION between queries; which is the only possible workaround
    Please do let me know if you have any queries related to the same.
    Regards,
    Sarbhjeet Kaur

  • Max number of records to hold in explicit cursor

    Hi Everyone,
    What is the maxmimum number of records that could be holded in
    an explicit cursor for manipulation. I need to process millions of records.
    Can I hold it in cursors or use temp table to hold those records and
    do fixes with volume control.
    Thanks

    Hi Kishore sorry for the delayed response,
    Table1
    prim_oid     sec_oid          rel_oid
    pp101     cp102          101
    pp101     cp103          101
    pp102     cp104          101
    pp102     cp105          101
    Table2
    ID     p_oid     b_oid     rel_oid
    1     pp101     -51     102
    2     pp102     -51     102
    3     cp102     52     102
    4     cp103     53     102
    5     cp104     54     102
    6     cp105     54     102
    From table1 I get the parent and child recs based on rel_oid=101,
    the prim_oid and sec_oid are related to another col in table2 again
    with a rel_oid. I need to get all the prim_oid that are linked to -ive b_oid
    in table2 whose child sec_oid are linked with +ive b_oid.
    In the above case, parent pp101 linked to 2 child cp102 & cp103 and
    pp102 linked to 2 child cp104 & cp105. Both pp101 and pp102 are linked
    to -ive b_oid (table2), but the children of these parents are linked to +ive b_oids.
    But pp101's children are linked to 2 diff b_oid and pp102's childrend are linked
    to same b_oid. For my requirement I can only update b_oid of pp102 with that
    of its children b_oid whereas cannot update pp101's b_oid as it children are
    linked to diff b_oid's.
    I've a sql that will return prim_oid, b_oid, sec_oid, b_oid as a record as below
    1     pp101     -51     3     cp102     52
    1     pp101     -51     4     cp103     53
    2     pp102     -51     5     cp104     54
    2     pp102     -51     6     cp105     54
    with a cursor sql that returns records as above, it would be difficult to process
    distinct parent and distinct child. So I've a cursor that returns only the parent
    records as below,
    1     pp101     -51
    2     pp102     -51
    and then for each parent I get the distinct child b_oid, if I get only one child
    b_oid I update parent else dont. but the problem is table2 has 8 million parent recs
    with link to -ve b_oid but child of only 2 million recs have link to only one distinct
    b_oid.
    If i include volume control in the cursor sql chances are all might returns like
    pp101 for which update is not required, so I should not have volume control in
    curosr sql which will now return all the 8 million record. (my assumption).
    is there any other feasible solution? Thanks

  • Max number of records in a cube & architectural issues

    Hi,
    Sorry if my question was already done but i can't find the same question with the research button (may be i have not the good words for search, I'm not english).
    I am on a BIG BIG IP project. The forecast volume of planned records is about 1.000.000.000 records a year. So we choose to split records toward many cubes. 
    1) Is there a maximum number of record supported by a cube to be planned? We planned to put 100.000.000 records maximum on a cube to be planned. Is it too much?
    2) If I make 100 cubes (one for each organizational entity) with 10.000.000 records per cube and if I "plug" a planning layout on these 100 cubes with a multiprovider, will IP:
    spread time to search the good cube to write in (due to the selected entity) in the 100 cubes (too much time !), or,
    search directly in the good cube (thanks to user exit that match cube with selected entity), so the response time will be about the same as the one for 1 layout plugged on 1 cube of 10.000.000 records?
    Thanks a lot, and sorry for my english language level
    Georges

    Hi Georges,
    Having too many records in the cube should not be very detrimental for performance of planning application as long as you can ensure that the data volume that you fetch at one go is restricted to reasonable limits, using restrictions in filters (the more restrictive the better). Take care of this while modelling both your planning functions/sequences and input-ready queries.
    I understand that you'll need to create a multiprovider for reporting purposes, but if you dont need the data from more than one cube for planning purpose, it will be better to create the aggregation level (and rest of the planning model) on top of individual cube. In case, you want to use the same planning function/input queries for multiple cubes (which will probably be the case), you can create the aggregation levels on the multiprovider but make sure you restrict the characteristic 'infoprovider' properly in the filter restrictions to avoid the function reading unnecessary data from many cubes.
    Hope this helps.

  • Max Number of records for BAPI 'BAPI_PBSRVAPS_GETDETAIL'

    Hi All,
    Can you suggest me the number of records to be fed to the 'BAPI_PBSRVAPS_GETDETAIL'.
    I am using a few location products for 9 key figures.Whenever number of records
    in selection table increases BAPI behaves in a strange way and the code written below it does not get executed.
    Please guide me to get full points.
    Thanks in Advance,
    Chandan Dubey

    Hi Uma,
        It comes out of the program after this code is executed.I have 50 location product combinations in vit_selection table.
      CALL FUNCTION 'BAPI_PBSRVAPS_GETDETAIL'
        EXPORTING
          planningbook                = planning_book
          period_type                 = 'B'
          date_from                   = l_from_week
          date_to                     = l_to_week
          logical_system              = logical_system
          business_system_group       = business_system_group
        TABLES
          selection                   = vit_selection
          group_by                    = vit_group_by
          key_figure_selection        = vit_kf_selection
          time_series                 = vit_t_s
          time_series_item            = vit_t_s_i
          characteristics_combination = vit_c_c
          return                      = vit_return.
      LOOP AT vit_return.

  • Max number of records for  'BAPI_PBSRVAPS_GETDETAIL'.

    Hi All,
    Can you suggest me the number of records to be fed to the 'BAPI_PBSRVAPS_GETDETAIL'.
    I am using a few location products for 9 key figures.Whenever number of records
    in selection table increases BAPI behaves in a strange way and the code written below it does not get executed.
    Please guide me to get full points.
    Thanks in Advance,
    Chandan Dubey

    Server memory issue !

  • Reg: Find Duplicate Records and select max number of record in Table

    Hi Guys,
         This is Nagendra, India.
    my table structure is
    id  name  tempid  temptime
    1  xxx     123        date
    1 yyy       128       date
    1 sdd     173       date
    14 ree    184      date
    14 fded   189     date
    This is Table Structure, totally 15000+ records is there.
    My Requirement is showing id and max(tempId) value.
    id  name  tempid  temptime
    1   sdd    173      date
    14  fded   189     date
    Like that, I want to show all record(after hiding duplicate values ) like that Could you please solve this issue.
    With Regards,
    Nagendra

    ; WITH numbering AS (
        SELECT id, name, tempid, temptime,
               rowno = row_number() OVER(PARTITION BY id ORDER BY tempid DESC)
        FROM  tbl
    SELECT id, name, tempid, temptime
    FROM   numbering
    WHERE  rowno = 1
    The WITH thing defines a Common Table Expression which is a locally defined view which only exists for this query. The row_number function numbers the rows, re-starting on 1 for every id, and they are numbering in falling tempid order. Thus, by
    selecting all rows from the CTE with rono = 1, we get the row with the highest tempid for each id.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • ODS Activation - Number of Records Per Data Package

    Hello,
    We are using the default Global ODS Setting called "Min Num of Data Records"... which defaults to 10,000.
    However, when I activate the ODS and look at the job log, I see that less than 10,000 records are in most of the data packages.  Below is part of the job log.  If the setting says the minimum must be 10,000 then why am i seeing data packages with less?
    Thanks!
    Data pkgs 000027; Added records 2,536; Changed records 0; Deleted records 0
    Data pkgs 000028; Added records 2,436; Changed records 0; Deleted records 0
    Data pkgs 000029; Added records 1,171; Changed records 0; Deleted records 0
    Data pkgs 000030; Added records 1,743; Changed records 0; Deleted records 0
    Data pkgs 000031; Added records 1,552; Changed records 0; Deleted records 0
    Data pkgs 000032; Added records 8,048; Changed records 0; Deleted records 0
    Data pkgs 000033; Added records 10,001; Changed records 0; Deleted records 0
    Data pkgs 000034; Added records 10,001; Changed records 0; Deleted records 0
    Data pkgs 000035; Added records 10,001; Changed records 0; Deleted records 0
    Data pkgs 000036; Added records 10,001; Changed records 0; Deleted records 0

    Hi Fong,
    The settings for 'no of data records' specifies how many records should be transfered in a data package to BW for processing .
    However added records specifies how many records of that data package actually loaded to your data target .
    This may be because of
    1.Key fields:  Your key field settings may lead to elimination of any duplicate records
    2.Update rule: Your update rule may have some filteration of records.
    Hope this makes you clear.
    Regards,
    Prema

  • Maximum number of records which can be added to custom list

    HI,
    What is the maximum number of records added to custom list, without increasing the list throttling?
    Thanks

    its two differnt thing you are asking.
    1) Max Number of Record MSFT supported is 30,000,000 per library/List
    http://technet.microsoft.com/en-us/library/cc262787.aspx#ListLibrary
    For List Throttling.
    To minimize database contention, SQL Server often uses row-level locking as a strategy to ensure accurate updates without adversely
    impacting other users who are accessing other rows.
    check this one to understand more about throttling.
    http://blogs.msdn.com/b/spses/archive/2013/12/02/sharepoint-2010-2013-list-view-lookup-threshold-uncovered.aspx
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • Issue when number of records in a DSO exceeds DTP package size

    Hi all,
    I'm having a strange problem with the package size of my DTPs.
    I move data from a DSO to another one while performing some transformations in an Export routine. The DTP has a package size of 6 million records.
    When the number of records in the first DSO exceeds the package size, some of the records seem not to be processed properly by the transformation and this forces me to choose a very large number for the package size so that the ABAP code in the transformation is processed for all the records of the source DSO.
    I can't understand why this is happening because package size is only supposed to determine the number of records to be processed in a single step and nothing else.
    Am I right?
    Thanks

    My tip would also be the summary item

  • How do I find the max number of sessions ever recorded?

    Hi,
    I see sessions_highwater column in v$license shows the max number of sessions since the database startup time, but I'm wondering if there is a view that shows the max number of sessions ever recorded? (i.e. before the last database restart)
    Thanks!
    Jon

    You can check the number of sessions from the below query using hist views. But you can only go back upto  the retention policy setting.
    SELECT   res.snap_id, s.begin_interval_time, s.end_interval_time,
             res.instance_number, res.resource_name, res.current_utilization,
             res.max_utilization
        FROM dba_hist_resource_limit res, dba_hist_snapshot s
       WHERE s.snap_id = res.snap_id AND res.resource_name = 'sessions'
    ORDER BY s.begin_interval_time, res.instance_number

  • REPORT BUILDER: MAX. NUMBER OF RECORDS PER PAGE

    Dear Consultant,
    Using DEVELOPER 2000 R2.1 REPORT BUILDER, is there any way to set the " MAXIMUM NUMBER OF RECORDS PER PAGE" dynamically by user input when calling the report, without the need to modify the report at Design time and recompile it?
    Thanks for your help.
    Ghia Haidar
    Developer
    Astrolabe IT - Professional Consultants
    Beirut, Lebanon
    Phone: + 961 1 841.050
    Fax: + 961 1 841.055 www.astrolabe.com.lb
    null

    Refer Manish question.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Ali DHAINI ([email protected]):
    Dear Consultant,
    Using DEVELOPER 2000 R2.1 REPORT BUILDER, is there any way to set the " MAXIMUM NUMBER OF RECORDS PER PAGE" dynamically by user input when calling the report, without the need to modify the report at Design time and recompile it?
    Thanks for your help.
    Ghia Haidar
    Developer
    Astrolabe IT - Professional Consultants
    Beirut, Lebanon
    Phone: + 961 1 841.050
    Fax: + 961 1 841.055 www.astrolabe.com.lb
    <HR></BLOCKQUOTE>
    null

Maybe you are looking for

  • Creation of Blank CR for MDG-S

    Hi I need to need blank CR (Without proving any value for Mandetory fields also but with CR Description and Attachment). For this I selected BP Role BBP000 and assigned to Vendor Account Group 0001 .We are using Vendor Like UI with template WS5310044

  • I have two drives in my macbook pro. How can I download my apps to my second data hard drive instead of my OS SSD drive?

    I have two drives in my macbook pro. How can I download my apps to my second data hard drive instead of my OS SSD drive? Is there a way to point the default directory in the app store to a different hard drive? Certainly for a Mac Pro this makes sens

  • NX8600GTS T2D256EZ NO HDMI Signal Single Monitor

    The above card refuses to work correctly with an LCD TV using a DVI to HDMI connector, it will not display anything using it as a single display and requires going into windows to move the display across to get it working. The adaptor, cable and TV s

  • Store Wouldn't Let Me Use Coupon?!?

    I'd appreciate any help that could be offered on this.  I went into the BB in Newington, NH yesterday to purchase the MacBook Air 13" for the advertised price of $899.99.  I recently moved and had a 10% off coupon that was with my change of address f

  • Plugin-container.exe : what is it, why is it necessary, can I delete it

    Continued from the "plugin-container.exe : what is it, why is it necessary, can I delete it" thread: 1) When people say Firefox uses up a lot of CPU/Memory, they are telling the truth. Tools like Task Manager and Process Explorer show you how much CP