Info package Settings

Hi Dear All,
I request some one to through some knowledge on info package setting for data loading from Source system.
I had gone to Info package-> Scheduler->DataS. Default Data Transfer
Now here I See some settings as below
Maximum size of a data packet in kByte                            20000       (Non edit)
Maximum number of dialog processes for sending data               2    (Non edit)
Number of data packets per Info-IDoc                              10               (Non edit)
Max number of data packet in K bytes = some value that can be edited (Please explain how this works and how to use)
Number of info packets for data records = some value that can be edited (Please explain how this works and how to use)
Here I have a case below for better understanding.
1. I had created a Generic data source by function module.
2. replicated in Bi
3. I had created a info package.
Note: I know that I have only 80000 records to be extracted from source system.
4. Case: a) Max number of data packet in K bytes = 20000
                   Number of info packets for data records = 10
     Result: Found that in the manage of the request no single record is added (0 From 0 Records) after 2hr also.
    Case: b) Max number of data packet in K bytes = 300
                   Number of info packets for data records = 10
    Result: it brings in 3 minutes 517(517 From 0 Records)  records and after ten sets it added to records as (5170 From 5170 Records)
Does making changes to "Max number of data packet in K bytes ' will make any changes terms of TIME.
Thanks in advance ......
It is really appreciated.......
Edited by: sreekanth goud on Feb 15, 2012 7:02 PM

Hi,
Read this OSS note for good understanding on this subject.
Note 417307 - Extractor package size: Collective note for applications

Similar Messages

  • Flatfile loading-info packager settings

    hi guys,
    Im trying to load one flatfile into infocube.
    when Im doing settings in infopackage,I came across a tabstrip with name EXTERNAL DATA....in that I didnot understand
    1.
    the two options they gave where I have to choose one.Can u explain the meaning of it.
    File is..Data file or Control file......
    what does these to options mean and where do we use them in real time...
    2.
    Currency conversion for external system ....what will happen if i choose this option...in which scenarios do we use this in real time....
    3.In DATA TARGETS tabstrip,there are some options to choose...
                         a.Update in all targets for which active rules exist
                          b.delete entire content of datatarget..
         what will happen if i choose these options.....where do we use them in real time....
    thanks and rgds
    Surekha
    Edited by: SChandx200 on Jul 3, 2009 5:47 AM

    1.File is..Data file or Control file......
    In this we have to choose Data File
    Please refer
    http://help.sap.com/saphelp_nw04/helpdata/en/80/1a6567e07211d2acb80000e829fbfe/frameset.htm
    2.
    Currency conversion for external system ....what will happen if i choose this option...in which scenarios do we use this in real time....
    This is important when we do flat file and there are Amounts with Currency.There are many currency related settings calculations that need to performed like determining the No of Decimal places for a Currency.
    For Eg.
    Whenever the amount 123 KRW is loaded in BW, it checks the TCURX table for the KRW currency. After finding the entry in TCURX table, it stores the amount value in target as 123 / {10** (2 u2013 (0))} that is 1.23 KRW. (The amount is divided by 100)
    For this to happen, we have to make a specific setting in the InfoPackage thru which we schedule the loads. The checkbox for u201CCurrency Conversion for External Systsu201D should be ticked as shown below

  • Error while transporting PCs and Info packages frm dev to qa

    Hello Guyz
    First of all, I am new to this forum. Heard a lot about it, and I hope I can learn a lot by the exchange of information in this forum. I ve already gone through so many threads, and have gained a lot of insight in BW.
    I have an issue, where I am transporting process chains and info packages from dev to qa. But the transports have thrown errors, saying that the 'Source System XXXCLNTXXX does not exist and info package has been deleted'.
    My question: How can I transport these objects in such a way that the loads which were going from ECCDEV to BWDEV, now go from ECCQA to BWQA? Are there any settings in SE09 or SE10 or somewhere in BW where I can do this?
    Thanks in advance.
    Atul.

    Hi Atul........
    There is no such settings........check this.......
    How to transport BW object
    Anyways.....I think there is some problem in the Client name.......Just check the Source system name in RSA1...........if it has been changed somehow........then run transaction BDLS and change the source system assignment........
    Regards,
    Debjani......

  • Info package in a PC failed

    Hi Experts,
    An Info Package in a PC has failed.
    When I see the status tab it shows that "Request Successfully Updated in the data target"
    But the status is Red.
    It is a Master Data
    Please help me in this regards.
    Should I Manually change the status to green ?
    Regards,
    Ashwin

    Hi Ashwin,
    In the Infopackage monitor screen, go to the details tab. In that check if the following mentioned tabs have completed successfully
    1) Extraction
    2) Transfer
    3) Processing
    If everything is green and processed sucessfully, then it means that ur request has been updated successfully in the datatarget
    Also sometimes it may occur that your Request might have run for a longer time than usual and crossed the maximum wait time settings defined..to cross check this... goto
    In the infopackage monitor screen goto Setting--> Incorret waiting time for status, and check what value is present there, if the IP run time is more than the value defined there, change it temporarily and refresh, then the request should turn to green..
    This should help i believe...
    Regards
    Deepak

  • Any log for Info Package

    Dear Experts,
    I have a info package which has been included in the process chain. It scheduled everyday.
    Now somebody changed some settings in InfoPackage  like data selection and processing tabs.How do i know that who has changed the settings.
    If i look at the info package it says ALEREMORE becoz everytime its been triggered by process chain.That means you have last change name only.
    If i want to find out last one weeks changes, HOW DO I ?
    Please help me.
    Rgds,
    Sin

    Dear Aadil,
    I have checked in the given table, Unfortunately there was only one record there which was done by ALEREMOTE(i.e Today's log). Since i told you that i need a log for entire last week.
    That means rsldpio table maintaining last changed time only.but i need complete log list for the last one week.
    Please help me
    Edited by: sinner icici on Jan 21, 2008 2:17 PM

  • Error Handling in info-package strangely doesn't work

    Dear All,
    I'm trying to load master-data from R3 containing time-dependent attributes which, I know, are sometimes overlapping. As I still want to load the correct part of data, i've set Processing - via PSA and Error Handling - "Updating valid records, reporting possible" with number of errors more than the dataset itself. However, the load still fails and what is more interesting, all records in PSA are marked as "green" whereas the errors are reported in a log.
    And what makes me wonder even more is that the same info-package from the same DS is loaded in a test system OK - with errors handled and "trashed" in PSA and remaining updated in Info-Object MD.
    Does anybody can help to resolve this strange problem as i can't honestly see the difference in global settings of test and productive systems that could allow this...
    Thank you in advance

    Hi,
    Here you are extracting data for master data with one of the attribute as time dependent . So there are two more fileds are neccessary in extraction ,i.e, 'from date' and 'to date'.
    And it seems there is overlapping of this range for two records. Make sure it in PSA with record numbers. And make sure there will be no overlapping.
    With rgds,
    Anil Kumar Sharma .P

  • Failed info package, failed request turns from red to green after 1 hour so

    hi,
    There is a failed info package and if u go to monitor screen, the request status will be red in colour but then the same request after 1 hour or so turn from red to green automatically, why is this, what could be the possible reason??????????

    Hi Lieu.
    The system automatically sets the status for the load to incorrect if the waiting time is exceeded. The load will still be running even though it might look incorrect, and that is why it turns green when it is complete (and everything is correct).
    You can set the waiting time if you go to the monitor -> choose a request -> settings -> waiting time up to status incorrect (ctrl + F9).
    Hope it helps (assign points if helpfull!).
    BR
    Stefan

  • )How can we schedule the info package daily run at 6 AM, with out useing in

    Hi
       Can any one explain 1) what is information broadcasting &how can we use this.
    2)What are the settings we have to do when we use the charts in WAD's
    3)How can we schedule the info package daily run at 6 AM, with out useing in process chains
    Thanks
    Bharath

    Hi,
      Information broadcaster :
          you can precalculate and distribute(thru mail) the query, workbook and webtemplate through online link or html file to the receipents (users).
    have a look at the below link.
         http://help.sap.com/saphelp_nw04/helpdata/en/3a/0e044017355c0ce10000000a1550b0/frameset.htm
    Infopackage scheduling:
         you can schedule the infopackage daily at your desired time .In the schedule tab ,select the start later in background and in the scheduling option give the date and time and give the period values as daily.
    Regards,
    Siva.

  • Setting in info package

    Hi,
    When we create info package on 3.x data source, we can find the setting like Separator for Thousands and  'Character Used for Decimal Point' in info package but where can I find these setting in info package when it is created on 7.0 data source?
    Thanks in advance.

    Hi,
    You will never get such option in Infopackage level for any datasource be it 3.X or 7.0 except for Extraction tab of Flat File data source.
    The option for decimal character and separator comes in the EXTRACTION tab of the Flat File data source IP only and no where else. I think you are getting confused with this
    In the flatfile data source you require these settings to be done before hand so that while extraction system can recognize the correct field values and field sequence.
    Hope this clarifies your doubts.
    Navesh
    Edited by: navsamol on Nov 16, 2011 12:10 PM

  • Delta update in Info package

    Hai all
    what is the us of delta update in info package
    I had intially loaded data from R/3 to psa using
    (INFO PACKAGE->UPDATE-->Initilize delta process)
    Then I had created to more records in  r/3 side .
    this time I scheduled using DELTA UPDATE I am getting 0 records
    then what is the use of delta update
    How can I get only delta records with out use of selection condition in DATA SELECTION of INFO PACKAGE
    thanks
    chaithanya

    Hi,
    You must specify "1" (day) in the lower limit of the delta maintenance.
    I could not post my previous reply so here a new try :
    Hi Chaithanya,
    Let me explain how the generic  delta mechanism with "date"-field works on the R/3 side:
    When doing a delta extraction, the current date is stored in table ROOSGENDLT or ROOSGENDLM (anyway not so important).
    This date will serve as a "from" date for the next delta extraction, though excluding the current value.
    This in order to prevent duplicate records to be extracted during a second delta run on the same day.
    Without extra settings, it will thus not be possible to extract two different delta's on date the same day.
    Solution :
    1. If you do not care about duplicate data (e.g. only field assignments with overwrite functionality in the transformation) goto RSO2 in change mode for your extractor and go into the generic delta maintenance screen. In the settings, enter "1" as safety interval lower limit. Like this, the delta processing mechanism will always subtract 1 day from the day stored as 'low-value'.
    In your example this means that the initialisation will transfer 4 records, the delta will transfer 6 (although only 2 were added).
    2. Use timestamps to determine delta (timestamp should be updated on savetime of record)
    I hope this helps.
    regards,
    Olav

  • Process Chain error during info package execution

    I'm trying to load master data into 0CURTYPE using process chain.
    I have created a start process and included the info package and corresponding DTP in the chain to load master data into the info object.
    On scheduling the process chain I get an error during the execution of the info package
    "Could not find code page for receiving system
    For the logical destination Q4900CL010, you want to determine the code page in which the data is sent with RFC. However, this is not currently possibe, and the IDoc cannot yet be dispatched."
    Is it relatred to some setting whihc I have missed out?
    Any help on this please?
    Thanks,

    Hi, this is a issue related to idocs in the source....
    U can contact ur BASIS team to solve this...
    From their end they have to chk the unprocessed/errored idocs in BD87...
    this issue belongs to source...
    They have to run a report:RBDAGAIN  to process the errore idoc's in the source which are in STATUS:'02'.
    This is a solution for temporarilrly basis...
    if this occurs regularly they have to change the settings in SM59..
    refer exact note for this: 784381
    rgds,

  • Status of Info packages and process changes become yellow after Unicode con

    Dear Friends,
    We have been facing very critical issue in Production system after Unicode conversion.
    Sequences of Details are provided for your reference.
    1.  We suspended all BW loads for 48(2 days) hours before Unicode starts and Handover the BW system to Basis.
    2.  Basis started doing Unicode conversion on BW and they found some standard jobs ( BI_* , Ex: BI_PROCESS_LOADING) which are released status and they assumed that these jobs would run any time so that they suspended all of them but they didnu2019t inform to BW team that they had done. We didnu2019t know that when they did that like pre-Unicode installation, during the Unicode conversion and post Unicode.
    3. Basis handover the system to BW team for Unicode testing (we called five finger testing) and we didnu2019t find any issues in the testing. Again we handover the system to Basis for post installations settings for Unicode and back up things. That time we didnu2019t release any jobs to load data to BW because Basis didnu2019t inform us that we can start loads. 
    4. After post installation is done and basis told us that you can start loads from Monday. Entire Unicode activity had been done on weekend.
    5. Before releasing all BW loads to run regularly, I found the all completed process chains status become yellow and I did further detailed analysis to find why; then found some more strange things those are as follows.
              1. Only info packages in process chains are become yellow and other process variants are fine like activation,
                   DTPu2019s, index and DB stats.
              2. Requests in PSA which were completed successfully are become yellow.
              3. Info packages in the process chains are successfully completed are not updating correct status instead of green it
                  showing yellow because of that other dependent objects are not triggering.
    We all thought that suspension of standard jobs must have been caused the problem.
    How can we change the status of IPs, PSA and process chains automatically?
    We reactivate all PCu2019s and released to run tomorrow.
    Could you please share your thoughts about the issue and help us resolve?
    Thanks in Advance,
    Nageswar.

    Nageswar,
    A couple of questions :
    1. If your previous  PC runs were in yellow status - it would not impact furute PC runs..
    2. If your IPAKs are in yellow - then you might have an issue with deltas happening since the last status would have been shown as incomplete.
    3. Suspension of the standard jobs will not affect the status since the jobs were suspended but not the actual loads i.e the loads were not stopped halfway but only the released jobs were stopped - you might have to schedule the process chains again to get back the old schedule.
    Usually in a Unicode conversion - the data is exported out of the DB and imported back - this might have caused some inconsistency in the database for the RS tables used for the monitors.
    Which version of BW are you in ..?
    were there any errors in the export / import procedure for unicode conversion ?
    As part of testing did you check if loads went through properly ..?
    Was stats run on the tables after unicode ?
    Also if possible regenerate all the indices on the files.
    Rerun stats and indices for all RS* tables

  • Data Load (Info package and DTP) performance

    Hi,
    We are planning to perform initialisation expecting 30 million records.
    To improve the data load performance i am planning to increase data package size from default value to double the default value
    Will that improve the performance

    Adjusting the info package size would be helpful...
    But don't go with INCREASING the numbers, It's all depends on the your extract structure size, source system memory settings.
    I would suggest to do some trail in DEV/QUALITY systems with enough data.
    I remember playing in production itself for CRM - 0BPARTNER extractor - we actually reduced the number of records to be extracted per package to improve the performance - we did see the impacts.
    First calculate how much is your current package size by number of records extracted per package now with your extract structure size..and then play around..sure it will help unless you DON'T have any excessive ABAP code user exits.
    Let us know if you need further help on this

  • Info package ABAP Routine to populate date from and To filed

    Hello Experts,
    I have requirement to populate the From and To dates with the: T-1 to T (Current Date u2013 1 to Current Run Date aka Sy-Datum) at info package level.I have written a code for this but see that From filed is not getting populated but TO filed is filled with current date.Can someone please tell me wats wrongwith my code?It is a delta info pacakage.
    data: l_idx like sy-tabix.
    read table l_t_range with key
         fieldname = 'CPUDT'.
    l_idx = sy-tabix.
      l_t_range-sign = 'I'.
      l_t_range-option = 'BT'.
      l_t_range-low = sy-datum - 1.
      l_t_range-High = sy-datum.
    modify l_t_range index l_idx.
    p_subrc = 0.
    Thanks,

    Was able to tackle this using the following code.
    data: l_idx like sy-tabix.
    data: V_date type sy-datum.
    v_date = sy-datum - 1.
    read table l_t_range with key
         fieldname = 'CPUDT'.
    l_idx = sy-tabix.
      l_t_range-sign = 'I'.
      l_t_range-option = 'BT'.
      l_t_range-low = v_date.
       l_t_range-High = sy-datum.
    modify l_t_range index l_idx.
    p_subrc = 0.
    Thanks,
    I am closing this thread

  • Info Package ABAP Routine

    Hi
    For FIGL_4 there is both open items and closed items. i want to filter only Open Items Accounting doc no's records for Customers,Vendors,GL account at info package level.
    There are 3 ODS's in BW side for Open items Customers,Vendors,GL accounts. i want to filter open items based on these 3 ODS's Active data table.But i am getting error when i execute the request in Info Package.
    *Error Message "For sel. field 'BELNR ', no selection with SIGN = ' '; OPTION ' ' allowed     RSM1     607     *
    Below is the Info package Routine code
    program conversion_routine.
    Type pools used by conversion program
    type-pools: rsarc, rsarr, rssm.
    tables: rssdlrange.
    Global code used by conversion rules
    $$ begin of global - insert your declaration only below this line  -
    TABLES: ...
    TABLES: /BIC/AZOCFFIGL00,/BIC/AZOCFBSID00,/BIC/AZOCFBSIS00
            ,/BIC/AZOCFBSIK00.
    DATA:   ...
    TYPES : BEGIN OF TY_ZOCFBSID,
            /BIC/ZMCFADCNO LIKE /BIC/AZOCFBSID00-/BIC/ZMCFADCNO,
            END OF TY_ZOCFBSID.
    TYPES : BEGIN OF TY_ZOCFBSIS,
            /BIC/ZMCFADCNO LIKE /BIC/AZOCFBSIS00-/BIC/ZMCFADCNO,
            END OF TY_ZOCFBSIS.
    TYPES : BEGIN OF TY_ZOCFBSIK,
            /BIC/ZMCFADCNO LIKE /BIC/AZOCFBSIK00-/BIC/ZMCFADCNO,
            END OF TY_ZOCFBSIK.
    DATA : TB_ZOCFBSIS TYPE TABLE OF TY_ZOCFBSIS WITH HEADER LINE,
           WA_ZOCFBSIS TYPE TY_ZOCFBSIS.
    DATA : TB_ZOCFBSID TYPE TABLE OF TY_ZOCFBSID WITH HEADER LINE,
           WA_ZOCFBSID TYPE TY_ZOCFBSID.
    DATA : TB_ZOCFBSIK TYPE TABLE OF TY_ZOCFBSIK WITH HEADER LINE,
           WA_ZOCFBSIK TYPE TY_ZOCFBSIK,
           FLAG TYPE I.
    $$ end of global - insert your declaration only before this line   -
        InfoObject      = ZMCFADCNO
        Fieldname       = BELNR
        data type       = CHAR
        length          = 000010
        convexit        = ALPHA
    form compute_BELNR
      tables   l_t_range      structure rssdlrange
      using    p_infopackage  type rslogdpid
               p_fieldname    type rsfnm
      changing p_subrc        like sy-subrc.
          Insert source code to current selection field
    $$ begin of routine - insert your code only below this line        -
    data: l_idx like sy-tabix.
              read table l_t_range with key
                   fieldname = 'BELNR'.
              l_idx = sy-tabix.
           SELECT
           /BIC/ZMCFADCNO
           FROM /BIC/AZOCFBSID00
           INTO TABLE TB_ZOCFBSID.
            SELECT
            /BIC/ZMCFADCNO
            FROM /BIC/AZOCFBSIK00
            INTO TABLE TB_ZOCFBSIK.
            SELECT
            /BIC/ZMCFADCNO
            FROM /BIC/AZOCFBSIS00
            INTO TABLE TB_ZOCFBSIS.
           LOOP AT TB_ZOCFBSID INTO WA_ZOCFBSID.
           l_t_range-sign = 'I'.
           l_t_range-option = 'EQ'.
           l_t_range-low = WA_ZOCFBSID-/BIC/ZMCFADCNO.
           Append l_t_range.
           CLEAR l_t_range.
           ENDLOOP.
           LOOP AT TB_ZOCFBSIK INTO WA_ZOCFBSIK.
           l_t_range-sign = 'I'.
           l_t_range-option = 'EQ'.
           l_t_range-low = WA_ZOCFBSIK-/BIC/ZMCFADCNO.
           Append l_t_range.
           CLEAR l_t_range.
           ENDLOOP.
           LOOP AT TB_ZOCFBSIS INTO WA_ZOCFBSIK.
           l_t_range-sign = 'I'.
           l_t_range-option = 'EQ'.
           l_t_range-low = WA_ZOCFBSIS-/BIC/ZMCFADCNO.
           Append l_t_range.
           CLEAR l_t_range.
           ENDLOOP.
              modify l_t_range index l_idx.
              p_subrc = 0.
    $$ end of routine - insert your code only before this line         -
    Edited by: AtulMohan Mishra on Dec 27, 2010 1:11 PM

    Hi
    Now i am facing an error during data extraction from Source system
    error message in BW system "Error occurred in the data selection " Data Package 1 : arrived in BW ; Processing : 2nd processing step not yet finished
    when i checked in  SM37 (R3 system) i found the job has been cancelled and Log of this job says
    "Asynchronous sending of InfoIDOCs 2 in task 0001 (0 parallel tasks) DATASOURCE = 0FI_GL_4 
    ABAP/4 processor: SAPSQL_WHERE_CANT_SCAN                                   
    Job cancelled      "   
    it seems to me the Problem in Routine at Info Package level.
    My Requirement is to fetch only those Accounting Doc. No. RECORDS from 0FI_GL_4 data source which are present in Active data table of BSID,BSIK,BSIS ODS's.
    So i want to put logic at Info Package routine .
    1) fetch all Accounting Doc. No.  from Active data table of BSID,BSIK,BSIS Ods's and Match those Accounting Doc. No.'s with the 0FI_GL_4 data source Accounting Doc. No's.if a Accounting Doc. No. match with 0FI_GL_4 data source Accounting Doc no then that corresponding record from 0FI_GL_4 will go to the ODS
    Regards
    Atul

Maybe you are looking for