Problem in data extraction

Hello,
   I loaded the deltas through 2lis_11_vascl, 2lis_11_v_ssl into the ods. since the date format is incorrect , the deltas failed.
when i see the data for the data sources in rsa3, i could not find some records, but they are present in the psa of failed delta loads. How can this be possible?
And also i went to filling up of the set up tables and have given the document number for which i have to load data and executed it.
after that i loaded data using the document no in data selection of info package and the load is successful.
why this happens?
plz sugest me.
Thanks & Regards,
Sheetal

Sheetal,
PSA Tables are created at the time you create the infosource and activate it.
I think you can not copy from one PSA to another PSA.
But if your Infopack is containing both old and new data target then in that case both target will be using the same PSA and at the time of running the infopackage you can select the new data target and update it.
If you are unable to see the new data target in your infopack try this....
Create an update rule for the new data target based on the infosource that you have used to update the old data target and activate it.
Now you must be able to see the new data target in the infopack.
Regs
Gopi.

Similar Messages

  • Problem in Data extraction for NEW GL DSO 0FIGL_O10

    Hi ,
    I am facing Problem in extraction of records from SAP to BW.
    I have installed Business Content of NEW GL DSO  0FIGL_O10.
    When I extract the Data from SAP R/3, to this DSO  ( 0FIGL_O10 )  the reocrds are getting over written
    For Example  When I go the the Mange Option ( InfoProvider Administration)  the transferred Records and the Added Records are not same.  The Added records are less then the Transfered reocords.
    This is happening becuase of Key Filed Definations.
    I have 16 Characterisics in the KEY FIELD, which the maximum that I can have. But the Data comming from is Unique in some casses.
    As result the data get added up in the DSO, hence my balances are not matching with SAP R/3 for GL Accounts.
    There are total 31 Characteristics in the Datasource (0FI_GL_10) . Of which 16 Charactheristics i can include in the Key field area.
    Please suggest some solution.
    Regards,
    Nilesh Labde

    Hi,
    For safety, the delta process uses a lower interval setting of one hour (this is the default setting). In this way, the system always transfers all postings made between one hour before the last delta upload and the current time. The overlap of the first hour of a delta upload causes any records that are extracted twice to be overwritten by the after image process in the ODS object with the MOVE update. This ensures 100% data consistency in BW.
    But u can achive ur objective in different manner::
    Make a custom info object ZDISTINCT and populate it in transformation using ABAP code. In ABAP try and compound the values from different charactersitcs so that 1 compounded characterstic can be made. Use ZDISTINCT in ur DSO as key
    Just a thought may be it can solve ur problem.
    Ravish.

  • Problem in Data extraction into BI

    Hi Guys,
    I am having a strange problem and unable to idenitfy the root cause. I am hoping that some one can help me with this issue.
    System Details:
    BI 7.0
    SAP ECC 5.0
    Current Design:
    We have created a DSO in the BI system and created a custom extractor to pull the data from SAP ECC 5.0 into the DSO.
    The custom extractor is using the option "Extraction Frm SAP Query" and using a custom infoset. This is the transaction type data source.
    Problem Statement:
    When I run the RSA3 transaction for this extractor, the extractor brings 1870 records. However when i run the Infopackage in the BI system to bring the records into the DSO --> PSA then i am only getting 823 records.
    Question:
    Why am i seeing this difference in the number of records in RSA3 and in the BI system? I am NOT using any selection or filter conditions in the Infopackage. And i am only getting the data till PSA. So i was expecting that in PSA i can see 1870 records same as that i see in RSA3. Any idea what is missing?
    Thanks in advance !!

    Hi,
    The infopackage ran for full load and not for delta load.
    I looked at the rsmo and at first glance everything looks fine. But when i look at the details tab i can see that the overall status shows as RED and it says
    Overall Status: Errors occured: or : Missing Messages
    I am not clear what this means. I dont see any other nodes in RED. Every node is in green except the top most node which is in RED and shows the above message.
    Any idea what could be the problem?

  • Problem during data extraction program (tran - RSA3) in SC system

    Hi Experts,
    We are extracting data from one of the data source (9ALS_SBC) through trans RSA3. We are using active version 000 for the same.
    but, the transaction is showing error below:
    Error when generating program
    Message no. /SAPAPO/TSM141
    Diagnosis
    Generated programs are programs that are generated based on individual data objects, such as planning object structures, planniung areas and InfoCubes. These programs are then executed in the transaction. An error occurred during the generation of such a program.
    There are two possible causes:
    The template has been corrupted.
    The object that the template uses to generate the program contains inconsistencies, for instnace an InfopCube has not been activated.
    could someone tell me how to solve this error?
    i hv corrected all the inconsistencies but still it is showing same error!
    thanks in advance for this help!
    with best rgds/
    Jay

    Dear Jay,
    the error can have several reasons, it depends on your business scenario and system settings.
    I send you some possible reasons and solutions:
    1.
    The problem can be that the technical datasource name has changed inside the planning area. The planning area has other extract structure assigned as the datasource.
    This could have happened due to a planning area or datasource transport.
    The best way to repair this is to create a dummy-datasource for the planning area with transaction /N/SAPAPO/SDP_EXTR. This will update the relevant tables with the new datasource structure.
    Please check if the other datasources will work correctly as well. If not you may create similar dummy-datasources for the other aggregates as well.
    Please do not forget to delete the dummy-datasources as well in  /N/SAPAPO/SDP_EXTR.
    2.
    You can try to re-generate the data source at transaction /SAPAPO/SDP_EXTR and recheck the issue.                   
    3.
    Run /sapapo/ts_pstru_gen via transaction se38.         
    In the selection, enter the Planning Object Structure used and select  'Planning area extractor' and set the flags on the 2 bottom checkboxes 'Reset generation time stamp' and 'Generate'. 
    4.
    Maybe the extract structure which is assigned to planning area doesn't exist in your system. This could  happend at the time of transport.             
    Please refer to following content from note 549184:                                                                               
    =======================================================================
    Q4: Why could I have extraction problem after transport of DataSource? 
    A4: DataSources for DP/SNP planning areas depend directly on the                                                                               
    structure of the planning areas. That's why the planning area MUST                                                                               
    ALWAYS be transported with or before the DataSource.                   
    ========================================================================
    To solve this inconsistency please try the below:                      
    -Please reactivate the datasoure on Planning Area.                     
    And activate all active transfer structures for a source system with  RS_TRANSTRU_ACTIVATE_ALL programme on se38 transaction.                
    I hope these proposals could solve the error.
    Regards,
    Tibor

  • Problem in Data Extracting from BEx to Webi

    Hi guys..this is sathish, we have a problem in webi reports.Actually we created webi reports from using BEx,the problem is the data in bex is showing correct data but in webi reports some cells are not showing the exact data which is showing the in the BEx.
    How to resolve the problem.please help me.

    BOXIR3.1 we are using..actually webi reports are prepared 3 months before.now the problem is in that particular report some fields are showing wrong information like Ex: in Bex it showing 1200545 but in webi report it showing like 346432.it is showing like that in 3 to 4 fields only remaining fields are correct. we tried to create new report but it showing correct information.but that particular report itself showing wrong information.
    Thanks for your reply Stratos
    Sathish
    Edited by: M.V.Sathish on Mar 27, 2011 1:26 PM

  • Problem in data extraction using 2LIS_11_VAITM

    Hi,
    For sales order item detail I am using a standard data source 2LIS_11_VAITM.
    And for this I have not written any code in exit.
    But for a particular date some of Sales Document: Item Data has pulled to BI using delta upload and some of not pulled.
    As some of sales order items are missing in BI. So i dentified those sales order detail in R/3
    So the Problem is coming for only those records which have created on 04.10.2008,27.10.2008, 01.11.2008.
    And the process chain for 05.10.2008, 28.10.2008, 02.11.2008 has not executed which suppose to pull the data for 04.10.2008,27.10.2008, 01.11.2008. The error is in delta uploading.
    On 05.10.2008, 28.10.2008, 02.11.2008  the server is down at the scheduled time of process chain.
    So the data suppose to come on next delta upload.
    So 0n 06.10.2008, 29.10.2008, 03.11.2008 the delta executed successfully.
    But all the records which are created on  04.10.2008,27.10.2008, 01.11.2008 hv not pulled to BI which I checked in PSA.
    Only some of records which are created on  04.10.2008,27.10.2008, 01.11.2008. are in PSA.
    Please suggest your idea.
    Point will be awarded for this.

    Hi,
    Are you loading to the DSO??
    Then
    1)fill the set up table for the following dates only for which you have misssed the records.
    2) Do a full repair to the targets for these date after the set up table is filled.
    3) schedule a normal delta to the further targets from this DSO.
    If loading to the cube directly??
    1)do a selective deletion from the cube for these dates
    2)fill the set up table for the following dates only
    3) Do a full repair to the targets for these date.
    No need to delete or do anything with the existing init in any case.
    Thanks
    Ajeet

  • Problem in data extractrion

    hi all,
    i am facing some problem in data extraction from r/3 system.
    for one of the sales requirement i have created a z data source based on table. its activated and properly replicated in bw system. when i try to extract data from r/3 system it has some 78125 records in total, data is posted from r/3 system comes to bw but it fails in bw system.
    under extraction tab it extracts data and shows like this ;
    Extraction (messages): Errors occurred
    Data request received
    Data selection scheduled
    78125 Records sent ( 0 Records received )
    Error occurred in the data selection
    Transfer (IDocs and TRFC): Errors occurred
    Data Package 000001 : sent, not arrived
    Data Package 000002 : sent, not arrived
    when i expand the node for message it gives below error messages;
    Error                                                 7 when sending an IDoc
    An SQL error occurred when accessing a table.
    An SQL error occurred when accessing a table.    3
    An SQL error occurred when accessing a table.    4
    Errors in source system
    but when i extract data with some restriction upto 2000 records it has extracted easily .... from that i can conclude that there is no problem with datasource, there is no rfc issue, no idoc issue problem is with data sizing....
    can anyone suggest me some paramerts to increase which can help me in removing this problem or any inputs are welcomed.
    thanks
    dipika

    same data source is working fine when i extract the data by putting some selection like sales org etc. so it seems to be that there is no problem with datasource,... but im clue less i have no idea why its not working when i am extracting full data without any selection and record numbers are also not very high.
    its only 78125 records.

  • Data Extraction.Init & Delta for datasource 0CRM_CONF_COSTS ControllingData

    Dear Experts,
    We are facing problem with Data Extraction for BW.  Prerequisites all checked and done.
    No data for Initial & Delta upload for datasource 0CRM_CONF_COSTS,
    0CRM_SRV_CO_PLAN and 0CRM_C_ITEM
    Other related data source 0CRM_SRV_COSTS and 0CRM_SERVICE_COSTS we are getting data.
    Request your help on this issue.
    Regards,
    Dilip W

    Hi krishna,
                       yes we did enhancement to this data source. In PRDOCTION its taking 27 hours time  to load data.
              For extraction from source system (CRM) its taking  nearly 24 hours time  as per JOb LOG .
    08/05/2010 11:53:08 Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 10,849 records
    08/05/2010 12:02:02 Result of customer enhancement: 10,849 records
    08/05/2010 12:02:03 PSA=0 USING & STARTING SAPI SCHEDULER
    08/05/2010 12:02:03 Asynchronous send of data package 1 in task 0002 (1 parallel tasks)
    08/05/2010 12:02:06 IDOC: Info IDoc 2, IDoc No. 1576298, Duration 00:00:01
    08/05/2010 12:02:06 IDoc: Start = 08/05/2010 10:26:37, End = 08/05/2010 10:26:38
    08/05/2010 13:02:38 Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 10,958 records
    08/05/2010 13:11:48 Result of customer enhancement: 10,958 records
    08/05/2010 13:11:52 Asynchronous send of data package 2 in task 0003 (1 parallel tasks)
    08/05/2010 13:11:53 tRFC: Data Package = 1, TID = 0AA00D0301984C5AEE8E44DB, Duration = 00:16:
    08/05/2010 13:11:53 tRFC: Start = 08/05/2010 12:02:19, End = 08/05/2010 12:18:27
    08/05/2010 14:30:13 Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 11,296 records
    08/05/2010 14:39:53 Result of customer enhancement: 11,296 records
    like this its taking 24 hours in extraction its self.
    Edited by: kotha123 on Aug 10, 2010 12:08 PM
    Edited by: kotha123 on Aug 10, 2010 12:09 PM

  • BODS 3.1 : SAP R/3 data extraction -What is the difference in 2 dataflows?

    Hi.
    Can anyone advise as to what is the difference  in using the data extraction flow for extracting Data from SAP R/3 ?
    1)DF1 >> SAPR/3 (R3/Table -query transformation-dat file) >>query transformation >> target
    This ABAP flow generates a ABAP program and a dat file.
    We can also upload this program and run jobs as execute preloaded option on datastore.
    This works fine.
    2) We also can pull the SAP R/3 table directly.
    DF2>>SAPR/3 table (this has a red arrow like in OHD) >> Query transformation >> target
    THIS ALSO Works fine. And we are able to see the data directly into oracle.
    Which can also be scheduled on a job.
    BUT am unable to understand the purpose of using the different types of data extraction flows.
    When to use which type of flow for data extraction.
    Advantage / disadvantage - over the 2 data flows.
    What we are not understanding is that :
    if we can directly pull data from R/3 table directly thro a query transformation into the target table,
    why use the Flow of creating a R/3 data flow,
    and then do a query transformation again
    and then populate the target database?
    There might be some practical reasons for using these 2 different types of flows in doing the data extraction. Which I would like to understand.  Can anyone advise please.
    Many thanks
    indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 3:25 PM

    Hi Jeff.
    Greetings. And many thanks for your response.
    Generally we pull the entire SAP R/3 table thro query transformation into oracle.
    For which we use R/3 data flow and the ABAP program, which we upload on the R/3 system
    so as to be able to use the option of Execute preloaded - and run the jobs.
    Since we do not have any control on our R/3 servers nor we have anyone on ABAP programming,
    we do not do anything at the SAP R/3 level
    I was doing this trial and error testing on our Worflows for our new requirement
    WF 1 : which has some 15 R/3 TABLES.
    For each table we have created a separate Dataflow.
    And finally in between in some dataflows, wherein, the SAP tables which had lot of rows, i decided to pull it directly,
    by-passing the ABAP flow.
    And still the entire work flow and data extraction happens ok.
    In fact i tried creating a new sample data flow and tested.
    Using direct download and - and also execute preloaded.
    I did not see any major difference in time taken for data extraction;
    Because anyhow we pull the entire Table, then choose whatever we want to bring into oracle thro a view for our BO reporting or aggregate and then bring data as a table for Universe consumption.
    Actually, I was looking at other options to avoid this ABAP generation - and the R/3 data flow because we are having problems on our dev and qa environments - giving delimiter errors.  Whereas in production it works fine. Production environment is a old set up of BODS 3.1. QA and Dev are relatively new enviornments of BODS. Which is having this delimiter error.
    I did not understand how to resolve it as per this post : https://cw.sdn.sap.com/cw/ideas/2596
    And trying to resolve this problem, I ended up with the option of trying to pull directly the R/3 table. Without using ABAP workflow.  Just by trial and error of each and every drag and drop option. Because we had to urgently do a POC and deliver the data for the entire e recruiting module of SAP. 
    I dont know whether i could do this direct pulling of data - for the new job which i have created,
    which has 2 workflows with 15 Dataflows in each worflow.
    And and push this job into production.
    And also whether i could by-pass this ABAP flow and do a direct pulling of R/3 data, in all the Dataflows in the future for ANY of our SAP R/3 data extraction requirement.  And this technical understanding is not clear to us as regards the difference between the 2 flows.  And being new to this whole of ETL - I just wanted to know the pros and cons of this particular data extraction. 
    As advised I shall check the schedules for a week, and then we shall move it probably into production.
    Thanks again.
    Kind Regards
    Indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 7:02 PM

  • Open data extraction orders -  Applying Support Packs

    Dear All,
    I have done the IDES 4.6C SR2 installation.
    While updating the support packs, i get the message saying
    CHECK_REQUIREMENTS phase.
    Open data extraction orders
    There are still open data extraction orders in the system
    process these before the start of the object import because changes to the ABAP Dictionary structures could lead to data extraction orders not being able to be read after the import and their processing terminating
    For more details about this problem, see Note 328181.
    Go to the Customizing cockpit for data extraction and start the processing of all open extraction orders.
    I have checked the Note.
    But this is something m facing for the first time.
    Any suggestion!!!
    Rgds,
    NK

    The exact message is :
    Phase CHECK_REQUIREMENTS: Explanation of the Errors
    Open Data Extraction Requests
    The system has found a number of open data extraction requests. These
    should be processed before starting the object import process, as
    changes to DDIC structures could prevent data extraction requests from
    being read after the import, thus causing them to terminate. You can
    find more information about this problem in SAP Note 328181.
    Call the Customizing Cockpit data extraction transaction and process all
    open extraction requests.

  • Urgent: Problems in Generic Extraction by Function Module

    Hi BW Gurus,
    I am new to SDN and also new to generic extraction using function module. My requirement is to extract long text(142 char) from CRM to BW as the text is not stored in database table I used function module read_text with in another ZXXX function module copy of (RSAX_BIW_GET_DATA_SIMPLE). In my extract structure I used GUID(char,32), Langu, long text(142 char) and 2 placeholders. Text can be extracted by passing STXH table fields(Tdname, Tdid, Tdobject, Tdspars) to read_text as parameters and i also need to use CRMD_ORDERADM_H field GUID(32 char) to compare 1st 32 chars of tdname(70 char) with Guid to select Guids and loop thru this Guids and for each Guid i need to append lines of text to e_t_data but as i donot know ABAP i unable to write the code for this. Through my friends help i wrote code when i check in RSA3 it is displaying the text but when i replicate into BW and load into data target in monitor the status is red with records initially but afterwards it will be red status again with 0 from 0 records for initial load again.when i check on job logs the errors i have are:
    The background job has created a job log file of 2Gb size and it is currently on a infinite loop writing entries into the SAP System Log that it cannot write to the Job log file due to “Error 22 for write/read access to a file” this is because of the datasource i have created. Please find my Function module and if anyone would please correct FM and send me that will be really great.I appreciate it in advance.
    MY Function Module is:
    FUNCTION Z_CRMORDERH_STR_TXT.
    ""Local Interface:
    *"  IMPORTING
    *"     VALUE(I_REQUNR) TYPE  SRSC_S_IF_SIMPLE-REQUNR
    *"     VALUE(I_DSOURCE) TYPE  SRSC_S_IF_SIMPLE-DSOURCE OPTIONAL
    *"     VALUE(I_MAXSIZE) TYPE  SRSC_S_IF_SIMPLE-MAXSIZE OPTIONAL
    *"     VALUE(I_INITFLAG) TYPE  SRSC_S_IF_SIMPLE-INITFLAG OPTIONAL
    *"     VALUE(I_READ_ONLY) TYPE  SRSC_S_IF_SIMPLE-READONLY OPTIONAL
    *"     VALUE(I_REMOTE_CALL) TYPE  SBIWA_FLAG DEFAULT SBIWA_C_FLAG_OFF
    *"  TABLES
    *"      I_T_SELECT TYPE  SRSC_S_IF_SIMPLE-T_SELECT OPTIONAL
    *"      I_T_FIELDS TYPE  SRSC_S_IF_SIMPLE-T_FIELDS OPTIONAL
    *"      E_T_DATA STRUCTURE  YCRM_TEXT_STR OPTIONAL
    *"  EXCEPTIONS
    *"      NO_MORE_DATA
    *"      ERROR_PASSED_TO_MESS_HANDLER
    ***"  EXCEPTIONS     NO_MORE_DATA
    *"      ERROR_PASSED_TO_MESS_HANDLER
      Tables: CRMD_ORDERADM_H, STXH.
    Auxiliary Selection criteria structure
    data: l_s_select type srsc_s_select.
    Maximum number of lines for DB table
      Statics: s_s_if type srsc_s_if_simple,
    counter
             s_counter_datapakid like sy-tabix,
    Cursor
             s_cursor type cursor.
    data: i_crmtext type standard table of TLINE .
      types: begin of xsreph ,
              GUID type CRMD_ORDERADM_H-guid,
            end of xsreph.
       data: i_guid type standard table of xsreph.
      data: I_TEXT type STXH-TDNAME.
      data: xempl like  YCRM_TEXT_STR occurs 0 with header line.
      data: t_tab like dd03l-tabname.
    Select ranges
      ranges: l_r_guid for CRMD_ORDERADM_H-guid.
             l_r_connid  for sflight-connid.
    Initialization mode (first call by SAPI) or data transfer mode
    (following calls) ?
      if i_initflag = sbiwa_c_flag_on.
    Initialization: check input parameters
                    buffer input parameters
                    prepare data selection
    Check DataSource validity
        case i_dsource.
          when 'yCRM_TEXT'.  " for S_SREPH1
          when others.
            if 1 = 2. message e009(r3). endif.
    this is a typical log call. Please write every error message like this
            log_write 'E'                  "message type
                      'R3'                 "message class
                      '009'                "message number
                    i_dsource   "message variable 1
                      ' '.                 "message variable 2
            raise error_passed_to_mess_handler.
        endcase.
       append lines of i_t_select to s_s_if-t_select.
    Fill parameter buffer for data extraction calls
        s_s_if-requnr    = i_requnr.
        s_s_if-dsource = i_dsource.
        s_s_if-maxsize   = i_maxsize.
    Fill field list table for an optimized select statement
    (in case that there is no 1:1 relation between InfoSource fields
    and database table fields this may be far from beeing trivial)
       append lines of i_t_fields to s_s_if-t_fields.
    we will do our selection based on what is in the p table for the
    infoobject
      else.                 "Initialization mode or data extraction ?
    Data transfer: First Call      OPEN CURSOR + FETCH
                   Following Calls FETCH only
    First data package -> OPEN CURSOR
        if s_counter_datapakid = 0.
    Fill range tables BW will only pass down simple selection criteria
    of the type SIGN = 'I' and OPTION = 'EQ' or OPTION = 'BT'.
        LOOP AT S_S_IF-T_SELECT INTO L_S_SELECT WHERE FIELDNM = 'GUID'.
            MOVE-CORRESPONDING L_S_SELECT TO L_R_GUID.
            APPEND L_R_GUID.
          ENDLOOP.
          case i_dsource.
            when 'YCRM_TEXT'.  " for S_SREPH1
              t_tab = 'CRMD_ORDERADM_H'.
          endcase.
          select GUID
          from (t_tab)
          into table i_guid where   PROCESS_TYPE = 'ZACI'  and ( OBJECT_ID < '0000000042').
         select tdname from stxh into i_text where tdobject = 'TEXT'.
         if sy-subrc ne 0.
           message e009(r3).
    this is a typical log call. Please write every error message like this
           log_write 'E'                  "message type
                     'R3'                 "message class
                     '009'                "message number
                     i_dsource   "message variable 1
                     'No master data found'.           "message variable 2
           raise error_passed_to_mess_handler.
         endif.
    Determine number of database records to be read per FETCH statement
    from input parameter I_MAXSIZE. If there is a one to one relation
    between DataSource table lines and database entries, this is trivial.
    In other cases, it may be impossible and some estimated value has to
    be determined.
         open cursor with hold s_cursor for
         select (s_s_if-t_fields) from CRMD_ORDERADM_H
                                  where GUID in L_R_GUID .
                                   ENDIF.
    Fetch records into interface table.
      named E_T_'Name of extract structure'.
       fetch next cursor s_cursor
                  appending corresponding fields
                  of table e_t_data
                  package size s_s_if-maxsize.
        IF SY-SUBRC <> 0.
         CLOSE CURSOR S_CURSOR.
         RAISE NO_MORE_DATA.
       ENDIF.
    as we are doing this only once can use the select statement.
    ***data: crmtext like tline occurs 0 with header line.
    **data: i_crmtext type standard table of TLINE.
    **data: i_guid type standard table of xsreph.
    data: l_guid type THEAD-TDNAME.
    data: st_guid type xsreph.
    data: st_crmtext type TLINE.
    data: lan type THEAD-TDSPRAS.
    lan = 'E'.
    loop at i_guid into st_guid.
    l_guid = st_guid-guid.
    CALL FUNCTION 'READ_TEXT'
       EXPORTING
       CLIENT                        = SY-MANDT
         ID                            = 'A002'
         LANGUAGE                      = lan
         NAME                          = l_guid
         OBJECT                        = 'CRM_ORDERH'
       ARCHIVE_HANDLE                = 0
       LOCAL_CAT                     = ' '
    IMPORTING
       HEADER                        =
       TABLES
         LINES                         = i_crmtext.
    EXCEPTIONS
       ID                            = 1
       LANGUAGE                      = 2
       NAME                          = 3
       NOT_FOUND                     = 4
       OBJECT                        = 5
       REFERENCE_CHECK               = 6
       WRONG_ACCESS_TO_ARCHIVE       = 7
       OTHERS                        = 8
    e_t_data-guid = l_guid.
    loop at i_crmtext into st_crmtext.
    move lan to e_t_data-langu.
    move st_crmtext-tdline to e_t_data-description.
    append e_t_data.
    endif.
    endloop.
    clear: st_guid,l_guid.
    refresh: i_crmtext.
    endloop.
    S_COUNTER_DATAPAKID = S_COUNTER_DATAPAKID + 1.
    endif.
    ENDFUNCTION.
    please Gurus as I donot know ABAP i appreciate if anyone would write a FM based on requirement and send me that will be really great this is my request. I gurantee of award points for good answers.
    Regards
    Kishore

    Hi,
    The statement <b>RAISE NO_MORE_DATA</b> should be active (uncommented) in your code. Otherwise, the infinte loop occurs.
    See also, the Siggi's blog:
    /people/siegfried.szameitat/blog/2005/09/29/generic-extraction-via-function-module
    BTW, was it your thread here:
    Re: Urgent: problems in extracting Long Text
    Best regards,
    Eugene

  • Steps for Data extraction from SAP r/3

    Dear all,
    I am New to SAP Bw.
    I have done data extraction from Excel into SAP BW system.
    that is like
    Create info objects > info area> Catalog
                                                --> Character catalog
                                                --> Key catalog
    Create info source
    Upload data.
    create info cube
    I need similar steps for data extraction for SAP R/3
    1. when data is in Ztables ( using Views/Infosets/Function etc)
    2. When data is with Standard SAP using Business Content.
    Thanks and Regards,
    Gaurav Sood

    hi,
    chk the links
    Generic Extraction
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/84bf4d68-0601-0010-13b5-b062adbb3e33
    CO-PA
    http://help.sap.com/saphelp_46c/helpdata/en/7a/4c37ef4a0111d1894c0000e829fbbd/content.htm
    CO-PC
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/fb07ab90-0201-0010-c489-d527d39cc0c6
    iNVENTORY
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f83be790-0201-0010-4fb0-98bd7c01e328
    Extractions in BI
    https://www.sdn.sap.com/irj/sdn/wiki
    LO Extraction:
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    /people/sap.user72/blog/2005/02/14/logistic-cockpit--when-you-need-more--first-option-enhance-it
    /people/sap.user72/blog/2004/12/23/logistic-cockpit-delta-mechanism--episode-two-v3-update-when-some-problems-can-occur
    /people/sap.user72/blog/2005/04/19/logistic-cockpit-a-new-deal-overshadowed-by-the-old-fashioned-lis
    Remya

  • 'Error 8 occurred when starting the data extraction program'

    Hello Experts,
    I am trying to pull master data (Full upload) for a attribute. I am getting an error on BW side i.e. 'The error occurred in Service API .'
    So I checked in Source system and found that the an IDOC processing failure has occurred. The failure shows 'Error 8 occurred when starting the data extraction program'.
    But when I check the extractor through RSA3, it looks fine.
    Can someone inform what might be the reason of the failure for IDOC processing and how can this be avoided in future. Because the same problem kept occurring later as well.
    Thanks
    Regards,
    KP

    Hi,
    Chk the idocs from SM58 of source system are processing fine into ur target system(BI system)...?
    Chk thru Sm58 and give all * and target destination as ur BI system and execute and chk any entries pending there?
    rgds,
    Edited by: Krishna Rao on May 6, 2009 3:22 PM

  • R/3 SP Import:Open data Extraction requests Error

    We are getting the below error when Basis doing the upgrade support packeage in R/3 source system.
    Open Data Extraction Requests
    The system has found a number of open data extraction requests. These
    should be processed before starting the object import process, as
    changes to DDIC structures could prevent data extraction requests from
    being read after the import, thus causing them to terminate. You can
    find more information about this problem in SAP Note 328181.
    Call the Customizing Cockpit data extraction transaction and process all
    open extraction requests.
    Intially we have cleared the entries in LBWQ,RSA7 and SM13.But after clearing these entries also we are getting the same above error.
    For support package upgrade in R/3 do we need to delete the SETUP table also in production environment?.
    Is there any other way around without deleting the setup table to upgrade support package in R/3 system?
    Please help us with your inputs urgently..
    Thanks
    Message was edited by:
            Ganesh S

    Thanks Siggi for the suggestion.
    We have already cleared the v3 updates by running RMBWV* job manually.After clearing all the entries in LBWQ,RSA7 and SM13 still we have got the same above error.
    When we imported In R/3 development system after deleting the setup table we didn't get the above error.But we are concerned about deleting the setup table in production system
    though already deltas are running fine.
    Is there any other way around? or deleting the setup table is the only option?
    Please help...

  • Invalid data status error during the data extraction

    Hi,
    while extracting capacity data from the SNP Capacity view to BW. i get the "invalid data status error" and the data extraction fails.
    when debugged the bad requests of the ODS object, i found that for a certain product(which has both positive and negative input and out qtys) co-product manufacturing orders were created. but this product was not marked as the co-product and functionally its fine.
    how can i rectify the data extraction problem..can you advice.
    Thanks,
    Dhanush

    Sir,
    In my company for some production order status some are having "errors in cost calculation" ie "cser" .how to deal these kind of errors.

Maybe you are looking for

  • How to get font's information from PDF of page item?

    The PDF file is placed in a document of InDesign.How to get the list of font's name from the PDF.I would like to know the font which is used in the PDF file.

  • Can't change the font on titles?

    I know the usual way to change the fonts on movie titles is to click the title in the Project window, and then a "Show Fonts" button appears in the upper left of the viewer. I have actually done this before, but now when I click a title, that button

  • How do I get to my old profile? Had to reinstall Windows due to a crash after install reinstalled Firefox lost all my bookmarks" HELP"

    Installed Windows XP from C.D. 's re downloaded Firefox. Have gone through HELP section with no luck. Was hoping to get back to old profile so i can transfer bookmarks. All efforts have failed HELP.

  • Howto make a Performance Test

    Hi, I use UCM 11g. I want to make a performance test on the customer side. That is why my plan was to upload 1.5 million files to the ucm before I start with the tests. The problem is that I don't know how to upload 1.5 million files. I wanted to use

  • Problems with the compiler?

    Hi all-, Can someone look into this?: In the below test, if I uncomment the first line, JSR014's javac fails, but CodeGuide (www.omnicore.com, features its own compiler) is OK. If I uncomment the second comment, both compilers complain: import java.u