Lo-Extraction Job Failed in SAP ECC 6.0

Hello,
My Lo-extraction job failed yesterday because of this I'm unable to extract the delta data from SAP ECC 6.0 to BI. When I extract the Delta data from 2LIS_11_VAITM to BI there is no data because of Job failed in SAP ECC 6.0.
How to get the data back to 2LIS_11_VAITM data source?
Thanks

Error log display as follows
If the error occures in a non-modified SAP program, you may be able to
find an interim solution in an SAP Note.
If you have access to SAP Notes, carry out a search with the following
keywords:
"MESSAGE_TYPE_X" " "
"SAPLMCEX" or "LMCEXU14"
"MCEX_UPDATE_11"
Actually I add ZZ fields to structure if that is the case do i need to delete the  data  from MCEX11 ?
I delete the data from setup tables but i did not delete the LUW entries?
When do we delete the LUW entries?
If the structure changes do I need to cancel the old job and re-schedule job?
Thanks

Similar Messages

  • Set up for Extraction the data from SAP ECC system

    Hi!
    I would like to analyze the transaction and master data from SAP ECC system into SAP BI system.
    I have created the connection (entry for SAP ECC system) within SAP NetWeaver System.
    Unfortunately when I try to load the transaction data for Info source the process does not end successfully.
    There are no information within SM58 on source system.
    The ST22 does not contain some suitable information.
    Question:
    Which settings do I miss and how can solve my problem?
    (user authorization, etc.)
    Thank you very much!
    regards
    Holger

    Hi!
    many thanks for your reply.
    The error I get:
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source
    Can you please help me to identify the error?
    Thank you!

  • Jobs failing in SAP

    Hi,
    SAP re-org, performance monitor  and TMS jobs are failing continuously in the system with the message
    'No logon possible (no hw ID received by mssg server)'.
    what could be the possible reason for this and the steps to correct this?
    Thanx
    Balaji Srinivas Singh

    Hello,
    As per your update you have recently refreshed the system from Production.
    After that as Post-System copy process, have you installed license again as per your system.
    As yoir system is of release 4.7, you must be having normal license key and not the digital license.
    Hence please check in transaction "SLICENSE" in your SAP system, it will show you license details of your system.
    Also you can login to OS level with sidadm user (where sid is SID of your system).
    if your OS is unix, then give command
    saplicense -get
    This will show you hardware key of your system.
    The command
    saplicense -show
    will give you details about license installed in your system.
    If your OS is Windows, open CMD, and then you can use above two commands.
    Please let us know if this wont solve your problem, you can put error details.
    Regards,
    Abhay

  • Master data extraction from SAP ECC

         Hi All,
    I am a newbie here and teaching myself SAP BI. I have looked through the forums regarding master data extraction from SAP ECC in all forums but could not answers for my question. Please help me out.
    I want to extract customer attributes from SAP ECC. i have identified the standard data source 0customer_attr and replicated in SAP BI. I have created infopackage for full update. I validated the extractor checker(RSA3) and found 2 records for 0customer_attr.
    When I run the info package, the info package remains in yellow state until it times out. Please let me know in case i am missing anything, Please let me know if there is any other process for master data extraction similar to transaction data extraction.

    Hi All,
    i did the below and afte clicking execute in the Simple job, it takes me back to the Monitor Infopackage screen.
    From your info pack monitor --> menu environment-->job overview--> job in the source system--> prompts for source system(ECC) user id and password, enter them, you will get your job at SM37(ECC), select job and click on log details icon. there you can see details about your error. please share that error screen shot.
    Please find the screenshots.
    I did an Environment -->Check connection and found the below,

  • Is there any Data Mapping document Between S&OP ( model that uses supply planning operator ) and SAP ECC that I will help the client in Data Mapping activity.

    Hello All,
    Is there any Data Mapping document Between S&OP ( model that uses supply planning operator ) and SAP ECC that I will help in Data Mapping activity.
    Thanks,
    Mownesh

    There are standard templates in HCI data sources.
    e.g. 1) Customer Master data template is SOP_MD_CustomerMaster for extracting master data from SAP ECC and load it to S&OP
    KNVP is the table for customer in ECC from that you can select the fields as required
    KUNNR for customer Number
    ADRNR for Address
    List of a few commonly used table names of ECC:
    Product Related:
    MARA – Material Master (MATNR)
    MARC – Material Master with Plant Data (MATNR, WERKS)
    MARD – Material Master with Storage Location Data (MATNR, LGORT, WERKS)
    MAKT – Material Master Material Descriptions (MATNR, MATKL)
    MBEW – Material Valuation Data (MATNR, BWTAR)
    MVKE – Material Master : Sales related Data
    MDKP, MDTB – MRP related Data( Header, Item)
    MCHA, MCHB – Material Batches (Header, Item) (MATNR, WERKS, LGORT, CHARG)
    Vendor/Supplier related:
    LFA1 – vendor data (LIFNR)
    LFB1 --  Company Code Segment : Vendor Data(LIFNR, BUKRS)
    LFC1 --  FI Related Vendor Data (LIFNR, BELNR)
    LFM1 – Pur. Orgn. Related Vendor Data (LIFNR, EKORG)
    PReq/PO, BOM Related:
    EBAN – Pur. Req. Data( BANFN, BNFPO, BADAT, MATNR)
    EINA – Purchase Info. Record(General Data)(INFNR, MATNR, LIFNR)
    EINE – Purchase Info. Record (pur. Orgn. Data) (INFNR, EKORG)
    ELBK, ELBN, ELBP – Vendor Evaluation Related Data
    EKKO – PO Data (Header) (EBELN, BSTYP, BSART)
    EKPO – PO Data (Item) (EBELN, EBELP, MATNR)
    Pur. Req., RFQ and PO are differentiated by Doc Type (BSTYP) in EKKO table.
    For RFQ it is ‘A’ and for PO it is ‘F’
    MKPF – GRN Data (Header) (EBELN, BLDAT, BUDAT, XBLNR, BKTXT)
    MSEG – GRN Data(Item) MBLNR, BWART, LIFNR, MATNR, EBELN)
    Apart from this there are lot of tables which begin with ‘M’ & ‘E’, but we
    use the following very often.
    EQUK – Quota (Header)(QUNUM, MATNR)
    EQUP – Quota (Item) (QUNUM, QUPOS, LIFNR)
    EKBE – PO History Data (EBELN, EBELP, BELNR, BLDAT, MATNR, VGABE)
    EKBZ – PO History with Delivery Costs(EBELN, BELNR, LIFNR, XBLNR)
    EKET – Schedule lines data of a PO(EBELN, EINDT, SLFDT)
    EKES – Vendor Confirmations Data (EBELN, EBTYP, EINDT, XBLNR)
    T163F – Confirmation Texts (EBTYP, EBTXT)
    T156 – Movement Types (BWARE)
    T024 – Purchasing Groups
    T024E – Purchase Organizations
    T163 – Item Category’s in Purchasing Documents(PSTYP)
    T149D – Valuation Types
    T134 – Material Types
    FVLK – Delivery Types
    STKO, STPO – BOM(Bill Of Material) related Data (Header & Item)
    STPU, STPN, STST, STZU – BOM Related Tables
    RKPF, RBKP, RSEG (Header & Item) – MM – FI Related Data
    KONO, KONH – Pricing data
    T006 – Basic Unit Of Measurements
    Customer/Sales Order Related:
    VBAK : Sales Document(Header Data) (VBELN)
    VBAP : Sales Document(Item Data) (VBELN, POSNR, MATNR, ARKTX, CHARG)
    Enquiry, Quotation, Sales Order are differentiated based on Doc.
    Type(VBTYP Field) in VBAK, VBAP Tables for Enquiry VBTYP = A, for Quotation ‘B’ & for Order it is ‘C’.)
    LIKP : Delivery Table(Header Data) (VBELN, LFART, KUNNR, WADAT, INCOL)
    LIPS : Delivery Table(Item Data)(VBELN, POSNR, WERKS, LGORT, MATNR, VGBEL)
    (LIPS – VBGELN = VBAK- VBELN, LIPS-VGPOS = VBAP-POSNR)
    VTTK : Shipment Table(Header Data) (TKNUM)
    VTTP : Shipment Table (Item Data)(TKNUM, TPNUM, VBELN)
    (VTTP – VBELN = LIKP – VBELN)
    VBRK : Billing Table(Header Data) (VBELN, FKART, BELNF)
    VBRP : Billing Table(Item Data) (VBELN, POSNR, FKIMG, NEWR, VGBEL, VGPOS)
    (VERP – AUBEL = VBAK- VBELN, VBRP – VBEL = LIKP – VBELN)
    Apart from these tables there are lot of other tables which starts with ‘V’, but we use the
    following tables frequently.
    VBUK: All Sales Documents status & Admn. Data(Header) (VBELN, VBTYP)
    VBTYP = ‘C’ (Sales Order) VBTYP = ‘L’(Delivery) VBTYP = ‘M’(Invoice)
    VBUP: Sales Documents status & Admin. Data(Item) (VBELN, POSNR)
    VBEP : Sales Document Schedule Lines Data (VBELN, POSNR, EDATU, WMENG)
    VBKD: To get sales related Business data like Payment terms etc.(VBELN, ZTERM)
    VBFA: Sales Document flow data(VBELV, VBELN, POSNV, VBTYP)
    VBPA: Partner functions Data(VBELN, PARVW, KUNNR, LIFNR)
    TVLKT: Delivery Type: Texts(LFART, VTEXT)
    KNA1, KNB1, KNC1 : Customer Master Data and Other Partner’s Data(KUNNR,
    NAME1,LAND1)
    KNVK: Customer Master Contact Person(PARNR, KUNNR)
    KNVV: Customer Master Sales Data.
    LFA1, LFB1, LFC1: Vendor Master Data(To get Transporter data)(LIFNR, NAME1, ORT01)
    MARA, MARC, MARD : Material Master Data(Basic, Plant, St. Location Views)
    TVKO: Sales Organizations(VKORG)
    TVKOV: Distribution Channels(VTWEG)
    TVTA: Divisions(SPART)
    TVKBZ: Sales Office(VKBUR)
    TVBVK: Sales Group(VKGRP)
    T077D: Customer Account Group(KTOKD)
    T001W: Plants(WERKS)
    T001L: Storage Locations(LGORT)
    TWLAD: To get address of Storage Location and Plant(LGORT, ADRNR)
    TVAU: Sales Document (Order) Types
    KONV: Condition Types (pricing) (KNUMV, KSCHL, KWETR)
    T685T: Condition Types Texts.
    ADRC: To get Addresses of Partners
    VBBE, VBBS: Sales Requirements Data
    VBKA: Sales Activities Data
    VBPV: Sales Document Product Proposal
    Based on the functionality you can search ECC table names and fields
    Hope this information is helpful for you.
    Thanks and Regards,
    Anjali

  • Extracting delta's with ODP-data source 0MATERIAL_ATTR from SAP ECC

    We are using ODP-replication to extract data from SAP ECC to a MS SQL-database by SAP BusinessObjects Data Services.
    For extracting Material-data we use the standard data source 0MATERIAL_ATTR in SAP ECC.
    This data source delivers a full load with all Materials in it.
    In the properties of this data source (tab attributes) we saw that this data source is Delta_Extractor_Capable.
    Therefore we set the extraction mode from 'Query' to CDC (= Change Data Capable) and did a new initial load to get all Materials again.
    After changing a Material in ECC, we did a delta-load, but unfortunately we received ZERO records.
    Is there anyone with experience on extracting delta's with data source 0MATERIAL_ATTR or another standard masterdata data source in ECC?
    Thank you for your reaction in advance,
    Best regards,
    Jan-Hendrik

    Hi,
    You can check the delta records in RSA3 see teh article. in https://wiki.sdn.sap.com/wiki/display/profile/Surendra+Reddy
    Checking the Data using Extractor Checker (RSA3) in ECC Delta Repeat Delta etc...
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/80f4c455-1dc2-2c10-f187-d264838f21b5&overridelayout=true 
    Data Flow from LBWQ/SMQ1 to RSA7 in ECC and Delta Extraction in BI
    http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/d-f/data%20flow%20from%20lbwq_smq1%20to%20rsa7%20in%20ecc%20and%20delta%20extraction%20in%20bi.pdf
    Delete the Queue in LBWQ/RSA7 and then freshly load Init and Check the Delta.
    Once you load the Init then if there are no new records then it won't get any data. so 0 records
    Check in SM37 in ECC why the job is terminated. See in ST22 also.
    Thanks
    Reddy

  • Loading of transaction data from SAP ECC system failed

    Hi!
    I successfully connected SAP ECC system to SAP BI system.
    The following steps have been executed:
    - user ALEREMOTE with max. authorization
    - RFC destination
    - Distributing Data model
    - Generated Partner profile
    - Maintaining message types in WE20
    Now when I try to load any data from SAP ECC system the loading process in hanging in status "yellow" and never comletes.
    [0FI_AR_4|http://www.file-upload.net/view-1447743/0FI_AR_4.jpg.html]
    The following steps within Load process are yellow:
    Extraction (messages): Missing messages
      Missing message: Request received
      Missing message: Number of sent records
      Missing message: Selection completed
    Transfer (IDocs and TRFC): Missing messages or warnings
      Request IDoc : Application document posted (is green)
      Data Package 1 : arrived in BW ; Processing : 2nd processing step not yet finished
      Info IDoc 1 : sent, not arrived ; Data passed to port OK
      Info IDoc 2 : sent, not arrived ; Data passed to port OK
      Info IDoc 3 : sent, not arrived ; Data passed to port OK
      Info IDoc 4 : sent, not arrived ; Data passed to port OK
    Subseq. processing (messages) : Missing messages
        Missing message: Subseq. processing completed
        DataStore Activation (Change Log) : not yet activated
    Question:
    Can some one give me some technical steps (tcode, report) to solve this problem?
    Thank you very much!
    Holger

    Hi!
    Many thanks for your answer.
    Via BD87 on BW system I detect that all the IDOC's (type: RSRQST) will be received from SAP ECC system.
    Via tcode SM58 I could not detect any entries.
    However the loading status from yesterday is set to "red".
    The errors are:
    Extraction (messages): Missing messages
    Data Package 1 : arrived in BW ; Processing : 2nd processing step not yet finished
    Info IDoc 1 : sent, not arrived ; Data passed to port OK
    Info IDoc 2 : sent, not arrived ; Data passed to port OK
    Can you investigate my issue again?
    Thank you very much!

  • Extracting Values of a Field from a Database Table in SAP ECC System

    Hi,
    I downloaded Extracting Values of a Field from a Database Table in SAP ECC System Using MII 12.0
    senario from sdn. I'm trying to do that senario in MII 12.05. But I have problem with section 6 in page 7 (you can supply senario from sdn)
    "6- Under the loop of Repeater, use action u2018Rowu2019 to append just the string part of the WA which will display only values for field u2018Batchu2019"
    I did not find WA elemen in Output element of Repeater_0
    How can I create WA element?
    Thanks.

    Cemil,
    Set up a SAP JCo Interface action block.  Use the RFC name RFC_READ_TABLE.
    In the link editor map the table to "MARA", set RowCount to something small (20 is good sample size) and create an xml transaction property named FIELDS and copy the following into it:
    <?xml version="1.0" encoding="UTF-8"?><FIELDS>
          <item>
            <FIELDNAME>MATNR</FIELDNAME>
            <OFFSET/>
            <LENGTH/>
            <TYPE/>
            <FIELDTEXT/>
          </item>
          <item>
            <FIELDNAME>MTART</FIELDNAME>
            <OFFSET/>
            <LENGTH/>
            <TYPE/>
            <FIELDTEXT/>
          </item>
          <item>
            <FIELDNAME>BSTME</FIELDNAME>
            <OFFSET/>
            <LENGTH/>
            <TYPE/>
            <FIELDTEXT/>
          </item>
          <item>
            <FIELDNAME>XCHPF</FIELDNAME>
            <OFFSET/>
            <LENGTH/>
            <TYPE/>
            <FIELDTEXT/>
          </item>
          <item>
            <FIELDNAME>DATAB</FIELDNAME>
            <OFFSET/>
            <LENGTH/>
            <TYPE/>
            <FIELDTEXT/>
          </item>
        </FIELDS>
    Then link the Transaction.FIELDS to SAP_JCo_Interface_0.Request{/RFC_READ_TABLE/TABLES/FIELDS}.  You may run into problems with two other fields and optionally they can be removed (set link type to remove xml).  I usually remove them initially for testing.  The two fields are:
    SAP_JCo_Interface_0.Request{/RFC_READ_TABLE/INPUT/NO_DATA}
    SAP_JCo_Interface_0.Request{/RFC_READ_TABLE/INPUT/DELIMITER} (or you can set this to something like a semicolon,";" or tilda,"~".  I find it easier to caclulate position by length, but that is my own idiosyncrasy.)
    Once you get this one working, we can explore how to do filtering on the dataset.  Your output should be something like this:
    <?xml version="1.0" encoding="utf-8"?>
    <RFC_READ_TABLE>
      <INPUT>
        <DELIMITER />
        <NO_DATA />
        <QUERY_TABLE>MARA</QUERY_TABLE>
        <ROWCOUNT>20</ROWCOUNT>
        <ROWSKIPS>0</ROWSKIPS>
      </INPUT>
      <TABLES>
        <DATA>
          <item>
            <WA>000000000000000023ROH 00000000</WA>
          </item>
          <item>
            <WA>000000000000000038HALB 00000000</WA>
          </item>
          <item>
            <WA>000000000000000043HAWA 00000000</WA>
          </item>
          <item>
            <WA>000000000000000058HIBE 00000000</WA>
          </item>
          <item>
            <WA>000000000000000059HIBE 00000000</WA>
          </item>
          <item>
            <WA>000000000000000068FHMI 00000000</WA>
          </item>
          <item>
            <WA>000000000000000078DIEN 00000000</WA>
          </item>
          <item>
            <WA>000000000000000088FERT 00000000</WA>
          </item>
          <item>
            <WA>000000000000000089FERT 00000000</WA>
          </item>
          <item>
            <WA>000000000000000098HALB 00000000</WA>
          </item>
          <item>
            <WA>000000000000000170NLAG 00000000</WA>
          </item>
          <item>
            <WA>000000000000000178NLAG 00000000</WA>
          </item>
          <item>
            <WA>000000000000000188NLAG 00000000</WA>
          </item>
          <item>
            <WA>000000000000000288HALB 00000000</WA>
          </item>
          <item>
            <WA>000000000000000358HAWA 00000000</WA>
          </item>
          <item>
            <WA>000000000000000359HAWA 00000000</WA>
          </item>
          <item>
            <WA>000000000000000521HAWA 00000000</WA>
          </item>
          <item>
            <WA>000000000000000578FERT 00000000</WA>
          </item>
          <item>
            <WA>000000000000000597HAWA 00000000</WA>
          </item>
          <item>
            <WA>000000000000000598VERP 00000000</WA>
          </item>
        </DATA>
        <FIELDS>
          <item>
            <FIELDNAME>MATNR</FIELDNAME>
            <OFFSET>000000</OFFSET>
            <LENGTH>000018</LENGTH>
            <TYPE>C</TYPE>
            <FIELDTEXT>Material Number</FIELDTEXT>
          </item>
          <item>
            <FIELDNAME>MTART</FIELDNAME>
            <OFFSET>000018</OFFSET>
            <LENGTH>000004</LENGTH>
            <TYPE>C</TYPE>
            <FIELDTEXT>Material Type</FIELDTEXT>
          </item>
          <item>
            <FIELDNAME>BSTME</FIELDNAME>
            <OFFSET>000022</OFFSET>
            <LENGTH>000003</LENGTH>
            <TYPE>C</TYPE>
            <FIELDTEXT>Purchase Order Unit of Measure</FIELDTEXT>
          </item>
          <item>
            <FIELDNAME>XCHPF</FIELDNAME>
            <OFFSET>000025</OFFSET>
            <LENGTH>000001</LENGTH>
            <TYPE>C</TYPE>
            <FIELDTEXT>Batch management requirement indicator</FIELDTEXT>
          </item>
          <item>
            <FIELDNAME>DATAB</FIELDNAME>
            <OFFSET>000026</OFFSET>
            <LENGTH>000008</LENGTH>
            <TYPE>D</TYPE>
            <FIELDTEXT>Valid-From Date</FIELDTEXT>
          </item>
        </FIELDS>
        <OPTIONS />
      </TABLES>
    </RFC_READ_TABLE>
    Add a repeater sourced on:
    SAP_JCo_Interface_0.Response{/RFC_READ_TABLE/TABLES/DATA/item}
    Link your repeater output to a tracer with this:
    Repeater_0.Output{/item/WA}
    What you will see in each tracer message is a single line of data with all the fields contents concatenated together.  You can look up what each field in the string represents by the length of the field as returned in the Response segment of the RFC_READ_TABLE rfc.  Then you can parse out the data you are interested in.
    Give this a try and let me know how you succeeded.
    By the way, I could not find the scenario you referred to.  Can you post a link?
    Regards,
    Mike
    Edited by: Michael Appleby on Jan 12, 2009 5:16 PM

  • SAP ECC 6.0 - Using BAPIs for C# to extract data out

    I have had a client recommend a strategy of extracting data usin C# (Visual Studio 2005) to extract data from SAP ECC using BAPIs.
    We do have an existing methodology in place using flat files to extract data to non SAP systems
    We have purchase XI which we intend to implement next year.
    I basically wanted to keep things tidy and continue with the flat file interface strategy just to keep things simple going forward to XI.
    THen there is also the security piece, the developer has had to slowly build a profile / roles that alows him to come in thru visual studio to access these BAPIs
    I am looking for anyone who has had experience with this and what your experiences are.  As well, what are the SAP best practicies concerning this strategy
    Thank YOU ALL who will reply

    I think that SAP's strategy toward exposing BAPI is leaning toward leveraging their AS-JAVA stack.  With the AS-JAVA Enterprise Services and standard delivered web content, SAP is already exposing many existing BAPI/Business Functionality outside of the ABAP world.   However, that's not saying that traditional flat files have no place in the future.  Matter of fact, I don't see flat/delimited files going away any time soon.
    It all really comes down to what you are doing, you may even end up with a landscape that's a mix of files/RFCs.  Here inhouse, we leverage file based transfers for large data sets such as SAP BI OHS extracts, and daily R/3 FI extracts.  Why file?  because it's simple, easy to control, and you can see the data being transferred.  A file can be "hold in my hands" if you will, at best there's just a few characters in-flight.
    As for RFCs, any type of RFCs in fact (Anything from sap .net connector to AS-JAVA ESOA Based ES), are basically designed for transactional data.  Here, you are looking at things that are high in transactional count but small in individual size.  Your client's call using C#, it all really depends on how you are doing it.  MOST IMPORTANTLY:  Make sure you are using a proven, standarized, and SAP supported way, PERIOD.  NEVER EVER go into production with a "hack".
    For the high-transactional-count-small-individual transactions, we leverage BizTalk Server (BTS) here.  I'll be frank, we are not using BTS because we wrote it.  We are using BTS because it just works better for us, in our environment.  BTS now support SAP 2.0 connector (registered program ID) and SAP 3.0 (WCF based, direct SAP RFC call, supported by Microsoft AND SAP).  We are leveraging these adapters and BTS (as distribution and transformation) in our environment more and more.
    So, long story short:  Determine the transaction type, multiple solutions is ok, make sure the solution is supported!

  • Real time job doesn't receive automatically changes in SAP ECC 6

    Hi Experts,
    <br/>
    <br/>I'm trying to (using Data Integrator) automatically retrieve data from SAP ECC and store it in a local database table. I followed the steps written in this article: http://wiki.sdn.sap.com/wiki/display/BOBJ/Receiving+IDOCs. I'm modifying some records in the Cost Center Master Data (CSKS) and using COSMAS01 as the IDOC in which I'm trying to send the information.
    <br/>
    <br/>All seems to be OK, since I manually sent a Cost Center data row using the /nbd16 transaction, SAP ECC displayed a message telling that an IDOC had been generated. Also, when I checked the IDOC status in BD87 it had status 03, meaning that it had been sent correctly.
    <br/>
    <br/>In Data Services when I clicked on 'View Data' in the local database table that I put in the dataflow I could see the rows that I manually sent from the ECC. However, after truncating the local table, it bothered me that when I tried to send manually a second IDOC with another row from the Cost Center Master Data table, my real-time job didn't receive the request and consequently didn't insert the desired rows. I didn't change anything in the configuration, so my first question to any of you is that if you know what could it be that causes my real-time job to not receive the request from ECC?. I tried making a third try, but at my company they had to shut down the ECC server, so it's going to be a while before I make another try.
    <br/>
    <br/>Now, the REAL question of this post is that after I sent the first IDOC successfully (and before sending manually the second one mentioned earlier), I changed a record in the CSKS table directly in the ECC, hoping that it would automatically generate an COSMAS01 IDOC and send all the data in the table to the Data Integrator. This didn't happen and no IDOC was generated, so do any of you know why the automatic change didn't trigger the sending of the IDOC with the data?
    <br/>
    <br/>As I said before, I made the configuration in ECC and DI following the steps in the link written at the start of the post, and it was tested OK by successfully sending a COSMAS01 IDOC once, manually using bd16.
    <br/>
    <br/>In advance I thank you all for your cooperation. This is my first thread in the SDN forum so also please excuse any mistakes in my english.
    <br/>
    <br/>Best regards...

    You do not need to schedule this job every 10 min.  Why?
    If we can't schedule this job every 10 mints how can I able to retrive the delta records into Queue(RSA7)
    What is your advice I mean how to schedule in order to get the delta?
    Thanks

  • Extraction from SAP ECC to SAP BI

    What are the steps for Extraction from SAP ECC to SAP BI? Thanks.

    Hi,
    we have two types of extractions 1) Application specific 2)cross application
    under applicatin specific -->we have customer genenrated and business content extractions
    under customer generated means LIS, CO-PA, FI-SL(i.e no ready made data source in business content so we have to generate datasource)
    Business content means there is ready made business content datasource so we can use it
    LIS,LO --> when ever we go for extracting logistics informatin
    LO steps
    http://www.sap-img.com/business/lo-cockpit-step-by-step.htm
    LO Cockpit (Use transaction LBWE)
    /people/vikash.agrawal/blog/2006/12/18/one-stop-shop-for-all-your-lo-cockpit-needs
    https://www.sdn.sap.com/irj/sdn/weblogs?blog=/pub/u/21406 [original link is broken] [original link is broken]
    /people/sap.user72/blog/2005/09/05/sap-bw-and-business-content-datasources-in-pursuit-of-the-origins
    To understand LO extraction go through these weblogs by Roberto
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2004/12/23/logistic-cockpit-delta-mechanism--episode-two-v3-update-when-some-problems-can-occur
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    /people/sap.user72/blog/2005/02/14/logistic-cockpit--when-you-need-more--first-option-enhance-it
    /people/sap.user72/blog/2005/04/19/logistic-cockpit-a-new-deal-overshadowed-by-the-old-fashioned-lis
    under cross Applicaion
    Generic extraction-->table, view, infoset query, function module
    Generic extractors are of 3 types:
    1. Based on table/view
    2. Based on Infoset Query
    3. Based on Function module
    GenericSteps
    Log on sap R/3.
    Step 1. create a table or view for generic extraction in se11.
    Step 2. Goto  t-code RSO2
    Step 3. Here u have to decide whether to extract transaction data or master data attributes or texts.
    Step 4. suppose if u have opted for transaction data,then give name in the column ex:ztd_m(data source name)
    Step 5. select create button,this will take u to another screen.
    Step 6. Here u have to decide from which application component u r extacting data.Ex: SD,MM,..
    Step 7. in this screen u have to fill short disciption,medium,long( these r mandatory).
    Step 8. then U have to Table name or view name which u have created in se11.
    Step 9. I f u want to maintain generic delta then u can select generic delta in top left hand side corner.
    Step 10. in next screen u have to give some field which is primary key.
    step 11. here u to specify whether time stamp or cal day or numeric pointer depending on u r requirement.
    step 12. then u have to specify  whether new status for changed records or additive delta.
            If u choose additive delta ,then u have to load data to infocube or ods object.
            If u choose new status for changed records, then u to load data  to ods object only.
    Step 13. then save it.
    Step 14. then logon to sap bw ,then replicate the data source then as usuall.
    COPA
    CO-PA--> for financial info
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/COPA/COPA.pdf
    You can browse forum for more detailed info..
    Regards,
    B

  • Loadind transaction data from SAP ECC 6.0 system failed

    Hi!
    I have SAP ECC 6.0 IDES system with the following clients:
    100 DEV client
    200 BI client of SAP 6.0
    Now I would like to load the transaction data from InfoSources 0FI_AR_6 and 0FI_AR_4 from SAP ECC client 100.
    The extraction process has been frozen in status "yellow" and after a while changed the status to "red" with the following information
    Errors while sending packages from OLTP to BI
    Diagnosis
    No IDocs could be sent to BI using RFC.
    System Response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further analysis:
    Check the TRFC log.
    You can access this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Error handling:
    If the TRFC is incorrect, check whether the source system is fully connected to BI. In particular, check the authorizations of the background user in the source system.
    The users BWREMOTE and ALEREMOTE seems to be ok and have the suitable authorization.
    The tcode SM58 does not show any entries.
    The tcode BD87 shows that the IDocs was sent (source system) and has been received on target system.
    Question:
    How can I solve my problem?  Please give me some technical information (tcode, report).
    Thank you very much!
    regards

    Walk through your basis set up and  check out
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/c0751ba5-b7a8-2b10-6d97-e91e85c0fafa
    pts appreciated.

  • How to extract data from SAP 4.7 and upload data to SAP ECC 6.0

    Hi,
        How to extract data from SAP 4.7 and upload data to SAP ECC 6.0? Can i use BDC,BAPI,LSMW? Help me please.

    hi
    good
    both works not possible simultaneously.
    If you want to do it in two separate task than you can use the GUI_UPLOAD function module to fulfill your requirement.
    thanks
    mrutyun^

  • Batch Jobs fail because User ID is either Locked or deleted from SAP System

    Business Users releases batch jobs under their user id.
    And when these User Ids are deleted or locked by system administrator, these batch jobs fail, due to either user being locked or deleted from the system.
    Is there any way that these batch jobs can be stopped from cancelling or any SAP standard report to check if there are any batch jobs running under specific user id.

    Ajay,
    What you can do is, if you want the jobs to be still running under the particular user's name (I know people crib about anything and everything), and not worry about the jobs failing when the user is locked out, you can still achieve this by creating a system (eg bkgrjobs) user and run the Steps in the jobs under that System User name. You can do this while defining the Step in SM37
    This way, the jobs will keep running under the Business User's name and will not fail if he/she is locked out. But make sure that the System User has the necessary authorizations or the job will fail. Sap_all should be fine, but it really again depends on your company.
    Kunal

  • SAP ECC 6.0 extractor 0FI_GL_4 -extracting GL data in delta mode into MSSQL

    Hi,
    Currently we are having the following need. We have to build delta interface from SAP ECC into MSSQL for G/L data. If possible, we would like to use for this standard ECC extractor: 0FI_GL_4. As far as I know this extractor is dedicated for SAP BW, but can we apply it for MSSQL? Does anyone has any experience with similar task? Any help will be appreciated.
    Regards,
    Marek

    Hi,
    I'm using the data source 0SEM_BCS_10.
    I have created my staging cube similar to my Totals cube (0BCS_C11).
    I'm stuck in creating update rules for this staging cube, particularly with respect to the follwing feilds:
    1) Posting level
    Pls let me know how to update this.
    Also, help me understand in general, how to write update rules to feilds which exist only in BCS.
    2) F Yr of the first acquisition posted
    3) Period in which first acq was posted
    (Mapping for these will be constant values (actual dates) - correct me if wrong)
    4) Mov Type
    I guess we need to just map this with the transaction type in ECC.
    Many Thanks,
    Kumar

Maybe you are looking for