Update rule problem - while data load

Hi friends,
I got the following error while doing initialisation for 2lis_02_sgr.
"ABORT was set in the customer routine 9998
Error 1 in the update "
In the forum i searched for this error and this error is something related to the start routine in my update rule.
But i dont know whats wrong with my routine.
Im giving the start routine below,pls go through this and give me your suggestions..
PROGRAM UPDATE_ROUTINE.
$$ begin of global - insert your declaration only below this line  -
TABLES: ...
<i>TABLES /bic/AZMM_PUR100 .
DATA:  T_PUR1 LIKE /bic/AZMM_PUR100 OCCURS 0 WITH HEADER LINE.</i>
$$ end of global - insert your declaration only before this line   -
The follow definition is new in the BW3.x
TYPES:
  BEGIN OF DATA_PACKAGE_STRUCTURE.
     INCLUDE STRUCTURE /BIC/CS2LIS_02_SGR.
TYPES:
     RECNO   LIKE sy-tabix,
  END OF DATA_PACKAGE_STRUCTURE.
DATA:
  DATA_PACKAGE TYPE STANDARD TABLE OF DATA_PACKAGE_STRUCTURE
       WITH HEADER LINE
       WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
FORM startup
  TABLES   MONITOR STRUCTURE RSMONITOR "user defined monitoring
           MONITOR_RECNO STRUCTURE RSMONITORS " monitoring with record n
           DATA_PACKAGE STRUCTURE DATA_PACKAGE
  USING    RECORD_ALL LIKE SY-TABIX
           SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS
  CHANGING ABORT LIKE SY-SUBRC. "set ABORT <> 0 to cancel update
$$ begin of routine - insert your code only below this line        -
fill the internal tables "MONITOR" and/or "MONITOR_RECNO",
to make monitor entries
if abort is not equal zero, the update process will be canceled
  CLEAR: T_PUR1[] ,
         T_PUR1,
         ABORT.
  SELECT * INTO TABLE T_PUR1 FROM /bic/AZMM_PUR100.
  IF SY-SUBRC EQ 0.
    SORT T_PUR1 BY DOC_DATE
                   DOC_ITEM
                    DOC_NUM.
  ELSE.
    MONITOR-msgid = sy-msgid.
    MONITOR-msgty = sy-msgty.
    MONITOR-msgno = sy-msgno.
    MONITOR-msgv1 = sy-msgv1.
    MONITOR-msgv2 = sy-msgv2.
    MONITOR-msgv3 = sy-msgv3.
    MONITOR-msgv4 = sy-msgv4.
    append MONITOR.
  if abort is not equal zero, the update process will be canceled
         ABORT = 1.
  ENDIF.
   ABORT = 0.
$$ end of routine - insert your code only before this line         -
ENDFORM.
Thanks & Regards
Ragu

thanks gimmo and a.h.p,
i have done the correction as you said,pls verify that.
And also kindly explain me what is the reason for this start routine,what exactly it does???
CLEAR: T_PUR1[] ,
         T_PUR1,
         ABORT.
  SELECT * INTO TABLE T_PUR1 FROM /bic/AZMM_PUR100.
  IF SY-SUBRC EQ 0.
    SORT T_PUR1 BY DOC_DATE
                   DOC_ITEM
                    DOC_NUM.
abort = 0.    (  added  abort = 0 as per your suggestion )
  ELSE.
    MONITOR-msgid = sy-msgid.
    MONITOR-msgty = sy-msgty.
    MONITOR-msgno = sy-msgno.
    MONITOR-msgv1 = sy-msgv1.
    MONITOR-msgv2 = sy-msgv2.
    MONITOR-msgv3 = sy-msgv3.
    MONITOR-msgv4 = sy-msgv4.
    append MONITOR.
  if abort is not equal zero, the update process will be canceled
         ABORT = 1.
exit. ( added exit as per your suggestion )
  ENDIF.
   ABORT = 0.
$$ end of routine - insert your code only before this line         -
ENDFORM.
Thanks & Regards
ragu

Similar Messages

  • Update rule problem for date in Prod

    My scenario is like this:-
    ODS 2 is loaded from ODS 1, in ODS 1 there is a data field calendar day (DATS, time characteristic) and there is a data field posting date (DATS, characteristic) in ODS 2. In the update rule, the posting date is updated from calendar day by a formula source. 
    The problem is the posting date data field is updated into ODS 2 correctly in development box but it is not updated (blank) in production box.  I can't figure out what is the cause, hopefully someone can give me some help.  Thanks.
    Cheers!
    Cecil

    The only different is the development system is having this part of code in the update rules program compare to production system.  Do I need to compile the update rules formula manually?
    *This ABAP Code was generated automatically          *
    *Formula Calculator                                  *
    *Generated :2008:09:12-10:47
    *User: XXX
    *Calculation:
    result = COMM_STRUCTURE-CALDAY.
      ENDCATCH.
      if sy-subrc <> 0.
        perform error_message using 'RSAU' 'E' '507'
                'ROUTINE_0004' g_s_is-recno
                rs_c_false rs_c_false g_s_is-recno
                changing c_abort.
      endif.
    Cheers!

  • IKM Problem While Data Loading

    Hi Everyone,
    I am new to ODI. And trying to change somethings in exixting ODI IKM.
    I have an IKM with several steps. In a step of insert into I$ table i am encountering a problem.
    Firstly you will need to know i have two different schema Temp_schema for creating temp tables and Target_schema for target tables.
    And our client wants to have all temporary tables in temp_schema.
    What happening over here is When i am trying to run following code it is taking correct schema that is temp_schema for odiRef.getTable("L","INT_NAME","A")%> this code.
    While for <%=odiRef.getFrom(i)%> code it takes target schema.
    I have specified temp_schema at steps in IKM.
    Code :-
    insert into     *<%=odiRef.getTable("L","INT_NAME","A")%>*
         <%=odiRef.getColList("", "[COL_NAME]", ",\n\t", "", "(((INS or UPD) and !TRG) and REW)")%>,
         IND_UPDATE
    <%for (int i=0; i < odiRef.getDataSetCount(); i++){%>
    <%=odiRef.getDataSet(i, "Operator")%>
    select      <%=odiRef.getUserExit("OPTIMIZER_HINT")%>
         <%=odiRef.getPop("DISTINCT_ROWS")%>
         <%=odiRef.getColList(i,"", "[EXPRESSION]", ",\n\t", "", "(((INS or UPD) and !TRG) and REW)")%>,
    <% if (odiRef.getDataSet(i, "HAS_JRN").equals("1")) { %>
         JRN_FLAG IND_UPDATE
    <%} else {%>
         'I' IND_UPDATE
    <%}%>
    from     <%=odiRef.getFrom(i)%>
    where     (1=1)
    <%=odiRef.getJoin(i)%>
    <%=odiRef.getFilter(i)%>
    <%=odiRef.getJrnFilter(i)%>
    <%=odiRef.getGrpBy(i)%>
    <%=odiRef.getHaving(i)%>
    <%}%>
    I really need to have temp_schema in odiRef.getFrom(i) step.
    Any help would be thankful.
    Edited by: MaheshGx on Dec 7, 2011 1:12 AM
    Edited by: MaheshGx on Dec 7, 2011 1:14 AM

    Hi
    Go to target physical schema and select workschema as OMPODI..Now use default LKM..Go to IKM change the step 200,210,220 to your target schema that is OMP (Not recommended).As you are giving your workschema different in your physical schema so no need to change the IKM.use the default one.
    Thanks

  • Problem while  data processing TRANSACTION data from DSO to CUBE

    Hi Guru's,
    we are facing problem while  data processing TRANSACTION data from DSO to CUBE. data packets  processing very slowly  and updating .Please help me regarding this.
    Thanks and regards,
    Sridhar

    Hi,
    I will suggest you to check a few places where you can see the status
    1) SM37 job log (give BIrequest name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
    2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. See if its accessing/updating some tables or is not doing anything at all.
    If its running and if you are able to see it active in SM66 you can wait for some time to let it finish.
    3) RSMO see what is available in details tab. It may be in update rules.
    4) ST22 check if any short dump has occured.
    You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
    If you feel its active and running you can verify by checking if the number of records has increased in the cube.
    Thanks,
    JituK

  • BW 3.5 update rules for master data

    Hi all,
             I create a dummy report in that report lifnr is charactics and name its att. after  this i create a application component and assign data source and create infopkg   data coming in lifnr and name field properly <b>i assign data source using direct master data update</b> after this i create infocube and assign dim.
    but when i rgt click on infocube and display dataflow it shows only one level and when i open bex and open that it shows no application data found why this happened after creating infocube it is nessory that i creat update rule for this am i right plz send your openion.
    thx

    Hi Ankit,
                  Yoy have to load data to the cube.For that you need update rules.
    Steps to load data to the cube
    1.Replicate the datasource.
    2.Assign Infosource
    3.Create transfer rules
    4.Maintain the update rules to the cube.
    5.Right click on the datasource and schedule an infopackage for it.
    Execute the Infopackage and pull the data to the cube.
    Hope this helps
    Regards
    Karthik

  • Update rule problem - validation of "sales/cost w/ tax" keyfigure

    BW Gurus,
        Hi to all, i have this update rule problem at "sales/cost w/tax" keyfigure here is the senario.
        Our Goverment mandatory implemented an additional 2% tax from the original 10%, this will affect our report on sales, and also the BW "sales/cost w/ tax" key figure.
        My question is How can I validated the effectivity of the new tax? i have tax 10% from previous sales and 12% on the current sales. What "date field" can I use to validated this. I am using /BIC/CS2LIS_13_VDITM stucture to get the data i need.
    Thanks in Advance
    Joven

    Hi,
    Till to day how are extracting the data for tax(original 10%) is it available directly in 2LIS_13_VDITM ?
    Usually all taxes( condition types) can be extracted by the data source 2LIS_13_VDKON.Discuss with SD team, they may give the condition type used for different taxes.
    With rgds,
    Anil Kumar sharma .P

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Problem in data loading through UD Connect.

    Hello All,
    i have problem in data loading...I have created the UD connect source systen and its working properly,then i have created UD Connect InfoSources and DataSources, But when i am trying to create Infopackages then its giving following error..
    "Error occurred in the source system" and "no record found".
    Thanks
    Shivanjali

    Hello Shivanjali,
    Mostly UDC is used for RemoteCube to access data outside of SAP. What is your external system? Make sure that there is data in the source system.
    Please check following links
    [Transferring Data with UD Connect|http://help.sap.com/saphelp_nw04s/helpdata/en/43/e35b3315bb2d57e10000000a422035/frameset.htm]
    [SAP BW Universal Data Integration SAP NW Know-How Call|https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f094ba90-0201-0010-fe81-e015248bc5dd]
    [SAP BW and ETL A Comprehensive Guide|https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/11e1b990-0201-0010-bf9a-bf9d0ca791b0]
    Thanks
    Chandran

  • Error in Unit conversion while data loading

    Hi,
    I have maintained DSO in Material (0MATERIAL) info object > Bex tab > Base unit of measure > ZDSO; Then loaded this ZDSO from std data source  0MAT_UNIT_ATTR so that all conversion factors into different units which are maintained in material master data in ECC would get loaded to this DSO.
    Then I have created one Conversion type (ZCON) to read source unit from record and convert it to fixed unit "ST" with reference info object as 0MATERIAL. ST is customized UOM here.
    I am using ZCON conversion type to convert Qty in base UOM to Qty in ST in Bex reports under conversion tab of key figure. Now as this is std functionally, conversion would automatically takes place using ZDSO (mentioned above) as source and target UOM are not from same dimension (Base UOM is EA and target UOM is ST).
    If conversion factor to ST is not found in ZDSO then conversion to base UOM would happen automatically. Now this functionality is happening perfectly in Bex but its giving error if I use the same conversion type ZCON while data loads. Its giving error for those material for which ST conversion is not maintained. But when its not maintained by default it should convert it to base UOM, but its not converting and giving error in data loads.
    Hope I am able to explain the issue.
    Please help me on on this issue or any way around.
    Thanks in advance.
    Prafulla

    Ganesh,
    Can you please check out the Alpha Conversion Routine and also nodeid for that infoobject..
    There might be some inconsistencies in the code..
    Hope it helps
    Gattu

  • Performance while data load

    Hello Friends,
    I an having a query regarding performance in BW 3.5.
    while data load from R3 to BW we have 4 options.
    Only PSA
    Only Data Target
    PSA and later data target
    PSA and Data target in parallel.
    In system performance point of view, which is the best option with less system resources and How.
    Your help aprreciated.
    Thanks
    Tony

    Hi ,
    for performance point of view ..
    PSA and later data target will be better option ..
    for more info check this link ..
    http://help.sap.com/saphelp_nw04/Helpdata/EN/80/1a6567e07211d2acb80000e829fbfe/frameset.htm
    Regards,
    shikha

  • Update Rule error while loading the data

    Hello All,
    I tried to do the init load for the datasource 2LIS_02_VAKON (Sales Document Condition), But it error's out in the Update rule saying:
    Short text: Error when generating the update program
    Long Text:  Diagnosis
         An error occurred during program generation for InfoSource 2LIS_11_VAKON
         and InfoProvider 0SD_C15 . This may be extrapolated to incorrect update
         rules.
    Procedure
         Check and correct your update rules and then generate the update program
         again. You can find help on the error in the error log.
    The update rule looks very fine. I even deleted it and re created, but no success..... Please give me your thoughts abt it..... Any suggestions will be great for me.....

    Hello Gokul,
    I could'nt see any analyse button in RSMON. This is the only error message I could get from the Process monitor:
    Short text: Error when generating the update program
    Long Text: Diagnosis
    An error occurred during program generation for InfoSource 2LIS_11_VAKON
    and InfoProvider 0SD_C15 . This may be extrapolated to incorrect update
    rules.
    Procedure
    Check and correct your update rules and then generate the update program
    again. You can find help on the error in the error log.
    Do you expect something else ? Also I went through your OSS note, they are asking to apply support pack 18. Is this a serious issue to apply SP? Because in Quality and Prod it loads fine......
    Please let me know your idea

  • Update Rule Error ( Invalid Data) while load 0VENDOR

    Hello All,
    I have attributes like  0POSTAL_CD of 10 Char (Postal Code) and 0SORTL 0f 10 Char (Sort Field) of Vendor. From yesterday onwards the Production master data loads for 0VENDOR started failing saying:
    Record 211 :0POSTALCODE : Data record 211 ('03 2VEN4000040 '): Version 'HK HONG KONG ' is not valid
    Record 136 :0SORTL : Data record 136 ('0002000175 '): Version 'SATTLER, I ' is not valid
    Solutions tried:
    Interstingly what we found is: when u edit the PSA data for the above records and again input the same values, the load runs fine.
    I found  this thread some what relating to my problem, but I could'nt get the solution.
    Please help to solve this issue which I am facing alone....... Thanks for your support

    Hi Ram,
    Really add more valid char is a possible option, but in your case I dont believe that was a good idea...
    I believe that this msg is an automatic master data validation.
    Ex:
    0SORT:  is not valid use comma ( , ) or sapce in this field
    0POSTALCODE: the value  'HK HONG KONG '  dont seems to be a correct value for a zip code
    I belive that this validation is not about allowed extra-char
    Check both InoObject and see wich datatype support.
    Hope this Help!

  • Master Data lookup in Update Rule problem

    Hi all,
    I am currently having a problem loading data to an InfoCube using flat files.
    The architecture is as follows:
    1) The source of the data is a flat file
    2) The data is loaded thru an Update Rule and is of type Full-Update
    3) The Update Rules determines the Profit Center using the Master Data of the WBS-Element
    4) The data is written in an InfoCube
    This solution however does not always work as planned. In the following situation a problem occurs:
    1) The flat file contains WBS-element RD.00753.02.01, which has a Profit Center attribute value 8060
    2) When I load the flat file, the PC value 8060 is written into the row in the InfoCube, which is correct
    3) Then I change the master data of the WBS-element by setting the Profit Center attribute value to 8068
    4) I run the Attribute Change Run
    5) Then i load a flat file again, which also contains WBS-element RD.00753.02.01
    6) The master data attribute value should now write the value 8068 into the InfoCube. HOWEVER, this is when the evil occurs. BW does not write a PC value of 8068, but it write the value 8060. This is wrong.
    Why does BW not take the newest version of the Master Data to performe the attribute value look-up? Or why doesn't BW write the correct Profit Center into the cube?
    Thanks,
    Onno

    Hi Ricardo,
    The debug via PSA simulation of the update indicates that the CORRECT Profit Center value is to be written into the InfoCube.
    However, if I check the contents of the cube (after the load has finished) using the request-id the WRONG Profit Center value is shown. This indicates that the correct Master Data is used, however the update of the Cube is wrong. Why does this happen. the load is of type full-update, so should add a new row in the cube using the value in the data from the UR.
    Onno

  • Performance Problem While Data updating In Customize TMS System

    Dear guys
    I have developed Customize time management system,but there is performance problem while updating machine date in monthly roster table,there is if else condition,which check record status either is late or Present on time etc.
    have any clue to improve performance.
    Thanks in advance
    --regard
    furqan

    Furqan wrote:
    Dear guys
    I have developed Customize time management system,but there is performance problem while updating machine date in monthly roster table,there is if else condition,which check record status either is late or Present on time etc.
    have any clue to improve performance.From that description and without any database version, code, explain plans, execution traces or anything that would be remotely useful.... erm... No.
    Hint:-
    How to Post an SQL statement tuning request...
    HOW TO: Post a SQL statement tuning request - template posting

  • Problem while down loading data into Excel

    Hi friend's,
    I am getting a error while down loading header
    internal table to excel.
    since in my header internal table there is no p
    type field..but i am getting this error..
    Only flat, character-type data objects are supported at the argument
    position "dest" for the statement
      "WRITE src TO dest".
    In this case, the operand "dest" has the non-character-type or deep type
    "P". The current program is flagged as a Unicode program. In the Unicode
    context, type X fields are seen as non-character-type, as are
    structures that contain non-character-type components.
    i have coded like this
    REPORT  ZACG_CCA_EXCEL                          .
    Tables : zacg_cca.
    data: P_file like RLGRAP-FILENAME.
    Data: Begin of it_header occurs 0,
          ccode(10) ,
          mat_cd(12) ,
          ingr_desc(30),
          con(10),
          quan(10),
          percqty(10),
          DATE(10),
          rsamnos(10),
          flag(5),
          end of it_header.
    Data : begin of it_final occurs 0,
           ccode type zacg_cca-ccode,
           mat_cd type zacg_cca-mat_cd,
           ingr_desc type zacg_cca-ingr_desc,
           conc type zacg_cca-conc,
           quantity type zacg_cca-quantity,
           percqty type zacg_cca-percqty,
           APP_DATE type zacg_cca-app_date,
           rsamnos type zacg_cca-rsamnos,
           flag ,
           end of it_final.
           SELECTION-SCREEN : BEGIN OF BLOCK blk WITH FRAME TITLE text-000.
           select-options : s_Date for zacg_cca-app_date.
           SELECTION-SCREEN  : END OF BLOCK blk.
           it_header-ccode = 'Samp_code'.
           it_header-mat_cd = 'Mat_code'.
           it_header-ingr_desc = 'Ingr_Desc'.
           it_header-con = 'Conc'.
           it_header-quan = 'Quantity'.
           it_header-percqty = 'Perc'.
           it_header-date = 'Date'.
           it_header-rsamnos = 'Rsamnos'.
           it_header-flag = 'Flag'.
           Append it_header.
           select ccode mat_cd ingr_desc conc quantity percqty
            app_date rsamnos from zacg_cca  into corresponding
            fields of table
           it_final where  zacg_cca~app_date in s_date.
           loop at it_final.
           it_final-flag = 'T'.
           modify it_final.
           it_final-quantity = it_final-quantity * 2 .
           Modify it_final.
           endloop.
           CALL FUNCTION 'MS_EXCEL_OLE_STANDARD_DAT'
             EXPORTING
               FILE_NAME                       = 'C:\amol\p_file'
              CREATE_PIVOT                    = 0
             DATA_SHEET_NAME                 = ' '
             PIVOT_SHEET_NAME                = ' '
             PASSWORD                        = ' '
             PASSWORD_OPTION                 = 0
              TABLES
             PIVOT_FIELD_TAB                 =
              DATA_TAB                        =   it_final
             FIELDNAMES                      =  it_header
            EXCEPTIONS
              FILE_NOT_EXIST                  = 1
              FILENAME_EXPECTED               = 2
              COMMUNICATION_ERROR             = 3
              OLE_OBJECT_METHOD_ERROR         = 4
              OLE_OBJECT_PROPERTY_ERROR       = 5
              INVALID_PIVOT_FIELDS            = 6
              DOWNLOAD_PROBLEM                = 7
              OTHERS                          = 8
           IF SY-SUBRC <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
           ENDIF.
    can any one....

    hii
    for title this FM have some restriction as it can not take value if its more then 10 length so you need to give length as 10 only for every string then append it with FIELDNAMES = it_header
    regards
    twinkal

Maybe you are looking for