Critical error with data load

Hello Gurus,
I was trying to load data ffrom ODS to a Cube using DTP when I got the following error:
Exceptions in Substep: End Routine - RSBK231
The database returned a value containing an error
Do you know what this error message means. It refers to a end routine and I had no syntaxt errors with the endroutine. It also worked fine with previous data loads.
Any help in this regards is appreciated.
Thanks
Rishi

Hi Rishi,
There might be some syntactical error in your end routine e.g division by a zero. This is just an example. Try running the DTP in Debug mode (you will get the option under execute tab of DTP) and see what the end routine code is doing. There might be some exception which you need to take care of.
Thanks..
Shambhu

Similar Messages

  • Error is data loading from 3rd party source system with DBCONNECT

    Hi,
    We have just finished an upgrade of SAP BW 3.10 to SAP NW 7.0 EHP1.
    After the upgrade, we are facing a problem with data loads from a third party Oracle source system using DBConnect.
    The connection is working OK and we can see the tables in the source system. But we cannot load the data.
    The error in the monitor is as follows:
    'Error message from the source system
    Diagnosis
    An error occurred in the source system.
    System Response
    Caller 09 contains an error message.
    Further analysis:
    The error occurred in Extractor .
    Refer to the error message.'
    But, unfortunately, the error message has no further information.
    If we look at the job log in sm37, the job finished with the following log -                                                                               
    27.10.2009 12:14:19 Job started                                                                                00           516          S 
    27.10.2009 12:14:19 Step 001 started (program RSBATCH1, variant &0000000000119, user ID RXSAHA)                    00           550          S 
    27.10.2009 12:14:23 Start InfoPackage ZPAK_4FMNJ2ZHNNXC6HT3A2TYAAFXG                                              RSM1          797          S 
    27.10.2009 12:14:24 Element NOAUTHORITYCHECK is not available in the container                                     OL           356          S 
    27.10.2009 12:14:24 InfoPackage ZPAK_4FMNJ2ZHNNXC6HT3A2TYAAFXG created request REQU_4FMXSQ6TLSK5CYLXPBOGKF31G     RSM1          796          S 
    27.10.2009 12:14:24 Job finished                                                                                00           517          S 
    In a BW 3.10 system, there is no  message related to element NOAUTHORITYCHECK. So, I am wondering if this is something new in NW 7.0.
    Thanks in advance,
    Rajib

    There will be three things to get the errors like this
    1.RFC CONNECTION FAILED
    2.CHECK THE SOURCE SYSTEM
    3.CHECK IT OUT WITH Oracle Consultants WEATHER THEY ARE FILLING UP THE LOADS.TELL THEM TO STOP
    4.CHECK I DOC PROCESSING
    5.FINALLY MEMORY ISSUES.
    6.CATCH THE DATA SOURCE FIRST CHANGE IT AND THEN ACTIVATE AND RUN THE LOAD
    7.LAST IS MEMORY ISSUE.
    and also Check the RFC connection in SM59 If  it is ok then
    check the SAP note : 692195 for authorization
    Santosh

  • Essbase Error(1003050): Data Load Transaction Aborted With Error (1220000)

    Hi
    We are using 10.1.3.6, and got the following error for just one of the thousands of transactions last night.
    cannot end dataload. Essbase Error(1003050): Data Load Transaction Aborted With Error (1220000)
    The data seems to have loaded, regardless of the error. Should we be concerned, and does this suggest something is not right somewhere?
    Your assistance is appreciated.
    Cheers

    Hi John
    Not using a load rule.
    There were two other records which rejected based on absentee members. The error message was different for them, and easily fixed.
    But this error doesn't tell us much. I will monitor the next run to see if the problem persists.
    Thanks

  • A Critical error encountered while loading webpage.

    Was running 2.1.51  and when browsing to the status link while logged in to the web gui I got a blank web page with the following text:
    A Critical error encountered while loading webpage.
    Reboot router - log back in - can go anywhere except the status page -- get same error page
    made the mistake of downloading and upgrading firmware to 2.1.71   = upgrade completed successfully = presented with the new login screen confirming that the box is running the new code -- now when I log in - I don't get in rather the previous error page - and cant get anywhere.
    A Critical error encountered while loading webpage.
    the url is https://192.168.75.1/scgi-bin/platform.cgi   but it doesn't load only displays the error.
    could someone share another valid management url for the box so I can at least try to get in somewhere before attempting to factory reset this POS ;-(
    thanks.

    Thanks for the response.   Based on the number of crickets that I heard after asking the question, I did indeed factory reset the box.  Luckily I had a backup of the config so it wasn't too painful to restore.
    I beleive the root cause co-incided with the expiration of the trend micro web protection gateway service.
    For whatever reason, I did not receive any notification that the service was to expire but the error experienced was noticed just two days after the subscription expired...  I had successfully logged in without issue and was able to view the status tab only a week earlier.
    be sure to pay the cisco tax on time or they'll lock you out of your box was the lesson learned...

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Common errors in Data Loading

    Can anybody tell me what are the common error we come across in BW/BI for following:-
    1] Data Loading in Infocube/ODS/DSO
    2] Aggregates
    3] PSA
    4] DTP
    5] Transformations
    Thanks in advance

    Hi,
    Here are the list of common issues we face while data loading in to BW:
    Data Loading in Infocube/ODS/DSO:
    1) Missing SID issue or missing master data
    2) Data load cancelled becasue of various exceptions. Eg: Message Type 'x', DBIF SQL Error,
    3) Data records found in duplicate
    4) Alpha Confirming Vaue error
    5) Invalid time interval
    6) Time overlap error
    7) Job Cancelled in source system
    8) RFC Connection Error
    9) Data Source Replication error
    10) Data load locked by active change run
    11) Attributes to target by user 'xyz'
    12) Job cancelled as the previous load is still running
    13) Target locked by a change run
    (Some times data loads dont fail but run for long time then usual with out any progess, because of struck tRFCs in source sytem, source sytem performance is poor and the job is in release state for long time, etc...)
    Aggregate Rollups:
    1) Nofilled aaggregates available, rollup not possible
    2) Rollup locked by active change run
    3) Job cancelled because of various exceptions
    PSA Updations:
    1) Job cancelled because of various exception
    2) Missing SID value
    ODS Activaitons:
    1) Request to be activate shoud be green - (Loading of data in to ODS in still going on)
    2) Key exists in duplicate
    3) Job cancelled because of various exceptions (Short Dumps)
    Attribue Change Run:
    1) Locked by a rollup
    2) Job cancelled because of various exceptions
    3) Locked by another change run
    Hope it helps....
    Cheers,
    Habeeb

  • Insert OR Update with Data Loader?

    Hello,
    can i Insert OR Update at same time with Data Loader?
    How can i do this?
    Thanks.

    The GUI loader wizard does allow for this including automatically adding values to the PICKLIST fields.
    However, if you mean the command line bulk loader, the answer is no. And to compound the problem, the command line version will actually create duplicates for some of the objects. It appears that the "External Unique Id" is not really "unique" (as in constrained via unique index) for some of the objects. So be very careful when you prototype something with the GUI loader and then reuse the map on the command line version.
    You will find that some objects can work well with the command line loader (some objects will not).
    Works well (just a few examples):
    Account (assuming your NAME,LOCATION fields are unique).
    Financial Product
    Financial Account
    Financial Transaction
    Will definitely create duplicates via command line bulk loader:
    Contact
    Asset
    Also be aware that you might hear that during a go-live that Oracle will remove the 30k record limit on bulks loads (temporarily). I have not had any luck with Oracle Support making that change (2 clients specifically in the last 12 months).

  • Need help with Rollback data if there is error in Data load

    Hi All,
    We are trying to load data to Oracle 11g database. We want a trigger, procedure or something like that to rollback the data if there are errors in load. Is it possible to do rollback after all the records has been parsed ? So if we try to load 100 records and if there are 30 records with error, we want to rollback after all the 100 records are parsed.
    Please advice.

    >
    Thanks for the suggestion. I'll try that option. So currently we are only loading data that is validated and erroneous records are rejected using trigger. So we don't get any invalid data in table. But Now users are saying that all the records should be rejected if there is even one error in the data load.
    >
    I generally use a much simpler solution for such multi-stage ETL processes.
    Each table has a IS_VALID column that defaults to 'N'. Each step of the process only pulls data with the flag set to 'Y'.
    That allows me to leave data in the table but guarantee that it won't be processed by subsequent stages. Since most queries that move data from one stage to another ultimately have to read table rows (i.e. they can't just use indexes) it is not a performance issue to add a predicate such as "'AND IS_VALID = 'Y'" to the query that accesses data.
    1. add new unvalidated data - automatically flagged an invalid by default of 'N" on IS_VALID column
    2. run audit step #1 - capture the primary key/rowid of any row failing the audit.
    3. run audit steps #2 through #n - capture error row ids
    4. Final step - update the data setting IS_VALID to 'Y' only if a row passes ALL audits: that is, only if there are NO fatal errors for that row captured in the error table.
    That process also allows me to capture every single problem that any row has so that I can produce reports for the business users that show everything that is wrong with the data. There are some problems that the business wan't to ignore, others that can be fixed in the staging tables then reprocessed and others that are rejected since they must be fixed in the source system and that can take several days.
    For data that can be fixed in the staging tables the data is fixed and then the audit is rerun which will set the IS_VALID flag to 'Y' allowing those 'fixed' rows to be included in the data that feeds the next processing stage.

  • Segmentation fault error during data load in parallel with multiple rules

    Hi,
    I'm trying to do sql data load in parallel with multiple rules (4 or 5 rules, maybe), i'm getting a "segmentation fault" error. I tested 3 rules file and it worked fine. we're using Essbase system 9.3.2., with UDB (v8) as the sql data source. ODBC driver is DataDirect 5.2 DB2 Wire Protocol Driver (ARdb222). Please let me know if you have any information on this.
    thx.
    Y

    Hi Thad,
    I was wondering, if system is unicode or non unicode that should not matter the amount and currency field . As currencies are defined by SAP and it is in pure English at least a currency code part of 3 Chars. 
    Could this because of some incosistency of data ??
    I would like to know for Which currency  had some special characters it it in that particular record ??
    Hope that helps.
    Regards
    Mr Kapadia

  • Error in data load from application server

    Well, this problem again!
    In datasource for dataload by flatfile, tab Extraction i selected Load Text-Type File From Application Server.
    In tab Proposal:
    Converter: Separated with Separator (for Example, CSV).
    No. of Data Records: 9198
    Data load sucefull.
    I add 1 record:
    Converter: Separated with Separator (for Example, CSV).
    No. of Data Records: 9199
    So appear a  message error "Cannot convert character sets for one or more characters".
    I look in line 9199 in file and not have any special character. I use AL11 and debug to see this line.
    When i load file from local workstation, this error not occur.
    What's happen?

    Hi Rodrigo,
    What type of logical file path have yopu created in the application server for loading this file.
    Is it a UNIX file path or an ASCII file path.
    Did you check how the file looks in the application server AL11?
    Prathish

  • Error while data loading in real time cube

    HI experts,
    I have a problem. I am loading data from a flat file.The data is loading correctly till the DSO but when i am trying to load it into the cube it is giving an error.
    The cube is  a real time cube for PLANING. I have chnaged the status to allow data loading but still the DTP is giving an error.
    It shows an error "error while extracting from DataStore" and some RSBK 224 ERROR and rsar 051 error.

    What was the resolution to this issue.  We rae having the same issue only with external system (not a flat file).  We get the  RSAR 051 with a return code of 238 error message, like it is not even getting to the rfc connection (DI_SOURCE).  We have been facing this issue for a while and even opened up a message with SAP.

  • Error in Data Load - Transformation inactive

    Hi
    While running DTPs to DSO or cube I am getting error saying that a  Transformation called up by the request is inactive and hence the request cannot be executed.
    But when I checked, I found all the transformations in the flow for the Data Load  were active with Executable version = Active version.
    I had been getting such error before also but at the time I executed the DTP using a Process Chain, then it ran fine without any errors, but now even the Process Chauin run is giving error.
    Does anyone have an idea about what this error exactly is or the solution for the same?
    Is this got anything to do with authorization issues?
    Thanks,
    Ninad

    I had an experience with similar issue
    Transformation Inactive;request cannot be executed
    Message no. RSBK257
    By debugging that DTP load, we found the root cause of problem, which is the following.
    Error when writing the SID 10033950 (value "10033950") in table /BIC/SZPERSON
    Since, a particular record can't be deleted from PSA, I modified that record in PSA and tried a load, but no use.
    To fix this, I deleted (from SE14) complete data from the above SID table, and scheduled the main PC in Dev, and it worked fine this time.

  • Short Dump Error While Data load

    Hi Experts,
    Data load to an ODS in production server has been failed because of Short dump error. The error message shows " OBJECTS_OBJREF_NOT_ASSIGNED ".
    Request you to please help me out with a solution ..
    Regards,
    Vijay

    Hi Vijay,
    follow the steps below (May Help you)
    Goto Monitor screen>Status Tab> using the wizard or the menu path >Environment -> Short dump> In the warehouse
    Select the Error and double click
    Analyse the error from the message.
    1.-->Go to Monitor
    -->Transactional RFC
    -->In the Warehouse
    -->Execute
    -->EDIT
    -->Execute LUW
    Refresh the Transactional RFC screen and go to the Data Target(s) which have failed and see the status of the Request-->Now it should be green.
    2.In some cases the above process will not work (where the Bad requests still exists after the above procedure).
    In that case we need to go to the Data Targets and need to delete the bad requests and do the update again
    Regards,
    BH

  • Error in data loading of hierarchy 0GL_ACCOUNT

    Dear experts,
    I have one issue related to the hierrachy data loading. I have one info object 0GL_ACCOUNT on which i activated the hierarchy data source 0GL_ACCOUNT_T011_HIER. When i start to upload the data for hierarchy i got an error message :
    Error in the hierarchy structure
    Node [00000020, 00000017 ]: Leaf '1000 ' already exists as child of node 00000014
    Please help me out from this problem.
    I also searched SDN for the same issue, but get the solution.
    Thanks in advance
    Neha

    Hii
    Its not a daily delta load, its a full load & i am working on this info object hierarchy for the first time. it successfully uploads the data in devt system but not on other ones. Also I have checked with the r/3 system, the hierarchy is correct there.
    When i uload the hierarchy upto PSA level, it does not show any error, but if i upload it till info object it displays an error message as in first thread.
    What to do??
    Thanks
    Neha

  • Critical Error with Bank Information Process (Personal Information Area)

    Hi All,
    I am experiencing a critical error when I open the Bank Information details in the portal.
    A critical error has occured. Processing of the service had to be terminated. Unsaved data has been lost.
    Please contact your system administrator.
    Syntax error in program /1PAPA/FEAT000DTAEL, error key: RFC_ERROR_SYSTEM_FAILURE
    I have read the following threads Error in the Bank Details in the Personal information and Error in Change own data but none have assisted us with our problem.
    We have also looked at SAP Note 1018036 but this only applies to HP14 and we are currently on HP26 (ECC6.0).
    Everyone's help will be much appreciated.
    Thanks in advance.
    Oloy.

    Resolved, but can't remember how - need to close thread to raise new issues.

Maybe you are looking for