Request failing while data loading

Hi Experts,
I am extracting some data for a data source 0CO_PC_01 upto PSA but the load is failing everytime with a message "Error occurred in data selection".
When I checked the source system there is a short dump with error " Source no does not exist".
This is happening when I am loading for fisc per > 001.2014. Prior to this period the loads are getting successful. It's a FULL load.
Do you have any insights on this issue "Source no. does not exist ". Any insights/ideas will be useful.
Regards,
Naveen

Hi Naveen ,
I had similar issue, starting of this year . Can you please check if you have done initialization for this year for this datasource .
Go to Scheduler<initialization option for source system .
Check if the fiscal year has 001.2014 - 012.2014 .
If you don't have this initialized for this year , you will not be able to pull this year's data for this data source in BW . Then in that case you need to re-initialize  for fiscal year 001.2014 - 012.2030 , so that you don't need to keep doing this activity every year .
Please check and let me know so that i can share the next steps .
Thanks
Zeenath

Similar Messages

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Error in Unit conversion while data loading

    Hi,
    I have maintained DSO in Material (0MATERIAL) info object > Bex tab > Base unit of measure > ZDSO; Then loaded this ZDSO from std data source  0MAT_UNIT_ATTR so that all conversion factors into different units which are maintained in material master data in ECC would get loaded to this DSO.
    Then I have created one Conversion type (ZCON) to read source unit from record and convert it to fixed unit "ST" with reference info object as 0MATERIAL. ST is customized UOM here.
    I am using ZCON conversion type to convert Qty in base UOM to Qty in ST in Bex reports under conversion tab of key figure. Now as this is std functionally, conversion would automatically takes place using ZDSO (mentioned above) as source and target UOM are not from same dimension (Base UOM is EA and target UOM is ST).
    If conversion factor to ST is not found in ZDSO then conversion to base UOM would happen automatically. Now this functionality is happening perfectly in Bex but its giving error if I use the same conversion type ZCON while data loads. Its giving error for those material for which ST conversion is not maintained. But when its not maintained by default it should convert it to base UOM, but its not converting and giving error in data loads.
    Hope I am able to explain the issue.
    Please help me on on this issue or any way around.
    Thanks in advance.
    Prafulla

    Ganesh,
    Can you please check out the Alpha Conversion Routine and also nodeid for that infoobject..
    There might be some inconsistencies in the code..
    Hope it helps
    Gattu

  • Performance while data load

    Hello Friends,
    I an having a query regarding performance in BW 3.5.
    while data load from R3 to BW we have 4 options.
    Only PSA
    Only Data Target
    PSA and later data target
    PSA and Data target in parallel.
    In system performance point of view, which is the best option with less system resources and How.
    Your help aprreciated.
    Thanks
    Tony

    Hi ,
    for performance point of view ..
    PSA and later data target will be better option ..
    for more info check this link ..
    http://help.sap.com/saphelp_nw04/Helpdata/EN/80/1a6567e07211d2acb80000e829fbfe/frameset.htm
    Regards,
    shikha

  • Short Dump Error While Data load

    Hi Experts,
    Data load to an ODS in production server has been failed because of Short dump error. The error message shows " OBJECTS_OBJREF_NOT_ASSIGNED ".
    Request you to please help me out with a solution ..
    Regards,
    Vijay

    Hi Vijay,
    follow the steps below (May Help you)
    Goto Monitor screen>Status Tab> using the wizard or the menu path >Environment -> Short dump> In the warehouse
    Select the Error and double click
    Analyse the error from the message.
    1.-->Go to Monitor
    -->Transactional RFC
    -->In the Warehouse
    -->Execute
    -->EDIT
    -->Execute LUW
    Refresh the Transactional RFC screen and go to the Data Target(s) which have failed and see the status of the Request-->Now it should be green.
    2.In some cases the above process will not work (where the Bad requests still exists after the above procedure).
    In that case we need to go to the Data Targets and need to delete the bad requests and do the update again
    Regards,
    BH

  • Error while data loading in real time cube

    HI experts,
    I have a problem. I am loading data from a flat file.The data is loading correctly till the DSO but when i am trying to load it into the cube it is giving an error.
    The cube is  a real time cube for PLANING. I have chnaged the status to allow data loading but still the DTP is giving an error.
    It shows an error "error while extracting from DataStore" and some RSBK 224 ERROR and rsar 051 error.

    What was the resolution to this issue.  We rae having the same issue only with external system (not a flat file).  We get the  RSAR 051 with a return code of 238 error message, like it is not even getting to the rfc connection (DI_SOURCE).  We have been facing this issue for a while and even opened up a message with SAP.

  • Message while data load is in progress

    Hi,
    We are loading the data in the infocube every day twice i.e in the morning and the afternoon.
    Loading methodology is always delete and load. At any given point we have only one request in the cube. Data Load takes arround 20-30 minutes.
    When the user runs the query during the data load, he gets the message 'No Applicable Data Found'. Can anyone please advise, how do we show the proper message like 'Data is updating in the system..Please try after some time...' etc.
    We are using BEx Browser with a template and a query attached to the template.
    Please advise.
    Regards
    Ramesh Ganji

    Hi,
    Tell the time of the data load to the users so that they are aware that the loads are in progress and data will not be available for reportting as of now and prohibit themselves from running the report give a buffer time of around 15-20 mins as there might be some issue some where down the line. Ask them to run the report other than the time data loads are happening
    You could also reschedule the timings of the process chain to finish before the users comes in.
    As far as the functionaly you are referring to i am not sure if we are able to acheive this..
    Regards
    Puneet

  • Number Ranges consideration while Data Load

    Hello team
    we are upgrading from 4.6 to ecc 6.0. abaper is gonna try test load..I already configured number ranges in new system with respect to what is in 4.6. should the number ranges be deleted before the data load tesing? How will the system react if number ranges are already there?
    Thanks

    If u r number range series are not Changes than there will be no issues
    As u will mention the seiries of No Ranges in PA04 and set that series no as default value in NUMKR
    in case if u miss any of the above things in customisation than we may get of mismatch of No
    say u have a series with 100  200 but in the number ranges u given 200 300 than it may leads to trouble
    i never come across this situtation but this is just an INFO

  • Disable a portlet while data loads

    Hi,
    Is there a way to disable a portlet programatically? We load data at certain times of the day and would like to temporarily disable the portlets. Is there a call in the Soap API that I could use to do this?
    Thanks
    Rick

    Thanks for responding
    We are looking for a way to have the process that makes the changes disable the appropriate gadgets using the soapapi. In 4.5 we did it with a stored proc. This allows the folks modifying the data decide when they need to shut off access. I agree that this is an option but would mean modifying lots of gadget code versus one data load service.
    Thanks
    Rick

  • How to skip an entire data packet while data loading

    Hi All,
    We want to skip some records based on a condition while loading from PSA to the Cube, for which we have written a ABAP code in Start Routine .
    This is working fine.
    But there is a Data packet where all the records are supposed to be skipped and here it is giving Dump and Exception CX_RSROUT_SKIP_RECORD.
    The ABAP Code written is
    DELETE SOURCE_PACKAGE WHERE FIELD = 'ABC' .
    And for a particular data packet all the records satisfy the condition and gets deleted.
    Please advice how to skip the entire data packet if all the reocrs satisfy the condition to be deleted and handle the exception CX_RSROUT_SKIP_RECORD .
    Edited by: Rahir on Mar 26, 2009 3:26 PM
    Edited by: Rahir on Mar 26, 2009 3:40 PM

    Hi All,
    The Dump I am getting is :
    The exception 'CX_RSROUT_SKIP_RECORD' was raised, but it was not caught
    anywhere along
    the call hierarchy.
    Since exceptions represent error situations and this error was not
    adequately responded to, the running ABAP program 'GPD4PXLIP83MFQ273A2M8HU4ULN'
    has to be terminated.
    But this  comes only  when all the records in a particular Data Packet gets skipped.
    For rest of the Data Packets it works fine.
    I think if the Data Packet(with 0 records) itself can be skipped this will be resolved or the Exception will be taken care of.
    Please advice how to resolve this and avoid 'CX_RSROUT_SKIP_RECORD'  at earliest .
    Edited by: Rahir on Mar 27, 2009 6:25 AM
    Edited by: Rahir on Mar 27, 2009 7:34 AM

  • Error while data loading in BI

    Hi gurus,
    Our BI team is unable to load data from BI staging to BI Target.There is no data in BI Targets because of this Users can not test the BI reports. when they run it the header file status shows yellow instead of green.
    Please help.
    Regards,
    Priyanshu Srivastava

    The problem is that jobs logs cannot be written. for example for job
    BIDTPR_6018_1 .
    sm51
    M *** ERROR => ThCallHooks: event handler rsts_before_commit for event
    sm21 F6 1 TemSe input/output to unopened file.
    Can anybody tell how to resolve that.
    Regards,
    Priyanshu Srivastava

  • Update rule problem - while data load

    Hi friends,
    I got the following error while doing initialisation for 2lis_02_sgr.
    "ABORT was set in the customer routine 9998
    Error 1 in the update "
    In the forum i searched for this error and this error is something related to the start routine in my update rule.
    But i dont know whats wrong with my routine.
    Im giving the start routine below,pls go through this and give me your suggestions..
    PROGRAM UPDATE_ROUTINE.
    $$ begin of global - insert your declaration only below this line  -
    TABLES: ...
    <i>TABLES /bic/AZMM_PUR100 .
    DATA:  T_PUR1 LIKE /bic/AZMM_PUR100 OCCURS 0 WITH HEADER LINE.</i>
    $$ end of global - insert your declaration only before this line   -
    The follow definition is new in the BW3.x
    TYPES:
      BEGIN OF DATA_PACKAGE_STRUCTURE.
         INCLUDE STRUCTURE /BIC/CS2LIS_02_SGR.
    TYPES:
         RECNO   LIKE sy-tabix,
      END OF DATA_PACKAGE_STRUCTURE.
    DATA:
      DATA_PACKAGE TYPE STANDARD TABLE OF DATA_PACKAGE_STRUCTURE
           WITH HEADER LINE
           WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
    FORM startup
      TABLES   MONITOR STRUCTURE RSMONITOR "user defined monitoring
               MONITOR_RECNO STRUCTURE RSMONITORS " monitoring with record n
               DATA_PACKAGE STRUCTURE DATA_PACKAGE
      USING    RECORD_ALL LIKE SY-TABIX
               SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS
      CHANGING ABORT LIKE SY-SUBRC. "set ABORT <> 0 to cancel update
    $$ begin of routine - insert your code only below this line        -
    fill the internal tables "MONITOR" and/or "MONITOR_RECNO",
    to make monitor entries
    if abort is not equal zero, the update process will be canceled
      CLEAR: T_PUR1[] ,
             T_PUR1,
             ABORT.
      SELECT * INTO TABLE T_PUR1 FROM /bic/AZMM_PUR100.
      IF SY-SUBRC EQ 0.
        SORT T_PUR1 BY DOC_DATE
                       DOC_ITEM
                        DOC_NUM.
      ELSE.
        MONITOR-msgid = sy-msgid.
        MONITOR-msgty = sy-msgty.
        MONITOR-msgno = sy-msgno.
        MONITOR-msgv1 = sy-msgv1.
        MONITOR-msgv2 = sy-msgv2.
        MONITOR-msgv3 = sy-msgv3.
        MONITOR-msgv4 = sy-msgv4.
        append MONITOR.
      if abort is not equal zero, the update process will be canceled
             ABORT = 1.
      ENDIF.
       ABORT = 0.
    $$ end of routine - insert your code only before this line         -
    ENDFORM.
    Thanks & Regards
    Ragu

    thanks gimmo and a.h.p,
    i have done the correction as you said,pls verify that.
    And also kindly explain me what is the reason for this start routine,what exactly it does???
    CLEAR: T_PUR1[] ,
             T_PUR1,
             ABORT.
      SELECT * INTO TABLE T_PUR1 FROM /bic/AZMM_PUR100.
      IF SY-SUBRC EQ 0.
        SORT T_PUR1 BY DOC_DATE
                       DOC_ITEM
                        DOC_NUM.
    abort = 0.    (  added  abort = 0 as per your suggestion )
      ELSE.
        MONITOR-msgid = sy-msgid.
        MONITOR-msgty = sy-msgty.
        MONITOR-msgno = sy-msgno.
        MONITOR-msgv1 = sy-msgv1.
        MONITOR-msgv2 = sy-msgv2.
        MONITOR-msgv3 = sy-msgv3.
        MONITOR-msgv4 = sy-msgv4.
        append MONITOR.
      if abort is not equal zero, the update process will be canceled
             ABORT = 1.
    exit. ( added exit as per your suggestion )
      ENDIF.
       ABORT = 0.
    $$ end of routine - insert your code only before this line         -
    ENDFORM.
    Thanks & Regards
    ragu

  • Increse No of BGP while data load and how to bypass the DTPin Process Chain

    Hello  All,
    We want to improve the performance of the loads. Currently we are loading the data from external Data Base though DB link. Just to mention we are on BI 7 system.  We are by passing the PSA to load the data quickest. Unfortunately we cannot use PSA.  Because loads times are more when we use PSA. So we are directly accessing views on external data base.  Also external data base is indexed as per our requirement.
    Currently our DTP is set to run on the 10 parallel processes (on DTP settings for batch Batch Manager with job class A). Even though we set to 10 we can see loads are running on 3 or 4 Back ground parallel processes only. Not sure why. Does any one know why it is behaving like that and how to increase them?
    If I want to split the load into three. (Diff DTPs with Different selections).  And all three will load the data into same info provider parallel. We have the routine in the selection that will look a table to get the respective selection conditions and all three DTPs will kick off parallel as part of the process chain.
    But in some cases we only get the data for two or oneDTPs(depends on the selection conditions). In this case is there any way in routine or process chain to say that if there is no selection for that DTP then ignore that DTP or set to success for that DTP and process chain should continue.
    Really appreciate your help.

    Hi
    Sounds like a nice problemu2026
    Here is a response to your questions:
    Before I start, I just want to mention that I do not understand how you are bypassing the PSA if you are using a DTP? Be that as it may, I will respond regardless.
    When looking at performance, you need to identify where your problem is.
    First, execute your view directly on the database. Ask the DBA if you do not have access. If possible perform a database explain on the view (this can also be done from within SAPu2026I think). This step is required to ensure that the view is not the cause of your performance problem. If it is, we need to implement steps to resolve that.
    If the view performs well, consider the following SAP BI ETL design changes:
    1. Are you loading deltas or full loads. When you have performance problems u2013 the first thing to consider is to make use of the delta queue (or changing the extraction to just send deltas to BI)
    2. Drop indexes before load and re-create them after the load 
    3. Make use of the BI 7.0 write optimized DSO. This allows for much faster loads.
    4. Check if you do ABAP lookups during the load. If you do, consider loading the DSO that you are selecting on in memory and change the lookup to refer to the table in memory rather. This will save tremendous time in terms of DB I/O
    5. This will have cost implications but the BI Accelerator will allow for much faster loads
    Good luck!

  • Aggregation while data loading

    Hi All,
    I need some help for understanding how I can load the following data :
    I have following DB table structure:
    Jan ABC Company 1 01/01/2011
    Jan ABC Company 1 01/10/2011
    Feb ABC Company 1 02/15/2011
    Mar ABC Company 1 03/20/2011
    When I am loading this data to Essbase, it aggregates Row 1 and row 2 , so only 3 records will be present in Essbase. one for jan , feb and mar each.
    But since they have different dates i want them to be loaded separately.
    Please advice how can I achieve this.
    Thanks

    Essbase databases are typically used for summary level information not detail. If you want all of that detail I can give you two suggestions
    1. Use another tool
    2. Use Essbase studio to build your cube at a summary level and use drill-through to a relational source to allow the users to get to detail data to see what makes up the summary data.
    (Note, I mention Essbase studio, but you could also use FDM or a thrid party tool like Dodeca)

  • While data loading

    hello,
    i load data using rule file but when i navigate for data file
    c:/dataload file/sample.xls
    seems there is space between dataload file so when ever i load data i need to edit the name and again i need to load any solution or it mandatory
    cheers

    Essbase does not like spaces in the path names. Try changing the Dataload Files to be Dataload_Files or DataloadFiles and it will solve your problem

Maybe you are looking for