Shortdump problem for loadinf data from ODS to InfoCube

hai
im trying to load the data from ODS to InfoCube.But i got the following error like below
Short dump in the Warehouse
Diagnosis
The data update was not completed. A short dump has probably been logged in BW providing information about the error.
<b>System response
"Caller 70" is missing.
Further analysis:
Search in the BW short dump overview for the short dump belonging to the request. Pay attention to the correct time and date on the selection screen.
You get a short dump list using the Wizard or via the menu path "Environment -> Short dump -> In the Data Warehouse".
Error correction:
Follow the instructions in the short dump.</b>
I looked at the shortdump.But it says that there is no shortdump for that particular date selection.
pls tell me wht i have to do
i ll assing the points
bye
rizwan

Hi Rizwan,
Why does the error occurs ?
• This error normally occurs whenever BW encounters error and is not able to classify them. There could be multiple reasons for the same
o Whenever we are loading the Master Data for the first time, it creates SID’s. If system is unable to create SID’s for the records in the Data packet, we can get this error message.
o If the Indexes of the cube are not deleted, then it may happen that the system may give the caller 70 error.
o Whenever we are trying to load the Transactional data which has master data as one of the Characteristics and the value does not exist in Master Data table we get this error. System can have difficultly in creating SIDs for the Master Data and also load the transactional data.
o If ODS activation is taking place and at the same time there is another ODS activation running parallel then in that case it may happen that the system may classify the error as caller 70. As there were no processes free for that ODS Activation.
o It also occurs whenever there is a Read/Write occurring in the Active Data Table of ODS. For example if activation is happening for an ODS and at the same time the data loading is also taking place to the same ODS, then system may classify the error as caller 70.
o It is a system error which can be seen under the “Status” tab in the Job Over View.
What happens when this error occurs ?
• The exact error message is “System response "Caller 70" is missing”.
• It may happen that it may also log a short dump in the system. It can be checked at "Environment -> Short dump -> In the Data Warehouse".
What can be the possible actions to be carried out ?
• If the Master Data is getting loaded for the first time then in that case we can reduce the Data Package size and load the Info Package. Processing sometimes is based on the size of Data Package. Hence we can reduce the data package size and then reload the data again. We can also try to split the data load into different data loads
• If the error occurs in the cube load then we can try to delete the indexes of the cube and then reload the data again.
• If we are trying to load the Transactional and Master Data together and this error occurs then we can reduce the size of the Data Package and try reloading, as system may be finding it difficult to create SID’s and load data at the same time. Or we can load the Master Data first and then load Tranactional Data
• If the error is happening while ODS activation cause of no processes free, or available for processing the ODS activation, then we can define processes in the T Code RSCUSTA2.
• If error is occurring due to Read/Write in ODS then we need to make changes in the schedule time of the data loading.
• Once we are sure that the data has not been extracted completely, we can then go ahead and delete the red request from the manage tab in the InfoProvider. Re-trigger the InfoPackage again.
• Monitor the load for successful completion, and complete the further loads if any in the Process Chain.
(From Re: caller 70 missing).
Also check links:
Caller 70 is missing
Re: Deadlock - error
"Caller 70 Missing" Error
Caller 70 missing.
Bye
Dinesh

Similar Messages

  • Error while loading the data from ODS to InfoCube

    hai
    Im trying to load the data from ODS to InfoCube for particular year .
    But it says that there is a source system problem .
    why it is like that .
    pls tell me
    i ll assing the points
    rizwan

    Hi Rizwan,
    you didn't mention the error message in details. there could be a few places to be checked:
    - check if BW itself source system is active and in tact and reactivate if necessary
    - check if update rule is active and reactivate if necessary
    - check if ODS is active and reactivate if necessary
    Regards,
    Lilly

  • Error While Loading data from ODS to Infocube

    I am trying to load data from ODS to Infocube  thru Update ODS data into data target. My requirement was to take a small subset of fields from ODS and design IC and load the data.
    My load fails at Extraction process as I get 0 records from total number of records sent in each package. Please let me know if you need more information.
    Please advise.
    Thanks,
    RR

    In Details tab of monitor, in extraction step,
    Extraction(messages): Errors occured
    Green Light for Data Request Received
    Green Light for Data Selection Scheduled
    Yellow for 25000 Records sent(0 records received)
    Yellow for 25000 Records sent(0 records received)
    Yellow for 15000 Records sent(0 records received)
    Green Light for Data Selection Ended
    Please let me know, If you need more information.
    Thanks,
    R R

  • Index used or not for selecting data from ODS in a start routine

    Dear friends,
    In the start routine of the update rules to a cube, I am reading some data of an ODS in to an internal table .
    The ODS is indexed. But, I am not sure if the index is at all used in the Select statement (that gets the data from ODS to the internal table in the start routine) while loading data to the cube.
    Any help is highly appreciated.
    regards,
    atlaj

    Hi Atlaj
    You can findout this is display execution plan for SQL statement in DB02.
    Goto DB02. and under diagnostic, you find explain. Select that and enter your query. Make sure that everything here is in capital format. Below is a sample query which I have entered.
    SELECT "CRM_SALORG" "SALESORG" FROM "/BI0/QORGUNIT"
    WHERE "SALESORG" = ? AND "OBJVERS" = ? AND "DATETO" >= ?
    AND "DATEFROM" <= ?
    The select parameters should be inside qoutes and in caps and even the from table. Once you enter your query in this format, click on explain. It will show the index scan if the index is present. the output for my query will be something like
    0 SELECT STATEMENT ( Estimated Costs =  1,348E+01 [timerons] )
            1 (COOR) RETURN
                2 (    0) TQ
                    3 (    0) FETCH /BI0/QORGUNIT
                        4 (    0) IXSCAN /BI0/QORGUNIT~Z1
    Where my last statement (line 4) is showing index scan and the name of index read is Z1.
    Hope this helps.
    Please let me know if you have any problems entering the query in the specified format and u get any error.
    Regards
    Sriram

  • Problem while loading data from ODS to infoobject

    Hi guys,
    I am facing problem while loading data from <b>ODS to infoobject</b>.
    If I load data via PSA it works fine but
    if I load data without PSA its giving error as Duplicate records.
    Do u have any idea why it is so.
    Thanks in advance
    savio

    Hi,
    when you load the data via the PSA, what did you select? Serial or Paralel?
    If you select serial most likely you don't have duplicates within the same datapackage and your load can go through.
    Loading directly in the IObj will happen thefore if you have the same key in two different packages, the "duplicate records" will be raised; you can perhaps flag your IPack with the option "ignore duplicate records" as suggested...
    hope this helps...
    Olivier.

  • Error while laoding data from ODS to infocube

    Hi all,
    I am trying to load data from ODS to Cube but its giving error.
    I am getting following error
    An error occurred while executing a transformation rule:
    The exact error message is:
    The argument 'Result dependent booking' cannot be interpreted as a
    number
    The error was triggered at the following point in the program:
    GPD5CD2GMGXQYJ77HKMS0C0CQVE 2510
    System Response
         Processing the data record has been terminated.
    Procedure
          The following additional information is included in the higher-level
         node of the monitor:
         o   Transformation ID
         o   Data record number of the source record
         o   Number and name of the rule which produced the error
    Procedure for System Administration.
    How to resolve this error, it will be great help if anyone post the steps

    Thanks Ananda  for your suggestion,
    CAn you please tell me How to identify which rule has a problem ..is there any way which tell me which rule and which record is creating a problem

  • Problem in transfering data from DSO to InfoCube

    Hi,
    I am trying to transfer data from DSO to Infocube. But it is giving the following message.
    Last delta update is not yet completed
    Therefore, no new delta update is possible.
    You can start the request again
    if the last delta request is red or green in the monitor (QM activity)
    It is not allowing to load data to InfoCube from DSO. I checked the request number. But there is no request number in InfoCube. If request no is there I can delete the request from InfoCube and delete the data mart status in DSO and reload again.But there is no request id in IC. So plz guide me how I can tranfer data to InfoCube overcoming this problem.
    Actually further updata was running from DSO to IC. But it failed. Executing Function Module : RSPC_PROCESS_FINISH, I set the chain  green before loading
    data to InfoCube. So it starts executing the next chain. So there is no request id. So plz guide me how I can retain this request id.
    Edited by: Purushotham Reddy on Jan 28, 2008 11:19 AM

    Hi Reddy,
    Make the QM Status Red of the failed request even if it is red.
    and then run the same infopackage used in the chain. This will ask for repeat of last delta.
    If this process is understood to u then its fine otherwise read below:
    the other option is check if this failed request is present in PSA. If it is present in PSA then update the failed request from PSA, here u wont need to do a repeat of last delta. This Option is easier. this u can do from the ,manage tab of the data source I think this u might be surely knowing....
    Assign Points if it helps u.....
    Thanks,
    Nagesh.

  • JTable Refreshing problem for continuous data from socket

    Hi there,
    I am facing table refreshing problem. My table is filled with the data received from socket . The socket is receiving live data continuously. When ever data received by client socket, it notify table model to parse and refresh the data on specific row. The logic is working fine and data is refresh some time twice or some thrice, but then it seem like java table GUI is not refreshing. The data on client socket is continuously receiving but there is no refresh on table. I am confident on logic just because when ever I stop sending data from server, then client GUI refreshing it self with the data it received previous and buffered it.
    I have applied all the available solutions:
    i.e.
         - Used DefaultTableModel Class.
         - Created my own custem model class extend by AbstractTableModel.
         - Used SwingUtilities invokeLater and invokeAndWait methods.
         - Yield Stream Reader thread so, that GUI get time to refresh.
    Please if any one has any idea/solution for this issue, help me out.
    Here is the code.
    Custom Data Model Class
    package clients.tcp;
    import java.util.*;
    import javax.swing.table.*;
    import javax.swing.*;
    public class CustomModel extends DefaultTableModel implements Observer {
    /** Creates a new instance of CustomModel */
    public CustomModel() {
    //super(columnNames,25);
    String[] columnNames = {"Sno.","Symbol","BVol","Buy","Sale",
    "SVol","Last","..","...","Total Vol...",
    "..","High","..","Low","Change","Average","..."};
    super.setColumnIdentifiers(columnNames);
    super.setRowCount(25);
    /*This method called by client socket when ever it recived data as line/record*/
    public synchronized void update(Observable o, Object arg){
    collectData(arg.toString());
    /*This method used to parse, collect and notify table for new data arrival*/
    public synchronized void collectData(String rec){              
    String validSymbol="";
    System.out.println("collectDataRecord :-"+rec);
    StringTokenizer recToken=new StringTokenizer(rec," ");
    Vector rowVector=new Vector();
    for(int i=0;i<25;i++){
    validSymbol=(String)getValueAt(0,1);
    if(rec.indexOf(validSymbol) != -1){
    for(int j=0;recToken.hasMoreTokens();j++){
    super.setValueAt(recToken.nextToken(),i,j);
    //break;
    Client Socket Class
    package clients.tcp;
    import java.io.*;
    import java.net.*;
    import java.util.*;
    import javax.swing.*;
    public class DataReceiver extends Observable implements Runnable{
    private Socket TCPConnection = null;
    private BufferedReader br = null;
    private String serverName = "Salahuddin";
    private int serverPort = 5555;
    public DataReceiver()
    public void setServerName(String name){
    this.serverName=name;
    public void setServerPort(int port){
    this.serverPort=port;
    //03-20-04 12:30pm Method to open TCP/IP connection
    public void openTCPConnection()
    try
    TCPConnection = new Socket(serverName, serverPort);
    br = new BufferedReader(new InputStreamReader(
    TCPConnection.getInputStream()));           
    System.out.println("Connection open now....");
    //System.out.println("value is: "+br.readLine());
    }catch (UnknownHostException e){
    System.err.println("Don't know about host: ");
    }catch (IOException e){
    System.err.println("Couldn't get I/O for "
    + "the connection to: ");
    //03-20-04 12:30pm Method to receive the records
    public String receiveData(){
    String rec = null;
    try
    rec = br.readLine();          
    catch(IOException ioe)
    System.out.println("Error is: "+ioe);
    ioe.printStackTrace(System.out);
    catch(Exception ex)
    System.out.println("Exception is: "+ex);
    ex.printStackTrace(System.out);
    return rec;     
    public void run()
    while(true)
    String str=null;
    str = receiveData();
    if(str != null)
    /*try{
    final String finalstr=str;
    Runnable updateTables = new Runnable() {
    public void run() { */
    notifyAllObservers(str);
    SwingUtilities.invokeAndWait(updateTables);
    }catch(Exception ex){
    ex.printStackTrace(System.out);
    System.out.println("Observer Notified...");*/
    try{
    Thread.sleep(200);
    }catch(Exception e){
    public synchronized void notifyAllObservers(String str){
    try{
    setChanged();
    notifyObservers(str);
    }catch(Exception e){
    e.printStackTrace(System.out);
    Regards,
    Salahuddin Munshi.

    use your code to post codes more easy to read, lyke this:
    package clients.tcp;
    import java.util.*;
    import javax.swing.table.*;
    import javax.swing.*;
    public class CustomModel extends DefaultTableModel implements Observer {
    /** Creates a new instance of CustomModel */
    public CustomModel() {
    //super(columnNames,25);
    String[] columnNames = {"Sno.","Symbol","BVol","Buy","Sale",
    "SVol","Last","..","...","Total Vol...",
    "..","High","..","Low","Change","Average","..."};
    super.setColumnIdentifiers(columnNames);
    super.setRowCount(25);
    /*This method called by client socket when ever it recived data as line/record*/
    public synchronized void update(Observable o, Object arg){
    collectData(arg.toString());
    /*This method used to parse, collect and notify table for new data arrival*/
    public synchronized void collectData(String rec){
    String validSymbol="";
    System.out.println("collectDataRecord :-"+rec);
    StringTokenizer recToken=new StringTokenizer(rec," ");
    Vector rowVector=new Vector();
    for(int i=0;i<25;i++){
    validSymbol=(String)getValueAt(0,1);
    if(rec.indexOf(validSymbol) != -1){
    for(int j=0;recToken.hasMoreTokens();j++){
    super.setValueAt(recToken.nextToken(),i,j);
    //break;
    Client Socket Class

  • Scheduling Problem for uploading Data from Flat file to SAP

    Hi guys,
    I am facing a weared problem in uploading some leave records in z table. The code is working fine if we run it through se38 after selecting the file from a shared location from production server which has all the access rights.
    This folder lies in the \usr folder of SAP Production.
    I have kept all the Flat files in the shared path "
    Tis-mum-iz-s1\migration\SAP-INT\leave\" ...
    To give u exact directory structure..
    Tis-mum-iz-s1 is the Server Name
    usr is the SAP System folder used for uploads and downloads
    usr |
    ...-> Migration |
                      -> SAP-INT |
                                 -> leave -> (Flat Files)
    Migration folder is shared with all rights.
    Obviously, we cannot give shared drive as the variant in the scheduler.
    So i use the system path i.e. \usr\sap\tmp\migration\sap-int\leave\ as the variant.
    All my other download programs are working fine with this path as a variant...
    But my this particular upload program does not work with this path....
    I am giving u my code...
    TATA INTERACTIVE SYSTEMS (A Division of TATA INDUSTRIES LIMITED)
    REPORT      :  ZMIGRATE_ZLEAVE
    DESCRIPTION :  To Upload the Leave data. (ZLEAVE)
    CREATED BY  :  Abhishek Bachhawat
    CREATED ON  :  01.09.2005
    CONSULTANT  :  ANAND
    REPORT  ZMIGRATE_ZLEAVE.
    TABLES: ZLEAVE.
    data: begin of wtab,
              MANDT(3),
              ZLVID(8),
              PERNR(8),
              ZSTDT(8),
              ZENDT(8),
              ZDAYS(4),
              AEDAT(8),
              ERDAT(8),
          end of wtab,
          itab like WTAB occurs 0 WITH HEADER LINE.
    data: temp like zleave occurs 0 WITH HEADER LINE.
    SELECTION-SCREEN BEGIN OF BLOCK file
                   WITH FRAME TITLE text-005.
    parameters: file like rlgrap-filename Obligatory.
    Concatenate File SY-DATUM '_Leave.txt' into File.
    SELECTION-SCREEN END OF BLOCK file.
    at SELECTION-SCREEN ON VALUE-REQUEST FOR file .
      CALL FUNCTION 'WS_FILENAME_GET'
        IMPORTING
          FILENAME = file.
      IF SY-SUBRC <> 0.
      ENDIF.
    start-of-selection.
      if file ne space.
        CALL FUNCTION 'WS_UPLOAD'
          EXPORTING
            FILENAME = FILE
            FILETYPE = 'DAT'
          TABLES
            DATA_TAB = ITAB.
      else.
        message e000(zps) with 'Specify a file'.
      endif.
      SORT ITAB BY ZLVID.
      LOOP AT ITAB.
        REFRESH TEMP.
        CLEAR TEMP.
        TEMP-MANDT = sy-mandt.
        TEMP-ERDAT = SY-DATUM.
        TEMP-ZLVID = ITAB-ZLVID.
        TEMP-PERNR = ITAB-PERNR.
        TEMP-ZSTDT = ITAB-ZSTDT.
        TEMP-ZENDT = ITAB-ZENDT.
        TEMP-ZDAYS = ITAB-ZDAYS.
        TEMP-AEDAT = ITAB-AEDAT.
        TEMP-ERDAT = ITAB-ERDAT.
        APPEND TEMP.
        SELECT SINGLE *
               FROM   ZLEAVE
               WHERE  ZLVID = TEMP-ZLVID
               AND    PERNR = TEMP-PERNR.
        IF SY-SUBRC = 0.
          UPDATE ZLEAVE SET ZSTDT = TEMP-ZSTDT
                            ZENDT = TEMP-ZENDT
                            ZDAYS = TEMP-ZDAYS
                            AEDAT = TEMP-AEDAT
                            ERDAT = TEMP-ERDAT
                 WHERE ZLVID = TEMP-ZLVID
                 AND   PERNR = TEMP-PERNR.
        ELSE.
          INSERT ZLEAVE FROM TABLE TEMP.
          COMMIT WORK.
        ENDIF.
      ENDLOOP.

    Hi,
    open dataset file for input in text mode.
    check sy-subrc = 0.
    while sy-subrc = 0.
      read dataset file into wa.
      if sy-subrc = 0.
      append wa to itab.
      else.
        exit.
      endif.
    endwhile.
    close dataset file.
    regards
    Siggi
    PS: check also the F1-help for open, read and close statements!

  • Updating data from ODS to InfoCube

    Hello experts,
    I am not sure if my approach works so let me elaborate:
    ODS1 contains trips on participant level.
    The data in ODS1 looks like this:
    ODS1:
    trip_id___participant_id___price_____trip date
    __101_____1___________25.00_____20070112
    __101_____27__________12.50_____20070112
    __102_____21__________55.00_____20070124
    __102_____7___________55.00_____20070124
    __102_____15__________35.00_____20070124
    Now I want to update the data to InfoCube IC_TP which looks like this:
    IC_TP:
    dimensions:
    - time
    - participant_id (has nav. attribute participant_type on which I want to report or do I have to select participant_type instead of participant_id as dimension???)
    measures:
    - price
    - number of trips (according to the data in ODS1 it should be "2" for jannuary)
    - number of participants (according to the data in ODS1 it should be "5" for jannuary)
    is this updating approach possible?
    thanx
    hiza

    Hi,
    You have take these prperties in  the edit screen of the info objects. In the <i>Change</i> screeen , go to <i>Aggregation </i>Tab page. Here  you can find these settings.
    And try like this .
    Hi,
    Try this:
    For the KF Info object : number of trips,
    Take Exception Aggregat = <b><i>Counter(Values unequal to zeros)</i></b>
    Agg.referen.char = Trip_id
    And
    For the KF Info object : number of participants,
    Take Exception Aggregat = <b><i>Counter(Values unequal to zeros)</i></b>
    Agg.referen.char = participant_id
    With rgds,
    Anil Kumar Sharma .P
    Message was edited by:
    Anil Kumar Sharma
    With rgds,
    Anil Kumar Sharma .P<b></b>
    Message was edited by:
            Anil Kumar Sharma

  • Any table for the data in ODS, PSA, INFOCUBE

    Hi All,
          We are loading data to PSA and then to ODS and then to INFOCUBE , in such condition is there any table where we can check the number of data that can be loaded in each run into  PSA and then to ODS and then to INFOCUBE .
       When u check the data in the PSA you will see (Suppose):
                        Processing          PSA
                        Update Mode       Full update
                        Records Check    32 Records Received
    Now here it is showing the number of data loaded to PSA then there must be one table where it is loading the data ..
    Do anyone have the idea  about the table.
    Thanks,
    Gautam

    Hi,
    You can check the data in the PSA table and Active table of ODS.
    PB

  • Error passing data from ODS to IC

    Hi,
    I have a problem when passing data from ODS to Infocube. According to the messages, the problem is there is erroneous characters in one record's ZANLAGE characteristic. I check and found the problematic record. Mi question is Why the problem showed up when passing data from ODS to IC instead of when passing data from Infosource to ODS? I check de ODS and the problematic record is in: "       " (hex.). How this characteristic value entered in the ODS correctly, but it can not enter the IC??
    Mauricio

    Hi,
    I check the ODS settings and the BEx option is not checked. This may explain whay the problem showed up passig from ODS to IC and not before.
    What should I do to catch the problem and pass the data from the ODS to the IC?
    My problem is how to identify this record, can I use:
    if ZANLAGE = "       "
      ZANLAGE = "XXXXXX".
    end if.
    Remember that "       " is Hex. code!!.
    regards
    Mauricio

  • Data Extraction from ODS to Infocube

    Hi.
    How will you extract data from ODS to Infocube. what is process ?
    please explain....

    Hi,
    as the no.. of records arelot more then your usual loading ..
    so it will take mmuch more time today ..but make sure that the job is running fine..
    You can create load the data by creating selection  but at this point i dont think you can do this.. because for this you have to kill the job ..may be some changes would have been happened
    in the Ip thats why its picking more records..
    hope this helps you ...
    Reagds,
    shikha

  • Data tranfer from ods to infocube

    Hi Friends,
    I have created ODS and infocube, the data was loaded on ods sucessfully , but when tried to transfer the data from ods to infocube, the error comes. the update rules and data flows are shown correctly as in FUFU book.
    the errors comes as check the load and after that the infopackage ask to delete the init request in same cube.
    can anyone suggest the right way of tranferring the data from ODS to INFOCUBE and the issues to check.
    Kind Regards,
    Irfankhan

    Hi
    Recheck whether you have the request Activate data in ods (odsrcactivate data & Activate Data automatically while ods creation & odsmanageRequest tab -- request -- Activate)
    If the Data is there in Active Data table then try to update the same with Update Rules.
    While Info Package from ods--Cube -- the Info Package Selection should be Delta ( Init is already there in your cube as I have seen the error message you mentioned)
    the update From ods to Cube
    Via Info Package -- Data Target Tab -- Schedule it with options Delta --
    Info Package -- Monitor -- Update Manually (Wheel)
    Adding it in Process chain
    Hope it helps and clear

  • Data transfer from ODS to InfoCube

    How to transfer data from ODS to InfoCube ? Where exactly do I have to create the
    Infopackage or is it auto-generated ?

    Hi,
    If you are using BI 7.0 then do the following:
    Create Transformation from ODS to Cube by right clicking on cube.
    then create DTP in cube and execute it.
    No need to create infopackage again.
    regards,
    Pruthvi R

Maybe you are looking for

  • Windows 7 doesn't awaken with Ipod Nano 5G connected

    Brand new HP computer w/Windows 7 and brand new Ipod Nano 5G. Left ipod connected to computer to charge. After computer powered down it would not restart. HP figured out that the Ipod was blocking the restart. If I unplug Ipod computer restarts fine.

  • Sd: rebate

    Hi all, I'm working on rebate. I found this situation: for the agreement I'm working on, the Rebate recipient is different from the customer in the condition type. How that is possible? I think they must be the same could you please explain?? thanks

  • Director 11.5 and Windows 7

    Has anyone tried Director 11.5 on Windows 7 (release candidate)? Is it compatible? Windows 7 will be out in October, so it would be nice to have a compatible version of Director available at that time.

  • Sound quality and coolers  on Macbook Air

    Hello everyone ...I just bought a macbook air (1.86GHz, 128 SSD...) and I'm quite disappointed by its sound quality .... when I pull up the volume it sounds like worn-out speaker. My old macbook white sounds much much better . Also it's getting hot a

  • Save As missing from drop down menu.

    I don't have the Save As option in my file dropdown menu. I would like to save pages documents as word, I am unable to check that box its grayed out. if anyone can let me know how to resolve this issue I would greatly appreciate it. It only has a Sav