Delta loading from ODS to InfoCube

Hai
Im extracting the data from the custom table into ODS and doing full load everytime for loading delta.
how can load the only delta from the ODS to InfoCube. what are the setting for this one
pls let me know
kumar

Hai <b>Ravi</b>...
<u><b>Load the Data From ODS to InfoCube</b></u>:
Clck as right Click in the Specified ODS.....
Select the Option " Update ODS Data in Data Target "......
Then You able to load the Data to any type of DataTarget's(either ODS, InfoCube)
<i><b>removed</b></i>...
<u><b>BalajeeKannan</b></u>

Similar Messages

  • Load from ODS into InfoCube gives TIME-OUT runtime error after 10 minutes ?

    Hi all,
       We have a full load from ODS into InfoCube and it was working fine till the last week upto with 50,000 records. Now, we have around 70,000+ records and started failing with TIME_OUT runtime error.
       The following is from the Short Dump (ST22):
       The system profile "rdisp/max_wprun_time" contains the maximum runtime of a
    program. The current setting is 600 seconds. Once this time limit has been exceeded, the system tries to terminate any SQL statements that are currently being executed and tells the ABAP processor to terminate the current program.
      The following are from ROIDOCPRMS table:
       MAXSIZE (in KB) : 20,000
       Frequency       :  10
       Max Processes : 3
      When I check the Data Packages under 'Details' tab in Monitor, there are four Data Packages and the first three are with 24,450 records.  I will right click on each Data Package and select 'Manual Update' to load from PSA. When this Manual Update takes more than 10 minutes it is failing with TIME_OUT again.
      How could I fix this problem, PLEASE ??
    Thanks,
    Venkat.

    Hello A.H.P,
    The following is the Start Routine:
    PROGRAM UPDATE_ROUTINE.
    $$ begin of global - insert your declaration only below this line  -
    TABLES: /BIC/AZCPR_O0400, /BIC/AZCPR_O0100, /BIC/AZCPR_O0200.
    DATA: material(18), plant(4).
    DATA: role_assignment like /BIC/AZCPR_O0100-CPR_ROLE, resource like
    /BIC/AZCPR_O0200-CPR_BPARTN.
    $$ end of global - insert your declaration only before this line   -
    The follow definition is new in the BW3.x
    TYPES:
      BEGIN OF DATA_PACKAGE_STRUCTURE.
         INCLUDE STRUCTURE /BIC/CS8ZCPR_O03.
    TYPES:
         RECNO   LIKE sy-tabix,
      END OF DATA_PACKAGE_STRUCTURE.
    DATA:
      DATA_PACKAGE TYPE STANDARD TABLE OF DATA_PACKAGE_STRUCTURE
           WITH HEADER LINE
           WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
    FORM startup
      TABLES   MONITOR STRUCTURE RSMONITOR "user defined monitoring
               MONITOR_RECNO STRUCTURE RSMONITORS " monitoring with record n
               DATA_PACKAGE STRUCTURE DATA_PACKAGE
      USING    RECORD_ALL LIKE SY-TABIX
               SOURCE_SYSTEM LIKE RSUPDSIMULH-LOGSYS
      CHANGING ABORT LIKE SY-SUBRC. "set ABORT <> 0 to cancel update
    $$ begin of routine - insert your code only below this line        -
    fill the internal tables "MONITOR" and/or "MONITOR_RECNO",
    to make monitor entries
       clear DATA_PACKAGE.
       loop at DATA_PACKAGE.
          select single /BIC/ZMATERIAL PLANT
             into (material, plant)
             from /BIC/AZCPR_O0400
             where CPR_EXT_ID = DATA_PACKAGE-CPR_EXT_ID
             and ( MATL_TYPE = 'ZKIT' OR MATL_TYPE = 'ZSVK' ).
           if sy-subrc = 0.
              DATA_PACKAGE-/BIC/ZMATERIAL = material.
              DATA_PACKAGE-plant = plant.
              modify DATA_PACKAGE.
              commit work.
           endif.
           select single CPR_ROLE into (role_assignment)
                         from /BIC/AZCPR_O0100
                         where CPR_GUID = DATA_PACKAGE-CPR_GUID.
            if sy-subrc = 0.
              select single CPR_BPARTN into (resource)
                         from /BIC/AZCPR_O0200
                         where CPR_ROLE = role_assignment
                         and CPR_EXT_ID = DATA_PACKAGE-CPR_EXT_ID.
                   if sy-subrc = 0.
                      DATA_PACKAGE-CPR_ROLE = role_assignment.
                      DATA_PACKAGE-/BIC/ZRESOURCE = resource.
                      modify DATA_PACKAGE.
                      commit work.
                   endif.
              endif.
           clear DATA_PACKAGE.
           endloop.
    if abort is not equal zero, the update process will be canceled
      ABORT = 0.
    $$ end of routine - insert your code only before this line         -
    Thanks,
    Venkat.

  • Delta load from ODS to cube failed - Data mismatch

    Hi all
    We have a scenario where the data flow is like
    R/3 table - >dataSrc -- > pSA - >InfoSrc -> ODS ->Cube.
    The cube has an additional field called "monthly version"
    and since it is a history cube , it is supposed to hold data
    snapshots of the all data in current cube for each month .
    We are facing the problem that the Data for the current month
    is there in the history ODS but not in the cube . In the ODS ->Manage->requests
    tab i can see only 1 red request that too with 0 recs.
    However ,In the cube -> manage-> reconstruction tab , i can see 2 Red request
    with the current month date. Could these red requests be the reason for
    data mismatch in ODS and cube .
    Please guide as to how can i solve this problem .
    thanks all
    annie

    Hi
    Thanks for the reply.
    Load to Cube is Delta and goes daily .
    Load to ODS is a full on daily basis .
    Can you help me how to sort this issue . I have to work directly in production env . so has to be safe and full proof .
    Thanks
    annie

  • Debugging data load from ODS to InfoCube?

    Hi Expert,
    I got some errors when I load data from ODS to Cube. I tried to simulate the load, but I got No data in PSA table. I checked mt Infopakage, it is only update Datatarget directly. How can I debug in this situation. I have only that choice "Data Target only" How can I select update Cube through PSA? Thanks in advance!
    Weidong

    Hi robert,
    Thanks for your reply! Do I have to delete the old one? It looks like that system generated two Infopackages - One is full update and another one is Initial update. I think that I have to delete both of them and re-created both, right? Thanks  a lot!
    Weidong

  • Data Load from ODS to Infocube

    Hi All,
    I have been trying to load Data from Standard ODS (activated from Business Content) to InfoCube (activated from Business Content). The records are present in ODS and it has been activated. However, when the load was scheduled and run for the Infocube, the data did not show up in the "Infocube" content and "Fact Table" as well. When I looked at Infocube->Manage->Requests, it showed transferred records as 21 (same as shown in the "monitor" of the existing request) and added records as 0. Similarly for other requests in the cube, it showed the same way. The only difference was the number of records (eg 178 or 48, etc). The number of entries for Fact table was 0 and the it is not compressed. The cube name is 0BBP_C02. Can any of you come up with an idea towards the solution to this problem.
    Thanks

    Hi Saura,
    Check start routines in the cube URs.
    For example, in my system for 0BBP_SC ODS object as infosource I see the start routine:
      LOOP AT DATA_PACKAGE.
        IF NOT ( DATA_PACKAGE-BBP_RELSC = 'I1129'
    *    if shopping cart is approved
          AND DATA_PACKAGE-BBP_ORIGIN = 'B' ).
    * and record originates in BBP
          DELETE DATA_PACKAGE.
        ENDIF.
      ENDLOOP.
    As you can see, records not conforming some conditions are deleted.
    The similiar routines you can find in another URs.
    Best regards,
    Eugene

  • Error while loading the data from ODS to InfoCube

    hai
    Im trying to load the data from ODS to InfoCube for particular year .
    But it says that there is a source system problem .
    why it is like that .
    pls tell me
    i ll assing the points
    rizwan

    Hi Rizwan,
    you didn't mention the error message in details. there could be a few places to be checked:
    - check if BW itself source system is active and in tact and reactivate if necessary
    - check if update rule is active and reactivate if necessary
    - check if ODS is active and reactivate if necessary
    Regards,
    Lilly

  • Error While Loading data from ODS to Infocube

    I am trying to load data from ODS to Infocube  thru Update ODS data into data target. My requirement was to take a small subset of fields from ODS and design IC and load the data.
    My load fails at Extraction process as I get 0 records from total number of records sent in each package. Please let me know if you need more information.
    Please advise.
    Thanks,
    RR

    In Details tab of monitor, in extraction step,
    Extraction(messages): Errors occured
    Green Light for Data Request Received
    Green Light for Data Selection Scheduled
    Yellow for 25000 Records sent(0 records received)
    Yellow for 25000 Records sent(0 records received)
    Yellow for 15000 Records sent(0 records received)
    Green Light for Data Selection Ended
    Please let me know, If you need more information.
    Thanks,
    R R

  • Direct vs Indirect data load from ODS - InfoCube

    We know that there are two options to load data from ODS to InfoCube:
    1. Direct load:
    If the ODS setting is checked with the option of "Update data targets from ODS object automatically", then when loading data to ODS, then the data also can be loaded to the InfoCube, in this way, only one InfoPackage is needed to load data to the ODS, then the InfoCube can be filled with data automatically.
    2. Indirect load:
    If the ODS setting is NOT checked with the option of "Update data targets from ODS object automatically", then not only the InfoPackage to load the data to the ODS is needed, but also the InfoPackage to load data from the exported datasource of the ODS to the InfoCube is needed to be created in order to load the data from ODS to the InfoCube.
    I wonder what are the pro and con between the above two loading methods from ODS to InfoCube.
    Welcome to every BW expert inputs!

    HI Kevin,
    Direct Loads or rather Automated Loads are usually used where u need to load data automatically into the final data target. And it has no dependencies.This process has less of maintenance involved and u need to execute a single infopackage to send data to the final data target.
    But Indirect Loads are usually used when u have a number of dependencies and this ODs object is one of the object in a process chain. So if u require the this ODS data load wait till some other event has been executed , then indirect loads are used through process chains.
    Regards,
    JAsprit

  • Some records not transfer from ODS to infocube

    Hello BW folks ,
    We have an ODS which stores the various sales doc. types.
    We are transfering all the data from this ODS to infocube. We do not have any routine or filter conditions while loading data from ODS to infocube.
    In the update rules of infocube we do not have any routine written .
    Also no start routine is present in Infocube.
    The data is loaded successfully from ODS to infocube.
    But still a particular 'sales doc type' is not transfer from ODS to infocube.
    While this 'sales doc type' is present in ODS.
    The Sales doc type is maintained with respect to the Ordernumber. So If I check the Ordernumber of that sales doc type in ODS then the same Order Number is not present in infocube.
    Means some / few Ordernumbers also getting deleted while transfering data from ODS to infocube.
    We do not have any object in update rules as master data attribute of.
    Please suggest me what to do in this case.
    Amol.

    hello VC,
    I have checked that in ODS update rules , for key figures we have update type as Addition and the data from ODS to infocube is INIT and then delta.
    Is this the reason for Overwritting the particular sales doc type ?
    I thing we should do full upload only from ODS to infocube in this case.
    Regards,
    Amol.

  • Data tranfer from ods to infocube

    Hi Friends,
    I have created ODS and infocube, the data was loaded on ods sucessfully , but when tried to transfer the data from ods to infocube, the error comes. the update rules and data flows are shown correctly as in FUFU book.
    the errors comes as check the load and after that the infopackage ask to delete the init request in same cube.
    can anyone suggest the right way of tranferring the data from ODS to INFOCUBE and the issues to check.
    Kind Regards,
    Irfankhan

    Hi
    Recheck whether you have the request Activate data in ods (odsrcactivate data & Activate Data automatically while ods creation & odsmanageRequest tab -- request -- Activate)
    If the Data is there in Active Data table then try to update the same with Update Rules.
    While Info Package from ods--Cube -- the Info Package Selection should be Delta ( Init is already there in your cube as I have seen the error message you mentioned)
    the update From ods to Cube
    Via Info Package -- Data Target Tab -- Schedule it with options Delta --
    Info Package -- Monitor -- Update Manually (Wheel)
    Adding it in Process chain
    Hope it helps and clear

  • How to debug start routines of update rules from ODS to InfoCube

    Dear gurus,
      I have an update rule from ODS to InfoCube. I wrote a start routine in the update rule. Now I want to debug it. I went to monitor and simulate update the data package and only got the prompt "No data exists in the corresponding PSA table". So how can I debug this start routine?
      Thanks in advance.
    Jin Ming

    Jin,
    In order to use PSA between ODS and InfoCube, you may have to use an exclusive InfoPackage and load separately. In that InfoPackage, choose the radio button to use a PSA.
    I think you are currently updating the InfoCube directly without using a separate InfoPackage.
    Look for an InfoSource under DataMarts (search for 8<ODS Technical name>) and create your InfoPackage there.
    Good luck.

  • Error when loading from ODS to Cube

    Hello Friends,
    I am having trouble loading data from ODS to Infocube in my 2004S system.When loading data i get this message
    07/03/2007     13:10:25     Data target 'ODSXYZ ' removed from list of loadable targets; not loadable.
    07/03/2007     13:28:42     Data target 'ODSXYZ ' is not active or is incorrect; no loading allowed     
    I checked for ODSXYZ in my data target but there is nothing by that name.Even the infopackage doesnt have it.What needs to be done.Please help.
    Thanks.

    Its a expected behavior. When you migrate ur DS, the infopacks associated with it, will grey out all the data targets that they were feeding before, that applies to any infopack you create even after the migration. You womt be able to delete it.
    Having said this, this shouldnt impact ur loads from ODS to Cube. As this shoudl be taken care of by ur DTP's rather than ur infopacks.
    A few questions:
    How are you loading ur cube?
    Did the data get thru fine to PSA with the infopack in question?
    How did you load ur DSO(assuming the load was successful)?
    Message was edited by:
            voodi

  • Shortdump problem for loadinf data from ODS to InfoCube

    hai
    im trying to load the data from ODS to InfoCube.But i got the following error like below
    Short dump in the Warehouse
    Diagnosis
    The data update was not completed. A short dump has probably been logged in BW providing information about the error.
    <b>System response
    "Caller 70" is missing.
    Further analysis:
    Search in the BW short dump overview for the short dump belonging to the request. Pay attention to the correct time and date on the selection screen.
    You get a short dump list using the Wizard or via the menu path "Environment -> Short dump -> In the Data Warehouse".
    Error correction:
    Follow the instructions in the short dump.</b>
    I looked at the shortdump.But it says that there is no shortdump for that particular date selection.
    pls tell me wht i have to do
    i ll assing the points
    bye
    rizwan

    Hi Rizwan,
    Why does the error occurs ?
    • This error normally occurs whenever BW encounters error and is not able to classify them. There could be multiple reasons for the same
    o Whenever we are loading the Master Data for the first time, it creates SID’s. If system is unable to create SID’s for the records in the Data packet, we can get this error message.
    o If the Indexes of the cube are not deleted, then it may happen that the system may give the caller 70 error.
    o Whenever we are trying to load the Transactional data which has master data as one of the Characteristics and the value does not exist in Master Data table we get this error. System can have difficultly in creating SIDs for the Master Data and also load the transactional data.
    o If ODS activation is taking place and at the same time there is another ODS activation running parallel then in that case it may happen that the system may classify the error as caller 70. As there were no processes free for that ODS Activation.
    o It also occurs whenever there is a Read/Write occurring in the Active Data Table of ODS. For example if activation is happening for an ODS and at the same time the data loading is also taking place to the same ODS, then system may classify the error as caller 70.
    o It is a system error which can be seen under the “Status” tab in the Job Over View.
    What happens when this error occurs ?
    • The exact error message is “System response "Caller 70" is missing”.
    • It may happen that it may also log a short dump in the system. It can be checked at "Environment -> Short dump -> In the Data Warehouse".
    What can be the possible actions to be carried out ?
    • If the Master Data is getting loaded for the first time then in that case we can reduce the Data Package size and load the Info Package. Processing sometimes is based on the size of Data Package. Hence we can reduce the data package size and then reload the data again. We can also try to split the data load into different data loads
    • If the error occurs in the cube load then we can try to delete the indexes of the cube and then reload the data again.
    • If we are trying to load the Transactional and Master Data together and this error occurs then we can reduce the size of the Data Package and try reloading, as system may be finding it difficult to create SID’s and load data at the same time. Or we can load the Master Data first and then load Tranactional Data
    • If the error is happening while ODS activation cause of no processes free, or available for processing the ODS activation, then we can define processes in the T Code RSCUSTA2.
    • If error is occurring due to Read/Write in ODS then we need to make changes in the schedule time of the data loading.
    • Once we are sure that the data has not been extracted completely, we can then go ahead and delete the red request from the manage tab in the InfoProvider. Re-trigger the InfoPackage again.
    • Monitor the load for successful completion, and complete the further loads if any in the Process Chain.
    (From Re: caller 70 missing).
    Also check links:
    Caller 70 is missing
    Re: Deadlock - error
    "Caller 70 Missing" Error
    Caller 70 missing.
    Bye
    Dinesh

  • Table on delta loads from data mart

    Hi,
    I am loading data from two DSO's (let's call them A and B) to another DSO (B) in BI 7 with a BW3.5 delta infopackage.
    Now I want to know where I can find information on the timestamp or last request from the last delta load from A and B to C.
    So in fact I would like to know how the system knows which requests in A en B have not been loaded to C yet at the next delta load.
    In which table can I find the information for ODS A and B that is used by the system to define what data in the change log has been loaded or not to ODS C or other targets. (In fact I should have a comparison table in BW as ROOS* in R3)
    Thanks in advance!
    Kind regards,
    Bart

    Hi Guys,
    Thanks for the answers.
    I know how to check everything in the Workbench, but I want to know where the information of the delta is stored technically.
    Just for the sake of completeness:
    Due to some issues; several successive loads from A and B were correctly loaded into DSO C (the 'new' table), but could not be activated. It is not possible to do a repeat or anything else. I am not going into too much detail, but just take this for granted.
    The only way we can 'solve' the problem is to make the system believe that the 3 last loads (activated data) in A and the two last loads in B have not been loaded to C yet. Just deleting the last delta's in C and do a new delta from A and B to C will not work.
    Therefore I want to 'manipulate' the table that is being read by a delta load. If I can change the timestamp or request numbers in that table, I can make the system believe that some requests have not been loaded to C yet.
    Dirty, but it should work I think. But I am still figuring out what table contains information about the datamart data source (8A and 8B) and the last delta load to the targets.
    Hope this is more clear.
    Thanks in advance!
    Kind regards,
    Bart

  • Error in Process Chain From ODS to InfoCube

    Hi There
    What variants should follow to load data from ods to infocube in process chains?
    I created Process Chain,first delete content of ods and data load,Activate ods , further update ods , delete index , load from ods to IC and generate index and it's warning message.
    Advice me regarding this.
    Regards,
    Chandu.

    HI Chandu,
    Please remove the further update process from the chain.
    The chain should look like...
    startdelete contents of the ODS-load to ODSactivate ODS-delete index of cubeload from ODS to Cubereconstruct indices.
    If its only a warning msg ...u can neglect...try activating the process chain and run it with an immediate option.
    By the way what does the warning msg say??
    There should be a process.....something like this ...then neglect it..
    Regards,
    Marc
    Message was edited by: Marc
    Message was edited by: Marc

Maybe you are looking for