Updating Data from OLAP Cube

Hi guys,
i have completed this tutorial ["Building OLAP 11g Cubes" |http://www.oracle.com/technology/obe/olap_cube/buildicubes.htm] followed by ["Lesson 1: Creating OBIEE Metadata for OLAP 11g Cubes"|http://www.oracle.com/technology/obe/olap_biee/CreateBIEEMetadata.htm] . But i notice whenever new data is added to my database , the Data isnt showing on my dashboard. Unless i go back to Analytic Workspace manager and redo the step of maintaining the cube and recreate the obiee metadata repository again with Oracle BI adminstration. New Data will be then show on my dashboard. Is a way for me to edit the settings or something so that the dashboard will be able to show the data once i just update the database.
Thanks for the help
Jerald

Hi,
When any data is added to the relational db you need to run the cube maintainance to load that data into the cube. While loading itself the materialized views are refreshed to represent the latest dataset.
I think for OBIEE you need to refresh the repository to reflect the changes but you neednot to rebuild it.
Thanks,
Brijesh

Similar Messages

  • Update data from cube to cube in BW 3.5

    Dear Experts,
            Can we update data from one cube to another cube in BW 3.5. i tried it, I created the connection between two cubes as i did in ods to cube through update rules but there is no option like in ods to send the data(update dso data to data targets).Please help me on this.

    hi arvind,
           Thanks, but i have some doubt again, i did it and also data transfered to the second cube, but after creating the connection i went to bi7 in infosource and there i found that cube name starting with 8(cube name), but my question is can i do it in bw3.5 without going to rsa1.

  • Move data from one cube to another cube

    Hi,
    I am on BW 3.5 and I have moved the data from one cube to another cube and found that the number of records in the original cube does not match to the newly created cube. for eg. if the original cube contains 8,549 records then the back up cube contains 7,379 records.
    Please help me on what I need to look on and if in case the records are getting aggregated then how do I check the aggregating record.
    Regards,
    Tyson

    Dear tyson m ,
    check with any update rules in ur transfer.If so check in it.
    Just go through these methods for making transfer from one cube to another cube fully without missing data.
    Update rules method
    if it's updated from ods, you can create update rules for cube2 and update from ods
    or you can try datamart scenario
    cube1 right click 'generate export datasource'
    create update rules for cube2, assign with cube1
    rsa1->source system->bw myself, right click 'replicate datasource'
    rsa1-> infosource -> search 8cube1 name
    (if not get, try right click root note 'infosource'->insert lost node(s)
    from that infosource, you will find assigned with datasource, right click and 'create infopackage', schedule and run.
    Copy from
    While creating the new cube give the cube name in the "Copy from" section. It would copy all the characteristics and Key figures. It would even copy the dimensions and Navigational attributes
    Another option is:
    The steps for copying the contents of one cube to another:
    1. Go to Manage -> Recontruct of the new cube.
    2. Select the "selection button"(red , yellow, blue diamond button).
    3.In the selection screen you can give the technical name of the old cube, requests ids you want to load, from & to date.
    4.Execute and the new cube would be loaded.
    Its all that easy!!!!!!
    Refer this link:
    Copying the structure of an Infocube
    Reward if helpful,
    Regards
    Bala

  • Can I store only the aggregate data in OLAP cube

    Hi All,
    I know that the OLAP cubes store the leaf data and then builds the aggregate data on top of it and stores with in it. I have huge amounts of data ( like billions of data in my fact table and 6-8 dimension tables). If you keep the leaf data along with the agg data with in the cube would be too large to build.
    So I am thinking to have to store only the agg data within the OLAP cube and for the leaf data it should still read from the Relational tables. Something like Hybrid OLAP.
    (what I mean is
    1. create the dimensions and cube in the AWM on 11g.
    2. I will also specifiy the levels that I want the agg data to be calculated and stored in the OLAP cube.
    (what I want is
    1. I want to store only the agg data in the cube and leaf data still in the relatlional tables.
    2. when I read the cube and drill down to the leaf level , it should fetch the leaf level data.
    Is this possible in Oracle OLAP, if yes please provide some pointers
    Thanks
    S

    First you should try out storing and aggregating data to atleast see if the cube-loading time, query-time and AW-size is within acceptable limit or not. 11gOLAP especially 11gR2 OLAP is very efficient on all three fronts.
    Regarding specifying levels, you can either use Cost-based-aggregation and pick the %age that should be pre-aggregated OR you can use Level-based-aggregation and pick the levels from each dimension that should be pre-aggregated.
    Try these out before venturing into anything else. You will be surprised by the results. There are other ways to store the data in smaller multiple-cubes and joining those cubes through formulas. Make sure you don't mistake an attribute for a dimension.
    Not sure what you mean by just storing aggregated data in olap cubes. You can just do a SUM .. GROUP BY .. of relational data before loading it into olap cubes. For example, if you source data is at DAY level, you can SUM.. GROUP BY .. at MONTH-level and then load month-level data into olap cubes where leaf-level is month-level.
    The sql view (used by reporting tools) could then be a join between month-level "olap" data and "day-level" relational data. When you are drilling-down on the data, the sql view will pick up data from appropriate place.
    One more thing. Drill-Thru to detail data is a functionality of reporting tools also. If you are using OBIEE or Discoverer Plus OLAP, then you can design the reports in a way that after you reach the olap leaf-level, then it will take the user to a relational detail-report.
    These are all just quick suggestions (on a Friday evening). If possible, you should get Oracle OLAP Consulting Group help, who can come up with good design for all your reporting needs.

  • Reading aggregated data from a cube/multiprovider

    Hi BI people
    My project is currently looking for a functionmodule that reads aggregated data from a cube/multiprovider.
    I already have a functionmodule that reads data from a cube and returns it in a flat format. I have debugged this, but have not found any flags that can enable the OLAP functionality needed to perform the aggregation. The functionmodule is "RSDRI_INFOPROV_READ_RFC".
    The situation is that I need to apply the aggregation logic of a profit center hierrarchy to the data I read from RSDRI_INFOPROV_READ_RFC, this means manually replicating the the OLAP engine functionality (keyfigure aggregation exception, ect.) and this is not an option with the available time/budget.
    Please have a look at the example below:
    Say that I have a profit center hierarchy as displayed below (with postable nodes).
    PC1 - $10
         |---- PC2  - $30
         |---- PC3  - $20
    The data I'm getting back from the functionmodule RSDRI_INFOPROV_READ_RFC looks like this:
    PC1 $10
    PC2 $30
    PC3 $20
    But I need the data aggregated. An aggregation utilizing the hierarchy above will make the data look like this:
    PC1 $60
    PC2 $30
    PC3 $20
    Instead of building an aggregation program, it would be usefull if it was possible to extract aggregated data.
    Any comments appreciated.
    Regards
    Martin

    Thx Olivier,
    The problem is that I need a functionmodule that can apply the OLAP aggregation for a hierarchy to the data outpu from RSDRI_INFOPROV_READ_RFC.
    ... or the best alternative would be if there were a fm/class that could provide me with the hierarchy aggregation of the data.
    /Martin

  • Deleting data from a cube

    Hi,
    I added a info object to a dimension in a cube. now i want to delete the info object from the cube. But since the cube has data in it, i cannot do it unless i delete all the data from the cube. When i select 'delete data', i'm given 2 options, delete fact table only and 2nd option delete fact table and dimension table. what is the difference between the two and what option should i select.
    i then need to load the cube back from ODS. the cude right now is setup in the proce4ss chain to do a delta load every day. would this be affected after i delete the data and load it again from the ODS?
    Thanks
    Sameer

    The way I understand your question is...
    1) You have data in the cube from which you have to remove an infoobject
    Ans: To acheive this, you have to delete the data from the cube
    2) You have deleted the data and then deleted the IO, activated the cube and the update rules from the ODS which feeds this cube
    3) Now you have to reload the deleted data from the ODS. But you already did an init and you are loading delta's from this ODS to the cube through process chain.
    Ans: As you've already deleted the whole data from the cube, do a full load from the ODS (you can do this even when delta initialization is done) to your cube by creating a new infopackage under the export datasource of the ODS. This will capture all the records from the ODS. Then to be on the safe side so as not to load any duplicate records...go the infopackage that you've used to do the full load (you can even create a new Info package)...SCHEDULER -> Initilization options for source system -> Delete the existing INIT -> Then do Init again...
    Leave the delta infopackage as is and let the process chain take care of the deltas from now on...
    Hope this helps. Let me know I am missing something here...

  • Error while updating data from DataStore object

    Hi,
    Currently we are upgrading BW3.5 to BI7.0 for technical only,
    we found and errors during process chain run in further processing step. This step is basically a delta loading from DSO to Cube.
    The error message are:
    Error while updating data from DataStore object 0GLS_INV
    Message no. RSMPC146
    Job terminated in source system --> Request set to red
    Message no. RSM078
    That's all no further errors message can be explained clearly here from system.
    I have applied SAP note 1152453 and reactivate the datasource, infosource, and data target.
    Still no help here!?
    Please advise if you encountered these errors before.
    Thanks in advance.
    Regards,
    David
    Edited by: David Tai Wai Tan on Oct 31, 2008 2:46 PM
    Edited by: David Tai Wai Tan on Oct 31, 2008 2:50 PM
    Edited by: David Tai Wai Tan on Oct 31, 2008 2:52 PM

    Hi Vijay,
    I got this error:
    Runtime Errors         MESSAGE_TYPE_X      
    Date and Time          04.11.2008 11:43:08 
    To process the problem further, contact you SAP system       
    administrator.                                                                               
    Using Transaction ST22 for ABAP Dump Analysis, you can look  
    at and manage termination messages, and you can also         
    keep them for a long time.                                   
    Short text of error message:                                             
    No start information on process LOADING                                                                               
    Long text of error message:                                              
      Diagnosis                                                               
          For process LOADING, variant ZPAK_0SKIJ58741F4ASCSIYNV1PI9U, the    
          end should be logged for instance REQU_D4FIDCEKO82JUCJ8RWK6HZ9KX    
          under the log ID D4FIDCBHXPLZMP5T71JZQVUWX. However, no start has   
          been logged for this process.                                       
      System Response                                                         
          No log has been written. The process (and consequently the chain)   
          has been terminated.                                                
      Procedure                                                               
          If possible, restart the process.                                   
      Procedure for System Administration                                                                               
    Technical information about the message:                                 
    Message class....... "RSPC"                                              
    Number.............. 004                                                 
    Variable 1.......... "D4FIDCBHXPLZMP5T71JZQVUWX"                         
    Variable 2.......... "LOADING"                                           
    Variable 3.......... "ZPAK_0SKIJ58741F4ASCSIYNV1PI9U"                    
    Variable 4.......... "REQU_D4FIDCEKO82JUCJ8RWK6HZ9KX" 
    Any idea?

  • Loading Data from one Cube into another Cube

    Hi Guys,
    I am trying to load data from one cube A to another cube B. Cube A has data around 200,000 records. I generate export datasource on Cube A. Replicated the datasource and created InfoSource and activated it.
    I created update rules for Cube B selecting Source as Cube A. I do have a start routine to duplicate records in Cube A. Now when I schedule load,
    It stops at Processing Datapacket and says no data. Is there something wrong with the update routine or is there any other way to load form cube to cube in a simpler way?
    Thanks in advance

    This is the start routine to duplicate records in two currencies.
    DATA: datew   TYPE /bi0/oidateto,
          datew2  TYPE rsgeneral-chavl,
          fweek   TYPE rsgeneral-chavl,
          prodhier TYPE /bi0/oiprod_hier,
          market  TYPE /bic/oima_seg,
          segment TYPE /bic/oizsegment.
    DATA: BEGIN OF S_DATA_PACK OCCURS 0.
            INCLUDE STRUCTURE /BIC/CS8ZSDREV.
    DATA: END OF S_DATA_PACK.
    S_DATA_PACK[] = DATA_PACKAGE[].
      REFRESH DATA_PACKAGE.
      LOOP AT S_DATA_PACK.
        move-corresponding s_data_pack to DATA_PACKAGE.
        if DATA_PACKAGE-loc_currcy = 'EUR'.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalloc.
          DATA_PACKAGE-CURRENCY = 'EUR'.
          APPEND DATA_PACKAGE.
        else.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
        endif.
      ENDLOOP.
    This is to load Quantity field
    RESULT = COMM_STRUCTURE-BILL_QTY.
    This is to load Value field
    RESULT = COMM_STRUCTURE-NETVAL_INV.
    UNIT = COMM_STRUCTURE-currency.

  • Error while updating data from PSA to ODS

    Hi Sap Gurus,
    I am facing the error while updating data from PSA to ODS in BI 7.0
    The exact error message is:
    The argument 'TBD' cannot be interpreted as a number
    The error was triggered at the following point in the program:
    GP44QSI5RV9ZA5X0NX0YMTP1FRJ 5212
    Please suggest how to proceed on this issue.
    Points will be awarded.

    Hi ,
    Try to simulate the process.That can give you exact error location.
    It seems like while updating few records may be no in the format of the field in which it is updated.
    Regards
    Rahul Bindroo

  • Update data from a view

    Hi,
    trying to update data from a view with:
    - Company (table)
    - Products (table)
    In a form, the user wants to update
    e.g: both products.product_name and Company.company_name.
    Is there a way to update a view records built on 2 tables ?
    Any idea will be really appreciated
    Thks

    An other question on INSTEAD OF Trigger:
    Base Tables:
    1.
    SQL> desc pcs_companies;
    Name Null? Type
    COMPANY_ID NOT NULL NUMBER(12)
    COUNTRY VARCHAR2(350)
    COMPANY_NAME VARCHAR2(320)
    COMPANY_PHONE VARCHAR2 (320)
    COMPANY_FAX VARCHAR2(320)
    COMPANY_URL VARCHAR2(150)
    UPDATED_DATE DATE
    2.
    SQL> desc pcs_individuals;
    Name Null? Type
    INDIVIDUAL_ID NOT NULL NUMBER(12)
    COMPANY_ID NUMBER(12)
    FIRST_NAME VARCHAR2(320)
    LAST_NAME VARCHAR2(320)
    LOB VARCHAR2(300)
    JOB_ROLE VARCHAR2(300)
    TITLE VARCHAR2(300)
    GENDER VARCHAR2(3)
    EMAIL VARCHAR2(720)
    FAX VARCHAR2(720)
    PHONE_NO VARCHAR2(720)
    UPDATED_DATE DATE
    COUNTRY VARCHAR2(150)
    ADDRESS_1 VARCHAR2(720)
    ADDRESS_2 VARCHAR2(720)
    ADDRESS_3 VARCHAR2(720)
    CITY VARCHAR2(720)
    3. pcs_individuals.COMPANY_ID = FK, ref pcs_companies.
    4.
    SQL> CREATE VIEW V_PCS_COMPANY_IND
    AS
    SELECT
       i.INDIVIDUAL_ID,
       c.company_id,
       c.country,
       c.Company_name,
       c.company_phone,
       i.Company_id indiv_company_id,
       i.gender,
       i.first_name,
       i.last_name,
       i.lob,
       i.job_role,
       i.title,
       i.email_address,
       i.fax,
       i.phone_no ,
       i.address_1,
       i.address_2,
       i.address_3,
       i.city
    FROM
      pcs_individuals i,
      pcs_companies c
    WHERE
      i.company_id = c.company_id
    5.
    CREATE OR REPLACE TRIGGER PCS_ADMIN.PCS_COMP_IND_UPDATE_TR
    INSTEAD OF UPDATE ON PCMS_ADMIN.V_PCS_COMPANY_IND
    FOR EACH ROW
    begin
    update PCS_COMPANIES
    set
         Company_name = nvl(:new.company_name,company_name),
         company_phone = nvl(:new.company_phone,company_phone)
    where company_id = :new.company_id;
    update PCS_INDIVIDUALS
    set
         gender = nvl(:new.gender,gender),
    first_name = nvl(:new.first_name,first_name),
         last_name = nvl(:new.last_name,last_name),
         lob = nvl(:new.lob,lob),
         title = nvl(:new.title,title),
         email_address = nvl(:new.email_address,email_address),
    phone_no = nvl(:new.phone_no,phone_no),
         fax = nvl(:new.fax,fax),
         country = nvl(:new.country,country),
         address_1 = nvl(:new.address_1,address_1),
         address_2 = nvl(:new.address_2,address_2),
         address_3 = nvl(:new.address_3,address_3),
         city = nvl(:new.city,city)
    where company_id = :new.company_id;
    end PCMS_COMP_IND_UPDATE_TR;
    6.
    CREATE OR REPLACE TRIGGER PCS_ADMIN.NEW_COMPANY_ID_INSERT
    INSTEAD OF INSERT ON PCS_ADMIN.V_PCS_COMPANY_IND
    DECLARE
    ID number;
    BEGIN
    INSERT INTO pcs_companies (org_id)
    select v_pcs_comp_id_seq.nextval
    into ID
    from dual;
    :new.company_id = ID;
    INSERT INTO pcs_individuals (company_id)
    select v_pcs_comp_id_seq.nextval
    into ID
    from dual;
    :new.company_id = ID;
    end;
    My question
    On point 6:
    Assumption:
    - Company_id is PK of pcs.Company and FK in pcs.individuals
    - It should be feed by sequence (v_pcs_comp_id_seq)
    Now how can i insert the same value for company_id (current.v_pcs_comp_id_seq) in both pcs_companies and pcs_individuals ? I've tested it in the above INSTEAD OF Trigger.It failed.
    Thks for any advice,
    lamine

  • Update data from ODS to ODS with infopackage selection

    Hi,
    I am trying to update data from one ODS to another ODS with selection criteria in InfoPackage which is created manually.For Full load I can give selection criteria in InfoPackage. When I initialize data Selection is greyed out even selections for Full load exists. Please advise me how to give selections for delta loads from ODS to ODS loads.
    Thanks in advance.
    Ram

    Once you started an ODS as destination in FULL mode from a DS you cannot get back.
    So if you want to update from ODS to ODS using Change Log but considering only some data records you could create an Update Routine with a Start Routine that DELETES undesired records (e.g. DELETE DATA_PACKAGE WHERE ...) and then start an Init-Delta Loading.
    Hope it helps
    GFV

  • To load data from a cube in SCM(APO) system to a cube in BI system.

    Experts,
         Please let me know whether it is possible to load data from a cube in SCM(APO) system to a cube in BI system.If so explain the steps to perform.
    Thanks,
    Meera

    Hi,
    Think in this way,
    To load the data fro any source we need datasource Ok. You can genare Export data source for Cube in APO,  then use that datasource for BW extraction, try like this. I think it will work, in my case I'm directly loading data from APO to BW using the DataSource sthat are genaraed on Planning Area.
    Why you need to take data from APO cube?. Is there any condition for that?. If it is not mandatory, you can use the dame datasource and load the data to BW, if they have any conditions while loading the data from APO to APO cube, they you try check wherther it is possible in BW or not. If possible then you use DataSource and do the same calculation in BW directly.
    Thanks
    Reddy

  • Automatically trigger the event to load data from Planning cube to Standard Cube

    Hello,
    We have a below set up in our system..
    1. A Planning BEx query using which user makes certain entries and writes back data to the planning cube.
    2. An actual reporting cube which gets data from the planning cube above.
    Now, what we want to do is to automate the data load from Planning cube to Reporting cube.
    This involves 2 things..
    1. Change the setting " Change real time load behaviour " of the planing cube to Planning.
    2. Trigger the DTP which loads data from Planning cube to reporting cube.
    We want to automate the above two steps...
    I have tried few things to achieve the same..
    1. Created an event in SM64,
    2. In the Planning cube "Manage" Screen,  clicked on "Subsequent Processing" and provided the event details. ( Not sure if its the correct place to provide the event detail ).
    3. Wrote a ABAP program which changes the setting of the planning cube (  " Change real time load behaviour " to Loading )
    4. Created a process chain, where we have used the event as a start variant, used the ABAP program as a next step, DTP run as the last step..
    This i hoped, would trigger the event as soon as a new request comes and sits in planning cube, which in turn would trigger the process chain which would load the data from Planning cube to reporting cube..
    This is not working. I dont think the event is triggering, and even if does i am not sure whether it will start the process chain automatically. Any ideas please ?

    hi,
    try to do the transformation directly in the input cube by using CR of type exit, more details :
    http://help.sap.com/saphelp_nw70ehp2/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
    hope it helps.

  • Data from planning cube to TEXT FILE IN directories of BW

    hi all
    i have 2 transactional planning cubes in BW. the data in these cubes in bw needs to be uploaded monthly in to a single text file so that APO system can access the data for some purpose.
    my question is how to upload the data from transactional cubes to the text file, which is more detailed. is that use ful to construct a ODS or CUBE(basic) on top of the transactional cubes.
    if i keep the file in BW directory whethere APO can access that.?..
    please help me out
    praveen.

    Hi,
    You could create a multicube over your 2 transactional cubes and extract your data with a query on the multicube.
    Make sure you restrict 0REQUID 'request id' using 0S_RQMRC 'most current data' to be sure you extract the most recently planned data.
    No need to convert the transactional cubes to basic cubes.
    To download the query I think you have to use the reporting agent of the broadcaster.
    Thanks for you points if you found this helpful or if this solved your problem.
    Best Regards,
    Filip

  • Deleting Data from SSAS CUBE

    Hello Everyone,
    I have a question, we developed a Basic Cube by building dimensions & Fact from the Source table(Everything in DSV, no Physical DIM and Fact Tables). Now everything goes well.  The Source table will only have CUrrent day data.
    My Question is: Need to delete  particular day data from the cube without disturbing the Existing Data.
    Eg: When I process the data on 31st of March, cube will have only 31st. But when I process 1st , 2nd data CUbe should have 31st +2nd data. 1st Dated data should be deleted from cube,when ever I process the data I just need month end data need to be stored
    with current day data. all the other stuff should be deleted from the cube.
    if I process the Cube on 1st of May, I should only have 31st March, April 30th and May 1st data.
    Hope the Question is clear, please let me know.
    Any help/suggestions would be appreciated.
    Thanks In Advance.
    Thanks, Please Help People When they need..!!! Mark as answered if your problem is solved.

    Hi BKomm,
    I Guess the only way to handle this scenario is by using partitions.
    Create partitions for every last day of months + One additional partition for  current_date.
    your where clause for current_date partition should be somewhat like this
      Where Date = current_date And Date <> Last_Day_of_current_Month
    so that it does not duplicate data for the last day of current month.
    Saurabh Kamath
    Hello Kamat, I was looking for something else inorder to delete the existing data from a cube. But your Approach is far better than making it complex, it did not strike to my mind. Thanks will implement practically and check.
    Thanks Again.
    Thanks, Please Help People When they need..!!! Mark as answered if your problem is solved.

Maybe you are looking for

  • Trying to make a clone backup of my internal HD but my iMac no longer recognizes my external HD.

    Short story: I am trying to make a clone of my internal drive using Super Duper! but the 3 TB La Cie external hard drive I want to copy to isn't being recognized by finder or in disk utility. The external hard drive is powered and attached to the des

  • Some fields missing from the Information interface

    On iTunes 12 I can't find the "Description" field that was in the Video page. The field still shows in the list of songs, but it doesn't appear on the new "Information" interface. And I don't think this is the only field missing compared to the previ

  • IDOC outbound port - No service for system X, client Y in Integration Direc

    Hi all, We have a problem when sending an IDOC from ECC to XI to Legacy. At ECC, tx SMQS, we got the error "No service for system SAPEQ0, client 400, in Integration Directory". I don't know what we need to do in ID to provide this service and solve t

  • About free non-production Directory Server

    Hi, there, We have downloaded the free SUN ONE Directory Server 5.2 from an ISV site. They claim that it supports up to 3000 entries. We are looking in the Admin Console and could not find any indicator for how many entries we have already used and h

  • How to check file exist in applet?

    I would like to check the existance of a file from an applet. How can I achieve that? I'd tried many attempts but still fail to get an answer. Please kindly help. Below are my file structures:- folder/applet/myApplet.java folder/appletView.jsp folder