Modify data in a cube with bi beans (ROLAP)

Hi,
I have put a setCellEditingAllowed(true) on my matrix
but nothing happend when i change a value.
Can i modify data of my database (ROLAP cube) with BI Beans?
Thanks
Yann

Here is some information that may help you:
If your want to use the Crosstab or Table Bean on a Java client, there are a few things to keep in mind. There is a GridView boolean property, CellEditingAllowed, that controls whether the Crosstab or Table Beans provide cell-level editing GUI in the Databody. For example, if you have a Crosstab instance m_crosstab, you can run the following code to turn on editing:
m_crosstab.setCellEditingAllowed(true);
When the Java client Crosstab or Table goes to write back the value, it calls the setValue method on the underlying DataAccess instance. If you are implementing your own data source, you can hook your custom code into your implementation of setValue. You can also consider implementing the setAutoSubmit, isAutoSubmit, submitChanges, and dropChanges methods in the oracle.dss.util.DataAccess interface if your application needs to provide bulk updates.
If you are working with our OLAP implementation, then you can retrieve the Crosstab's DataSource property (of type oracle.dss.util.dataSource), cast it to an oracle.dss.dataSource.common.Query instance (i.e., our OLAP implementation), and then add yourself as a QueryListener. The QueryListener will get a CellsSubmittingEvent with the list of updates to be processed. You'll find the various classes and interfaces in the JavaDoc for the oracle.dss.dataSource.common package.
In our current release, the OLAP implementation does not fire a CellsSubmittedEvent because the OLAP Data Source does not support writeback to the underlying data store. We are looking at providing native writeback support for Analytic Workspaces in a later release.

Similar Messages

  • Help to fill BPS Cube with same data in a Cube with these conditions

    Hi,
    I need some help in implementing BPS in a small project. (Integrated Planning is not available).
    An existing cube, Cube1 has: Year/month, Year, char1, char2, keyfig1, keyfig2
    Keyfig1 is filled directly from R3 with actuals; keyfig2 (planned values) is  filled manually filled with a monthly flat file load.
    Now, there is a change in direction to fill keyfig2 through BPS features and bring in additional key figures all based on keyfig1.
    Cube2 has been created only for the purpose of this BPS project. Cube2 was a copy of Cube1(with no data). For Cube1, I have created a Planningarea1 and PlanningLevel1; and for Cube2, Planningarea2 and PlanningLevel2 in BPS0.
    How do I fill the BPS Cube2 with the same data as in BPS Cube1 with the following conditions:
    keyfig1 : same as source value from R3 (not modifiable)
    keyfig2 : modifiable by users only on the first and second of the month.
    keyfig3 : keyfig1 * 1.1
    keyfig4 : keyfig1 of previous Year/month 
    keyfig5 : same as source value from R3 (But modifiable)
    keyfig6 : same as keyfig5 as of last day of 20th of the current month (not modifiable)
    The goal is to create a multi planning area to join the two cubes. Hints all that will also be appreciated.
    Thanks

    Your thought of having a multi area is right.
    Create a multi area and being the basic areas to which you have assigned cube 1 and cube 2 underneath the multi area.
    UNder your planing package, create a function of type Formula and create a parameter set like this:
    = * 1.1.
    Just this one line will is enough.
    TO get keyfig 4 as previous month's key fig; you need another fox. To do this, you need to have a BPS variable to et previous month and use this variable in the parameter set.
    Your fox will be like this.
    DATA CURRMONTN TYPE 0CALMONTH.
    DATA PREMONTH TYPE 0CALMONTH.
    {KEYFIG4, CURRMONTH} = {KEYFIG1, PREMONTH}.
    To make users modify only on days 1 and 2, you need to define a dara slice.
    Ravi Thothadri

  • Loading data into Fact/Cube with surrogate keys from SCD2

    We have created 2 dimensions, CUSTOMER & PRODUCT with surrogate keys to reflect SCD Type 2.
    We now have the transactional data that we need to load.
    The data has a customer id that relates to the natural key of the customer dimension and a product id that relates to the natural key of the product dimension.
    Can anyone point us in the direction of some documentation that explains the steps necessary to populate our fact table with the appropriate surrgoate key?
    We assume that we need to have an lookup table between the current version of the customer and the incoming transaction data - but not sure how to go about this.
    Thanks in advance for your help.
    Laura

    Hi Laura
    There is another way to handling SCD and changing Facts. This is to use a different table for the history. Let me explain.
    The standard approach has these three steps:
    1. Determine if a change has occurred
    2. End Date the existing record
    3. Insert a new record into the same table with a new Start Date and dummy End Date, using a new surrogate key
    The modified approach also has three steps:
    1. Determine if a change has occurred
    2. Copy the existing record to a history table, setting the appropriate End Date en route
    3. Update the existing record with the changed information giving the record a new Start Date, but retaining the original surrogate key
    What else do you need to do?
    The current table can use the surrogate key as the primary key with the natural key being being a unique key. The history table has the surrogate key and the end date in the primary key, with a unique key on the natural key and the end date. For end user queries which in more than 90% of the time go against current data this method is much faster because only current records are in the main table and no filters are needed on dates. If a user wants to query history and current combined then a view which uses a union of the main and historical data can be used. One more thing to note is that if you adopt this approach for your dimension tables then they always keep the same surrogate key for the index. This means that if you follow a strict Kimball approach to make the primary key of the fact table be a composite key made up of the foreign keys from each dimension, you NEVER have to rework this primary key. It always points to the correct dimension, thereby eliminating the need for a surrogate key on the fact table!
    I am using this technique to great effect in my current contract and performance is excellent. The splitter at the end of the map splits the data into three sets. Set one is for an insert into the main table when there is no match on the natural key. Set two is when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the current row needs to be copied into history, setting the End Date to the system date en route. Set three is also when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the main record is simply updated with the Start Date being reset to the system date.
    By the way, I intend to put a white paper together on this approach if anyone is interested.
    Hope this helps
    Regards
    Michael

  • Read Data from "virtual" Cube with different ConsUnit Hier. Version

    Dear all,
    I got a odd request.
    I need to load data from the virutal Reporting-Bapi Cube for a certain Data Version.
    This Version has been attached to a new consUnit Hierachy a few weeks ago.
    Now we get a request to read data in the past for units which are not present in the current hierarchy anymore.
    I know that this is a general issue... but does anyone got a workaround for this?
    Thanks in advance,
    regards
    Oliver

    Dear All,
    Restamtent functionality did not help at this issue.
    What we did a Copy of the Version and attached the "old" hierachy.
    This was the only solution.
    Regards
    Oliver

  • Changing the Modified Date of a List Item using the SharePoint Client Object Model (C#) with Contribute Permission

    I have a small snippet of code that I use to update the Modified Date of a list item and it works great for users with Full Control permissions.  However, for users with just Contribute access to the site the code doesn't work.  Instead, SharePoint
    just updates the Modified Date to now.
    I did some testing, and narrowed down the specific permission level that allows updating of Modified dates and oddly enough, it's the "Manage Permissions" level.
    Has anyone run into this issue? If so, how do I work around this and update the Modified date as a user with only Contribute access to a site/library?
    Here's the code:
    DateTime Test = new DateTime(2012, 5, 4);
    ListItem li = list.GetItemById(itemID);
    li["Modified"] = Test;
    li.Update();
    ct.ExecuteQuery();
    Thanks,
    Max

    Hello,
    As a workaround you can pass admin credential in your code because as per my knowledge contributor can't update default columns like: created by, modified by, modified, created.
    ClientContext clientContext = newClientContext(siteUrl);
    ClientContext.Credentials = newNetworkCredential(UserName, Password, Domain);
    Hope it could help
    Hemendra:Yesterday is just a memory,Tomorrow we may never see
    Please remember to mark the replies as answers if they help and unmark them if they provide no help

  • Can not find data in virtual cube

    Experts,
    I can not find data in virtual cube with services function module (copying Z_RS_BCT_FIGL_DATA_GET_VC10 from RS_BCT_FIGL_DATA_GET_VC10).The standard cube is maintained with data.
    Can anybody tell me what could be the reasons the data is coming through Virtua cube.
    Thanks,
    AK

    Hi,
    i_basic_infoprov is the name of the provider where the data is selected from. This must be a basic provider that contains data. What I meant was, create a copy of fm RS_BCT_FIGL_DATA_GET_VC,1 maybe as Z_BCT_FIGL_DATA_GET_VC1, and enter the name of this fm in the properties of your virtual cube. For that fm it might also be a good idea to enter a default value for i_infoprov. And this should be the name of your virtual cube.
    regards
    Siggi

  • Loading data from Info cube to PA

    Hi All,
    I am trying to load the CVC from Infocube to Planning Area in Development environment.
    I am getting error message for mismatch of periodicity of infocube and Planning Area.
    how to resolve this. Where do we define the periodicty of infocube ? I don't think so ?
    Please help me
    Thanks a lot
    Prabhat Sahay

    Prabhat,
    I understand that you already have the data in the infocube and want to load it into the planning area.
    However it is important to ensure that the data in the cube and the planning area have the same periodicity failing which we will run into such issues.
    So pls go back to the step of defining the infocube.
    In RSA1 if you double click on the infocube name which you are using as source, you can see the details in the right. In that, look for the dimensions section, under that look for the time section. If that contains a calweek or a calmonth and you are using a fiscal variant, this issue occurs.
    You have to drop your infocube data, redefine the cube to use FISCAL PERIOD and FISCAL VARNT instead of calweek or calmonth and then load the data back again. You need to ensure you populate the fields FISCAL PERIOD and FISCAL VARNT. You might want to do that with routines in your update rules or your infosource communication structure. REmember to activate the entire structure from cube to data source.
    Once you load the data into the cube with the new setting, and then try to copy the data into planning area, it will be successful.
    To answer your other question. Yes Time Characteristics is necessary for loading data between cube and planning area.
    Hope this helps.
    Thanks
    Mani Suresh

  • MDX query performance on ASO cube with dynamic members

    We have an ASO cube and we are using MDX queries to extract data from that cube. We are doing some performance testing on the MDX data extract.
    Recently we made around 15-20 account dimension members dynamic in the ASO cube, and it is taking around 1 and a half hour for the query to run on an empty cube. Earlier the query was running in 1 minute on the empty cube when there were no dynamic members in the cube.
    Am not clear why it takes so much time to extract data from MDX on an empty cube when there is nothing to extract. The performance has also degraded while extracting data on the cube with data in it.
    Does dynamic members in the outline affect the MDX performance? Is there a way to exclude dynamic members from the MDX extract?
    I appreciate any insights on this issue.

    I guess it depends on what the formulas of those members in the dynamic hierarchy are doing.
    As an extreme example, I can write a member formula to count every unique member combination in the cube and assign it to multiple members, regardless if I have any data in the database or not, that function is going to resolve itself when you query it and it is going to take a lot of time. You are probably somewhere in between there and a simple function that doesn't require any over head. So without seeing the MDX it is hard to say what about it might be causing an issue.
    As far as excluding members there are various function in MDX to narrow down the set you are querying
    Filter(), Contains(), Except(), Is(), Subset(), UDA(), etc.
    Keep in mind you did not make members dynamic, you made a hierarchy dynamic, that is not the same thing and it does impact the way Essbase internally optimizes the database based on Stored vs dynamic hierarchies. So that alone can have an impact as well.

  • Compare data in R/3 with data in a BW Cube after the daily delta loads

    Hi Friends,
    How can I compare data in R/3 with data in a BW Cube after the daily delta loads? Are there any standard procedures for checking them or matching the number of records?

    Hi Sunil,
    If you want to check the records daily instead of checking the data in R/3 manually ......
    You can try this...
    If you have staging DSO(level 1) that means whatever data is in source system load it to Staging DSO without any routines or any modifications.
    Now load this DSO data to Cube or DSO(level 2) as per your requirement with routines etc.
    Now Staging DSO contains Source system data.
    Now the level 2 Cube or DSO contains BW data with some modifications.
    Now create a Multiprovider based on level 1 and level 2 data targets.
    Now create a report on which keyfigures you want to test the data.
    In Multiprovider there is a field called 0infoprovider in data packet dimension.
    you can drag this infoprovider to the columns and restict your keyfigures with level 1 and level 2 data targets.
    In the first column you can see the level 1 DSO data ( source system data),in the 2nd column you can see the BW data.
    Now create a formula which gives the diffrence b/n level 1 and level2.
    that is R/3 data - BW data.
    If the diffrence is zero both R/3 and BW data are same.
    if the diffrence is not eqaul to zero check whether any routine is there or not.

  • What is the use of Data packet dim with Virtual cube with services

    Hi Friends,
    Could anybody tell me what is the use of Packet dimension in Virtual cube with services , It cannot contain data persistently ,Just fetch at runtime,Then there is no Physical load happening,Then why it should be automatically created for this??
    Message was edited by: Pradeep choudhari

    Hi Maik,
    Unit dimension will not be created automatically until and unless you have keyfigure referning units..
    But my question is : Why should we have data pack dimension , which is useless in this case of Virtual cube with services.It should not be created automatically!! What do you say??

  • How to load data from a virtual cube with services

    Hello all,
    we have set up a virtual cube with service and create a BEx report to get the data from an external database. That works fine. The question is now:
    Is it some how possible to "load" the data from this virtual cube with service (I know that there are not really data...) into an other InfoCube?
    If that is possible, can you please give my some guidance how to set up this scenario.
    Thanks in advance
    Jürgen

    Hi:
    I don't have system before me, so try this.
    I know it works for Remote Cube.
    Right Click on the Cube and Select Generate Export Data Source.
    If you can do this successfully, then go to Source Systems tab and select the BW. Here, Right CLick on select Replicate DataSources.
    Next, go to InfoSOurces, click on Refresh. Copy the name of Virtual Cube and add 8 as a prefix and search for the infosource.
    If you can see it, that means, you can load data from this cube to anywhere you want, just like you do to ODS.
    ELSE.
    Try and see if you can create an InfoSpoke in Virtual Cube. Tran - RSBO.
    Here, you can load to a database table and then, from this table, you can create datasource, etc.
    ELSE.
    Create query and save it as CSV file and load it anywhere you want. This is more difficult.
    Good luck
    Ram Chamarthy

  • How to tune performance of a cube with multiple date dimension?

    Hi, 
    I have a cube where I have a measure. Now for a turn time report I am taking the date difference of two dates and taking the average, max and min of the date difference. The graph is taking long time to load. I am using Telerik report controls. 
    Is there any way to tune up the cube performance with multiple date dimension to it? What are the key rules and beset practices for a cube to perform well? 
    Thanks, 
    Amit

    Hi amit2015,
    According to your description, you want to improve the performance of a SSAS cube with multiple date dimension. Right?
    In Analysis Services, there are many tips to improve the performance of a cube. In this scenario, I suggest you only keep one dimension, and only include the column which are required for your calculation. Please refer to "dimension design" in
    the link below:
    http://www.mssqltips.com/sqlservertip/2567/ssas--best-practices-and-performance-optimization--part-3-of-4/
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • Problem while adding a new dimension in a existing cube with data in BW3.5

    Hi,
    We are trying to create a new dimension with new characteristics in a existing cube. But while activating we are getting error that "InfoCube contains data; intentional changes not permitted". Can we add a new dimension in a existing cube with data in BW3.5? If it is not possible then we can take the backup and delete the data from cube and then change the cube. But my question is that, when we will transport the cube with changes to quality then will it transport successfully or same procedure needs to be done there as well.
    Regards,
    Saikat

    hi,
    Adding new infoobject /dimension to the cube containing data in 3.5 is not possible.
    easier solution is create a new cube similar to old and add the new dimension and proceed.
    else you have to drop data and transport.
    Ramesh

  • This database was connected to a Microsoft Access Services site that is not compatible with this version of Microsof Access. You cannot make design changes or modify data in database.

    Hello All,
    I am getting this error while I am connecting my access database website for the change
    "This database was connected to a Microsoft Access Services site that is not compatible with this version of Microsof Access. You cannot make design changes or modify data in database."
    I am using SharePoint 2010 and Microsoft Office 2010.
    I have created one Contact Database Site and now I am trying to modify the layout of the form in Access. and I am getting the error that I have mentioned.
    Please provide help on it.
    Thanks in Advance....

    Hello All,
    I got the answer, we have to enable the port 80 or where the access service is deployed and we have to change Trust setting and also, enable the macro from Access.
    Hope this will help to other....

  • Modified date changes with get

    Hello,
    I am having trouble with synchronizing files with Dreamweaver CS6.
    When I Get files from the server, the modified date for all the local files changes to the date/time that I did the Get, they do not reflect the modified date/time of those on the server, even though I have not modified any of them.
    Then if I modify some of them (say 2 of 10 files) and go to synchronize, Dreamweaver will want to update all 10, not just the modified files.
    I have tried checking and unchecking "Maitain synchronizatin information" but it doesn't seem to make any difference.
    Sorry if this has been answered. I've been looking and have not found a solution.
    Any suggestions?
    Thanks in advance!

    If you "Get" all your files from the server, you are essentially saving over the top of the files you have locally, which changes the file modification date as would be expected.
    After that, all of your local files would then have the same, newer modificaiton date than anything on the server, so it would make sense that they all want to be uploaded again when you "Put" the entire site.
    If you are using Get and Put each time to try to synchronize your files, stop. Use the Synchronize function (circular arrows on the right end of the Files Window toolbar) instead. That will check the server files without transferring them to your computer and saving over your local copy and allow you to upload only the files that have changed (after the first use of Synchronize that is).

Maybe you are looking for

  • Filling checkboxes with data from database

    Hello. I need help setting a checkbox value to checked if the corresponding field has "Yes" saved in the database. I can do this with textboxes, but haven't figured out how to do it with checkboxes or drop down lists. If you can help, please do. Than

  • Iphoto will not quit or force quit

    iPhoto will not quit, will not force quit, and if I turn off my mac and turn it on again, it is still trying to import photos from my disconnected iphone6. What shall I try next??

  • CS2 Transparency

    I'm fairly green when it comes to graphic design, and even moreso regarding transparency. So, please forgive me if what I'm about to ask has been discussed ad nauseum, or is a particularly stupid question... A piece I'm working on at the moment (usin

  • Scanning a photo with my printer into Elements 8

    I know I have done this before, but now it's not working and I can't figure out why. The printer and computer talk to each other just fine for printing jobs. But when I try to scan something, the printer says "wireless or USB not connected", and Elem

  • WDS migration from 2008R2 to 2012R2

    Hi everyone, I would like to know how can I move the WDS service from server A to server B? The server A is a Windows 2008R2 with WDS service active. The server B is a Windows 2012R2. Do I just: 1) install the WDS service into 2012R2 2) import boot i