Best Practice of Essbase Cube Join Relational Table

Hi,
As the topic mentioned, what's the best practice of Essbase cube join relational table?
Please notice that I want description columns, not measure columns, from relational table.
Thanks.

As a starting point, check out what Ed wrote together a while ago:
[https://community.altiusconsulting.com/blogs/altiustechblog/archive/2008/10/24/are-essbase-and-oracle-bi-enterprise-edition-obiee-a-match-made-in-heaven.aspx|https://community.altiusconsulting.com/blogs/altiustechblog/archive/2008/10/24/are-essbase-and-oracle-bi-enterprise-edition-obiee-a-match-made-in-heaven.aspx]
[http://community.altiusconsulting.com/blogs/altiustechblog/archive/2008/11/03/enhance-your-essbase-month-dimension-to-aid-sorting.aspx|http://community.altiusconsulting.com/blogs/altiustechblog/archive/2008/11/03/enhance-your-essbase-month-dimension-to-aid-sorting.aspx]
Cheers,
C.

Similar Messages

  • How to define join in physical layer between cube and relational table

    Hi
    I have a aggregated data in essbase cube. I want to supplement the information in the cube with data from relational source.
    I read article http://community.altiusconsulting.com/blogs/altiustechblog/archive/2008/10/24/are-essbase-and-oracle-bi-enterprise-edition-obiee-a-match-made-in-heaven.aspx which describes how to do it.
    From this article I gather that I have to define a complex join between the cube imported from essbase to my relational table in physical layer.
    But when I use Join Manager I am only able to define jooin between tables from relation source but not with the imported cube.
    In My case I am trying to join risk dimension in the cube based on risk_type_code (Gen3 member) with risk_type_code in relation table dt_risk_type.
    How can I create this join?
    Regards
    Dhwaj

    Hi
    This has worked the BI server has joined the member from the oracle database to cube. So Now for risk type id defined in the cube I can view the risk type code and risk type name from the relational db.
    But now if I want to find aggregated risk amount against a risk type id it brings back nothing. If I remove the join in the logical model then I get correct values. Is there a way by which I can combine phsical cube with relational model and still get the aggregated values in the cube?
    I have changed the column risk amount to be sum in place of aggr_external both in logical and phsical model.
    Regards,
    Dhwaj

  • Best practice for SD Cube

    Hi Gurus!
    I have to active this cubes:
    Deliveries        
    0SD_C02
    Delivery Service 0SD_C04
    I can´t find the best practices for that.
    Any ideas?
    I will assign full points.
    Thanks in advance!

    Did you wanted Best Practice for Installation.
    Find out datasources feeding 0SD_C02 from                                                        <a href="http://help.sap.com/saphelp_nw2004s/helpdata/en/3d/5fb13cd0500255e10000000a114084/frameset.htm">Help</a> .
    Activate the same in ECC  - RSA5
    Replicate in BW.
    Install 0SD_C02 with Grouping as Dataflow Before

  • Integrating Essbase cubes with Oracle Tables in BI Server

    I'm trying to link together data from an aggregated Essbase Cube with some static table data from our oracle system. Both the essbase and oracle data exist correctly in their own right in the physical, business and presentation levels. Aggragted data is client sales, static data is client details
    Within the OBIEE Administration tool I've tried to drag the physical oracle table for clients onto the clients essbase section in the business area, and it seems to work OK, until you try and report on them together and I get the following error:
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 42043] An external aggregate is found in an outer query block. (HY000)
    Can anyone advise on what I'm doing wrong?
    Thanks

    Thanks Christian I found some very useful articles (one or two by you) - I'll have to look harder on the net before posting.
    One thing I found out, with respect to vertical federation, that others may benefit from, is that I fuond it much easier to start from the most detailed level and then attach the less detailed source rather start with the less detailed and add the additional on.

  • Best Practice to save record in a table with appropriate trigger

    Hi,
    At the same time, five users are inserting client information in a client table through Oracle form. Client table has the following fields:
    CLIENT ID
    CLIENT NAME
    CLIENT ADDRESS
    CLIENT ID is generating automatically by calling a procedure. In this procedure I am using MAX function to get maximum value of CLIENT ID. After that, I store newly generated CLIENT ID in a DATA BLOCK item say :MASTER.CLIENT ID.
    The problem is that all five users do get same MAX value(suppose 40) of CLIENT ID at the same time, and oracle form will surely throw an exception at the time of insertion the record in client table. CLIENT ID is PK and it is a member of MASTER DATA BLOCK.
    I hope all above will clearly illustrate the problem, further please guide, can PRE-INSERT trigger may handle this problem efficiently ? , if so, then how ? ...
    Thanks,

    Hello,
    Welcome to the forum!
    CLIENT ID is generating automatically by calling a procedure.So, in which trigger you are calling that procedure?
    After that, I store newly generated CLIENT ID in a DATA BLOCK item say :MASTER.CLIENT ID. I would guess that you are using the code generation procedure in the WHEN-CREATE-RECORD trigger block-level. Because this trigger normally intialize the value for new record insertion. If no then specify.
    further please guide, can PRE-INSERT trigger may handle this problem efficiently ? ,Yes, PRE-INSERT will work without any problem.
    if so, then how ? ... Because, PRE-INSERT will fetch the MAX number from the table at the time of save not at the time of creation. So, suppose five user will enter record at the same time but when they will save then PRE-INSERT will get the max sequence number. And it will never create problem. Because in five user's insertion sure there will be some time difference. So, max number can be get easily without any problem.
    -Ammad

  • Survey on best practice Business Intelligence Strategy - Join in

    I have created a survey (link below) and will share the results back to this thread.  Many of SAP's customers are focusing on establishing or reenergizing a Business Intelligence strategy and platform.  Many questions arise and we often wonder how others are approaching the challenges.  This is a quick 10 question survey.  If the response is well received, we can have follow-up surveys or polls.  Please take a moment to complete the survey and respond back with other questions we might add to future surveys.
    Thanks.
    [Survey|http://polldaddy.com/s/C63FAE8827A1432B]

    Looks like there is some interest.  Looking for others to take the survey too.  Thanks.

  • Essbase vs. Relational Data Warehouse (Which one is the fact table in DW)?

    Guys, thanks in advance for your feedback but below is a simple question i am trying to get feedback on. I am trying to compare an Essbase cube to Relational Data Warehouse containing the same set of information:
    Essbase Dimensions
    Time, Account, Product, Scenario (making it easy)
    Relational Data Warehouse
    Time (dim), Account (dim), Product (dim), Scenario (Fact table)
    OR
    Time (dim), Product (dim), Scenario (dim), Account (Fact Table)
    Which of the relational lines is correct? Is Account the fact table? or Scenario the fact table? Account will contain your usual P&L accounts. Scenario will contain your usual Actual, Budget and Forecast scenarios.
    Thanks,

    I am so not a DW guy, it's amazing, but I've never let little more than a brush with a product stop me from posting...
    Wouldn't all of your dimensions need to be in your fact table? How else would you join from the fact table to the dimensions?
    In either layout, wouldn't you have the keys for Product, Time, Scenario, Acount, and then data in the fact table?
    Or are you talking about the last dimension in your layouts be the columns? If that were the case (and I don't know that it is), I would guess that Scenario changes less, so it would be in columns, although I can definitely see that not being efficient as you are likely to pull all or some of the Accounts for a given time, product and scenario versus all of the scenarios for a given time, product, and account.
    I'm really curious about this as I am just the consumer of star schemas, never (thankfully, and obviously, given the above insane ramblings) the designer of them.
    Regards,
    Cameron Lackpour

  • Best practice "changing several related objects via BDT" (Business Data Toolset) / Mehrere verbundene Objekte per BDT ändern

    Hallo,
    I want to start a
    discussion, to find a best practice method to change several related master
    data objects via BDT. At the moment we are faced with miscellaneous requirements,
    where we have a master data object which uses BDT framework for maintenance (in
    our case an insured objects). While changing or creating the insured objects a
    several related objects e.g. Business Partner should also be changed or
    created. So am searching for a best practices approach how to implement such a
    solution.
    One Idea was to so call a
    report via SUBMIT AND RETURN in Event DSAVC or DSAVE. Unfortunately this implementation
    method has only poor options to handle errors. Second it is also hard to keep LUW
    together.
    Another idea is to call an additional
    BDT instance in the DCHCK-event via FM BDT_INSTANCE_SELECT and the parameters
    iv_xpush_classic = ‘X’ and iv_xpop_classic = ‘X’. At this time we didn’t get
    this solution working correctly, because there is always something missing
    (e.g. global memory is not transferred correctly between the two BDT instances).
    So hopefully you can report
    about your implementations to find a best practice approach for facing such
    requirements.
    Hallo
    ich möchte an der Stelle eine Diskussion starten um einen Best Practice
    Ansatz zu finden, der eine BDT Implementierung/Erweiterung beschreibt, bei der
    verschiedene abhängige BDT-Objekte geändert werden. Momentan treffen bei uns
    mehrere Anforderungen an, bei deinen Änderungen eines BDT Objektes an ein
    anderes BDT Objekte vererbt werden sollen. Sprich es sollen weitere Objekte geänderte
    werden, wenn ein Objekt (in unserem Fall ein Versicherungsvertrag) angelegt
    oder geändert wird (zum Beispiel ein Geschäftspartner)
    Die erste unserer Ideen war es, im Zeitpunkt DSAVC oder DSAVE einen
    Report per SUBMIT AND RETURN aufzurufen. Dieser sollte dann die abhängigen Änderungen
    durchführen. Allerdings gibt es hier Probleme mit der Fehlerbehandlung, da
    diese asynchrone stattfinden muss. Weiterhin ist es auch schwer die Konsistenz der
    LUW zu garantieren.
    Ein anderer Ansatz den wir verfolgt hatten, war im Zeitpunkt
    DCHCK per FuBA BDT_INSTANCE_SELECT und den Parameter iv_xpush_classic = ‘X’ and
    iv_xpop_classic = ‘X’ eine neue BDT Instanz zu erzeugen. Leider konnten wir diese
    Lösung nicht endgültig zum Laufen bekommen, da es immer Probleme beim
    Übertragen der globalen Speicher der einzelnen BDT Instanzen gab.
    Ich hoffe Ihr könnt hier eure Implementierungen kurz beschreiben, dass wir
    eine Best Practice Ansatz für das Thema finden können
    BR/VG
    Dominik

  • What is the best practice for creating primary key on fact table?

    what is the best practice for primary key on fact table?
    1. Using composite key
    2. Create a surrogate key
    3. No primary key
    In document, i can only find "From a modeling standpoint, the primary key of the fact table is usually a composite key that is made up of all of its foreign keys."
    http://download.oracle.com/docs/cd/E11882_01/server.112/e16579/logical.htm#i1006423
    I also found a relevant thread states that primary key on fact table is necessary.
    Primary Key on Fact Table.
    But, if no business requires the uniqueness of the records and there is no materilized view, do we still need primary key? is there any other bad affect if there is no primary key on fact table? and any benifits from not creating primary key?

    Well, natural combination of dimensions connected to the fact would be a natural primary key and it would be composite.
    Having an artificial PK might simplify things a bit.
    Having no PK leads to a major mess. Fact should represent a business transaction, or some general event. If you're loading data you want to be able to identify the records that are processed. Also without PK if you forget to make an unique key the access to this fact table will be slow. Plus, having no PK will mean that if you want to used different tools, like Data Modeller in Jbuilder or OWB insert / update functionality it won't function, since there's no PK. Defining a PK for every table is a good practice. Not defining PK is asking for a load of problems, from performance to functionality and data quality.
    Edited by: Cortanamo on 16.12.2010 07:12

  • What would be the best practice?

    Hello,
    Sorry for the question, but I am new in JSP. I am concerned about the architeture of my application, and looking for the best practices.
    What are the best practices for persistence in a relational database? I need to make queries and map the tables into objects, but wich approach should I take?
    Certainly I do not want to make a page full of scriptlets... ;)
    Thanks

    There are several technologies available to help map objects from relational dbs to objects.
    Object Relational Bridge - http://db.apache.org/ojb/
    Jaxor - http://jaxor.sourceforge.net/
    Hibernate - http://hibernate.bluemars.net/
    Castor - http://castor.exolab.org/
    Just to name a few. Each has its strengths and weaknesses.
    Those will help you create objects from your relational tables. From there it becomes a matter of creating a model for your application. Here are a few good links which explain some strategies for accomplishing this.
    Front Controller Pattern - http://java.sun.com/blueprints/corej2eepatterns/Patterns/FrontController.html
    Data Access Objects - http://java.sun.com/blueprints/corej2eepatterns/Patterns/DataAccessObject.html
    Or if your app is very simple you could probably get away with calling your data objects directly from your JSP. If you decide something like a front-controller is appropriate for your app you should look into the Apache Struts framework (http://jakarta.apache.org/struts/) which will facilitate the ubiquitous MVC pattern. There is also plenty of info available on the site as to what MVC is if you aren't familar with it.
    Hope that helps,
    gamlin

  • How to build an Essbase cube

    Hi Everyone
    I am new to Essbase , i looking for step by step process of creating essbase cubes . Please suggest me what tools do i need to use to build essbase cubes on relational database(Oracle 10g) . How hard is it for fresh guy to build according to a business requirement. Do i need to learn any scripting language to build a cube which involves in lot of calculation. I have googled a lot to get information on this, but i didn't get full information .Please provide me a link or doc or video through which i will get complete understanding of building essbase cube.
    Thanks in advance.
    Regards
    Sandeep Artham

    To start with Oracle by Example is one the best source:
    http://www.oracle.com/technology/obe/hyp_ana/ess11.1.1.htm
    It contains proper audio/video lessons as well.
    Hyperion Essbase provides powerful analytical processing to fulfill a broad spectrum of business performance management needs across all levels and functional areas of business. Information is stored in multiple dimensions and hierarchies that provide a concise and intuitive application model, because they define multiple views of information in the same way users organize, navigate and select information for analysis. It means business understanding plays a vital role in heading towards Essbase.
    Essbase has our very own Calc scripts, No doubt it becomes easy for you to learn it if you have prior programming experience. Not so hard but its fun. Its really exciting how you extract information from the data, which the help of such a stong tool.
    Best of Luck.

  • Unicode Migration using National Characterset data types - Best Practice ?

    I know that Oracle discourages the use of the national characterset and national characterset data types(NCHAR, NVARCHAR) but that is the route my company has decide to take and I would like to know what is the best practice regarding this specifically in relation to stored procedures.
    The database schema is being converted by changing all CHAR, VARCHAR and CLOB data types to NCHAR, NVARCHAR and NCLOB data types respectively and I would appreciate any suggestions regarding the changes that need to be made to stored procedures and if there are any hard and fast rules that need to be followed.
    Specific questions that I have are :
    1. Do CHAR and VARCHAR parameters need to be changed to NCHAR and NVARCHAR types ?
    2. Do CHAR and VARCHAR variables need to be changed to NCHAR and NVARCHAR types ?
    3. Do string literals need to be prefixed with 'N' in all cases ? e.g.
    in variable assignments - v_module_name := N'ABCD'
    in variable comparisons - IF v_sp_access_mode = N'DL'
    in calls to other procedures passing string parameters - proc_xyz(v_module_name, N'String Parameter')
    in database column comparisons - WHERE COLUMN_XYZ = N'ABCD'
    If anybody has been through a similar exercise, please share your experience and point out any additional changes that may be required in other areas.
    Database details are as follows and the application is written in COBOL and this is also being changed to be Unicode compliant:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    NLS_CHARACTERSET = WE8MSWIN1252
    NLS_NCHAR_CHARACTERSET = AL16UTF16

    ##1. while doing a test convertion I discovered that VARCHAR paramaters need to be changed to NVARCHAR2 and not VARCHAR2, same for VARCHAR variables.
    VARCHAR columns/parameters/variables should not by used as Oracle reserves the right to change their semantics in the future. You should use VARCHAR2/NVARCHAR2.
    ##3. Not sure I understand, are you saying that unicode columns(NVARCHAR2, NCHAR) in the database will only be able to store character strings made up from WE8MSWIN1252 characters ?
    No, I meant literals. You cannot include non-WE8MSWIN1252 characters into a literal. Actually, you can include them under certain conditions but they will be transformed to an escaped form. See also the UNISTR function.
    ## Reason given for going down this route is that our application works with SQL Server and Oracle and this was the best option
    ## to keep the code/schemas consistent between the two databases
    First, you have to keep two sets of scripts anyway because syntax of DDL is different between SQL Server and Oracle. There is therefore little benefit of just keeping the data type names the same while so many things need to be different. If I designed your system, I would use a DB-agnostic object repository and a script generator to produce either SQL Server or Oracle scripts with the appropriate data types or at least I would use some placeholder syntax to replace placeholders with appropriate data types per target system in the application installer.
    ## I don't know if it is possible to create a database in SQL Server with a Unicode characterset/collation like you can in Oracle, that would have been the better option.
    I am not an SQL Server expert but I think VARCHAR data types are restricted to Windows ANSI code pages and those do not include Unicode.
    -- Sergiusz

  • Best practice Warnings in OBIEE 10G

    Hi
    Today i suddenly checked in my RPD that tere are 250 best practice warnings. Most are related to keys not defined etc.
    I want to know what is the impact of best practice warnings. as to resolve 250 warnings, i will hve to invest lot of time.
    If its helpful and do somethign good to OBIEE dashboards etc, then only i will be spening more time to this.
    Please suggest.

    Hi,
    Warnings may not be harmful to your code but perfect code may give you the best performance.Defining keys is the best practice becos it will increase the query performance.May be you didnt define keys on unused objects ,so the unused stuff in the RPD may degrade the performance.Anyway we are not always follow the best practices but do which is possible.
    mark if helpful/correct...
    thanks,
    prassu

  • Join large external external Tables (best practice ?)

    Hi there,
    I have three external tables in a master-detail relation: table A (10000000 rows) is master of table B (20000000 rows), Table B is master of table C (100000000 rows). Can you tell my the best way:
    - directly join the external tables, or
    - copy the external tables into tables, create an index, and join them
    What is more efficient, and why ?
    Thanks for your help and ideas.
    Gerhard.

    In general, if the joins you are doing can benefit from indexes, you will want to copy the data to database tables. If the joins will end up doing full table scans anyway, it will matter far less.
    For data integrity, you will likely also want to be able to enforce foreign key constraints.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Best practice when deleting from different table simultainiously

    Greetings people,
    I have two tables joined with a foreign key contrraint. They are written at the same time to keep the constraint happy but I don't know the best way of deleting them as far as rowsets and datamodels are concerned. Are there "gotchas" like do I delete the row in the foreign key table first?
    I am reading thread:http://swforum.sun.com/jive/thread.jspa?forumID=123&threadID=49918
    and getting my head around it.
    Is there a tutorial which deals with this topic?
    I was wondering the best way to go.
    Many Thanks.
    Phil
    is there a "best practice" method for

    Without knowing many details about your specifics... I can suggest a few alternatives -
    You can definitely build coordinating the deletes into your application - you can automatically delete any FK related entries prior to deleting the master, or, refuse to delete the master until the user goes and explicitly deletes the children... just depends on how you want to manage it.
    Also in many databases you can build the cascading delete rules into your database tables themselves.... so that when you delete the master the deletes automatically cascade. I think this is something you typically declare when creating the FK constrataint (delete cascade and update cascade rules).
    hth,
    v

Maybe you are looking for

  • Generate script from SQL for database creation

    Hi, as we have DBMS_METADATA to get scripts for constraints , is there any way by which we can get scripts used for database creation from existing databases Thanks kedar

  • Where to store the logo?

    Hi all! Our problem is to find a way to store our webdynpro login screen logo. The logo is a .gif file. In the customizing of the webdynpro login screen there is a field for putting in an URL to the logo. Is there a way to store the logo inside SAP a

  • 20" iMac with Dual Monitor?

    I have a 20" iMac (early 2009) and i love it! However I've learned lately of dual monitor setups. Can the iMac handle a dual monitor setup? And what are the advantages of a dual monitor setup? I use my iMac for email, web browsing, ichat, ilife, iwor

  • Échec de la connexion sécurisée pour les modules complémentaires avec Thunderbierd

    J'ai TB 24.5.0 en version portable. Quand je vais sur outils->Modules complémentaires j'obtiens l'erreur suivante (copie du texte) : Échec de la connexion sécurisée services.addons.mozilla.org utilise un certificat de sécurité invalide. Le certificat

  • SDK Advanced Training

    Hello, I will like to know were is the best place to get advanced training on SDK for SAP Business One. Best Regards, Yaco