Siebel to OBIEE migration - Nested Aggregation Issue

Hi
I recently migrated from Siebel to OBIEE, now after migrating i found that i am having consistency errors at RPD level
Error_:
[14041] Error in measure definition for column X. Nested aggregate measure definitions are currently not supported.
Formula for Column X:
X= Sum(A*B)
Where A and B are two columns from 2 different logical tables, and both A and B have aggregates like Max or Min or Avg ...
I need someone to guide me how to fix the error and get the right data, because sometimes when we fix Error we still don't get correct values.
regards

Hy,
Yes, you can't have a sum from a sum as :
sum(sum(a)*sum(b))I suppose that you want :
sum(a*b)You need to create a Calculation Measure Using Physical Columns to have
this calculation a*b
Check here how to do :
http://www.oracle.com/technology/obe/obe_bi/bi_ee_1013/bi_admin/biadmin.html#t5s5
And then apply the sum aggregate definition on this column
Success
Nico

Similar Messages

  • OBIEE Error[14041] Nested Aggregation measure are currently not supported

    Hi,
    Please provide workaround for the OBIEE Error[14041] Error in measure definiion. Nested aggregation measures are currently not supported.
    I have two logical columns - Current YTD Invoice Quantity and Prior YTD Invoice Price. I want to create a column which is -
    (Current YTD Invoice Quantity * Prior YTD Invoice Price)/ sum(Current YTD Invoice Quantity * Prior YTD Invoice Price)
    sum in the denominator should be the sum of all rows returned by the report. So level based measures cannot be used as there are multiple report and dimension used may vary.
    Columns Current YTD Invoice Quantity and Prior YTD Invoice Price are already aggregated to Sum.
    If I try to do the sum on the physical column while creating logical column, the query is running for 6-7 hours and not giving any result. Both the columns belong to different alias table in physical layer.
    Please let me know if you guys know any work around for this issue. Also let me know whether this type of nested aggregation are supported in 11g or not?
    Thanks.

    Hi Adil,
    Yes, I did create a hierarchy for the time dimesion.
    No, I was not able to specify the aggregate rule since the source of the logical column is another logical column
    But the time hierarchy works for rows that is not based upon the SUM aggregate at the answers level.
    Say I have 3 columns: Income,Expense, Bottom Line (formula : Income- Expense)
    Income is created based upon a case statement in the Logical Column and
    Expense is also created based upon a case statement in the Logical Column.
    Income 1000 Rupees
    Expense -300 Rupees
    Bottom Line 700 Rupees (where the Bottom line is a row based upon the Answers aggregate formula
    which is SUM(Income+Expense))
    When I add a column year, I get the following output
    Income 2009 500 Rs
    Income 2008 500 Rs
    Expense 2009 -150 Rs
    Expense 2008 -150 Rs
    Bottom Line 2009 700 Rs
    Bottom Line 2008 700 Rs
    The Bottom Line doesn't spread across the Year
    Hope this helps you to understand what my problems is.
    Thank you!

  • Problem regarding Discoverer to OBIEE Migration

    Hi,
    I am trying Discoverer to OBIEE Migration  by using simple apps tables for practicing purpose facing some issue here , so i would like to get help from you all experts.
    The tables that used in .exx file are as follows:
            po_requisition_headers_all ,
            po_requisition_lines_all ,
            po_req_distributions_all ,
            po_distributions_all ,
            po_headers_all .
    I have created .eex in Discoverer it contains above tables, using OBIEE migration utility I have migrated .exx file into the .rpd  . Create a connection pool so physical table views the  data  in physical layer tables upto here i followed  the following link :
    Article: Migrating Oracle Discoverer Data to Oracle Business Intelligence 11g
    when .eex is migrated into .rpd facing following problems:
    1. The joins between the tables are not visible into the Physical Layer. So i need to create it manually or what?
    2. About Hierarchy,Complex folders Hierarchies are not migrated as above link so I need to create it manually or what?
    3.
        - Complex folders based on other complex folders are not migrated.
        - Summary folders are not migrated.
        -The system cannot migrate complex folders that are dimensions.
        -Hierarchies based on Complex Folders are not migrated.
    So what are the steps that I need to follow to overcome this all.
    Please reply me all or some points that you know or you already overcome this types of senario.
    Thanks

    Its the same as you'd get if you were creating the repository from scratch. Username: Administrator with no password.
    Full details of what is done can be found here...
    Disco Conversion

  • How to integrate Siebel with OBIEE ?

    Hi,
    I need a favor from you guys. Kindly let me know the procedure to integrate Siebel with OBIEE.
    According to our requirement we have to integrate Siebel 8.0 version with OBIEE 10.1.3.4 version. Kindly provide me the necessary steps/inputs so that i can integrate Siebel with OBIEE.
    Thanks in advance.
    Regards,
    Samdev

    Hi all,
    @Stijn: With regards to OBIA-specific screens I definitely agree with you. Not so sure about the basic integration though. The Integration Administration screens are normally available.
    Key item to have is the ROW_ID (or "INTEGRATION_ID" if you're talking OBIA context) in order for the integration to work...i.e. be able to jump from OBIEE to Siebel.
    The responsibilities are the other issue. Your approach is 100% sound and I'd keep that one in mind (@user8681498). I have never tested Siebel integrated OBIEE without OBIA where the security is concerned so I somehow doubt that the hand-over of usernames and the attached responsibilities for transformation into web groups works immediately.
    @user8681498: For the tables missing from the Siebel schema, those are mostly OBIA related tables - i.e. for data loading purposes rather than for authorization or integration mechanisms.
    Cheers,
    C.

  • Discoverer 4i to OBIEE migration methodology - Oracle Recommended Way

    Hi,
    We are looking into a scenario where we are required to migrate large number of Discoverer 4i reports to OBIEE. We have good number of custom reports(approx 200) and views too that are to be migrated.
    Am aware about the migration assistant that comes along with OBIEE 10.1.3.4. It's a very useful tool, BUT we need to consider the following facts before proceeding:
    1. The tool is not for migrating Oracle E-Business Suite Discoverer reports (BIS).
    2. The minimum version of Discoverer EUL that can be used with the Assistant is 10.1.2.n.
    3. The version of the RPD file that is generated is dependent upon the installed version of Oracle BI EE that the Assistant is run against but must be a minimum of 10.1.3.4.
    What is the Oracle recommended approach... complete rebuild of the repository in OBIEE/ Migration from 4i -> 9i ->10g and then to 10.1.3.4/...?
    Appreciate inputs from the product gurus and experts.
    Regards
    KSK
    Edited by: user633620 on Sep 30, 2008 10:04 PM

    Ok firstly I'm not sure there is an Oracle recommended approach for migration,.
    However based on my experience at an eBS client - the Discoverer workbooks were very different from what we wanted to achieve with our new OBIEE dashboards. The Discoverer folders were based on complex views which joined relevant OLTP tables together and contained both Dimension attributes and Facts in the same folder. They were used by our users mainly to extract data at the transaction level into Access databases, this is something we were keen to move away from, and adopt a more Analytic top down approach, drilling into the detail as in Disco. This did require some additional work with our user base to see what they were actually doing in Access and if we could do this in OBIEE, which in most cases we could (although we did use Apex in places too).
    We looked at the migration tool, and it seemed to do a decent enough job technically, however for the reasons above ie. the move towards a top down solution we decided on a manual build approach.
    The model for OBIEE had to be different, we broke down the views into their component Dimensions and Facts for - whilst the Physical model was 3NF, we ensured the Business model was a pure Star Schema (this was recommended by Oracle). The Subject Areas were kept quite small, and aligned to the business process, no more than 10-12 logical tables with a couple of fact logical tables, and most of all our dev teams had a set of guidelines to follow to ensure we had consistent repositories (Object Names, Ordering, Logical Levels etc..)
    If you are using eBS i would have a look at some of the BI Applications Oracle have produced, you may get a good idea for best practise and standards for producing a repository from them. Also the new 10.1.3.4 dashboard has some examples of good layout, and some nice tips (ie. One Answer per Dashboard page, Prompts at the top etc...etc..)
    Good Luck.

  • Aggregation issue Bex report

    Hi all,
    I am facing the following aggregation issue at reporing level. BW system 3.5
    Cube1
    Material, Company code, Cost center, Material, Month,   Volume KF
    Cube2
    Material, Company code, Cost center, Material, Month,   Price KF
    Multi provider
    Material, Company code, Cost center, Material, Month,   Volume KF, Price KF
    Report
    - Global Calculated key figure 'Value' is  based on basic KF's  Volume KF, Price KF
    - Time of aggregation is set to " Before aggregation" in propery of Calculated Key Figure.
    -  There is only one characteristics 'Company code' is used in report.
    When, I execute this report, Calculated KF is not working (No values), If I change Time of aggregation " After aggregation" in propery of Calculated Key Figure, then It works but wrong values.Price gets aggregated(add ups) and multiply with Volume which is wrong.
    Can you please give me a Ideal solution to resolve this.
    Thanks,
    Harry

    Hi all,
    Can I assume that there is no solution for this issue ??
    Thanks,
    Harry

  • Nested Aggregation Problem

    I am using OBI 11g.
    Nested aggregation is not supported in some forms in the BI Server (RPD), but appears to be possible by putting the second aggr rule in an Answers column formula or pivot view column. However, I cannot get this to work. Looks like can even be done in the RPD with aggregation based on dimensions, as long as there is a standard aggregation function on the outside of the expression.
    The biggest problem with any of the above techniques is the BI Server does not push the outer aggregation rule to the DB engine (the generated SQL).
    In my case, consider a Referral Fact with Customer Dim and Referral Dim. I need to get # of Referrals per customer, filter that with a case statement to "bin" 1 Referral and >1 Referral, and then get # of Customers in each bin. So the first measure aggregation looks like:
    Other: COUNT (DISTINCT "Referral Key")
    Customer: "SUM( CASE WHEN "Referral Key" = 1 THEN 1 ELSE 0 END )"
    Or the logical measure just has the COUNT DISTINCT aggregation rule and an Answers column has the CASE statement with a SUM aggregation rule. Or use CASE WHEN "Referral Key" = 1 THEN "Customer Key" END and use COUNT DISTINCT instead of SUM.
    All these appear to return correct results, but they all perform the outer aggregation in the BI Server or Pivot engine instead of pushing to the generated SQL (DB engine).
    I can't find any problem in the DB Features. We are using SQL Server 2010.
    Thanks in advance for help.

    Hi AL,
    here is my requirement what i have been asked to get this output result.
    i have keyfigures KF1, KF2 and total KF.
    three characteristics dist,inch,load.
    dist-inch-load--KF1-KF2-Total KF                         
      5---A--010-0-----10=10                         
      5---A--120-20----20+10=30                         
    10--B---050-0-----50                         
    12--C---160-60----60                         
    13--D---270-70----70                         
    14--E---080-0-----80                         
    15--E---120-20----20+80=100                         
    15--E---230-30----302080=130     
    KF1 is the initial volume coming from the file directily.based on this keyfigure i have to calculate KF2,Total KF.     
    In order to calcuate KF2 and Total KF i have some conditions.which are mentioned below;
    KF2---> if  load=0 then KF2=0 elseif load>0 then KF2=KF1 ;
    Total KF--->if load=0 then Total KF=KF2KF1 elseif load>0 then KF2=KF2KF1
    How to achieve this dynamic summation.Do i have to do nested exception aggregation based on the above three characteristics. what would be the open options.please do help me.

  • Aggregation issue for report with bw structure

    Hi,
    I am facing aggregation issue while grouping reports in webi.
    We have a BW query with 16 values, which we bring to bw as structure. Out of 16, 8 are percentage values (agg. type should be average).
    If we bring the data at site level, data is comming properly. But if we use same query and try sum/grouping( on region level), then percentage is getting added.
    Since it's a dashboard report with lots of filters, we cannot go for seperate query at each level(site, region, zone).
    How we can resolve this.. please give me suggestion.
    Regards
    Baby

    Hi,
    Since we were using structure, it was not possible to produce the required result in BO.
    We change structure to keyfigures and bring all of them in to BO. All the column formulas are now in BO side.
    Now it is working fine.
    Regards
    Baby
    Edited by: Baby on May 10, 2010 11:39 AM

  • Nested aggregation and additional sql in query

    Hi experts,
    I need to create a measure in the rpd that will look like:
    measure 1 = SUM(count distinct (ID)) over(order by T11687.MON_CODE asc rows unbounded preceding)
    Is it possible to have a nested aggregation and the over (...) in the query that willl be generated?
    NB: I can't use the evaluate function
    because measure 1 will be just a factor in a complex calculation where other measures are derived from different tables.
    Regards

    In doesn't work that way.
    this query is a bit "stupid", but try this
    select * from topping where topping_id = 1
    union all
    select * from topping where topping_id = 1
    union all
    select * from topping where topping_id = 1Or better yet, let it return once, and then do
    var = rs.getString(1);
    var2 = var;
    var3 = var;

  • Issue with nested aggregated in time series

    Hi Experts,
    I have a small requiremnt using time series.
    eg : (sum(current year4:Quarters bookingamt)-sum(previous year4:Quarters bookingamt))/sum(previous year4:Quarters bookingamt)
    Bookingamt is My measure Name.
    how do we implement above logic can please help regarding this.
    Thanks,
    C Mahesh

    Hi Mahesh,
    These functions help to obtain the data analyzed based on the time. There are two types of Time Series in OBIEE viz..
    1) Ago
    2) To Date
    Following are the prerequisites and procedure to create time series function in OBIEE RPD.
    a) Create a dimensional Hierarchy (ex: YearàQTRàMonth) and convert it to the time dimension. DC on hierarchy and check the box time dimension.
    b) Define the Chronological Key. A chronological key should always be the lowest level in the table and the data in it should be in the format of 20090102 or 200109 (year, month format). DC on level (preferred/last—detail) go to KEYS tab and check chronological key.
    c) Duplicate the existing column, rename and specify the formula for the column.
    1) Ago: This function when specified in any column displays data that is month ago, 2 months ago depending on the specified formula. DC on new column check use logical columns and click on exp. Builder. In the left column select functions and then time series and then AGO. Then in metric go to logical tables and select metric and in level go to time dimension and select month (1 month ago) and give the period as 1(since it is 1 month ago) then transfer two new columns to presentation layer
    Ex: sales: 100,200,300 à -, 100, 200 (1 month AGO).
    2) To Date: This functions allows us to display the aggregated measures based on the formula specified. DC on new column check use logical columns and click on exp. Builder. In the left column select functions and then time series then select TO DATE. Select metric from logical tables (revenue) and as level go to time dimension and select year(YTD) and OK.
    YTD- year to date – Displays aggregated values that is with respect to month since it is YTD. Similarly for QTD and MTD.
    Ex: sales: 100,200,300 à YTD Sales: 100,300,600
    http://www.rittmanmead.com/2007/04/obi-ee-time-dimensions-and-time-series-calculations/ -- How to implement step by step with screen shots.
    http://www.oraclebidwh.com/2009/12/time-series-wizard-in-obiee/ -- you can understand easily how to implement this with screen shot.
    Hope it help's
    Thanks,
    satya

  • OBIEE 11g: Dashboard Javascript Issue

    Hi Gurus,
    We have upgraded obiee from 10g to 11g and finding issues with javascript in a dashboard.
    Functionality: There are some custom labels showing prompt values in it with large font. When user change prompt value and Apply, it should change the value of those text as well.
    In 10g its running fine, but in 11g its not happening after we change the value of prompts. I found the following Javascript is responsible for this functionality. Even I saw one thread to suggest the exactly same code, but in 10g.
    <script type="text/javascript">
    (function(){
    var tblTag = document.getElementsByTagName('table');
    var tdElem= document.getElementsByTagName('td');
    for(m=0;m<tdElem.length;m++){
    if(tdElem[m].className=='GFPSubmit'){tdElem[m].childNodes[0].tBodies[0].rows[0].cells[0].childNodes[0].childNodes[0].innerHTML='Run Report';
    }//close if statement
    }//close for loop
    }// close function clickVal()
    </script>
    Thread:
    Change Go Button Text on Prompt Only
    I need to understand what its actually doing? and Does it really work in 11g? Whats the alternative code?
    Thanks in advance.

    In 10g these are the html objects, after upgrade you need to know the html objects for that report based on that you need to modify javascript code.
    From given code, activity is doing on these objects; You need to find out the equivalent object name for 'GFPSubmit' in 11g.
    GFPSubmit
    table
    td
    If make sense mark

  • OBIEE 11g SSL Configuration Issue : Unable to import the Server certs

    Hello All,
    We are trying to configure OBIEE 11.1.1.6.0 with SSL using Windows server 2003 (IIS) and facing some issues with that.
    Followed the document : OBIEE11g SSL Setup and Configuration [1326781.1]
    http://obieedue.blogspot.sg/2012/08/obiee11g-ssl-setup-and-configuration.html
    and also completed generating the required certificate signing request and keystores for SSL communication and sent it to the CA (IT Admin team) to to have the certificate signed by CA. The issue comes when I am trying to import the CA certificate (Root certificate) and Server Certificate into the Java Keystore.
    I am importing the Root CA Certificate first which is successfully added to the keystore.
    keytool -import -trustcacerts -alias mycacert -file cacert.pem -keystore mykeystore.jks -storepass Welcome1
    Trust this certificate? [no]: yes
    Certificate was added to keystore.
    But when trying to add the Server Certificate to the keystore using the command below :
    keytool -import -v -alias testserver -file server.cer -keystore mykeystore.jks -keypass Welcome1 -storepass Welcome1
    Certificate reply was installed in keystore
    I get the following error:
    keytool error: java.lang.Exception: Failed to establish chain from reply
    java.lang.Exception: Failed to establish chain from reply
    at sun.security.tools.KeyTool.establishCertChain(KeyTool.java:2662)
    at sun.security.tools.KeyTool.installReply(KeyTool.java:1870)
    at sun.security.tools.KeyTool.doCommands(KeyTool.java:807)
    at sun.security.tools.KeyTool.run(KeyTool.java:172)
    at sun.security.tools.KeyTool.main(KeyTool.java:166)
    Read many forums and tried to convert it to the PKCS#7 format and import the cert to the identity keystore, but was not successful in that either. I have also checked with the IT Admin team and found there is only one RootCA and no other intermediate CA's.
    Please advice if any one has similar issues or suggestions.
    Thanks in advance,
    SVS

    Hi,
    One obvious reason would be that you did not specify -trustcacerts, and the root CA is not included in the present server keystore. In that case, using the -trustcacerts option would solve the problem, if the root CA is indeed in the JDK cacerts.
    To print out the certificates present in the JDK cacerts, use the following command:
    keytool -list -keystore <JAVA_HOME>/jre/lib/security/cacerts -storepass changeit -v
    Then check if the root CA that signed your server certificate is present, and has not expired (in which case,you would need to re-import a newer one into cacerts).
    Another common reason for that error message is when you have used a proprietary CA to sign your server certificate. Then it would obviously not be in the JDK cacerts. The solution in that case is to import your proprietary root CA into the JDK cacerts, using the following command:
    keytool -import -keystore <JAVA_HOME>/jre/lib/security/cacerts -file yourRootCA.pem -storepass changeit -alias youralias
    A third reason for that error message is when your server was signed by an intermediate certificate. In that case, you would have received from your CA a chain of certificates. One way to solve this (not the only one, but this one works well): Prepend your intermediate CA file to your server cert file, and import the obtained concatenated file into the server keystore. Be careful, the intermediate CA must be BEFORE the server cert. Example:
    copy rootca.cer certchain.p7b
    type server.cer >> certchain.p7b
    The file certchain.p7b will be the concatenation of the intermediate CA and the signed server cert. Then import the newly created file under the key alias as follows:
    keytool -import -keystore serverks.jks -file certchain.p7b -alias yourkey -trustcacerts
    If you only prepend the intermediate root CA, you must make sure the the final root CA is in cacerts. But you can also prepend your whole chain of trust inside the server keystore.
    Regards,
    Kal

  • Aggregation issue on a T5220

    I have a Enterprise T5220 server, running Solaris 10 that I am using as a backup server. On this server, I have a Layer 4, LACP-enabled link aggregation set up using two of the server's Gigabit NICs (e1000g2 and e1000g3) and until recently I was getting up to and sometimes over 1.5 Gb/s as desired. However, something has happened recently to where I can now barely get over 1 Gb/s. As far as I know, no patches were applied to the server and no changes were made to the switch that it's connected to (Nortel Passport 8600 Series) and the total amount of backup data sent to the server has stayed fairly constant. I have tried setting up the aggregation multiple times and in multiple ways to no avail. (LACP enabled/disabled, different policies, etc.) I've also tried using different ports on the server and switch to rule out any faulty port problems. Our networking guys assure me that the aggregation is set up correctly on the switch side but I can get more details if needed.
    In order to attempt to better troubleshoot the problem, I run one of several network speed tools (nttcp, nepim, & iperf) as the "server" on the T5220, and I set up a spare X2100 as a "client". Both the server and client are connected to the same switch. The first set of tests with all three tools yields roughly 600 Mb/s. This seems a bit low to me, I seem to remember getting 700+ Mb/s prior to this "issue". When I run a second set of tests from two separate "client" X2100 servers, coming in on two different Gig ports on the T5220, each port also does ~600 Mb/s. I have also tried using crossover cables and I only get maybe a 50-75 Mb/s increase. After Googling Solaris network optimizations, I found that if I double tcp_max_buf to 2097152, and set tcp_xmit_hiwat & tcp_recv_hiwat to 524288, it bumps up the speed of a single Gig port to ~920 Mb/s. That's more like it!
    Unfortunately however, even with the TCP tweaks enabled, I still only get a little over 1 Gb/s through the two aggregated Gig ports. It seems as though the aggregation is only using one port, though MRTG graphs of the two switch ports do in fact show that they are both being utilized equally, essentially splitting the 1 Gb/s speed between
    the two ports.
    Problem with the server? switch? Aggregation software? All the above? At any rate, I seem to be missing something.. Any help regarding this issue would be greatly appreciated!
    Regards,
    Jim
    Output of several commands on the T5220:
    uname -a:
    SunOS oitbus1 5.10 Generic_137111-07 sun4v sparc SUNW,SPARC-Enterprise-T5220
    ifconfig -a (IP and broadcast hidden for security):
    lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 8232 index 1
    inet 127.0.0.1 netmask ff000000
    aggr1: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
    inet x.x.x.x netmask ffffff00 broadcast x.x.x.x
    ether 0:14:4f:ec:bc:1e
    dladm show-dev:
    e1000g0 link: unknown speed: 0 Mbps duplex: half
    e1000g1 link: unknown speed: 0 Mbps duplex: half
    e1000g2 link: up speed: 1000 Mbps duplex: full
    e1000g3 link: up speed: 1000 Mbps duplex: full
    dladm show-link:
    e1000g0 type: non-vlan mtu: 1500 device: e1000g0
    e1000g1 type: non-vlan mtu: 1500 device: e1000g1
    e1000g2 type: non-vlan mtu: 1500 device: e1000g2
    e1000g3 type: non-vlan mtu: 1500 device: e1000g3
    aggr1 type: non-vlan mtu: 1500 aggregation: key 1
    dladm show-aggr:
    key: 1 (0x0001) policy: L4 address: 0:14:4f:ec:bc:1e (auto) device address speed
    duplex link state
    e1000g2 0:14:4f:ec:bc:1e 1000 Mbps full up attached
    e1000g3 <unknown> 1000 Mbps full up attached
    dladm show-aggr -L:
    key: 1 (0x0001) policy: L4 address: 0:14:4f:ec:bc:1e (auto) LACP mode: active LACP timer: short
    device activity timeout aggregatable sync coll dist defaulted expired
    e1000g2 active short yes yes yes yes no no
    e1000g3 active short yes yes yes yes no no
    dladm show-aggr -s:
    key: 1 ipackets rbytes opackets obytes %ipkts %opkts
    Total 464982722061215050501612388529872161440848661
    e1000g2 30677028844072327428231142100939796617960694 66.0 59.5
    e1000g3 15821243372049177622000967520476 64822888149 34.0 40.5
    Edited by: JimBuitt on Sep 26, 2008 12:04 PM

    JimBuitt wrote:
    I have a Enterprise T5220 server, running Solaris 10 that I am using as a backup server. On this server, I have a Layer 4, LACP-enabled link aggregation set up using two of the server's Gigabit NICs (e1000g2 and e1000g3) and until recently I was getting up to and sometimes over 1.5 Gb/s as desired. However, something has happened recently to where I can now barely get over 1 Gb/s.Is this with multiple backup streams or just one?
    I would not expect to get higher throughput with a single stream. Only with the aggregate throughput of multiple streams.
    Darren

  • OBIEE 11G catalog security issue

    I have the following (simplified) requirement:
    - People with application role FIN should have access to analysis XYZ
    - People with application role HR should have access to analysis ABC
    - People with both application roles should have access to XYZ and ABC
    The above seems pretty obvious and per http://docs.oracle.com/cd/E29505_01/bi.1111/e10543/mgrgrpsusers.htm#autoId16 this should work without further ado, since the least restrictive access should be granted when roles with different permissions (FIN has nog permissions on ABC and HR no permissions on XYZ) are granted, however...
    We find that if role HR is not explicitly excluded from access to XYZ then it will always have access to XYZ. The same goed for FIN and ABC. But then when both roles are granted (with these explicit exclusions), user has access to neither XYZ or ABC. This actually makes sense, since explicit exclusions take precedence, but then how do I achieve my goal as stated above?
    The same logic applies to subject areas. If not explicitly excluded for either of the two roles, access is always granted. But if explicitly excluded, granting both roles excludes both subject areas.
    I have already checked inherited roles. The FIN and HR roles are based on standard BI Consumer roles, so they are both members of Authenticated User and BI Consumer. But even taking these into account, the same problem persists. Pretty sure I am overlooking something small and simple.

    Hi there. Does anybody noticed this issue on calendars?! That only occurs when the month is october....
    Just updating, the correct version of OBIEE is 11.1.1.5.0.
    Thanks in advance.
    Marcos

  • OBIEE 11g Union Report Issue

    Hi all,
    Did one any of you started development work on 11g? If so have you created Union reports in 11g?
    I am facing two issues afte creating Union reports.
    1.I am unable to sort the columns. (Report is automatically sorted according to alphabatical order)
    2. I am unable to create New calculation Item.
    Thanks

    Hi there. Does anybody noticed this issue on calendars?! That only occurs when the month is october....
    Just updating, the correct version of OBIEE is 11.1.1.5.0.
    Thanks in advance.
    Marcos

Maybe you are looking for