Definition of Dimensions

HI All
What is the criteria for defining the dimensions..Account,Category,..Datasource...what in case it changes, I have to define company , costcenter ,version , product and keyfigures like new price . Again what is the advantage of using them ? why not go for user defined ones..?
thanks
Rak

Hi Rakshita,
By design, you will need 4 dimensions of type A, C, E and T. Other than these, you can have your own user defined dimensions as per the requirement.

Similar Messages

  • Pixel dimensions and resolution

    When designing for an iphone4 we set up the canvas to 640x960 pix, but we discussed if we should define the resolution to 326ppi, as described in the iphone tech spec.
    When the pixeldimensions are defined - why does a pixelbased program like Fireworks ask for resolution? I should think that only non-pixel based output wanted definition of dimensions and resolution.
    Can you help?
    Nina Storm

    Yes. Go ahead with the reccomendations for PPI for iphone. Dimensions are merely the height and width of the image. Those are independant of resolution. Fireworks default resolution is 72 PPI. Since PPI doesn't affect the exported file size, it doesn't matter, at least for web pages displayed in a browser, if you use 72 PPI or 500.

  • Aggregation content Group by

    Can someone explain to me the use of the aggregation content group by tab.
    when we set up the aggregation content group by tab say for a fact logical table source ( Revenue fact LTS1 ) what does it actually imply..
    in the sales sample example that ships with the Oracle BI
    In the Revenue facts table source , we do the following in the aggregation content group by
    logical dimension logical level
    H1 Time Time Detail
    H2 Product Product detail
    H3 Customer Customer detail
    and in the respective logical sources of time, product & customer we set up aggregation content which corresponds to the level we set up in the revenue fact table logical source.
    I fail to understand what is the implied meaning of this aggregation content setting ?? Can someone please help me understand.

    The content tab plays major role for BI server about the physical sources. Based on the columns select in Answers criteira tab BI server uses its inteligence to pick table(or list) to fetch records from db.
    We can go by logical level (using dimension hierarchies) and column level (use table columns when we not using dimensions).
    Be default these are at the lowest level to the fact.
    When a logical (fact)table have multiple sources we set content tab equalent to the level of information is contented.
    Ex: Fact table and its Aggregate table; for fact we set (generally) lowest level and for Aggregate some where mid level or based on its definition wrt dimension tables.
    The level represents group by cluase in the physical query.
    These level can go for measures(Level base measure), when set the specific level to the metric, that metric will execute as indipendent query(with group by clause) and sites with other columns. These will not come under influence of drill down at any cost.
    If you read Administrator help for both logical tables and dimension hierarchies you can get complete piture.
    If helps pls mark

  • How to deduce type variable

    First, sorry for my poor English.
    This piece of code compiled with no problem nor errors.
    interface Singularity<T> {
        public T unique();
    class UniqueArrayList<K, V extends Singularity<K> > {
        Map<K, V> map = new HashMap<K, V>();
        public void add(V value) {
            map.put(value.unique(), value);
        public boolean contains(K key) {
            return map.containsKey(key);
    class Dimension implements Singularity<String> {
        public String unique() {
            return null;
    class Test {
        UniqueArrayList<String, Dimension> list = new UniqueArrayList<String, Dimension>();
    }My problem is simple, in this line:
    UniqueArrayList<String, Dimension> list = new UniqueArrayList<String, Dimension>(); I need to specify both types K and V, but this is a burden: from definition of Dimension it is clear that K is String
    How can i define UniqueArrayList in such way that K will be automatically deduced from V.
    Thanks
    Boaz

    How can i define UniqueArrayList in such way that K will be automatically deduced from >V.I don't know if this is possible. But I have played around a little bit with your code and I think there is bug. Look at the following:
    import java.util.HashMap;
    import java.util.Map;
    interface Singularity<T> {
        public T unique();
    class UniqueArrayList<K, V extends Singularity<K> > {
        Map<K, V> map = new HashMap<K, V>();
        public void add(V value) {
            map.put(value.unique(), value);
        public boolean contains(K key) {
            return map.containsKey(key);
    class Dimension<T> implements Singularity<T> {
        private T unique;
        public Dimension(T unique){
            this.unique=unique;
        public T unique() {
            return (T)null;
    public class Test {
        public static void main(String[] args) {
            Dimension dim=new Dimension<Integer>(new Integer(0));    
            Map<String, Dimension> map = new HashMap<String, Dimension>();
            map.put(dim.unique(), dim);
            UniqueArrayList<String, Dimension> list = new UniqueArrayList<String, Dimension>(); 
            list.add(dim);
    }When using the map the compiler complains about a not compatible method call. When using the array List he doesn't although I think he should.
    I am not sure if this really is bug or if I missunderstand something about the genreics in Java.

  • Query rewrites with Nested materialized views with different aggregations

    Platform used : Oracle 11g.
    Here is a simple fact table (with measures m1,m2) and dimensions (a) Location (b) Calendar and (c) Product. The business problem is that aggregation operator for measure m1,m2 are different along location dimension and Calendar dimension. The intention is to preaggregate the measures for a product along the calendar dimension and Location dimension and store it as materialized views.
    The direct option is to define a materialized view with Inline queries (Because of the different aggrergation operator, it is not possible to write a query without Inline query). http://download-uk.oracle.com/docs/cd/B28359_01/server.111/b28313/qradv.htm#BABEAJBF documents the limitations that it works only for 'Text match' and 'Equivalent queries' and that is too limiting.
    So decided to have nested materialized view, with first view having just joins(my_dim_mvw_joins), the second view having aggregations along Calendar dimension (my_dim_mvw_calendar) and third view having aggregations along the Location dimension(my_dim_mvw_location). Obviously I do not want the query I fire to know about materialized views and I fire it against the fact table. I see that for the fired query (Which needs aggregations along both Calendar and Location), is rewritten with just second materialized view but not the third. (Had set QUERY_REWRITE_INTEGRITY as TRUSTED) .
    Wanted to know whether there are limitations on Query Writes with nested materialized views? Thanks
    (Have given a simple testable example below. Pls ignore the values given in 'CALENDAR_IDs', 'PRODUCT_IDs' etc as they are the same for all the queries)
    -- Calendar hierarchy table
    CREATE TABLE CALENDAR_HIERARCHY_TREE
    (     "CALENDAR_ID" NUMBER(5,0) NOT NULL ENABLE,
    "HIERARCHY1_ID" NUMBER(5,0),
    "HIERARCHY2_ID" NUMBER(5,0),
    "HIERARCHY3_ID" NUMBER(5,0),
    "HIERARCHY4_ID" NUMBER(5,0),
    CONSTRAINT "CALENDAR_HIERARCHY_TREE_PK" PRIMARY KEY ("CALENDAR_ID")
    -- Location hierarchy table
    CREATE TABLE LOCATION_HIERARCHY_TREE
    (     "LOCATION_ID" NUMBER(3,0) NOT NULL ENABLE,
    "HIERARCHY1_ID" NUMBER(3,0),
    "HIERARCHY2_ID" NUMBER(3,0),
    "HIERARCHY3_ID" NUMBER(3,0),
    "HIERARCHY4_ID" NUMBER(3,0),
    CONSTRAINT "LOCATION_HIERARCHY_TREE_PK" PRIMARY KEY ("LOCATION_ID")
    -- Product hierarchy table
    CREATE TABLE PRODUCT_HIERARCHY_TREE
    (     "PRODUCT_ID" NUMBER(3,0) NOT NULL ENABLE,
         "HIERARCHY1_ID" NUMBER(3,0),
         "HIERARCHY2_ID" NUMBER(3,0),
         "HIERARCHY3_ID" NUMBER(3,0),
         "HIERARCHY4_ID" NUMBER(3,0),
         "HIERARCHY5_ID" NUMBER(3,0),
         "HIERARCHY6_ID" NUMBER(3,0),
         CONSTRAINT "PRODUCT_HIERARCHY_TREE_PK" PRIMARY KEY ("PRODUCT_ID")
    -- Fact table
    CREATE TABLE RETAILER_SALES_TBL
    (     "PRODUCT_ID" NUMBER,
    "PRODUCT_KEY" VARCHAR2(50 BYTE),
    "PLAN_ID" NUMBER,
    "PLAN_PERIOD_ID" NUMBER,
    "PERIOD_ID" NUMBER(5,0),
    "M1" NUMBER,
    "M2" NUMBER,
    "M3" NUMBER,
    "M4" NUMBER,
    "M5" NUMBER,
    "M6" NUMBER,
    "M7" NUMBER,
    "M8" NUMBER,
    "LOCATION_ID" NUMBER(3,0),
    "M9" NUMBER,
    CONSTRAINT "RETAILER_SALES_TBL_LOCATI_FK1" FOREIGN KEY ("LOCATION_ID")
    REFERENCES LOCATION_HIERARCHY_TREE ("LOCATION_ID") ENABLE,
    CONSTRAINT "RETAILER_SALES_TBL_PRODUC_FK1" FOREIGN KEY ("PRODUCT_ID")
    REFERENCES PRODUCT_HIERARCHY_TREE ("PRODUCT_ID") ENABLE,
    CONSTRAINT "RETAILER_SALES_TBL_CALEND_FK1" FOREIGN KEY ("PERIOD_ID")
    REFERENCES CALENDAR_HIERARCHY_TREE ("CALENDAR_ID") ENABLE
    -- Location dimension definition to promote query rewrite
    create DIMENSION LOCATION_DIM
    LEVEL CHAIN IS LOCATION_HIERARCHY_TREE.HIERARCHY1_ID
    LEVEL CONSUMER_SEGMENT IS LOCATION_HIERARCHY_TREE.HIERARCHY3_ID
    LEVEL STORE IS LOCATION_HIERARCHY_TREE.LOCATION_ID
    LEVEL TRADING_AREA IS LOCATION_HIERARCHY_TREE.HIERARCHY2_ID
    HIERARCHY PROD_ROLLUP (
    STORE CHILD OF
    CONSUMER_SEGMENT CHILD OF
    TRADING_AREA CHILD OF
    CHAIN
    -- Calendar dimension definition
    create DIMENSION CALENDAR_DIM
    LEVEL MONTH IS CALENDAR_HIERARCHY_TREE.HIERARCHY3_ID
    LEVEL QUARTER IS CALENDAR_HIERARCHY_TREE.HIERARCHY2_ID
    LEVEL WEEK IS CALENDAR_HIERARCHY_TREE.CALENDAR_ID
    LEVEL YEAR IS CALENDAR_HIERARCHY_TREE.HIERARCHY1_ID
    HIERARCHY CALENDAR_ROLLUP (
    WEEK CHILD OF
    MONTH CHILD OF
    QUARTER CHILD OF
    YEAR
    -- Materialized view with just joins needed for other views
    CREATE MATERIALIZED VIEW my_dim_mvw_joins build immediate refresh complete enable query rewrite as
    select product_id, lht.HIERARCHY1_ID, lht.HIERARCHY2_ID, lht.HIERARCHY3_ID, lht.location_id, cht.HIERARCHY1_ID year,
    cht.HIERARCHY2_ID quarter, cht.HIERARCHY3_ID month, cht.calendar_id week, m1, m3, m7, m9
    from retailer_sales_tbl RS, calendar_hierarchy_tree cht, location_hierarchy_tree lht
    WHERE RS.period_id = cht.CALENDAR_ID
    and RS.location_id = lht.location_id
    and cht.CALENDAR_ID in (10,236,237,238,239,608,609,610,611,612,613,614,615,616,617,618,619,1426,1427,1428,1429,1430,1431,1432,1433,1434,1435,1436,1437,1438,1439,1440,1441,1442,1443,1444,1445,1446,1447,1448,1449,1450,1451,1452,1453,1454,1455,1456,1457,1458,1459,1460,1461,1462,1463,1464,1465,1466,1467,1468,1469,1470,1471,1472,1473,1474,1475,1476,1477)
    AND product_id IN (5, 6, 7, 8, 11, 12, 13, 14, 17, 18, 19, 20)
    AND lht.location_id IN (2, 3, 11, 12, 13, 14, 15, 4, 16, 17, 18, 19, 20)
    -- Materialized view which aggregate along calendar dimension
    CREATE MATERIALIZED VIEW my_dim_mvw_calendar build immediate refresh complete enable query rewrite as
    select product_id, HIERARCHY1_ID , HIERARCHY2_ID , HIERARCHY3_ID ,location_id, year, quarter, month, week,
    sum(m1) m1_total, sum(m3) m3_total, sum(m7) m7_total, sum(m9) m9_total,
    GROUPING_ID(product_id, location_id, year, quarter, month, week) dim_mvw_gid
    from my_dim_mvw_joins
    GROUP BY product_id, HIERARCHY1_ID , HIERARCHY2_ID , HIERARCHY3_ID , location_id,
    rollup (year, quarter, month, week);
    -- Materialized view which aggregate along Location dimension
    CREATE MATERIALIZED VIEW my_dim_mvw_location build immediate refresh complete enable query rewrite as
    select product_id, year, quarter, month, week, HIERARCHY1_ID, HIERARCHY2_ID, HIERARCHY3_ID, location_id,
    sum(m1_total) m1_total_1, sum(m3_total) m3_total_1, sum(m7_total) m7_total_1, sum(m9_total) m9_total_1,
    GROUPING_ID(product_id, HIERARCHY1_ID, HIERARCHY2_ID, HIERARCHY3_ID, location_id, year, quarter, month, week) dim_mvw_gid
    from my_dim_mvw_calendar
    GROUP BY product_id, year, quarter, month, week,
    rollup (HIERARCHY1_ID, HIERARCHY2_ID, HIERARCHY3_ID, location_id)
    -- SQL Query Fired (for simplicity have used SUM as aggregation operator for both, but they will be different)
    select product_id, year, HIERARCHY1_ID, HIERARCHY2_ID,
    sum(m1_total) m1_total_1, sum(m3_total) m3_total_1, sum(m7_total) m7_total_1, sum(m9_total) m9_total_1
    from
    select product_id, HIERARCHY1_ID , HIERARCHY2_ID , year,
    sum(m1) m1_total, sum(m3) m3_total, sum(m7) m7_total, sum(m9) m9_total
    from
    select product_id, lht.HIERARCHY1_ID , lht.HIERARCHY2_ID , lht.HIERARCHY3_ID ,lht.location_id, cht.HIERARCHY1_ID year, cht.HIERARCHY2_ID quarter, cht.HIERARCHY3_ID month, cht.calendar_id week,m1,m3,m7,m9
    from
    retailer_sales_tbl RS, calendar_hierarchy_tree cht, location_hierarchy_tree lht
    WHERE RS.period_id = cht.CALENDAR_ID
    and RS.location_id = lht.location_id
    and cht.CALENDAR_ID in (10,236,237,238,239,608,609,610,611,612,613,614,615,616,617,618,619,1426,1427,1428,1429,1430,1431,1432,1433,1434,1435,1436,1437,1438,1439,1440,1441,1442,1443,1444,1445,1446,1447,1448,1449,1450,1451,1452,1453,1454,1455,1456,1457,1458,1459,1460,1461,1462,1463,1464,1465,1466,1467,1468,1469,1470,1471,1472,1473,1474,1475,1476,1477)
    AND product_id IN (5, 6, 7, 8, 11, 12, 13, 14, 17, 18, 19, 20)
    AND lht.location_id IN (2, 3, 11, 12, 13, 14, 15, 4, 16, 17, 18, 19, 20)
    GROUP BY product_id, HIERARCHY1_ID , HIERARCHY2_ID , HIERARCHY3_ID , location_id, year
    ) sales_time
    GROUP BY product_id, year,HIERARCHY1_ID, HIERARCHY2_ID
    This Query rewrites only with my_dim_mvw_calendar. (as saw in Query Plan and EXPLAIN_MVIEW). But we would like it to use my_dim_mvw_location as that has aggregations for both dimensions.

    blackhole001 wrote:
    Hi all,
    I'm trying to make my programmer's life easier by creating a database view for them to query the data, so they don't have to worry about joining tables. This sounds like a pretty horrible idea. I say this because you will eventually end up with programmers that know nothing about your data model and how to properly interact with it.
    Additionally, what you will get is a developer that takes one of your views and see's that of the 20 columns in it, it has 4 that he needs. If all those 4 columns comes from a simple 2 table join, but the view has 8 tables, you're wasting a tonne of resources by using the view (and heaven forbid they have to join that view to another view to get 4 of the 20 columns from that other view as well).
    Ideally you'd write stored routines that satisfy exactly what is required (if you are the database resource and these other programmers are java, .net, etc... based) and the front end developers would call those routines customized for an exact purpose.
    Creating views is not bad, but it's by no means a proper solution to having developers not learn or understand SQL and/or the data model.

  • Complex Aggregate Aware Problem

    I am facing issue with @AA() in following the scenario.
    I have a measure with following definition:
    m1= @AA(FT1\ftm1,AT1\atm1,AT2\atm2,FT2\ftm2)
    Where,
    AT1\atm1=@AA(ATA1\atam1,ATA2\atam2)
    Similar definition is also applicable for dimensions from AT1
    AT1\atd1=@AA(ATA1\atad1,ATA2\atad2)
    I have marked the ATA1 and ATA2 incompatible with all the objects derived from FT1 and FT2 and vise verse.
    Definition of Dimension object from AT1 is as follows:
    atd1= @AA(ATA1\atad1,ATA2\atad2)
    There is another dimension table related to all the tables.
    FT1->DimA1
    FT2->DimA1
    ATA1->DimA1
    ATA2->DimA1
    Now, When I drag dimension DimA1,AT1 and measure m1 it creates 2 queries as:
    Q1: select DimA1,ATA2 from DimA1,ATA1
    Q2: select DimA1,FT1\ftm1 from DimA1,FT1
    Its very tedious but this is the most simplest version in which I can describe...

    Hi
    Look what I have done:
    1) Here is view of my JSF Page and components: http://img695.imageshack.us/img695/4218/21964358.jpg
    On the left there is my Data Control for Web Service. The input is simple Integer and the results are many rows (here is result form EM console test: http://img41.imageshack.us/img41/8803/23008039.jpg )
    The input value form is made by drag and drop entire execute(Integer) as ADF Parameter Form. The id of execute button is cb1.
    The results are made by drag and drop PassDetails as ADF Read Only Table.
    2) I did a Create Method Binding For Action for execute button as you suggested. Adding New Bean: [Bean Name - BindingMethod] [Class Name - BindingMethod] [Scope - request]. It was created a java file as seen on this picture - http://img11.imageshack.us/img11/8192/17868604.jpg I've added your code.
    3) I've also put Refresh Condition: #{adfFacesContext.initialRender} for PassDetailsIterator - http://img35.imageshack.us/img35/2271/85080775.jpg
    Thats all I've done. When running page the ADF Table is not fetching data. When I enter Input Value there are no results in ADF Table, but the instance of my web service is completed.
    When I put Refersh Condition #{!adfFacesContext.initialRender} it automatically fetchind data.
    What should I change in my application? Please help?
    Best regards MK.

  • Not opening the excel workbook from web application

    I am trying the load excel work book from web application and have configured the http authentication while configuring the ADF Secutity. It is asking for the login (weblogic/weblogic1) while load excel workbook, but throws the following exception/error. Please let me know how to resolve this.
    ADFDI-05530: unable to initialize worksheet: Sheet1
    and details are:
    ADFDI-00108: user session required
    The remote server returned an error: (401) Unauthorized.
    ~~~~~~~~~~
    UserSessionRequiredException: ADFDI-00108: user session required
    Source: adfdi-datamanager
    Stack:
    at oracle.adf.client.windows.datamanager.servletrequest.SyncRequest.BeginResponse()
    at oracle.adf.client.windows.datamanager.servletrequest.TamperCheckRequest.InternalCallSyncServlet()
    at oracle.adf.client.windows.datamanager.servletrequest.SyncServletRequest.CallSyncServlet(Boolean retry)
    at oracle.adf.client.windows.datamanager.ADFBindingContext.SendSyncServletRequest(SyncServletRequest request)
    at oracle.adf.client.windows.datamanager.ADFBindingContext.CheckForTampering()
    at oracle.adf.client.windows.datamanager.ADFBindingContext.PreSyncServletTamperCheck()
    at oracle.adf.client.windows.datamanager.ADFBindingContext.SyncModel(BindingContainer bc, String contentType)
    at oracle.adf.client.windows.datamanager.BindingContainer.ReloadMetadata()
    at oracle.adf.client.windows.datamanager.BindingContainer.LoadMetadata()
    at oracle.adf.client.windows.excel.runtime.DIWorksheet.Initialize()
    Inner:
    WebException: The remote server returned an error: (401) Unauthorized.
    Source: System
    Stack:
    at System.Net.HttpWebRequest.GetResponse()
    at oracle.adf.client.windows.datamanager.servletrequest.http.ManagedHttpResponse..ctor(HttpWebRequest request)
    at oracle.adf.client.windows.datamanager.servletrequest.http.ManagedHttpRequest.GetResponse()
    at oracle.adf.client.windows.datamanager.servletrequest.SyncRequest.BeginResponse()

    Thanks for your quick reply. I have tried giving permission to all the page definition for the test-all role. But still I am getting the following error.
    [JpsAuth] Check Permission
    PolicyContext: [UI#V2.0]
    Resource/Target: [getSubjectFromDomainCombiner]
    Action: [null]
    Permission Class: [javax.security.auth.AuthPermission]
    Result: [SUCCEEDED]
    Subject: [null]
    Evaluator: [SM]
    [JpsAuth] Check Permission
    PolicyContext: [UI#V2.0]
    Resource/Target: [sessiondef._FDMApplicationVO_0_DynReg_com_hyperion_aif_gl_common_setup_userinterface_pageDefs_regFDMApplicationsPageDef]
    Action: [view]
    *Permission Class:     [oracle.adf.share.security.authorization.RegionPermission]*
    *Result:               [FAILED]* Evaluator: [ACC]
    Failed ProtectionDomain:ClassLoader=sun.misc.Launcher$AppClassLoader@47858e
    CodeSource=file:/C:/Oracle/Middleware/oracle_common/modules/oracle.adf.share_11.1.1/adf-share-support.jar Principals=total 3 of principals(
    1. JpsPrincipal: oracle.security.jps.internal.core.principals.JpsAnonymousUserImpl "anonymous" GUID=null DN=null
    2. JpsPrincipal: oracle.security.jps.internal.core.principals.JpsAnonymousRoleImpl "anonymous-role" GUID=null DN=null
    3. JpsPrincipal: oracle.security.jps.service.policystore.ApplicationRole "test-all" GUID=48AD91A06D7511DFBFDBE5A03915DD7D DN=null)
    Permissions=(
    oracle.adf.controller.security.TaskFlowPermission//data/data-task-flow-definition.xml#data-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//dim/dimension-task-flow-definition.xml#dimension-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/Access-Denied-task-flow-definition.xml#Access-Denied-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/setup-flow-definition.xml#setup-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//writeback/writeBack-task-flow-definition.xml#writeBack-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/regApp-task-flow-definition.xml#regApp-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//writeback/valueMapping-task-flow-definition.xml#valueMapping-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//process/processDetails-task-flow-definition.xml#processDetails-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/importProfile-task-flow-definition.xml#importProfile-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/categoryMapping-task-flow-definition.xml#categoryMapping-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/srcAcctEnties-task-flow-definition.xml#srcAcctEnties-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//hr/hr-task-flow-definition.xml#hr-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/regSrcSystem-task-flow-definition.xml#regSrcSystem-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/regFDMApps-task-flow-definition.xml#regFDMApps-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/periodMapping-task-flow-definition.xml#periodMapping-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//setup/location-task-flow-definition.xml#location-task-flow-definition/customize,grant,personalize,view
    oracle.adf.controller.security.TaskFlowPermission//dim/memberMapping-task-flow-definition.xml#memberMapping-task-flow-definition/customize,grant,personalize,view
    (java.io.FilePermission \C:\Oracle\Middleware\oracle_common\modules\oracle.adf.share_11.1.1\adf-share-support.jar read)
    (java.net.SocketPermission localhost:1024- listen,resolve)
    (oracle.security.jps.service.credstore.CredentialAccessPermission context=SYSTEM,mapName=*,keyName=* *)
    (java.util.PropertyPermission line.separator read)
    (java.util.PropertyPermission java.vm.version read)
    (java.util.PropertyPermission java.vm.specification.version read)
    (java.util.PropertyPermission java.vm.specification.vendor read)
    (java.util.PropertyPermission java.vendor.url read)
    (java.util.PropertyPermission java.vm.name read)
    (java.util.PropertyPermission os.name read)
    (java.util.PropertyPermission java.vm.vendor read)
    (java.util.PropertyPermission path.separator read)
    (java.util.PropertyPermission java.specification.name read)
    (java.util.PropertyPermission os.version read)
    (java.util.PropertyPermission mds.store.filesystem.path read)
    (java.util.PropertyPermission os.arch read)
    (java.util.PropertyPermission java.class.version read)
    (java.util.PropertyPermission java.version read)
    (java.util.PropertyPermission file.separator read)
    (java.util.PropertyPermission java.vendor read)
    (java.util.PropertyPermission java.vm.specification.name read)
    (java.util.PropertyPermission java.specification.version read)
    (java.util.PropertyPermission java.specification.vendor read)
    (oracle.security.jps.service.policystore.PolicyStoreAccessPermission context=APPLICATION,name=* getApplicationPolicy)
    (java.lang.RuntimePermission stopThread)
    (java.lang.RuntimePermission exitVM)
    oracle.adf.share.security.authorization.RegionPermission/com.hyperion.aif.gl.drilldown.userinterface.pageDefs.drilldownPageDef/view
    oracle.adf.share.security.authorization.RegionPermission/com.hyperion.aif.gl.common.setup.userinterface.pageDefs.MainPageDef/grant,view
    Call Stack: java.security.AccessControlException: access denied oracle.adf.share.security.authorization.RegionPermission/sessiondef._FDMApplicationVO_0_DynReg_com_hyperion_aif_gl_common_setup_userinterface_pageDefs_regFDMApplicationsPageDef/view java.security.AccessControlContext.checkPermission(AccessControlContext.java:323)
    java.security.AccessController.checkPermission(AccessController.java:546)

  • Essbase Calculation Time

    Hi All,
    There is a history cube, there is calc script which used to complete withnin 2 hours.We added a member to a sparse dimension.The cube load process was initiated and the calc script took 20 hoours to complete :(.Its using parelle calculation,kindly advice how to find the root cause for this.
    Thanks in advance.

    Will this be helpful sorry i couldnt reply soon.I added a member to band dimension,thanks for your concern
    ----- Database Information -----
    Database Type : NORMAL
    Status : Loaded
    Elapsed Db Time : 07:12:28:11
    Users Connected : 1
    Blocks Locked : 0
    Dimensions : 15
    Data Status : Data has not been mo
    since last calculati
    Data File Cache Size Setting : 0
    Current Data File Cache Size : 0
    Data Cache Size Setting : 4188672
    Current Data Cache Size : 4188672
    Index Cache Size Setting : 1048576
    Current Index Cache Size : 1048576
    Index Page Size Setting : 8192
    Current Index Page Size : 8192
    Cache Memory Locking : Disabled
    Database State : Read-write
    Data Compression on Disk : Yes
    Data Compression Type : BitMap Compression
    Retrieval Buffer Size (in K) : 20
    Retrieval Sort Buffer Size (in K) : 20
    Isolation Level : Uncommitted Access
    Pre Image Access : No
    Time Out : Never
    Number of blocks modified before internal commit : 3000
    Number of rows to data load before internal commit : 0
    Number of disk volume definitions : 0
    Dimension Name Type Declared Size Actual Size
    ===================================================================
    Measures DENSE 37 24
    Fiscal Years DENSE 5 4
    Category DENSE 8 6
    Final DENSE 3 3
    SI SPARSE 5 5
    ABC SPARSE 12 12
    Product Type SPARSE 14 14
    Periods SPARSE 13 13
    Brand SPARSE 141641 139332
    PT Curr Mo SPARSE 13 13
    Business Division SPARSE 5 5
    Biz 1 SPARSE 8 8
    ABC Curr Mo SPARSE 13 13
    SI Curr Mo SPARSE 4 4
    ABC Defaults SPARSE 3 3
    Number of dimensions : 15
    Declared Block Size : 4440
    Actual Block Size : 1728
    Declared Maximum Blocks : 1546719720
    Actual Maximum Blocks : 1521505440
    Number of Non Missing Leaf Blocks : 4280052
    Number of Non Missing Non Leaf Blocks : 16752323
    Number of Total Blocks : 21032375
    Index Type : B+ TREE
    Average Block Density : 0.9878472
    Average Sparse Density : 1.38234
    Block Compression Ratio : 0.01882556
    Average Clustering Ratio : 0.3272385
    Average Fragmentation Quotient : 15.65848
    Edited by: Umar on Sep 8, 2008 11:33 AM

  • [Target] - Partition definition is not valid: [Cell count mismatch: [1814427] area for slice [1] members per dimension [67 1 1 1 1 1 1 1 51 531 ]]

    Hello Everyone,
    I am receiving the error : [Target] - Partition definition is not valid: [Cell count mismatch: [1814427] area for slice [1] members per dimension [67 1 1 1 1 1 1 1 51 531 ]] when trying to validate my transparent partition.
    Because of some issues the data is not getting reflected at the target database.
    Please suggest some solution.
    Thanks in advance.

    I would exercise the following steps
    1.Copy the original partition script in a notepad
    2.Open one source outline and target outline applications in EAS Console in vertical mode.
    3.Check the partition members coming from both source  to Target application are in sync, by checking the source members and target members in the respective cubes ,if  any new member got added in your source partition intersection then you need the member in the target too,if they are not ,then cell count mismatch will occur.
    4.Since its a transparent partition target the shared transparent areas of the data source and data target outlines need not be identical.The dimensions and member in the data source maps to each dimension and member in the data target.You need to define the definition of mapping
    5.Create partition either using maxl or using create partition in EAS .
    In your case 1 partitions has the cell mismatch counts ,since you are able to retrieve the data for 3 partitions in the target applications
    Thanks,
    Sreekumar Hariharan

  • Definitions Siebel Analytics Dimensions

    I'm new to Siebel Analytics.
    I want to setup a report, and i can't find a list of definitions wich are standard included in Siebel Analytics. (The dimensions section)
    Can anyone tell me where i can find this?
    I can't build a proper report of dashboard when i'm not shure what de definition is of the included items.
    Thanks for the help!
    Best regards, Dennis

    no they cannot be on same machine under same user account

  • I exported a video in .MOV format at 1920x1080p 29.97fps and even though the video file still says it has the dimensions of 1920x1080, it plays back in squished, standard definition. Why is it doing that?

    I don't understand why the video playback is squishing when the .mov file's dimensions are the right aspect ratio. When I exported with the codec .h264, the video played back the way it was supposed to, with HD/16:9 dimensions. Why is the .mov file not exporting/playing back properly?

    Check your Pixel Aspect Settings in Export Panel

  • Can we retrieve Data of Single Dimension in multiplecolumns in Excel Add-in

    Hi All,
    I have strange requirement of having single dimension in two columns while retrieving data from essbase using Excel Add-In.
    Logically I feel it is not possible however we are looking if there is any possibilty of work arround for this requirement.
    Whenevr i am trying to retrieve data using a macro on a single column we are able to successfully do it. But in case of multiple columns we are unable to do it.Do let us know if there is any work arround or alternative method to handle this requirement.
    Regards,
    Krishna

    Not completely sure what you mean by "single dimension in two columns". Normally this request comes down to wanting to show multiple levels from a dimension, e.g...
    _______________"Sales"
    "100" "100-10" 12345
    "100" "100-20" 34567...which is definitely not possible.
    If you just mean data like this (with multiple columns from the same dimension)...
    ________"100-10"   "100-20"
    "Sales"  12345      34567...then that should work fine.
    If you could post a sample of what you'd like to see, it would be easier to answer.

  • Not able to see ikm oracle incremental update and ikm oracle slowly changing dimensions under PHYSCIAL tab in odi 12c

    not able to see ikm oracle incremental update and ikm oracle slowly changing dimensions under PHYSCIAL tab in odi 12c
    But i'm able to see other IKM's please help me, how can i see them

    Nope, It has not been altered.
    COMPONENT NAME: LKM Oracle to Oracle (datapump)
    COMPONENT VERSION: 11.1.2.3
    AUTHOR: Oracle
    COMPATIBILITY: ODI 11.1.2 and above
    Description:
    - Loading Knowledge Module
    - Loads data from an Oracle Server to an Oracle Server using external tables in the datapump format.
    - This module is recommended when developing interfaces between two Oracle servers when DBLINK is not an option.
    - An External table definition is created on the source and target servers.
    - When using this module on a journalized source table, the Journaling table is first updated to flag the records consumed and then cleaned from these records at the end of the interface.

  • I would appreciate it if someone could advise me as to the optimum resolution, dimensions and dpi for actual photographic slides that I am scanning for use in a Keynote Presentation, that will be projected in a large auditorium.  I realize that most proje

    I would appreciate it if someone could advise me as to the optimum resolution, dimensions and dpi for actual photographic slides that I am scanning for use in a Keynote Presentation, that will be projected in a large auditorium. I realize that most projectors in auditoriums that I will be using have 1024 x 1200 pixels, and possibly 1600 x 1200. There is no reference to this issue in the Keynote Tutorial supplied by Apple, and I have never found a definitive answer to this issue online (although there may be one).
                Here’s my question: When scanning my photographic slides, what setting, from 72 dpi to 300 dpi, would result in the best image quality and use up the most efficient amount of space? 
                Here’s what two different photo slide scanning service suppliers have told me: 
    Supplier No. 1 tells me that they can scan slides to a size of 1544 x 1024 pixels, at 72 dpi, which will be 763 KB, and they refer to this as low resolution (a JPEG). However, I noticed when I looked at these scanned slides, the size of the slides varied, with a maximum of 1.8 MB. This supplier says that the dpi doesn’t matter when it comes to the quality of the final digital image, that it is the dimensions that matter.  They say that if they scanned a slide to a higher resolution (2048 x 3072), they would still scan it at 72 dpi.
    Supplier No. 2: They tell me that in order to have a high quality image made from a photographic slide (starting with a 35 mm slide, in all cases), I need to have a “1280 pixel dimension slide, a JPEG, at 300 dpi, that is 8 MB per image.” However, this supplier also offers, on its list of services, a “Standard Resolution JPEG (4MB file/image – 3088 x 2048), as well as a “High Resolution JPEG (8 MB file/image – 3088x2048).
    I will be presenting my Keynotes with my MacBook Pro, and will not have a chance to try out the presentations in advance, since the lecture location is far from my home, so that is not an option. 
    I do not want to use up more memory than necessary on my laptop.  I also want to have the best quality image. 
    One more question: When scanning images myself, on my own scanner, for my Keynote presentations, would I be better off scanning them as JPEGs or TIFFs? I have been told that a TIFF is better because it is less compressed. 
    Any enlightenment on this subject would be appreciated.
    Thank you.

    When it comes to Keynote, I try and start with a presentation that's 1680 x 1050 preset or something in that range.  Most projectors that you'll get at a conference won't project much higher than that and if they run at a lower resolution, it's better to have the device downsize your Keynote.  Anything is better than having the projector try and upsize your presentation... you work hard to make it look good, and it's mangled by some tired Epson projector.
    As far as slides go, scan them in at 150 dpi or better, and make them at least the dimensions of your presentation.  Keynote is really only wanting 72dpi, but I do them at 150, just in case I need to print out the presentation as a handout later, and having the pix at 150 dpi gives me a little help with their quality on a printer.
    You'd probably have to drop in the 150 versions again if you output the Keynote to .pdf or Word or something, but at least you have the option.
    And Gary's right (above) go ahead and scan them as TIFFs.  Sooner or later you'll want to do something else with these slides (like make something for an iPad or the like) and having them as TIFFs keeps your presentation looking good.
    Finally, and this is a big one, get to the location for your presentation ahead of time if you can, and plug the laptop in and see what you get.  There's always connection problems. Don't let the AV bonehead tell you everything will work just fine ('... I don't have any adapters for a Mac...') .  See it for yourself... you're the one that's standing up there.  Unless it's your boss, then you better be really sure it works.

  • Ssrs 2014 errorrule 'Create list of measure group-dimension pairs

    I'm getting this error generating a model for a ssas tabular model in ssrs 2014.  When I test the tabulur definition in Excel under SSAS everything works fine and there are no errors.
    Thanks
    From the log
    semanticmodelgenerator!ReportServer_0-1!b58!06/13/2014-15:39:25:: e ERROR: Semantic Model Generator: ModelGeneratorExtention
    An error occurred while executing rule 'Create list of measure group-dimension pairs':Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'..
    semanticmodelgenerator!ReportServer_0-1!b58!06/13/2014-15:39:25:: e ERROR: Semantic Model Generator: ModelGeneratorExtention
    An error occurred during the generation of semantic model : [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'..
    library!ReportServer_0-1!b58!06/13/2014-15:39:25:: e ERROR: Throwing Microsoft.ReportingServices.Diagnostics.Utilities.ModelGenerationErrorException: , Microsoft.ReportingServices.Diagnostics.Utilities.ModelGenerationErrorException: An error occurred while
    generating model. ---> System.InvalidOperationException: An error occurred during the generation of semantic model : [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'..
    ---> System.InvalidOperationException: [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'. ---> Microsoft.AnalysisServices.Modeling.Generation.ModelGenerationException:
    Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'.
       at Microsoft.AnalysisServices.Modeling.Generation.ModelGenerationDiagnostics.Assert(Boolean condition, String message, Boolean throwOnFail)
       at Microsoft.AnalysisServices.Modeling.Generation.Rules.MapMeasureGroupsToDegenerateDimensions.Process(RuleFlowRecords& generatedRules, RuleExecutionRequirement& rulesExecutionRequirement)
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.ProcessRule(IRule rule, RuleContext ruleContext)
       --- End of inner exception stack trace ---
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.ProcessRule(IRule rule, RuleContext ruleContext)
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.ProcessRuleFlowRecord(RuleFlowRecord ruleFlowRecord)
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.ProcessRule(IRule rule, RuleContext ruleContext)
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.ProcessRuleFlowRecord(RuleFlowRecord ruleFlowRecord)
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.Process()
       --- End of inner exception stack trace ---
       at Microsoft.AnalysisServices.Modeling.Generation.Flow.RuleProcessor.Process()
       at Microsoft.AnalysisServices.Modeling.Generation.ModelGenerator.Generate()
       at Microsoft.AnalysisServices.Modeling.Generation.ModelGeneratorExtention.ReGenerateModel(IDbConnection connection, XmlReader currentModelReader, XmlWriter newModelWriter)
       at Microsoft.ReportingServices.Library.GenerateModelAction.GenerateModel(DataSourceInfo dataSource)
       --- End of inner exception stack trace ---;
    ui!ReportManager_0-2!3f3c!06/13/2014-15:39:25:: e ERROR: System.Web.Services.Protocols.SoapException: System.Web.Services.Protocols.SoapException: An error occurred while generating model. ---> Microsoft.ReportingServices.Diagnostics.Utilities.ModelGenerationErrorException:
    An error occurred while generating model. ---> System.InvalidOperationException: An error occurred during the generation of semantic model : [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for
    cube dimension '[$DimDate].[DimDate]'.. ---> System.InvalidOperationException: [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'. ---> Microsoft.AnalysisServices.Modeling.Generation.ModelGenerationException:
    Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'.
       at Microsoft.ReportingServices.Library.ReportingService2010Impl.GenerateModel(String DataSource, String Model, String Parent, Property[] Properties, CatalogItem& ItemInfo, Warning[]& Warnings)
       at Microsoft.ReportingServices.WebServer.ReportingService2010.GenerateModel(String DataSource, String Model, String Parent, Property[] Properties, CatalogItem& ItemInfo, Warning[]& Warnings)
       at Microsoft.ReportingServices.UI.NewDependentItem.ApplyChanges()
       at Microsoft.ReportingServices.UI.BaseApplyPage.ApplyBtn_Click(Object sender, EventArgs e)
       at System.Web.UI.WebControls.Button.OnClick(EventArgs e)
       at System.Web.UI.WebControls.Button.RaisePostBackEvent(String eventArgument)
       at System.Web.UI.Page.RaisePostBackEvent(IPostBackEventHandler sourceControl, String eventArgument)
       at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint)
    ui!ReportManager_0-2!3f3c!06/13/2014-15:39:25:: e ERROR: HTTP status code --> 200
    -------Details--------
    System.Web.Services.Protocols.SoapException: System.Web.Services.Protocols.SoapException: An error occurred while generating model. ---> Microsoft.ReportingServices.Diagnostics.Utilities.ModelGenerationErrorException: An error occurred while generating model.
    ---> System.InvalidOperationException: An error occurred during the generation of semantic model : [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'..
    ---> System.InvalidOperationException: [Create list of measure group-dimension pairs]Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'. ---> Microsoft.AnalysisServices.Modeling.Generation.ModelGenerationException:
    Unexpected error occurred : Unable to obtain cube schema object for cube dimension '[$DimDate].[DimDate]'.
       at Microsoft.ReportingServices.Library.ReportingService2010Impl.GenerateModel(String DataSource, String Model, String Parent, Property[] Properties, CatalogItem& ItemInfo, Warning[]& Warnings)
       at Microsoft.ReportingServices.WebServer.ReportingService2010.GenerateModel(String DataSource, String Model, String Parent, Property[] Properties, CatalogItem& ItemInfo, Warning[]& Warnings)
       at Microsoft.ReportingServices.UI.NewDependentItem.ApplyChanges()
       at Microsoft.ReportingServices.UI.BaseApplyPage.ApplyBtn_Click(Object sender, EventArgs e)
       at System.Web.UI.WebControls.Button.OnClick(EventArgs e)
       at System.Web.UI.WebControls.Button.RaisePostBackEvent(String eventArgument)
       at System.Web.UI.Page.RaisePostBackEvent(IPostBackEventHandler sourceControl, String eventArgument)
       at System.Web.UI.Page.ProcessRequestMain(Boolean includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint)
    library!ReportServer_0-1!b58!06/13/2014-15:39:25:: i INFO: Call to GetSystemPropertiesAction().
    ui!ReportManager_0-2!3f3c!06/13/2014-15:39:25:: e ERROR: System.Threading.ThreadAbortException: Thread was being aborted.
       at System.Threading.Thread.AbortInternal()
       at System.Threading.Thread.Abort(Object stateInfo)
       at System.Web.HttpResponse.End()
       at Microsoft.ReportingServices.UI.ReportingPage.ShowErrorPage(String errMsg)
    library!WindowsService_2!bd0!06/13/2014-15:41:25:: i INFO: Call to CleanBatch()
    library!WindowsService_2!bd0!06/13/2014-15:41:26:: i INFO: Cleaned 0 batch records, 0 policies, 0 sessions, 0 cache entries, 0 snapshots, 0 chunks, 0 running jobs, 0 persisted streams, 0 segments, 0 segment mappings, 0 edit sessions.
    library!WindowsService_2!bd0!06/13/2014-15:41:26:: i INFO: Call to CleanBatch() ends
    rshost!rshost!3f3c!06/13/2014-19:37:27:: i INFO: Currently registered url http://+:80/ReportServer_SQLEXPRESS/ on endpoint 2
    rshost!rshost!3f3c!06/13/2014-19:37:27:: i INFO: Currently registered url http://+:80/Reports_SQLEXPRESS/ on endpoint 3
    rshost!rshost!3f3c!06/13/2014-19:37:27:: i INFO: Endpoint 4 is disabled and no url is registered vdir=/ReportServer_SQLEXPRESS/ReportBuilder, pdir=C:\Program Files\Microsoft SQL Server\MSRS12.SQLEXPRESS\Reporting Services\ReportServer\ReportBuilder.
    rshost!rshost!e38!06/13/2014-19:37:27:: i INFO: Derived memory configuration based on physical memory as 16487384 KB
    appdomainmanager!DefaultDomain!e38!06/13/2014-19:37:27:: i INFO: Appdomain:5 ReportServer_SQLEXPRESS_0-1-130471616741701355 unregistered.
    library!WindowsService_2!22c!06/13/2014-19:37:29:: i INFO: Call to CleanBatch()
    appdomainmanager!ReportServer_0-1!e38!06/13/2014-19:37:30:: i INFO: Appdomain:5 ReportServer_SQLEXPRESS_0-1-130471616741701355 unloading
    appdomainmanager!DefaultDomain!e38!06/13/2014-19:37:33:: i INFO: Appdomain:6 ReportManager_SQLEXPRESS_0-2-130471623126349107 unregistered.
    appdomainmanager!ReportManager_0-2!e38!06/13/2014-19:37:33:: i INFO: Appdomain:6 ReportManager_SQLEXPRESS_0-2-130471623126349107 unloading
    servicecontroller!DefaultDomain!e38!06/13/2014-19:37:41:: i INFO: Recycling the service from the default domain
    dbpolling!WindowsService_2!376c!06/13/2014-19:37:41:: i INFO: EventPolling polling service stopped
    dbpolling!WindowsService_2!376c!06/13/2014-19:37:41:: i INFO: NotificationPolling polling service stopped
    dbpolling!WindowsService_2!376c!06/13/2014-19:37:41:: i INFO: SchedulePolling polling service stopped
    dbpolling!WindowsService_2!376c!06/13/2014-19:37:41:: i INFO: UpgradePolling polling service stopped
    runningjobs!WindowsService_2!e38!06/13/2014-19:37:41:: e ERROR: Timer Database Cleanup (NT Service) still executing, breaking...
    rshost!rshost!e38!06/13/2014-19:37:48:: i INFO: Application domain type WindowsService statistics: created: 3, unloaded: 3, failed: 0, timed out: 0.
    appdomainmanager!DefaultDomain!e38!06/13/2014-19:37:48:: i INFO: Appdomain:7 WindowsService_3 started.
    library!WindowsService_2!22c!06/13/2014-19:37:52:: e ERROR: Throwing Microsoft.ReportingServices.Diagnostics.Utilities.ReportServerStorageException: , An error occurred within the report server database.  This may be due to a connection failure, timeout
    or low disk condition within the database.;
    library!WindowsService_2!22c!06/13/2014-19:37:52:: w WARN: Transaction rollback was not executed connection is invalid
    dbcleanup!WindowsService_2!22c!06/13/2014-19:37:52:: e ERROR: Error in CleanExpiredSessions: Microsoft.ReportingServices.Diagnostics.Utilities.ReportServerStorageException: An error occurred within the report server database.  This may be due to a connection
    failure, timeout or low disk condition within the database. ---> System.Data.SqlClient.SqlException: Database 'ReportServer$SQLEXPRESSTempDB' is being recovered. Waiting until recovery is finished.
    library!WindowsService_3!e38!06/13/2014-19:37:55:: i INFO: SQLDump flags: ReferencedMemory, AllThreads, SendToWatson
    library!WindowsService_3!e38!06/13/2014-19:37:55:: i INFO: MiniDump flags: DataSegs, UnloadedModules, ProcessThreadData
    library!WindowsService_3!e38!06/13/2014-19:37:55:: i INFO: Dump on: Microsoft.ReportingServices.Diagnostics.Utilities.InternalCatalogException,Microsoft.ReportingServices.Modeling.InternalModelingException,Microsoft.ReportingServices.ReportProcessing.UnhandledReportRenderingException
    library!WindowsService_3!e38!06/13/2014-19:37:55:: i INFO: Do not dump on: System.Threading.ThreadAbortException,System.Web.UI.ViewStateException,System.OutOfMemoryException,System.Web.HttpException,System.IO.IOException,System.IO.FileLoadException,Microsoft.SharePoint.SPException,Microsoft.ReportingServices.WmiProvider.WMIProviderException
    library!WindowsService_3!e38!06/13/2014-19:37:55:: i INFO: Minidump location: C:\Program Files\Microsoft SQL Server\MSRS12.SQLEXPRESS\Reporting Services\Logfiles
    resourceutilities!WindowsService_3!e38!06/13/2014-19:37:56:: i INFO: Reporting Services starting SKU: Developer

    Hi shiftbit,
    According to your description, you get an empty Measure Group Columns list when defining relationship. Right?
    Based on your screenshot, you may notice that Measure group table is empty, so definitely it can't show any column in the dropdown list. If there's columns in the measure group that are related to the dimension columns, at least it will show the fact
    table for Measure group.
    In this scenario, please check if the fact table for the Measure group is existing. Re-process the Measure group and try again. If the issue persists, please re-create that measure group.
    Reference:
    Define Relationship Dialog Box (Analysis Services - Multidimensional Data)
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou
    TechNet Community Support

Maybe you are looking for

  • What is the maximum resolution for an external monitor for a 13" BMP

    I have a new MBP 13 and a new Samsung LED 23" monitor that is spec'd at 1920x1080 resolution. However the maximum resolution I can get (using Display) for the external monitor is 1600x900 (or 1600x1200 stretched). What could be causing this limitatio

  • Some fields from InfoPath form not showing up in Sharepoint

    I created an InfoPath form containing a checklist that sends data to an InfoPath document library in Sharepoint 2010. Most of the fields that I have linked to the sharepoint transfer their data just fine, but 2 of the fields come out blank in SharePo

  • Issue with calling procedure remotely

    Hello, Within the stored procedure, I am calling a procedure remotely but the '@v_remote_db' variable does not resolve: begin for i in cur_database_list loop v_remote_db := i.database_name; p_trunc_remote_table@v_remote_db; <<<------does not resolve

  • All drivers for hp 2000-2311 tu for windows 7 64 bit

    hello sir   i need driver of bluetooth & wifi for hp 2000-2311 tu for windows 7 64 bit so plz give me a link 

  • Select Statment condition

    Hi All, I have a following select statement. SELECT TSTC~PGMNA   TSTC~TCODE   TSTCT~TTEXT   TRDIR~SECU   INTO CORRESPONDING FIELDS OF TABLE IT_TEST   FROM ( TSTC inner join tstct on tstcttcode = tstctcode          INNER JOIN TRDIR ON TRDIRNAME = TSTC