View Cube Partitioned Data...

Hi
I partitioned my Cube from Jan2010 to Dec2010. Now i want to see its data individually for each partition. Is it possible?
Thanks...

check the links in this post.
Display of data after cube partition
bottom line : partition are logical features of the cube, not physical. they are not new tables.
M.

Similar Messages

  • I used a partitioned HDD for time machine, using a partition already containing other data files. I am now no longer able to view that partition in Finder. Disk Utility shows it in grey and "not mounted". Any suggestions of how to access the files?

    I used a partitioned HDD for time machine, using a partition already containing other data files. I am now no longer able to view that partition in Finder. Disk Utility shows it in grey and "not mounted". Any suggestions of how to access the files? Does using time machine mean that that partition is no longer able to be used as it used to be?
    HDD is a Toshiba 1TB, partitioned into two 500GB partitions.
    OS X version 10.9.2

    Yes, sharing a TM disk is a bad idea, and disks are cheap enough so that you don't need to.
    Now
    Have you tried to repair the disk yet

  • BIB-9509 OLAPI error viewing cubes data

    Hi, I'm using 11.1.0.7 Oracle Server and AWM 11.1.0.7.0B. Last week I created a cube, mantained it and I could view all the data. Today when I try to view the cube data I get this error:
    An error has occurred on the server
    Error class: OLAPI
    Server error descriptions:
    DPR: Unable to create server cursor, Generic at TxsOqDefinitionManager::crtCurMgrs4
    OES: ORA-00600: internal error code, arguments: xscbTest, 18446744073709551615], [, ], [, ], [, ], [, ], [, []
    , Generic at TxsRdbSQLQuery::absolute()
    at oracle.dss.dataSource.QueryEvaluator.setUpCursors(QueryEvaluator.java:3979)
    at oracle.dss.dataSource.QueryEvaluator._setUpCursorsForMainQuery(QueryEvaluator.java:2822)
    at oracle.dss.dataSource.QueryEvaluator.getCursorForCube(QueryEvaluator.java:2854)
    at oracle.dss.dataSource.QueryEvaluator.createCubeAndCursor(QueryEvaluator.java:1234)
    at oracle.dss.dataSource.MainQueryEvaluator.createCubeAndCursor(MainQueryEvaluator.java:171)
    at oracle.dss.dataSource.QueryServer.createCubeAndCursor(QueryServer.java:3863)
    at oracle.dss.dataSource.QueryServer._createCommonQuery(QueryServer.java:3822)
    at oracle.dss.dataSource.QueryServer._initCubeQuery(QueryServer.java:3702)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at oracle.dss.util.Operation.execute(Operation.java:72)
    at oracle.dss.dataSource.OperationQueue.update(OperationQueue.java:68)
    at oracle.dss.dataSource.common.BaseOperationQueue.addOperation(BaseOperationQueue.java:176)
    at oracle.dss.dataSource.common.BaseOperationQueue.addOperation(BaseOperationQueue.java:146)
    at oracle.dss.dataSource.QueryServer.queueOperation(QueryServer.java:7079)
    at oracle.dss.dataSource.QueryServer.initCubeQuery(QueryServer.java:3652)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at oracle.dss.util.Operation.execute(Operation.java:72)
    at oracle.dss.dataSource.QueryManagerServer.sendQueue(QueryManagerServer.java:2001)
    at oracle.dss.dataSource.common.OperationQueue.update(OperationQueue.java:207)
    at oracle.dss.dataSource.common.BaseOperationQueue.addOperation(BaseOperationQueue.java:176)
    at oracle.dss.dataSource.common.BaseOperationQueue.addOperation(BaseOperationQueue.java:146)
    at oracle.dss.dataSource.client.QueryClient._initCubeQuery(QueryClient.java:2385)
    at oracle.dss.dataSource.client.QueryClient.initCubeQuery(QueryClient.java:2368)
    at oracle.olap.awm.dataobject.dialog.olapi.UBiBeanUtil.doQueryTop3(Unknown Source)
    at oracle.olap.awm.dataobject.dialog.olapi.UBiBeanUtil.showCrosstab(Unknown Source)
    at oracle.olap.awm.dataobject.dialog.olapi.UBiBeanUtil.showData(Unknown Source)
    at oracle.olap.awm.navigator.node.olapi.UCubeNode.actionPerformed(Unknown Source)
    at javax.swing.AbstractButton.fireActionPerformed(Unknown Source)
    at javax.swing.AbstractButton$Handler.actionPerformed(Unknown Source)
    at javax.swing.DefaultButtonModel.fireActionPerformed(Unknown Source)
    at javax.swing.DefaultButtonModel.setPressed(Unknown Source)
    at javax.swing.AbstractButton.doClick(Unknown Source)
    at javax.swing.plaf.basic.BasicMenuItemUI.doClick(Unknown Source)
    at javax.swing.plaf.basic.BasicMenuItemUI$Handler.mouseReleased(Unknown Source)
    at java.awt.Component.processMouseEvent(Unknown Source)
    at javax.swing.JComponent.processMouseEvent(Unknown Source)
    at java.awt.Component.processEvent(Unknown Source)
    at java.awt.Container.processEvent(Unknown Source)
    at java.awt.Component.dispatchEventImpl(Unknown Source)
    at java.awt.Container.dispatchEventImpl(Unknown Source)
    at java.awt.Component.dispatchEvent(Unknown Source)
    at java.awt.LightweightDispatcher.retargetMouseEvent(Unknown Source)
    at java.awt.LightweightDispatcher.processMouseEvent(Unknown Source)
    at java.awt.LightweightDispatcher.dispatchEvent(Unknown Source)
    at java.awt.Container.dispatchEventImpl(Unknown Source)
    at java.awt.Window.dispatchEventImpl(Unknown Source)
    at java.awt.Component.dispatchEvent(Unknown Source)
    at java.awt.EventQueue.dispatchEvent(Unknown Source)
    at java.awt.EventDispatchThread.pumpOneEventForFilters(Unknown Source)
    at java.awt.EventDispatchThread.pumpEventsForFilter(Unknown Source)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
    at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
    at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
    at java.awt.EventDispatchThread.run(Unknown Source)
    anyone can help me??

    It is clearly a database bug. It could be related to
    Bug 7329946 - ORA-600 [XSCBTEST] WHEN TRYING TO VIEW CUBE DATA WITH AWM DATA VIEWER
    The AWM data viewer uses the BI Beans code, so it is probably the same thing. It is related to
    Bug 8719947 - CC CRASHES WHEN THERE IS NEGATIVE VALUES IN CCINCR TUPLES
    My advice is to raise a service request so that you can supply proper trace file information and can get a patch for your specific platform.
    For a faster response to OLAP server related problems you could post to the OLAP Forum
    OLAP

  • Creation in cube base date wise partitions in sssiss

    hi
     i was  created the cube.    the partitions are created  date wise in dynamical with ssis and   process the  the partitions also dynamically  and
     dimension also  process and   how to refresh the cube and partitions and measures and
      dimension process 
       how to create the that package 

    Hi sheshu0022,
    According to your description, you want to use SSIS package to create cube partition. Right?
    In Integrated Services, you can create cube partition with script task. Using the script task, you can save and modify partition through AMO (Analysis Management Objects). Please refer to the link below:
    http://consultingblogs.emc.com/jamiethomson/archive/2008/09/15/ssis-create-analysis-services-partitions-from-a-ssis-package.aspx 
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • Cannot view cube data in mmeasure data viewer

    I am using AWM(10.2.3A) with Oracle EE 10g(10.2.1)with the appr patchset(for enabling OLAP option).
    I have created dimensions(time,product),cube (cube1),measures (sales) in a AW,
    and mapped it to the source tables as well.
    in time dimension,i created levels like,year,quarter,month
    in product i created just 1 level,drink.
    time i mapped to a year_table table which i created as follows:
    create table year_table(yid number,year varchar2(30),qid number,qtr varchar2(30),mid number primary key,month varchar2(30));
    product i mapped to product_table which i created as follows:
    create table product_table(pid number primary key,product varchar2(30));
    sales i mapped to sales_table which i created as follows:
    create table sales_table(id number primary key,mid number,pid number,sales number));
    for all tables i inserted values.
    when i did a maintain on AW,it was successfull.
    after maintaining ,when i try to view data on my cube,using Measure Data Viewer,
    I was able to see my dimensions,but no data.
    also in AWM right hand canvas in mappings when i do a right click on my fact table(for cube mapping) i am not able to see the. data.
    Can anybody help me in this regarrd.

    If you check the OLAPSYS.XML_LOAD_LOG table you will find the log for your build. Did you manage to load any dimension members? The log will tell you how many records were loaded and how many were rejected.
    It is strange that you cannot view the source data from your fact table. Try launching AWM using the AWMC.EXE file. This will launch AWM via a command window. All error messages will be recorded in this window. Try viewing the data from your source table and see if any additional error messages are recorded in the log window.
    One thought - when you connect to AWM do you provide a full JDBC connect string (host:port:sid) or do you supply a TNSNAME entry? Try using the JDBC connection string and see if that resolves the problem.
    Keith Laker
    Oracle EMEA Consulting
    BI Blog: http://oraclebi.blogspot.com/
    DM Blog: http://oracledmt.blogspot.com/
    BI on Oracle: http://www.oracle.com/bi/
    BI on OTN: http://www.oracle.com/technology/products/bi/
    BI Samples: http://www.oracle.com/technology/products/bi/samples/

  • Remote cube on Data Source 0FI_GL_4

    HI,
    I have a requirement of  reconciling  the Line Items from R/3 to BW.
    So i tried to create Remote cube with Direct access, but coudnt due to 0FI_GL_4 data source doesnt support Direct access and reconciliation.
    I can see that Data Source  0FI_GL_4  is being fetched by Function module BWFID_GET_FIGL_ITEM and the base table is DTFIGL_4.
    My question is can I create the Generic data source based on the function module by taking the copy of existing FM BWFID_GET_FIGL_ITEM  and change Direct access option to 2(Supports Preaggregation).
    Do i need to change any settings at table level and parameters to get the required solution.
    Please sugget me how to proceed the development to reconcile the Line Items from R3 to BW.
    Regards,
    Siva..
    Edited by: Siv Kishore on Apr 7, 2011 4:59 PM

    HI Banerjee,
    Thanks for the update.
    As you said i cant use the data source 0FI_GL_1, Because we need to reconcile many fields from R/3 to BW, but  those fields are not available in this data source and it will be complicated to enhance such fields.
    can I create View based generic data source on BSEK and BSEG tables, Will it work? if yes do i need to follow any criteria to get the Correct Line Items.
    Please suggest me is there any SAP suggested solution or if you worked on such kind of development.
    Regards,
    Siva Thottempudi..
    Edited by: Siv Kishore on Apr 12, 2011 4:06 PM

  • Authorization issue to view cube contents

    Hi Gurus,
      I am getting Authorization issue to view cube contents in Production server, When I execute the cube it is showing me the following statement.
    "You do not have sufficient authorization for the infoprovider ZMMG_C05".
    Please provide me a possible solution for this.
    Thanks,
    Jackie.

    Hi,
    Two things to be checked with respect to authorization for this one.
    1) Functional Roles: Check whether Info cube is present in the functional roles that are assigned to you.
                                  If not you need to get the functional role in which the Infocube is assigned.
    2) Data Access Roles: Check in the data access roles assigned to you, whether you have the access
                                      to the selection that you are using to see the data in the info cube. Else, request
                                      BASIS team to assign the appropriate data access roles to you.
    Hope this helps.
    Regards,
    Bharat

  • Materialized view and Partition

    Hi,
    I want to discuss a scenario, with which i am dealing nowadays. there is only a single table (named livecdr ) has 22-23 fields and the records are updating in every sec (on daily basis) and in 24hrs livecdr contains approximately more than 2 crore records.
    Problem is that the query processing time is so slow due to large number of records and query is fetching records from a view created on livecdr table. I want to know the options, how the query would fetch these records in a short time. Firstly i did range partitioning on livecdr and then i created materlized view but the view didn't create on partition table and when i removed partitioning it was created. After view creation, query performance is much better. i also applied indexing on livecdr.
    Can you please tell me, whether this approach is okay or i am doing mistake. should i create partitioning on materialized view and how would i select records from view if partitioning would be applied on materialized view. Please guide me with best approaches.
    thankyou.

    Hello,
    Here is how I created partitioned table (monthly partitions) and partitioned mview on the table. This is just an example you might have to change the mview partitions from monhtly to yearly or quartely.
    CREATE TABLE live_cdr
       CALLID              VARCHAR2 (20),
       CD                  VARCHAR2 (25),
       CG                  VARCHAR2 (25),
       RE                  VARCHAR2 (25),
       OPC                 VARCHAR2 (12),
       DPC                 VARCHAR2 (12),
       STIME               VARCHAR2 (25),
       SDATE               DATE,
       ATIME               VARCHAR2 (25),
       ADATE               DATE,
       ETIME               VARCHAR2 (25),
       EDATE               DATE,
       schangeTime         VARCHAR2 (25),
       schangeDate         DATE,
       STATUS              VARCHAR2 (25),
       bucket              NUMBER (20),
       CIC                 NUMBER (20),
       charge              VARCHAR2 (30),
       ported              VARCHAR2 (30),
       jurisdiction_Info   VARCHAR2 (30),
       origLine_Info       CHAR (1),
       Variant             NUMBER,
       auto_inc            NUMBER
    PARTITION BY RANGE (EDATE)
      PARTITION LIVE_CDR_JAN_2009 VALUES LESS THAN (TO_DATE(' 2009-02-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
        LOGGING
        NOCOMPRESS, 
      PARTITION LIVE_CDR_FEB_2009 VALUES LESS THAN (TO_DATE(' 2009-03-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
        LOGGING
        NOCOMPRESS
    );Mview definition
    CREATE MATERIALIZED VIEW KLONDIKE.LIVE_CDR_MV
    LOGGING
    PARTITION BY RANGE (EDATE)
      PARTITION LIVE_CDR_MV_JAN_2009 VALUES LESS THAN (TO_DATE('20090201',
          'YYYYMMDD'))
        LOGGING
        NOCOMPRESS,
      PARTITION LIVE_CDR_MV_FEB_2009 VALUES LESS THAN (TO_DATE('20090301',
          'YYYYMMDD'))
        LOGGING
        NOCOMPRESS
    BUILD IMMEDIATE
    REFRESH COMPLETE
    START WITH SYSDATE
    NEXT SYSDATE  + 2 /24
    AS
    SELECT   DISTINCT (ic.callid) callid,
                      ic.CG,
                      ic.CD,
                      ic.RE,
                      ic.OPC,
                      ic.DPC,
                      ic.SDATE,
                      ic.STIME,
                      ac.ADATE,
                      ac.ATIME,
                      ec.ETIME,
                      ec.EDATE,
                      ec.STATUS
      FROM   live_cdr ic, live_cdr ac, live_cdr ec
    WHERE       ic.callid = ac.callid
             AND ic.callid = ec.callid
             AND ic.stime = ac.stime
             AND ic.stime = ec.stime
             AND ic.sdate = ac.sdate
             AND ic.sdate = ec.sdate
             AND ac.ADATE IS NOT NULL
             AND ec.EDATe IS NOT NULL
             AND ec.etime IS NOT NULL
             AND ic.CD IS NOT NULL
             AND ic.cg IS NOT NULL;Regards

  • DATACOPY with partitioned data source

    Hello,
    I am actually working on the Hyperion Essbase EPM 11.1.2
    I am trying to perform a Datacopy command in a HBR.
    Data source are partitioned with a transparent partition.
    This datacopy do not work. The HBR is really fast but when I retrieve data no data are copied.
    I am wondering if this is possible to perform a datacopy command, in a HBR with partitioned data source.
    I have managed to perform this datacopy throw a HBR using the command:
    SET CREATEBLOCKONEQ
    Target = Source;
    But this method is really long.
    Thanks for your answer,
    Arthur.

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • PCA cube partitioning how to determine based on 0fiscper/0calmonth

    Dear Gurus,
         Can anyone give me information, how the partition criteria will be decided. What things need to consider to determine go for partition on 0calmonth or 0fiscper and how to determine no of partitions? And once we partitioned the cube, the data stores in different partitioned tables or its a single table where it maintains the different indicators as partitions.
    Thanks in advance,
    Ram

    Hi Ram,
    The partitioning is a process done internally in the Database. This creates partition areas for each of the fiscal periods/calendar months and stores the data. Which would help the SQL query to fetch the records pertaining to each of the periods.
    Since the cube which you are talking about is a PCA cube, its advisable to use 0fiscper as partitioning.
    Let me know if this helps.
    Regards,
    Avinash

  • Help needed in BI content sales over view cube

    We have activated sales over view cube. Just wondering about update routines for each of KFs. They all have similar logical patterns in examples below
    Could any one explain me why this logic of multiplying by -1 is applied for each KF.
    Also, I need to know, if I should use same routines for updating in ODS in place of updating directly in sales overview cube.
    Example - 1
    RESULT = COMM_STRUCTURE-/BIC/ZSHIPQTY.
      IF COMM_STRUCTURE-DOC_CATEG CA DEB_CRED.
        RESULT = RESULT * ( -1 ).
      ENDIF.
      RETURNCODE = 0.
    Example - 2
    RESULT = COMM_STRUCTURE-ACT_DL_QTY.
      IF COMM_STRUCTURE-DOC_CATEG CA CREDIT.
        RESULT = RESULT * ( -1 ).
      ENDIF.
    Example -3
    CLEAR RESULT.
      RESULT = COMM_STRUCTURE-NO_DEL_IT.
      IF COMM_STRUCTURE-DOC_CATEG CA CREDIT.
        RESULT = RESULT * ( -1 ).
      ENDIF.
      RETURNCODE = 0.
    Points will be awarded for good answers.
    Thanks
    Jeff
    Edited by: Jeff Edwards on Aug 11, 2008 7:00 AM
    Edited by: Jeff Edwards on Aug 11, 2008 7:00 AM

    So instead of putting update routine in ODS update rule, I can put it between ODS and cube.
    yes.
    Now, question is.. is 1:1 KF mapping in ODS is fine.
    yes.
    Initial ODS KF value:     30
    ODS set to overwrite: credit posting 10
    KF value in ODS:  10
    Update to cube: -10
    values of KF in cube:    30
    second value:             -10
    during reporting 30 + (-10) = 20.
    All this is hypothetical for your understanding. The real data model should be based on ur requirements. This is one way of doing it.

  • I see 'enq: JI - contention' when building multiple cubes/partitions

    Version 11.2.0.3
    I can successfully build multiple partitions of a cube simultaneously by supplying the degree of parallelism that I want. I can also build multiple cubes and multiple partitions of multiple cubes by submitting separate jobs (one per cube) with parallelism set in the job (for number of partitions per job/cube).
    My goal was to refresh 2 cubes simultaneously, 2 partitions in parallel each, so that 4 partitions total were refreshing simultaneously. There were sufficient hardware resources (memory and processes) to do this. I tried to submit 2 jobs, one for each cube, with parallel 2 on each.
    What happens is that 3 partitions start loading, not 4. The smaller of the 2 cubes loads 2 partitions at a time, but the larger of the cubes starts loading only 1 partition and the other partition process waits with JI - contention.
    I understand that JI contention is related one materialized view refresh blocking another refresh of the same MV. Yet simultaneous refresh of different partitions is supported for cube MVs.
    Because I see the large cube having the problem but not the smaller one, I wonder if adding more hash partitions to the AW$ (analytic workspace) table would allow more concurrent update processes. We have a high enough setting for processes and job_queue_processes, and enough available threads, etc.
    Will more hash subpartitions on the AW$ table allow for more concurrency for cube refreshes?

    It looks like the JI contention was coming from having multiple jobs submitted to update the SAME cube (albeit different partitions). Multiple jobs for different cubes (up to one job/cube each) seems to avoid this issue. I thought there was only one job per cube, but that was not true.
    Still, if someone has some insight into creating more AW hash subpartitions, I'd like to hear it. I know how to do it, but I am not sure what the impact will be on load or solve times. I have read a few sources online indicating that it is a good idea to have as many subpartitions as logical cube partitions, and that it is a good idea to set the subpartition number to a power of two to ensure good balance.

  • Problem in viewing cubes with Cube Viewer

    I have installed oracle 9.2.0.1.0 in windows 2000 platform and created a oracle's standard data warehousing database. When I try to use the cube viewer in Enterprise manager Console to view a cube(SALES_CUBE or COST_CUBE which are the sample cubes given with the database) a dialog box is displayed while reading cube meta data. Problem is this dialog does not display any error message(It is empty except for the red icon) but when ok is clicked cube viewer terminates and also the windows service of the database stops.
    Can you please help.

    hmm.... I had experience with Oracle and Macoye DB. They save the data in ISO. In order to view the data in IE u need to get them as UTF8. So I think thats the only problem u see them as ????. In oder to test this, please view it in IE and when u see ????. Select View -> Encoding and select Arabic (Window) from IE toolbar. if you see the correct Arabic, than it simply means that you need to convert ur Arabic Charset into UTF8 after getting it from DB, but before displaying it in IE.

  • Allowing parallel processing of cube partitions using OWB mapping

    Hi All,
    Iam using an OWB mapping to load a MOLAP cube partitioned on TIME dimension. I configured the OWB mapping by checking the 'Allow parallel processing' option with the no.of parallel jobs to be 2. I then deployed the mapping.The data loaded using the mapping is spread across multiple partitions.
    The server has 4 CPU's and 6 GB RAM.
    But, when i kick off the mapping, i can see only one partition being processed at a time in the XML_LOAD_LOG.
    If i process the same cube in AWM, using parallel processing, i can see that multiple partitions are processed.
    Could you pls suggest if i missed any setting on OWB side.
    Thanks
    Chakri

    Hi,
    I have assigned the OLAP_DBA to the user under which the OWB map is running and the job started off.
    But, it failed soon with the below error:
    ***Error Occured in __XML_MAIN_LOADER: Failed to Build(Refresh) XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace. In __XML_VAL_MEASMAPS: In __XML_VAL_MEASMAPS_VAR: Error Validating Measure Mappings. In __XML_FND_PRT_TO_LOAD: In __XML_SET_LOAD_STATUS: In ___XML_LOAD_TEMPPRG:
    Here is the log :
    Load ID     Record ID     AW     Date     Actual Time     Message Time     Message
    3973     13     SYS.AWXML     12/1/2008 8:26     8:12:51     8:26:51     ***Error Occured in __XML_MAIN_LOADER: Failed to Build(Refresh) XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace. In __XML_VAL_MEASMAPS: In __XML_VAL_MEASMAPS_VAR: Error Validating Measure Mappings. In __XML_FND_PRT_TO_LOAD: In __XML_SET_LOAD_STATUS: In ___XML_LOAD_TEMPPRG:
    3973     12     XPRO_OLAP_NON_AGG.OLAP_NON_AGG     12/1/2008 8:19     8:12:57     8:19:57     Attached AW XPRO_OLAP_NON_AGG.OLAP_NON_AGG in RW Mode.
    3973     11     SYS.AWXML     12/1/2008 8:19     8:12:56     8:19:56     Started Build(Refresh) of XPRO_OLAP_NON_AGG.OLAP_NON_AGG Analytic Workspace.
    3973     1     XPRO_OLAP_NON_AGG.OLAP_NON_AGG     12/1/2008 8:19     8:12:55     8:19:55     Job# AWXML$_3973 to Build(Refresh) Analytic Workspace XPRO_OLAP_NON_AGG.OLAP_NON_AGG Submitted to the Queue.
    Iam using AWM (10.2.0.3 A with OLAP Patch A) and OWB (10.2.0.3).
    Can anyone suggest why the job failed this time ?
    Regards
    Chakri

  • Can not view 24 hours data in Websphere performance page

    We configured EM 10.2.0.5 to monitor Websphere 6.1 on AIX.
    Now we can see current data in the performance page.
    But when we switch to 24 hours view, there is not any data shown in the page.
    Does EM 10.2.0.5 support viewing Websphere historical data?

    This issue is solved.
    The OMR parameter job_queue_processes is set to 0.
    It should be 10 or above.

Maybe you are looking for

  • Report for deviation of yield

    Hello Gurus, Is there a standard report available in ECC6.0 to check all the process orders that had deviated from the yields. Thanking you in anticipation Sri.

  • Capture Excise Invoice w.r .t  Without Purchase order

    HI When i try to post J1IEX i.e Capture Excise Invoice w.r .t  Without Purchase order i am getting the following error message : Balance in transaction currency Message no. F5702 Diagnosis A balance has occurred in transaction currency 'INR' with the

  • Middle click paste no longer works correctly zsh

    After a recent update my middle click to paste no longer works right. Instead or just pasting "x" it will paste "00~x01~". I tried terminator and rxvt-unicode and they do the same. This is a recent issue that just started after an update. Is anyone e

  • Binding Apple Xserver with XSAN 3 to AD

    Have anyone had any issue binding Apple Server 10.8.5 running Xsan 3 to Active Directory?

  • What is an iso image and how do i get it

    I just completed downloading the windows support software but still dont know where do i get the windows 8 iso image