Autosolve during cube build

Hello!
Several days ago 14th of December 2008 we had run several cube builds. During these builds OLAP engine performs Data loading and Auto solve it was looking like this in XML_LOAD_LOG
bq. +20:19:45 Running Jobs: AWXML$_2260_2120. Waiting for Tasks to Finish...+ \\ +20:19:45 Started 20 Finished 19 out of 20 Tasks.+ \\ +20:18:25 Finished Auto Solve for Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT11 Partition.+ \\ +20:15:32 Started Auto Solve for Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT10 Partition.+ \\ +20:15:32 Finished Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT10 Partition. Processed 644523 Records. Rejected 1071 Records.+ \\ +20:14:03 Finished Auto Solve for Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT09 Partition.+ \\ +20:12:19 Started Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT10 Partition.+ \\ +20:12:15 Attached AW SALES.SALES3 in MULTI Mode.+
Now OLAP engine refuses to perform Auto solve during build. So log looks like this:
bq. +15:28:19 Finished Parallel Processing.+ \\ +15:28:18 Finished Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT10 Partition. Processed 15910 Records. Rejected 0 Records.+ \\ +15:28:12 Finished Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT11 Partition. Processed 12104 Records. Rejected 0 Records.+ \\ +15:28:09 Started Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT10 Partition.+ \\ +15:28:08 Attached AW SALES.SALES3 in MULTI Mode.+ \\ +15:28:08 Finished Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT09 Partition. Processed 15193 Records. Rejected 0 Records.+ \\ +15:28:07 Running Jobs: AWXML$_2801_2412, AWXML$_2801_2413, AWXML$_2801_2414.+ \\ +15:28:07 Started 14 Finished 11 out of 14 Tasks.+ \\ +15:28:03 Finished Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT07 Partition. Processed 21047 Records. Rejected 0 Records.+ \\ +15:28:02 Attached AW SALES.SALES3 in MULTI Mode.+ \\ +15:28:02 Started Load of Measures: CAHTCA, CATCCA, MAPCCA, MCPCCA, QTPVCA, QTUNCA from Cube SALES.CUBE. PRT11 Partition.+
We did not change anything.
Cube settings are the same - Use Default Aggregation plan for Cube Aggregation
Scripts are the same with TrackStatus=true and RunSolve=true -
bq. declare \\ +     xml_clob clob;+ \\ +     xml_str varchar2(4000);+ \\ +     isAW number;+ \\ +     begin+ \\ +     DBMS_LOB.CREATETEMPORARY(xml_clob,TRUE);+ \\ +     dbms_lob.open(xml_clob, DBMS_LOB.LOB_READWRITE);+ \\ +     dbms_lob.writeappend(xml_clob, 183, ' <BuildDatabase Id="Action6" AWName="SALES.SALES3" BuildType="EXECUTE" RunSolve="true" CleanMeasures="false" CleanAttrs="false" CleanDim="false" TrackStatus="true" MaxJobQueues="0">');+ \\ +     dbms_lob.writeappend(xml_clob, 39, ' <BuildList XMLIDref="SALES.CUBE" />');+ \\ +     dbms_lob.writeappend(xml_clob, 18, ' </BuildDatabase>');+ \\ +     dbms_lob.close(xml_clob);+ \\ +     xml_str := sys.interactionExecute(xml_clob);+ \\ +     dbms_output.put_line(xml_str);+ \\ +     end;+
Why OLAP engine refuses do Solve measures during build? Of course it is possible to do solve separately, but why it suddenly stops to perform it?
Thank you for your help.
Regards,
Kirill Boyko

Hello!
It seems VERY strange but PRSALES performs normally. It does Autosolve.
22-JAN-09 12.17.34     12:17:34 Finished Parallel Processing.
22-JAN-09 12.17.34     12:17:34 Completed Build(Refresh) of SALES.SALES3 Analytic Workspace.
22-JAN-09 12.17.33     12:17:33 Finished Auto Solve for Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT10 Partition.
22-JAN-09 12.16.49     12:16:49 Finished Auto Solve for Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT11 Partition.
22-JAN-09 12.14.07     12:14:07 Finished Auto Solve for Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT09 Partition.
22-JAN-09 12.13.40     12:13:40 Finished Auto Solve for Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT07 Partition.
22-JAN-09 12.12.30     12:12:30 Finished Load of Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT10 Partition. Processed 1696 Records. Rejected 0 Records.
22-JAN-09 12.12.30     12:12:30 Started Auto Solve for Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT10 Partition.
22-JAN-09 12.12.22     12:12:22 Started Load of Measures: CAHTPR, CATCPR, MAPCPR, MCPCPR, QTPVPR, QTUNPR from Cube PRSALES.CUBE. PRT10 Partition.
22-JAN-09 12.12.22     12:12:22 Running Jobs: AWXML$_2935_2495, AWXML$_2935_2496, AWXML$_2935_2497, AWXML$_2935_2498.
Regards,
Kirill Boyko

Similar Messages

  • ORA-01426: Numeric Overflow During Cube Build (Doc ID 1494869.1)

    Our cube builds starting failing and we received this error after the cube log table sequence reached 32787.
    After recreating the sequence the builds run successfully. This seems to be an unacceptable bug for an enterprise level product.
    I have been unable to find a patch on the Oracle support site for it. Is a patch avaliable?
    COMP_NAME
    VERSION
    OLAP Analytic Workspace
    11.2.0.4.0
    Oracle OLAP API
    11.2.0.4.0
    OLAP Catalog
    11.2.0.4.0

    I found what appears to be a related bug Bug 14627371 - ORA-01426: NUMERIC OVERFLOW DURING CUBE BUILD
    .  This bug is marked as fixed in 12.1 of the database. 
    I think that you may want to pursue this issue with support to see whether this can be back-ported ton 11.2.0.4.0.
    --Ken Chin

  • OLAP error leading to  a corrupted AW during cube build

    Hi all
    My cube load keeps on crashing and the result is a corrupted AW. Sometimes it happens, sometimes it doesn't, thus it is very hard to pin point the cause of the problem
    I would like to emphasize that the Alert log and trace files contain general info about the issue but does not help in resolving the problem.
    Let me explain what is happening.
    I have 6 cubes A, B, C, D , E and F in "MY_AW" and all the cubes are populated via a single dbm_cube.build call as shown below
    DBMS_CUBE.BUILD('CUBE_A, CUBE_B, CUBE_C, CUBE_D, CUBE_E , CUBE F', 'C', FALSE, 5, TRUE, TRUE, FALSE);
    Here is a brief description of the cube structure
    a) CUBE_A, CUBE_B CUBE_C are partitioned along dimension A_A ( i.e. along the same dimension)
    b) CUBE_D is partitioned along dimension B_B ( i.e. along a different dimension)
    c) CUBE_E and CUBE_F' are NOT partitioned ( These cubes are really tiny, thus no need partitioning them)
    All 6 cubes are loaded on a daily basis. Initial reload load is always without any issues. Then the next , say 3 incremental loads would run successfully too.
    After that, any further incremental load generates an Oracle error ( as shown below) and the result is a corrupted AW.
    It is really frustrating because the error is inconsistent( i.e does not surface always) , although the cube structure is constant across all loads..
    Can anybody figure out what is happening here?
    Any help would be highly appreciated.
    ERROR at line 1:
    ORA-37162: OLAP error
    XOQ-01707: Oracle job "JOB$_7041" failed while executing slave build "*CUBE_F* USING (LOAD NO SYNCH, SOLVE) AS OF SCN 11027959160835" with error "37162: ORA-37162: OLAP error
    XOQ-00703: error executing OLAP DML command "(AW ATTACH MY_AW RO : ORA-01403: no data found )"
    ORA-06512: at "SYS.DBMS_CUBE", line 234
    ORA-06512: at "SYS.DBMS_CUBE", line 316
    ORA-06512: at line 1
    ORA-06512: at "SYS.DBMS_CUBE", line 234
    ORA-06512: at "SYS.DBMS_CUBE", line 287
    ORA-06512: at line 3
    Edited by: user9018701 on Jun 11, 2011 4:50 AM

    One recent source of AW corruption is
    BUG 8664896 - ORA-20000: ORU-10027: BUFFER OVERFLOW - AW VALIDATE
    A characteristic error of this bug (which you may see in the alert log) is
    +ORA-00600: internal error code, arguments: [xspggepGenPSErase05], [], [], []+
    This is fixed in the most recent 11g patches, as described by metalink note 1078454.1, so apply these may clear up the issue.
    I also see that you selecting the 'atomic' option in your call to dbms_cube.build. This means you may be a victim of
    BUG 11795047 - USING OLAP "ROLLBACK TO FREEZE" CAN DAMAGE AW
    The issue is that if you choose to run an atomic build and any error occurs during a slave process, then your AW may end up being corrupted. Choosing FALSE for the atomic argument would not fix the underlying error, but it may stop the subsequent AW corruption. As of writing the fix for this bug is not part of any public OLAP patch, so you would need to get a one-off patch through a service request.
    In either case I think a service request would be appropriate here since this is a serious error. You can mention my name if you like and I will help direct the problem to the appropriate developers.

  • Multiple columns (named the same originally) and mapped to the same lookup table are causing a Cube Build issue

    Hey folks, looking for some insight here.
    I've an implementation that contains some custom Enterprise columns mapped to lookup tables.  In the instance I'm working with now, it looks like there was/is an issue with one of those columns.  In this scenario, I have a column named
    ProjectType, created initially with that name, mapped to a lookup table.  This field's name was then changed to
    Project Type.  After that, it looks like another column was created, also called
    ProjectType.  So now, we have what I would have originally thought was two distinct columns, even though the names used are the same.
    Below is the error we're currently getting during the Cube Build Process...
    PWA:http://ps2010/PWA, ServiceApp:Project Web App, User:DOMAIN\user, PSI: SqlException occurred in DAL:  <Error><Class>1</Class><LineNumber>1</LineNumber><Number>4506</Number><Procedure>MSP_EpmProject_OlapView_B8546719-4D4C-473A-84B1-89DEDA2307E0</Procedure> 
    <Message>  System.Data.SqlClient.SqlError: Column names in each view or function must be unique. Column name 'ProjectType' in view or function 'MSP_EpmProject_OlapView_B8546719-4D4C-473A-84B1-89DEDA2307E0' is specified more than once.  </Message> 
    <CallStack>   
     at System.Data.SqlClient.SqlConnection.OnError(SqlException exception, Boolean breakConnection)   
     at System.Data.SqlClient.TdsParser.ThrowExceptionAndWarning(TdsParserStateObject stateObj)   
     at System.Data.SqlClient.TdsParser.Run(RunBehavior runBehavior, SqlCommand cmdHandler, SqlDataReader dataStream, BulkCopySimpleResultSet bulkCopyHandler, TdsParserStateObject stateObj)   
     at System.Data.SqlClient.SqlCommand.FinishExecuteReader(SqlDataReader ds, RunBehavior runBehavior, String resetOptionsString)   
     at System.Data.SqlClient.SqlCommand.RunExecuteReaderTds(CommandBehavior cmdBehavior, RunBehavior runBehavior, Boolean returnStream, Boolean async)   
     at System.Data.SqlClient.SqlCommand.RunExecuteReader(CommandBehavior cmdBehavior, RunBehavior runBehavior, Boolean returnStream, String method, DbAsyncResult result)   
     at System.Data.SqlClient.SqlCommand.InternalExecuteNonQuery(DbAsyncResult result, String methodName, Boolean sendToPipe)   
     at System.Data.SqlClient.SqlCommand.ExecuteNonQuery()   
     at Microsoft.Office.Project.Server.DataAccessLayer.DAL.SubDal.ExecuteStoredProcedureNoResult(String storedProcedureName, SqlParameter[] parameters)  </CallStack>  </Error>
    I've tried deleting the one column, but the build still gives the above error.
    Any thoughts as to how the above could be resolved?
    Thanks! - M
    Michael Mukalian | Jan 2010 - Dec 2010 MVP SharePoint Services | MCTS: MOSS 2007 Configuration | http://www.mukalian.com/blog

    We tried taking it out of the cubes, and it builds fine.  The challenge we're having is in building the cubes with that custom field "ProjectType".  It's as if the cubes still hold some reference to it even when it's deleted.
    Since the OLAP View ('MSP_EpmProject_OlapView_{guid}') is recreated, would it be as simple as deleting that View, and trying to recreate?
    Thanks - M
    Michael Mukalian | Jan 2010 - Dec 2010 MVP SharePoint Services | MCTS: MOSS 2007 Configuration | http://www.mukalian.com/blog

  • How to use HTML Cube builder in SSM 7.0 SP4

    Hi,
    I've just installed SSM 7.0 SP4 (with hotfix 5, 6)
    A blank page is opened when I open cube builder page for building cube/PAS model.
    My configuration settings are below:
    1. Add a new database through the link "Add a new database" in the Strategy Management application tools page.
    2. Create Link ID in PAS
    3. Create model connection in the Strategy Management administration page.
    4. Open cube builder page
    "http://<nwce>:<port>/strategy/cubebuilder/cube_builder.htm"
    Please, let me know if my configuration settings are incomplete or incorrect.
    Or, where can I get information (or document) of it?
    Have a nice day!
    - System environment
    OS: Windows 2003 Server Enterprise Edition R2 SP2 64bit.
    DB: MS SQL Server 2005 Enterprise Edition
    Netweaver CE 7.1 SP6
    SSM 7.0 SP4 (Hot fix 5, 6)

    Santiago,
    Cube Builder is designed to build simple cubes without using PAS. Until very recently Cube Builder was only available to internal or consultants during the Ramp Up process. Now that Strategy Management is out of Ramp Up it has been made available.
    I know that Cube Builder is covered in the training courses for Strategy Management, but not aware of any courses scheduled until next year. Advise you to use the documentation with Cube Builder Help.
    Regards,
    Bob

  • Rejcted records cube build log.

    I am on 11.1.0.7 db with 11.1.0.7B AWM and loading the dimensions and cube. The olapsys.xml_load_log table is not populated with the build log. I can query cube_build_log but xml_load_log give me better information(rejected/processed records). I know there is a bug for this issue on 11.1.
    Can I somehow know the rejected records information other than doing a minus between fact and dimensions? Any workaroud for this?
    Thanks,

    With 11gr2 we have the following logs which will give all the information including rejected dimensions without having to do a minus from fact table.
    Maintenance Logs
    The first time you load data into a cube or dimension using Analytic Workspace Manager, it creates several logs. These logs are stored in tables in the same schema as the analytic workspace:
    •     Cube Build Log: Contains information about what happened during a build. Use this log to determine whether the build produced the results you were expecting, and if not, why not. The log is continually updated whenever a cube or dimension is refreshed, whether by Analytic Workspace Manager, the database materialized view refresh subsystem, or a PL/SQL procedure. You can query the log at any time to evaluate the progress of the build and to estimate the time to completion. The default table name is CUBE_BUILD_LOG.
    •     Cube Dimension Compile Log: Contains errors that occur during the validation of the dimension hierarchies when OLAP is aggregating a cube. The default table name is CUBE_DIMENSION_COMPILE.
    •     Cube Operations Log: Contains messages and debugging information for all OLAP engine events. The default table name is CUBE_OPERATIONS_LOG.
    •     Cube Rejected Records Log: Identifies any records that were rejected because they did not meet the expected format. The default table name is CUBE_DIMENSION_COMPILE.
    You can also run the $ORACLE_HOME/olap/admin/utlolaplog.sql script to create the build log with some useful views.

  • Error while creating a cube using Cube Builder

    we have installed the Cube builder for SSM 7.0
    while creating the cube I am facing a error and the .TRC file contents are as follows :-
    LSS>
    LSS> set control LINKID SSM_CB_EA
    LSS>
    LSS> .... capture cube id in a control var to simplify queries on other tables
    LSS> access lslink
    LSLink> connect SSM_CB_EA
    LSLink>
    LSLink> SELECT ID FROM CPMS_CB_CUBES WHERE NAME = 'CUBEBUILDER1'
    [Microsoft][ODBC SQL Server Driver][SQL Server]Invalid object name 'CPMS_CB_CUBES'.
    SQLSTATE: S0002
    SQL System code: 208
    LSLink> lss create code = 'set control CUBEID ' + ID
    LSLink> output proc setcubeid;PIPADMINDEFAULT over
    LSLink> peek create nohead nonumb
    No Fields currently Selected
    ACC004:
    No Record Has Been Accessed From the Database.
    ACC004:
    No Record Has Been Accessed From the Database.
    CHE FRE
    CHE UPD
    Kindly help me out........

    Hi Martin,
    I believe that the issue you are encountering now is that you're using the wrong set of procedures for accessing MS SQL Server.
    In section 5.4.1 (Setting up SAP NetWeaver System database software) in the Strategy Management configuration guide, there is a section which discusses steps that are required depending on whether or not you are using MaxDB or MS SQL Server as your system database behind NetWeaver CE.  My guess is that you have not followed the steps described for SQL Server here.  This is simple to do thoughm, because all you need to do is copy the files in the "<install-dir>\SAP\SSM\InternetPub\procs\sqlsrvr_procs" into the <install-dir>\SAP\SSM\InternetPub\procs" directory.
    This is required because, when using Cube Builder or Entry and Appropval, the PAS component communicates directly with the system database in order to load data from there into PAS.  Because MaxDB and SQL Server have different SQL syntax, specifically regarding field concatenation, slightly different SQL is required.  So if you copy the SQL Server specific procedures across, you should be able to successfully build your model from Cube Builder.
    Of course, having read through your question again I'm not 100% sure that this is what you are hitting.  The problem you're seeing seems to be that, when we internally concatenate some selected fields together, we aren't finding a field called ID.  However, the SQL select being issued doesn't seem to have returned any error messages, so I'm not sure why there is a problem referencing the field.
    The only thing that comes to mind is a problem that I ran into at one point with the type of driver I was using to connect to SQL Server.  The PAS component of Strategy Management, even though it runs on Windows 2003 Server 64bit, is still only a 32bit component.  This means that, when configuring the System Data Sources within the ODBC applet in Control Panel, it is critical that the 32bit version of that ODBC applet be used - as documented in the note at the end of section 5.4.3 in the configuration guide.  The reason this is important is because, if you use an OLEDB connection to SQL Server, for some reason the field names are not returned correctly.  It is only when using the ODBC driver that the field names are returned correctly.
    So please ensure that you are using the ODBC driver for your Link ID which connects to SQL Server, and not an OLEDB connection.
    Hope this helps,
    Robert
    Edited by: Robert Holland on Dec 19, 2008 3:23 PM

  • Error message during cube creation

    Hi expert,
    System show the following error message during cube creation, please info how to solve this problem. Thanks.
    Define the characteristics of the validity table for non-cumulatives
    Message no. R7846
    Diagnosis
    The InfoCube contains non-cumulative values. A validity table is created for these non-cumulative values, in which the time interval is stored, for which the non-cumulative values are valid.
    The validity table automatically contains the "most detailed" of the selected time characteristics (if such a one does not exist, it must be defined by the user, meaning transfered into the InfoCube).
    Besides the most detailed time characteristic, additional characteristics for the InfoCube can be included in the validity table:
    Examples of such characteristics are:
    A "plan/actual" indicator, if differing time references exist for plan and actual values (actual values for the current fiscal year, plan values already for the next fiscal year),
    the characteristic "plant", if the non-cumulative values are reported per plant, and it can occur that a plant already exists for plant A for a particular time period, but not yet for plant B.
    Procedure
    Define all additional characteristics that should be contained in the validity table by selecting the characteristics.
    In the example above, the characteristics "plan/actual" and  "plant" must be selected.
    The system automatically generates the validity table according to the definition made. This table is automatically updated when data is loaded.

    Hi,
    Go to the Extras Tab in the definition of the Cube and there select " Maintain Non-Cumulatives" and there select the Plant check box. You can also check Material but it is not recommended. Now try to activate the cube. I think it will work for you.

  • Not able to see model name in Cube Builder

    Hello,
    I am working on SSM 10.1, I have successfully Created Model connection CBDEFAULT and Created Link id SSM_CB_EA.
    When I open cube builder link, I am not able to see any Model.
    When I try to created a new model to in cube builder, it is not able to SAVE and no option is appearing in Fiscal Year Start Tab. While saving it neither give any error nor create any model.
    Can any body help me out to solve this issue. It is very urgent.
    Thanks
    Thanks and regards
    Rahul Sharma

    Hi Isabel,
    Now I able to add new model in Cube builder by Having model connection with Pas user "MMADMIN", but when I created model and pressed on "BUILD PAS MODEL" tab, it is giving me error.

  • Slide Hyperlinks not working during a Build In?

    Hi all,
    I'm working on a Keynote presenation set up as a mulitple choice quiz. Below is the idea for how each question is set up.
    Question Slide
    Incorrect answers links to 'Next Slide' - Ex. Slide 7 ('Incorrect' Slide)
    Correct answer links to 'Applicable Slide #' - Ex. Slide 8 ('Correct' Slide)
    'Incorrect' Slide
    Everything on the slide links to 'Previous Slide' to take the user back to the question to try again.
    ^ This is where I'm having an issue that I'm not sure is solvable.
    I've got the 'Flame' Build In on the word 'Incorrect', which seems to negate the links put on that slide if the user taps/clicks while the Build In is in effect.
    For example:
    It runs correctly when -
    On the Question Slide I click the incorrect answer - Keynote jumps to 'Incorrect' Slide
    On the 'Incorrect' Slide I click AFTER the Flames Build In on the word Incorrect is complete - Keynote jumps to the linked Previous Slide (Question Slide). User can now attempt the quetion again.
    It runs incorrectly when -
    On the Question Slide I click the incorrect answer - Keynote jumps to 'Incorrect' Slide
    On the 'Incorrect' Slide I click WHILE the Flames Build In on the word Incorrect is happening - Keynote jumps FORWARD to the 'Correct' Slide (NOT linking to the linked Previous Slide).
    User is forced to move forward to next question.
    So essentially the following happens: Pick incorrect answer> 'Incorrect' Slide appears w/ Flames Build In> Quickly click during flames Build In> 'Correct' Slide appears
    Anyone know if this is just due to the Build In occurring and the slide not registering as having a link to Previous until it's completed? Or something along those lines?

    Links do not work when a transition, action or build effect is in progress, there is no work around.

  • Oracle OLAP cube build question

    Hello,
         I am trying to build a reasonably large cube (around 100
    million rows from the underlying relational fact table). I am using
    Oracle 10g Release 2. The cube has 7 dimensions, the largest of which
    is TIME (6 years of data with the lowest level day). The cube build
    never finishes.
    Apparently it collapses while doing "Auto Solve". I'm assuming this
    means calculating the aggregations for upper levels of the hierarchy
    (although this is not mentioned in any of the documentation I have).
    I have two questions related to this:
    1. Is there a way to keep these aggregations from being performed at
    cube build time on dimensions with a value-based hierarchy? I already
    have the one dimension designated as level-based unchecked in the
    "Summarize To" tab in AW manager (TIME dimension).
    2. Are there any other tips that might help me get this cube built?
    Here is the log from the olapsys.xml_load_log table:
    RECORD_ID LOG_DATE AW XML_MESSAGE
    1. 09-MAR-06 SYS.AWXML 08:18:51 Started Build(Refresh) of APSHELL Analytic Workspace.
    2. 09-MAR-06 SPADMIN.APSHELL 08:18:53 Attached AW APSHELL in RW Mode.
    3. 09-MAR-06 SPADMIN.APSHELL 08:18:53 Started Loading Dimensions.
    4. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members.
    5. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for ACCOUNT.DIMENSION (1 out of 9 Dimensions).
    6. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for ACCOUNT.DIMENSION. Added: 0. No Longer Present: 0.
    7. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for CATEGORY.DIMENSION (2 out of 9 Dimensions).
    8. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for CATEGORY.DIMENSION. Added: 0. No Longer Present: 0.
    9. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for DATASRC.DIMENSION (3 out of 9 Dimensions).10. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for DATASRC.DIMENSION. Added: 0. No Longer Present: 0.
    11. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for ENTITY.DIMENSION (4 out of 9 Dimensions).
    12. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for ENTITY.DIMENSION. Added: 0. No Longer Present: 0.
    13. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for INPT_CURRENCY.DIMENSION (5 out of 9 Dimensions).
    14. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for INPT_CURRENCY.DIMENSION. Added: 0. No Longer Present: 0.
    15. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for INTCO.DIMENSION (6 out of 9 Dimensions).
    16. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for INTCO.DIMENSION. Added: 0. No Longer Present: 0.
    17. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for RATE.DIMENSION (7 out of 9 Dimensions).
    18. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for RATE.DIMENSION. Added: 0. No Longer Present: 0.
    19. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for RPTCURRENCY.DIMENSION (8 out of 9 Dimensions).
    20. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Finished Loading Members for RPTCURRENCY.DIMENSION. Added: 0. No Longer Present: 0.
    21. 09-MAR-06 SPADMIN.APSHELL 08:18:54 Started Loading Dimension Members for TIME.DIMENSION (9 out of 9 Dimensions).
    22. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Finished Loading Members for TIME.DIMENSION. Added: 0. No Longer Present: 0.
    23. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Finished Loading Dimension Members.
    24. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Started Loading Hierarchies.
    25. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Started Loading Hierarchies for ACCOUNT.DIMENSION (1 out of 9 Dimensions).
    26. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Finished Loading Hierarchies for ACCOUNT.DIMENSION. 1 hierarchy(s) ACCOUNT_HIERARCHY Processed.
    27. 09-MAR-06 SPADMIN.APSHELL 08:18:55 Started Loading Hierarchies for CATEGORY.DIMENSION (2 out of 9 Dimensions).
    28. 09-MAR-06 SPADMIN.APSHELL 08:18:56 Finished Loading Hierarchies for CATEGORY.DIMENSION. 1 hierarchy(s) CATEGORY_HIERARCHY Processed.
    29. 09-MAR-06 SPADMIN.APSHELL 08:18:56 Started Loading Hierarchies for DATASRC.DIMENSION (3 out of 9 Dimensions).
    30. 09-MAR-06 SPADMIN.APSHELL 08:18:56 Finished Loading Hierarchies for DATASRC.DIMENSION. 1 hierarchy(s) DATASRC_HIER Processed.
    31. 09-MAR-06 SPADMIN.APSHELL 08:18:56 Started Loading Hierarchies for ENTITY.DIMENSION (4 out of 9 Dimensions).
    32. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for ENTITY.DIMENSION. 2 hierarchy(s) ENTITY_HIERARCHY1, ENTITY_HIERARCHY2 Processed.
    34. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for INPT_CURRENCY.DIMENSION. No hierarchy(s) Processed.
    36. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for INTCO.DIMENSION. 1 hierarchy(s) INTCO_HIERARCHY Processed.
    37. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Hierarchies for RATE.DIMENSION (7 out of 9 Dimensions).
    38. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for RATE.DIMENSION. No hierarchy(s) Processed.
    39. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Hierarchies for RPTCURRENCY.DIMENSION (8 out of 9 Dimensions).
    40. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for RPTCURRENCY.DIMENSION. No hierarchy(s) Processed.
    41. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Hierarchies for TIME.DIMENSION (9 out of 9 Dimensions).
    42. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies for TIME.DIMENSION. 2 hierarchy(s) CALENDAR, FISCAL_CALENDAR Processed.
    43. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Hierarchies.
    44. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes.
    45. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for ACCOUNT.DIMENSION (1 out of 9 Dimensions).
    46. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for ACCOUNT.DIMENSION. 6 attribute(s) ACCTYPE, CALC, FORMAT, LONG_DESCRIPTION, RATETYPE, SCALING Processed.
    47. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for CATEGORY.DIMENSION (2 out of 9 Dimensions).
    48. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for CATEGORY.DIMENSION. 2 attribute(s) CALC, LONG_DESCRIPTION Processed.
    49. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for DATASRC.DIMENSION (3 out of 9 Dimensions). 50. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for DATASRC.DIMENSION. 3 attribute(s) CURRENCY, INTCO, LONG_DESCRIPTION Processed.
    51. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for ENTITY.DIMENSION (4 out of 9 Dimensions).
    52. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for ENTITY.DIMENSION. 3 attribute(s) CALC, CURRENCY, LONG_DESCRIPTION Processed.
    53. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for INPT_CURRENCY.DIMENSION (5 out of 9 Dimensions).
    54. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for INPT_CURRENCY.DIMENSION. 2 attribute(s) LONG_DESCRIPTION, REPORTING Processed.
    55. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for INTCO.DIMENSION (6 out of 9 Dimensions).
    56. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Finished Loading Attributes for INTCO.DIMENSION. 2 attribute(s) ENTITY, LONG_DESCRIPTION Processed.
    57. 09-MAR-06 SPADMIN.APSHELL 08:18:57 Started Loading Attributes for RATE.DIMENSION (7 out of 9 Dimensions).
    58. 09-MAR-06 SPADMIN.APSHELL 08:18:58 Finished Loading Attributes for RATE.DIMENSION. 1 attribute(s) LONG_DESCRIPTION Processed.
    59. 09-MAR-06 SPADMIN.APSHELL 08:18:58 Started Loading Attributes for RPTCURRENCY.DIMENSION (8 out of 9 Dimensions).
    60. 09-MAR-06 SPADMIN.APSHELL 08:18:58 Finished Loading Attributes for RPTCURRENCY.DIMENSION. 2 attribute(s) LONG_DESCRIPTION, REPORTING Processed.
    61. 09-MAR-06 SPADMIN.APSHELL 08:18:58 Started Loading Attributes for TIME.DIMENSION (9 out of 9 Dimensions).
    62. 09-MAR-06 SPADMIN.APSHELL 08:20:26 Finished Loading Attributes for TIME.DIMENSION. 3 attribute(s) END_DATE, LONG_DESCRIPTION, TIME_SPAN Processed.
    63. 09-MAR-06 SPADMIN.APSHELL 08:20:26 Finished Loading Attributes.
    64. 09-MAR-06 SPADMIN.APSHELL 08:20:26 Finished Loading Dimensions.
    65. 09-MAR-06 SPADMIN.APSHELL 08:20:26 Started Updating Partitions.
    66. 09-MAR-06 SPADMIN.APSHELL 08:20:26 Finished Updating Partitions.
    67. 09-MAR-06 SPADMIN.APSHELL 08:20:40 Started Loading Measures.
    68. 09-MAR-06 SPADMIN.APSHELL 08:20:40 Started Load of Measures: SIGNEDDATA from Cube FINANCE.CUBE.
    69. 09-MAR-06 SPADMIN.APSHELL 10:54:06 Finished Load of Measures: SIGNEDDATA from Cube FINANCE.CUBE. Processed 100000001 Records. Rejected 0 Records.
    70. 09-MAR-06 SPADMIN.APSHELL 10:54:06 Started Auto Solve for Measures: SIGNEDDATA from Cube FINANCE.CUBE.

    Hi, I've taken a few minutes to do a quick analysis. I just saw in your post that this isn't "real data", but some type of sample. Here is what I'm seeing. First off, this is the strangest dataset I've ever seen. With the exception of TIME, DATASOURCE, and RPTCURRENCY, every single other dimension is nearly 100% dense. Quite truthfully, in a cube of this many dimensions, I have never seen data be 100% dense like this (usually with this many dimensions its more around the .01% dense max, usually even lower than that). Is it possible that the way you generated the test data would have caused this to happen?
    If so, I would strongly encourage you to go back to your "real" data and run the same queries and post results. I think that "real" data will produce a much different profile than what we're seeing here.
    If you really do want to try and aggregate this dataset, I'd do the following:
    1. Drop any dimension that doesn't add analytic value
    Report currency is an obvious choice for this - if every record has exactly the same value, then it adds no additional information (but increases the size of the data)
    Also, data source falls into the same category. However, I'd add one more question / comment with data source - even if all 3 values DID show up in the data, does knowing the data source provide any analytical capabilities? I.e. would a business person make a different decision based on whether the data is coming from system A vs. system B vs. system C?
    2. Make sure all remaining dimensions except TIME are DENSE, not sparse. I'd probably define the cube with this order:
    Account...........dense
    Entity..............dense
    IntCo...............dense
    Category.........dense
    Time...............sparse
    3. Since time is level based (and sparse), I'd set it to only aggregate at the day and month levels (i.e. let quarter and year be calculated on the fly)
    4. Are there really no "levels" in the dimensions like Entity? Usually companies define those with very rigid hierarchies (assuming this means legal entity)
    Good luck with loading this cube. Please let us know how "real" this data is. I suspect with that many dimensions that the "real" data will be VERY sparse, not dense like this sample is, in which case some of the sparsity handling functionality would make a huge benefit for you. As is, with the data being nearly 100% dense, turning on sparsity for any dimension other than TIME probably kills your performance.
    Let us know what you think!
    Thanks,
    Scott

  • [svn] 3347: -remove the javascript directory since it is getting cleaned and recreated during the build process .

    Revision: 3347
    Author: [email protected]
    Date: 2008-09-25 06:27:28 -0700 (Thu, 25 Sep 2008)
    Log Message:
    -remove the javascript directory since it is getting cleaned and recreated during the build process.
    -this should make SVN happy and get the builds working again
    Removed Paths:
    flex/sdk/trunk/frameworks/javascript/FABridge/samples/srcview/source/fabridge/javascript/

    - Where? That depends upon what you what. But, going by the letter of what you asked, you shouldn't deploy any applications or packages if you want them on your newly images systems, you should make them a part of the TS using Install Software and Install
    Application tasks.
    - How many times? You should only need one although depending upon what you are deploying and if you are using offline updates, some folks add more. I typically use two one right after the Setup Windows and ConfigMgr task (just like the task sequence wizard
    builds for you) to update the OS and "things" in the image and one at the end to catch additional updates for apps and components added during the TS. Note that you also have to initiate an update scan cycle for additional install update tasks to work. Do
    this by adding a run-command line task before the second Install Updates task with the proper WMIC incantation.
    - Offline Updates only injects CBS updates into the image which are typically just core OS updates. You should also update your image though to include non-CBS updates. Using a build and catpure task sequence makes this easy.
    Jason | http://blog.configmgrftw.com

  • Error during warehouse builder repository assistant

    hi every body ,
    i have Oracle 11g and working on warehouse builder 11g too .
    ( OS = windows Xp Sp2 )
    both DB and warehouse have the same oracle home .
    i have created the first db called : DB1
    during warehouse builder repository assitant i saw this error :
    OWBSYS is not granted access to c:\oracle\product\11.1.0\db_1\owb\bin\admin\rtrepos.properties: Please run Unified/Repos/reset_owbcc_home.sql specifying the path of the Oracle Home from which the Control Center Servie is being run.
    i fixed it , by adding one entry to reset_owbcc_home.sql and every thing goes well .
    i have created another Oracle DB called DB2 .
    both DB are in one oracle home .
    when i run OWB repository assistant for DB2 i see that message again and i can not continue .
    pls guid me
    thanks
    Padideh ( [email protected] )

    hi David ,
    i didn't get your mean . for DB1 i had these entries in reset_owbcc_home.sql :
    1- dbms_java.grant_permission('OWBSYS','java.io.FilePermission', path||'\owb\bin\admin\rtrepos.properties','read,write');
    2- dbms_java.grant_permission('OWBSYS','java.io.FilePermission', unix_path||'/owb/bin/admin/rtrepos.properties', 'read,write');
    3- dbms_java.grant_permission('OWBSYS','java.io.FilePermission', path||'/owb/bin/unix/run_service.sh', 'read,execute');
    4- dbms_java.grant_permission('OWBSYS','java.io.FilePermission', path||'\owb\bin\win32\run_service.bat', 'read,execute');
    the first entry is for windows that i have added to current entries .
    i can not understand how come oracle can not give permission to another DB for that properties file . all situation is the same as for DB1 .
    regards
    Padideh

  • PAS Data Load  iin Cube Builder Dimensions

    Hi All,
      Two important questions about SSM Implementation , that are impacting our development
      1 - Is it possible to develop a data load to fill dimensions created in Cube Builder ? For example, the user wants to be able to fill dimensions manually in SSM BUT      wants to load some aditional information via data load un the same Cube created via Cube Builder Tool  What I see is that, when we fill the dimension manually in Cube Builder a internal code is created in PAS Database
          INPUT
          L0M1281099797397 '001'
          but I am in doubt about to reprocude the same code via data load in a PAS Procedure
      2 - My customer in his original system maintains a relationship between Iniatitives Is it possibile to do the same in SAP SSM ? I looked for the documentation anda I had not found anyting
          associated with this
      Regards,
         Cristian

    Hi Cristian,
    Jus for clarification: do you want to modify a dimension that was created through Cube Builder? Or do you want to create and maintain a new dimension without going through Cube Builder?
    If you are trying to create a new dimension, how many times would this dimension changed? And would the change be made manually or would the structure be available in a table?
    I usually would suggest to have the dimension structure in a table and use PAS procedures to create/re-create the dimension. You can assign whatever technical name and label you want, and as long as you maintain the same technical name for each member, you should be able to recreate dimensions without losing any data.
    If this is a dimension that you won't change anymore you can also code it directly in PAS with the structure you find in the other dimensions:
    INPUT
    input_member1_technical_name 'input_member1-label',
    input_member2_technical_name 'input_member2-label',
    input_member3_technical_name 'input_member3-label',
    input_member4_technical_name 'input_member4-label'
    OUTPUT
    output_member1_technical_name 'output_member1-label',
    output_member2_technical_name 'output_member2-label',
    RESULT
    result_member_technical_name 'result_member-label'
    output_member1_technical_name = SUM
    input_member1_technical_name 'input_member1-label',
    input_member2_technical_name 'input_member2-label'
    output_member2_technical_name = SUM
    input_member3_technical_name 'input_member3-label',
    input_member4_technical_name 'input_member4-label'
    result_member_technical_name = SUM
    output_member1_technical_name,
    output_member2_technical_name
    Best regards,
    Ricardo Vieira

  • 10g: Cube  build performance:Waiting for tasks to finish

    I had a question related to cube build performance. We see the following in the OLAPSYS.XML_LOAD_LOG prior to completion of parallel processing. So basically the loading of records etc. happens in 2-3 mins but then the part for "running jobs and started 6 finished 5 out of 6 tasks takes almost 30-40 minutes to complete". At this point in background it invokes the xmlloader and xml_parallel_loader. Any insights as to what specifically this steps do (solving of measures etc.) ? Also what we need to modify in terms of xml tag or anything else to improve these timings?
    14:10:56 Started 6 Finished 5 out of 6 Tasks.
    14:10:56 Running Jobs: AWXML$_4361_77872. Waiting for Tasks to Finish...
    14:01:04 Started Auto Solve for Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition.
    14:01:04 Finished Load of Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition. Processed 189616 Records. Rejected 0 Records.
    Thanks,
    Sudip

    When we checked the scheduler jobs we saw xmlparallel_loader. in terms of v$session_wait we saw a db file sequential read. Following is a snapshot from the stage of finished updating partitions to starting of parallel load. Please advise.
    XML_MESSAGE
    14:49:12 Completed Build(Refresh) of AWM_CR2.CRICUBE Analytic Workspace.
    14:49:12 Finished Parallel Processing.
    14:49:10 Finished Auto Solve for Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition.
    14:40:59 Started 6 Finished 5 out of 6 Tasks.
    14:40:59 Running Jobs: AWXML$_4361_77872. Waiting for Tasks to Finish...
    14:30:58 Started 6 Finished 5 out of 6 Tasks.
    14:30:58 Running Jobs: AWXML$_4361_77872. Waiting for Tasks to Finish...
    14:20:57 Running Jobs: AWXML$_4361_77872. Waiting for Tasks to Finish...
    14:20:57 Started 6 Finished 5 out of 6 Tasks.
    14:10:56 Running Jobs: AWXML$_4361_77872. Waiting for Tasks to Finish...
    14:10:56 Started 6 Finished 5 out of 6 Tasks.
    14:01:04 Started Auto Solve for Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition.
    14:01:04 Finished Load of Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition. Processed 189616 Records. Rejected 0 Records.
    14:01:02 Finished Auto Solve for Measures: TOT_AMT from Cube CR_BASE_FIN.CUBE. 2008 Partition.
    14:01:00 Finished Load of Measures: TOT_AMT from Cube CR_BASE_FIN.CUBE. 2008 Partition. Processed 1494 Records. Rejected 0 Records.
    14:01:00 Finished Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_OSUC.CUBE. 200901 Partition.
    14:01:00 Started Auto Solve for Measures: TOT_AMT from Cube CR_BASE_FIN.CUBE. 2008 Partition.
    14:01:00 Finished Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_ENR.CUBE. 200901 Partition.
    14:01:00 Finished Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_FIN.CUBE. 200901 Partition.
    14:01:00 Finished Auto Solve for Measures: TOT_AMT from Cube CR_BASE_ENR.CUBE. 2008 Partition.
    14:00:58 Finished Load of Measures: TOT_AMT from Cube CR_CURR1_FIN.CUBE. 200901 Partition. Processed 1494 Records. Rejected 0 Records.
    14:00:58 Started Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_FIN.CUBE. 200901 Partition.
    14:00:58 Finished Load of Measures: TOT_AMT from Cube CR_BASE_ENR.CUBE. 2008 Partition. Processed 1577 Records. Rejected 0 Records.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_CURR1_OSUC.CUBE. 200901 Partition.
    14:00:58 Finished Load of Measures: TOT_AMT from Cube CR_CURR1_OSUC.CUBE. 200901 Partition. Processed 4521 Records. Rejected 0 Records.
    14:00:58 Started Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_OSUC.CUBE. 200901 Partition.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_CURR1_FIN.CUBE. 200901 Partition.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_BASE_ENR.CUBE. 2008 Partition.
    14:00:58 Started Auto Solve for Measures: TOT_AMT from Cube CR_CURR1_ENR.CUBE. 200901 Partition.
    14:00:58 Finished Load of Measures: TOT_AMT from Cube CR_CURR1_ENR.CUBE. 200901 Partition. Processed 1577 Records. Rejected 0 Records.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_CURR1_ENR.CUBE. 200901 Partition.
    14:00:58 Started Auto Solve for Measures: TOT_AMT from Cube CR_BASE_ENR.CUBE. 2008 Partition.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_BASE_FIN.CUBE. 2008 Partition.
    14:00:58 Started Load of Measures: TOT_AMT from Cube CR_BASE_OSUC.CUBE. 2008 Partition.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:57 Attached AW AWM_CR2.CRICUBE in MULTI Mode.
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871, AWXML$_4361_77872, AWXML$_4361_77873.
    14:00:55 Started 4 Finished 0 out of 6 Tasks.
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871, AWXML$_4361_77872, AWXML$_4361_77873, AWXML$_4361_77874.
    14:00:55 Started 6 Finished 0 out of 6 Tasks.
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871, AWXML$_4361_77872, AWXML$_4361_77873, AWXML$_4361_77874, AWXML$_4361_77875.
    14:00:55 Started 6 Finished 0 out of 6 Tasks.
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871, AWXML$_4361_77872, AWXML$_4361_77873, AWXML$_4361_77874, AWXML$_4361_77875. Waiting for Tasks to Finish...
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871, AWXML$_4361_77872.
    14:00:55 Started 3 Finished 0 out of 6 Tasks.
    14:00:55 Running Jobs: AWXML$_4361_77870, AWXML$_4361_77871.
    14:00:55 Started 2 Finished 0 out of 6 Tasks.
    14:00:55 Running Jobs: AWXML$_4361_77870.
    14:00:55 Started 1 Finished 0 out of 6 Tasks.
    14:00:55 Starting Parallel Processing.
    14:00:55 Detached AW AWM_CR2.CRICUBE.
    14:00:55 Started 5 Finished 0 out of 6 Tasks.
    14:00:22 Finished Updating Partitions.

Maybe you are looking for