Issue in loading data into OLAP dim/cube.

All,
Facing this issue while executing the dimension in OLAP.
Warning:-_
O/*RA-20101: 17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.
TRUNCATE_LOAD=false
AW Execution status: Success
17:39:12 Started Build(Refresh) of MIS Analytic Workspace.
17:39:12 Attached AW MIS in RW Mode.
17:39:12 Started Loading Dimensions.
17:39:12 Started Loading Dimension Members.
17:39:12 Started Loading Dimension Members for DIM_SST.DIMENSION (1 out of 1 Dimensions).
17:39:12 Failed to Build(Refresh) MIS Analytic Workspace.
17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.

Hi there,
What version of 10g OLAP are you using?
It can be very important to apply the latest patches where possible - more details can be found here:- http://www.oracle.com/technology/products/bi/olap/collateral/olap_certification.html
Thanks,
Stuart Bunby
OLAP Blog: http://oracleOLAP.blogspot.com
OLAP Wiki: http://wiki.oracle.com/page/Oracle+OLAP+Option
OLAP on OTN: http://www.oracle.com/technology/products/bi/olap/index.html
DW on OTN : http://www.oracle.com/technology/products/bi/db/11g/index.html

Similar Messages

  • Error encountered while loading data into an Essbase cube from oracle DB

    Hi All,
    I am trying to load data into an Essbase cube from Oracle db. Essbase is installed in an 64 bit HP-UX machine and oracle database on HP -UX 11.23 B server. While performing the above I am getting the below error messages:
    K/INFO - 1021013 - ODBC Layer Error: [08001] ==> [[DataDirect][ODBC Oracle Wire Protocol driver][Oracle]ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    HP-UX Error: 2: No such file or directory].
    OK/INFO - 1021014 - ODBC Layer Error: Native Error code [1034] .
    ERROR - 1021001 - Failed to Establish Connection With SQL Database Server. See log for more information.
    Can you please help me out in identifying the issue, as why it could have occured , because of network problem or is there anything related to databse?
    Regards,
    Priya

    The error is related to Oracle, I would check that it is all up and running
    ORA-01034: ORACLE not available
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Issue with loading data into cube with duplicate members in different dimensions

    Have a cube with allowed dublicate members only in two dims (Period1 and Period2).
    Outlines saves with no errors, but when i try to load data, i get some errors:
    "Member 2012-12 is a duplicate member in the outline
    A6179398-68BD-7843-E0C2-B5EE81808D0B    01011    cd00    st01    2905110000    EK    fo0000    NNNN-NN-NN    2012-12    cust00000$$$    1" 
    These dims represent two similar periods of time. Am I to change member names and aliases a little in one of them(f.e. yyyy1-mm --> yyyy1 - mm1)?
    Users wouldnt like it...
    Period1
      yyyy1
        yyyy1/q1
          yyyy1-mm1
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Period2
      yyyy1
        yyyy1/q1
          yyyy1-mm1
            yyyy1-mm1-dd1 
            yyyy1-mm1-dd2
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Tnanx

    You may have to use fully qualified name something like
    [Period1].[2012-12]
    [Period2].[2012-12]
    You can refer to ESSBASE admin guide on "Creating and Working With Duplicate Member Outlines"
    Regards,
    Sunil

  • Semantic Partitioning Delta issue while load data from DSO to Cube -BW 7.3

    Hi All,
    We have created the semantic partion with the help of BADI to perform Partitions. Everthing looks good,
    first time i have loaded the data it is full update.
    Second time i init the data load. I pulled the delta  from DSO to Cube . DSO is standard where as Cube is partition with the help of semantic partition. What i can see is the records which are updated in Latest delta are shown into the report rest all are ignored by the system.
    I tried compression but still it did not worked.
    Can some has face this kind
    Thanks

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue while loading data from DSO to Cube

    Gurus,
    I'm facing a typical problem while loading Cube from a DSO. The load is based upon certain conditions.
    Though the data in DSO satisfies those condition, but the Cube is still not being loaded.
    All the loads are managed through Process Chains.
    Would there be any reason in specific/particular for this type of problem?
    Any pointers would be of greatest help !
    Regards,
    Yaseen & Soujanya

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Unable to load data into Planning cube

    Hi,
    I am trying to load data into a planning cube using a DTP.
    But the system throws a message saying that, real time data loading cannot be performed to the planning cube.
    What would be the possible reason for the same?
    Thanks & Regards,
    Surjit P

    Hi Surjit,
    To load data into cube using DTP, it should be put in loading mode only.
    The real time cube behaviour is set to planning mode only when the data is input to the cube via the planning layouts / or through the Fileupload program.
    You can change the behaviour of the cube in RSA1 -> Right click on the cube -> Change Real-time load behaviour and select the first option(real time cube can be loaded with data; planning not allowed)
    Best Rgds
    Shyam
    Edited by: Syam K on Mar 14, 2008 7:57 AM
    Edited by: Syam K on Mar 14, 2008 7:59 AM

  • Getting an error while trying to load data into cube

    Hello,
    I am trying to load data into cube but it gives me following error:
    fiscal year variant not expected.
    Can someone tell what could be done on fiscal year
    Cheers
    Jim.

    Hi,
    May be you have not mapped anything to fiscal variant or not included fiscal variant in the cube but you have used fiscal period inside it.
    You should load fiscal variant if you wanna use fiscal period.
    Map fiscal variant to a constant K4 or from the source if present and then see the result.
    Generally it is K4 except for FI-CO cubes .
    In the transfer rules, select field '0fiscvarnt' 'Tp' click
    Enter K4 for Constant................Save and Activate
    Iam sure you'd fix the issue.
    Regards,
    Ray

  • Short Dump when loading Data into Cube UNCAUGHT_EXCEPTION(Pls Help)

    Dear Guru's
    i am trying to load data into a Cube from a ODS and the load gets dump...
    In Dump Analysis.....
    The exception 'ZCX_EDW_MAP_ROOT' was raised, but it was not caught anywhere         
    along   
    "UNCAUGHT_EXCEPTION" "ZCX_EDW_MAP_ROOT"                                       
    "ZCL_EDW_MAP_ROOT==============CP" or "ZCL_EDW_MAP_ROOT==============CM002"   
    "LOAD_MAPPING_INFOPROV"                                                                               
    So it would be great if some one can give me some suggestion to handle this issue
    Thanks in Adv
    Dev

    Dear Arun,
    i have checked the routine there exists some code
    Eg: data: vref_monitor type ref to zcl_edw_monitor,
    But to say fact i ddnot i understood it....
    Any suggestions dude...pls
    thanks in adv
    Dev

  • Error while loading data into cube

    hi BW gurus,
    when ever i am trying to load data into the cube from flat file after scheduling iam getting short dump in BW system. I checked it in st22 it is giving me a error as exception add_partition_failed. please help me to sort out this problem. If you know the error recovery please give me the answer in detail.
    I will assign points for good answers.

    This is what the note says:
    Symptom
    The process of loading transaction data fails because a new partition cannot be added to the F fact table.The loading process terminates with a short dump.
    Other terms
    RSDU_TABLE_ADD_PARTITION_ORA, RSDU_TABLE_ADD_PARTITION_FAILED, TABART_INCONSITENCY, TSORA, TAORA , CATALOG
    Reason and Prerequisites
    The possible causes are:
    SQL errors when creating the partition
    Inconsistencies in the Data Dictionary control tables TAORA and TSORA
    Solution
    BW 3.0A & BW 3.0B
    In the case of SQL errors:Analyze the SQL code in the system log or short dump and if possible, eliminate the cause. The cause is often a disk space problem or lock situations on the database catalog or, less frequently: the partitioning option in the ORACLE database is not installed.
    The most common cause of the problem is inconsistencies in the TAORA and TSORA tables. As of Support Package 14 for BW 3.0B/Support Package 8 for BW 3.1C, the TABART_INCONSITENCY exception is issued in this case. The reason is almost always missing entries in TSORA for the tablespaces of the DDIM, DFACT and DODS data classes.
    The TAORA table contains the assignment of data classes to data tablespaces and their attributes, for example:
    Tablespace Data class
    DDIM PSAPDIMD ........
    DFACT PSAPFACTD ........
    DODS PSAPODSD .......
    Foreach data tablespace, the TSORA table must contain an entry for the corresponding index tablespace, for example:
    TABSPACE INDSPACE
    PSAPDIMD PSAPDIMD
    PSAPFACTD PSAPFACTD
    PSAPODSD PSAPODSD
    In most cases, these entries are missing and have to be added. See also notes 502989 and 46272.

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Error when loading data into Planning

    Unable to load data into planning 11.1.2.3.200 using ODI 11.1.1.7
    Please find the errors from below logs:
    INFO  [SimpleAsyncTaskExecutor-2]: Oracle Data Integrator Adapter for Hyperion Planning
    INFO  [SimpleAsyncTaskExecutor-2]: Connecting to planning application [xxxxxxx] on [xxxxxxxxxxx]:[xxxx] using username [admin].
    INFO  [SimpleAsyncTaskExecutor-2]: Successfully connected to the planning application.
    INFO  [SimpleAsyncTaskExecutor-2]: The load options for the planning load are
      Dimension Name: Account Sort Parent Child : false
      Load Order By Input : false
      Refresh Database : false
    INFO  [SimpleAsyncTaskExecutor-2]: Begining the load process.
    DEBUG [SimpleAsyncTaskExecutor-2]: Number of columns in the source result set does not  match the number of planning target columns.
    INFO  [SimpleAsyncTaskExecutor-2]: Load type is [Load dimension member].
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A603010, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -100, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A601060, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -250, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    log.err
    Account,Data Load Cube Name,Budget,Point-of-View,Error_Reason
    A603010,xxxxx,-100,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    A601060,xxxxx,-250,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    FDMEE log:
    ERROR [AIF]: Error: No records exist for Period 'Pd 2 - 2014-08-01'
    ERROR [AIF]: Error: No records exist for Period 'Pd 3 - 2014-09-01'
    FDMEE Logging Level is set to 5

    Are you sure that you haven't got Planning 11.1.2.3.500 in your environment? This sounds spookily similar to the issue described in note 1678759.1 which was seen after applying Planning 11.1.2.3.500. If it's definitely Planning 11.1.2.3.500 is there any chance that someone has applied ODI patch 18687916 to the ODI home (C:\Oracle\Middleware\odi)?
    If you /*are*/ running Planning 11.1.2.3.500 then you might need to apply patch 18687916 to C:\Oracle\Middleware\odi - you might need note 1683307.1 to take some of the pain out of applying this patch though.
    Given the error that you're seeing though there's a mismatch in versions of JAR files between ODI and the Planning server somewhere.
    Regards
    Craig

  • Job cancelled While loading data from DSO to CUBE using DTP

    Hi All,
      while i am loading data frm DSO to CUBE the job load is getting cancelled.
    in the job overview i got the following messages
        SYST: Date 00/01/0000 not expected.
       Job cancelled after system exception ERROR_MESSAGE
    wt can be the reason for this error as i have successfully loaded data into 2 layers of DSO before loading the data into the CUBE

    hi,
    Are u loading the flat file to DSO?
    then check the data in the PSA in the date field and replace with the correct date.
    i think u r using write optimised DSO which is similar to PSA it will take all the data from PSA to DSO.
    so clear the data in PSA and then load to DSO and then to CUBE.
    If u dont need a date field in CUBE then remove the mapping in transformation in cube and activate and make DTP activate and trigger it it will work.

  • Unable to load data from DSO to Cube

    Good morning all,
    I was trying to load data from DSO to Cube for validation. Before loading the new data, I deleted all from DSO and Cube. They contain no request at all. Cube has "Delta Update". First DSO was loaded 138,300 records successfully. Then Activated the DSO. The last when I clicked Exectue (DSO --> Cube), it loaded 0 record. I was able to load the data yesterday. What might be the reasons for this situation?
    Thank you so much!

    Hi BI User,
    For loading delta upload into the data target, there should be an initialization request in the data target with same selection criteria.
    So..first do the initialization and perform delta upload into the cube.
    Regards,
    Subhashini.

  • FDMEE Import error "No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'

    Hi,
    We are having trouble while importing one ledger 'GERMANY EUR GGAAP'. It works for Dec 2014 but while trying to import data for 2015 it gives an error.
    Import error shows " RuntimeError: No periods were identified for loading data into table 'AIF_EBS_GL_BALANCES_STG'."
    I tried all Knowledge docs from Oracle support but no luck. Please help us resolving this issue as its occurring in our Production system.
    I also checked all period settings under Data Management> Setup> Integration Setup > Global Mapping and Source Mapping and they all look correct.
    Also its only happening to one ledger rest all ledgers are working fine without any issues.
    Thanks

    Hi,
    there are some Support documents related to this issue.
    I would suggest you have a look to them.
    Regards

  • How to load data from PSA to CUBE & DSO at a time using DTP in BI 7 ?

    HI all,
    I am new to BI 7 . How to load the data at same time to DSO & INFO CUBE using DTP.
    Please provide me steps to load & plz specify which update mode I have to use ( FULL OR DELTA ) which one is best.
    Plz Suggest me.
    Thanks & Regards,
    Kiran m.
    Message was edited by:
            kiran manyam

    Below are the basic steps which we follow in any BI 2004S system:
    1)Create datasource. Here u can set/check the Soucre System fields.
    2)Create Transformation for that datasource. (no more update rules/transfer rules)
    2.1) While creating transformation for DS it will ask you for data target name, so just assign where u want to update ur data.
    DataSource -> Transformation -> Data Target
    Now if you want to load data into data target from Source System Datasource:
    1) Create infopackage for that data source. If you are creating infopackage for new datasources, it will only allow you update upto PSA, all other options u can see as disabled.
    2)Now Create DTP (Data Transfer Process) for that data source.
    3) NOw schdule the Infopackage, once the data is loaded to PSA, you can execute your DTP which will load data to data target.
    If you are loading data from one one data target to other, no need to use PSA, you can directly execute DTP in that case.
    Data Source -> Transformation (IP/DTP) -> Data Target1 -> DTP ->Data Target 2
    Use the below link for detailed example:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/fc61e12d-0a01-0010-2883-e2fc63ef729b
    Infosources are no more mandatory with BI 7.0, below is the link to scenarios where we use infosources:
    http://help.sap.com/saphelp_nw04s/helpdata/en/44/0243dd8ae1603ae10000000a1553f6/content.htm
    Full or delta depends on your requirement...
    chk the below thread to know better
    difference between the various loads
    hope it helps
    Message was edited by:
            sriram viswanathan

Maybe you are looking for