ERROR WHILE PROCECESSING A SSAS CUBE

the following exception occured while the managed IDataReader interface was being used[.NET Data Provider for Teradata] The session is not open, send or receive operation is not allowed.
I am getting the above error while processing my cube agasint Production DATABASE( We use TD as the source) from my SQL DEV SERVER where as same cube is running fine from my PRODUCTION SQL SERVER towards TD database with out any issues. Can any one advise
what could be the reason?

Hi Grangar,
According to the error message, it seems that the issue is related to the provider for Treadata database. To connect to a data source, we must have the appropriate provider installed on your computer. You said that the same cube is running fine from my production
SQL Server towards Treadata database with out any issues. We can use TDOLEDB OLE DB provider or .Net Data Provider for Teradata to connect to Treadata database. In your scenario, please install the corresponding provider in the Dev server and check if this
issue persists or not.
Reference
http://msdn.microsoft.com/en-IN/library/ms175608.aspx
http://msdn.microsoft.com/en-us/library/gg471695.aspx
Regards,
Charlie Liao
TechNet Community Support

Similar Messages

  • ENT-06954: Error while Displaying BIBean for Cube/Dimension Dataviewer.

    Hi all,
    I have defined a cube with three dimensions. All elements are deployed and the mappings is executed successfully. If I try to open the dataviewer either for the cube or for the dimensions I receive the following errors:
    ENT-06954: Error while Displaying BIBean for Cube Dataviewer.
         at oracle.wh.ui.owbcommon.dataviewer.dimensional.CubeDataViewerMain.getDataviewObject(CubeDataViewerMain.java:391)
         at oracle.wh.ui.owbcommon.dataviewer.dimensional.CubeDataViewerEditor._init(CubeDataViewerEditor.java:66)
         at oracle.wh.ui.editor.Editor.init(Editor.java:1115)
         at oracle.wh.ui.editor.Editor.showEditor(Editor.java:1431)
         at oracle.wh.ui.owbcommon.IdeUtils._tryLaunchEditorByClass(IdeUtils.java:1431)
         at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1344)
         at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1362)
         at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:864)
         at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:851)
         at oracle.wh.ui.console.commands.DataViewerCmd.performAction(DataViewerCmd.java:19)
         at oracle.wh.ui.console.commands.TreeMenuHandler$1.run(TreeMenuHandler.java:188)
         at java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:178)
         at java.awt.EventQueue.dispatchEvent(EventQueue.java:454)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchThread.java:201)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:151)
         at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:145)
         at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:137)
         at java.awt.EventDispatchThread.run(EventDispatchThread.java:100)
    ENT-06972: Error while Displaying BIBean for Dimension Dataviewer.
         at oracle.wh.ui.owbcommon.dataviewer.dimensional.DimDataViewerMain.getDimensionListObject(DimDataViewerMain.java:479)
         at oracle.wh.ui.owbcommon.dataviewer.dimensional.DimDataViewerEditor._init(DimDataViewerEditor.java:89)
         at oracle.wh.ui.editor.Editor.init(Editor.java:1115)
         at oracle.wh.ui.editor.Editor.showEditor(Editor.java:1431)
         at oracle.wh.ui.owbcommon.IdeUtils._tryLaunchEditorByClass(IdeUtils.java:1431)
         at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1344)
         at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1362)
         at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:864)
         at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:851)
         at oracle.wh.ui.console.commands.DataViewerCmd.performAction(DataViewerCmd.java:19)
         at oracle.wh.ui.console.commands.TreeMenuHandler$1.run(TreeMenuHandler.java:188)
         at java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:178)
         at java.awt.EventQueue.dispatchEvent(EventQueue.java:454)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchThread.java:201)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:151)
         at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:145)
         at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:137)
         at java.awt.EventDispatchThread.run(EventDispatchThread.java:100)
    In another warehouse this function works properly. I am using OWB 10gR2 (10.2.0.1.31).
    Any hints to fix this issue would be appreciated.
    Thanks in Advance, C.

    If you run OWB using the supplied shell script or batch file are there other errors on the console before the exception. OWB uses a component (OLAPI does some validation over and above any other) which dumps the errors on the the standard output which often illustrates some invalid metadata in the OLAP catalog.

  • ENT-06954: Error while Displaying BIBean for Cube Dataviewer.

    Hello,
    I need urgent help.
    I build a ROLAP cube using OWB 11g. First it works, but after 2 days when I want to view data cube, using Data Viewer I receive this error. Please help me.
    ENT-06954: Error while Displaying BIBean for Cube Dataviewer.
    ENT-06954: Error while Displaying BIBean for Cube Dataviewer.
    at oracle.wh.ui.owbcommon.dataviewer.dimensional.CubeDataViewerMain.getDataviewObject(CubeDataViewerMain.java:391)
    at oracle.wh.ui.owbcommon.dataviewer.dimensional.CubeDataViewerEditor._init(CubeDataViewerEditor.java:66)
    at oracle.wh.ui.editor.Editor.init(Editor.java:1115)
    at oracle.wh.ui.editor.Editor.showEditor(Editor.java:1431)
    at oracle.wh.ui.owbcommon.IdeUtils._tryLaunchEditorByClass(IdeUtils.java:1436)
    at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1349)
    at oracle.wh.ui.owbcommon.IdeUtils._doLaunchEditor(IdeUtils.java:1367)
    at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:869)
    at oracle.wh.ui.owbcommon.IdeUtils.showDataViewer(IdeUtils.java:856)
    at oracle.wh.ui.console.commands.DataViewerCmd.performAction(DataViewerCmd.java:19)
    at oracle.wh.ui.console.commands.TreeMenuHandler$1.run(TreeMenuHandler.java:188)
    at java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:209)
    at java.awt.EventQueue.dispatchEvent(EventQueue.java:461)
    at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchThread.java:242)
    at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:163)
    at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:157)
    at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:149)
    at java.awt.EventDispatchThread.run(EventDispatchThread.java:110)

    Hi,
    Check this Metalink Note 455940.1
    Thanks,
    Sutirtha

  • Error while updating to target CUBE

    Hi,gurus here.
    I upload data from ODS into CUBE via DTP. The error happens:
    Error while updating to target ZICMM0200 (type INFOCUBE)
    Message no. RSBK241
    Package 3 / 2011.11.21 07:05:07 / Status 'Processed with Errors'
    Message no. RSBK257
    I've implemented the note 0001148007 and 0001159978,they don't work.
    Thanks,for your help.

    Hi,
    Please try this....
    Data is not updated to Infocube (BI 7.0) when i load from flat file
    RSBK257 - Status 'Processed with Errors' message in DTP

  • Error while checking RSRV of Cube

    hi all, I am getting following error while checking RSRV (DataBase indexses) of one of the cube.
    ORACLE: The status of index /BIC/FZPS_C02~020 is INVALID
    Please suggest something on it.
    amit

    Amit,
    Delete the cube indexes from Performance Tab of the Manage of the cube manually and check again.
    REDDY

  • Error while loading data into cube 0calday to 0fiscper (2lis_13_vdcon)

    Hi all,
    I m getting following error while loading the data into cube.
    "Time conversion from 0CALDAY to 0FISCPER (fiscal year V3 ) failed with value 10081031"
    amit shetye

    Hi Amit,
    This conversion problem. Calender not maintained for Fiscal variant "V3", for Year: 1008.
    Maintain calender for year: 1008 and transfer global setting from soruce(R/3).
    RSA1--> Source systems --> from context menu --> transfer global settings > choose fiscal year variants and calender> execute
    Hope it Helps
    Srini

  • Error while maintaing olap 11g cube

    Hi,
    Iam getting the below error while trying to maintain the cube
    An error has occurred on the server
    Error class: Express Failure
    Server error descriptions:
    INI: Error creating a definition manager, Generic at TxsOqConnection::generic<BuildProcess>
    INI: XOQ-01724: Can not specify a USING clause with a rewrite MV cube, Generic at TxsOqBuildProcessor::initializeForDimension
    at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
    at oracle.olapi.data.source.DataProvider.callGeneric(Unknown Source)
    at oracle.olapi.data.source.DataProvider.executeBuild(Unknown Source)
    at oracle.olap.awm.wizard.awbuild.UBuildWizardHelper$1.construct(Unknown Source)
    at oracle.olap.awm.ui.SwingWorker$2.run(Unknown Source)
    at java.lang.Thread.run(Unknown Source)
    ALso when I try to drop the cube/dimension , I get the following error
    An error has occurred on the server
    Error class: Express Failure
    Server error descriptions:
    DPR: Unable to create server cursor, Generic at TxsOqDefinitionManager::generic<CommitRoot>
    INI: XOQ-00703: Error executing OLAP DML command "(AW ATTACH CBU_UNIFIED.RECHARGES RW : ORA-34738: A severe problem has been detected. Analytic workspace operations have been disabled.
    )", Generic at TxsOqAWManager::executeCommand

    The two errors are, I suspect, unrelated.
    The XOQ-01724 error happens if you attempt to execute dbms_cube with a script that contains a USING clause. This is allowed for most cubes, but not for cubes associated with a materialized view. If this is what you tried to do, then the error is expected and the solution is to not specify a USING clause or to drop the MV.
    The second error is much more serious since it means your AW is probably corrupted. Did this happen after a failed parallel build? What version are you using.

  • Error while loading data to cube.

    Hi,
    I have a DTP which has failed few days back .... so since then no data request has been reportable/active.
    well this data package has failed after completing records update to the cube.  The error massage indicates problem is in material......
    I want to know, how this error can be rectified as data has been written to cube and (I feel ) while finding the SID request got failed.

    Hi
    I am putting the DTP progress as below.
    Data Package 1: Errors During Processing
    Extraction DataSource 2LIS_02_ITM : 2184 Data Records
    Prepare Error Handling : 2184 -> 2184 Data Records
    RSDS 2LIS_02_ITM RP1CLNT500 -> TRCS Z2LIS_02_ITM : 2184 -> 2184 Data Records
    TRCS Z2LIS_02_ITM -> CUBE ZPUR_C01 : 2184 -> 1524 Data Records
    Updating to InfoCube ZPP_FC_01 : 1524 -> 1523 Data Records
    Load and Generation (Where Necessary) of Write Program
    Conversion of Characteristic Values to SIDs
    Write to Fact Table
    InfoCube Update Completed
    Messages for data records saved; request is red acc. to configuration
    at 'Messages for data records saved; request is red acc. to configuration' I am getting error for material.

  • Dead lock error while updating data into cube

    We have a scenario of daily truncate and upload of data into cube and volumes arrive @ 2 million per day.We have Parallel process setting (psa and data targets in parallel) in infopackage setting to speed up the data load process.This entire process runs thru process chain.
    We are facing dead lock issue everyday.How to avoid this ?
    In general dead lock occurs because of degenerated indexes if the volumes are very high. so my question is does deletion of Indexes of the cube everyday along with 'deletion of data target content' process help to avoiding dead lock ?
    Also observed is updation of values into one infoobject is taking longer time approx 3 mins for each data packet.That infoobject is placed in dimension and defined it as line item as the volumes are very high for that specific object.
    so this is over all scenario !!
    two things :
    1) will deletion of indexes and recreation help to avoid dead lock ?
    2) any idea why the insertion into the infoobject is taking longer time (there is a direct read on sid table of that object while observed in sql statement).
    Regards.

    hello,
    1) will deletion of indexes and recreation help to avoid dead lock ?
    Ans:
    To avoid this problem, we need to drop the indexes of the cube before uploading the data.and rebuild the indexes...
    Also,
    just find out in SM12 which is the process which is causing lock.... Delete that.
    find out the process in SM66 which is running for a very long time.Stop  this process.
    Check the transaction SM50 for the number of processes available in the system. If they are not adequate, you have to increase them with the help of basis team
    2) any idea why the insertion into the infoobject is taking longer time (there is a direct read on sid table of that object while observed in sql statement).
    Ans:
    Lie item dimension is one of the ways to improve data load as well as query performance by eliminationg the need for dimensin table. So while loading/reading, one less table to deal with..
    Check in the transformation mapping of that chs, it any rouitne/formula  is written.If so, this can lead to more time for processing that IO.
    Storing mass data in InfoCubes at document level is generally not recommended because when data is loaded, a huge SID table is created for the document number line-item dimension.
    check if your IO is similar to doc no...
    Regards,
    Dhanya

  • Hyperion Planning migration error while refreshing/creating Essbase cube

    Hi,
    I'm migrating a Hyperion Planning application (11.1.1.4) to a complete other environment (due to the split of the companies). I have read most post on otn, but unfortunately nothing that helped me to solve the issue, yet... I hope someone can help me. Thanks in advance!
    I have executed the following steps:
    1. Create a new Oracle database
    2. Create a datasource to the Oracle database and to Essbase (both tested succesfully)
    3. Create a fresh planning application (I have used a native Admin account and not the original owner)
    4. Go into the database tables for the planning app and look at table HSP_USERS, store the SID value against user 50001
    5. Go into the database planning system tables and HSPSYS_APPLICATION, clear out the version field for the application you have created - I did not do this since I could not delete this value. null is not accepted.
    6. Stop Planning
    7. Restore the planning application database over the one created in the first step (I had to drop the existing tables first)
    8. Go into the HSP_USERS table and update the SID for user id 50001
    9. Start planning
    10. Log on to planning and I did see the webforms, etc as expected.
    11. Now, when I try to refresh the database, I receive an error:
    Cube refresh/create in progress
    Status failed
    Step 4 Adding dimensions
    java.lang.NoSuchMethodError: getAttributeType
    Does anyone has an idea? Hope someone can help me since the deadline is getting closer and closer.
    Thanks!
    Kind Regards
    Martin

    Hi John,
    Thanks for your reply.
    I have exactly done what you have described. I droped the schema and recreated it with an impdp to import.
    But, in the meantime I have been investigating the dimensions and there's indeed something very wrong. Some dimensions have been mixed up. In the 'old' environment I have 10 stored dimensions and 4 attribute dimensions (as well as in EAS as in Planning, offcourse as you expect). In the 'new' environment, two stored dimensions have been mixed up with two attribute dimensions.... How is that possible?
    I have never seen that. Could this be caused by the reason that we use a different version of Oracle (11G release 2 vs 10G (10.2.0.4))? Or went something wrong with exporting the schema...? Can you tell me what I can do to investigate this?
    Thanks!
    Kind Regards,
    Martin

  • Error while loading data into cube

    hi BW gurus,
    when ever i am trying to load data into the cube from flat file after scheduling iam getting short dump in BW system. I checked it in st22 it is giving me a error as exception add_partition_failed. please help me to sort out this problem. If you know the error recovery please give me the answer in detail.
    I will assign points for good answers.

    This is what the note says:
    Symptom
    The process of loading transaction data fails because a new partition cannot be added to the F fact table.The loading process terminates with a short dump.
    Other terms
    RSDU_TABLE_ADD_PARTITION_ORA, RSDU_TABLE_ADD_PARTITION_FAILED, TABART_INCONSITENCY, TSORA, TAORA , CATALOG
    Reason and Prerequisites
    The possible causes are:
    SQL errors when creating the partition
    Inconsistencies in the Data Dictionary control tables TAORA and TSORA
    Solution
    BW 3.0A & BW 3.0B
    In the case of SQL errors:Analyze the SQL code in the system log or short dump and if possible, eliminate the cause. The cause is often a disk space problem or lock situations on the database catalog or, less frequently: the partitioning option in the ORACLE database is not installed.
    The most common cause of the problem is inconsistencies in the TAORA and TSORA tables. As of Support Package 14 for BW 3.0B/Support Package 8 for BW 3.1C, the TABART_INCONSITENCY exception is issued in this case. The reason is almost always missing entries in TSORA for the tablespaces of the DDIM, DFACT and DODS data classes.
    The TAORA table contains the assignment of data classes to data tablespaces and their attributes, for example:
    Tablespace Data class
    DDIM PSAPDIMD ........
    DFACT PSAPFACTD ........
    DODS PSAPODSD .......
    Foreach data tablespace, the TSORA table must contain an entry for the corresponding index tablespace, for example:
    TABSPACE INDSPACE
    PSAPDIMD PSAPDIMD
    PSAPFACTD PSAPFACTD
    PSAPODSD PSAPODSD
    In most cases, these entries are missing and have to be added. See also notes 502989 and 46272.

  • Error while uploading data into cube

    I am trying to upload data into my content cube but I got an error it says"
    "Time conversion from 0CALDAY to 0FISCPER (fiscal year S1 ) failed with value 20040303"     
    I checked the data in the PSA it's there but the first record is not green light it has red light. Could you please give me some idea how to solve this problems.
    Thank you in advance
    sajita

    If you don't know if you want to take over all settings (especially exchange rates may be critical) the problem is probably found in the fiscal year variant. So if you just take over the fiscal year variants.
    If the problem remains you could check the following things:
    In SPRO -> Global Settings -> Fiscal Year Variants (or similar) check:
    Does a fiscal year variant S1 exist?
    Is it time dependent? If yes, is it valid for Mar 3rd 2004?
    If it is a self defined variant check if there is a period defined for March 3rd 2004.
    Best regards
       Dirk

  • Source system not Active error while installing BI content cube

    Hi.,
      When I try to install the BI content cube, In logs it is showing the error message as "Source client not Active". I activated the source system in modelling tab. Still the problem continues.  In source system assignment of BI content the source client is also checked. Please give me a solution.

    Hi Selva,
    If you want to select only infocube, select "Only necessary objects" in grouping. If you want everything related to Info cube, then system will collect data source also and for that it will check if you are authorized in source system. Please check if the source system is connected to BI or you can do one thing activate the data source in source system and replicate it in BI.
    Regards,
    Kams

  • Error while loading data from cube

    while ia m trying to load into the infocube i am unable to load data into the infocube. it is getting an error as raise_exception in the short dump analysis.  how to solve the problem if u know give me clear clarification.

    Hi,
    Can you check what is the difference in the PSA Data in the previous load and the latest load.
    May be by that way you can find what is wrong in any fields if any chance. It may be the case that it would be giving you that error because of some data mismatch in the fields. And in the Short Dump analyse it completely there would be definitely something which will be evident of the error.
    It may also very well be a data base table space issue as well.
    Hope this helps,
    Pradip Parmar

  • Process Chain error while creating index on cube

    Hi All,
    I have two process chain P1 daily(except sunday) and P2 weekly(sunday), both have same process, i.e. cube C1 is loaded with same infopackage in both infopackage.
    Now, on saturday after laoding the content index creating process on cube c1 is failed by process chain P1, but on sunday P2 have laoded C1 correctly and created indexes.
    Now, my doubt is wheather i repeat process chain P1 from creating indexes or there is no need as indexes on cube are created by PC 2.
    kindly suggest,
    Regards,
    Rajvinder

    Hi Raj,
    There is no need of running the process chain P1, the data has been bought by the process chain P2.
    If there are any filters/data selection in infopackage level then you need to run the process chain P1.
    Be make sure about the process chains and there infopackages/dtp's.
    With Regards,
    Ravi Kanth

Maybe you are looking for