OWB R2 / Paris: importing source metadata (tables) results in LONG error

Hello,
Ik hope anyone knows a solution to this, as I can't find anything on OTN, Metalink and google...
When I create a new module to function as 'source' for my datawarehouse I try to import the table definitions from this schema using the import wizard. The wizard lists the correct tables, views etc. But when I select a table to import as metadata I get the following error: Repository Error Message: ORA-00997: illegal use of LONG datatype.
Odd enough; functions do import.
I've tried recreating the metadata repository and different source schema's to no avail. Even re-importing a target table doesn't work. Besides that, everything works just fine but there are too many source tables to manually insert.
Please help!
Regards.
The complete error is :
SQL Exception
SQL Exception
Repository Error:SQL Exception..
Class Name: CacheMediator.
Method Name: getDDEntryFromDB.
Repository Error Message: ORA-00997: illegal use of LONG datatype
     at oracle.wh.repos.pdl.foundation.CacheMediator.getSQLResult(CacheMediator.java:504)
     at oracle.wh.repos.impl.CMPWBAPIFactory.getSQLResult(CMPWBAPIFactory.java:2137)
     at oracle.wh.service.sdk.integrator.RepositoryUtils.getSQLResult(RepositoryUtils.java:448)
     at oracle.wh.ui.integrator.sdk.RelationalEntityAccessor.getTableColumnsSQLResult(RelationalEntityAccessor.java:2541)
     at oracle.wh.ui.integrator.sdk.RelationalEntityAccessor.getItems(RelationalEntityAccessor.java:680)
     at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.getColumnsToImport(ImportEntityAlgorithm.java:2925)
     at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importColumns(ImportEntityAlgorithm.java:2914)
     at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importTable(ImportEntityAlgorithm.java:1343)
     at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.dispatchElement(ImportEntityAlgorithm.java:561)
     at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importElement(ImportEntityAlgorithm.java:366)
     at oracle.wh.ui.integrator.sdk.EntityAccessor.importElement(EntityAccessor.java:80)
     at oracle.wh.ui.integrator.common.ImportService.importElement(ImportService.java:1046)
     at oracle.wh.ui.integrator.common.wizards.ImportElementTransaction.run(ImportWizardDefinition.java:578)

Diagnosis:
        - Check the SQL statement being issued.
        - Note that LONGs cannot be used in a WHERE clause, in
          INSERT into ... SELECT ... FROM constructs, in snapshots
          etc..
        - For 'INSERT into ... SELECT ... FROM constructs' use the
          SQL*Plus "COPY" command.
Articles:
Bugs: (Fix version in brackets)
        Grant on local view of remote table with LONG (7.1.3)  Bug 196300Regards,
Marcos

Similar Messages

  • BODS 4.2 Cannot import the metadata table, RFC_ABAP_INSTALL_AND_RUN syntax error

    Hi all, we installed BODS 4.2 server to substitute a 4.1, but we are facing the error:
    Error: Cannot import the metadata table <name=T001>
    RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE -(Exception Key: Syntax error in program /BODS/SAPLBODS....
    We already tried the solution for when people get the error related to unicode.
    Also, we are able to pull data via extractors, it only fails when loading Tables....
    Any help is greatly appreciated.

    Dear,
    You will have to import the new ABAP Function Group "BODS/BODS".
    Here are some details:
    Installing Functions on the SAP Server
    SAP BusinessObjects Data Services provides functions that support the use of the ABAP, BAPI, and
    IDoc interfaces on SAP servers. You will need some or all of these functions in the normal operation
    of the software in an SAP environment. These functions perform such operations as dynamically loading
    and executing ABAP programs from Data Services, efficiently running preloaded ABAP programs,
    allowing for seamless metadata browsing and importing from SAP servers, and reporting the status of
    running jobs. Some of these functions read data from SAP NetWeaver BW sources.
    You must upload the provided functions to your SAP server in a production environment. It is
    recommended that you always upload the functions to your SAP server whether you are in a
    development, test, or production environment. The functions provide seamless integration between
    Data Services and SAP servers.
    The default installation places two function module files for SAP servers in the ...\Data
    Services\Admin\R3_Functions\transport directory. You then upload these files to SAP servers
    using the SAP Correction and Transport System (CTS) or manually. Using CTS allows for version
    control as the functions evolve across releases.
    The installation provides two versions of transport files (depending on the server version you are using)
    to install the functions on the SAP server. To obtain the names of the latest transport files for installing
    or upgrading these SAP server functions, see the readme.txt file
    And I've found those files and text files in the local install folder....in:
    Program Files\SAP BusinessObjects\Data Services\admin\R3_Functions
    (that's where I've installed it).
    There you'll find some descriptive txt as how to proceed.
    After installing, it might happen that the executing user is missing some authorizations.
    Here my authorizations team helped me by tracing the user and then adding the necessary rights.
    Sure hope this will help you.

  • Is this a bug in OWB 11.2 - importing table metadata for character columns

    The Oracle® Warehouse Builder Data Modeling, ETL, and Data Quality Guide provides an overview of the data types supported.
    http://docs.oracle.com/cd/E11882_01/owb.112/e10935/orcl_data_objx.htm
    It says that for VARCHAR2 data type it saws (http://docs.oracle.com/cd/E11882_01/owb.112/e10935/orcl_data_objx.htm#CHDFIADI )
    "Stores variable-length character data. How the data is represented internally depends on the database character set. The VARCHAR2 data type takes a required parameter that specifies a maximum size up to 4,000 characters"
    That means , I guess, it says that when I import a table, any columns of type VARCHAR2(10) in the database should have its length show as characters in OWB, so a column of type Varchar2(10) in the Oracle database, should be shown as Varchar2(10) when imported into OWB table metadata via the OWB import function.
    However, if I have a database that set-up as a single-byte and import a table using the OWB import function a column that has a size of e.g. 10 in the database, is imported as OWB table metadata and the size is 10. Correct, I am happy.
    However, if the database is modified to support multi-byte characters, ALTUF16 encoding with the semantics set to "CHAR", then when I import the same table into OWB, OWB reports the size as 40, I guess its 40 bytes as in 10 characters @ 4 bytes per character.
    Is this a bug in OWB, as the datatype in the Oracle DB is varchar2(10), should OWB after importing a table not also report the column as VARCHAR2(10) ? Currently, is shows the column as varchar2(40).

    I noticed that myself in our project.
    Our varchars2 are defined as VARCHAR2(xxx CHAR) - OWB puts the size*4
    In fact if you have special characters like umlauts (ü,ä,ö,...) it will use 4 bytes per character.
    You can try it yourself. Define a Varchar2(1 CHAR) and manually change the size of the Column in your mapping inside OWB (in filters, joins or your target table).
    Then shoot an umlaut through the mapping and will end up with a "too small" error.
    Dont mind the size*4 issue - we totally ignored it and run without error since 4 years now.

  • Source file structure when importing a qualified table

    Hello MDM:ers
    I need to import a qualified table into the following repository
    (it is simplified to illustrate the problem):
    Main table: Products
    Fields: ID (int), Name (Text), Part number (Qualified)
    Qualified Table: Part Numbers
    Fields: Manufacturer (Lookup), Part Number (Text)
    Lookup table: Manufacturers
    Fields: Name (Text)
    The source file is a very large text file that has the following structure:
    Part ID          Name     Part no. Manuf 1     Part no. Manuf2
    123          Nut     ABC               XYZ
    456          Bolt     DEF               QWV
    The result I’m looking for Is to map a file structured a above so that two main table records are created (Nut and Bolt) which both have two qualified table records for each manufacturer. I have not been able to create such a map. I have tried to partitioning and combining fields without success.
    I have no problem creating the map if I manually restructure the file to look like:
    Part ID          Name     Manufacturer           Part Number
    123          Nut     Manufacturer 1          ABC
    123          Nut     Manufacturer 2          XYZ
    456          Bolt     Manufacturer 1          DEF
    456          Bolt     Manufacturer 2          QWV
    It where a way to import the file directly in MDM without restructuring it?

    Hi Alexander,
    MDIS will not work fine if you have multiple tables as the mapping is saved only for 1 table. If you have multiple tables then after performing the process of Pivoting/Reverse pivoting, you will have to use joins.
    Join the other tables as Lookup in the Main Table
    Save all your configuration in a map. And then configure the map in the port.
    It should then work with the MDIS also.
    hope it helps.
    Thanks and Regards
    Niitn Jain

  • Cannot retrieve table metadata - Table does not exist: ODP source 0WRKCNT_CATG_TEXT does not exist

    Hi, when i able to import the 0WRKCNT_CATG_TEXT extractor into source system i am getting the above bug Cannot retrieve table metadata - Table does not exist: ODP source <0WRKCNT_CATG_TEXT> does not exist, i have been checked in RSA5 T code to check the object is active or not, its active and its available in ROOSATTR table with enabled mode,but still its showing the error, can anyone help on this ..

    Hi Airings,
    'ORA-00942: table or view does not exist'
    According to the error message, it seems that the migrating table or view does not exist in the database, or SSMA does not have access to it. To troubleshoot the issue, please check the following things.
     1. Verify that if the spelling of the table or view name is correct.
     2. If the table or view exists but is in a different schema from the current schema where the SQL is executing (in other word, the table doesn’t own by you, but owned by other user), the ORA-00942 error will return too. Resolve this by
    explicitly reference the table or view by specifying the schema name (schema_name.table_name).
    3. SSMA queries some additional catalog tables that you may not have permission to, please make sure that you grant the account permission to
     read sys.mlog$. For more details, please review this similar thread:
    Bug in SSMA For Oracle 6.0 for non-dba Oracle user.
    Reference:
    ORA-00942 Table or View Does Not Exist Oracle Error
    Thanks,
    Lydia Zhang
    Lydia Zhang
    TechNet Community Support

  • OMB plus Creating Locations Modules And Importing the Metadata?

    Hi All,
    First of all I would like to thank you all of you in advance for all your time and concern. I have serious of questions and I will ask them in 3 different portions as they are slightly different.
    How can I implement the steps below using OMB plus? If you could provided sample script and notation it will be highly appreciated.
    -- Creating an Oracle Location for a Schema and Registering it
    -- Creating an Module and relating it to the location that was created.
    -- Importing the metadata for all the objects in this Schema in to this module.
    For example,
    You have a schema called ORACLE_SRC. I need to create the location and module for this schema and tie them each other. Then I need to import all the metadata for all the objects in this ORACLE_SRC schema. During this import all the preserve statements should be unchecked. And I need to accomplish all these using OMB plus and TCL scripting.
    -- Creating an Non Oracle (SQL server 2005) Location for a database and Registering it
    -- Creating an Module and relating it to the location that was created.
    -- Importing the metadata for all the objects in this Schema in to this module.
    For example,
    You have a schema called SQLSERVER_SRC. I need to create the location and module for this schema and tie them each other. Then I need to import all the metadata for all the objects in this SQLSERVER_SRC schema. During this import all the preserve statements should be unchecked. And I need to accomplish all these using OMB plus and TCL scripting.
    -- Creating an Flat File Location for a folder structure and Registering it
    -- Creating an Module and relating it to the location that was created.
    For example,
    You have a schema called FLATFILE_SRC. I need to create the location and module for this schema and tie them each other.
    Thanks again for all your help and concern!
    Kind Regards,
    Mike

    In addition to the performance issue above I have faced an extremely more important problem. When I have used the same logic for importing Materialized views I am getting an error at the last step of the script. It is exactly the same case also for EXTERNAL_TABLE objects. TABLE and VIEW objects are fine. OMBIMPORT command throw the error message below.
    PUB04201: Encountered an internal fatal error during processing of oracle module XXX_DMY_DWPROD_TST: unsupported obj type.
    I would like to mention that MATERIALIZED_VIEW is the object name I took from OWB scripting Reference guide. I have tried couple other naming conventions too but it didn't work.
    Strange thing is If I were to change MATERIALIZED_VIEW to TABLE it works but the materialized view gets imported into 'Tables' folder under the module.
    Please help?
         #IMPORT FOR MATERIALIZED VIEWS STARTS
    #pull the names of the object inthe schema
         set MVIEWLIST {}
         set sqlStr "select distinct Mview_Name from all_mviews where owner = upper('$DbNm') order by Mview_Name"
         #execute the query
         set oraRs [oracleQuery $oraConn $sqlStr]
         #Loop through the list of objects
         while {[$oraRs next]} {
              set mvwName [$oraRs getString Mview_Name]
              lappend MVIEWLIST $mvwName
         #and close the query
         $oraRs close
         set listCntr 0
         set listTotal [llength $MVIEWLIST]
         set mviewName [lindex $MVIEWLIST $listCntr]
         #First object
         OMBALTER IMPORT_ACTION_PLAN 'IMPORT_ALL' ADD ACTION 'MAT_VIEWS' SET REF SOURCE MATERIALIZED_VIEW '$mviewName' SET REF TARGET ORACLE_MODULE '$ModNm'
         OMBCOMMIT
         puts "$mviewName is being imported"
         incr listCntr
         while {$listCntr<$listTotal} {     
              #Rest of the objects
              set mviewName [lindex $MVIEWLIST $listCntr]
              incr listCntr
              OMBALTER IMPORT_ACTION_PLAN 'IMPORT_ALL' MODIFY ACTION 'MAT_VIEWS' SET REF SOURCE MATERIALIZED_VIEW '$mviewName' SET REF TARGET ORACLE_MODULE '$ModNm'
              OMBCOMMIT
              puts "$mviewName is being imported"
         set listCntr 0
         set listTotal 0     
         #IMPORT FOR MATERIALIZED VIEWS ENDS
              OMBIMPORT FROM METADATA_LOCATION FOR IMPORT_ACTION_PLAN 'IMPORT_ALL'
              OMBDROP IMPORT_ACTION_PLAN 'IMPORT_ALL'

  • Issue in OWB mapping - when changing source and target database

    Hi,
    I need help for resolution of the issue I am facing when moving mapping from development environment to QA.
    Here is situation,
    We develop ETL using one source, one staging and one target database.
    In development we use one control_center for Source to staging and another control center staging to target.
    All works fine in development.
    Now I have created new runtime repository and imported all OWB projects (with full dependencies, exact replica of development). Now I need to change source and staging and target as different database.
    I have created new database location connections and defined/attached DB connectors with stage and target location.
    Now issues are
    1.     Two Staging mapping are not able to bound with source table (giving different error
    a.     One mapping show error for source synonym translation no longer valid when deploying this mapping , but validation comes without any issue)
    b.     Other mapping show error for source table/object not bound to repository
    2.     All the target mapping show validation successful, but when deploying says “table or view does not exist”. But tables are exists on source stage and target. (also permission are set correctly for target user to read from staging tables).
    Not sure how to proceed from here.
    I have recreated new repository and re-imported all project/mappings and defined all connection but still same issue.
    Thanks in advance,
    Vipin

    1. Two Staging mapping are not able to bound with source table (giving different error
    a. One mapping show error for source synonym translation no longer valid when deploying this mapping , but validation comes without any issue)
    b. Other mapping show error for source table/object not bound to repository
    The above error were resolved when re-synchronized the table (for few I have to reimport the table) and mapping.
    2. All the target mapping show validation successful, but when deploying says “table or view does not exist”. But tables are exists on source stage and target. (also permission are set correctly for target user to read from staging tables).
    The above error still pending. My target mapping are not able to deployed/compiled.
    For the above I have defined one Staging location to one target target location and target location have connector to staging (not sure if I have to define connector name same as staging location, as I have created DB connector with different name but reference database is same as staging location).
    Mapping are assoicated with desired data locatoin and meta data.
    control center is also have that data location.
    Mapping are configured for the desired location.

  • OWB object creation/import error

    I installed OWB for evaluation on Windows XP.
    OWB version is 10.1.0.4 and the database is 10g release 2.
    While trying to create a target or trying to import source table, owb fails with
    a "NULL" error. The detailed error is given below, pls help.
    null
    null
         at oracle.wh.repos.impl.foundation.DirtyCache.persist(DirtyCache.java:442)
         at oracle.wh.repos.impl.foundation.DirtyCache.persist(DirtyCache.java:355)
         at oracle.wh.repos.impl.foundation.DirtyCache.persist(DirtyCache.java:233)
         at oracle.wh.repos.impl.foundation.CacheMediator.postChanges(CacheMediator.java:458)
         at oracle.wh.repos.impl.foundation.CacheMediator.postChanges(CacheMediator.java:451)
         at oracle.wh.repos.impl.foundation.NestedTransactionManager.removeLatestSavepoint(NestedTransactionManager.java:81)
         at oracle.wh.repos.impl.transaction.TransactionManager.commitNestedTransaction(TransactionManager.java:392)
         at oracle.wh.repos.impl.transaction.TransactionManager.endModalInteraction(TransactionManager.java:483)
         at oracle.wh.ui.integrator.common.ImportElementTransaction.run(ImportWizard.java:1058)

    963510 wrote:
    What am I doing wrong??Posting this in the APEX forum instead of the {forum:id=76} one?
    You also appear to be trying to reinvent the wheel, as all of these types are already defined in Oracle Spatial.

  • Import source from server machine

    hi,
    i want to import source data into my localhost system.
    source data is residing in server system in BD2 database;
    i want to design mapping using these source.
    so i want to know how we can import data from server system data
    and how we can use these data in mapping
    plz tell me procedure
    Edited by: user10373357 on Nov 3, 2008 10:49 PM

    LEARN OWB !
    1. I havent heard of BD2 i only know DB2(if thats what you are referring to).
    2. Create a repository under NON-ORacle ==> DB2 in design center.
    3. create setup locations in connection explorer(under non-oracle).
    4. Import the metadata.
    5. create mappings.
    6. create a control centre location
    7. deploy mappings
    8. execute your mappings(if required use schedules and process flows)..

  • Error while importing a DW table into DAC

    Hi,
    We are on OBIEE 7.9.6 and we have a requirement of adding new DW table. I created a new table in DW and getting error "Could connect to data source process. Process failed during creation of connection pool", while trying to import this new table into DAC. I would like to know why I am getting this error.
    Any help is greatly appreciated.
    Thanks,
    Chandra

    Thank you for your response.
    I added this table in DW database and have imported into informatica. When I look at the traffic light it is red. Looks like thats the problem.
    Thanks again.

  • Best practice of metadata table in data warehouse environment ?

    Hi guru's,
    In datawarehouse, we have 1. Stage schema 2. DWH(Data warehouse reporting schema). In stageing we have about 300 source tables. In DWH schema, we are creating the tables which are only required from reporting prespective . some of the tables in stageing schema, have been created in DWH schema as well with different table name and column names. The naming convention for these same tables and columns in DWH schema is more based on business names.
    In order to keep track of these tables we are creating metadata table in DWH schema say for example
    Stage                DWH_schema
    Table_1             Table_A         
    Table_2             Table_b
    Table_3             Table_c
    Table_4              Table_DMy question is how do we handle the column names in each of these tables. The stage_1, stage_2 and stage_3 column names have been renamed in DWH_schema which are part of Table_A, Table_B, Table_c.
    As said earlier, we have about 300 tables in stage and may be around 200 tables in DWH schema. Lot of the column names have been renamed in DWH schema from stage tables. In some of the tables we have 200 column's
    so my concern is how do we handle the column names in metadata table ? Do we need to keep only table names in metadata table not column names ?
    Any idea will be greatly appriciated.
    Thanks!

    hi
    seems quite a buzzing question.
    In our project we designed a hub and spoke like architecture.
    Thus we have 3 layer, L0 is the one closest to the source and L0 table's name are linked to the corresponding sources names by mean of naming standard (like tabA EXT_tabA tabA_OK1 so on based on implementation of load procedures).
    At L1 we have the ODS , normalized model , we use business names for table there and standard names for temporary structures and artifacts
    Both L0 an L1 keep source's column names as general rule, new columns like calculated one are business driven and metadata are standard driven.
    Datamodeler fits perfect for modelling L1 purpose.
    L2 is the dimensional schema business names take place for tables and columns eventually rewritten at presentation layer ( front end tool )
    hope this helps D.

  • Import OLAP Metadata in OBIEE 11.1.1.5 failing

    We are in the process of integrating our existing Oracle OLAP infrastructure with OBIEE 11.1.1.5. We are currently on 11.2.0.1 DB with our 11g cubes in 11.2.0.0.0 compatibility mode.
    We have two AWs within the same schema that are tightly integrated (Eg. Q12_AW and Q12_HIST_AW). When I import the metadata using OBIEE 11g BI Administrator, I can only see Q12_HIST_AW but not Q12_AW. Also, when I copy Q12_HIST_AW to OBIEE, it errors out. Looking at java host logs, it looks like I am getting parse errors but not sure how to pursue further on why the parser is failing.
    Logs:
    [2011-08-16T10:34:45.168-05:00] [init] [WARNING] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEjD6FS8O6yjMaMG1EIco2000006,0] Bad AW -- Q12_AW.Q12_AW
    [2011-08-16T10:34:45.175-05:00] [init] [WARNING] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEjD6FS8O6yjMaMG1EIco2000006,0] Errors have occurred during xml parse[[
    <Line 1181, Column 26>: Encountered "(" at line 1, column 83.
    Was expecting one of:
    "DIMENSION" ...
    <Line 1210, Column 26>: Encountered "(" at line 1, column 83.
    Was expecting one of:
    "DIMENSION" ...
    <Line 1529, Column 26>: Encountered "(" at line 1, column 58.
    Was expecting one of:
    "DIMENSION" ...
    <Line 1558, Column 26>: Encountered "(" at line 1, column 58.
    Was expecting one of:
    "DIMENSION" ...
    <Line 3025, Column 23>: Encountered "(" at line 1, column 54.
    Was expecting one of:
    "DIMENSION" ...
    <Line 4020, Column 24>: Encountered "(" at line 1, column 81.
    Was expecting one of:
    "DIMENSION" ...
    <Line 9516, Column 24>: Encountered "(" at line 1, column 101.
    Was expecting one of:
    "DIMENSION" ...
         at oracle.olapi.xml.TagHandler.createRootException(Unknown Source)
         at oracle.olapi.xml.TagHandler.getRootException(Unknown Source)
         at oracle.olapi.xml.TagHandler.reportException(Unknown Source)
         at oracle.olapi.xml.TagHandler.processException(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataXMLReader.resolveDeferredProperties(Unknown Source)
         at oracle.olapi.metadata.MetadataXMLReaderMetadataInitialState.exit(Unknown Source)
         at oracle.olapi.metadata.MetadataXMLReaderMetadataInitialState.exit(Unknown Source)
         at oracle.olapi.xml.TagHandler.endElement(Unknown Source)
         at org.xml.sax.helpers.ParserAdapter.endElement(ParserAdapter.java:626)
         at oracle.xml.parser.v2.XMLContentHandler.endElement(XMLContentHandler.java:211)
         at oracle.xml.parser.v2.NonValidatingParser.parseElement(NonValidatingParser.java:1359)
         at oracle.xml.parser.v2.NonValidatingParser.parseRootElement(NonValidatingParser.java:376)
         at oracle.xml.parser.v2.NonValidatingParser.parseDocument(NonValidatingParser.java:322)
         at oracle.xml.parser.v2.XMLParser.parse(XMLParser.java:226)
         at org.xml.sax.helpers.ParserAdapter.parse(ParserAdapter.java:405)
         at oracle.olapi.xml.XMLProcessor.parse(Unknown Source)
         at oracle.olapi.metadata.MetadataFetcher.processXML(Unknown Source)
         at oracle.olapi.metadata.MetadataFetcher.fetchBaseMetadataObjects(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataProvider.fetchMetadataObjects(Unknown Source)
         at oracle.olapi.metadata.MetadataListProperty.getObjects(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataObjectState.getPropertyListValues(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataObject.getPropertyListValues(Unknown Source)
         at oracle.olapi.metadata.mdm.MdmSchema.getCubes(Unknown Source)
         at oracle.olapi.metadata.deployment.AW.getCubes(Unknown Source)
         at oracle.bi.integration.aw.v11g.AW11gUtil.getAWImportInfo(AW11gUtil.java:1035)
         at oracle.bi.integration.aw.v11g.AW11gUtil.getAWImportInfo(AW11gUtil.java:1113)
         at oracle.bi.integration.aw.v11g.service.AW11gService.execute(AW11gService.java:83)
         at oracle.bi.integration.javahost.ServiceRpcCall.processMessageInternal(ServiceRpcCall.java:55)
         at com.siebel.analytics.javahost.AbstractRpcCall.processMessage(AbstractRpcCall.java:251)
         at com.siebel.analytics.javahost.MessageProcessorImpl.processMessage(MessageProcessorImpl.java:193)
         at com.siebel.analytics.javahost.Listener$Job.run(Listener.java:223)
         at com.siebel.analytics.javahost.standalone.SAJobManagerImpl.threadMain(SAJobManagerImpl.java:207)
         at com.siebel.analytics.javahost.standalone.SAJobManagerImpl$1.run(SAJobManagerImpl.java:155)
         at java.lang.Thread.run(Thread.java:662)
    [2011-08-16T10:34:46.359-05:00] [init] [NOTIFICATION] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEjD6FS8O6yjMaMG1EIco2000006,0] Reading AW -- Q12_AW.Q12_HIST_AW
    [2011-08-16T10:34:46.419-05:00] [init] [NOTIFICATION] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEjD6FS8O6yjMaMG1EIco2000006,0] [Thread 21] Service done -- AWImportService11G
    [2011-08-16T10:34:50.149-05:00] [workmanager] [NOTIFICATION] [] [saw.workmanager] [tid: 15] [ecid: 0000J7JElO_FS8O6yjMaMG1EIco2000007,0] Thread started
    [2011-08-16T10:35:22.340-05:00] [init] [NOTIFICATION] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEtF^FS8O6yjMaMG1EIco200000C,0] [Thread 21] calling service -- AWImportService11G
    [2011-08-16T10:35:22.340-05:00] [init] [NOTIFICATION] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEtF^FS8O6yjMaMG1EIco200000C,0] Reading AW UDML -- Q12_HIST_AW
    [2011-08-16T10:35:25.768-05:00] [init] [ERROR] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEtF^FS8O6yjMaMG1EIco200000C,0] Errors have occurred during xml parse[[
    <Line 1181, Column 26>: Encountered "(" at line 1, column 83.
    Was expecting one of:
    "DIMENSION" ...
    <Line 1210, Column 26>: Encountered "(" at line 1, column 83.
    Was expecting one of:
    "DIMENSION" ...
    <Line 9516, Column 24>: Encountered "(" at line 1, column 101.
    Was expecting one of:
    "DIMENSION" ...
         at oracle.olapi.xml.TagHandler.createRootException(Unknown Source)
         at oracle.olapi.xml.TagHandler.getRootException(Unknown Source)
         at oracle.olapi.xml.TagHandler.reportException(Unknown Source)
         at oracle.olapi.xml.TagHandler.processException(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataXMLReader.resolveDeferredProperties(Unknown Source)
         at oracle.olapi.metadata.MetadataXMLReaderMetadataInitialState.exit(Unknown Source)
         at oracle.olapi.metadata.MetadataXMLReaderMetadataInitialState.exit(Unknown Source)
         at oracle.olapi.xml.TagHandler.endElement(Unknown Source)
         at org.xml.sax.helpers.ParserAdapter.endElement(ParserAdapter.java:626)
         at oracle.xml.parser.v2.XMLContentHandler.endElement(XMLContentHandler.java:211)
         at oracle.xml.parser.v2.NonValidatingParser.parseElement(NonValidatingParser.java:1359)
         at oracle.xml.parser.v2.NonValidatingParser.parseRootElement(NonValidatingParser.java:376)
         at oracle.xml.parser.v2.NonValidatingParser.parseDocument(NonValidatingParser.java:322)
         at oracle.xml.parser.v2.XMLParser.parse(XMLParser.java:226)
         at org.xml.sax.helpers.ParserAdapter.parse(ParserAdapter.java:405)
         at oracle.olapi.xml.XMLProcessor.parse(Unknown Source)
         at oracle.olapi.metadata.MetadataFetcher.processXML(Unknown Source)
         at oracle.olapi.metadata.MetadataFetcher.fetchBaseMetadataObjects(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataProvider.fetchMetadataObjects(Unknown Source)
         at oracle.olapi.metadata.MetadataListProperty.getObjects(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataObjectState.getPropertyListValues(Unknown Source)
         at oracle.olapi.metadata.BaseMetadataObject.getPropertyListValues(Unknown Source)
         at oracle.olapi.metadata.mdm.MdmSchema.getCubes(Unknown Source)
         at oracle.olapi.metadata.deployment.AW.getCubes(Unknown Source)
         at oracle.bi.integration.aw.v11g.AW11gUtil.getAWUdml(AW11gUtil.java:914)
         at oracle.bi.integration.aw.v11g.AW11gUtil.getAWUdml(AW11gUtil.java:876)
         at oracle.bi.integration.aw.v11g.service.AW11gService.getAWUdmlObject(AW11gService.java:157)
         at oracle.bi.integration.aw.v11g.service.AW11gService.getAWUdml(AW11gService.java:137)
         at oracle.bi.integration.aw.v11g.service.AW11gService.execute(AW11gService.java:78)
         at oracle.bi.integration.javahost.ServiceRpcCall.processMessageInternal(ServiceRpcCall.java:55)
         at com.siebel.analytics.javahost.AbstractRpcCall.processMessage(AbstractRpcCall.java:251)
         at com.siebel.analytics.javahost.MessageProcessorImpl.processMessage(MessageProcessorImpl.java:193)
         at com.siebel.analytics.javahost.Listener$Job.run(Listener.java:223)
         at com.siebel.analytics.javahost.standalone.SAJobManagerImpl.threadMain(SAJobManagerImpl.java:207)
         at com.siebel.analytics.javahost.standalone.SAJobManagerImpl$1.run(SAJobManagerImpl.java:155)
         at java.lang.Thread.run(Thread.java:662)
    [2011-08-16T10:35:25.784-05:00] [init] [ERROR] [] [saw.init.application] [tid: 14] [ecid: 0000J7JEtF^FS8O6yjMaMG1EIco200000C,0] [Thread 21] Service failed - AWImportService11G. Details:Q12_HIST_AW
    Any help to diagnose the problem is appreciated.
    Swapan.
    Edited by: Swapan on Aug 16, 2011 9:28 AM

    It looks like OBIEE 11.1.1.5 ships with 11.1 jars and since my DB was running 11.2, I faced this issue. The fix is described below for folks who encounter this issue.
    The OLAP API jars on the middle tier need to be updated to version 11.2.x. The OLAP API libraries are found in your Oracle Database home: [oracledb home]\product\11.2.0\dbhome_1\olap\api\lib. BI EE provides an 11.1.x version of these files in [obiee home]\Oracle_BI1\bifoundation\javahost\lib\obisintegration\aw\11g. Backup the BI EE version of the OLAP API jars and replace them with the version provided by the database.
    Swapan.
    Edited by: Swapan on Aug 16, 2011 1:33 PM

  • Error when Importing a External table

    When I am trying to import a external table I am getting a error as SQL Exception. Any ideas as to how to fix this error.
    The external table is under one schema. I have a locator in OWB pointing to that schema and I am trying to Import the external table. I am able to see this table in TOAD.
    Thanks

    Repository Error:SQL Exception..
    Class Name: CacheMediator.
    Method Name: getDDEntryFromDB.
    Repository Error Message: ORA-22992: cannot use LOB locators selected from remote tables
         at oracle.wh.repos.pdl.foundation.CacheMediator.getSQLResult(CacheMediator.java:504)
         at oracle.wh.repos.impl.CMPWBAPIFactory.getSQLResult(CMPWBAPIFactory.java:2137)
         at oracle.wh.service.sdk.integrator.RepositoryUtils.getSQLResult(RepositoryUtils.java:448)
         at oracle.wh.service.sdk.integrator.RepositoryUtils.getSQLResult(RepositoryUtils.java:413)
         at oracle.wh.ui.integrator.oracle.WBOracleDatabaseEntityAccessor.getAccessParametersSQLResult(WBOracleDatabaseEntityAccessor.java:518)
         at oracle.wh.ui.integrator.oracle.WBOracleDatabaseEntityAccessor.getAccessParameters(WBOracleDatabaseEntityAccessor.java:459)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importAccessParameters(ImportEntityAlgorithm.java:2724)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importExternalTable(ImportEntityAlgorithm.java:1499)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.dispatchElement(ImportEntityAlgorithm.java:583)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importElement(ImportEntityAlgorithm.java:366)
         at oracle.wh.ui.integrator.sdk.EntityAccessor.importElement(EntityAccessor.java:80)
         at oracle.wh.ui.integrator.common.ImportService.importElement(ImportService.java:1046)
         at oracle.wh.ui.integrator.common.wizards.ImportElementTransaction.run(ImportWizardDefinition.java:578)

  • OIM and OIA11g Integration(error while importing Resource metadata into OIA

    Hi,
    I followed the below link and trying to integrate OIM 11.1.1.5 and OIA 11.1.1.3.6
    And both are deployed different weblogic domains..
    http://wikis.sun.com/display/OIA11gDocs/Integrating+With+Oracle+Identity+Manager,+Preferred+Method
    I am getting error in Step 6: Import the Oracle Identity Manager (OIM) Data Into Oracle Identity Analytics (OIA) -> To Import Resource Metadata
    10:56:21,484 DEBUG [DBIAMSolution] publishing import starting event...
    10:56:21,623 DEBUG [OimUtilityFactory] ************** OIM Connection Params *************
    10:56:21,628 DEBUG [OimUtilityFactory] XL Home ---> C:\Oracle\Middleware\Oracle_IDM1
    10:56:21,630 DEBUG [OimUtilityFactory] login config ---> C:\Oracle\Middleware\Oracle_IDM1\server\config\authwl.conf
    10:56:21,631 DEBUG [OimUtilityFactory] ****************************************************
    10:56:21,636 DEBUG [OimUtilityFactory] Xellerate Discovery Settings {}
    10:56:21,638 DEBUG [OimUtilityFactory] ********** Connecting to OIM Server **********
    10:56:21,774 ERROR [IamDbNamespaceImporterHelperImpl] Error connecting to OIM
    Thor.API.Exceptions.tcAPIException: javax.security.auth.login.LoginException: java.lang.SecurityException: [Security:090304]Authentication Failed: User xelsysadm javax.security.auth.login.FailedLoginException: [Security:090302]Authentication Failed: User xelsysadm denied
         at Thor.API.tcUtilityFactory.<init>(tcUtilityFactory.java:166)
         at com.vaau.rbacx.iam.util.oracle.oimapi.OimUtilityFactory.getUtilityFactory(OimUtilityFactory.java:67)
         at com.vaau.rbacx.iam.db.helpers.IamDbNamespaceImporterHelperImpl.readNamespaces(IamDbNamespaceImporterHelperImpl.java:87)
         at com.vaau.rbacx.iam.db.DBIAMSolution.readResourceMetadata(DBIAMSolution.java:642)
         at com.vaau.rbacx.iam.service.impl.RbacxIAMServiceImpl.importResourceMetadata(RbacxIAMServiceImpl.java:472)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:597)
         at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:307)
         at org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
         at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)
         at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:106)
         at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
         at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
         at $Proxy274.importResourceMetadata(Unknown Source)
         at com.vaau.rbacx.scheduling.executor.iam.IAMJobExecutor.execute(IAMJobExecutor.java:107)
         at com.vaau.rbacx.scheduling.manager.providers.quartz.jobs.AbstractJob.execute(AbstractJob.java:72)
         at org.quartz.core.JobRunShell.run(JobRunShell.java:202)
         at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:534)
    10:56:21,784 ERROR [DBIAMSolution] Error Importing Namespaces : javax.security.auth.login.LoginException: java.lang.SecurityException: [Security:090304]Authentication Failed: User xelsysadm javax.security.auth.login.FailedLoginException: [Security:090302]Authentication Failed: User xelsysadm denied
    10:56:21,836 DEBUG [DBIAMSolution] publishing import completed event...
    10:56:21,903 INFO [VaauSchedulerEventListenerImpl] Job executed: oim tst, IAM
    10:56:21,904 INFO [VaauSchedulerEventListenerImpl] Job run time: 0s
    10:56:21,904 INFO [VaauSchedulerEventListenerImpl] Next Run: null
    Thanks..
    Edited by: IDMuser19 on Aug 24, 2011 4:06 PM
    Edited by: IDMuser19 on Aug 25, 2011 10:58 PM

    Thanks rajiv for the reply..
    No i am not able to import users too..getting same error in the rbacx.log
    I tried with below details
    Xellerate Home: C:\Oracle\Middleware\Oracle_IDM1
    Login Config : C:\Oracle\Middleware\Oracle_IDM1\server\config\authwl.conf
    User Name: XELSYSADM
    Password : ●●●●●●●●
    ( in version 11.1.1.3.6 does not prompt for these two Provider URL: t3://localhost:14000 , Initial Context Factory: weblogic.jndi.WLInitialContextFactory)
    here the OIA and OIM are in different domains.. i am able loginto OIM with xelsysadm without any problem just followed the instruction in the integration doc not sure where i did a mistake..
    Please let know if have any idea..Thanks..
    Edited by: IDMuser19 on Aug 30, 2011 8:25 AM

  • Problem in import of MetaData file in Oracle Warehouse Builder 9.2

    My Problem is Related to Oracle WareHouse Builder.My system's configuration is following :-
    Machine : P4
    Operating System :Windows XP
    ORacle : Version 9.2
    Oracle WareHouse Builder 9.2
    Ram :1GB
    SGA Size :Approximately 650M
    I have 2 MetaData files which I have to import into OWB.
    First file's size is 9M.It got imported properly in 4 minutes without any error.
    2nd File's Size is 30M.when
    I start importing this MEtaData file.It goes upto 35% in Progress Bar.But after that it does not show any improvement.
    I have left the machine running with this option upto 24 hours.
    But nothing happened .It still shows only 35%.
    I have checked log file/Trace file/Alert log files ,but there is no error in these files.
    If any body have any solution about it,Please reply ASAP.I am stucked in this problem from last 7-8 days.
    Waiting for reply
    Thanks & Regards
    Harvinder Singh

    Thanks for your reply Igor.
    As posted the OWB 10.1.0 software is running against a 9.2.0.8.0. database, so OWB still uses DBMS_JOB instead of DBMS_SCHEDULER in this database. In addition I already tested scheduling a simple job to run an OS command as suggested in note 365539.1 , and job scheduling is working fine.
    There are no invalid objects in this database.
    Yasmin

Maybe you are looking for

  • How to get swf file of any mxml file in flex1.5

    hi all, can anybody tell i have one mxml file now when i run this i file server generate swf file dynemically ....now my requirement is i need that swf file so how can i get that swf file from which location i get this file...... plz tell me aasp tha

  • 17" apple ADC--- DVI enabled video card

    I'm having to replace my video card on my g5, the new one I'm getting does not have an ADC connection; I think its just dvi/vga. Is there any lower cost options other then the $99 apple adapter?

  • Can I configure WS-Sec authentication via Active Directory with OSB or OWSM

    Hi I'm planning a project where I need to add security to a group of proxy services in OSB. I need to authenticate them via WS-Security using Active Directory. Is this possible with OSB or adding OWSM? Regards, Néstor Boscán

  • Where can I find an article on how to publish CI as Web Services in PT 8.46

    We have PeopleTools 8.46, and we want to integrate it to a customized application that we are building using ADF, we are considering the option the above option, of publishing the needed components interfaces to web services and consuming them from t

  • Starcam 370i - LEDs on all the time?

    Hi All, Having just purchased a Starcam 370i, I've already posted in another thread about the colour issues I'm having. But my question - is there some way you can stop the white LEDs from automatically coming on each time the computer is turned on?