Issue in migrating dimension from EPMA

Hi,
I have been trying to migrate an EPMA shared dimension from UAT to Dev. New members were created in Dev. But the properties, alias names were not updated. I am migrating the Shared Library dimensions using LCM EPM Architect in Foundation.
Is there a step I am missing?
Regards,
Ragav.

First, Please validate that the properties are truly not getting updated (manually extract them via the File Generator).
Then, I would recommend making a copy of the app, toggling one of the dimensions that isn't migrating properly to Local, and then extracting that from the file generator (as a comparison) - and see if those properties are different than what you expect.
Lastly, I would do 2 things:
1. LCM the 'local' copy over to the target environment
2. Extract hierarchies in target using the file generator to confirm the properties truly are not being updated.
The trick with EPMA is that it keeps multiple copies of members in the relational repository, what I am trying to get you to do is see if a single copy is broken. EPMA 11.1.2.0 is overall less stable than 11.1.1.3.xx, 11.1.1.4.xx and 11.1.2.1+
Also, make sure the properties were set for your appropriate product... Essbase BSO, Essbase ASO, Planning, etc. Each set of properties has a different header and cannot be used across products
Edited by: MattRollings on Aug 22, 2012 8:27 AM

Similar Messages

  • Webi report: Issue Cannot retrieve dimensions from Cube

    Hello Experts,  We're facing an issue with a partical set of Webi reports in Quality environment. The reports are BW reports based on OLAP BICS connectivity on top of Bex queries.y  Below is the scenario and the issue:  1. The backed Bex queries have been trasported to Quality BW system and they're running fine when run from RSRT or Bex analyzer. 2.The Webi reports along with the BICS connections have been migrated to the Quality box and the BICS connections have been editied so that they point to the Quality BOX bex queries and they're testing and responding to the connection successfully. 3. The Webi reports pointed to the Dev Box in the Dev BOBJ system are working fine, but when the same reports for which the connections were repointed in Quality as above, are giving the error "Cannot Retrieve Dimensions from Cube".  Steps taken: a. Deleted the reports and the BICS connection from Quality environment and remigrated the reports,connecitons and changed the connecitons but still the issue persists. b.Tried finding the root cause and got the following issue in the Bex query in Quality: There are filters put in the Bex query . All other filters are working fine except 0CALMONTH. We tried putting all sorts of variables on 0CALMONTH, but the moment a variable is put on 0CALMONTH, this issue occurs. But when we tried removing the variable from 0CALMONTH in filters and restricted it to a single value, the Report works fine.  However the same is not an issue in the Dev Box and all queries with similar variables are working fine.  So we're thinking this might be an issue on the BW backend and the cube might have to be transported again to the Quality Box.  ANy quick suggestions or inputs on this would be greatly appreciated.  Thanks and regards, Abhishek

    Hi Abhishek,
    I am facing same issue.Have you resolve this issue?
    Regards,
    Rajesh

  • Issue with migration DB from SQL Server 2005 to Oracle 11.2 using SQL Developer Migration workbench

    Hi,
    We face an issue while migrating an SQL Server 2005 DB to Oracle 11.2.  It fails during the process.  I hope someone on the forum has seen this before and can give us some advice.
    I use the latest version of SQL Developer with JRE included (3.2.20.09).  The JDBC driver to connect to SQL Server 2005 is 1.2.0
    Here are the steps we take in the Migration Workbench wizard:
    I made an online extract of the SQL Server database.
    Using the workbench in SQL Developer I created a migration repository in schema PDM_MIGRATION.
    Next I start the migration, it captures the tables fine, but immediately after the start of the conversion I get a failed message without further explanations.
    This is the content of the error xml:
    <?xml version="1.0" encoding="windows-1252" standalone="no"?>
    <log>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@4c12ab</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>SEVERE</level>
      <class>oracle.dbtools.migration.workbench.core.logging.MigrationLogUtil</class>
      <message>Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
      <param>oracle.dbtools.migration.convert.ConverterWorker.copyModel(ConverterWorker.java:1078)</param>
      <param>oracle.dbtools.migration.convert.ConverterWorker.runConvert(ConverterWorker.java:316)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doConvert(FullMigrateTask.java:1002)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doMaskBasedActions(FullMigrateTask.java:303)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:205)</param>
      <param>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask.doWork(FullMigrateTask.java:159)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTask.call(RaptorTask.java:193)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask.run(RaptorTaskManager.java:515)</param>
      <param>java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)</param>
      <param>java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)</param>
      <param>java.util.concurrent.FutureTask.run(FutureTask.java:138)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)</param>
      <param>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)</param>
      <param>java.lang.Thread.run(Thread.java:662)</param>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@5dc1bc</param>
      <exception>
        <message>oracle.dbtools.migration.convert.ConvertException: Ongeldig naampatroon.: PDM_MIGRATION .MIGR_FILTER</message>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>1078</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.convert.ConverterWorker</class>
          <line>316</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>1002</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>303</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>205</line>
        </frame>
        <frame>
          <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
          <line>159</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTask</class>
          <line>193</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>oracle.dbtools.raptor.backgroundTask.RaptorTaskManager$RaptorFutureTask</class>
          <line>515</line>
        </frame>
        <frame>
          <class>java.util.concurrent.Executors$RunnableAdapter</class>
          <line>441</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask$Sync</class>
          <line>303</line>
        </frame>
        <frame>
          <class>java.util.concurrent.FutureTask</class>
          <line>138</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>886</line>
        </frame>
        <frame>
          <class>java.util.concurrent.ThreadPoolExecutor$Worker</class>
          <line>908</line>
        </frame>
        <frame>
          <class>java.lang.Thread</class>
          <line>662</line>
        </frame>
      </exception>
    </record>
    <record>
      <date>2013-08-14T16:23:32</date>
      <logger>oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle</logger>
      <level>WARNING</level>
      <class>oracle.dbtools.migration.workbench.core.ui.FullMigrateTask</class>
      <message>Building converted model: FAILED : Database Migration : FAILED</message>
      <param>oracle.dbtools.migration.workbench.core.logging.LogInfo@15a3779</param>
    </record>
    Does anybody know what this error means and what steps we should take to continue the migration?
    I see the PDM_MIGRATION.MIGR_FILTER is a type.
    Many thanks in advance,
    Kris

    Hi Wolfgang,
    Thanks for your reply.
    This is how the type MIGR_FILTER looks like:
    create or replace
    TYPE MIGR_FILTER IS OBJECT (
      FILTER_TYPE INTEGER, -- Filter Types are 0-> ALL, 1->NAMELIST, 2->WHERE CLAUSE, 3->OBJECTID LIST
      OBJTYPE VARCHAR2(40),
      OBJECTIDS OBJECTIDLIST,
      NAMES NAMELIST,
      WHERECLAUSE VARCHAR2(1000));
    I think the repository user has the correct privileges.  This is the overview of privileges it has:
    SQL> select * from dba_sys_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE PRIVILEGE ADM
    PDM_MIGRATION ALTER SESSION NO
    PDM_MIGRATION CREATE CLUSTER NO
    PDM_MIGRATION CREATE DATABASE LINK NO
    PDM_MIGRATION CREATE PROCEDURE NO
    PDM_MIGRATION CREATE SEQUENCE NO
    PDM_MIGRATION CREATE SESSION NO
    PDM_MIGRATION CREATE SYNONYM NO
    PDM_MIGRATION CREATE TABLE NO
    PDM_MIGRATION CREATE TRIGGER NO
    PDM_MIGRATION CREATE VIEW NO
    PDM_MIGRATION UNLIMITED TABLESPACE NO
    SQL> select * from dba_role_privs where GRANTEE in ('PDM_MIGRATION') order by GRANTEE;
    GRANTEE GRANTED_ROLE ADM DEF
    PDM_MIGRATION CONNECT NO  YES
    PDM_MIGRATION RESOURCE NO  YES
    Best regards,
    Kris

  • Issue in Migrating Attachments from Siebel to SAP CRM

    Hi All,
    We are doing Data Migration for Attachments into SAP CRM from Siebel. and have an issue for migrated attachments showing blank pages while opening in SAP CRM.
    We  have used all the formats of converting the data into binary and loading the attachments from Application Server.
    But the PDF/File open with blank pages.  The no. of pages in the original attachment and the migrated file from application server matches.
    May i know the best approach of meeting the above Requirement .. Its prioritized issue !!
    Thanks a lot !!

    Go to sm58-->Select IDOC_INBOUND_ASYNCHRONOUS this function module, in menu mar select edit--> execute LUW.
    or take help from sap basis team, can you please check user in the source system have authorizations profile. 'S_BI-WX_RFC' for this u can refer sap note :150315.
    if you need more information refer this document:
    How to check a BW - SAP source system connection - SAP NetWeaver Business Warehouse - SCN Wiki

  • Linking issues while Migrating code from SUN CC 4.3 to SUN CC 5.8

    Hello,
    I am migrating code written in SUN CC 4.3 to SUN CC 5.8
    I have compiled entire code with SUN CC 5.8 with -compact=4 option ,on machine with OS SunOS 5.8 Generic_Virtual sun4v sparc sun4v.
    during linking phase i am facing issues like
    (cd msggen; make CG="-fast -O3 -compat=4 -lC -features=extensions -i" all)
    /opt/SUNWspro/bin/CC -o msggen msggen.o licensemessage.o ../../../../libs/lib/libgeneral.a
    Undefined first referenced
    symbol in file
    cerr ../../../../libs/lib/libgeneral.a(string.o)
    unsafe_ostream::operator<<(long) ../../../../libs/lib/libgeneral.a(string.o)
    operator delete(void*) msggen.o
    operator new(unsigned int) licensemessage.o
    Iostream_init::Iostream_init(void) msggen.o
    Iostream_init::~Iostream_init(void) msggen.o
    istream::operator>>(char*) ../../../../libs/lib/libgeneral.a(string.o)
    setw(int) ../../../../libs/lib/libgeneral.a(string.o)
    unsafe_ostream::outstr(const char*, const char*) ../../../../libs/lib/libgeneral.a(string.o)
    exrethrow_q msggen.o
    ld: fatal: Symbol referencing errors. No output written to msggen
    I felt like these issues are due to the standard header files. I am including the headers which comes with SUN CC 5.8 and I am using Rogue Wave rwav/ver09.1.
    I am not getting any pointers to solve this issue,any pointer will be highly appriciated.
    Thanks in advance,
    Revathi R

    975741 wrote:
    Yes i did migration by following the Migration Guide, while Migration i ve come across situations where a class method is invisible at many situations like...
    C.h
    class C
    f1();
    B.h
    #include<C.h>
    class B
    f2();
    A.C
    #include<B.h>
    class A:B
    f1();-----> Error Not accessible so I changed to C::f1() then worked like this I ve came across many will it be a Problem, in migration Guide i haven't seen any of such.
    }When you post a code example, it is best to copy/paste from actual code that you verified shows the problem. This example is nowhere near valid for any compiler, even with your suggested change.
    - The class declarations are not terminated by a semicolon
    - You can't call C::f1() using that syntax outside of class C or a class derived from class C because it is not a static function, and requires an object. You could call it as c.f1() if you had a c object of type C, except that f1() would still not be accessible.
    The rules about accessibility have never changed. Please post an example that actually works with C++ 4.x or with C++ 5.x in -compat=4 mode, but that fails with -compat=5 mode.
    >
    Like this for operator overloading also , I haven't got any way to make function body available to function call, so i ve changed those functions to normal functions...I don't understand what you mean. An example would help.
    >
    what is a template cache?? How to delete it??
    The template cache is also known as the template repository. See section 3.3.5 Template Repository in the Migration Guide. To delete the cache (repository) just remove the entire cache directory.
    >
    >
    Another task of mine is, i am compiling a code written in SUN CC 5.5 to 5.8, would you suggest me the migration guide, I could get from 4 to 5 and 3 to 5 but
    not from 5.5 to 5.8, I am not sure about in which compiler code is written but, I could able to compile the code with out compact option with SUN CC 5.5, CC 5.5 and 5.8 are fully compatible. There should be no migration issues. The exceptions would be invalid code that compiled due to a bug in the earlier compiler but is correctly diagnosed as an error with the later compiler. In rare cases, the later compiler has a bug not present in the earlier compiler (a regression). When found, regressions are fixed in patches. If you have an Oracle service contract, you can download patches from My Oracle Support. Otherwise, get the most recent compiler, which should not have those regressions.
    but while trying to compile with 5.8 i am facing issues like
    1. typecastingI would have to see an example.
    2.virtual vector<int> getProcedureArgs() { return 0; } -----> Error: Cannot use int to initialize std::vector<int>
    The compiler error message is correct. The function returns a vector<int>, but you can't create a vector<int> from an integer 0. That is, the vector template has several constructors, but none that accept a single integer argument.
    At this point we have moved away from the original topic. If you have further questions, please start a new thread.

  • Issues with migrating lookups from 9.0.3 to 9.2

    Below is a text from a TAR we've submitted with Oracle. Has anyone else run into this issue when upgrading from 9.0.3 to 9.2?
    Lookups conditions not being properly migrated is a known issue when upgrading from OWB 9.0.3 to 9.2,
    documented in Bug 3989847 - Generated Code Has Invalid Aliases After Upgrade From 9.0.3 to 9.2

    Did you try putting the toplinkProject.xml into the META-INF directory? Does the same error occur?

  • Issues when migrating workbook from one EUL to another EUL

    Dear Members,
    We have two types of EUL. One is an standard EUL and another one is Apps based EUL. We have many reports created in standard eul and my task is to move these workbooks to an apps based eul.
    Standard EUL is nothing but a database based eul. Users who created these workbooks in this EUL are the database users. Now i exported these workbooks to an .eex file using the command line interface. But when i am trying to import these workbooks in an apps based eul i cannot see them after import. I know this might be an previlages issue.
    My question is these two eul's are completly different from one another. One is database based and another one is apps based and even the users in both eul's are different as they have different accounts in database and application.
    We have a lot of reports created using the database based eul. Our goal is to move these to apps based eul so that users can use there responsibilities and run the workbooks rather than using desktop by logging in as a database user.
    How should i proceed in migrating these reports. It would be great if any one guides me.
    Thanks
    Sandeep

    Hi,
    You should ensure that you take ownership of the workbooks when they are imported and that you log in as an apps user (not the EUL owner) when the workbooks are imported.
    Rod West

  • Issues while migrating data from a disk based table to a memory optimized table

    Hi All,
    I have a Disk based table with 400000000 rows in it, We are trying to convert it into a memory optimized table.
    We have  already created a memory optimized table with similar structure and trying to import data into this mem optimized table using 'insert into' from the disk table.
    I am trying to Migrate around 10000000 rows at a time,  but I am getting an error 'There is insufficient system memory in resource pool 'default' to run this query.' Altough we have 128 GB RAM on the server and SS is utilizing more than 120 GB RAM.
    Altough the query has been cancelled.
    Wanted to Know how could we migrate the table with the available RAM or do we have increase our RAM?
    aa

    Josh,
    Microsoft's documentation on this subject isn't at its best right now (I believe there will be incremental improvements for better understanding), but here is what I read so far.
    http://msdn.microsoft.com/en-us/library/dn133190.aspx
    "A hash index consists of a collection of buckets organized in an array. A hash function maps
    index keys to corresponding buckets in the hash index."
    Judging by this statement, a hash index is a hash table just like the ones used as work tables for hash operators in queries (hash matching or grouping). Doesn't contain (or include) other columns, i. e. it doesnt store any data.
    "Multiple index keys may be mapped to the same hash bucket."
    This means there is some kind of mapping, but this is not explained in the article above. However...
    http://msdn.microsoft.com/en-us/library/dn282389.aspx
    "For each hash index in the table, each row has an 8-byte address pointer to the next row in the index. Since there are 4 indexes, each row will allocate 32 bytes for index pointers (an 8 byte pointer for each index)."
    Each row (in the table) has a pointer (for every index, 1:1 ratio) that points to a row (also known as bucket) in the hash index. So that is how the aforementioned mapping works huh!
    > What happens if you include a column in two or three different indexes, or is that not allowed?
    My conclusion is that the hash indexes works the same way as a hash worktable, with the addition of the column in the base table that is added to store pointers to the hash index.
    When you create a new index, even if you use the same column twice, a hash table is created, hash calculations are distinctly made for each key and stored on it, and while this is done, the column that is exclusively used for this new index is populated
    with pointers to this index. You can add a given column to the set of keys of different hash indexes as many times as you want. Correct if i'm wrong, I'm also new on this subject :D

  • Issue after migrate  workbook from 3.1 to BI Netweaver 7.0

    Hy experts,
    We have a workbook in v.3.1 with several querys in only one sheet, one below the other.
    When we migrate this workbook to 7.0 and refresh it, the system shows this message:
    <i>"Some item(s) overlap with each other".</i>
    Then an unassigned navigation pane (no characteristic is asiggned to it), corresponding to the first query,  is put on the analysis Grid, corresponding to the second query, so some rows are not visible.
    How can we do to avoid that unassigned navigation pane?
    Regards.

    Hi,
    Issue may be due to the Excle formulae used in the workbook which do not get migrated.
    One way around it is to migrated the queries first and apply Excel validations on top of it in 7.0
    Hope this helps.
    Thx,
    SK

  • Export Dimensions using EPMA Batch Client

    HI,
    Does anyone know if there is a way to export the dimensions from EPMA using a command line tool like the EPMA Batch Client. I have gone through the guide and there are commands to import the dimensions but none to export them from EPMA. We are on EPM 11.1.2.1 version and all our apps are using EPMA with planning so unfortunately i cannot use the planning outline extractor tool either. Thanks

    You can use LCM to export a hierarchy, and that can be scripted LCM batch utility. Create a migration to the file system with LCM. Select the hierarchies you want to export. When you get to the point where it asks you to execute the migration, there is also an option to save an XML file with the options you selected. You can use that xml file with the utility. Details are found at http://docs.oracle.com/cd/E17236_01/epm.1112/epm_lifecycle_management/frameset.htm?launch.htm.
    The exports are in the import/export folder under the ID of the person executing the migration. The folder/files can be copies anywhere.
    Kyle Goodfriend
    http://www.in2hyperion.com
    h4. Please make sure to assign the post as answered and the responses as answer or helpful so others may benefit from your experience.

  • Excluding Members from EPMA application - performance issues?

    We are building a new application using dimensions from our Shared Library and I had a question about excluding some members.
    For one dimension, I was able to use the "Add to App View" functionality, which was able to bring in only the one tree of members I wanted, without having to exclude anything, which seems to be the best way of doing this.
    For another dimension, I need to pick and choose members from within the hierarchy, so I will need to add almost all parents, and then go through and exclude some members. Will this cause any performance issues having excluded members? Will the excluded members get deployed to planning or do they just sit in EPMA?
    Any help would be appreciated.

    Hi,
    Excluded members are not deployed to Planning, therefore would not impact performance negatively.
    Cheers,
    Alp

  • Migrating Application from JDev 10.1.3.5 to JDev 11.1.2.2.; WebLogic issues

    Hi,
    I've been in the process of migrating an application we have from JDev 10.1.3.5. to the most current release. This involved migrating from OC4J to the WebLogic Server. After clearing up code-related issues, I have been trying to mitigate errors that came from deploying the application to the WebLogic server. I fixed (or so I think) the original error(see immediately below in italics) by deleting the servlet tag in both of the web.xml files (ViewController and Deploy) that are within the application. I also changed it to 'adfLibResources' per a previous thread on this topic.
    +<Error> <HTTP> <BEA-101371> <There was a failure when processing annotations for application C:\Users\priestem\AppData\Roaming\JDeveloper\system11.1.2.2.39.61.83.1\o.j2ee\drs\Muster\ModelWebApp.war. Please make sure that the annotations are valid. The error is oracle.adf.view.faces.webapp.ResourceServlet>+
    Now I am getting the following errors:
    <Aug 6, 2012 3:24:28 PM CDT> <Error> <HTTP> <BEA-101371> <There was a failure when processing annotations for application C:\Users\priestem\AppData\Roaming\JDeveloper\system11.1.2.2.39.61.83.1\o.j2ee\drs\Muster\ViewControllerWebApp.war. Please make sure that the annotations are valid. The error is >
    <Aug 6, 2012 3:24:28 PM CDT> <Error> <Deployer> <BEA-149265> <Failure occurred in the execution of deployment request with ID '1344284600393' for task '0'. Error is: 'weblogic.application.ModuleException: Failed to load webapp: 'Muster''
    weblogic.application.ModuleException: Failed to load webapp: 'Muster'
         at weblogic.servlet.internal.WebAppModule.prepare(WebAppModule.java:393)
         at weblogic.application.internal.flow.ScopedModuleDriver.prepare(ScopedModuleDriver.java:176)
         at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
         at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:517)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
         Truncated. see log file for complete stacktrace
    Caused By: java.lang.ClassNotFoundException: oracle.adfinternal.view.faces.taglib.UIXPanelTag
         at weblogic.utils.classloaders.GenericClassLoader.findLocalClass(GenericClassLoader.java:297)
         at weblogic.utils.classloaders.GenericClassLoader.findClass(GenericClassLoader.java:270)
         at weblogic.utils.classloaders.ChangeAwareClassLoader.findClass(ChangeAwareClassLoader.java:64)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:305)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:246)
         Truncated. see log file for complete stacktrace
    >
    <Aug 6, 2012 3:24:28 PM CDT> <Warning> <Deployer> <BEA-149004> <Failures were detected while initiating deploy task for application 'Muster'.>
    <Aug 6, 2012 3:24:28 PM CDT> <Warning> <Deployer> <BEA-149078> <Stack trace for message 149004
    weblogic.application.ModuleException: Failed to load webapp: 'Muster'
         at weblogic.servlet.internal.WebAppModule.prepare(WebAppModule.java:393)
         at weblogic.application.internal.flow.ScopedModuleDriver.prepare(ScopedModuleDriver.java:176)
         at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
         at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:517)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
         Truncated. see log file for complete stacktrace
    Caused By: java.lang.ClassNotFoundException: oracle.adfinternal.view.faces.taglib.UIXPanelTag
         at weblogic.utils.classloaders.GenericClassLoader.findLocalClass(GenericClassLoader.java:297)
         at weblogic.utils.classloaders.GenericClassLoader.findClass(GenericClassLoader.java:270)
         at weblogic.utils.classloaders.ChangeAwareClassLoader.findClass(ChangeAwareClassLoader.java:64)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:305)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:246)
         Truncated. see log file for complete stacktrace
    >
    [03:24:28 PM] Deployment cancelled.
    [03:24:28 PM] ---- Deployment incomplete ----.
    [03:24:28 PM] Remote deployment failed (oracle.jdevimpl.deploy.common.Jsr88RemoteDeployer)
    #### Cannot run application Muster due to error deploying to IntegratedWebLogicServer.
    [Application Muster stopped and undeployed from Server Instance IntegratedWebLogicServer]
    Can someone please provide insight/direction? I restarted this whole process because I was not sure if I was making mistakes. But I have received this same exact error both times in both copies of the application code.
    Thanks,
    M. Priester

    I think your web.xml that is a part of the application had the following:
    <taglib>
    <taglib-uri>
    http://xmlns.oracle.com/adf/faces
    </taglib-uri>
    <taglib-location>
    /WEB-INF/lib/adf-faces-impl.jar
    </taglib-location>
    </taglib>
    I think this is not even your configuration but something that JDev added, thus you can consider testing removing this part of the code and test your application and verify the functionality.
    Regarding the possible issue during migration from 10g to 11g of Jdev, you can review the following thread:
    Class not found : oracle.adfinternal.view.faces.taglib.UIXPanelTag
    The missing class is a part of "adf-faces-impl-1013.jar" which seems to be specific to the JDev 10g, thus might not be even required by JDev11g.
    Also regarding the fact that how this seems to be an issue with missing class, please review the following API doc for better understanding of the classnotfoundexception:
    http://docs.oracle.com/javase/6/docs/api/java/lang/ClassNotFoundException.html
    Cheers,
    AJ

  • JMS issues when migration from weblogic 9.2 to 10.3.5

    We are facing some issues when migration from weblogic 9.2 to 10.3.5
    In  weblogic 9.2 :_
    BMP Entity EJBs used in our project are read-only in nature using entity cache, below is the configuration details
    <!DOCTYPE weblogic-ejb-jar PUBLIC "-//BEA Systems, Inc.//DTD WebLogic 6.0.0 EJB//EN" "http://www.bea.com/servers/wls600/dtd/weblogic-ejb-jar.dtd">
    <weblogic-ejb-jar>
    <weblogic-enterprise-bean>
    <ejb-name>
    Company
    </ejb-name>
    <entity-descriptor>
    <pool>
    <max-beans-in-free-pool>300</max-beans-in-free-pool>
    <initial-beans-in-free-pool>150</initial-beans-in-free-pool>
    </pool>
    <entity-cache>
    <max-beans-in-cache>3500</max-beans-in-cache>
    <idle-timeout-seconds>100000</idle-timeout-seconds>
    <read-timeout-seconds>0</read-timeout-seconds>
    <concurrency-strategy>ReadOnly</concurrency-strategy>
    </entity-cache>
    Entity beans will get refreshed using the JMS messges. with in the MDB descriptor files(weblogic-ejb-jar.xml) we are using the provider URL directly and XA enabled connection factory is set to false.
    migration to Weblogic 10.3.5_
    With the same configurations MDB are not not getting deployed in weblogic 10 with some exception, so we removed the provider URL from weblogic-ejb-jar.xml and changed the JMS configuration to use foreign JMS and XA enable connection factory is set to true. Now when ever the JMS message is triggered Entity bean is not getting refreshed with the updated values. i.e values are stale.
    Can some one look into this and provide your inputs to resolve this issue.

    I think the Entity bean refresh problem appears to be unrelated to MDBs. The MDB is only responsible for getting the message to your application (which in turn interacts with Entity beans). You might want to try posting your question to an EJB newsgroup.
    Tom

  • FR Studio - Is there a workaround for the issue with referencing data from a grid with a different DB connection and different # of dimensions?

    Hi,
    I am trying to reference a column in a different grid with a different database connection. I did some research in the forums, and it seems like this is a known issue with FR Studio. The issue is that if the 2 different database connections have different # of dimensions, then the reference will throw an error.
    For example, what I'm trying to do is below:
    In grid2.column[A], I have the following formula:
    grid1.column[A]
    The result is an error.
    The ultimate goal is to have a basic variance formula in the grid between grid1 and grid2, which have different database connections with a different # of dimensions. Is there a workaround at all to this bug? The example of the basic formula is below:
    grid1.column[A] - grid2.column[A] 
    Any help is much appreciated!
    Thanks,
    Tiffany

    Thanks for the detailed response Mike.
    The issue with trying it in Grid1 vs Grid2 is that Grid1's DB connection has more dimensions than Grid2's DB connection. I can't change the DB connection to a database with more dimensions because you have to be able to map all dimensions from the existing DB connection  (Grid1) to the new DB connection (Grid2). This isn't possible, so instead I have to map Grid2 (less dimensions) to Grid1 (more dimensions). It just ignores the extra dimensions from Grid1. Does that make sense?
    Thanks again for your responses. It's very much appreciated.
    Tiffany

  • EPMA very slow to import large dimension from interface tables

    I am attempting to import a dimension into the master library from the EPMA interface tables. The dimension is roughly 255,000 members. The import from the interface tables into an empty dimension member is taking close to 3 hours. CPU utilization on the EPMA server is a steady 6% (epma_server.exe) for the entire time. We are on 11.1.2.1.001 of EPMA. The rest of the suite is at 11.1.2.1. The dimension generic type. I have had the same result when importing into a local dimension. The performance degrades after about 6000 members.
    Your thoughts would be much appreciated.

    I have found my answer from EPMA guide:
    In addition to the dimension interface tables created by the template script, you can add interface tables for additional dimensions. For example, the template script contains one set of tables for the Entity dimension. You can add more Entity dimensions as needed. For each dimension added to the interface tables, you must also include the dimension in the IM_Dimension system table so that the dimension is available during profile creation.

Maybe you are looking for

  • HT1363 iPod nano 5th generation is not recognized by computer

    I've uninstalled and installed iTunes, gone through the 5Rs, attempted all the USB ports, restarted the computer, attempted putting the iPod in Disk Mode, and I'm now going to put my iPod on YouTube with a shameful sign on it to punish it until it co

  • Why is my Mac storing graphics and photos from the web in iPhoto?

    How can I prevent my Mac from downloading graphics and photos from the web?   How do I remove the thousands of images that the Mac moved into iPhoto when I browse the web? 

  • MS Office Professional Outlook 2013 Calendar, Tasks, People Tabs Disfunctional

    After installing iCloud on Windows 7 Professional 64-bit OS, it worked well with my Outlook calendar except for the fact that the publish function was no longer an option. So I uninstalled iCloud. After uninstalling, my Outlook was permanently change

  • Audio stopped working through dock connector

    I recently installed an iPhone dock in my car (used the composite video kit, and an approved car charger), but today the iPhone stopped outputting audio through the dock connector, so I can't get music or GPS through my car stereo. Thinking it might

  • Banner with customer logo in top of E-Recruiting

    <b>I am trying to find out how to add a banner in the top of the E-Recruiting BSP applications containing a customer logo</b> I have looked into various documentation, older topics and blogs on how to modify themese for BSP - but I have not been able