10.1.3 Migration Issue with data-sources.xml

Hi:
Running into lots of migration issues for an application built in 10.1.3 EA, migrating to 10.1.3. production, mostly in the BC project. Here's one: the database connections built in EA were not carried over to prod, so they needed to get re-built. No biggie. But now the data-sources.xml file in the model project has duplicate entries for my data sources. I have tried directly editting that file and removing the duplicates, but when I go to run my application (ADF Faces & ADF BC) JDeveloper hangs. When I abort JDev and re-start it, the duplicate entries are back in data-sources.xml (and yes, I did save the file after making the changes). It seems that the xml file is built dynamically from definitions stored elsewhere. Where is the source? What will I need to modify to get the changes to take?
Thanks.
Johnny Lee

There is new control over when/whether JDeveloper add/updates your data-sources.xml file for the embedded server. The settings are located here:
Tools | Embedded OC4J Server Preferences | Current Workspace | Data Sources
There are some settings there that control whether/when we update the data-sources.xml

Similar Messages

  • Issue with data source after deploying

    We are experiencing an issue with our data source after deployment of a cube. On the datasource properties in Visual Studio 2012, we have the max connections set to 0 before the deployment. Once the cube is deployed, I can navigate to the <name>.0.ds.xml
    file and open it and see that the <MaxActiveConnections>0</MaxActiveConnections> is indeed set to 0. At some point over the next couple days, a process of the cube or some other action causes that value to get updated to some number too large to
    be converted to an int, and makes the datasource invalid. At that point we cannot view the datasource properties in SSMS, we cannot open the cube project from Visual Studio, and we’ve even had failures when trying to process the cube.  Is there a config
    somewhere that would cause this value to get overwritten, or some other behind the scenes process that we can look at?
    Our server information is:
    Microsoft SQL Server 2012 (SP1) - 11.0.3153.0 (X64)
                    Jul 22 2014 15:26:36
                    Copyright (c) Microsoft Corporation
                    Enterprise Edition: Core-based Licensing (64-bit) on Windows NT 6.2 <X64> (Build 9200: ) (Hypervisor)
    Chad Dotzenrod SWC | TECHNOLOGY PARTNERS 1420 Kensington Road, Suite 110 Oak Brook, Illinois 60523-2144 http://www.swc.com

    Typically you would import the metadata from the source location and either use that location as the data source (and so not need to redeploy), or deploy it to a separate target location.
    The replace action is destructive as you've found, and effectively performs a drop table followed by create table. Hence any data in the table is lost.
    If you just want the Control Center Manager to correctly display that the table is deployed, try setting the action to "Upgrade". This will try to upgrade the deployed object to match the definition in OWB, but as the two are identical this will result in no changes. However, it will update the deployment records to indicate that the object is deployed.
    Nigel.

  • Issue with Data sources

    Hi Gurus,
               I have issue like can I use one data source for more than one info source.
               And one more thing is it necessary to add key figures in Comm Struc, and in Tran. Struct.
               Please I am still in starting stage of BW. If any have idea give me a advice or email me to [email protected]
    Regds
    Dayakar Reddy.

    Hi SK,
          Acutally I need only two fields from data source like Document Type and Purchasing group. Can I select only this two field  from the list which already existing in Comm Struct and in Tran Struct.
          Does it work out and one more thing about key figures Here I need to include key figures as well......
           Please give me good advice where I come acros this issue....
    Regds
    Dayakar Reddy

  • Failed to acquire DAS with data sources that match the specified OCE

    Please help, Thank you all.
    We are migrating Oracle databases from Windows to Linux servers. Currently, I have no problem to use my desktop Hyperion Intelligence Designer 8.5 with new oce file to connect to Linux. However, I published the same oce file to Hyperion Performance Suite 8.5 then I got “Sever Error [2018]: Failed to acquire Data Access Service with data sources that match the specified OCE”.
    I launched Service Configurator, in the “Local Service Configurator - DAS1_host”, the “Database JDBC URL is point to Linux server. Also, I added ODBC connection to Oracle database on Linux in the “Properties of Data Access Service: DAS1_host’s Data Source”. Database of JDBC URL in BI1_host, AZ1_host, PUB1_host, SM1_host, LS1_host, UT1_host, BPS1_host, AN1_host are also point to Linux server.
    The “Remote Service Configurator” has ES1_host’s” Storage” with JDBC URL point to Linux server. The same for RM1_host’s and NS1_host’s “Storage” with JDBC URL point to Linux server.
    Is there some missing configuration on the server? Thank you in advance.

    You should login to my oracle support to look into this ID.
    This is the solution mentioned
    Solution
    1) In the .profile file, you will find the definition of the LD_LIBRARY_PATH variable. However, the path to 64 bit libraries is probably appended before the path to the 32 bit libraries(%ORACLE_HOME%/lib32). Therefore, please reverse the order in the definition so that the 32 bit libraries get loaded before the 64 bit libraries.
    Then, restart all services.
    2) Oracle 9.2 was the first Oracle release which defaults to using the 64 bit libraries in the 'lib' directory and the 32-bit libraries that we need are in the 'lib32' directory.
    Therefore, please have the%ORACLE_HOME%/lib32 placed in front of the LD_LIBRARY_PATH environment variable definition.Then, restart all services.
    3) Please check that users have read and write access to the files located in the %ORACLE_HOME%/lib32 folder.
    4) Check the permissions of the libclntsh.a file(located in the %ORACLE_HOME%/lib32 folder) are set to rwxr-xr-x
    5) Recreate the OCE connection in DAS and restart DAS in CMC.
    6) Restart all the BIPlus Services and process the query.

  • Issue with Data Provider name in variable screen for BEx Analyzer

    Hello all,
    We got an issue with Data Provider name in Variable screen in BEx Analayzer.
    We want to change the DataProvider name there to Description of the report instead of its Technical name.
    Any inputs are appreciated.
    Thanks
    Kumar

    You have to create a workbook to do this.
    Refresh your query/report. In Bex analyser, there is one toolbar named BEx design toolbox, If you are not able to see it in analyser, right click on the toolbar space of BEx analyser and click on BEx design toolbox. Here, goto to design mode, by clicking on a sysbol like 'A'. after that place the curser where you want to see the Query description. and click on insert text (T) in BEx toolbox. click on it and check "Query description" in constant tab. in the general tab you need to assign a dataprovider, for that assign your query name in workbook settings (in Bex design toolbox). also check the "display caption" in general tab.
    Pravender

  • Issue with data dictionary -Table maintanance generator

    Hi all,
    I have an issue with Data dictionary, table maintenance generator. I have entered some records in a custom table (ZBCSECROLETOGRP) and changed the delivery class from C to A. When I create the table maintainance generator, I am encountered with the following errors:
    1)Field ZBCSECROLETOGRP-PORTALGROUP shortened (new visible length: 000032)
    2)0012 could not be generated
    3)In TCTRL_ZBCSECROLETOGRP field LENGTH has the invalid value 01
    My main motto is to create the table maintainace generator and transport to the furthur systems .
    Please help.
    ThnX in advance,
    Vishal..

    HI,
    Regenerate the table maintenance by selecting the checkbox of "Modified field structure" => new entry & then save.
    Also ensure that the new changes are not affecting old data bcz of data type changes. If that is the case, then delete the old records, regenerate table maint. & re-enter those records which you had deleted.
    Thanks,
    Best regards,
    Prashant

  • Issue with Date Format

    Hi All,
    I m facing an issue with Date format in the prompt. I have used date presentation variables in my column formula as shown below:
    FILTER("SKU Order Details"."Fulfilled Quantity" USING Time."Calendar Date" <= DATE '@{todate}{2900-01-01}')
    The report returns data when I don't select any date range for start & end date prompts on the page. But when I select the start & end date values in the prompt, I m getting the following error:
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 46047] Datetime value 10/22/2009 12:00:00 AM from 10/22/2009 12:00:00 AM does not match the specified format. (HY000)
    I included the following formulas for start & end date prompts:
    Start Date prompt: case when 1=2 then License."Ips Creation Date" else cast ('1.1.1900' as date) end
    End Date prompt: case when 1=2 then License."Ips Creation Date" else cast ('1.1.2900' as date) end
    Can you please help me resolve the issue.
    Thanks,
    Kartik

    Hi Nico,
    I tried putting the format that you mentioned, I m getting an error message.
    My prompts have the following formula :
    Start: case when 1=2 then License."Creation Date" else cast('1.1.1900' as DATE) end
    End: case when 1=2 then Time."Calendar Date" else cast('1.1.2900' as DATE) end
    My column formula has the following syntax:
    FILTER("SKU Order Details"."Fulfilled Quantity" USING Time."Calendar Date" between DATE '@{start}{1900-01-01}' AND DATE '@{end}{2999-01-01}')
    Error Message:
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 46047] Datetime value 11/17/2009 from 11/17/2009 does not match the specified format. (HY000)
    Can you please let me know if something needs to be changed.
    Thanks,
    Kartik

  • Missing data-sources.xml deploying war with jdeveloper 9i RC

    When deploying a war file to an OC4J 1.0.2.2.1 server and error reports during the deployment that data-sources.xml is missing in the application-deployments/xxxx directory. Only the orion-application.xml file is present - which references a data-sources.xml file in the same directory. The problem is : it isn't there.
    When I copy the xml files from the applications/xxxx directory to the application-deployments/xxxx directory and redeploy the war file everything works fine.
    Is this a configuration issue or a bug ? Anyway, can somebody tell me how I can work around this problem ?
    Thanks !

    this is a bug and you have already found the workaround.

  • Problems with OC4J EJB deployment and the data-sources.xml file

    I am running 2 Windows 2000 Machines one with the 8.1.7 database another with iAS 1.0.2.2.1 and OC4J.
    I am trying to deploy a 3rd party EJB-based application whic seems to have deployed successfully except when I try to test the EJB deployment via a jsp it can't connect to the database, giving the error:
    1/16/02 4:52 PM VerySimple: Servlet error
    java.lang.NoClassDefFoundError: com.netexp.user.UserManagerHome
    at com.netexp.beans.BeanHelper.class$(Unknown Source)
    at com.netexp.beans.BeanHelper.getUserManagerBean(Unknown Source)
    at /very_simple.jsp._jspService(/very_simple.jsp.java:48) (JSP page line 27)
    at com.orionserver[Oracle9iAS (1.0.2.2.1) Containers for J2EE].http.OrionHttpJspPage.service(OrionHttpJspPage.java:54)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.HttpApplication.serviceJSP(HttpApplication.java:5459)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.JSPServlet.service(JSPServlet.java:31)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.ServletRequestDispatcher.invoke(ServletRequestDispatcher.java:508)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.ServletRequestDispatcher.forwardInternal(ServletRequestDispatcher.java:177)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.HttpRequestHandler.processRequest(HttpRequestHandler.java:576)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].server.http.HttpRequestHandler.run(HttpRequestHandler.java:189)
    at com.evermind[Oracle9iAS (1.0.2.2.1) Containers for J2EE].util.ThreadPoolThread.run(ThreadPoolThread.java:62)
    I try to update the file using the installdatasource switch thus:
    C:\Oracle\iSuites\j2ee\home>java -jar admin.jar ormi://localhost admin adm_pwd -application apptricityII -installDataSource -jar %ORACLE_HOME%\jdbc\lib\classes12.zip -url jdbc:oracle:thin:@db_host.unitas.com:1521:db_name -connectionDriver oracle.jdbc.driver.OracleDriver -location jdbc/pool/OracleDataSource -username scott -password tiger
    And get the following error:
    Exception passing by from remote server: java.lang.InstantiationException: No class specified for jdbc/pool/OracleDataSource
    java.lang.InstantiationException: No class specified for jdbc/pool/OracleDataSource
    <<no stack trace available>>
    Error adding source: No class specified for jdbc/pool/OracleDataSource
    Please advise.
    I have followed the instructions in 'EJB Primer' and 'Using Oracle9iAS Containers for J2EE' to no avail. And I can't find any meaningfull data-sources.xml samples. I have been struggling with this for some time.
    Is there any other file, do I need to bind anything?
    I'd appreciate your assistance.
    Thank you
    Louiza

    Hi Louiza,
    Show us your web.xml and ejb-jar.xml files as well as your
    data-source.xml file.
    Thanks,
    Avi.

  • Problem with embedded data-sources.xml and custom UserManager

    Hi all,
    Our application uses a custom UserManager, which is basically extended from the JDBC UserManager, declared as follows in orion-application.xml:
         <user-manager class="com.infocorpnow.a2g.security.oracle.A2GUserManager">
              <property name="table" value="pos.users" />
              <property name="userNameField" value="username" />
              <property name="passwordFiled" value="password" />
              <property name="dataSource" value="jdbc/A2GDS" />
              <property name="groupMemberShipTableName" value="pos.user_roles" />
              <property name="groupMemberShipGroupFieldName" value="role_name" />
              <property name="groupMemberShipUserNameFieldName" value="login_id" />
         </user-manager>
    Since we want to be able to deploy the application several times on the application server, and therefore have each deployment of the ear point to its own datasource (i.e. its own local "A2GDS"), we've found out how to embed data-sources.xml inside the EAR file we're deploying, and modify the orion-application.xml as follows:
         <data-sources path="./data-sources.xml" />
    And then place data-sources.xml in the same meta-inf folder as the orion-application.xml.
    This has worked fine when deploying to the standalone OC4J.
    Now when I try to deploy the exact same EAR file in Oracle 9iAS, and I get to the User Manager screen, the Custom User Manager does not show up correctly. It did show up prior to me embedding the data-sources.xml. Please help? This is fairly urgent.
    Thanks
    Jason

    I should also mention I'm using the Java Edition of 9iAS R2 (9.0.3 container) on Solaris.

  • ConnectionCacheName not read from data-sources.xml

    I created a simple web application to test the use of the ConnectionCacheBean (from the jdbc 3.0 examples page). I read through the code and some other articles regarding data sources and noted that you "should" be able to set the connection cache name and caching enabled, as well as failover support, even in oc4j 9.0.4 and 10.1.2. I tried this (at least the first two), and it would NOT set the cache name.
    The data-sources.xml entry was the same as in the one documented here:
    http://www.oracle.com/technology/tech/java/newsletter/articles/oc4j_data_sources/oc4j_ds.htm
    What gives?
    Also, I tried to simulate the database going down - I killed all the sessions rather than doing a shutdown, but anyway... even with a connection cache property of validatelogin (i think that's the name) to true, and a refresh of invalid connections (they'd all be invalid if I killed them), it would only reactivate ONE SESSION. If I refreshed ALL Sessions, then it would do 9 of the 10 sessions (10 being my initial cache limit).
    Any ideas?
    I can post the code if that is an issue.

    A response I got back from Frances Zhao regarding the two issues that I experienced was:
    Issue#1: Cannot set the connectionCacheName via data-sources.xml
    1. This is very likely because of an OC4J bug that affects the actual order of
        setting each <property> specified in data-sources.xml.  The connection cache
        in OracleDataSource requires the cache-name property be set after the cache
        is enabled; but OC4J may not always follow this order, causing the name
        to be null.  If you really need the cache name, you have to do it in the code
        for now.
    Issue#2: Can't refresh the invalid connections
    2. These are known issues in the Oracle JDBC driver in both 10.1.0.4 and
        10.2.0.1.  The fix for the REFRESH_INVALID case will be in the next
        patch release for both 10.1.x and 10.2.x; the fix for the REFRESH_ALL
        case will come later.

  • rib-rms /application-deployments/rib-rms/data-sources.xml

    Hi,
    Support note 1314377.1 "Call To GetNextPublisher Service In RIB-RMS Fails With "Unable To Publish Message To JMS" Exception" says:
    Solution
    "The issue was resolved with the below suggestion:
    Modify $ORACLE_HOME/j2ee/<rib-rms>/application-deployments/rib-rms/data-sources.xml to have below entries and bounce the <rib-rms> instance:
    min-connections="10"
    max-connections="20"
    initial-limit="5"
    used-connection-wait-timeout="10"
    inactivity-timeout="10"
    login-timeout="0"
    connection-retry-interval="1"
    max-connect-attempts="3"
    validate-connection="true"
    validate-connection-statement="select count(*) from dual"
    num-cached-statements="10"
    time-to-live-timeout="-1"
    abandoned-connection-timeout="6"
    disable-server-connection-pooling="false"
    property-check-interval="2"
    lower-threshold-limit="20" "
    However it is not at all clear to me how to enter these elements in the data-sources.xml XML structure.
    I have searched OTN and Google for it but can't find a good/complete reference. E.g. http://download.oracle.com/docs/cd/B25016_08/doc/dl/web/B14012_02/datasrc.htm documents some of the above attributes, but not all.
    Not all attributes go into the same XML element in data-sources.xml.
    Can someone give an example of how to set this?
    Thanks,
    Erik
    Edited by: ErikYkema on Jul 19, 2011 4:29 PM

    Hi Erik,
    All of these are attributes to <connection-pool> tag.
    So it will ideally look like
    <connection-pool name="connectionPool" min-connections="10"  max-connections="20" >
    <connection-factory factory-class="oracle.jdbc.pool.OracleDataSource"      user="RPM_ADAPTOR_AQ"
                   password="RPM_ADAPTOR_AQ"
                   url="jdbc:oracle:thin:@//+hostname+/RMS131K" >
              </connection-factory>
    Please modify the same are restart rib oc4j instance.
    Regards,
    PPS
    Edited by: ppskannan on Jul 19, 2011 8:19 PM
    Edited by: ppskannan on Jul 19, 2011 8:21 PM

  • 903/902/BC4J can't get data-sources.xml conn pooling to work in production; help

    I have several BC4J ears deployed to a 903 instance of OC4J being configured as a standalone
    instance. I've had this problem since I started deploying in development on 902. So it's
    some basic problem that I've not mastered.
    I can't get data-sources.xml managed connection pooling to actually pool conn's. I'm wanting
    to declare my jndi jdbc source connection pool in j2ee/home/config/data-sources.xml.
    Have all BC4J apps get conns from this JNDI JDBC pool. I've removed all data-sources.xml from my BC4J ears,
    and published the jndi jdbc source in my oc4j common data-sources.xml. I've tested that this is
    the place controlling the conn URL/login passwd by commenting it out of config/data-sources.xml
    and my BC4J apps then throw exceptions, can't get conn.
    I've set the oc4j startup cmd line with the BC4J property to enabled connection pooling:
    -Djbo.doconnectionpooling=true
    symptom
    Connections are created and closed. Instead of being put back into the pool managed by oc4j,
    what ever BC4J is doing or my data-sources.xml is doing, the connections are just being created and
    closed.
    I can verify this via (solaris) lsof and netstat, where I see my oc4j instance under test load
    with only 1 or 2 conns to the db box, and the ephemeral port is tumbling, meaning a new socket is
    being opened for each conn. ;( grrrrrrr
    Does anyone have a clue as to why this is happening?
    Thanks, curt
    my data-sources.xml
    <data-sources>
         <data-source
            class="com.evermind.sql.DriverManagerDataSource"
            connection-driver="oracle.jdbc.driver.OracleDriver"
            ejb-location="jdbc/DEVDS"
            location="jdbc/DEVCoreDS"
            name="DEVDS"
            password="j2train"
            pooled-location="jdbc/DEVPooledDS"
            url="jdbc:oracle:thin:@10.2.1.30:1521:GDOC"
            username="jscribe"
            xa-location="jdbc/xa/DEVXADS"
            inactivity-timeout="300"
            max-connections="50"
            min-connections="40"
        />
    </data-sources>

    I've run another test using local data-source.xml, that's packaged in the .ear. Still
    pooling under BC4J doesn't work??
    A piece of info is that the 903 oc4j release notes states that global conn pooling doesn't
    work. Infering that the j2ee/home/config/data-sources.xml data sources aren't pooled or ??
    I just tested so called local connection pooling, where I edited the data-sources.xml that
    gets packaged in the ear, to include the min/max params and re-ran my test.
    Still, the AM creates a new conn, it's to a new socket, and closes the conn when done. Causing
    each conn to not be pooled, rather opened then closed to the DB box. As verified with lsof and
    netstat, checking the ephemeral port # on the DB box side, always changes, meaning it's a
    new socket and not an old pooled conn socket.
    ???? What the heck??
    Surely if the AM conn check out / return code works properly, OC4J's pooling JDBC driver would
    pool and not close the socket??
    Has anywone gotten JDBC Datasource connections in BC4J to actually be pooled under OC4J??
    Since I couldn't get this to work in my early 902 oc4j testing, and now can't get it to work
    still under 903 OC4J, either it's my config or BC4J AM's code or OC4J?
    Any thoughts on how to figure out what's not configed correctly or has a bug?
    Thanks, curt

  • 903/902/BC4J can't get OC4J data-sources.xml conn pooling to work in production: help

    [cross posted to the j2ee forum]
    I have several BC4J ears deployed to a 903 instance of OC4J being configured as a standalone
    instance. I've had this problem since I started deploying in development on 902. So it's
    some basic problem that I've not mastered.
    I can't get data-sources.xml managed connection pooling to actually pool conn's. I'm wanting
    to declare my jndi jdbc source connection pool in j2ee/home/config/data-sources.xml and
    have all BC4J apps get conns from this JNDI JDBC pool. I've removed all data-sources.xml from
    my BC4J ears, and published the jndi jdbc source in my oc4j common data-sources.xml.
    I've tested that this is the place controlling the conn URL/login passwd by commenting it
    out of config/data-sources.xml and my BC4J apps then throw exceptions, can't get conn.
    I've set the oc4j startup cmd line with the BC4J property to enabled connection pooling:
    -Djbo.doconnectionpooling=true
    symptom
    Connections are created and closed. Instead of being put back into the pool managed by oc4j,
    what ever BC4J is doing or my data-sources.xml is doing, the connections are just being created and
    closed.
    I can verify this via (solaris) lsof and netstat, where I see my oc4j instance under test load
    with only 1 or 2 conns to the db box, and the ephemeral port is tumbling, meaning a new socket is
    being opened for each conn. ;( grrrrrrr
    Does anyone have a clue as to why this is happening?
    Thanks, curt
    my data-sources.xml
    <data-sources>
         <data-source
            class="com.evermind.sql.DriverManagerDataSource"
            connection-driver="oracle.jdbc.driver.OracleDriver"
            ejb-location="jdbc/DEVDS"
            location="jdbc/DEVCoreDS"
            name="DEVDS"
            password="j2train"
            pooled-location="jdbc/DEVPooledDS"
            url="jdbc:oracle:thin:@10.2.1.30:1521:GDOC"
            username="jscribe"
            xa-location="jdbc/xa/DEVXADS"
            inactivity-timeout="300"
            max-connections="50"
            min-connections="40"
        />
    </data-sources>

    Thanks Leif,
    Yes, set it to the location jndi path.
    A piece of info is that the 903 oc4j release notes states that global conn pooling doesn't
    work. Infering that the j2ee/home/config/data-sources.xml data sources aren't pooled or ??
    I just tested so called local connection pooling, where I edited the data-sources.xml that
    gets packaged in the ear, to include the min/max params and re-ran my test.
    Still, the AM creates a new conn, it's to a new socket, and closes the conn when done. Causing
    each conn to not be pooled, rather opened then closed to the DB box. As verified with lsof and
    netstat, checking the ephemeral port # on the DB box side, always changes, meaning it's a
    new socket and not an old pooled conn socket.
    ???? What the heck??
    Surely if the AM conn check out / return code works properly, OC4J's pooling JDBC driver would
    pool and not close the socket??
    Has anywone gotten JDBC Datasource connections in BC4J to actually be pooled under OC4J??
    Since I couldn't get this to work in my early 902 oc4j testing, and now can't get it to work
    still under 903 OC4J, either it's my config or BC4J AM's code or OC4J?
    Any thoughts on how to figure out what's not configed correctly or has a bug?
    Thanks, curt

  • Shared-library in OC4J - load data-sources.xml throws error

    H
    To share common libraries (framework.jar, JAXB jars, axis.jar, log4j, tibco*.jars)
    across application(app1/app2/app3) in the OC4J container.
    Currently all these jars are loaded with common jars separately application, a total 3 common
    footprints of these jars.
    1 - Placed all common jars under $J2EE_HOME/shared-lib folder, updated server.xml in $J2EE_HOME/config.
    2 - Connecting Tibco queues using generic JMS resource adapter (connectors - JCA).
    3 - Connecting to databases using data-sources.xml.
    4 - With following entries in orion-application.xml
    <imported-shared-libraries>
    <import-shared-library name="my.shared.libraries" max-version="1.0"/>
    </imported-shared-libraries> ---> Entry 1
    <connectors path="oc4j-connectors.xml"/> ---> Entry 2
    <data-sources path="data-sources.xml"/> ---> Entry 3
    Getting the following error while deploying the code.
    Error occurred initializing connectors. Exception is: Exception creating Manage
    d DataSource ConnectorPropertySet. Exception: Error setting up resource adapter
    for embedded resource archive inside application 'deployGood1'. ManagedConnecti
    onFactory implementation class 'oracle.oc4j.sql.spi.ManagedConnectionFactoryImpl
    ' cannot be set up: Error setting JavaBean property 'managedDataSourceConfigXML'
    for ManagedConnectionFactory class 'oracle.oc4j.sql.spi.ManagedConnectionFactor
    yImpl for embedded resource archive inside application 'deployGood1'. Exception
    is: java.lang.reflect.InvocationTargetException
    However orion-application.xml
    Option A
    Entry 1 (imported-shared-libraries)
    Entry 2 (connectors)
    Works perfectly fine - however the Service complains of not finding the datasource JNDI
    5 - If all jars are loaded separately with each application
    With following entries in orion-application.xml
    <library path="C:/soft/jar/framework.jar" />
    <library path="C:/soft/jar/log4j-1.2.14.jar" />
    <library path="C:/soft/jar/axis.jar"/>
    <library path="C:/soft/jar/commons-discovery-0.2.jar"/>
    <library path="C:/soft/jar/commons-logging-1.0.4.jar"/>
    <library path="C:/soft/jar/fscontext.jar"/>
    <library path="C:/soft/jar/gjra.jar"/>
    <library path="C:/soft/jar/jax-qname.jar"/>
    <library path="C:/soft/jar/jaxb-api.jar"/>
    <library path="C:/soft/jar/jaxb-impl.jar"/>
    <library path="C:/soft/jar/jaxb-libs.jar"/>
    <library path="C:/soft/jar/jaxb-xjc.jar"/>
    <library path="C:/soft/jar/jaxrpc.jar"/>
    <library path="C:/soft/jar/jmxri.jar"/>
    <library path="C:/soft/jar/jmxtools.jar"/>
    <library path="C:/soft/jar/namespace.jar"/>
    <library path="C:/soft/jar/providerutil.jar"/>
    <library path="C:/soft/jar/relaxngDatatype.jar"/>
    <library path="C:/soft/jar/saaj.jar"/>
    <library path="C:/soft/jar/tibcrypt.jar"/>
    <library path="C:/soft/jar/tibjms.jar"/>
    <library path="C:/soft/jar/tibjmsadmin.jar"/>
    <library path="C:/soft/jar/tibjmsapps.jar"/>
    <library path="C:/soft/jar/tibrvjms.jar"/>
    <library path="C:/soft/jar/wsdl4j-1.5.1.jar"/>
    <library path="C:/soft/jar/wss4j-1.5.2.jar"/>
    <library path="C:/soft/jar/xercesImpl.jar"/>
    <library path="C:/soft/jar/xmlsec-1.4.0.jar"/>
    <library path="C:/soft/jar/xsdlib.jar"/> -------> Entry 1
    <connectors path="oc4j-connectors.xml"/> ---> Entry 2
    <data-sources path="data-sources.xml"/> ---> Entry 3
    This configuration works perfectly fine as expected, but this configuration defeats the
    purpose of shared-library
    Thanks
    sunder

    H
    To share common libraries (framework.jar, JAXB jars, axis.jar, log4j, tibco*.jars)
    across application(app1/app2/app3) in the OC4J container.
    Currently all these jars are loaded with common jars separately application, a total 3 common
    footprints of these jars.
    1 - Placed all common jars under $J2EE_HOME/shared-lib folder, updated server.xml in $J2EE_HOME/config.
    2 - Connecting Tibco queues using generic JMS resource adapter (connectors - JCA).
    3 - Connecting to databases using data-sources.xml.
    4 - With following entries in orion-application.xml
    <imported-shared-libraries>
    <import-shared-library name="my.shared.libraries" max-version="1.0"/>
    </imported-shared-libraries> ---> Entry 1
    <connectors path="oc4j-connectors.xml"/> ---> Entry 2
    <data-sources path="data-sources.xml"/> ---> Entry 3
    Getting the following error while deploying the code.
    Error occurred initializing connectors. Exception is: Exception creating Manage
    d DataSource ConnectorPropertySet. Exception: Error setting up resource adapter
    for embedded resource archive inside application 'deployGood1'. ManagedConnecti
    onFactory implementation class 'oracle.oc4j.sql.spi.ManagedConnectionFactoryImpl
    ' cannot be set up: Error setting JavaBean property 'managedDataSourceConfigXML'
    for ManagedConnectionFactory class 'oracle.oc4j.sql.spi.ManagedConnectionFactor
    yImpl for embedded resource archive inside application 'deployGood1'. Exception
    is: java.lang.reflect.InvocationTargetException
    However orion-application.xml
    Option A
    Entry 1 (imported-shared-libraries)
    Entry 2 (connectors)
    Works perfectly fine - however the Service complains of not finding the datasource JNDI
    5 - If all jars are loaded separately with each application
    With following entries in orion-application.xml
    <library path="C:/soft/jar/framework.jar" />
    <library path="C:/soft/jar/log4j-1.2.14.jar" />
    <library path="C:/soft/jar/axis.jar"/>
    <library path="C:/soft/jar/commons-discovery-0.2.jar"/>
    <library path="C:/soft/jar/commons-logging-1.0.4.jar"/>
    <library path="C:/soft/jar/fscontext.jar"/>
    <library path="C:/soft/jar/gjra.jar"/>
    <library path="C:/soft/jar/jax-qname.jar"/>
    <library path="C:/soft/jar/jaxb-api.jar"/>
    <library path="C:/soft/jar/jaxb-impl.jar"/>
    <library path="C:/soft/jar/jaxb-libs.jar"/>
    <library path="C:/soft/jar/jaxb-xjc.jar"/>
    <library path="C:/soft/jar/jaxrpc.jar"/>
    <library path="C:/soft/jar/jmxri.jar"/>
    <library path="C:/soft/jar/jmxtools.jar"/>
    <library path="C:/soft/jar/namespace.jar"/>
    <library path="C:/soft/jar/providerutil.jar"/>
    <library path="C:/soft/jar/relaxngDatatype.jar"/>
    <library path="C:/soft/jar/saaj.jar"/>
    <library path="C:/soft/jar/tibcrypt.jar"/>
    <library path="C:/soft/jar/tibjms.jar"/>
    <library path="C:/soft/jar/tibjmsadmin.jar"/>
    <library path="C:/soft/jar/tibjmsapps.jar"/>
    <library path="C:/soft/jar/tibrvjms.jar"/>
    <library path="C:/soft/jar/wsdl4j-1.5.1.jar"/>
    <library path="C:/soft/jar/wss4j-1.5.2.jar"/>
    <library path="C:/soft/jar/xercesImpl.jar"/>
    <library path="C:/soft/jar/xmlsec-1.4.0.jar"/>
    <library path="C:/soft/jar/xsdlib.jar"/> -------> Entry 1
    <connectors path="oc4j-connectors.xml"/> ---> Entry 2
    <data-sources path="data-sources.xml"/> ---> Entry 3
    This configuration works perfectly fine as expected, but this configuration defeats the
    purpose of shared-library
    Thanks
    sunder

Maybe you are looking for