Copy the JDBC data sources to another weblogic domain?!

we have a weblogic domain that has around 20 JDBC data sources
is there an easy way to copy the JDBC data sources to another weblogic domain?!

You can make a WLST script that creates the datasource(s) and target it to whatever servers/clusters you like,
but as implemented, for security reasons, certain aspects of the datasources themselves are encrypted to
operate only in their originating environment, so for instance, you cannot inadvertently email access to the
company's DBMS to your new friend in Nigeria.

Similar Messages

  • Standalone application can't get JDBC data source from Weblogic 10.3

    We have the following configuration :
    A Weblogic server 10.3 (default installation).
    The server contains a JMS queue (jndi name: "DMQ") and JDBC data sources (jndi names: Oracle thin XA - "dataSource", MS SQL - "dataSource1")
    We have built wlfullclient5.jar for Java 1.5 ([http://edocs.bea.com/wls/docs103/client/jarbuilder.html#wp1078122]) (according to the docs)
    And now we use a test standalone application with the wlfullclient5.jar :
    public static void main (String[] args) throws NamingException {
    bq. Hashtable<String, String> env = new Hashtable<String, String>(); \\     env.put(Context.+INITIAL_CONTEXT_FACTORY+, "weblogic.jndi.WLInitialContextFactory"); \\     env.put(Context.+PROVIDER_URL+, "http://serv1:7001"); \\     env.put(Context.+SECURITY_CREDENTIALS+, "weblogic"); \\     env.put(Context.+SECURITY_PRINCIPAL+, "weblogic"); \\     InitialContext ic = new InitialContext(env); \\ \\ System.+out+.println("Get DMQ"); \\     ic.lookup("DMQ"); \\ System.+out+.println("Get dataSource"); \\     ic.lookup("dataSource");
    bq. System.+out+.println("Get dataSource1"); \\     ic.lookup("dataSource1"); \\ System.+out+.println("Done"); \\
    Here is the output when connected to WLS 10.3:
    bq. Get DMQ \\ Get dataSource \\ Exception in thread "Main Thread" java.lang.AssertionError: Failed to generate class for weblogic.jdbc.common.internal.RmiDataSource_1030_WLStub \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:790_) \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:779_) \\ at weblogic.rmi.extensions.StubFactory.getStub(_StubFactory.java:74_) \\ at weblogic.rmi.internal.StubInfo.resolveObject(_StubInfo.java:213_) \\ at weblogic.rmi.internal.StubInfo.readResolve(_StubInfo.java:207_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke0(_Native Method_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke(_NativeMethodAccessorImpl.java:39_) \\ at sun.reflect.DelegatingMethodAccessorImpl.invoke(_DelegatingMethodAccessorImpl.java:25_) \\ at java.lang.reflect.Method.invoke(_Method.java:585_) \\ at java.io.ObjectStreamClass.invokeReadResolve(_ObjectStreamClass.java:1033_) \\ at java.io.ObjectInputStream.readOrdinaryObject(_ObjectInputStream.java:1728_) \\ at java.io.ObjectInputStream.readObject0(_ObjectInputStream.java:1305_) \\ at java.io.ObjectInputStream.readObject(_ObjectInputStream.java:348_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:197_) \\ at weblogic.rjvm.MsgAbbrevInputStream.readObject(_MsgAbbrevInputStream.java:564_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:193_) \\ at weblogic.rmi.internal.ObjectIO.readObject(_ObjectIO.java:62_) \\ at weblogic.rjvm.ResponseImpl.unmarshalReturn(_ResponseImpl.java:240_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:348_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:259_) \\ at weblogic.jndi.internal.ServerNamingNode_1030_WLStub.lookup(Unknown Source) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:392_) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:380_) \\ at javax.naming.InitialContext.lookup(_InitialContext.java:351_) \\ at test.main(_test.java:23_) \\ Caused by: java.lang.reflect.InvocationTargetException \\ at sun.reflect.NativeConstructorAccessorImpl.newInstance0(_Native Method_) \\ at sun.reflect.NativeConstructorAccessorImpl.newInstance(_NativeConstructorAccessorImpl.java:39_) \\ at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(_DelegatingConstructorAccessorImpl.java:27_) \\ at java.lang.reflect.Constructor.newInstance(_Constructor.java:494_) \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:788_) \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:779_) \\ at weblogic.rmi.extensions.StubFactory.getStub(_StubFactory.java:74_) \\ at weblogic.rmi.internal.StubInfo.resolveObject(_StubInfo.java:213_) \\ at weblogic.rmi.internal.StubInfo.readResolve(_StubInfo.java:207_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke0(_Native Method_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke(_NativeMethodAccessorImpl.java:39_) \\ at sun.reflect.DelegatingMethodAccessorImpl.invoke(_DelegatingMethodAccessorImpl.java:25_) \\ at java.lang.reflect.Method.invoke(_Method.java:585_) \\ at java.io.ObjectStreamClass.invokeReadResolve(_ObjectStreamClass.java:1033_) \\ at java.io.ObjectInputStream.readOrdinaryObject(_ObjectInputStream.java:1728_) \\ at java.io.ObjectInputStream.readObject0(_ObjectInputStream.java:1305_) \\ at java.io.ObjectInputStream.readObject(_ObjectInputStream.java:348_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:197_) \\ at weblogic.rjvm.MsgAbbrevInputStream.readObject(_MsgAbbrevInputStream.java:564_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:193_) \\ at weblogic.rmi.internal.ObjectIO.readObject(_ObjectIO.java:62_) \\ at weblogic.rjvm.ResponseImpl.unmarshalReturn(_ResponseImpl.java:240_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:348_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:259_) \\ at weblogic.jndi.internal.ServerNamingNode_1030_WLStub.lookup(Unknown Source) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:392_) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:381_) \\ at javax.naming.InitialContext.lookup(_InitialContext.java:351_) \\ at test.main(_test.java:26_) \\ Caused by: java.lang.ArrayIndexOutOfBoundsException: 6 \\ at weblogic.jdbc.common.internal.RmiDataSource_1030_WLStub.ensureInitialized(Unknown Source) \\ at weblogic.jdbc.common.internal.RmiDataSource_1030_WLStub.<init>(Unknown Source) \\ at sun.reflect.NativeConstructorAccessorImpl.newInstance0(_Native Method_) \\ at sun.reflect.NativeConstructorAccessorImpl.newInstance(_NativeConstructorAccessorImpl.java:39_) \\ at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(_DelegatingConstructorAccessorImpl.java:27_) \\ at java.lang.reflect.Constructor.newInstance(_Constructor.java:494_) \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:788_) \\ at weblogic.rmi.internal.StubGenerator.generateStub(_StubGenerator.java:779_) \\ at weblogic.rmi.extensions.StubFactory.getStub(_StubFactory.java:74_) \\ at weblogic.rmi.internal.StubInfo.resolveObject(_StubInfo.java:213_) \\ at weblogic.rmi.internal.StubInfo.readResolve(_StubInfo.java:207_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke0(_Native Method_) \\ at sun.reflect.NativeMethodAccessorImpl.invoke(_NativeMethodAccessorImpl.java:39_) \\ at sun.reflect.DelegatingMethodAccessorImpl.invoke(_DelegatingMethodAccessorImpl.java:25_) \\ at java.lang.reflect.Method.invoke(_Method.java:585_) \\ at java.io.ObjectStreamClass.invokeReadResolve(_ObjectStreamClass.java:1033_) \\ at java.io.ObjectInputStream.readOrdinaryObject(_ObjectInputStream.java:1728_) \\ at java.io.ObjectInputStream.readObject0(_ObjectInputStream.java:1305_) \\ at java.io.ObjectInputStream.readObject(_ObjectInputStream.java:348_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:197_) \\ at weblogic.rjvm.MsgAbbrevInputStream.readObject(_MsgAbbrevInputStream.java:564_) \\ at weblogic.utils.io.ChunkedObjectInputStream.readObject(_ChunkedObjectInputStream.java:193_) \\ at weblogic.rmi.internal.ObjectIO.readObject(_ObjectIO.java:62_) \\ at weblogic.rjvm.ResponseImpl.unmarshalReturn(_ResponseImpl.java:240_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:348_) \\ at weblogic.rmi.cluster.ClusterableRemoteRef.invoke(_ClusterableRemoteRef.java:259_) \\ at weblogic.jndi.internal.ServerNamingNode_1030_WLStub.lookup(Unknown Source) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:392_) \\ at weblogic.jndi.internal.WLContextImpl.lookup(_WLContextImpl.java:380_) \\ at javax.naming.InitialContext.lookup(_InitialContext.java:351_) \\ at test.main(_test.java:23_)
    But at the same time the output when connected to WLS 10.0 is :
    Get DMQ \\Get dataSource \\Get dataSource1 \\Done
    (so that the test passed)
    Could you give me a piece of advice ?
    Thanks,
    Sergey

    I hit the same problem as yours. This issue is caused by JDK version. The JDK used by Weblogic 10.3 is java 6, while your client program's jdk version is java5. So change the both of them to java 6. This issue will be fixed.
    Don't forget to rebuild the wlfullclient.jar which also should be java 6.
    Good luck!

  • How to copy the all data to another new table

    Hi,
    1) "How can I copy the data from one Data Dictionary Table to another new Table, in Dictionary"?
    for eg.  (<i>In Dictionary,  I have one table named 'sflight' and now I want to copy the all data from 'sflight' to another new table named 'zabc'</i>)
    How can I do this?
    2) What is Logical Unit of Work or LUW.
    Thanks

    HI,
    go to <b>se11</b>, give sflight (Your standard table name) in tables text,
    No go to dictionary objects --> copy --> give new <b>Ztable</b> in <b>to table</b> text.
    Like this you can copy the standard sap table to a new ztable..
    A Logical Unit of Work (LUW or database transaction) is an inseparable sequence of database operations which must be executed either in its entirety or not at all. For the database system, it thus constitutes a unit.
    LUWs help to guarantee database integrity. When an LUW has been successfully concluded, the database is once again in a correct state. If, however, an error occurs within an LUW, all database changes made since the beginning of the LUW are canceled and the database is then in the same state as before the LUW started.
    An LUW begins
    o each time you start a transaction
    o when the database changes of the previous LUW have been confirmed (database commit) or
    o when the database changes of the previous LUW have been cancelled (database rollback)
    An LUW ends
    o when the database changes have been confirmed (database commit) or
    o when the database changes have been canceled (database rollback)
    SAP memory and ABAP memoryUsing SAP memory. User-specific memory retained for the duration of the session. Can only be used for simple field values.
    Using ABAP memory. Can be used to transfer data between internal modi. Can be used to transfer any ABAP/4 variables ( Fields, strings, internal tables, complex objects)
    SAP memoryYou use the SPA/GPA parameters to pass data. These paramters are saved globally in memory. The paramters are indetified by a three-character code.
    In dialog programs you can SET and GET these parameters in the attribute window of the screen field, by marking the SET and GET fields and put the name of the parameter in the ParameterId field.
    In a program ( Dialog or Report ) you can use the GET PARAMETER and SET PARAMETER statements:
    set parameter id 'RV1' field <fieldname>
    get parameter id 'RV1' field <fieldname>
    rewards if useful,
    regards,
    nazeer

  • Using WLST to set up credential mapping on JDBC Data Source

    I am wanting to write a script (WLST) to set up a large initial set of credential mappings on a JDBC data source.
    I can create the datasource and do all the basic set ups, but I can't figure out how to do the credential mappings.
    I tried the "record" option from the Admin console, but it created a script that says that credential mappings are not recorded.
    This is the mapping function that you get to from the "Security" tab on the DataSource page, and then the "Credential Mappings" tab. This allows you to map a "WLS user" to a "remote user" - which in my case is an Oracle OLS user.
    I am using Weblogic 10.3
    Can anyone point me to a set of commands which can perform this function?
    Thanks!

    I wonder if there isn't a publicly visible MBean for the credential mapping. That would explain why the recorded script doesn't have the mbean method invocations.

  • WLS 10.3.4: How to use OS authentication for JDBC Data Source

    Hello all,
    As a preface, I've tried searching the forum/Google for "OS authentication" and reading the WLS JDBC doc to no avail - if it's documented somewhere, a RTFM link would be much appreciated.
    I'm trying to set up a JDBC data source on WLS that leverages the OS Authentication capability of the Oracle database. If it would help, I can go into the reasoning behind why I want to do this, but basically, it's to simplify the config/deployment of a COTS application. What I have in the database is an "identified externally" user that corresponds to the OS user that is running the WebLogic Server. Normally, in tools such as SQL*Plus, I would use "/@db" as the username/password (in other words, no username and no password specified), and I would be logged in as the "idenfitied externally" user. I want to configure the same thing for a WebLogic Data Source, but if I leave the username/password blank, testing the connection in the WLS console gives me "invalid username/password, login denied" I've also tried using "/" as the username, as was documented in a quite old WLS faq, but that gives me the same result.
    Is there some magic switch I need to flip?
    Thanks,
    John

    Hi John, there's no way to do that with connection pools, which is how WLS datasources get their
    connections, or middleware in general. WebLogic would have no way of knowing which if any of the
    pooled connections was appropriate for the current 'user', which is not the application user, but
    instead is the OS identity of the person who started the WebLogic server! If you start up your
    WebLogic server, and people start pointing their browsers to it, doing various stuff, the OS knows
    you started WebLogic, and maybe with the help of OCI, Oracle's JDBC might know it was you who
    started WebLogic's OS process, but what does the OS know about any user that may be running
    a browser or application elsewhere (even if on this same machine), when that browser or application
    connects to your WebLogic server process?
    HTH,
    Joe

  • JDBC Data Sources: Potential Issue with JDeveloper 10.1.3.4

    I think I found a bug or issue with the latest JDeveloper 10.1.3.4 release when using JDBC Data Sources on the Embedded OC4J container.
    To state the issule bluntly, If I use a JDBC Data Source in an ADF Faces application, I get the following error on the screen when I run my application if I create a simple page using a Form layout for database data:
    [http://cs.uwindsor.ca/~ruston7/jdbcError.jpg]
    Or if I use a simple drag and drop ADF Faces Table:
    javax.faces.el.PropertyNotFoundException: Error testing property '<<FIRST_FIELD_ON_THE_PAGE>>' in bean of type null
        at com.sun.faces.el.PropertyResolverImpl.isReadOnly(PropertyResolverImpl.java:274)
        at oracle.adfinternal.view.faces.model.FacesPropertyResolver.isReadOnly(FacesPropertyResolver.java:124)
        at com.sun.faces.el.impl.ArraySuffix.isReadOnly(ArraySuffix.java:236)
        at com.sun.faces.el.impl.ComplexValue.isReadOnly(ComplexValue.java:209)
        at com.sun.faces.el.ValueBindingImpl.isReadOnly(ValueBindingImpl.java:266)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.EditableValueRenderer.getReadOnly(EditableValueRenderer.java:211)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.FormElementRenderer.renderAsElement(FormElementRenderer.java:155)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.InputLabelAndMessageRenderer.getLabelFor(InputLabelAndMessageRenderer.java:53)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.LabelAndMessageRenderer$Label.getForId(LabelAndMessageRenderer.java:500)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.OutputLabelRenderer.encodeAll(OutputLabelRenderer.java:69)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.delegateRenderer(CoreRenderer.java:281)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.LabelAndMessageRenderer.encodeAll(LabelAndMessageRenderer.java:123)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.InputLabelAndMessageRenderer.encodeAll(InputLabelAndMessageRenderer.java:94)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.encodeEnd(CoreRenderer.java:169)
        at oracle.adf.view.faces.component.UIXComponentBase.encodeEnd(UIXComponentBase.java:624)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.encodeChild(CoreRenderer.java:246)When I change my Application Module connection to a JDBC URL, this all works perfectly. Also, everything works fine when I deploy to our Oracle AS 10.1.3 application servers.
    I also tried this on a different computer using a fresh install of JDeveloper just to make sure that the copy of JDeveloper that I downloaded didn't have a fluke in it.
    Thanks!

    M. Ruston,
    It must be something on your side. i just tried the same thing using the employees table from the HR sample schema (it has a date column). It works with JDBC URL and datasource both.
    Just out of curiosity - if you look at the properties for your Business Components project at the business components section, what does it show for the SQL Flavor and Type Map?
    John

  • Null pointer in Data Source deployment in WebLogic 8.1

    I'm using WebLogic 8.1 in Eclipse 3.0.
    In the Weblogic console, my datasource tests out fine. But when I startup the WebLogic Server in Eclipse I get the error:
    Data Source "Datasource-Secondary" deployment failed with the following error: null.
    Here are the details and Stack Trace:
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001177> <Creating Connection Pool named Pool-XI5G, URL = jdbc:db2://z555.mycomputer.com:5054/XI5G, Properties = {user=MYUSERID, driverType=4, readOnly=true}.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000626> <Free resources in pool "Pool-XI5G" will be tested every "600" seconds.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001068> <Connection for pool "Pool-XI5G" created.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001132> <Initialized statement cache of size "100" for connection in pool "Pool-XI5G".>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000628> <Created "1" resources for pool "Pool-XI5G", out of which "1" are available and "0" are unavailable.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001124> <Created Connection Pool named Pool-XI5G.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001177> <Creating Connection Pool named Pool-XT5G, URL = jdbc:db2://z555.mycomputer.com:5051/XT5G, Properties = {user=MYUSERID, driverType=4, readOnly=true}.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000626> <Free resources in pool "Pool-XT5G" will be tested every "600" seconds.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001068> <Connection for pool "Pool-XT5G" created.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001132> <Initialized statement cache of size "100" for connection in pool "Pool-XT5G".>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000628> <Created "1" resources for pool "Pool-XT5G", out of which "1" are available and "0" are unavailable.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001124> <Created Connection Pool named Pool-XT5G.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001175> <Creating TX Data Source named jdbc/ODR for Pool MultiPool-Primary.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Error> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001151> <Data Source "Datasource-Primary" deployment failed with the following error: null.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001156> <Stack trace associated with message 001151 follows:
    weblogic.common.ResourceException
         at weblogic.jdbc.common.internal.DataSourceManager.createDataSource(DataSourceManager.java:254)
         at weblogic.jdbc.common.internal.DataSourceManager.createAndStartDataSource(DataSourceManager.java:109)
         at weblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:192)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:331)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployments(DeploymentTarget.java:591)
         at weblogic.management.mbeans.custom.DeploymentTarget.updateServerDeployments(DeploymentTarget.java:569)
         at weblogic.management.mbeans.custom.DeploymentTarget.updateDeployments(DeploymentTarget.java:241)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:324)
         at weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl.java:731)
         at weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:710)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:484)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.private_invoke(RemoteMBeanServerImpl.java:985)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:943)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:946)
         at weblogic.management.internal.MBeanProxy.invokeForCachingStub(MBeanProxy.java:481)
         at weblogic.management.configuration.ServerMBean_Stub.updateDeployments(ServerMBean_Stub.java:7351)
         at weblogic.management.deploy.slave.SlaveDeployer.updateServerDeployments(SlaveDeployer.java:1304)
         at weblogic.management.deploy.slave.SlaveDeployer.resume(SlaveDeployer.java:347)
         at weblogic.management.deploy.DeploymentManagerServerLifeCycleImpl.resume(DeploymentManagerServerLifeCycleImpl.java:229)
         at weblogic.t3.srvr.SubsystemManager.resume(SubsystemManager.java:131)
         at weblogic.t3.srvr.T3Srvr.resume(T3Srvr.java:966)
         at weblogic.t3.srvr.T3Srvr.run(T3Srvr.java:361)
         at weblogic.Server.main(Server.java:32)
    >
    Any idea?
    - SB

    Steve Bialek wrote:
    I'm using WebLogic 8.1 in Eclipse 3.0.
    In the Weblogic console, my datasource tests out fine. But when I startup the WebLogic Server in Eclipse I get the error:
    Data Source "Datasource-Secondary" deployment failed with the following error: null.
    Here are the details and Stack Trace:
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001177> <Creating Connection Pool named Pool-XI5G, URL = jdbc:db2://z555.mycomputer.com:5054/XI5G, Properties = {user=MYUSERID, driverType=4, readOnly=true}.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000626> <Free resources in pool "Pool-XI5G" will be tested every "600" seconds.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001068> <Connection for pool "Pool-XI5G" created.>
    ####<Feb 28, 2006 3:13:09 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001132> <Initialized statement cache of size "100" for connection in pool "Pool-XI5G".>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000628> <Created "1" resources for pool "Pool-XI5G", out of which "1" are available and "0" are unavailable.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001124> <Created Connection Pool named Pool-XI5G.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001177> <Creating Connection Pool named Pool-XT5G, URL = jdbc:db2://z555.mycomputer.com:5051/XT5G, Properties = {user=MYUSERID, driverType=4, readOnly=true}.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000626> <Free resources in pool "Pool-XT5G" will be tested every "600" seconds.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001068> <Connection for pool "Pool-XT5G" created.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001132> <Initialized statement cache of size "100" for connection in pool "Pool-XT5G".>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <Common> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-000628> <Created "1" resources for pool "Pool-XT5G", out of which "1" are available and "0" are unavailable.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001124> <Created Connection Pool named Pool-XT5G.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001175> <Creating TX Data Source named jdbc/ODR for Pool MultiPool-Primary.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Error> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001151> <Data Source "Datasource-Primary" deployment failed with the following error: null.>
    ####<Feb 28, 2006 3:13:10 PM PST> <Info> <JDBC> <MYMACHINE_NAME> <myserver> <main> <<WLS Kernel>> <> <BEA-001156> <Stack trace associated with message 001151 follows:
    weblogic.common.ResourceException
         at weblogic.jdbc.common.internal.DataSourceManager.createDataSource(DataSourceManager.java:254)
         at weblogic.jdbc.common.internal.DataSourceManager.createAndStartDataSource(DataSourceManager.java:109)
         at weblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:192)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:331)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployments(DeploymentTarget.java:591)
         at weblogic.management.mbeans.custom.DeploymentTarget.updateServerDeployments(DeploymentTarget.java:569)
         at weblogic.management.mbeans.custom.DeploymentTarget.updateDeployments(DeploymentTarget.java:241)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:324)
         at weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl.java:731)
         at weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:710)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:484)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.private_invoke(RemoteMBeanServerImpl.java:985)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:943)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:946)
         at weblogic.management.internal.MBeanProxy.invokeForCachingStub(MBeanProxy.java:481)
         at weblogic.management.configuration.ServerMBean_Stub.updateDeployments(ServerMBean_Stub.java:7351)
         at weblogic.management.deploy.slave.SlaveDeployer.updateServerDeployments(SlaveDeployer.java:1304)
         at weblogic.management.deploy.slave.SlaveDeployer.resume(SlaveDeployer.java:347)
         at weblogic.management.deploy.DeploymentManagerServerLifeCycleImpl.resume(DeploymentManagerServerLifeCycleImpl.java:229)
         at weblogic.t3.srvr.SubsystemManager.resume(SubsystemManager.java:131)
         at weblogic.t3.srvr.T3Srvr.resume(T3Srvr.java:966)
         at weblogic.t3.srvr.T3Srvr.run(T3Srvr.java:361)
         at weblogic.Server.main(Server.java:32)
    Any idea?
    - SBI'd like to see the config.xml file, and also if yo could turn on JDBC logging,
    and show me the jdbc.log file? From the line numbers in the stacktrace, I am
    guessing this is 8.1sp4. If that's correct, can you take this attached debug
    jar, and then edit your start-weblogic script to add this jar in front of
    the CLASSPATH the script creates, just before calling java. Reboot, reproduce
    the problem, and the stack trace printed out will be more understandable.
    thanks
    Joe
    thanks
    Joe

  • Need Help on using CAS Incremental Crawl with JDBC data source

    Hi,
    As part of one of the e-commerce implementations, we are implementing a delta pipeline which reads the data from the database views. We plan to crawl the data with the help of CAS JDBC data source. In order to optimize the data reads, we want to use the CAS capabilities of incremental crawl. We rely on CAS to be able to identify the updates available in the subsequent incremental crawl and only read the updated data for any given crawl unless we force a full crawl.
         We have tried implementing the above setup using JDBC data source. CAS reads from the database and stores it in the record store. The Full crawl works fine however; we observed some unexpected behavior during the incremental crawl run. Even when there is no change in the database, crawl metrics show that certain number of records have been updated in the record store and the number of updates differ in every subsequent run.
    Any pointers what can be the issue ? Does CAS has the incremental crawl capability using JDBC data source ?
    Regards,
    Nitin Malhotra

    Hi Ravi,
    Generic Extraction is used to extract data from COPA tables. And delta method used to extract delta records (records created after initilization) is timestamp.
    What is this timestamp?
    Assume that, you have posted 10 records one after the other to the COPA tables. And we do initilization to move all these 10 records to BW system. Later another 5 records are added to COPA tables. How do you think the system identifies these new 5 records (delta records).?
    It identifies based on a timestamp field (eg : Document created on ,a 16 digit decimal field).
    Assume that, in our previous initilzation, say  "document created on" field for the last (or the latest) record is 14/11/2006, 18.00 hrs. and  timestamp is set to the 14/11/2006, 18.00hrs.  then when you do delta update, the system treats all the records with "document created on" field is greater than 14/11/2006, 18.00 hrs as delta records. This is how new 5 records are extracted to bw system and timestamp is again set to the new value besed in field in the records.(say 14/11/2006, 20.00hrs)
    Assume that, you have two records with "document created on" field as 14/11/2006, 17.55hrs and 14/11/2006, 17.57hrs. and they were updated to the COPA table after 14/11/2006,20.00hrs (at this time last delta load is done)  due to some reason. How can you bring these two records? For this, purpose we can reset the timestamp at KEB5. In this eg, we can reset as 14/11/2006, 17.50 hrs) and do the delta again. So, the system picks all the records which are posted after 14/11/2006,17.50hrs again.  But remember that, doing this, you sending some of the records again (duplicate records). So, make sure that you sending ODS object. otherwise you end up with inconsistant data due to duplicate records.
    Hope this helps!!
    Vj

  • Getting JDBC data source from PersistenceUnit

    I have an application accessing a database, and am using JPA to read/write/update records in the database.
    Some of the database tables use stored procedures to perform complex computations across entire
    database tables, and I need to call these stored procedures when I the user does certain operations,
    such as cloning an object and all its associated data. To call the stored procedures I use the same
    JDBC data source that is used to define the PersistenceUnit.
    Is there any way to get that DataSource from the PersistenceUnit?
    Since I do not know how to do this I am currently using @Resource DataSource annotations to get the JDBC
    datasource. That is ugly because I have the datasource specified in two different places in my code.
    Please help!

    I hit the same problem as yours. This issue is caused by JDK version. The JDK used by Weblogic 10.3 is java 6, while your client program's jdk version is java5. So change the both of them to java 6. This issue will be fixed.
    Don't forget to rebuild the wlfullclient.jar which also should be java 6.
    Good luck!

  • Difference between DataSource & JDBC Data Source Factory

    Hi,
    Can anyone tell me what the difference between a 'DataSource' and 'JDBC Data Source
    Factory' is ? I see both these on the Weblogic admin console. I have reviewed
    the documentation but it is not clear to me what the difference between the two
    is and what situations each on is appropriate for.
    Thanks,
    Raju

    Hi Sree,
    I have read literally ALL the online documents related to JDBC Data Source Factories
    including the one you provided below, but I still do not understand what they
    are for. I have also searched in Google but to no avail.
    I will try to be more specific as to where my confusion lies.
    1)Why would anyone ever choose a JDBCData Source Factory instead of a Tx Data
    Source or plain vanilla (not XA) Data Source? EJBs connect just fine with the
    latter 2.
    2)What should go inside the properties box?
    a) http://e-docs.bea.com/wls/docs70/ejb/EJB_environment.html#resourcefact states
    that connection properties should go in here
    b)Then, the document goes on to describe a syntax for binding a Connection Pool
    to a JNDI name, all in the properties box. I have never been able to successfully
    create a data resource factory with a JNDI name. I get no errors, but it appears
    no where in the JNDI tree
    c)If this mysterious syntax associates a resource factory with an existing Connection
    Pool, then why go through all the fuss of redefining the Driver Class Name, URL,
    etc., when those are already defined on the Connection Pool?
    When you setup either a Data Source or a TX Data Source, WebLogic does not require
    that information since it was all setup in the pool. Why is it needed here?
    Thanks in advance for you help.
    "Sree Bodapati" <[email protected]> wrote:
    Hi Raju,
    Please see,
    http://e-docs.bea.com/wls/docs70/ConsoleHelp/domain_jdbcdatasourcefactorytable.html
    sree

  • Excel SSAS Tabular error: An error occurred during an attempt to establish a connection to the external data source

    Hello there,
    I have an Excel report I created which works perfectly fine on my dev environment, but fails on my test environment when I try to do a data refresh.
    The key difference between both dev and test environments is that in dev, everything is installed in one server:
    SharePoint 2013
    SQL 2012: Database Instance, SSAS Instance, SSRS for SharePoint, SSAS POWERPIVOT instance (Powerpivot for SharePoint).
    In my test and production environments, the architecture is different:
    SQL DB Servers in High Availability (irrelevant for this report since it is connecting to the tabular model, just FYI)
    SQL SSAS Tabular server (contains a tabular model that processes data from the SQL DBs).
    2x SharePoint Application Servers (we installed both SSRS and PowerPivot for SharePoint on these servers)
    2x SharePoint FrontEnd Servers (contain the SSRS and PowerPivot add-ins).
    Now in dev, test and production, I can run PowerPivot reports that have been created in SharePoint without any issues. Those reports can access the SSAS Tabular model without any issues, and perform data refresh and OLAP functions (slicing, dicing, etc).
    The problem is with Excel reports (i.e. .xlsx files) uploaded to SharePoint. While I can open them, I am having a hard time performing a data refresh. The error I get is:
    "An error occurred during an attempt to establish a connection to the external data source [...]"
    I ran SQL Profiler on my SSAS Server where the Tabular instance is and I noticed that every time I try to perform a data refresh, I get the following entries:
    Every time I try to perform a data refresh, two entries under the user name ANONYMOUS LOGON.
    Since things work without any issues on my single-server dev environment, I tried running SQL Server Profiler there as well to see what I get.
    As you can see from the above, in the dev environment the query runs without any issues and the user name logged is in fact my username from the dev environment domain. I also have a separated user for the test domain, and another for the production domain.
    Now upon some preliminary investigation I believe this has something to do with the data connection settings in Excel and the usage (or no usage) of secure store. This is what I can vouch for so far:
    Library containing reports is configured as trusted in SharePoint Central Admin.
    Library containing data connections is configured as trusted in SharePoint Central Admin.
    The Data Provider referenced in the Excel report (MSOLAP.5) is configured as trusted in SharePoint Central Admin.
    In the Excel report, the Excel Services authentication settings is set as "use authenticated user's account". This wortks fine in the DEV environment.
    Concerning SecureStore, PowerPivot Configurator has configured it the PowerPivotUnnattendedAccount application ID in all the environments. There is
    NO configuration of an Application ID for Excel Services in any of the environments (Dev, test or production). Altough I reckon this is where the solution lies, I am not 100% sure as to why it fails in test and prod. But as I read what I am
    writing, I reckon this is because of the authentication "hops" through servers. Am I right in my assumption?
    Could someone please advise what am I doing wrong in this case? If it is the fact that I am missing an Secure Store entry for Excel Services, I am wondering if someone could advise me on how to set ip up? My confusion is around the "Target Application
    Type" setting.
    Thank you for your time.
    Regards,
    P.

    Hi Rameshwar,
    PowerPivot workbooks contain embedded data connections. To support workbook interaction through slicers and filters, Excel Services must be configured to allow external data access through embedded connection information. External data access is required
    for retrieving PowerPivot data that is loaded on PowerPivot servers in the farm. Please refer to the steps below to solve this issue:
    In Central Administration, in Application Management, click Manage service applications.
    Click Excel Services Application.
    Click Trusted File Location.
    Click http:// or the location you want to configure.
    In External Data, in Allow External Data, click Trusted data connection libraries and embedded.
    Click OK.
    For more information, please see:
    Create a trusted location for PowerPivot sites in Central Administration:
    http://msdn.microsoft.com/en-us/library/ee637428.aspx
    Another reason is Excel Services returns this error when you query PowerPivot data in an Excel workbook that is published to SharePoint, and the SharePoint environment does not have a PowerPivot for SharePoint server, or the SQL Server Analysis
    Services (PowerPivot) service is stopped. Please check this document:
    http://technet.microsoft.com/en-us/library/ff487858(v=sql.110).aspx
    Finally, here is a good article regarding how to troubleshoot PowerPivot data refresh for your reference. Please see:
    Troubleshooting PowerPivot Data Refresh:
    http://social.technet.microsoft.com/wiki/contents/articles/3870.troubleshooting-powerpivot-data-refresh.aspx
    Hope this helps.
    Elvis Long
    TechNet Community Support

  • Can we set the dynamic data source when using getReportParameters() ?

    Hello!
    I have a report where one of its parameters refers to a list of values (LOVs). This list of values is an SQL Query type. When the data source used in the report is defined in the BI Publisher server, I'm able to get the report parameters using the getReportParameters() function in my application. However, if the data source is not defined the function throws an exception, which is understandable.
    I decided to dynamically set the data source so that even if the data source used by the report is not defined in the BI Publisher server, it still will be able to get the LOVs for the parameter. I tried setting the JDBCDataSource of the dynamicDataSource for the ReportRequest object that I passed to the getReportParameters() function. Please see the sample code below:
    reportRequest.dynamicDataSource = new BIP10.BIPDataSource();
    reportRequest.dynamicDataSource.JDBCDataSource = new BIP10.JDBCDataSource();
    setReportDataSource(reportRequest.dynamicDataSource.JDBCDataSource, connectstr, jdbc, dc); //function to set the values for JDBCDataSource object
    reportParams = webrs.getReportParameters(reportRequest, uid, pwd); //call the getReportParameters
    I was expecting this to work as this is what I did to dynamically set the data source before calling the runReport function. So, my question is -- can we set the dynamic data source when using getReportParameters() ? I tried this both in versions 10g and 11g. It does not seem to work on both versions.
    Regards,
    Stephanie

    report_id column of apex_application_page_ir_rpt can help us uniquely identify each saved report.
    We can assign this report_id value to a page item and this page item can be put in the Report ID Item text box of the Advanced section of the Report Attributes page of the IR.
    This should load the saved report identified by report_id and you can get rid of javascript
    Regards,
    Vishal
    http://obiee-oracledb.blogspot.com
    http://www.packtpub.com/oracle-apex-4-2-reporting/book
    Kindly mark the reply as helpful/correct if it solves your problem

  • How to set JDBC Data Sources in Oracle MapViewer for Oracle database 12c Release 1 (12.1.0.1)

    How to set JDBC Data Sources in Oracle MapViewer for Oracle database 12c Release 1 (12.1.0.1)?
    The following is my configuration in the conf\mapViewerConfig.xml:
    <map_data_source name="mvdemo12"
    jdbc_host="127.0.0.1"
    jdbc_sid="orcl12c1"
    jdbc_port="1522"
    jdbc_user="mvdemo"
    jdbc_password="7OVl2rJ+hOYxG5T3vKJQb+hW4NPgy9EN"
    jdbc_mode="thin"
    number_of_mappers="3"
    allow_jdbc_theme_based_foi="true"
    editable="true"/>
    <!--  ****  -->
    But it does not work.
    After use "sqlplus mvdemo/[email protected]:1522/pdborcl", it connected to the Oracle database 12c.
    Does anyone know it?
    Thanks,

    For 11.1.1.7.1 use the syntax for jdbc_sid, i.e.
    //mypdb1.foo.com as described in the README,
    - MapViewer native (non-container) data sources can now use database service name in place of SID. To supply a db service name, you will use the same jdbc_sid attribute, but specify the service name with double slashes in front, such as follows:
      <map_data_source name="myds"
        jdbc_host="foo.com"
        jdbc_sid="//mypdb1.foo.com"
        jdbc_port="1522"
      />
    For 11.1.1.7.0 use a container_ds instead.
    i.e. instead of using
    <map_data_source name="my_12c_test"
                       jdbc_host="mydbinstance"
                       jdbc_sid="pdborcl12c"
                       jdbc_port="1522"
                       jdbc_user="mytestuser"
                       jdbc_password="m2E7T48U3LfRjKwR0YFETQcjNb4gCMLG8/X0KWjO00Q="
                       jdbc_mode="thin"
                       number_of_mappers="6"
                       allow_jdbc_theme_based_foi="false"
                       editable="false"
       />
    use
      <map_data_source name="my_12c_test"
                       container_ds="jdbc/db12c"
                       number_of_mappers="6"
                       allow_jdbc_theme_based_foi="false"
                       editable="false"
       />
    In my case the Glassfish 3.1.2.2 JDBC connection pool definition was
    Property
    url  jdbc:oracle:thin:@mydbinstance:1522/pdborcl12c.rest_of.service.name
    Uncheck the Wrap JDBC Objects option in Advanced panel, i.e. the Edit JDBC Connection Pool Advanced properties page.
    Add a JDBC resource for that newly created pool
    Use that in mapviewerconfig.xml as above

  • Using JDBC Data Sources with ADFBC, NoInitialContextException

    Using JDBC Data Sources with ADF Business Components, NoInitialContextException
    I follow the instruction in the link below to create an ADF Swing application using datasource. I am using JDeveloper version 10.1.3.
    http://www.oracle.com/technology/products/jdev/howtos/10g/usingdatasources/using_datasources.html
    The ADF generated code looks like this:
    JUMetaObjectManager.setErrorHandler(new JUErrorHandlerDlg());
    JUMetaObjectManager mgr = JUMetaObjectManager.getJUMom();
    mgr.setJClientDefFactory(null);
    BindingContext ctx = new BindingContext();
    ctx.put(DataControlFactory.APP_PARAM_ENV_INFO, new JUEnvInfoProvider());
    ctx.setLocaleContext(new DefLocaleContext(null));
    HashMap map = new HashMap(4);
    map.put(DataControlFactory.APP_PARAMS_BINDING_CONTEXT, ctx);
    mgr.loadCpx("datasource.view.DataBindings.cpx" , map);
    final FormMain frame = new FormMain();
    frame.setBindingContext(ctx);
    I got this error when executing the last line: frame.setBindingContext(ctx);
    (oracle.jbo.common.ampool.ApplicationPoolException) JBO-30003: The application pool (datasource.datamodel.AppModuleDS) failed to checkout an application module due to the following exception:
    ----- LEVEL 1: DETAIL 0 -----
    (oracle.jbo.JboException) JBO-29000: Unexpected exception caught: oracle.jbo.DMLException, msg=JBO-27200: JNDI failure. Unable to lookup Data Source at context jdbc/xe_hrDS
    ----- LEVEL 2: DETAIL 0 -----
    (oracle.jbo.DMLException) JBO-27200: JNDI failure. Unable to lookup Data Source at context jdbc/xe_hrDS
    ----- LEVEL 3: DETAIL 0 -----
    (javax.naming.NoInitialContextException) Need to specify class name in environment or system property, or as an applet parameter, or in an application resource file: java.naming.factory.initial
    If I configure the application module connection type as JDBC URL, everything works.
    If the connection type is JDBC Datasource, I got the above error.
    Can someone show me how to adjust the generated code by ADF to use datasource?

    ADF BC has a bug. With Data Source in Application Module, application module does not connect. Use JDBC Connection URL.
    Also refer
    ADF BC: JDBC URL vs JDBC DataSource

  • How to change the jta-data-source for projects in EJB 3

    Hi all,
    Sorry my english! :)
    I'm learning the EJB 3.0 and I'd like to change jta-data-source for projects in JDeveloper 10.1.3.1.
    I explain:
    In the same Application, I have a project with all Entitys and Session Beans. At the moment, the persistence.xml file is there with the jta-data-source tag.
    I make a new project with JUnit (Cactus) for test the EJB's that I'd like jta-data-source reference the test database.
    I make another project, with Servlet/JSP/JSF that use EJB's with interface and I'd like jta-data-source reference another database.
    What kind of changes (workaround) I could make to solve this problem?
    Thanks a lot!
    Luciano

    Hi,
    There are COOM data source which may suits your needs.
    please check data source
    0CO_OM_WBS_6
    0CO_OM_WBS_3
    Base tables are COSP and COSS.
    Check about coom data source, if not suitable then go for custom data source.
    About table cross verify with your source team.
    Thanks

Maybe you are looking for

  • Trying to move my iPhoto library to external drive

    I'm trying to clear up room on my internal drive and so trying to move my library to an external drive.  I've copied the photo library to the external drive.  How do i point the iPhoto application to this new location so that I can delete the library

  • Sync/async bridge with mapping

    Hello togehter, i have to define a bridge between an sync RFC an an async DB Receiver. Thus i also need a mapping since the message types are different. But where in the BPM Flow can i add the interface mapping?

  • Itunes library.itl can't be opened

    when i want to open itunes, the messages "itunes library.itl can't be opned since it had been configuered by a newer itunes version (or similar, it's written in German) does anybody know anything about this?

  • Room Booking - unavailable if start / end time is :15 or :45

    Hi all, First apologies if this is in the wrong forum. Due to the way our school day runs, we need to be able to book rooms from :15 and :45 (e.g. 10.15 to 11.15, 11.45 to 12.45). You can change the start / end time manually but if you do the rooms d

  • Black spot on the center screen

    Hi I have baught blackberry curve 9300 from Bahrain Airport last week.. now my berry shows a black spot on the center in 1.5 Cm lenght its visible when screen is dark but the white screen doesnt shows clearely. I am afraid whether any daamage happend