How-to use custom data sources with OracleAS TopLink

Does anyone know if there is an updated version of this dcoument ?
I am having difficulties attempting to extend OracleAS TopLink to Support Custom Type Conversions.
Some of the links on this tutorial are broken
http://www.oracle.com/technology/products/ias/toplink/technical/tips/customconversion/index.html
Please help
Thanks,
John

Doug,
I may have been making a mountain out of a molehill.....
My problem while generating classes & descriptors was that Numbers from my Tables were being converted to BigDecimals.
I want these to be converted to Integers automatically as I am importing from approx 100 tables & don't want to update each mapping.
Is there a quick & easy way to achieve this ?
(I'm not being lazy & have spent 1/2 a day attempting to find this solution)
Many thanks,
John

Similar Messages

  • How to set JDBC Data Sources in Oracle MapViewer for Oracle database 12c Release 1 (12.1.0.1)

    How to set JDBC Data Sources in Oracle MapViewer for Oracle database 12c Release 1 (12.1.0.1)?
    The following is my configuration in the conf\mapViewerConfig.xml:
    <map_data_source name="mvdemo12"
    jdbc_host="127.0.0.1"
    jdbc_sid="orcl12c1"
    jdbc_port="1522"
    jdbc_user="mvdemo"
    jdbc_password="7OVl2rJ+hOYxG5T3vKJQb+hW4NPgy9EN"
    jdbc_mode="thin"
    number_of_mappers="3"
    allow_jdbc_theme_based_foi="true"
    editable="true"/>
    <!--  ****  -->
    But it does not work.
    After use "sqlplus mvdemo/[email protected]:1522/pdborcl", it connected to the Oracle database 12c.
    Does anyone know it?
    Thanks,

    For 11.1.1.7.1 use the syntax for jdbc_sid, i.e.
    //mypdb1.foo.com as described in the README,
    - MapViewer native (non-container) data sources can now use database service name in place of SID. To supply a db service name, you will use the same jdbc_sid attribute, but specify the service name with double slashes in front, such as follows:
      <map_data_source name="myds"
        jdbc_host="foo.com"
        jdbc_sid="//mypdb1.foo.com"
        jdbc_port="1522"
      />
    For 11.1.1.7.0 use a container_ds instead.
    i.e. instead of using
    <map_data_source name="my_12c_test"
                       jdbc_host="mydbinstance"
                       jdbc_sid="pdborcl12c"
                       jdbc_port="1522"
                       jdbc_user="mytestuser"
                       jdbc_password="m2E7T48U3LfRjKwR0YFETQcjNb4gCMLG8/X0KWjO00Q="
                       jdbc_mode="thin"
                       number_of_mappers="6"
                       allow_jdbc_theme_based_foi="false"
                       editable="false"
       />
    use
      <map_data_source name="my_12c_test"
                       container_ds="jdbc/db12c"
                       number_of_mappers="6"
                       allow_jdbc_theme_based_foi="false"
                       editable="false"
       />
    In my case the Glassfish 3.1.2.2 JDBC connection pool definition was
    Property
    url  jdbc:oracle:thin:@mydbinstance:1522/pdborcl12c.rest_of.service.name
    Uncheck the Wrap JDBC Objects option in Advanced panel, i.e. the Edit JDBC Connection Pool Advanced properties page.
    Add a JDBC resource for that newly created pool
    Use that in mapviewerconfig.xml as above

  • Using JDBC Data Sources with ADFBC, NoInitialContextException

    Using JDBC Data Sources with ADF Business Components, NoInitialContextException
    I follow the instruction in the link below to create an ADF Swing application using datasource. I am using JDeveloper version 10.1.3.
    http://www.oracle.com/technology/products/jdev/howtos/10g/usingdatasources/using_datasources.html
    The ADF generated code looks like this:
    JUMetaObjectManager.setErrorHandler(new JUErrorHandlerDlg());
    JUMetaObjectManager mgr = JUMetaObjectManager.getJUMom();
    mgr.setJClientDefFactory(null);
    BindingContext ctx = new BindingContext();
    ctx.put(DataControlFactory.APP_PARAM_ENV_INFO, new JUEnvInfoProvider());
    ctx.setLocaleContext(new DefLocaleContext(null));
    HashMap map = new HashMap(4);
    map.put(DataControlFactory.APP_PARAMS_BINDING_CONTEXT, ctx);
    mgr.loadCpx("datasource.view.DataBindings.cpx" , map);
    final FormMain frame = new FormMain();
    frame.setBindingContext(ctx);
    I got this error when executing the last line: frame.setBindingContext(ctx);
    (oracle.jbo.common.ampool.ApplicationPoolException) JBO-30003: The application pool (datasource.datamodel.AppModuleDS) failed to checkout an application module due to the following exception:
    ----- LEVEL 1: DETAIL 0 -----
    (oracle.jbo.JboException) JBO-29000: Unexpected exception caught: oracle.jbo.DMLException, msg=JBO-27200: JNDI failure. Unable to lookup Data Source at context jdbc/xe_hrDS
    ----- LEVEL 2: DETAIL 0 -----
    (oracle.jbo.DMLException) JBO-27200: JNDI failure. Unable to lookup Data Source at context jdbc/xe_hrDS
    ----- LEVEL 3: DETAIL 0 -----
    (javax.naming.NoInitialContextException) Need to specify class name in environment or system property, or as an applet parameter, or in an application resource file: java.naming.factory.initial
    If I configure the application module connection type as JDBC URL, everything works.
    If the connection type is JDBC Datasource, I got the above error.
    Can someone show me how to adjust the generated code by ADF to use datasource?

    ADF BC has a bug. With Data Source in Application Module, application module does not connect. Use JDBC Connection URL.
    Also refer
    ADF BC: JDBC URL vs JDBC DataSource

  • How to use JavaMail 1.4 with Oracle Application Server 10g (9.0.4.0.0)

    Hi all,
    I'd like to know if it's possible and how to use JavaMail 1.4 with Oracle Application Server 10g (9.0.4.0.0), Windows version.
    With the following code, I can see that the mail.jar used by the server is the one included in the jdk installation :
    // I'm testing InternetAddress.class because I want to use commons-email-1.2.jar that requires mail.jar 1.4 (or higher) and activation.jar 1.1 (or higher)
    // and I know that inside the commons-email-1.2.jar file, I need to call the InternetAddress.validate() method that throws a java.lang.NoSuchMethodError: javax.mail.internet.InternetAddress.validate()V if it is used with mail.jar 1.2.
    Class cls = javax.mail.internet.InternetAddress.class;
    java.security.ProtectionDomain pDomain = cls.getProtectionDomain();
    java.security.CodeSource cSource = pDomain.getCodeSource();
    java.net.URL location = cSource.getLocation();
    System.out.println(location.toString());
    This code returns : file:/C:/oracle/app/jdk/jre/lib/ext/mail.jar and this mail.jar file has an implementation version number: 1.2
    - I've tried to include my own mail.jar (1.4.2) and activation.jar (1.1.1) files in the war file that I deploy, but it doesn't work (the server still uses the same mail.jar 1.2)
    - I've tried to put the mail.jar (1.4.2) and activation.jar (1.1.1) files in the applib directory of my OC4J instance, but it doesn't work (the server still uses the same mail.jar 1.2)
    - I know that a patch exists : I've read the following document: How to Make Libraries such as mail.jar and activation.jar Swappable ? [ID 552432.1]
    This article talks about the Patch 6514136, but this patch only applies to : Oracle Containers for J2EE - Version: 10.1.3.3.0
    Can you please help me ?
    Thanks in advance for your answers,
    Laurent

    I strongly suggest to upgrade to AS 10.1.3 to get this.
    Think of future support of AS 9.0.4. You will get not critical patch updates anymore.
    --olaf                                                                                                                                                                                                                                                                                                               

  • How to use Sql data source from Essbase 9.3.1

    Hi All,
    How to use Sql data source from Essbase 9.3.1 for ASO cube.Are there any rules and limitations for that.
    Do we need to create any data source connection for this purpose. If there please let me know the dteps to create that connection.
    Regards

    Yes you need to create one DSN connection and you have to use DSN name and login details at the time of building/loading of the outline.
    Create DSN
    Goto Administrative tools -> DataSources (ODBC) and add the DSN name and specify the Server name of SQL and login details and database.
    goto data prep editor and click on File Menu and Click on Open SQL option Next window opens.
    There you have to enter the details of the DSN connection and SQL query to build/load.
    Thanks,
    Prathap

  • How to use many data sources

    I am accessing the data base in my jsp using the datasource in data-sources.xml. I need to use more number of schema's in my database and I am in need of using more number of datasources.
    How to create the datasources ? Is it enough , if I specify my datasource in
    the data-sources.xml alone or do I need to make any other changes in the other configuration files.
    Thanks.

    Add <datasource/> elements to data-sources.xml.

  • How to use 'query data source Name '  property in a block

    I want to create a block using query and want to show the columns on a form.
    Could you please tell me how can I do this?
    Regards

    Hi!
    Create a new block manually.
    Set the Query Data Source Type Property to FROM clause query.
    In the Query Data Source Name Property write your query like you would
    write it in sql*plus or toad but without a order by clause.
    Set the DML Data Target Type Property to None.
    Create items for the columns in your query.
    May set the where clause and order by property of the block.
    Build the form and execute a query in your new block.
    Regards

  • Data truncation with oracle.toplink.internal.databaseaccess.OraclePlatform

    Hello everyone
    I use Toplink 10g(9.0.4.5) with Oracle Lite database. I use the oracle.toplink.internal.databaseaccess.OraclePlatform and everything seems to be fine, except that when I pass a String more than the length defined in the Database, data is getting truncated to the maximum possible length in the Database and no exception is thrown. If I run a SQL statement directly on the Oracle Lite database using msql "value too large for column" exception is thrown.
    The question is it the default behviour of oracle.toplink.internal.databaseaccess.OraclePlatform to truncate the data for the large Strings? How can i change this behaviour to throw an exception?
    Any information on this would be highly appreciated.
    Thanks in advance..

    Thanks for the reply.
    But my tests are pointing towards ToplinkPlatform only. I use the oracle.lite.poljdbc.POLJDBCDriver with Oracle Database Lite, and when I do a simple test to insert a large String wiht plaing JDBC and POLJDBCDriver, i receive the following error.
    [POL-2403] value too large for column
    However if i did the same using Toplink and oracle.toplink.internal.databaseaccess.OraclePlatform, the SQL statement generated by Toplink does not throw any errors. When i query the database, i see that the data is truncated. I use the same oracle.lite.poljdbc.POLJDBCDriver with Toplink.
    <session>
    <name>testLite</name>
    <project-class>com.entityobjects.mappings.testLite</project-class>
    <session-type>
    <server-session/>
    </session-type>
    <login>
    <driver-class>oracle.lite.poljdbc.POLJDBCDriver</driver-class>
    <connection-url>jdbc:polite:POLITE;DataDirectory=C:\;Database=testLite;</connection-url>
    <platform-class>oracle.toplink.internal.databaseaccess.OraclePlatform</platform-class>
    <user-name>test1</user-name>
    <password>test1</password>
    <should-bind-all-parameters>true</should-bind-all-parameters></login>
    <enable-logging>true</enable-logging>
    <logging-options/>
    </session>
    I do not want to add a check at the object setter level, as the same object model is used in different projects and published across.
    Any more ideas why data is getting truncated?
    Thanks in advance for any suggestions..

  • How to use custom iometer icf with multiple io analyzers

    so here's my setup
    I have deployed to the IO analyzer ova to 3 VMs
    1 controller and 2 workers
    on the controller I have created custom Iometer workloads, gave them proper names and placed them in
    /var/www/configs/
    then i did
    /etc/init.d/tomcat6 restart
    when I load up the analyzer webui, my custom workload shows up in the dropdown menu
    next I proceeded to add my workers using my custom workload
    but when I hit run and check each console session, IOmeter only loads up on the controller but not the 2 workers..
    did I miss something?
    thanks

    Hi,
    Sorry that we didn't make it clear in our User's Guide. For new/modified ICF files, please manually SCP them to all the worker VMs. The backend script starting Iometer doesn't verify whether an ICF file exists.
    We'll be updating the document along with our next release.
    Thanks,
    Chien-Chia

  • How to use customer-specific fields with FI-CA event 940?

    The following text can be found under FI-CA event 940:
    You can also set customer-specific fields if you have defined the predefined include for customer enhancements CI_FKK_ACCIT_CUS in the structure FKK_ACCIT_CUS. The fields included here must have the same field name as the fields included in the structure ACCIT. These can also originate from an enhancement of the coding block in General Ledger Accounting. The fields Reservation Class and Reservation Key of the reconciliation key and the totals record are available as import parameters.
    My requirement is to populate the field HZUON in this event, which will be called when running t-code FPG1 (Transfer of FI-CA total records to general ledger)
    The problem is the fact that the output of the event (i.e. output of function module FKK_SAMPLE_0940 at first) only have the following fields:
    E_ZUONR     LIKE     ACCIT-ZUONR     Zuordnung (BSEG-ZUONR)
    E_XREF2     LIKE     ACCIT-XREF2     Referenz  (BSEG-XREF2)
    E_SGTXT     LIKE     ACCIT-SGTXT     Positionstext
    E_BUPLA     LIKE     ACCIT-BUPLA     Businnes Place (Korea only)
    I tried adding a new field to structure CI_FKK_ACCIT_CUS and adding this field to the output of the event 940, but how do I ensure that the content of the field will be used to populate the HZUON field of the FI document during "transfer of FI-CA total records to general ledger"?
    Cheers,
    Teo

    Hi Teo,
    I happened to see this post from you when I searched for a similar scenario related to event 940. I also need to add two fields in CI_FKK_ACCIT_CUS and make these fields as outputof the event. Did you come across any solution for the same? If so, can you please share.
    Regards,
    Harikumar. S
    Edited by: Harikumar Sasidharan on Dec 30, 2009 12:14 PM

  • How to create JDBC data source w/o LDAP server

    I am trying to test using JDBC data source on a computer without a LDAP server. Is there an alterative JNDI solution? How about using file system or RMI registry JNDI service providers?

    Any J2EE container should be able to handle that. I use JNDI data sources with Tomcat 4.1.27. I'm sure any other J2EE app server (e.g., WebLogic, WebSphere, JBOSS, etc.) would be able to manage it, too. - MOD

  • Custom Data Source in CMC does not work

    Hi there,
    Server : Crystal Reports Server 2008 V1
    Designer : CR 2008
    I have several reports that are originally set located to ODBC DSN1. I deployed this report to the server and in CMC, updated the database configuration in CMC as below.
    Enabled "Use Custom Data Source Information"
    Specified a new ODBC DSN2, user id, password, fully qualified table prefix for the new database etc.
    I can see that the report is running against the new database, and the data is selected, but the output is always blank. 
    But if I open the report in CR 2008, set locate the report to the new database DSN2, deploy again and use "original data source information", it works just fine. I have noticed this issue mainly with reports having sub-reports.
    Is there a solution to this problem. I don't want to set locate each report and deploy to the CMC if I have an option.
    Thanks for the help.
    Ajith

    I believe that the connection details are stored as part of the CR itself. Therefore if you update a connection in the CMC that might not change the parameters and credentials, especially when using sub-reports.
    Hope this helps...
    Martijn van Foeken
    Focuzz BI Services
    http://www.focuzz.nl
    http://nl.linkedin.com/in/martijnvanfoeken
    http://twitter.com/mfoeken

  • Generic Data Source with Function Module data mismatch in BI

    Hi All,
    I'm using Generic Data Source with Function Module, When I execute the Function Module (Which I have Created), I'm getting 16000 records and when run extractor(in RSA3) im getting different no.of records(infact they are more no.).
    when I run the InfoPackage in BI im Getting more no. of records than what i got executing the function module..
    and single record is divided into 2 records in BI side(not all the records), how can it be possible???
    is there anything Im missing to explain you my issue???
    if understood please help me out.
    Thanks n Regards,
    ravi.

    the datasource frame work starts the function module several times.
    1. the initialization
    2. the serval times, until you "raise no_more_data".
    check you coding: have you refreshed necessary internal tables.
    Sven

  • Generic Data Source with Function Module data mismatch

    Hi All,
    I'm using Generic Data Source with Function Module, When I execute the Function Module (Which I have Created), I'm getting 16000 records and when run extractor(in RSA3) im getting different no.of records(infact they are more no.).
    when I run the InfoPackage  in BI im Getting more no. of records than what i got executing the function module..
    and single record is divided into 2 records in BI side(not all the records), how can it be possible???
    is there anything Im missing to explain you my issue???
    if understood please help me out.
    Thanks n Regards,
    ravi.

    HI rkiranbi,
    1. FIrst you excute function module according to your paramers, you will get some records. then goto tcode RSA3 --> excute
    Provide your Data source name and under setting we have options like Data records/calls, Display extractor calls and selections --> fields .
    in that options you have to increase the values. and then you have to pass paramers in RSA3 according to your function module
    selections in SE37. Now you will get equal values in both functin module selection and RSA3 Selection. if it fail means  you need to
    check coding logic in function module. 
    2. if your  are getting wrong values in BI System then check with
                  1. compare with PSA data and data target data (here you need to check with characterstic as well as keyfigures)
                  if you find any mistake you need change the coding in function module according to client requirement.
                  2. compare data with RSA3 and bi report data or data target data.
                                 check it properly above steps, you will get solution.
    thanks and regards,
    malli

  • Shared Data Source with prompted credentials - errors out.

    Using Data Tools (VS 2010) to build report.
    Have built a simple report while trying to diagnose this issue.  This report is created in a project that was originally built in BIDS2008 and migrated to Data Tools 2010.  This project contains an original shared dataset that does NOT
    error out.
    Report contains
         1) shared data source with properties set to prompted credentials.
         2) shared dataset using data source above
         3) 2 parameters (month and year)
         4) Report uses shared dataset
         5) 2 text boxes in the report body to display the parameters values.
    When I try to preview this report I get the error:
         The execution failed for the shared data set "xxx"
         Cannot create a connection to data source 'Data source for shared dataset'
         Security processing failed with reason "3" ("password missing")
    Now, if I changed that same shared dataset to use a static login and password; the report renders fine.
    What is going on???

    Hi RoseNeedsAVacation,
    When you use shared data source with "Prompt for credential" option, please use SQL Server login credential to u view the report in BIDS or SSDT report designer environment.
    Note: The shared data source will automatic switch to SQL Server authentication mode if we use "Prompt for credential" option.
    After you deploy your report and shared data source to report server, please remember to configure your shared data source to use "Credentials supplied by the user running the report" option. Furthermore, select "Use
    as Windows credentials when connecting to the data source" checkbox if the credentials that the user provides are Windows Authentication credentials. Do not select this check box if you are using database authentication (for example, SQL
    Server Authentication).
    For more information, please refer to the article below:
    New Data Source Page (Report Manager):
    http://technet.microsoft.com/en-us/library/ms180077(v=sql.100).aspx
    Regards,
    Elvis Long
    TechNet Community Support

Maybe you are looking for

  • I need a query

    Hi I've a table with 3(THREE)columns like the following structure: desc acc Name Null? Type MCODE NOT NULL NUMBER(2) TYPE CHAR(1) BALANCE NUMBER(14,2) Select * from acc MCODE TYPE BALANCE 10 I 0 11 L 0 12 E 0 12 L 0 13 L 0 14 L 0 15 L 0 16 E 15468971

  • Is Adobe Photoshop Element right for me? Please advise.

    I need an App to allow me to use a laptop and a remote mouse, and to make a formal presentation to a large audience. I need it to show both "photos" and a "video" in my presentation, allowing me to page through the photos one-at-a-time while I talk a

  • Primary key is null in CMP ejbLoad()

    Using WebLogic 7.0 with Oracle 9i Lite. We're logging the values of the CMP fields in ejbLoad(). All of the CMP values look fine, except for the primary key value; it's always null. This means that Foo.getKey() (the CMP field that returns the primary

  • EWA Report do not generate ..

    Hi , The early Watch Report is not generated. I have checked every thing , rfc are working fine and services are activated also. Created a new Refresh Session using Sdccn In my satellite system. BUT i can not see the job SM:EXEC SERVICES executing in

  • EPM 11.1.1.3 Drill back with Hyperion Planning and Financial Reporting

    Hi, I'm trying to drill back data with Planning thanks to FDM but it doesn't work. In fact Planning server and FDM server are not installed on the same machine. My user is both admin on Planning application and FDM one. When I drill back my data the