How to specify custom SQL in polling db adapter with logical delete option

Hi all,
I am writing a SOA composite app using JDeveloper SOA Suite 11.1.1.4 connecting to a SQL Server db using a polling DB Adapter with the logical delete option to send data to a BPEL process.
I have requirements which go beyond what is supported in the JDeveloper UI for DB Adapter polling options, namely:
* update more than one column to mark each row read, and
* specify different SQL for the logical delete operation based on whether bpel processing of the data polled was successful or not.
A complicating factor is that the polling involves two tables. Here is my full use-case:
1) Polling will select data derived from two tables: e.g. 'headers' and 'details' simplified for this example:
table: headers
hid - primary key
name - data label
status - 'unprocessed', 'processed', or 'error'
processedDate - null when data is loaded, set to current datetime when row is processed
table: details
hid - foreign key pointed at header.hid
attr - data attribute name
value - value of data attribute
2) There is a many:1 relationship between detail and header rows through the hid columns. The db adapter polling SELECT shall return results from an outer join consisting of one header row and the associated detail rows where header.status = 'unprocessed' and header.hid = details.hid. (This is supported by the Jdeveloper UI)
3) The polled data will be sent to be processed by a bpel process:
3.1) If the bpel processing succeeds, the logical delete (UPDATE) operation shall set header.status = 'processed', and header.processedDate = 'getdate()'.
3.2) If bpel processing fails (e.g. hits a data error while processing the selected data) the logical delete (UPDATE) operation shall set header.status = 'failed', header.processedDate = 'getdate()', and header.errorMsg = '{some text returned from bpel}'.
Several parts of #3 are not supported by the JDeveloper UI: updating multiple columns to mark the row processed, using getdate() to populate a value of one of those column updates, doing different update operations based on the results of the BPEL processing of the data (success or error), and using data obtained from BPEL processing as a value of those column updates (error message).
I have found examples which describe specifying custom SQL using the polling delete option to create a template then modifying the toplink file(s) to specify custom select and update SQL to implement a logical delete. (e.g. http://dlimiter.wordpress.com/2009/11/05/advanced-logic-in-oracle-bpel-polling-database-adapter/ and http://myexperienceswithsoa.blogspot.com/2010/06/db-adapter-polling-tricks.html). But none of them match what I've got in my project, in the first case because maybe because I'm using a higher version of JDeveloper, and in the second I think because in my case two tables are involved.
Any suggestions would be appreciated. Thanks, John

Hi John,
You've raised a good scenario.
First of all let me say that the purpose of the DB polling transaction, is to have an option to initiate a process from a DB table/view and not to update multiple fields in a table (or have other complex manipulation on the table).
So, when choose to update a field in a record, after reading it, you are "telling" the engine not to poll this record again. Sure, i guess you can find a solution/workaround for it, but I don't think this is the way....
The question now is what to do?
You can have another DB adapter where you can update the data after finishing the process. In that case, after reading the data (on polling transaction) - update the header.status = 'processed' for example, and after processing the selected data update the rest of the fields.
Hope it make some sense to you.
Arik

Similar Messages

  • How to add custom sql operations in Database adapters?

    I have a database adapter in ESB configured with insert, update and select operations.
    I wanted to add another select operation with pure sql. When I added the operation it gave me error "Start of the root element expected" while registering the service.
    Can't we do like this?
    I use SOA suite 10.1.3.4.
    - Sam

    Hi Thanks for reply.
    There is an option while configuring the DB adapter "Execute Custom SQL" where you can specify the query you want to execute. It will generate the schema as per the query you enter in the wizard.
    I have created adapters with this type of custom SQLs and even I have put multiple operations in the same adapters where all use Custom SQL and they are working fine.
    Thease custom SQL operations can be anything insert, update, select and they even work with other operations using procedures in the same adapter. But yes we need to modify the WSDL manuallly.
    My problem was that when I try to add the custom SQL operation in the adapter which is configured with normal select, insert and update operation, I was getting the error mentioned.
    - Sam

  • How can our customer edit our Captivate 6 content with Captivate 5?

    Our company has decided to dive into the e-learning business and create e-learnings.  We just purchased a Captivate 6 subscription, unfortunately before finding out our customer uses Captivate 5.  He would like to edit our files once in a while when he needs to if he doesn't have enough time to have us do it.   Is there anyway he can edit content produced in Captivate 6 with his Captivate 5.  Can I export files of a certain type that he can in turn import and add to his Captivate 5 and edit himself?  In other words, what kind of files can I export from Captivate 6 than he can import with Captivate 5 and make small changes to?
      Vielen Dank,
                 Ryan

    Thank you.  I see that I can export content as an .xml file.  Isn’t this something he can upload with Captivate 5?    Thanks for bearing with my naivety. : )
    Von: RodWard [email protected]
    Gesendet: Mittwoch, 1. August 2012 14:25
    An: Scharfer, Ryan
    Betreff: How can our customer edit our Captivate 6 content with Captivate 5?
    Re: How can our customer edit our Captivate 6 content with Captivate 5?
    created by RodWard<http://forums.adobe.com/people/RodWard> in Adobe Captivate - View the full discussion<http://forums.adobe.com/message/4591371#4591371

  • Duplicate processing by DBAdapter when using Distributed Polling with Logical Delete Strategy

    We have DBAdapter based polling services in OSB running across two Active-Active clusters (total 20 managed service across 2 clusters),
    listening to the same database table. (Both clusters read from the same source DB). We want to ensure distributed polling without duplication,
    hence in the DBAdapter we have selected Distributed Polling option, meaning we are using "Select For Update Skip Locking".
    But we see that sometimes, same rows are processed by two different nodes and transactions are processed twice.
    How do we ensure that only one managed server processes a particular row using select for update? We do not want to use the markReservedValue option which was preferred in older version of DBAdapter.
    We are using following values in DB Adapter configuration, the Jdev project for DBAdapter and the OSB proxy using DBAdapter are attached.
    LogicalDeletePolling Strategy
    MarkReadValue = Processed
    MarkUnreadValue = Initiate
    MarkReservedValue = <empty as we are using Skip Locking>
    PollingFrequency = 1 second
    maxRaiseSize = 1
    MaxTransactionSize = 10
    DistributionPolling = checked   (adds lock-n-wait in properties file and changes the SQL to SELECT FOR UPDATE SKIP LOCKED)
    Thanks and Regards

    Hi All,
    Actually I'm also facing the same problem.
    Step I follwed:
    1) Created a job_table in database
    create talbe job_table(id, job_name, job_desc, job_status)
    2)created a bpel process to test the Inbound distributed polling.
    3)Configure the DBAdapter for polling.
    a)update a field in the job_table with logical delete.
    b)select the field name form the drop down
    c) change the read value-->Inprogress and unRead value--->Ready
    d) dont change the value for Reserved value
    e) select the check box for "distributed polling".
    f) the query will be appended with "For update NoWait."
    g)click next and then finish.
    4) Then i followed the below steps.
    To enable pessimistic locking, run through the wizard once to create an inbound polling query. In the Applications Navigator window, expand Application Sources, then TopLink, and click TopLink Mappings. In the Structure window, click the table name. In Diagram View, click the following tabs: TopLink Mappings, Queries, Named Queries, Options; then the Advanced… button, and then Pessimistic Locking and Acquire Locks. You see the message, "Set Refresh Identity Map Results?" If a query uses pessimistic locking, it must refresh the identity map results. Click OK when you see the message, "Would you like us to set Refresh Identity Map Results and Refresh Remote Identity Map Results to true?Ó Run the wizard again to regenerate everything. In the new toplink_mappings.xml file, you see something like this for the query: <lock-mode>1</lock-mode>.
    5) lock-mose is not changed to 1 in toplink_mappingss.xml
    Can we edit the toplink_mappings.xml manually.
    If yes, what allt he values Ineed to change in toplink_mappings.xml file, so that it will not pick the same record for the multiple times in clustered environment.
    Please help me out this is urgent.
    Thanking you in advance.

  • DBAdapter polling with logical delete x distrib polling x DB rows per trans

    Hi all.
    I'm trying to configure a DBAdapter with "logical delete" polling strategy, distributed polling (cluster environment) and a defined number of "Database Rows per Transaction".
    When I check the box "Distributed Polling", the SQL generated gets appended by "FOR UPDATE NOWAIT"
    However, when I set a value for "Database Rows per Transaction" the "FOR UPDATE NOWAIT" sql clause disappear.
    Is this a bug, or some limitation related to the "logical delete" strategy???
    Thanks
    Denis

    Hi All,
    Actually I'm also facing the same problem.
    Step I follwed:
    1) Created a job_table in database
    create talbe job_table(id, job_name, job_desc, job_status)
    2)created a bpel process to test the Inbound distributed polling.
    3)Configure the DBAdapter for polling.
    a)update a field in the job_table with logical delete.
    b)select the field name form the drop down
    c) change the read value-->Inprogress and unRead value--->Ready
    d) dont change the value for Reserved value
    e) select the check box for "distributed polling".
    f) the query will be appended with "For update NoWait."
    g)click next and then finish.
    4) Then i followed the below steps.
    To enable pessimistic locking, run through the wizard once to create an inbound polling query. In the Applications Navigator window, expand Application Sources, then TopLink, and click TopLink Mappings. In the Structure window, click the table name. In Diagram View, click the following tabs: TopLink Mappings, Queries, Named Queries, Options; then the Advanced… button, and then Pessimistic Locking and Acquire Locks. You see the message, "Set Refresh Identity Map Results?" If a query uses pessimistic locking, it must refresh the identity map results. Click OK when you see the message, "Would you like us to set Refresh Identity Map Results and Refresh Remote Identity Map Results to true?Ó Run the wizard again to regenerate everything. In the new toplink_mappings.xml file, you see something like this for the query: <lock-mode>1</lock-mode>.
    5) lock-mose is not changed to 1 in toplink_mappingss.xml
    Can we edit the toplink_mappings.xml manually.
    If yes, what allt he values Ineed to change in toplink_mappings.xml file, so that it will not pick the same record for the multiple times in clustered environment.
    Please help me out this is urgent.
    Thanking you in advance.

  • How to specify custom table name for subclasses in Kodo 3?

    I'm migrating our codes from Kodo 2.5.x to version 3.
    In Kodo 2.x, specifying the class/subclass table name uses the extension
    like this "@jdo.extension vendor-name="kodo" key="table"
    value="DyeStuffs"" when using XDoclet to generate the .jdo metadata files.
    It no longer works in Kodo 3. Following the Kodo's documentation, I have
    upgraded XDoclet to 1.2b3 and changed all doclet tags to the ones that
    XDoclet understands.
    I noticed in the documentation, to specify custom table mapping, it uses
    nested vendor extensions. E.g
    <extension vendor-name="kodo" key="jdbc-class-map" value="vertical">
    <extension vendor-name="kodo" key="table" value="TABLOID"/>
    <extension vendor-name="kodo" key="ref-column.JDOID"
    value="MAG_ID"/>
    </extension>
    So how do I specify nested extensions when using with XDoclet's tags?
    Need help as I have to complete the migration by end of the week.
    Thanks,
    Makas

    Abe White wrote:
    Makas Tzavellas wrote:
    Does it mean I have to map all classes and fields to the existing tables?As I tried to point out earlier, in 3.0 all mapping information must be
    explicit all the time. So basically everything is like mapping to an
    existing schema (only Kodo can generate those mappings for you... as
    long as you're not using XDoclet for all your mapping! :)
    It seems quite a pain to manually map all the existing tables and
    fields, especially when Kodo gave the table and field names. I don't
    mind mapping the tables, but mapping the fields will be a horror with
    105 tables. Have I missed out anything in the migration documentation
    that might help me with this process? Also does Kodo require the
    ..mapping file to work or having everything in .jdo file is sufficient
    (looking at the documentation it seems ok to have .jdo file only)?
    If I do, I prefer to regenerate all tables without the base table,
    basically having tables to include all the fields from the base table.
    And migrate all existing data to the new table structure.I'm not entirely clear on what you want. Are you saying that you want
    each subclass to have its own table and repeate the columns from its
    superclass table in its own?
    Yes, that is exactly what I want.
    If so, then this is not supported by Kodo (yet... we're working on it
    though).
    Ok, I'll settle having Kodo 3 to only map back to the original tables,
    as I believe Kodo 3 will provide some performance improvement as
    compared to 2.x.

  • How to specify custom namespace in generating WSDL for BAPI

    In the creation of the Web Service Definition(WSDL) for a rfc-enabled function module, I would like to specify the custom namespace rather than the default in /sap/bc/srt/rfc/sap/. And I could not change the path in the Service Definition after it is generated(greyed out in change mode).
    I have looked all over the SAP help and could not find any documentation for the change.
    Any advise?
    regards,
    Wai-ming

    Hi John,
    You've raised a good scenario.
    First of all let me say that the purpose of the DB polling transaction, is to have an option to initiate a process from a DB table/view and not to update multiple fields in a table (or have other complex manipulation on the table).
    So, when choose to update a field in a record, after reading it, you are "telling" the engine not to poll this record again. Sure, i guess you can find a solution/workaround for it, but I don't think this is the way....
    The question now is what to do?
    You can have another DB adapter where you can update the data after finishing the process. In that case, after reading the data (on polling transaction) - update the header.status = 'processed' for example, and after processing the selected data update the rest of the fields.
    Hope it make some sense to you.
    Arik

  • How to specify custom query as well as a selection criteria

    Hi
    I'm struggling to specify a custom select query with variable no. of agruments through toplink.
    The custom query looks like:
    select outer.col1, outer.col2, outer.col3,
    outer.effective_date, outer.capacity_rate, outer.expiry_date, outer.comments, outer.create_date,
    outer.create_user_id, outer.modify_date, outer.modify_user_id, outer.vol_uom_type_code
    from tbl outer,
    ( select max(effective_date) effective_date, col1, col2, col3 from tbl
    where expiry_date is null
    or (trunc(effective_date) >= to_date('01/01/'||to_char(sysdate,'yyyy'), 'dd/mm/yyyy') )
    group by col1, col2, col3
    ) inner
    where outer.col1 = inner.col1
    and outer.col2 = inner.col2
    and outer.col3 = inner.col3
    and trunc(outer.effective_date) = trunc(inner.effective_date)
    My domain class TBL.class is mapped one to one with DB Table TBL.
    With this customer query, I want to add a selection criteria depending upon what is specified on the UI-screen - I can get col1, col2 or a combination of these as search criteria.
    So far I have tried multiple approaches:
    1. Specifying the custom query (without arguments) as a ReadAll Query in toplink descriptor and then using expression builder to add the criteria but this doesn't work as custom ReadAll query gets overridden and i get back all the results from
    TBL which match the criteria specified.
    expFinal = new ExpressionBuilder().get("col1").equal(srchCriteria.getCol1());
    results = (List) getTopLinkTemplate().readAll(TBL.class,expFinal,false);
    2. Using Named Query:
    Toplink doesn't allow me to specify a custom SQL along with a selection criteria - both for the same NamedQuery.
    3. Specifying the DB query in java layer and then adding arguments and argument values. When query is executed, i don't see arguments being added.
    ReadAllQuery query = new ReadAllQuery(TBL.class);
    query.setSQLString(READ_ALL_QUERY);
    ExpressionBuilder exp = new ExpressionBuilder();
    query.addArgument("col1", Long.class);
    Vector vect = new Vector();
    vect.add(srchCriteria.getCol1());
    query.addArgumentValues(vect);
    result = (List) getTopLinkTemplate().executeQuery(query, false);
    Please suggest. Would be great if you could share some code bits.
    Thanks
    Nitin

    Nitin,
    This style of query does require code to apply logic as it is created. If you wish to have this used as a named query you can use a query re-direct so that instead of executing a query we call a method of yours.
    You can setup your named query to use re-direction into a static method as:
            QueryRedirector redirector = new MethodBaseQueryRedirector(QueryRedirectorExample.class, "redirectMethod");
            ReadAllQuery raq = new ReadAllQuery(Employee.class);
            raq.addArgument("FIRST_NAME");
            raq.setRedirector(redirector);
            descriptor.getDescriptorQueryManager().addQuery("findByFirstNameLike", raq);The static method that then gets invoked where you can dynamically create a query looks like:
        public static Object redirectMethod(DatabaseQuery query, Record record, Session session) {
            String firstName = (String)record.get("FIRST_NAME");
            ReadAllQuery raq = new ReadAllQuery(Employee.class);
            if (firstName != null) {
                raq.setSelectionCriteria(raq.getExpressionBuilder().get("firstName").likeIgnoreCase(firstName));           
            return session.executeQuery(raq);
        }Doug

  • How to submit custom SQL to the DB in SAP system

    Hi
    In the project in which I had participated before, there were CRM server and WAS which had IPC server (price decision tool). There was a requirement that IPC referred CRM data, but there was no suitable BAPI in CRM. So, at the UserExit in IPC I made the programs that submit custom SQL to CRM DB via JCo.
    In a similar idea, Can I submit custom SQL to the DB of SAP environment (ERP, CRM and so on) from general WAS (don't have the IPC) ?
    I already know the way that I make the function module which throw SQL in SAP and call it from WAS. Could you tell me other methods if it exists ?
    Best regards,

    What's about RFC?

  • How do I use SQL statements to perform calculations with form fields????

    Please help!!! I don't know how to use a SQL statement within my APEX form......
    My form is below. The user will enter the values in the form. Click on Submit. Then we need to run a SQL select statement with those values.
    Our form looks like this:
    Start_Date ____________
    Per_Period ____________
    Period ____________
    [Submit Button]
    The user will enter these 3 values in the form.
    This is an example of an user providing the values:
    Start_Date 03/14/08_______
    Per_Period $200.00________
    Period 4____________
    [Submit Button]
    Then they will click the Submit Button.
    The SQL statement (BELOW) returns output based on the users selections:
    START_DATE PER_PERIOD PERIOD
    14-MAR-2008 00:00 200 Week 1 of 4
    21-MAR-2008 00:00 200 Week 2 of 4
    28-MAR-2008 00:00 200 Week 3 of 4
    04-APR-2008 00:00 200 Week 4 of 4
    Total 800
    This is the full text of the SQL that makes the output above:
    with criteria as (select to_date('03/14/08', 'mm/dd/rr') as start_date,
    4 as periods,
    'Week' as period,
    200 per_period from dual),
    periods as (select 'Week' period, 7 days, 0 months from dual
    union all select 'BiWeek', 14, 0 from dual
    union all select 'Month', 0, 1 from dual
    union all select 'ByMonth', 0, 2 from dual
    union all select 'Quarter', 0, 3 from dual
    union all select 'Year', 0 , 12 from dual
    t1 as (
    select add_months(start_date,months*(level-1))+days*(level-1) start_date,
    per_period,
    c.period||' '||level||' of '||c.periods period
    from criteria c join periods p on c.period = p.period
    connect by level <= periods)
    select case grouping(start_date)
    when 1 then 'Total'
    else to_char(start_date)
    end start_date,
    sum(per_period) per_period,
    period
    from t1
    group by rollup ((start_date, period))
    THANKS VERY MUCH!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

    You're just doing a parameterized report, where the input fields are your parameters.
    Check out the Advanced Tutorial titled Parameterized Report here:
    http://download.oracle.com/docs/cd/E10513_01/doc/appdev.310/e10497/rprt_query.htm#BGBEEBJA
    Good luck,
    Stew

  • How do I get the Korg NanoKontrol to work with Logic Studio?

    I just bought a Korg NanoKontrol control surface to use with Logic Studio but it isn't working right. I can get the faders and pan to work in one scene but the transport controls don't work in any scene. If anyone knows anything about how to get it to work please let me know.

    Hey Scrooge, why the attitude? Did anyone here mistreat you? And why hijack the thread?
    Anyway, if you loose settings minutes after they worked, I would say the hardware may be shaky. Or the cabling. Or the operators' skills.
    Ambient8 has gracefully and concisely pointed the way, and if something is not working just describe what and how it is, instead of complaining (on somebody else's thread) to people who voluntarily try to offer help & solutions.
    Ambient8:
    +"Pay special attention to "Assigning Key Commands to Control Surfaces" on page >164+
    +The rest need to be set up as a control surface by using the Learn (Command+L) function to assign faders to tracks and knobs to pans, etc.+
    +See the Logic Pro 8 Control Surfaces Support guide in the Help menu for info on controller assignments. Chapter 2 "Customizing Controller Assignments" has the info assigning controllers to Logic Pro parameters."+
    If you say I did exactly as described on these pages, and all your settings seemed to disappear, I can only assume that either something is wrong with your Logic prefs, the logic.cs file or the CS itself.
    Regards, Erik.

  • SQL Server 2008 Express Installation with command line options , FAILED.

    HI,
    I Have SQL Server 2005 Express instances on my PC. Now i am trying to install 2008 Express edition with Advance Services(Reporting.)
    I am trying to install the SQL Server 2008 Express with command line Options.
    /qs /Action=Install /Hideconsole /SecurityMode=SQL /Features=SQL,Tools,RS /InstanceName=XXXXXX /SAPWD="StrongPassword" /SQLSYSADMINACCOUNTS="Builtin\Administrators" /SQLSVCACCOUNT="NT AUTHORITY\SYSTEM" /RSSVCACCOUNT="NT
    AUTHORITY\SYSTEM" /BROWSERSVCSTARTUPTYPE=Automatic /TCPENABLED=1 /NPENABLED=1
    But installation exits. It says the previous version 2005 tool are already exists, Remove them to Continue the 2008 Server Installation.
    please suggest me how to avoid that check and continue installation by replacing the older tool with 2008 automatically.
    regards 
    Pratap M:)

    Have you found any error in event logs?
    In my case even after uninstalling all 2005 related components it didn't worked. I had to resort to a brute force way and
    remove following registry keys
    32 Bit OS: HKLM\SOFTWARE\Microsoft\Microsoft SQL Server\90
    64 Bit OS: HKLM\Software\Wow6432Node\Microsoft\Microsoft SQL Server\90
    Check below threads
    http://stackoverflow.com/questions/175881/sql-server-express-2008-install-side-by-side-w-sql-2005-express-fails
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/5fc58507-9f40-4213-acbd-32a57c8822d7/cant-uninstall-sql-server-2005-express-tools?forum=sqlsetupandupgrade
    Please click the Mark as answer button and vote as helpful if this reply solves your problem

  • Database Adapter using Logical Delete Polling Strategy not updating field

    I have an ESB database adapter defined against a table. The adapter is set to use the Logical Delete Polling Strategy, and the table has an extra column defined as the "delete" field. When I register the adapter to the ESB and add a record to the table, the adapter reads the contents of the record as expected. However, 15 seconds later (that being the polling interval) it reads it again, and again 15 seconds after that, ad infinitum.
    I have defined the logical delete field as a single character field and set the Read value for "T" (a record is normally inserted with that field having a null value). Results as outlined in the previous paragraph. I redefined the field as a number and set the Read value to "1" (with a record normally inserted with that field having a value of "0"). Results as outlined in the previous paragraph.
    Apparently the update of the logical delete value to the Read value is not occurring. My question (obviously) is: Why?
    Thanks for your time.

    Here is the adapter WSDL:
    <?xml version="1.0" encoding="UTF-8"?>
    <definitions
    name="Poll_PM_LOG"
    targetNamespace="http://xmlns.oracle.com/pcbpel/adapter/db/Poll_PM_LOG/"
    xmlns:tns="http://xmlns.oracle.com/pcbpel/adapter/db/Poll_PM_LOG/"
    xmlns:plt="http://schemas.xmlsoap.org/ws/2003/05/partner-link/"
    xmlns:jca="http://xmlns.oracle.com/pcbpel/wsdl/jca/"
    xmlns:pc="http://xmlns.oracle.com/pcbpel/"
    xmlns:top="http://xmlns.oracle.com/pcbpel/adapter/db/top/PollPMLOG"
    xmlns="http://schemas.xmlsoap.org/wsdl/">
    <types>
    <schema xmlns="http://www.w3.org/2001/XMLSchema">
    <import namespace="http://xmlns.oracle.com/pcbpel/adapter/db/top/PollPMLOG"
    schemaLocation="PollPMLOG_table.xsd"/>
    </schema>
    </types>
    <message name="LogCollection_msg">
    <part name="LogCollection" element="top:LogCollection"/>
    </message>
    <portType name="Poll_PM_LOG_ptt">
    <operation name="receive">
    <input message="tns:LogCollection_msg"/>
    </operation>
    </portType>
    <binding name="Poll_PM_LOG_binding" type="tns:Poll_PM_LOG_ptt">
    <pc:inbound_binding/>
    <operation name="receive">
    <jca:operation
    ActivationSpec="oracle.tip.adapter.db.DBActivationSpec"
    DescriptorName="PollPMLOG.Log"
    QueryName="Poll_PM_LOG"
    PollingStrategyName="LogicalDeletePollingStrategy"
    MarkReadFieldName="PROCESSED"
    MarkReadValue="T"
    SequencingFieldName="ID"
    MaxRaiseSize="1"
    MaxTransactionSize="unlimited"
    PollingInterval="15"
    NumberOfThreads="1"
    UseBatchDestroy="false"
    ReturnSingleResultSet="false"
    MappingsMetaDataURL="PollPMLOG_toplink_mappings.xml" />
    <input/>
    </operation>
    </binding>
    <service name="Poll_PM_LOG">
    <port name="Poll_PM_LOG_pt" binding="tns:Poll_PM_LOG_binding">
    <jca:address location="eis/DB/PM"
    UIConnectionName="PM"
    ManagedConnectionFactory="oracle.tip.adapter.db.DBManagedConnectionFactory"
    />
    </port>
    </service>
    <plt:partnerLinkType name="Poll_PM_LOG_plt" >
    <plt:role name="Poll_PM_LOG_role" >
    <plt:portType name="tns:Poll_PM_LOG_ptt" />
    </plt:role>
    </plt:partnerLinkType>
    </definitions>
    The field PROCESSED is defined as CHAR(1), and is normally null when a record is inserted. The value of the field is to be set to 'T' when the record is read by the polling adapter.

  • How to reset or change "what should firefox do with this file" option in Firefox 3.6.6

    How to reset or change "what should firefox do with this file" in Firefox 3.6.6.
    Accidentally I changed it to save file for .pdf files and now I want to reset it to default asking whether to open or save.
    any help please
    Thanks,
    raj

    Check the setting in Tools > Options > Applications
    See "Reset Download Actions" and "File handling in Firefox 3 and SeaMonkey 2":
    http://kb.mozillazine.org/File_types_and_download_actions

  • How to specify custom page size in points?

    Have done the usual searches and not hit anything.
    Has anyone worked out how to get the custom sizes in Print Setup… to be measured in Points?
    I have a job where I need to work to screen resolution.
    The two System choices are for Metric which oddly shows in mm when everything else is in cm, or US measurements which are in inches (at least they aren't cubits).
    Nothing I see will switch it to the most accurate, by pdf/PostScript standards, of showing the measure in points.
    I have a vague recollection of having done this years ago in an OS far, far away. Could have been OS9, Windows or Linux of some ilk.
    Peter

    Peter,
    I have looked at this again in Pages ’09, and via Page Setup > Paper: Manage Custom Sizes..., it will not let me override the inch measurements.
    Try this to get two decimal places. Double-click the mm value in the paper size to select all, and then just type your mm setting with two decimal places (no mm text designation) and then tab out of that entry window. The two decimal setting is retained.
    This is redundant knowledge, but here is the conversion fluff. Inches in Page Setup may need to be set to approximate closest points. In the Postscript (cough) world, the conversion is:
    1 Point = 0.01388888889 Inch
    And the calculator (with round option).

Maybe you are looking for

  • Logical Volumes Not Creating w lvcreate? Install??

    After following the arch raid guide i have gotten all tge way down to creating logical volumes and i get  this lvcreate -L 20G VolGroupArray -n lvroot /dev/VolGroupArray/lvroot: not found: device not cleared   Aborting. Failed to wipe start of new LV

  • Mapping xsql on tomcat

    hi I deployed an application on apache-tomcat server. but when I access the xsql page it's displayed without executing the sql statement. I modified the web.xml. to map the xsql page with XSQLservlet. I also modified the httpd.conf file to handle the

  • Motion 1 on Mac OSX 10.5.8 Crashes

    Hi, I recently bought Apple Motion (apparently version 1) off of Amazon... I have Final Cut Express and didn't want to pay for Studio but wanted Motion. Anyway, I installed it and opened it up. First, it tries to register but instead gives me a messa

  • Yikes! I accidently trashed iphoto library

    I see my iphoto library in the trash, and a bunch of my albums. I was cleaning stuff up and must have inadvertently done this. I'm trying to figure out the best way to get things back to normal. I dragged the library to my desktop, and it's copying i

  • Dreamweaver is unable to open my app in the emulator.

    Hi..! I am using Dreamweaver cs5.5 to develop my android apps but unfortunately I couldn't run the app in the emulator. When I go to Site -> Mobile Applications -> Build And Emulate -> Android then the emulator opens up but my app does not display th