What is the best approach to generate control numbers from bpel?

1. If we want to control ISA/GS/ST control numbers from bpel, what is the best approach to do that?
2.  how to generate these control numbers and where to store them to get a sequence out of it?
Thanks,
Kathar

Internally Oracle B2B uses DB sequence for generating the control numbers. It is the best approach but at the same time it is not very straight forward, specially in case of clustered database. So you may carefully implement same with BPEL.
Hi Anuj,
If we let B2B to generate control numbers in the clustered environment, is there any settings we have to do?
So you may carefully implement same with BPEL. BTW, what is the use case behind this?
We were thinking about using this to send out duplicate messages to two TPs but we decided to go with java callout as you suggested in another thread.
Thanks!
Kathar

Similar Messages

  • What is the best approach to return Large data from Stored Procedure ?

    no answers to my original post, maybe better luck this time, thanks!
    We have a stored proc (Oracle 8i) that:
    1) receives some parameters.
    2)performs computations which create a large block of data
    3) return this data to caller.
    compatible with both ASP (using MSDAORA.Oracle), and ColdFusion (using Oracle ODBC driver). This procedure is critical in terms of performance.
    I have written this procedure as having an OUT param which is a REF CURSOR to a record containing a LONG. In order to make this work, at the end of the procedure I have to store the working buffer (an internal LONG variable) into a temp table, and then open the cursor as a SELECT from the temp table.
    I have tried to open the cursor as a SELECT of the working buffer (from dual) but I get the error "ORA-01460: unimplemented or unreasonable conversion requested"
    I suspect this is taking too much time; any tips about the best approach here? is there a resource with REAL examples on returning large data?
    If I switch to CLOB, will it speed the process, be compatible with callers, etc ? all references to CLOB I saw use trivial examples.
    Thanks for any help,
    Yoram Ayalon

    Create a new farm in the secondary Data Center at the same patch level with the desired configuration. Replicate the databases using the method of choice (Mirroring, AlwaysOn, etc.). Create a downtime window where you can then attach the databases to the
    new farm's Web Application(s)/Service Application(s).
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • What is the best way to preserve page numbers from InDesign to EPUB?

    I have a project where the source files are InDesign based - the final step before the product goes to print.  We want to bring this content to reflowable EPUB format, but we need to keep the print page numbers in order to support citations.
    Has anyone faced this type of challenge previously, and what was the approach pursued?  While an InDesign 2014 export to EPUB Fixed will preserve the page numbers, there will be a lot of cleanup required to make these files reflowable again.
    Andre

    Hi Douglas,
    I agree that the common practice of removing the page numbers works with regards to trade content (fiction, non-fiction, magazines, etc) where pages generally are to communicate how far you have 'travelled' in a particular title.  With reference works, however, this particular type of metadata can have a greater meaning - communicating to other individuals the location of a reference in order to allow them to see the reference in its original context.  It becomes a way of locating relatively precise locations within a work that is being referenced.
    Currently, and we can expect for some time, the original print content will be the authoritative version that is referred to, or cited, as a means of communicating the exact location of a specific reference to another individual.  Although evident in many industries, perhaps this arises most commonly in legal texts, where a lawyer as part of an argument or a judge in writing a decision will cite a text reference by title, volume, possibly edition, date/year, etc. and, ultimately, the page number - which in the end is just another piece of metadata relative to the text that appears on that particular page.  In order to create digital copies of these types of texts that can still function relative to the print authority, an accurate page number that corresponds to the print needs to be available somewhere in the texts.  (This is true both ways, by the way - both in terms of following a cite to a specific location within the content; and citing a specific location within the content).
    For many publishers, InDesign will be the final, what I would call 'exportable' format that can maintain a certain degree of information about the context of the source content and will have page numbers represented that precisely reflect what will be found in the final print version.
    I have some ideas on how I would like to present this in a reflowable digital version (and reflowable text for this type of content is desirable for all the reasons it would be elsewhere - adaptability to screen size, ability to resize/change fonts for different reader experience, etc). And the new EPUB fixed export function will allow me to preserve the page numbers and through some programming I can get to a point where I can isolate the text that corresponds to a specific page number from the InDesign exporta manual solution would not be an improvement), but with a somewhat 'cleaner' export of the text of the content, where the precise fixed formatting is not so important.
    The perfect solution would be to allow page numbers to be preserved within a 'hidden' tag in the EPUB reflowable export - so that the content of each page would be preceeded by an <a id="Page_75" /> tag, with words that are hyphenated belonging to the page on which they started.  Note that the use of such a tag would also go a long way towards supporting the call for a "page-list" and hyperlinks as you've described them above.  If the publisher wanted to make use of such a tag, it would be available to them - otherwise the tag would be ignored by publisher and eReader software alike.
    Andre

  • What is the best approach to migrate SharePoint farm from one data center to other datacenter

    We have two web front server and one application server and two instances for database server and we have to migrate complete farm from one data center to other data center with minimal downtime and end user impact.
    Please provide your best input on this.
    Thanks in advance.

    Create a new farm in the secondary Data Center at the same patch level with the desired configuration. Replicate the databases using the method of choice (Mirroring, AlwaysOn, etc.). Create a downtime window where you can then attach the databases to the
    new farm's Web Application(s)/Service Application(s).
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • What is the best way to get phone numbers from an iPhone 3g to an iPhone 4s?

    I am planning on buying an iPhone 4s but i was just wondering how i would get my contacts on the new phone?  I am on o2 .#
    Thank you!

    That's not a problem.  With a new device, you could connect to iTunes, set it up and restore from the backup of the old device (either including, or not, the contacts).  Then, with the same iCloud account setup on it, you could enable contact syncing and go from there.  In the iCloud settings on the iPhone, you can selective enable/disable the things you want to use iCloud for.  Contact sync via iCloud is a separate on/off setting, so you can use iCloud just for email and contacts, and nothing else if you wished.  Then use iTunes sync for everything else.
    Personally, even if you use iCloud for all its possible features, I'd still suggest backing up in iTunes once in a while too - an iTunes backup is more complete than an iCloud backup.

  • What´s the best approach to work with Excel, csv files

    Hi gurus. I got a question for you. According to your experience what's the best approach to work with Excel or csv files that have to be uploaded through DataServices to you datawarehouse.
    Let's say your end-user, who is not a programmer, creates a group of 4 excel files with different calculations in a monthly basis, so they can generate a set of reports from their datawarehouse once the files have been uploaded to tables in your DWH. The calculations vary from month to month. The user doesn't have a front-end to upload the excel files directly to Data Services. The end user needs to keep a track of which person uploaded the files for a determined month.
    1. The end user should place their 4 excel files in a shared directory that will be seen by DataServices.
    2. DataServices will execute certain scheduled job that will read the four files and upload them to the Datawarehouse at a determined time, lets say at 9:00pm.
    It makes me wonder... what happens if the user needs to present their reports immediately so they can´t wait until 9:00pm.  Is it possible for the end user to execute some kind of action (out of the DataServices Environment) so DataServices "could know" that it has to process those files right now, instead of waiting for the night schedule?
    Is there a way that DS will track who was the person who uploaded those files?
    Would it be better to build a front-end for the end user so they can upload their four files directlyto the datawarehouse?
    Waiting for your comments to resolve this dilemma
    Best Regards
    Erika

    Hi,
    There are functions in DS that captures the input files automatically. You could use file_exists() or wait_for_file() option to do that. Schedule the job to run every certain minute and if the file exists then run. This could be done by using a certain file name with date and timestamp etc or after running move the old files to archive and DS wait for new files to show up.
    Check this - Selective Reading and Postprocessing - Enterprise Information Management - SCN Wiki
    Hope this helps.
    Arun

  • What is the best approach ??

    I want to create an XML document using dynamically (run-time) generated data
    based on a DTD. What is the best approach ?? Should I use xmlDocument or ...

    Hi,
    Have a look at JAXP:
    http://java.sun.com/xml/jaxp/dist/1.1/docs/tutorial/index.html
    Regards,
    Slava Imeshev
    "vignette" <[email protected]> wrote in message
    news:[email protected]..
    I want to create an XML document using dynamically (run-time) generateddata
    based on a DTD. What is the best approach ?? Should I use xmlDocument or...

  • What is the best approach for Continuously reading all DI lines of all the ports of a DIO device?

    Hi! I am new to Labview ..... I want to monitor all the DI lines of all the ports of a device .... What should be my approach?

    same as this
    http://forums.ni.com/t5/LabVIEW/What-is-the-best-approach-for-Continuously-reading-all-DI-lines/td-p...
    Tim
    Johnson Controls
    Holland Michigan

  • What's the best approach for handeling about 1300 connections in Oracle.

    What's the best approach for handling about 1300 connections in Oracle 9i/10g through a Java application?
    1.Using separate schema s for various type users(We can store only relevant data with a particular schema.     Then No. of records per table can be reduced by replicating tables but we have to maintain all data with a another schema     Then we need update two schema s for a given session.Because we maintain separate scheama for a one user and another schema for all data and then there may be Updating problems)
    OR
    2. Using single schema for all users.
    Note: All users may access the same tables and there may be lot of records than previous case.
    What is the Best case.
    Please give Your valuable ideas

    It is a true but i want a solution from you all.I want you to tell me how to fix my friends car.

  • What are the best approaches for mapping re-start in OWB?

    What are the best approaches for mapping re-start in OWB?
    We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    We have number of mappings. We built process flows for mappings as well.
    I like to know, what are the best approches to incorportate re-start options in our process. ie a failure of mapping in process flow.
    How do we re-cycle failed rows?
    Are there any builtin features/best approaches in OWB to implement the above?
    Does runtime audit tables help us to build re-start process?
    If not, do we need to maintain our own tables (custom) to maintain such data?
    How did our forum members handled above situations?
    Any idea ?
    Thanks in advance.
    RI

    Hi RI,
    How many mappings (range) do you have in a process flows?Several hundreds (100-300 mappings).
    If we have three mappings (eg m1, m2, m3) in process flow. What will happen if m2 fails?Suppose mappings connected sequentially (m1 -> m2 -> m3). When m2 fails then processflow is suspended (transition to m3 will not be performed). You should obviate cause of error (modify mapping and redeploy, correct data, etc) and then repeat m2 mapping execution from Workflow monitor - open diagram with processflow, select mapping m2 and click button Expedite, choose option Repeat.
    In re-start, will it run m1 again and m2 son on, or will it re-start at row1 of m2?You can specify restart point. "at row1 of m2" - I don't understand what you mean (all mappings run in Set based mode, so in case of error all table updates will rollback,
    but there are several exception - for example multiple target tables in mapping without corelated commit, or error in post-mapping - you must carefully analyze results of error).
    What will happen if m3 fails?Process is suspended and you can restart execution from m3.
    By having without failover and with max.number of errors=0, you achieve re-cycle failed rows to zero (0).This settings guarantee existence only two return result of mapping - SUCCSES or ERROR.
    What is the impact, if we have large volume of data?In my opinion for large volume Set based mode is the prefered processing mode of data processing.
    With this mode you have full range enterprise features of Oracle database - parallel query, parallel DML, nologging, etc.
    Oleg

  • What is the best option for generating .wmv video working on Mac?

    I am working on a mac. I need to generate my output in .wmv. I am exporting as .mov then using MacX converter to convert to .wmv. The quality degrades too much. What is the best way to generate .wmv?
    Thanks in advance.

    YOur statement is way too generic and vague. You have not provided any exact settings,  version info, details about the footage etc. That asude WMV is a compressed format and wil always degrade.
    Mylenium

  • What is the best approach to converting LV7.1 tags to LV2012 shared variables in multiple VIs?

    What is the best approach to upgrading from LV7.1/DSC tags to LV2012/DSC shared variables, in multiple VIs running on multiple platforms? Our system is composed of  about 5 PCs running Windows 2000/LV7.1 Runtime, plus a PLC, and a main controller running XP/SP3/LV2012. About 3 of the PCs publish sensor information via tags across the LAN to the main controller. Only the main controller is currently being upgraded. Rudimentary questions:
    1. Will the other PCs running the 7.1 RTE (with tags) be able to communicate with the main controller running 2012 (shared variables)?
    2. Is it necessary to convert from tags to shared variables, or will the deprecated legacy tag VIs from LV7.1 work in LV2012?
    3. Will all the main controller VIs need to be incorporated into a project in order to use shared variables?
    4. Is the only way to do this is to find all tag items and replace them with shared variable items?
    Thanks in advance with any information and advice!
    lb
    Solved!
    Go to Solution.

    Hi lb,
    We're glad to hear you're upgrading, but because there was a fundamental change in architecture since version 7.1, there will likely be some portions that require a rewrite. 
    The RTE needs to match the version of DSC your using.  Also, the tag architecture used in 7.1 is not compatible with the shared variable approach used in 2012.  Please see the KnowledgeBase article Do I Need to Upgrade My DSC Runtime Version After Upgrading the LabVIEW DSC Module?
    You will also need to convert from tags to shared variables.  The change from tags to shared variables took place in the transition to LabVIEW 8.  The KnowledgeBase Migrating from LabVIEW DSC 7.1 to 8.0 gives the process for changing from tags to shared variables. 
    Hope this gets you headed in the right direction.  Let us know if you have more questions.
    Thanks,
    Dave C.
    Applications Engineer
    National Instruments

  • Newbie: What is the best approach to integrate BO Enterprise into web app

    Hi
    1. I am very new to Business Objects and .Net. I need to know what's the best approach
    when intergrating bo into my web app i.e which sdk do i use?
    For now i want to provide very basic viewing functionality for the following reports :
    -> Crystal Reports
    -> Web Intellegence Reports
    -> PDF Reports
    2. Where do i find a standalone install for the Business Objects Enteprise XI .Net providers?
    I only managed to find the wssdk but i can't find the others. Business Objects Enteprise XI
    does not want to install on my machine (development) - installed fine on server, so i was hoping i could find a standalone install.

    To answer question one, you can use the Enterprise .NET SDK for each, though for viewing Webi documents it is much easier to use the opendocument method of URL reporting to view them.
    The Crystal Reports and PDF instances can be viewed easily using the SDK.
    Here is a link to the Developer Library:
    [http://devlibrary.businessobjects.com/]
    VB.NET XI Samples:
    [http://support.businessobjects.com/communityCS/FilesAndUpdates/bexi_vbnet_samples.zip.asp]
    C# XI Samples:
    [http://support.businessobjects.com/communityCS/FilesAndUpdates/bexi_csharp_samples.zip.asp]
    Other samples:
    [https://boc.sdn.sap.com/codesamples]
    I answered the provider question on your other thread.
    Good luck!
    Jason

  • What is the best approach to process data on row by row basis ?

    Hi Gurus,
    I need to code stored proc to process sales_orders into Invoices. I
    think that I must do row by row operation, but if possible I don't want
    to use cursor. The algorithm is below :
    for all sales_orders with status = "open"
    check for credit limit
    if over credit limit -> insert row log_table; process next order
    check for overdue
    if there is overdue invoice -> insert row to log_table; process
    next order
    check all order_items for stock availability
    if there is item that has not enough stock -> insert row to
    log_table; process next order
    if all check above are passed:
    create Invoice (header + details)
    end_for
    What is the best approach to process data on row by row basis like
    above ?
    Thank you for your help,
    xtanto

    Processing data row by row is not the fastest method out there. You'll be sending much more SQL statements towards the database than needed. The advice is to use SQL, and if not possible or too complex, use PL/SQL with bulk processing.
    In this case a SQL only solution is possible.
    The example below is oversimplified, but it shows the idea:
    SQL> create table sales_orders
      2  as
      3  select 1 no, 'O' status, 'Y' ind_over_credit_limit, 'N' ind_overdue, 'N' ind_stock_not_available from dual union all
      4  select 2, 'O', 'N', 'N', 'N' from dual union all
      5  select 3, 'O', 'N', 'Y', 'Y' from dual union all
      6  select 4, 'O', 'N', 'Y', 'N' from dual union all
      7  select 5, 'O', 'N', 'N', 'Y' from dual
      8  /
    Tabel is aangemaakt.
    SQL> create table log_table
      2  ( sales_order_no number
      3  , message        varchar2(100)
      4  )
      5  /
    Tabel is aangemaakt.
    SQL> create table invoices
      2  ( sales_order_no number
      3  )
      4  /
    Tabel is aangemaakt.
    SQL> select * from sales_orders
      2  /
            NO STATUS IND_OVER_CREDIT_LIMIT IND_OVERDUE IND_STOCK_NOT_AVAILABLE
             1 O      Y                     N           N
             2 O      N                     N           N
             3 O      N                     Y           Y
             4 O      N                     Y           N
             5 O      N                     N           Y
    5 rijen zijn geselecteerd.
    SQL> insert
      2    when ind_over_credit_limit = 'Y' then
      3         into log_table (sales_order_no,message) values (no,'Over credit limit')
      4    when ind_overdue = 'Y' and ind_over_credit_limit = 'N' then
      5         into log_table (sales_order_no,message) values (no,'Overdue')
      6    when ind_stock_not_available = 'Y' and ind_overdue = 'N' and ind_over_credit_limit = 'N' then
      7         into log_table (sales_order_no,message) values (no,'Stock not available')
      8    else
      9         into invoices (sales_order_no) values (no)
    10  select * from sales_orders where status = 'O'
    11  /
    5 rijen zijn aangemaakt.
    SQL> select * from invoices
      2  /
    SALES_ORDER_NO
                 2
    1 rij is geselecteerd.
    SQL> select * from log_table
      2  /
    SALES_ORDER_NO MESSAGE
                 1 Over credit limit
                 3 Overdue
                 4 Overdue
                 5 Stock not available
    4 rijen zijn geselecteerd.Hope this helps.
    Regards,
    Rob.

  • What is the best approach to install BI statistics in SAP BI ?

    Hello All,
    what is the best approach to install BI statistics in SAP BI ?
    by collecting objects in standard BI content- 0TCT*objects or
    by executing some standard tcodes.
    Regards,
    Siva

    the best approach is based on version of your BW system follow up install steps in notes:
    BW 3.x:
    309955 - BW statistics - Questions, answers and errors
    BW 7.x
    934848 - Collective note: (FAQ) BI Administration CockpitBW 7.x
    Cheers,
    m./

Maybe you are looking for

  • I am struggling to re-boot my time capsule

    I have an old Time Capsule ... have not used it for about a year (since moving home) I have now set up a home network (via power plugs) I wanted to set up the capsule by attaching it to my modem or directly into a pwer ethernet plug .... however my M

  • Closing command window....

    hi i have an application that is to be made into a jar and then run. I want to start the jar through the command prompt. is there some way i can close the command window from my code itself based on a variable in a properties file???? Please help....

  • No image with component cable through receiver?

    I have just purchased an Apple TV device, since my HDMI Input in my projector is not working everything is not plugged in my receiver with component cable. here is a list of what my audio/video setup is : - Sony PS3 (Plugged in Receiver in Component

  • Superdrive wont read

    My disk drive quit working today. I tried to burn a cd and a couple error messages popped up, and now when i try to insert any cd it will spin in the drive and then eject it. any suggestions?

  • Final Cut Server not storing clip proxy in the database

    FCS generates a clip proxy but doesn't store it in the database unless the user that installed FCS is logged into the server. The error message I get is: ERROR: EREQERROR Encountered error using Compressor: <Error>: kCGErrorFailure: Set a breakpoint