TCP/IP Timeout setting when using CLS

G'day community,
I'm in a project where we need to change the link-loss detection for DMQ connection using TCPIP, independent of the systemwide TCPIP/IP settings (we're talking DMQ v5.0-RP47 in OpenVMS 8.4).
There's no problem when using DMQ - DMQ connections: the logicals DMQLD_PING_TIMEOUT and DMQLD_PING_INTERVAL work fine here.
However, when connecting with non-OpenVMS systems via CLS, this does not work. It looks like the TCP driver is perfectly happy pinging CLS and CLS doesn't notice the communication partner went away (until the system-wide timeout is reached; but that has an higher value for other purposes!).
Is there a way to detect a connection was lost before TCP at systemlevel detects this?
Thanks in advance!
Oscar

hi mike...thanks again...
I created two parallel loops as you said...it works fine but one problem is there....as I said two parallel while loops are there with separate stop button...so when I start vi both are starting.... when I stop the my other code while loop it stops but my tcp vi does not stop even I stop the button because it is in the forever state...so it does not reach that button until the client initiate for once atleast....I just want to exit this forever state from server side.whenever server wants... 
how to terminate that forever state from tcp listen vi?
Attachments:
tcp.png ‏158 KB

Similar Messages

  • Using smart previews on iMac to control two sets of identical original files on two external drives --- render previews setting when using smart previews

    I’m just getting started with Lightroom and if someone can help me I’ll be very appreciative.
    I’m using an iMac with a 1TB drive and two external 2.5 TB drives for backup.  My active files are on all three drives and I’m running out of room on the iMac drive.
    With the smart previews capability of Lightroom 5, I’m hoping to remove all the active files from the iMac drive.  I hope to import existing images into the LR catalog from two identical drives with identical folders and have both sets of images reflect the changes I make using smart previews on my iMac.
    I just created two identical test folders of junk images on the external drives and was following a beginners tutorial on importing. However, I see nothing about using smart previews on my iMac that, when connected, will update more than one external drive at the same time.
    I’m also wondering if, when using smart previews, it matters how I set the render previews during the import process.
    Many thanks for any help,
    Karl Hentz

    I don't see any way to use smart previews to update both external drives at the same time. The smart previews are linked to a specific set of images. But it seems to me that if both drives are in fact identical, once the catalog is updated when one of those drives is connected those changes should also apply to the files on the other drive. It's actually the catalog that gets updated, not the images.  Smart previews and regular import previews are completely unrelated. How you have your import previews configured will have no effect on your smart previews.
    Incidentally, in order for my suggestion to work it would be necessary to only have one of those external drives connected. And when the other EHD is connected it would have to have the same drive letter or name. In other words it would have to appear to the iMac that it is precisely the same drive.

  • Is there any database parameters we need to set when using DBMS_PARALLE_EXECUTE?

    Hi,
    I am using dbms_parallel_execute package to do some process, I call a procedure to do the some tasks and insert data using the chunks by sql. Here is my code, I am calling p_process procedure, I am using 11g 11.2.0.3.0 - 64 bit  on windows server.
    DBMS_PARALLEL_EXECUTE.CREATE_TASK ('process a');
      DBMS_PARALLEL_EXECUTE.CREATE_CHUNKS_BY_SQL (
          task_name => 'process a',
         sql_stmt => 'SELECT startid, endid FROM chng_chunks',
         by_rowid => FALSE);
            dbms_parallel_execute.run_task
            ( task_name      => 'process a',
              sql_stmt       => 'begin P_process( :start_id, :end_id ); end;',
              language_flag  => DBMS_SQL.NATIVE,
             parallel_level => 24 );
    This code runs very fast on a one database and I can see it uses lots of cpus but it runs very slow on a copy of the same database on another server which has more cpus and memory. I compare v$parameter vlues and those are pretty much  identical between databases. I checked the disk spaces and both servers has plenty of free space on disks.
    Now my question is, is there any other parameters that we need to set/check when using dbms_parallel_execute package.
    Thanks in advance.
    gg

    I don't get this. Ever. Why developers insist on comparing server1 with server2, simply because their code is running on both.
    It is like comparing the athletic ability of two persons, arguing that h/w wise they are the same (i.e. human), and both have green eyes (your same software). And as because these are all the same, both persons should be able to run the 100m in the same time.
    Yes, the analogy is silly.. as is the warped concept amongst many developers that server1 and server2 should exhibit the same performance when running the same code.
    It. Does. Not. Work. Like. That.
    Want to know why server2 is exhibiting the performance it does when running your code?
    Do that by ignoring server1 as it is NOT RELEVANT.
    Do that by examining the workloads and resource usage of server2, and the run and wait states of your code on server2.

  • Af:inputListOfValues sets value of first item in result set when using enter key or tab and component set to autosubmit=true

    I'm using JDev 11.1.1.6 and when I type a value into an af:inputListOfValues component and hit either the enter key or the tab key it will replace the value I entered with the first item in the LOV result set. If enter a value and just click out of the af:inputListOfValues component it works correctly. If I use the popup and search for a value it works correctly as well. I have a programmatic view object which contains a single transient attribute (this is the view object which is used to create the list of value component from) and then I have another entity based view object which defines one of its attributes as a list of value attribute. I tried using an entity based view object to create the LOV from and everything works as expected so I'm not sure if this is a bug when using programmatic view objects or if I need more code in the VOImpl. Also, it seems that after the first time of the value being replaced by the first value in the result set that it will work correctly as well. Below are some of the important code snippets.
    Also, it looks like it only doesn't work if the text entered in the af:inputListOfValues component would only have a single match returned in the result set. For instance given the result set in the code: Brad, Adam, Aaron, Fred, Charles, Charlie, Jimmy
    If we enter Cha, the component works as expected
    If we enter A, the component works as expected
    If we enter Jimmy, the component does not work as expected and returns the first value of the result set ie. Brad
    If we enter Fred, the component does not work as expected and returns the first value of the result set ie. Brad
    I also verified that I get the same behavior in JDev 11.1.1.7
    UsersVOImpl (Programmatic View Object with 1 transient attribute)
    import java.sql.ResultSet;
    import java.util.ArrayList;
    import java.util.Iterator;
    import java.util.List;
    import oracle.adf.share.logging.ADFLogger;
    import oracle.jbo.JboException;
    import oracle.jbo.server.ViewObjectImpl;
    import oracle.jbo.server.ViewRowImpl;
    import oracle.jbo.server.ViewRowSetImpl;
    // ---    File generated by Oracle ADF Business Components Design Time.
    // ---    Wed Sep 18 15:59:44 CDT 2013
    // ---    Custom code may be added to this class.
    // ---    Warning: Do not modify method signatures of generated methods.
    public class UsersVOImpl extends ViewObjectImpl {
        private static ADFLogger LOGGER = ADFLogger.createADFLogger(UsersVOImpl.class);
        private long hitCount = 0;
         * This is the default constructor (do not remove).
        public UsersVOImpl () {
         * executeQueryForCollection - overridden for custom java data source support.
        protected void executeQueryForCollection (Object qc, Object[] params, int noUserParams) {
             List<String> usersList = new ArrayList<String>();
             usersList.add("Brad");
             usersList.add("Adam");
             usersList.add("Aaron");
             usersList.add("Fred");
             usersList.add("Charles");
             usersList.add("Charlie");
             usersList.add("Jimmy");
             Iterator usersIterator = usersList.iterator();
             setUserDataForCollection(qc, usersIterator);
             hitCount = usersList.size();
             super.executeQueryForCollection(qc, params, noUserParams);
        } // end executeQueryForCollection
         * hasNextForCollection - overridden for custom java data source support.
        protected boolean hasNextForCollection (Object qc) {
             Iterator usersListIterator = (Iterator)getUserDataForCollection(qc);
             if (usersListIterator.hasNext()) {
                 return true;
             } else {
                 setFetchCompleteForCollection(qc, true);
                 return false;
             } // end if
        } // end hasNextForCollection
         * createRowFromResultSet - overridden for custom java data source support.
        protected ViewRowImpl createRowFromResultSet (Object qc, ResultSet resultSet) {
             Iterator usersListIterator = (Iterator)getUserDataForCollection(qc);
             String user = (String)usersListIterator.next();
             ViewRowImpl viewRowImpl = createNewRowForCollection(qc);
             try {
                 populateAttributeForRow(viewRowImpl, 0, user.toString());
             } catch (Exception e) {
                 LOGGER.severe("Error Initializing Data", e);
                 throw new JboException(e);
             } // end try/catch
             return viewRowImpl;
        } // end createRowFromResultSet
         * getQueryHitCount - overridden for custom java data source support.
        public long getQueryHitCount (ViewRowSetImpl viewRowSet) {
             return hitCount;
        } // end getQueryHitCount
        @Override
        protected void create () {
             getViewDef().setQuery(null);
             getViewDef().setSelectClause(null);
             setQuery(null);
        } // end create
        @Override
        protected void releaseUserDataForCollection (Object qc, Object rs) {
             Iterator usersListIterator = (Iterator)getUserDataForCollection(qc);
             usersListIterator = null;
             super.releaseUserDataForCollection(qc, rs);
        } // end releaseUserDataForCollection
    } // end class
    <af:inputListOfValues id="userName" popupTitle="Search and Select: #{bindings.UserName.hints.label}" value="#{bindings.UserName.inputValue}"
                                                  label="#{bindings.UserName.hints.label}" model="#{bindings.UserName.listOfValuesModel}" required="#{bindings.UserName.hints.mandatory}"
                                                  columns="#{bindings.UserName.hints.displayWidth}" shortDesc="#{bindings.UserName.hints.tooltip}" autoSubmit="true"
                                                  searchDesc="#{bindings.UserName.hints.tooltip}"                                          
                                                  simple="true">
                              <f:validator binding="#{bindings.UserName.validator}"/>                      
    </af:inputListOfValues>

    I have found a solution to this issue. It seems that when using a programmatic view object which has a transient attribute as its primary key you need to override more methods in the ViewObjectImpl so that it knows how to locate the row related to the primary key when the view object records aren't in the cache. This is why it would work correctly sometimes but not all the time. Below are the additional methods you need to override. The logic you use in retrieveByKey would be on a view object by view object basis and would be different if you had a primary key which consisted of more than one attribute.
    @Override
    protected Row[] retrieveByKey (ViewRowSetImpl viewRowSetImpl, Key key, int i) {
        return retrieveByKey(viewRowSetImpl, null, key, i, false);
    @Override
    protected Row[] retrieveByKey (ViewRowSetImpl viewRowSetImpl, String string, Key key, int i, boolean b) {
        RowSetIterator usersRowSetIterator = this.createRowSet(null);
        Row[] userRows = usersRowSetIterator.getFilteredRows("UserId", key.getAttribute(this.getAttributeIndexOf("UserId")));
        usersRowSetIterator.closeRowSetIterator();
        return userRows;
    @Override
    protected Row[] retrieveByKey (ViewRowSetImpl viewRowSetImpl, Key key, int i, boolean b) {
        return retrieveByKey(viewRowSetImpl, null, key, i, b);

  • Conversions between character sets when using exp and imp utilities

    I use EE8ISO8859P2 character set on my server,
    when exporting database with NLS_LANG not set
    then conversion should be done between
    EE8ISO8859P2 and US7ASCII charsets, so some
    characters not present in US7ASCII should not be
    successfully converted.
    But when I import such a dump, all characters not
    present in US7ASCII charset are imported to the database.
    I thought that some characters should be lost when
    doing such a conversions, can someone tell me why is it not so?

    Not exactly. If the import is done with the same DB character set, then no matter how it has been exported. Conversion (corruption) may happen if the destination DB has a different character set. See this example :
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:01 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> create table test(col1 varchar2(1));
    Table created.
    TEST@db102 SQL> insert into test values(chr(166));
    1 row created.
    TEST@db102 SQL> select * from test;
    C
    ¦
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:47:55 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ©
    Typ=1 Len=1: 166
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ echo $NLS_LANG
    AMERICAN_AMERICA.EE8ISO8859P2
    [ora102 work db102]$ exp test/test file=test.dmp tables=test
    Export: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:47 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    server uses WE8ISO8859P15 character set (possible charset conversion)
    About to export specified tables via Conventional Path ...
    . . exporting table                           TEST          1 rows exported
    Export terminated successfully without warnings.
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:48:56 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> drop table test purge;
    Table dropped.
    TEST@db102 SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    [ora102 work db102]$ imp test/test file=test.dmp
    Import: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:15 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in EE8ISO8859P2 character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P15 character set (possible charset conversion)
    . importing TEST's objects into TEST
    . importing TEST's objects into TEST
    . . importing table                         "TEST"          1 rows imported
    Import terminated successfully without warnings.
    [ora102 work db102]$ export NLS_LANG=AMERICAN_AMERICA.WE8ISO8859P15
    [ora102 work db102]$ sqlplus test/test
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jul 25 14:49:34 2006
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    TEST@db102 SQL> select col1, dump(col1) from test;
    C
    DUMP(COL1)
    ¦
    Typ=1 Len=1: 166
    TEST@db102 SQL>

  • WRE54G v3 - WEP key setting when using ASCII passphrase

    When entering WEP information for this device you may notice that the range expander is asking for the WEP key in hex (for example: 128 bits 26 hex digits). If the key that you have set in your router is using a passphrase then you will need to convert it to hex. Entering the passphrase and clicking the generate button will not work if your router is set to use ASCII as the WEP key format. The result seems to be that the PC may attempt to connect (and/or show as connected) but will not receive an IP address and does not really connect to your network. Strangely, the WRE54G will show that it has connected sucessfully to the router. Instead, convert the passphrase to hex one character at a time then enter the hex digits as the hex key. For example: if you are using 128 bit and your passkey is: 1234567890123 then enter 31323334353637383930313233 as the 128bit hex key (1 in ASCII = 31 in hex, 2 in ASCII = 32 in hex, etc...). http://www.asciitable.com/ 

    Hi, The solution I posted above works for me . I can ping the WRE from the network and access the configuration page. The wireless router I am connecting to is a TP-LINK TL-WR340G. The extender is a WRE54G ver. 3 running v3.04.01 firmware. Instead of converting the passphrase to hex I could just set the TP-LINK router to use a hex key instead but thats a pain. Then I would have to type in a long hex key when adding another wireless device to the network or when allowing a guest device on the network. An ASCII passphrase is so much easier to remember and much less typing. If I leave the TP-LINK router set to accept an ASCII passphrase then convert the ASCII passphrase to hex and enter it into the WRE extender then I can use an ASCII passphrase when connecting a device through the extender .
    What I found confusing initially though (before finding the solution) was that if I enter the ASCII passphrase into the extender setup and click the generate button for it to generate the 4 hex keys, the extender claims to connect successfully. I cannot ping it from the network and a wireless device will not get a connection through it but the light does turn blue and the setup screen shows the correct mac address in the link status section. Seems like a bug...

  • Capture setting when using a firewire and camcorder

    does anyone know the capture setting I use when importing from a dvc camcorder using a firewire

    Typically it would be DV/DVCPro (NTSC or PAL), 48khz, 16 bit.
    -DH

  • Timeout errors when using high speed camera.

    Hi all.
    I'm currently trying to capture images using a Mikrotron EoSens MC1363 camera at high frame rates >500fps. The issue arises when the region of interest (ROI) is decreased in the microtron software (of which screenshots are attached). The ROI must be dropped in order to increase the fps. When MAX is opened and configured to match the camera settings, and grab is initiated, the timeout error occurs. I've attached screenshots of the mikrotron software, max settings, the error and PC used. The OS is windows 7 64 bit. We use the camera config file for the mono version of the camera which i've been informed will cross over to satisfy the colour version we use. Any questions feel free to ask. Thanks.
    Attachments:
    errors.docx ‏3016 KB
    997-EoSens 3CL-MC1362-Manual.pdf ‏1209 KB

    Hi Dom.
    As you said the maximum frame rate with the ROI set to 1280 x 1024 is 505fps but according to the general information for the camera series:
    The Mikrotron EoSens camera series features extremely sensitive high-speed
    CMOS sensors available in monochrome or in colour with a resolution of 1280 x
    1024 pixels and capture rates of 110 or 500 frames per second. Depending on
    the model the image data is transferred in 8 or 10 bits via either CameraLink
    (Base, Medium or Full) or Gigabit Ethernet.The region of interest (ROI) can be
    freely selected, and the cameras can achieve even higher capture rates up to
    120,000 fps when the ROI is reduced
    Therefore it was my thought that by reducing the width of the ROI increased frame rates may be reached. Realistically we're looking for as high frame rates as possible so that the camera has capability to pass between projects. Currently the moving object enters and leaves the ROI within a 4 frame window, so frame rates of up to 1000fps would generate more data. Naturally, a point will come where the data cannot be physically written as fast as it is generated (>600MBps). Therefore it would be useful to get a handle onto why the error occurs as well as methods to fix it. Hopefully the problem is being discussed between NI and Mikrotron and we may have some answers in the near future.

  • Error "Object refernce not set" when using http context in Eventreceiver for Document Library,

    Hi,
    Httpcontext returning null values when I used in item adding event of document library.
    I have a document set home page which has document set properties(by default) and one OOTB document library is placed in the same page.When user clicking on upload documents in the ribbon I need to read the document set properties.
    when uploading the docs if I pass the Hardcoded item id of the document set I am getting the values of the document set in the item added event.
    now the problem, i have the id in the query string if i want to read the query string value using httpcontext in the itemadding event and pass it to item added event using httpruntime it is throwing "object reference error" .
    if any body has any solution/workaround how to read the context in the document library event receiver,please help me out.
    Thanks ,G1

    Hi,
    Did you use the HttpContext.Current method to get the current context?
    You can use the HttpContext.Current to check whether it works.
    There is a similar thread for your reference.
    http://stackoverflow.com/questions/1601352/how-to-obtain-the-httpcontext-in-event-handler
    Thanks,
    Jason
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jason Guo
    TechNet Community Support

  • Is there any way to activate the keep identity setting when using a Transfer SQL Server Object Task?

    Is there any way to add the "Keep Identity" option to the tables selected for transfer in a Transfer SQL Server Object Task?  It seems that would be very useful, yet I can't find an easily available setting for it.  I would prefer to
    accomplish this within the Transfer SQL Server Object Task since it can copy any new fields added to the source tables, as opposed to having to rebuild a custom Data Flow Task any time a table structure is altered.  I thought setting "CopyPrimaryKeys"
    to True would do the trick, but I found that only ensures that the column settings are transferred; when the transfer happens, the destination id column is populated based on the specified seed and increment value, not what's actually in the source table. 
    I'd like to point out that replication isn't really an option here. This is a package that runs a few times a day to keep some data on our website up to date with data in our protected corporate environment.

    IIRW there is FastLoadKeepIdentity specifies
    whether to keep an identity in OLEDB destination 
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • Order Quantity (Config.) not set when using BAPI_SALESORDER_CREATEFROMDAT2

    Hi when I use BAPI_SALESORDER_CREATEFROMDAT2 order quantity is empty (missing also in the inclompent log in VA03)
    why ? what should be also set ? thanks a lot !
    I set this data:
    ls_order_header_in-doc_type    = 'TA'.
    ls_order_header_in-sales_org   = '1000'.
    ls_order_header_in-distr_chan  = '01'.
    ls_order_header_in-division    = '01'.
    ls_order_header_in-PURCH_NO_C  = 'test'.
    ls_order_partners-partn_role = 'AG'.
    ls_order_partners-partn_numb = 'CH10'.
    APPEND ls_order_partners TO lt_order_partners.
    ls_order_item_in-material = 'S-000'.
    APPEND ls_order_item_in TO lt_order_item_in.
    ls_order_cfgs_value-charc = 'S001'.
    ls_order_cfgs_value-value = '1'.
    APPEND ls_order_cfgs_value TO lt_order_cfgs_value.
    ls_order_cfgs_value-charc = 'S002'.
    ls_order_cfgs_value-value = '1'.
    APPEND ls_order_cfgs_value TO lt_order_cfgs_value.
    ls_order_cfgs_value-charc = 'S002'.
    ls_order_cfgs_value-value = '2'.
    APPEND ls_order_cfgs_value TO lt_order_cfgs_value.
    ls_order_cfgs_value-charc = 'S003'.
    ls_order_cfgs_value-value = '1'.
    APPEND ls_order_cfgs_value TO lt_order_cfgs_value.
    ls_order_cfgs_value-charc = 'S003'.
    ls_order_cfgs_value-value = '3'.
    APPEND ls_order_cfgs_value TO lt_order_cfgs_value.
    CALL FUNCTION 'BAPI_SALESORDER_CREATEFROMDAT2'
      EXPORTING
    *   SALESDOCUMENTIN               =
        order_header_in               = ls_order_header_in
    *   ORDER_HEADER_INX              =
    *   SENDER                        =
    *   BINARY_RELATIONSHIPTYPE       =
    *   INT_NUMBER_ASSIGNMENT         =
    *   BEHAVE_WHEN_ERROR             =
    *   LOGIC_SWITCH                  =
    *   TESTRUN                       =
    *   CONVERT                       = ' '
    * IMPORTING
    *   SALESDOCUMENT                 =
      TABLES
        return                        = lt_return
        order_items_in                = lt_order_item_in
    *   ORDER_ITEMS_INX               =
        order_partners                = lt_order_partners
    *   ORDER_SCHEDULES_IN            =
    *   ORDER_SCHEDULES_INX           =
    *   ORDER_CONDITIONS_IN           =
    *   ORDER_CONDITIONS_INX          =
    *   ORDER_CFGS_REF                =
    *   ORDER_CFGS_INST               =
    *   ORDER_CFGS_PART_OF            =
        order_cfgs_value              = lt_order_cfgs_value.
    *   ORDER_CFGS_BLOB               =
    *   ORDER_CFGS_VK                 =
    *   ORDER_CFGS_REFINST            =
    *   ORDER_CCARD                   =
    *   ORDER_TEXT                    =
    *   ORDER_KEYS                    =
    *   EXTENSIONIN                   =
    *   PARTNERADDRESSES              =
    *   NFMETALLITMS                  =
    CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'

    We use the same BAPI to process Web orders into SAP.  You have to have a corresponding ORDER_SCHEDULES_IN entry with matching ITM_NUMBER for every ORDER_ITEMS_IN entry.  Within ORDER_SCHEDULES_IN is where you put the quantity.  Should be stored in REQ_QTY.    We had to fill ITM_NUMBER and SCHED_LINE with the line number value in order to make it work.  Remember to fill in your corresponding X-structure fields as well.
    hope this helps,
    Chad

  • Conversions between character sets when using 'exp' and 'expdp' utilities

    When I make export with exp utility with NLS_LANG environment
    variable not set, export is done with US7ASCII character set,
    but when server uses some other character set e.g. EE8ISO8859P2,
    there can be some national specific characters lost with such a conversion.
    But when I use expdp utility and have not NLS_LANG set,
    in log file there is not mentioned that export is done in other character set
    than that on the server, does it mean that it uses character set on server?

    When I make export with exp utility with NLS_LANG environment
    variable not set, export is done with US7ASCII character set,
    but when server uses some other character set e.g. EE8ISO8859P2,
    there can be some national specific characters lost with such a conversion.
    But when I use expdp utility and have not NLS_LANG set,
    in log file there is not mentioned that export is done in other character set
    than that on the server, does it mean that it uses character set on server?

  • CFID and CFTOKEN still set when using J2EE sessions

    I'm using CF10 and "Use J2EE session variables" is selected in the CF admin.
    When I visit an application, I get the JSESSIONID cookie, but I also get the CFID and CFTOKEN persistent cookies. The app I'm working with is older and uses Application.cfm instead of Application.cfc, but the clientmanagement and setclientcookies application attributes are set to false.
    I'm not sure why CFID and CFTOKEN are still set. Are they set regardless of the client and session management settings?

    cherdt
    There are ways to force the CF create cookies SESSION simply change the cfcookie to be a session cookie and tag you find these two links:
    http://www.bennadel.com/blog/1131-ask-ben-ending-coldfusion-session-when-user-closes-brows er.htm
    http://www.johnwbartlett.com/cf_tipsntricks/index.cfm?TopicID=75
    BKBK
    I agree with his statement and is so even though the documentation says.

  • Size Property of OracleParameter is not set when used in constructor

    Hi
    I have an issue in using OracleParameter in my .net windows app(vb.net) . If I create a OracleParameter like this, ths size property is not set
    Dim opParam As New OracleParameter("v_Test", OracleDbType.Char, 120, ParameterDirection.Output)
    'Here 120 is the size I have set, but it is not getting assigned to the Size property
    MsgBox(opParam.Size) 'This will return 0
    'I have to assign like this explicitly to have the size property set.
    opParam .Size=120
    Is it a problem in ODP.Net or I am doing something stupid ?.
    Any answers would be greatly appreciated
    Cheers
    John

    Hi John,
    In looking at the constructor call you have this "format":
    string, OracleDbType, int, ParameterDirection
    This does not match to a constructor that does what you want. What it matches to is:
    string, OracleDbType, object, ParameterDirection (i.e. the value (object) is being set where it looks like you want to set the size).
    The closest matching constructor for what you want would be:
    string, OracleDbType, int, object, ParameterDirection
    i.e.
    Dim opParam As New OracleParameter("v_Test", OracleDbType.Char, 120, null, ParameterDirection.Output)Hope that helps a bit (assuming I have not misread, etc...)
    - Mark

  • Trouble with DAQmxReadDigital Timeout (200284) when used in a threaded loop

    Hi,
    I'm using a NI6537 for fast Digital Acquisition. The board is installed on a NIPXIe-1062 crate controlled by a NI PXIe-8130 controller, where the software is running. I'm reading all the 32 lines as a single channel, at 50MS/s, and the clock is the one provided by the board itself. I put the timeout as infinite, but I also tried various other settings. As I'm reading a finite samples bunch of 524288 samples, at 50 MS/s, a timeout of 1 second should be by far sufficient.
    As far as I put the DAQmxReadDigitalU32(...) call in a simple loop I have no troubles, everything working fine.
    for (...) {
       DAQmxReadDigitalU32(...);
    } // That loop works perfectly
    while, if a have a loop inside a multi-thread process, or simply I give the system some freedom inside the loop, sometimes I get the timeout error. Note that the time cost of the Sleep() function is just of some millisecond, so there is no reason for the call to timeout.
    for (...) {
       DAQmxReadDigitalU32(...);
       Sleep(1); // To release system resources...
    }// That loop stops erratically
    With the last code, I can also see that actually the board did not complete the reading, i.e. is returning less samples than what requested. It seems to me that someting is stopping the acquisition, but I'm not able to figure what. The same happens even in LabView, using the VI read: sometimes the read randomly generates a timeout error and stops, even if LabView is the only running application.
    Could you point me what I'm doing wrong?
    Thanks, Piero.
    DAQmxReadDigitalU32

    Duplicate Post
    ColeR
    Field Engineer

Maybe you are looking for

  • I'm trying to install iTunes 10.5 and I get an error message.

    The error says, "There is a problem with this Windows Installer package. A program required for this install to comeplete could not be run. Contact your support personnel or package vendor."

  • Changing the strategy when stocks exist

    Dear Gurus, Few of our FG (FERT)materials were handled as MTO materials and was using strategy 20, now the business case has changed for some of these materials and we want to change the material strategy to 40, These FG's have stock reserved against

  • Strange problems on Macbook Pro over past year...

    I'm about to do an erase and install after having numerous strange problems on my mid 2009 17" Macbook Pro (Leopard) over the past year, some of which I have opened threads for on these forums. They are listed bellow to see if anyone wants to give th

  • Regarding RFC Communication from SAP R/3 to SAP J2EE Engine 6.40

    Hello,      We have an application deployed in SAP J2EE engine 6.20. From SAP R3 (SAP R/3 version is 6.4C), We are communicating to the application deployed in SAP J2EE Engine 6.20 using RFC protocol. EJB module (Deployed in SAP J2EE Engine 6.20) whi

  • 'Display PDF in browser' whats the default?

    Can anyone tell me definitively what the current default setting has been for whether Preferences/Internet Options/'Display PDF in browser' is checked or unchecked? I'm curious to know as well if it has changed with 8, 9, and 10?