Running data at different rates within a SCTL for resource sharing

I have a design that contains several single cycle timed loops (SCTL):
(40MHz SCTL) -> (20MHz SCTL) -> (40MHz SCTL) -> (10MHz SCTL)
The first 2 SCTLs are decimation filters so that the data coming out of the 20MHz SCTL is
at a 10MHz sampling rate.  The 2ns 40MHz SCTL is filter that is 4X resource
shared so that I can use the available dsp48E slices.  I'm using target-scoped FIFOs to
pass the data between the loops.  In simulation, I have the timeout set to -1 so that it reads
from the FIFO whenever there is data available.
When I run the simulation, I expect/want to see my control counter running 4 times faster
than my incoming data.  I thought that since the data was written into the target-scoped
FIFO at a 10MHz rate and the control counter is in the 40MHz SCTL, I would see the
behavior I wanted:
    X     a     X     b      X      c     X     d     X  <- data from FIFO
    X0X1X2X3X0X1X2X3X0X1X2X3X0X1X2X3X  <- modulo-3 counter
Instead I see the data and counter essentially running at the same rate:
    X1X2X3X4X5X6X7X8X9XaXbXcXdXeXf X0X    <-data from FIFO
    X0X1X2X3X0X1X2X3X0X1X2X3X0X1X2X3X    <-modulo-3 counter
Also the iteration counter for the SCTL runs at the same rate which seems to indicate that the
loop only executes when data is pulled from the FIFO.   I was expecting the iteration counter
to be at 3 before the second sample was read from the FIFO and at 7 before the third sample
was read from the FIFO. I'm guessing that the simulator is event driven.
I have a couple of questions:
1. How can I  get the behavior I'm looking for?
2. If not, can I expect the behavior I want if I run this on the hardware?

update:
NI support suggested using the FIFO timeout signal to determine when the data is valid
coming out of the FIFO.  I was using a -1 on the FIFO read to read only when a sample
was available.  NI support said to use a 0 and then gate the data based on the timeout
signal.  The behavior of the design changed, but still did not give me the timing I was
expecting.  Instead of seeing my counter update on every new sample, I see the SCTL
"free running"  It's difficult to determine if the counter is behaving properly, but I do see it
running faster.  The problem is that the SCTL seems to be running much faster was well.
Using the -1 on the FIFO read I would see > 262K samples written out for every 1000 in.
Using the 0 on the FIFO read and gating the data based on the timeout signal, I see
>60K samples for every 1000 read in. I'm expecting 250 samples out for every 1000 in.
Can someone tell me what controls the behavior (in terms of execution0 of the SCTL?  It
seems like it is event driven, but I need the design to be clock or cycle driven for both
simulation and execution of the hardware.
Regards,

Similar Messages

  • How can I record several channels of data at different rates with the same card?

    I am building an application in which I need to record several channels of DC voltage data to disk at different rates, for example, Channel 1 at 40 Hz, Channels 3-5 at 10 Hz, Channels 6-8 at 1 Hz, etc.
    All these channels are time-referenced, so I need to make sure data for channels recorded at slower rates line up with their correct time stamp.
    Currently, I am building all data into a 2D array one row per loop iteration.

    Hello;
    At this time, the NI DAQ boards have only one scan clock, so, you can only define one valid scan rate for all of your channels.
    The better approach for you is to sample all channels at the highest rate you will need, and then discard the samples you don't need to save.
    Hope this helps.
    Filipe

  • Read and Log Data at Different Rates

    Hi all, I
    'm certain this a fairly easy task with the right training but I just can't seem to figure it out.  What I would like to do is be able to read from a source and display the values at one rate while simultaneously storing the data to be analyzed later at a different rate.  For example if I was to read a thermocouple and display the readings every 3 seconds on the screen but only log the data every ten seconds.  What's the most efficient way to accomplish this?  I've come up with ways that have given me the results I want but I know it wasn't good programming practice.  So in an effort to get better, does anyone have any suggestions?
    Thanks.
    LabVIEW 2012 - Windows 7
    CLAD

    You could try something like the attached vi.
    Cheers!
    CLA, CLED, CTD,CPI, LabVIEW Champion
    Platinum Alliance Partner
    Senior Engineer
    Using LV 2013, 2012
    Don't forget Kudos for Good Answers, and Mark a solution if your problem is solved.
    Attachments:
    Untitled 3.vi ‏11 KB

  • Different Layout within a webflow for one Portal Page

    Hi All,
    Can I set different layouts within one portal page journey depending on the hyperlink
    I click on one of the portal page with in a webflow.
    My requirements is such that if user clicks on one of the link on one of the portlet,
    I want to show them the next page with different content in a different layout
    with in same portal page journey.
    Can anyone post the sample code to achieve this.
    Thanks
    Nitin

    1 - you can not change the photo frame size if that is what you are asking
    2 - you can enlarge the photo within the frame by selecting it and using the slider above the frame
    3 - you can use a different page layout -- or even theme -- see Liz Castro's web site for examples of all available layouts for each theme -- http://www.lizcastro.com/iphotobookthemes/
    LN

  • Read and write data at different rates, parallel loops.

    Machine monitoring and datalogging application.
    imple question, maybe not so simple answer.
    I want to read fieldpoint channels at a certain rate, say every 2 seconds, so I can verify satisfactory operating conditions of a machine under test.
    I then want to write these values to a file at less frequent intervals, say once an hour, as I have no need to record all the data if the machine is running correctly.
    I thought parallel loops with "wait until next ms multiple" would be the way to go, but I am having trouble ensuring that the slower loop actually grabs the values I need. I am not clear on the intricacies of shared variables, local variable, and queues and not sure which way I need to go.
    Semaphores didn't seem to work for me as the faster loop grabbed control and the slower loop couldn't sneak in to read its values.
    Any help would be much appreciated,
    Thanks

     I would use a queue.  As you gather data, enqueue the data.  In the slower loop, process the data queued so far by reading from the queue until it is empty.  You can use the Queue Status vi and then flush the queue to get ready for new data.
    Here is some sample code to get you started.
    Message Edited by tbob on 08-06-2007 06:00 PM
    Message Edited by tbob on 08-06-2007 06:01 PM
    - tbob
    Inventor of the WORM Global
    Attachments:
    Queue.png ‏8 KB

  • Buffered Data Acquisition at different rate

    Hi,
    I have a pb with my data acquisition : I'm using a PXI-4472 AI card and continuous acquiring data to file.
    Until now all channels are acquired at the same frequency : Ok. I know this card does not support multi-frequency acquisition.
    But I want save data at different rate : ex : first channel at the frequency desired (ex 10 KHz), second channel at frequency /2 : 5 Khz ... to reduce amouth of data in the saved file.
    I know the amouth of data retrieve et each AI-Read (ex 2048 scaled data / channel), and the frequency divisor / channel.
    What's the best way (performance) to process the scaled data tab (2 dim) from the AI-READ to save it in the file (using Write File +) with frequency divisor consideration.
    And how to re
    ad the data in the saved file after.
    Can you send me an example please ???
    Thanks
    Eddy DUCHENE
    12 F Chemin de Boutary
    69300 CALUIRE ET CUIRE
    [email protected]

    Hi kdm,
    please stick in one thread for the same topic!
    Best regards,
    GerdW
    CLAD, using 2009SP1 + LV2011SP1 + LV2014SP1 on WinXP+Win7+cRIO
    Kudos are welcome

  • How to display rows of data into different columns?

    I'm new to SQL and currently this is what I'm trying to do: 
    Display multiple rows of data into different columns within the same row
    I have a table like this:
        CREATE TABLE TRIPLEG(
            T#              NUMBER(10)      NOT NULL,
            LEG#            NUMBER(2)       NOT NULL,
            DEPARTURE       VARCHAR(30)     NOT NULL,
            DESTINATION     VARCHAR(30)     NOT NULL,
            CONSTRAINT TRIPLEG_PKEY PRIMARY KEY (T#, LEG#),
            CONSTRAINT TRIPLEG_UNIQUE UNIQUE(T#, DEPARTURE, DESTINATION),
            CONSTRAINT TRIPLEG_FKEY1 FOREIGN KEY (T#) REFERENCES TRIP(T#) );
        INSERT INTO TRIPLEG VALUES( 1, 1, 'Sydney', 'Melbourne');
        INSERT INTO TRIPLEG VALUES( 1, 2, 'Melbourne', 'Adelaide');
    The result should be something like this:
    > T#  | ORIGIN  | DESTINATION1  |  DESTINATION2 
    > 1   | SYDNEY  | MELBORUNE     | ADELAIDE
    The query should include the `COUNT(T#) < 3` since I only need to display the records less than 3. How can I achieve the results that I want using relational views???
    Thanks!!!

    T#
    LEG#
    DEPARTURE
    DESTINATION
    1
    1
    Sydney
    Melbourne
    1
    2
    Melbourne
    Adelaide
    1
    3
    Adelaide
    India
    1
    4
    India
    Dubai
    2
    1
    India
    UAE
    2
    2
    UAE
    Germany
    2
    3
    Germany
    USA
    On 11gr2, you may use this :
      SELECT t#,
             REGEXP_REPLACE (
                LISTAGG (departure || '->' || destination, ' ')
                   WITHIN GROUP (ORDER BY t#, leg#),
                '([^ ]+) \1+',
                '\1')
        FROM tripleg
        where leg#<=3
    GROUP BY t#;
    Output:
    1 Sydney->Melbourne->Adelaide->India
    2 India->UAE->Germany->USA
    Cheers,
    Manik.

  • How to display the header data on different position in alv report

    hi all,
    how to display the header data on different position in alv report.
    for example ,
    customer                                                   name
      xxxx                                                         xxxx
                     vendor        name     street 
                      xxxx         xxxx      xxxxx
    pls   help me .

    hi
    as per my understanding you need to trnasfer header internal table data  to pdf..
    please check the following links for internal table to pdf..
    Convertion of Internal table data to PDF
    This link is related to ur issue
    Re: how to insert the calling of the FM: OTF to PDF
    Thanks

  • Background Job Still Running - Data Can Be Displayed Only Msg Populated in Maintain Versions (OKEQ) Screen

    Hi,
    I try to create 2013 FY in "Setting for Each FY" in OKEQ - Maintain Versions unfortunately I have created Null FY, try to delete Null FY on the time it will ask Job schedule, I have cancelled the job & reopen Maintain Versions (OKEQ) screen on the time system populated the msg "Background Job still running - Data can be displayed only"
    In Settings for Each FY screen there is no New Entries button, all FY are blocked status, changes was not allowed.
    How can I resolve it.
    Waiting for your valuable suggestions.
    Regards,
    Nageswar.

    Dear Rajneesh Saxena
    Thanks for your prompt reply.
    Yes you are correct, on the time of deletion it will ask Background Job Process, for immediate deletion must & should select "NO".
    While "Settings for Each FY" screen populates message "Background Job Still Running - Data Can Be Displayed Only" open  screen in display mode enter the command "=KILL" in command bar press enter, then you will allow to changes get New Entries button also.
    Problem is solved
    Regards,
    Nageswar.

  • Slow running data for heart rate variability VI Please help my thesis is due in 3days!! :(

    Hi there i'm having a great issue with a VI i have created to detect heart rate variabilty from ECG and PPG signals and seeing which one is most accurate with regards to HRV.
    I initially recorded 25 minutes worth of data and saved this into a text file. which has 2 columns the 1st being PPG data the 2nd ECG. There are at least 1000000 rows of data 
    I am now running this data into a VI which should give me results such a RMSDD PNN50, poincare plots and histograms etc  the trouble i'm having is the VI is initially starts of fast then becomes incredibly slow i left it running for over 10hours and it hardly read any data 
    and me not being such a big labview expert knows how to fix this. I would be most grateful if somone could help me find my error as my thesis is due in 3 days and i'm stuck on this last part. I have attached my VI below . please please help me!
    Attachments:
    CV Report1.vi ‏41 KB

    First question: does the code produce the correct result, just slowly?
    So you have a million data points and process them one point at a time in a FOR loop, updating several xy graphs and charts and building arrays. Most indicators will get overwritten with each iteration.
    You are heavily filtering the data (twice!) and take the derivative (twice=second derivative). You are doing a pt-by-pt peak finding on 50 points with every new point added, meaning that you mostly analyze the same data (differing by only one point with each iteration. Seems inefficient. Wouldn't you detect most peaks 50x?
    Since your data is filtered, maybe it would be sufficient to decimate the data before the loop processing. How correlated are adjacent points?
    Have you analyzed which part of the code is slowest? What is the current execution time ("slow" is subjective) and what speed would you like to have instead?
    Things like the histogram only needs to be done exactly once at the end. It seems pointless to do it over and over, a million times in a row even if the data did not change. Same problem for all the graphs.
    You seems to take the difference between adjacent points in the generated array. Again you are repeating the same calculation over and over. Wouldn't it be much more effiicent to keep the array of pairwise differences in the shift register and use a scalar shift register to retain the immediately previous point. Subtract the new point from the previous and add it to the array instead.
    LabVIEW Champion . Do more with less code and in less time .

  • Reports fail when run against a different data source

    Hello,
    We have a VB.NET 2008 WinForms application running on Microsoft .NET 3.5. We are using Crystal Reports 2008 runtime, service pack 3 -- using the CrystalDecisions.Windows.Forms.CrystalReportViewer in the app to view reports. In the GAC on all our client computers, we have versions 12.0.1100.0 and 12.0.2000.0 of CrystalDecisions.CrystalReports.Engine, CrystalDecisions.Shared, and CrystalDecisions.Windows.Forms.
    Please refer to another one of our posted forum issues, u201CCritical issue since upgrading from CR9 to CR2008u201D, as these issues seem to be related:
    Critical issue since upgrading from CR9 to CR2008
    We were concerned with report display slow down, and we seemed to have solved this by using the Oracle Server driver (instead of either Microsoft's or Oracle's OLEDB driver).  But now we must find a resolution to another piece of the puzzle, which is:  why does a report break if one data source is embedded in the .rpt file is different than the one you are trying to run the report against, in the .NET Viewer?
    Problem:
    If you have a production database name (e.g. "ProdDB") embedded in your .rpt file that you built your report from and try to run that report against a development database (e.g. "DevDB") (OR VICE VERSA -- it is the switch that is the important concept here), the report fails with a list of messages such as this:
        Failed to retrieve data from the database
        Details:  [Database vendor code: 6550 ]
    This only seems to happen if the source of the report data (i.e. the underlying query) is an Oracle stored procedure or a Crystal Reports SQL Command -- the reports run fine against all data sources if the source is a table or a view).  In trying different things to troubleshoot this, including adding a ReportDocument.VerifyDatabase() call after setting the connection information, the Crystal Reports viewer will spit out other nonsensical errers regarding being unable to find certain fields (e.g. "The field name is not known), or not able to find the table (even though the source data should be coming from an Oracle stored procedure, not a table).
    When the reports are run in the Crystal Reports Designer, they run fine no matter what database is being used; but the problem only happens while being run in the .NET viewer.  It's almost as if something internally isn't getting fully "set" to the new data source, or something -- we're really grasping at straws here.
    For the sake of completeness of information, here is how we're setting the connection information
            '-- Set database connection info for the main report
            For Each oConnectionInfo In oCrystalReport.DataSourceConnections
                oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
            Next oConnectionInfo
            '-- Set database connection info for each subreport
            For Each oSubreport In oCrystalReport.Subreports
                For Each oConnectionInfo In oSubreport.DataSourceConnections
                    oConnectionInfo.SetConnection(gsDBDataSource, "", gsDBUserID, gsDBPassword)
                Next oConnectionInfo
            Next oSubreport
    ... but in troubleshooting, we've even tried an "overkill" approach and added this code as well:
            '-- Set database connection info for each table in the main report
            For Each oTable In oCrystalReport.Database.Tables
                With oTable.LogOnInfo.ConnectionInfo
                    .ServerName = gsDBDataSource
                    .UserID = gsDBUserID
                    .Password = gsDBPassword
                    For Each oPair In .LogonProperties
                        If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                            oPair.Value = gsDBDataSource
                            Exit For
                        End If
                    Next oPair
                End With
                oTable.ApplyLogOnInfo(oTable.LogOnInfo)
            Next oTable
            '-- Set database connection info for each table in each subreport
            For Each oSubreport In oCrystalReport.Subreports
                For Each oTable In oSubreport.Database.Tables
                    With oTable.LogOnInfo.ConnectionInfo
                        .ServerName = gsDBDataSource
                        .UserID = gsDBUserID
                        .Password = gsDBPassword
                        For Each oPair In .LogonProperties
                            If UCase(CStr(oPair.Name)) = "DATA SOURCE" Then
                                oPair.Value = gsDBDataSource
                                Exit For
                            End If
                        Next oPair
                    End With
                    oTable.ApplyLogOnInfo(oTable.LogOnInfo)
                Next oTable
            Next oSubreport
    ... alas, it makes no difference.  If we run the report against a database that is different than the one specified with "Set Datasource Location" in Crystal, it fails with nonsense errors 

    Thanks for the reply, Ludek.  We have made some breakthroughs, uncovered some Crystal bugs and workarounds, and we're probably 90% there I hope.
    For your first point, unfortunately the information on the Oracle 6550 error was generic, and not much help in our case.  And for your second point, the errors didn't have anything to do with subreports at that time -- the error would manifest itself even in a simple, one-level report.
    However, your third point (pointing us to KB 1553921) helped move us forward quite a bit more.  For the benefit of all, here is a link to that KB article:
    Link: [KB 1553921|http://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/oss_notes_boj/sdn_oss_boj_bi/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/scn_bosap/notes%7B6163636573733d36393736354636443646363436353344333933393338323636393736354637333631373036453646373436353733354636453735364436323635373233443330333033303331333533353333333933323331%7D.do]
    We downloaded the tool referenced there, and pointed it at a couple of our reports.  The bottom line is that the code it generated uses a completely new area of the Crystal Reports .NET API which we had not used before -- in the CrystalDecisions.ReportAppServer namespace.  Using code based on what that RasConnectionInfo tool generated, we were able gain greater visibility into some of the objects in the API and to uncover what I think qualifies as a genuine bug in Crystal Reports.
    The CrystalDecisions.ReportAppServer.DataDefModel.ISCRTable class exposes a property called QualifiedName, something that isn't exposed by the more commonly-used CrystalDecisions.CrystalReports.Engine.Table class.  When changing the data source with our old code referenced above (CrystalDecisions.Shared.ConnectionInfo.SetConnection), I saw that Crystal would actually change the Table.QualifiedName from something like "SCHEMAOWNER.PACKAGENAME.PROCNAME" to just "PROCNAME" (essentially stripping off the schema and package name).  Bad, Crystal...  VERY BAD!  IMHO, Crystal potentially deserves to be swatted on the a** with the proverbial rolled-up newspaper.
    I believe this explains why we were also able to generate errors indicating that field names or tables were not found -- because Crystal had gone and changed the QualifiedName to remove some key info identifying the database object!  So, knowing this and using the code generated by the RasConnectionInfo tool, we were able to work around this bug with code that worked for most of our reports ("most" is the key word here -- more on that in a bit).
    So, first of all, I'll post our new code.  Here is the main area where we loop through all of the tables in the report and subreports:
    '-- Replace each table in the main report with new connection info
    For Each oTable In oCrystalReport.ReportClientDocument.DatabaseController.Database.Tables
        oNewTable = oTable.Clone()
        oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
        oCrystalReport.ReportClientDocument.DatabaseController.SetTableLocation(oTable, oNewTable)
    Next oTable
    '-- Replace each table in any subreports with new connection info
    For iLoop = 0 To oCrystalReport.Subreports.Count - 1
        sSubreportName = oCrystalReport.Subreports(iLoop).Name
        For Each oTable In oCrystalReport.ReportClientDocument.SubreportController.GetSubreportDatabase(sSubreportName).Tables
            oNewTable = oTable.Clone()
            oNewTable.ConnectionInfo = GetNewConnectionInfo(oTable)
            oCrystalReport.ReportClientDocument.SubreportController.SetTableLocation(sSubreportName, oTable, oNewTable)
        Next oTable
    Next iLoop
    '-- Call VerifyDatabase() to ensure that the tables update properly
    oCrystalReport.VerifyDatabase()
    (Thanks to Colin Stynes for his post in the following thread, which describes how to handle the subreports):
    Setting subreport connection info at runtime
    There seems to be a limitation on the number of characters in a post on this forum (before all formatting gets lost), so please see my next post for the rest....

  • How can I read published data from a datasocket server running in a different network?

    Hi all,
    I have been trying to solve this problem but without any success. I want to develop data acquisition Vi to run on a computer on a remote location. I want to use the datasocket technology to acquire and publish this data so that in my office (in a different network), I can read the published data and perform analysis with it. I can successfully use this approach for the two computers in the same network but not when they are in different networks. My real problem is how to specify the URL for the acquisition computer in one network while the analysis Vi runs on a different network.
    I have attached two Vis to illustrate my point (I simply want to be able to read the random numbers generated in the acquisition computer in a different network). The first Vi (RemoteDatasocketWrite.vi) will run on the remote computer with an IP address (192.168.0.110). My office computer on which RemoteDatasocketRead.vi runs has an IP address (192.168.0.11) and I can log in remotely into the remote computer using Remote Desktop Protocol with name (emelvin.001.eairlink.com) from my office computer.
     Is it possible to do what I am trying to do or is there a simpler way to solve this problem?
    I will really appreciate any help towards a solution to this problem.
    Thanks
    Attachments:
    RemoteDatasocketWrite.vi ‏9 KB
    RemoteDatasocketRead.vi ‏9 KB

    Are you getting an error? The way you have it set up, you can have an error and it will never be displayed. Put and error control on your front panel and see what it gives you. Also, shared variables in a project can be useful. Look at some examples for that.

  • How to run the Invoices in different groups within the batch in AP

    Need to know that how to run the Invoices in different groups within the batch. This will be of very helpful when we deal with lot of lines under one batch. (e.g) like somewhere we issued a Corporate Card to all the employees Via Bank Of America. Every Month they will send the complete details of all the employees who ever swiped the corporate provided BOA. Accordding to Natco all those lines should be loaded as One Invoice so that a Single Payment can be provided to BOA and it will also makes their life easier. This standard program sometimes it works normal sometimes it will run like a TORTOISE. So thats why in the manual they suggested to use GROUP ID to split the Invoice Load.
    So plz tell me how we can run it
    plz give me the solution
    thanks

    can you give me some material or document on that so that i can read actuaaly i need to make doc on it..

  • How data flow when SSIS packages are run on a different server than the DB server

    The scenario is that i have a dedicated SQL SErver 2014 SSIS machine that executes the packages.
    The database server is a separate machine with SQL Server 2008R2.
    1) Running SSIS packages that transfer data within SQL Server 2008R2 (same machine)
    2) Running SSIS packages that transfer data between 2 separate SQL Server servers.
    How the data flow in these two cases and what resource is being used where? (cpu,disk,ram,network)
    Elias

    When you have a dedicated SSIS server, all data read flows to that server, is processed using the resources of that ETL server and then sent back over the network to the destination server.
    It doesn't matter if source and destination are the same server. If you use a data flow, all data flows over the network twice.
    The only exception is when you don't use a data flow, but only SQL statements. In that case, data flows only between source and destination.
    MCSE SQL Server 2012 - Please mark posts as answered where appropriate.

  • How to read data with different XML schemas within the single connection?

    I have Oracle 11g database
    I access it through jdbc:oracle:thin, version 11.2.0.3, same as xdb.
    I have several tables, each has one XMLType column, all schema-based.
    There are three different XML schemata registered in the DB
    I may need to read the XML data from several tables.
    If all the XMLTypes have the same XML schema ,there is no problem,
    If the schemata are different, the second read throws BindXMLException.
    If I reset the connection between the reads of the XMLType column with different schemata, it works.
    The question is: how can I configure the driver, or the connection to be able to read the data with different XML schemata without resetting the connection (which is expensive).
    The code to get the XMLType data is textbook implementation:
    1   ResultSet resultSet = statement.executeQuery( sql ) ;
    2   String result = null ;
    3    while(resultSet.next()) {
    4   SQLXML sqlxml = resultSet.getSQLXML(1) ;
    5   result = sqlxml.getString() ;
    6   sqlxml.free();
    7   }
    8   resultSet.close();
    9    return result ;

    It turns out, that I needed to serialize the XML on the server and read it as Blob. Like this:
    1    final Statement statement = connection.createStatement() ;
    2    final String sql = String.format("select xmlserialize(content xml_content_column as blob encoding 'UTF-8') from %s where key='%s'", table, key ) ;
    3   ResultSet resultSet = statement.executeQuery( sql ) ;
    4   String result = null ;
    5    while(resultSet.next()) {
    6   Blob blob = resultSet.getBlob( 1 );
    7   InputStream inputStream = blob.getBinaryStream();
    8   result = new Scanner( inputStream ).useDelimiter( "\\A" ).next();
    9   inputStream.close();
    10   blob.free();
    11   }
    12   resultSet.close();
    13   statement.close();
    14
    15   System.out.println( result );
    16    return result ;
    17
    Then it works. Still, can't get it work with XMLType in resultset.On the client unwrapping XML blows up when trying to switch to different XML schema. JDBC/XDB problem?

Maybe you are looking for

  • Which port number to use in remote site definition? [was:Help!]

    I recently upgraded from Dreamweaver 8 to Dreamweaver cs6.  I'm having trouble accessing my website because I dont know which port to choose.  I never had to choose a port in Dreamweaver 8.  How does one decide which port to use?  Do you just keep pi

  • Dual 23-inch HD Mac Displays

    I currently have a 1.66mhz Intel Mac Mini with 2 GB of operating memory, and a 23-inch HD Display. I would like to add another display. What do I need technically???

  • Difficulty finding downloaded CS6 (upgraded to CS5) for Win after following all recommended steps

    Hi there - would appreciate a hand on what I am sure is a very basic request - I purchased upgrade CS5 to CS6 online about 3 days ago and signed in and located my order, clicked on it and proceeded to what I thought was 'download', waited for it to d

  • Event when execute CJ46

    Hi all body , I create a program(ZPS_TRANSFERT)  that transfers amounts to an external application. My problem is how to run this program(ZPS_TRANSFERT) through a job after the déclechement the event of execution of the transaction cj46. Thank you fo

  • The Print Screen Key under Windows

    I just picked up my Apple aluminum keyboard (with the numeric keypad) and absolutely think it's the bomb. I have gotten to the point where any other keyboard just isn't as easy to use, and I do a LOT of typing at work. That being said, there are a fe