Blobs and sql

i was given a new task at work today and it involves automating datbase results. We are given one record and i have to populate the rest of the database to a given amount of records.
i have this blob and i know its an image and i am trying to retrieve it but the getBlob() is giving me an error. I hope someone can help me.

It looks like I have to resort to using an OCIBlobLocator
as I would if not using ANSI Dynamic. I guess I assumed that
since I specify the exact amount I want out and the type to be
raw I wouldn't need the lob-locator. Can anyone tell me why
the code fragment below seg faults rather than returns some
ora/sql error?
I'm trying to retrieve a blob using ANSI Dynamic SQL in a Pro*C application.
However, when I attempt to use FETCH DESCRIPTOR I get a segmentation fault.
The precompiler picks up if the fetch size and the output structure size
don't match, but it seems that they don't actually match according to the
database.
char correctedinb[GAMMA_CHANNELS*4];
rtype = 113;
rsize = GAMMA_CHANNELS*4;
EXEC SQL SET DESCRIPTOR :out_desc VALUE :occurs
TYPE = :rtype,
LENGTH = :rsize,
REF DATA = :correctedinb,
REF INDICATOR = :spectrumind;
occurs++;
EXEC SQL FETCH g INTO DESCRIPTOR :out_desc;

Similar Messages

  • BLOB and SQL server

    BLOB
    suppose i want to insert image into SQL server table.....what would be table design ?
    should i write
    IMAGE   varchar(50);  // its wrong ..so whats right to put a blob in a column ?

    Blobs are a pain. First off there's nothing about
    them in the SQL99 standard, so every database handles
    them differently.
    Worse, whoever wrote JDBC specification doesn't seem
    to have understood how they work.
    A BLOB/CLOB is functionally very different from a
    varable text/binary field, because all that goes into
    the database record itself is some kind of reference
    number. BLOBs are stored outside the tables
    structure, and not normally through SQL. There can be
    several references to the same BLOB from different
    records and databases don't seem to keep track of use
    counts, so you have to delete them explicitly.
    For example in the database I'm using at the moment,
    Postgresql, The reference is declared as an "OID" and
    it's set in a preparedStatement with a setBlob
    method. Saddly there's no legitimate way to create a
    new Blob object. I've known people create an record
    with an empty blob first, then read and update it in
    order to get a Blob object.
    Some JDBC drivers, you can use a setOutputStream
    method, but the specification is pretty vague.Sir,
    While an excellent rant I'm not sure what this has to do with the topic at hand which is how to create a table with a BLOB like column in it in MS-SQL. The OP question really has nothing to with Java at all but we try not let such trivialities stop us.
    Good rant though. Have thought the same things myself many a time.
    Sincerely,
    Slappy

  • Reading BLOB from SQL Server 2005 using DB Link from Oracle 10g

    Hi All,
    I am trying to read a table's data from SQL Server 2005 using ODBC DB Link created in Oracle 10g (10.2.0.3/4) database. I am not using oracle gateway.
    I am able read all data except for the BLOB data from SQL server 2005. It gives error given below when I try to execute query SELECT * from T_TRANSACTION_DATA@sdeslink inside a stored procedure:
    ORA-22992: cannot use LOB locators selected from remote tables
    Kindly help how to read BLOB from SQL server 2005 inside Oracle 10g.
    Best Regards!
    Irfan

    Irfan,
    If you can read the blob data using 10.2 HSODBC then there should be no problem using the 11g gateway.
    I recommend you download the latest version which is 11.2.0.3 available from My oracle Support as -
    Patch 10404530: 11.2.0.3.0 PATCH SET FOR ORACLE DATABASE SERVER
    This is a full version and does not need a previous version to have bene installed. When installing it needs to be in a completlely separate ORACLE_HOME from the existing 11.2 install.
    For the Ora-22992 problem have a look at this note in My Oracle Support -
    Ora-22992 has a workaround in 10gR2 (Doc ID 436707.1)
    Regards,
    Mike

  • Compare blob in SQL statement

    Hi,
    Is it possible to compare blob in SQL statement?
    Example:
    Table1 (id number
    image long raw);
    Table2 (name varchar2
    pic long raw);
    select id, name from table1, table2
    where image = pic

    Hi,
    Try the following and see if it works :
    I have TEXT in the LONG RAW column and not image
    files.
    Method I
    You can use PL/SQL to check if the 2 Long RAW Columns are
    identical or not :
    NOTE :- PL/SQL however has a limitation of processing
    data size less than 32k in size. If the
    Long Raw column exceeds 32k in size then PL/SQL
    cannot be used.
    SQL> desc temp9;
    Name Null? Type
    ID NUMBER
    COL1 LONG RAW
    SQL> desc temp9a;
    Name Null? Type
    ID NUMBER
    COL1 LONG RAW
    SQL> @t
    Id = 1, Long Raw is not the same
    Id = 2, Long Raw is the same
    PL/SQL procedure successfully completed.
    SQL> select id from temp9;
    ID
    1
    2
    SQL> get t
    1 Set ServerOutput On Size 999999;
    2 Declare
    3 CURSOR c1 Is
    4 Select id, col1
    5 from temp9;
    6 xCol1 Temp9a.Col1%Type;
    7 BEGIN
    8 For x In c1 Loop
    9 Select col1
    10 Into xCol1
    11 From temp9a
    12 Where id = x.id;
    13 If ( x.col1 = xCol1) Then
    14 Dbms_Output.Put_Line ('Id = ' || x.id || ', Long Raw is the same');
    15 Else
    16 Dbms_Output.Put_Line ('Id = ' || x.id || ', Long Raw is not the same');
    17 End If;
    18 End Loop;
    19 END;
    20* /
    Method II
    You can create a Database function for the above script and use it within
    your SQL .
    -- Shailender Mehta --

  • Migrating LONG RAW to BLOB and optimizing extent size

    Hi all,
    I got a quite fragmented table with a LONG RAW column I want to migrate to BLOB and defragment.
    DB version is Oracle9i Release 9.2.0.4.0 and this is a production environment.
    I know MOVE and/or CTAS are not possible with LONG RAW columns
    So, how can I do that? Is ALTER TABLE MODIFY the only possibility to migrate from LOING RAW to BLOB?
    Since ALTER TABLE MODIFY will lock the whole table preventing any DML operation, I need at least a rough estimate of the time needed for this operation. How can I do that?
    Since this table is quite fragmented, I also want to rebuilt it using a different extent size.
    I think I should issue a ALTER TABLE MOVE... after having performed the "ALTER TABLE MODIFY".
    Can I do something better to minimize unavailability to DML operations?
    thanks,
    andrea

    Hi,
    Is this an OCCI question?
    I don't see that "to_blob" is documented anywhere. The "to_lob" function can be used to convert long raw columns, but its use is pretty specific and certainly not for general query use.
    Regards,
    Mark
    EDIT1: Well, my local documentation set does not have "to_blob" in it at all. However, it is in the 11.1 SQL Language Reference on OTN:
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28286/functions186.htm#sthref2358
    Despite the fact that the documentation mentions "long raw" the function appears to only work with "raw" data in my 11.1 tests.
    What's your goal here?
    Edited by: Mark Williams on Jun 8, 2009 7:15 PM

  • Saving a file to a BLOB and back...

    Hello everyone.
    I'm trying to save a file to a blob, and recover the blob value later to save to a file, but i'm having problems.
    I'm recording and retrieving the data sucessfuly, but the file written is like corrupted. But it is with the same filesize as the original one.
    This is how i'm doing:
    Recording the file to the database (BLOB):
    <?
    $conn = oci_connect('USER_NAME', 'PASSWORD', '(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=MYSERVER)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=MYSERVICENAME)))');
    $filePath = '/var/www/html/Something.png';
    $fileOut = '/var/www/html/fileback.png';
    $codigo = 5;
    $blob = oci_new_descriptor($conn, OCI_D_LOB);
    $sql =
    "begin
       insert into MYTABLE (COD, FILE) values (:COD, :FILE);
    exception
      when dup_val_on_index then
         update MYTABLE set FILE = :FILE where COD = :COD;
    end;";
    // Parse the query
    $stid = oci_parse($conn, $sql);
    // Bind values
    oci_bind_by_name($stid, ":COD",  $codigo);
    oci_bind_by_name($stid, ":FILE", $blob, -1, OCI_B_BLOB);
    // Get the file
    $file = fopen($filePath, 'rb');
    // Read file content
    $fileContent = fread($file, filesize($filePath));
    // Write file content to the BLOB
    $blob->writeTemporary($fileContent);
    // Execute the statement
    $r = oci_execute($stid, OCI_DEFAULT);
    // Commit
    oci_commit($conn);
    // Close BLOB descriptor
    $blob->close();
    // Release the statement and close the connection
    oci_free_statement($stid);
    oci_close($conn);
    Retrieving the BLOB value:
    $conn = oci_connect('USER_NAME', 'PASSWORD', '(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=MYSERVER)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=MYSERVICENAME)))');
    $filePath = '/var/www/html/Something.png';
    $fileOut = '/var/www/html/fileback.png';
    $codigo = 5;
    $sql = 'select * from MYTABLE where COD = :COD';
    $stmt = oci_parse($conn, $sql);
    oci_bind_by_name($stmt, ":COD", $codigo);
    $result = oci_execute($stmt);
    $a = oci_fetch_array($stmt, OCI_RETURN_NULLS);
    $a['FILE']->export($fileOut);   // Don't work
    //file_put_contents($fileOut, $a['FILE']->read($a['FILE']->size()), FILE_BINARY);   // Don't work
    //file_put_contents($fileOut, $a['FILE']->load(), FILE_BINARY);   // Don't work
    fwrite(fopen($fileOut, 'rwb'), $a['FILE']->load());
    oci_close($conn);If someone can help me with anything i appreciate, thanks.
    Edited by: user1977051 on 12/06/2009 10:33

    I solved it.
    The problem is because the execute after the blob write:
    // Write file content to the BLOB
    $blob->writeTemporary($fileContent);
    // Execute the statement
    $r = oci_execute($stid, OCI_DEFAULT);and should be:
    // Execute the statement
    $r = oci_execute($stid, OCI_DEFAULT);
    // Write file content to the BLOB
    $blob->writeTemporary($fileContent);
    // i'm using this instead
    $blob->saveFile($filePath);Thanks.

  • Help Please:  working with HTML BLOBS and workflow attachments

    simply put need to be able to:
    1. create html document using plsql and store as BLOB in a table.
    2. retrieve this BLOB and send as Document Attachment in workflow notification.
    This sounds so simple to be able to do and yet I've spent the past week trying to find easy to understand documentation on how to do this.
    The html i'm creating is too large for just the normal 32K (I believe) buffer for a regular (text) document attachment. If i limit the # of rows returned, I can use this method just fine. However, my assumption is that with the amount of data I need to return in the attachment is that the BLOB type is what i will need to be able to accomplish this.
    I would be more than happy to e-mail anyone samples of the workflow and code I am currently using to attempt this.
    Lastly, one of the main articles I've looked at is this:
    http://oracle.anilpassi.com/index2.php?option=com_content&do_pdf=1&id=22
    In my latest attempt, i picked an html file currently stored in fnd_lobs and tried to send the file as an attachment.
    wf_engine.setitemattrdocument ( itemtype => c_itemtype,
    itemkey => l_itemkey,
    aname => 'LOB_DOCUMENT',
    documentid =>'PLSQLBLOB:DANA_RLM_EXCEPTION_NOTIFY.xx_notif_attach_procedure/' || to_char(75871));
    I get the following error in Workflow when trying to open the attached document:
    ORA-06502: PL/SQL: numeric or value error: character to number conversion error
    Any help would be greatly appreciated... would consider naming our next child after you. ;)

    I had done some work trying to see how Lotus Notes handles HTML content.. to my dismay, Lotus Notes does not even handle some of the basic HTML correctly.. leave alone the embedded OA Fwk content in the workflow e-mail notifications.
    This seems to be a bug in Lotus notes.. I don't have the link to the IBM site that talks abt the bug. Will post it once I have it.

  • Create Oracle BLOB and Pass to Oracle Stored Procedure

    Hi All,
    I am using Oracle 10g and am dealing with a requirement where I have to upload a file to Oracle Portal from the client's local machine using a JSP.
    I am planning to convert the text file to a BLOB and pass it to an Oracle stored procedure which does the rest.
    I am unable to create a BLOB object. Can anybody help me with this please.
    Is there a better alternative to do it?
    Thanks in advance,
    Shardul

    u can create blob as below...
    java.sql.Blob blob=new Blob();
    File file=new File(fullPathTo UrFile);
    FileInputStream fInSteam = new FileInputStream(file);
    BufferedInputStream bInputStream = new BufferedInputStream(fInSteam);
    ByteArrayOutputStream bOutputStream = new ByteArrayOutputStream();
    int nextByte;
    while (( nextByte = bInputStream.read() ) != - 1)
    bOutputStream.write(nextByte);
    byte[] byteContent = bOutputStream.toByteArray();
    blob.setBytes(byte[]);
    i think it should work for u...

  • BLOB and CLOB data in BI Publisher

    Hello,
    I have an XML file with blob and clob data in it. I am using BI desktop publisher to load the xml file and create the rtf template. I am getting the following error:
    ConfFile: C:\Program Files\Oracle\BI Publisher\BI Publisher Desktop\Template Builder for Word\config\xdoconfig.xml
    Font Dir: C:\Program Files\Oracle\BI Publisher\BI Publisher Desktop\Template Builder for Word\fonts
    Run XDO Start
    Template: C:\work\employee.rtf
    RTFProcessor setLocale: en-us
    FOProcessor setData: C:\work\employee.xml
    FOProcessor setLocale: en-us
    Do i need to install anything or am i missing any configuration to read the blob and clob? Please advise
    thanks

    Check this out, hopefully it'll help you
    http://oraclebizint.wordpress.com/2007/11/12/oracle-bi-ee-101332-working-with-clob-fields/
    Fiston

  • Crystal Reports 9 and SQL Server 2005 default parameter values

    We're using Crystal Reports 9 and upgraded from SQL Server 2000 to SQL Server 2005.
    I'm noticing a very weird problem which I wonder if anyone else has experienced (and hopefully found a resolution for). It appears that in using Crystal with SQL Server 2005 stored procedures, if we have default parameter values in the stored procedures, the default parameter values get completely ignored if you pass in a NULL value from Crystal!
    For example, if you have a stored procedure that begins like this:
    ALTER          Procedure [dbo].[StoredProcedure]
    @Param1 VarChar(200) = '',
    @Param2 VarChar(200) = ''
    AS ...
    both @Param1 and @Param2 have a default value of an empty string, and therefore should become empty strings if nothing (NULL values) gets passed in for them.
    But, like I said, what I'm finding is that with Crystal calling the stored procedure with NULL @Param1 and @Param2 values, they never become empty strings, but rather remain as NULLs.
    This was never a problem with SQL 2000.
    Very perplexing. Anyone else every experience this?
    Thank you.

    Please ignore my earlier post -- answered my own question.
    NULL parameter values do not get replaced by default values in SQL -- that is normal behavior in both SQL 2000 and SQL 2005. Just goes to show, that no many how many years programming experience you have, you can still get tripped up sometimes : (

  • Reading the Blob and writing it to an external file in an xml tree format

    Hi,
    We have a table by name clarity_response_log and content of the column(Response_file) is BLOB and we have xml file or xml content in that column. Most probably the column or table may be having more than 5 records and hence we need to read the corresponding blob content and write to an external file.
    CREATE TABLE CLARITY_RESPONSE_LOG
      REQUEST_CODE   NUMBER,
      RESPONSE_FILE  BLOB,
      DATE_CRATED    DATE                           NOT NULL,
      CREATED_BY     NUMBER                         NOT NULL,
      UPDATED_BY     NUMBER                         DEFAULT 1,
      DATE_UPDATED   VARCHAR2(20 BYTE)              DEFAULT SYSDATE
    )The xml content in the insert statement is very small because of some reason and cannot be made public and indeed we have a very big xml file stored in the BLOB column or Response_File column
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (5, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (6, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (7, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (8, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (9, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');THe corresponding proc for reading the data and writing the data to an external file goes something like this
    SET serveroutput ON
    DECLARE
       vstart     NUMBER             := 1;
       bytelen    NUMBER             := 32000;
       len        NUMBER;
       my_vr      RAW (32000);
       x          NUMBER;
       l_output   UTL_FILE.FILE_TYPE;
    BEGIN
    -- define output directory
       l_output :=
          UTL_FILE.FOPEN ('CWFSTORE_RESPONCE_XML', 'extract500.txt', 'wb', 32760);
       vstart := 1;
       bytelen := 32000;
    ---get the Blob locator
       FOR rec IN (SELECT response_file vblob
                     FROM clarity_response_log
                    WHERE TRUNC (date_crated) = TRUNC (SYSDATE - 1))
       LOOP
    --get length of the blob
    len := DBMS_LOB.getlength (rec.vblob);
          DBMS_OUTPUT.PUT_LINE (len);
          x := len;
    ---- If small enough for a single write
    IF len < 32760
          THEN
             UTL_FILE.put_raw (l_output, rec.vblob);
             UTL_FILE.FFLUSH (l_output);
          ELSE  
    -------- write in pieces
             vstart := 1;
             WHILE vstart < len AND bytelen > 0
             LOOP
                DBMS_LOB.READ (rec.vblob, bytelen, vstart, my_vr);
                UTL_FILE.put_raw (l_output, my_vr);
                UTL_FILE.FFLUSH (l_output);
    ---------------- set the start position for the next cut
                vstart := vstart + bytelen;
    ---------- set the end position if less than 32000 bytes
                x := x - bytelen;
                IF x < 32000
                THEN
                   bytelen := x;
                END IF;
                UTL_FILE.NEW_LINE (l_output);
             END LOOP;
    ----------------- --- UTL_FILE.NEW_LINE(l_output);
          END IF;
       END LOOP;
       UTL_FILE.FCLOSE (l_output);
    END;The above code works well and all the records or xml contents are being written simultaneously adjacent to each other but we each records must be written to a new line or there must be a line gap or a blank line between any two records
    the code which I get is as follow all all xml data comes on a single line
    <?xml version="1.0" encoding="ISO-8859-1"?><emp><empno>7369</empno><ename>James</ename><job>Manager</job><salary>1000</salary></emp><?xml version="1.0" encoding="ISO-8859-1"?><emp><empno>7370</empno><ename>charles</ename><job>President</job><salary>500</salary></emp>But the code written to an external file has to be something like this.
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <emp>
      <empno>7369</empno>
      <ename>James</ename>
      <job>Manager</job>
      <salary>1000</salary>
    </emp>
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <emp>
    <empno>7370</empno>
    <ename>charles</ename>
    <job>President</job>
    <salary>500</salary>
    </emp>Please advice

    What was wrong with the previous answers given on your other thread:
    Export Blob data to text file(-29285-ORA-29285: file write error)
    If there's a continuing issue, stay with the same thread, don't just ask the same question again and again, it's really Pi**es people off and causes confusion as not everyone will be familiar with what answers you've already had. You're just wasting people's time by doing that.
    As already mentioned before, convert your BLOB to a CLOB and then to XMLTYPE where it can be treated as XML and written out to file in a variety of ways including the way I showed you on the other thread.
    You really seem to be struggling to get the worst possible way to work.

  • Workflows 2013 and SQL Report Services: Configure on Independent Servers or Existing App Servers?

    Currently, my SharePoint 2013 Farm has Workflows 2013 and SQL Report Services installed on app servers, along with other applications. Workflows 2013 isn't working correctly (installed prior to my coming on board with the organization), I'm assuming it
    is a configuration issue, but rather than re-configure, I've read that is recommended to be a standalone server. It was also recommended by a MSFT rep that our report services have its own independent server. Currently, Report Services is also installed on
    an app server that runs other applications on the farm.
    Has anyone had any experience with installing Workflows 2013 and/or SQL Report Services on independent vs. existing application servers?
    Additionally, my existing application servers have 24 GB of RAM (I believe the recommendation for SharePoint production environments is 12 GB), should I plan for the potentially new servers (Workflows and Report Services) to also have 24 GB of RAM? Not really
    sure if it will be needed if they have dedicated tasks.
    Thanks for any input you can provide!

    It entirely depends on load. WFM will work just fine installed on servers in the SharePoint farm. SSRS preferably should go on the end-user facing servers for better performance.
    The questions you're asking are all "it depends" :)
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • What is the different between Sharepoint fast search service and Sql server fulltext search?

    HI ,
    I want to kow what is the different between Sharepoint fast search service and Sql server fulltext search?
    Or Can I abstract the Sharepoint fast search from the Sharepoint platform as a isolate component?
    Thank you.
    James

    They are very, very different beasts.
    Firstly FAST Search for SharePoint is the old name for the product and is only relevant for SharePoint 2010 not 2013. It got merged into the standard SharePoint search for the 2013 release.
    SharePoint search is aimed at providing a Bing or Google like experience for your intranet content, as well as providing some nifty features that are purely SharePoint releated along the way. That means it can crawl SharePoint content, file shares,
    outlook mailboxes, internal and external websites and probably fifty other different things if you really tried. Whilst i'm not an expert on SQL full text search I believe it's intended to provide a search feature for content held within SQL databases
    and tables.
    Can you run SharePoint purely for Search? Yes, definitely.

  • Matching data types b/w oracle and sql server

    anyone here knows if there is a list of data types supported by both oracle and sql server (regardless of releases & versions?
    an immediate response would be highly appreciated.
    thanks.

    Hi,
    The following post might be of assistance to you:
    http://msdn.microsoft.com/en-us/library/ms151817.aspx
    If this is what you're looking for, then mark the question as answered and closed.
    Regards,
    Naveed.

  • SQL Report not showing data - available in SQL Workshop and SQL Developer

    I am having an issue with developing a SQL Report in APEX 3.2.1. I run the code in both SQL developer and SQL Workshop and I get data pulled back (both against my development environment). When I run the same code in a SQL Report region, it returns no data available. Does anyone have any idea what would be causing this? Other regions on the page accessing different tables in the same schema return data without issue. Any help would be appreciated.
    Thanks
    Freddie

    Could you explain the last comment a bit more. Here is a bit more info just in case I touch on the info with it. The db schema is BPAMGR, the Workspace is BPAMGR. We use the same schema for all of our reporting. All of our tables are in the same schema. We don't use any tables outside of this schema. Our APEX workspace has been associated to only this schema. The tables are able to be queried by SQL Workshop in the same APEX instance that the report application is under.
    Freddie

Maybe you are looking for