EA 2.1. Can't export CLOB columns.
Hello:
I don't know the reason why CLOB column export is not available or incorrect...
In EA 2.1. version SQL developer "truncates" the CLOB field longer than 4000 characters..
Why?
*I can do that succesfully with TOAD.....*
Because it's a missing feature.
Any enhancement requests should go to the SQL Developer Exchange, so others can vote and add weight for possible future implementation.
Regards,
K.
Similar Messages
-
How can I export image column from SQL Server 2000
Hi everybody,could someone tell me how can I export an entire column( pictures type) to a folder on disk?I've tried the textcopy.exe under the SQL installation's binn folder,it can only export one picture a time,I'm trying to find an automated way to export
the whole colum's images to a folder,I'm using SQL Server 2000.Vishal,thanks for the replay,I've tried the BCP method but it seems does not work on my Sql Server 2000,also the SSIS package method is demonstrated with Sql Server 2008 R2 and my machine does not have the visual application to try it.In the BCP method,I
got prompt indicating me that I wrote something wrong in the grammer,and I'm pretty sure that maybe I made something wrong in the below part,the grammer makes me confused
INSERT
INTO @sqlStatements
SELECT 'BCP "SELECT Photograph_Data FROM [ALBSCH_Trial].[dbo].[Photograph] WHERE Photograph_ID = '''
+ CAST(Photograph_ID AS VARCHAR(500)) + '''" queryout ' + @OutputFilePath
+ CAST(Photograph_ID AS VARCHAR(500)) + '.jpg -S localhost\SQLEXPRESS2008 -T -f C:\SQLTest\Images.fmt'
FROM dbo.Photograph -
Can we Export selected columns of a table in ADF 11g
Hi,
My problem is I want a table to be exported to Excel format.But my condition is only particular cloumns of a table to be exported not all the visible columns of the table.
For example, I have 8 cloumns in a table.I need pre-determined 5 cloumns only to be exported into Excel format.I dont want to use PanelCollection Views-->Collumns to be sellected.
Sorry,I dont know whether this is a valid question or not.But Im looking for this condition.
Can you please help me.
Regards,
FelixI dont know how to continue.
I have emloyee table in the HRSchema.
I used employee table in the PanelCollection
All the collumns are in display.
Here in a commandButton I use a ExportCollectionListener for which id i used the employee table id.
If I click the button it exports all collumns
I dont know what condition I can give in the EL expression in the render or visible property of the collumn.
can you please give further more ideas,
Regards,
Felix -
Can't export XML column via postgres odbc driver
Hi,
I'm developing a process to export rows from a table in SQL server 2012 to a table in a Postgres database (9.3).
I'm using the postgres odbc driver in a linked server is SQL server.
I can read and insert across this linked server fine, however I have an XML column in the in the table and the process is failing with the following error.
INSERT into openquery(postgres_linked_server, 'SELECT column1, xml_column from mytable')
select top 1 column1, xml_column data_from mytable
Implicit conversion from data type xml to nvarchar is not allowed. Use the CONVERT function to run this query.
I've tried casting the source row to varchar/nvarchar, I've tried inserting it into a staging table in postgres with a character varying field. I'm reaching the conclusion that I need to extract the xml from sql server and insert it into postgres using an
intermediary. I'll try this with SSIS.
I was wondering if anyone had any experience of doing this with a linked server.
any comments appreciated.
Seansorry for the delay in replying, I've tried the second type of syntax and get
The OLE DB provider "MSDASQL" for linked server "postgres" supplied inconsistent metadata for a column. The column "xml_column" (compile-time ordinal 6) of object "postgres_linked_server"."database.schema"."mytable"
was reported to have a "DBCOLUMNFLAGS_ISLONG" of 128 at compile time and 0 at run time.
I'll look into it a little more, but I will try using an SSIS package as well to see if
I have any more luck with that.
thanks
Sean
Sean -
Data Pump Export with Column Encryption
Hi All-
I have the following two questions:
- Can I export encrypted column data to a file, preserving the data encrypted? (ie. I don't want clear text in resulting file)
- If yes, what would another person need to decrypt the generated file? (ie. I want to give the resulting file to a customer)
Thanks a lot for your help!
JuanHi,
expdp and impdp works only on oracle database. You you need to have oracle database on source and destination to use expdp/impdp. You can also use expdp with ENCRYPTION_PASSWORD option so that password can be passed to client for doing impdp at their end with given password.
Regards -
Exporting Table with CLOB Columns
Hello All,
I am trying to export table with clob columns with no luck. It errors saying EXP-00011TABLE do not exist.
I can query the table and the owner is same as what i am exporting it from.
Please let me know.An 8.0.6 client definitely changes things. Other posters have already posted links to information on what versions of exp and imp can be used to move data between versions.
I will just add that if you were using a client to do the export then if the client version is less than the target database version you can upgrade the client or better yet if possilbe use the target database export utility to perform the export.
I will not criticize the existance of an 8.0.6 system as we had a parent company dump a brand new 8.0.3 application on us less than two years ago. We have since been allowed to update the database and pro*c modules to 9.2.0.6.
If the target database is really 8.0.3 then I suggest you consider using dbms_metadata to generate the DDL, if needed, and SQLPlus to extact the data into delimited files that you can then reload via sqlldr. This would allow you to move the data with some potential adjustments for any 10g only features in the code.
HTH -- Mark D Powell -- -
Export data from table with CLOB column
DB: Oracle10g
SQLDeveloper: 1.0
OS: windows xp for sqldeveloper
Linux for Oracle db
we have a table with two CLOB columns. We want to export the data to either text or csv. But the CLOB columns don't show up in the columns that can be exported.
How do I export CLOB data from SQL Developer?
Regards,
LonnekeI don't think this is a good protection: you can have these characters in VARCHAR anyway, and this way you "throw the baby out with the bath water". Not mentioning that right now also immune formats like HTML are also limited.
-
Hi everyone, again landed up with a problem.
After trying a lot to do it myself, finally decided to post here..
I have created a form in form builder 6i, in which on clicking a button the data gets exported to excel sheet.
It is working fine with a single table. The problem now is that i am unable to do the same with 2 tables.
Because both the tables have same number of columns & column names.
Below are 2 tables with column names:
Table-1 (MONTHLY_PART_1)
Table-2 (MONTHLY_PART_2)
SL_NO
SL_NO
COMP
COMP
DUE_DATE
DUE_DATE
U-1
U-1
U-2
U-2
U-4
U-4
U-20
U-20
U-25
U-25
Since both the tables have same column names, I'm getting the following error :
Error 402 at line 103, column 4
alias required in SELECT list of cursor to avoid duplicate column names.
So How can i export the data to excel which has 2 tables with same number of columns & column names?
Should i paste the code? Should i post this query in 'SQL and PL/SQL' Forum?
Help me with this please.
Thank You.You'll have to *alias* your columns, not prefix it with the table names:
$[CHE_TEST@asterix1_impl] r
1 declare
2 cursor cData is
3 with data as (
4 select 1 id, 'test1' val1, 'a' val2 from dual
5 union all
6 select 1 id, '1test' val1, 'b' val2 from dual
7 union all
8 select 2 id, 'test2' val1, 'a' val2 from dual
9 union all
10 select 2 id, '2test' val1, 'b' val2 from dual
11 )
12 select a.id, b.id, a.val1, b.val1, a.val2, b.val2
13 from data a, data b
14 where a.id = b.id
15 and a.val2 = 'a'
16 and b.val2 = 'b';
17 begin
18 for rData in cData loop
19 null;
20 end loop;
21* end;
for rData in cData loop
ERROR at line 18:
ORA-06550: line 18, column 3:
PLS-00402: alias required in SELECT list of cursor to avoid duplicate column names
ORA-06550: line 18, column 3:
PL/SQL: Statement ignored
$[CHE_TEST@asterix1_impl] r
1 declare
2 cursor cData is
3 with data as (
4 select 1 id, 'test1' val1, 'a' val2 from dual
5 union all
6 select 1 id, '1test' val1, 'b' val2 from dual
7 union all
8 select 2 id, 'test2' val1, 'a' val2 from dual
9 union all
10 select 2 id, '2test' val1, 'b' val2 from dual
11 )
12 select a.id a_id, b.id b_id, a.val1 a_val1, b.val1 b_val1, a.val2 a_val2, b.val2 b_val2
13 from data a, data b
14 where a.id = b.id
15 and a.val2 = 'a'
16 and b.val2 = 'b';
17 begin
18 for rData in cData loop
19 null;
20 end loop;
21* end;
PL/SQL procedure successfully completed.
cheers -
Can clob columns be selected across a database link (synonym)?
Hello,
I would like to do something like this...
insert into archive_table (id, msg, start_date)
select id, msg, startdate from synonym_table
where start_date < whatever;
...where msg represents a clob column and synonym_table represents a synonym in the current table space which accesses another Oracle table using a database link.
Is this possible?http://asktom.oracle.com/pls/ask/f?p=4950:8:15261108079829288196::NO::F4950_P8_DISPLAYID,F4950_P8_CRITERIA:950029833940
-
XML Import CLOB column size limit
I have a table with a CLOB column and some of the rows have up to 7000 characters in the column. I exported the table to XML and the XML file contains all the data. When I try to import the file into another APEX instance I get the error:
XML Load error.
After some experimentation, I found that If I manually edit the XML to reduce the size of the text to under 4000 characters (3700 in my test), it imports fine.
Is there a way around this limitation? The database I'm migrating has LOTS of CLOB columns (converted from MS Access "memo" fields).jlange,
Having converted a bunch of MS Access applications myself, I would encourage you to look at the Oracle Migration Workbench (OMWB): http://www.oracle.com/technology/tech/migration/index.html
This free tool can be downloaded from OTN, and provides a more streamlined approach to moving the data from MS Access to Oracle, including support for Memos to CLOBs.
Once you're data has been moved over, you can then use ApEx to re-create the UI.
Thanks,
- Scott - -
Hi,
Below question is for oracle 11g database on RHEL.
we are having a CLOB column of 8K. Now the developer is asking to get that increase to 32k how can i make that?
As i know we cannot change "chunk" size once after it is created. I thought we need to recreate a new table with chunk 32k and then export/import the data from old table and drop that old table.
But, the point where i'm struck is ,
the DB_BLOCK_SIZE is 8k. So, can we create a new table with 32k chunk size. Does it conflict?
If thats not possible, i thought an alternative of recreating a tablespace with 32k. I heard that, we can create new tablespace with different db_block_size.. i'm not sure if i'm right? and if you don't mind can you just explain me the lob in-line and out-line.. i read the documents but, i didn't get that well
Can you please suggest me the better way and correct me if i'm wrong..These steps might help:
How to alter LOB storage definition.
You may use the following to alter the table and the associated LOB allowing for more extents:
alter table owner.table_name
modify LOB(lob_column_name)
(storage (next 100M MAXEXTENTS 1000));
Example:
User is getting the following error:
APP-FND-00565: The export failed in APP_EXPORT.DUMP_DATA
PL/SQL ERROR: ORA-01691: unable to extend lob segment APPLSYS.SYS_LOB0000033586C00004$$ by 311075 in tablespace APPLSYSD
Looking at tablespace APPLSYSD I can see that it does have room to extent at datafile level for another 800 MB. And TEMP has plenty of room as well.
Using the SQL provided in this page: Looking at the LOB:
SYS_LOB0000033586C00004$$
It looks like Column FILE_DATA of datatype LOBSEGMENT in table APPLSYS.FND_LOBS has grown to be 5 GB in size and was trying to extent itself grabbing another 2.3 GB (NEXT_EXTENT is 2548301824) which of course it could not do, I modified it so it can extent only by another 100 MB.
SQL> alter table APPLSYS.FND_LOBS modify lob (FILE_DATA)
(storage (maxextents 1000 next 100M));
Table altered.
For more detail see: http://www.idbasolutions.com/database/find-lob.html -
Can't fetch clob and long in one select/query
I created a nightmare table containing numerous binary data types to test an application I was working on, and believe I have found an undocumented bug in Oracle's JDBC drivers that is preventing me from loading a CLOB and a LONG in a single SQL select statement. I can load the CLOB successfully, but attempting to call ResultSet.get...() for the LONG column always results in
java.sql.SQLException: Stream has already been closed
even when processing the columns in the order of the SELECT statement.
I have demonstrated this behaviour with version 9.2.0.3 of Oracle's JDBC drivers, running against Oracle 9.2.0.2.0.
The following Java example contains SQL code to create and populate a table containing a collection of nasty binary columns, and then Java code that demonstrates the problem.
I would really appreciate any workarounds that allow me to pull this data out of a single query.
import java.sql.*;
This class was developed to verify that you can't have a CLOB and a LONG column in the
same SQL select statement, and extract both values. Calling get...() for the LONG column
always causes 'java.sql.SQLException: Stream has already been closed'.
CREATE TABLE BINARY_COLS_TEST
PK INTEGER PRIMARY KEY NOT NULL,
CLOB_COL CLOB,
BLOB_COL BLOB,
RAW_COL RAW(100),
LONG_COL LONG
INSERT INTO BINARY_COLS_TEST (
PK,
CLOB_COL,
BLOB_COL,
RAW_COL,
LONG_COL
) VALUES (
1,
'-- clob value --',
HEXTORAW('01020304050607'),
HEXTORAW('01020304050607'),
'-- long value --'
public class JdbcLongTest
public static void main(String argv[])
throws Exception
Driver driver = (Driver)Class.forName("oracle.jdbc.driver.OracleDriver").newInstance();
DriverManager.registerDriver(driver);
Connection connection = DriverManager.getConnection(argv[0], argv[1], argv[2]);
Statement stmt = connection.createStatement();
ResultSet results = null;
try
String query = "SELECT pk, clob_col, blob_col, raw_col, long_col FROM binary_cols_test";
results = stmt.executeQuery(query);
while (results.next())
int pk = results.getInt(1);
System.out.println("Loaded int");
Clob clob = results.getClob(2);
// It doesn't work if you just close the ascii stream.
// clob.getAsciiStream().close();
String clobString = clob.getSubString(1, (int)clob.length());
System.out.println("Loaded CLOB");
// Streaming not strictly necessary for short values.
// Blob blob = results.getBlob(3);
byte blobData[] = results.getBytes(3);
System.out.println("Loaded BLOB");
byte rawData[] = results.getBytes(4);
System.out.println("Loaded RAW");
byte longData[] = results.getBytes(5);
System.out.println("Loaded LONG");
catch (SQLException e)
e.printStackTrace();
results.close();
stmt.close();
connection.close();
} // public class JdbcLongTestThe problem is that LONGs are not buffered but are read from the wire in the order defined. The problem is the same as
rs = stmt.executeQuery("select myLong, myNumber from tab");
while (rs.next()) {
int n = rs.getInt(2);
String s = rs.getString(1);
The above will fail for the same reason. When the statement is executed the LONG is not read immediately. It is buffered in the server waiting to be read. When getInt is called the driver reads the bytes of the LONG and throws them away so that it can get to the NUMBER and read it. Then when getString is called the LONG value is gone so you get an exception.
Similar problem here. When the query is executed the CLOB and BLOB locators are read from the wire, but the LONG is buffered in the server waiting to be read. When Clob.getString is called, it has to talk to the server to get the value of the CLOB, so it reads the LONG bytes from the wire and throws them away. That clears the connection so that it can ask the server for the CLOB bytes. When the code reads the LONG value, those bytes are gone so you get an exception.
This is a long standing restriction on using LONG and LONG RAW values and is a result of the network protocol. It is one of the reasons that Oracle deprecates LONGs and recommends using BLOBs and CLOBs instead.
Douglas -
Can we export DATA from all tables in a schema?
Hi,
I have a question; Can we export all the DATA from all the tables present in the schema to any source (eigther CSV, TXT, DAt, etc..)?
Or
Can I have a PL/SQL procedure to display DATA from all Tables in a schema?
With Best Regards,
- NaveedHi,
This is pretty much what you need.
DECLARE
V_COUNT NUMBER;
v_sql varchar2(1000);
IN_OWNER_NAME VARCHAR2(100) := 'AP' ; -- SCHEMA NAME
TYPE T_COL_NAME is table of varchar2(1000) index by binary_integer;
v_col_tbl t_col_name;
BEGIN
FOR i in
(SELECT object_name
FROM dba_objects
WHERE owner = IN_OWNER_NAME
AND object_type ='TABLE'
and rownum < 2)
LOOP
v_sql := 'SELECT COUNT(*) FROM ' || i.object_name ;
EXECUTE IMMEDIATE v_sql INTO V_COUNT;
if v_count > 0 then
v_sql := 'SELECT * FROM ' || i.object_name ;
select column_name
bulk collect
into v_col_tbl
from DBA_TAB_COLUMNS
WHERE TABLE_NAME = I.OBJECT_NAME
AND OWNER = IN_OWNER_NAME;
-- start selecting the column and exporting using the column names selected.
end if;
if v_col_tbl.count > 0 then
for i in v_col_tbl.first .. v_col_tbl.last
loop
DBMS_OUTPUT.PUT_lINE(v_col_tbl(i));
end loop;
end if;
DBMS_OUTPUT.PUT_lINE( i.object_name || '-' || v_count);
END LOOP;
END;
- Ronel -
How can we get Dynamic columns and data with RTF Templates in BI Publisher
How can we get Dynamic columns and data with RTf Templates.
My requirement is :
create table xxinv_item_pei_taginfo(item_id number,
Organization_id number,
item varchar2(4000),
record_type varchar2(4000),
record_value CLOB,
State varchar2(4000));
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'USES','fever','TX');
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'HOW TO USE','one tablet daily','TX');
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'SIDE EFFECTS','XYZ','TX');
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'DRUG INTERACTION','ABC','TX');
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'OVERDOSE','Go and see doctor','TX');
insert into xxinv_item_pei_taginfo values( 493991 ,224, '1265-D30', 'NOTES','Take after meal','TX');
select * from xxinv_item_pei_taginfo;
Item id Org Id Item Record_type Record_value State
493991 224 1265-D30 USES fever TX
493991 224 1265-D30 HOW TO USE one tablet daily TX
493991 224 1265-D30 SIDE EFFECTS XYZ TX
493991 224 1265-D30 DRUG INTERACTION ABC TX
493991 224 1265-D30 OVERDOSE Go and see doctor TX
493991 224 1265-D30 NOTES Take after meal TX
Above is my data
I have to fetch the record_type from a lookup where I can have any of the record type, sometime USES, HOW TO USE, SIDE EFFECTS and sometimes some other set of record types
In my report I have to get these record typpes as field name dynamically whichever is available in that lookup and record values against them.
its a BI Publisher report.
please suggestif you have data in db then you can create xml with needed structure
and so you can create bip report
do you have errors or .... ? -
Can I Export to Text File?
hi,
can i export my ical events to a text file?
thanks.
samanthaI have figured out a way to do this with free available programs:
Download this program:
http://www.nirsoft.net/utils/faview.html
Faview will allow the bookmarks exported in html format to be saved as a text file. This text file is in a uniform state.
Open the text file with a spreadsheet program. I used the free program called spread32.exe that can be found anywhere on the internet.
Highlight all the columns and do a sort by Column A. The urls will now be in alphabetical order at the bottom of the spreadsheet. Cut and paste into a new text file.
Maybe you are looking for
-
Lion Recovery from a Remote Time Machine Disk
I maintain the time machine backup for my MBP on a large USB drive hanging off of my Mac Mini. Yesterday I decided to upgrade the internal hard drive on my system to a faster, higher-capacity disk. I assumed that this would be very straightfoward
-
Trying to run Migration Assistant on a PC (Windows XP). Error message "Outlook Express could not be started. The application was unable to initialize the store directory. Your computer may be out of memory or your disk is full. Contact Microsoft s
-
How do I add new contacts from email in Gmail app
I recently just switched over from the native iPhone Mail app to the Gmail app and I'm enjoying the Gmail app a lot. However, I have one issue that I can seem to figure out. With the Apple Mail app, inside any email you can hold down on the sender an
-
Nokia n8 mp4!!!
how come the n8 wont play mp4 files when I select them in the videos app? It says "unable to play clip". However if I go to file manager and select a mp4 it plays perfect. My n95 had the same problem. What goes on?
-
hi , I am new to sap XI. now i am planning to install sap XI in my desktop. one of my friend suggested me to install ECC 6.0 as it contains both R/3 and netweaver components like XI. i just want to know which version XI is available in Ecc 6.0 and fu