[FIXED 1.5.3] 1.5: Export Data to clipboard can't handle extended ASCII

Hi,
Exporting table data to the clipboard through the Export Data entry in the context menu, extended ASCII shows as "�". Especially killing for languages with accented vocals like Spanish.
FWIW, this was working in 1.2...
Regards,
K.

Hi Kristof,
I have exported the following table structure successfully to clipboard, pasting into Microsoft word on an English language XP SP2 Windows box:
CREATE TABLE "TOTIERNE"."SPANISHMÁSÓN"
(     "SPANISHMÁSÓN" VARCHAR2(100 BYTE),
     "ID" NUMBER(38,0)
) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING
STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT)
TABLESPACE "USERS" ;
Can you give a more precise testcase?
-Turloch

Similar Messages

  • V2.1 - Exporting DATE columns produces invalid SQL

    If you export a table with a column of type DATE, the SQL produced is invalid:
    Left-click on table in tree view -> Click on "Data" tab -> Right click on data -> Export Data -> Insert -> Clipboard:
    to_timestamp('01-SEP-09 17:49:21','DD-MON-RR HH.MI.SSXFF AM')
    I have seen this work correctly, but I can't find the right combination now... but in this case the date format should be:
    to_timestamp('01-SEP-09 17:49:21','DD-MON-RR HH24.MI.SSXFF')
    Edited by: gclough on Jan 4, 2010 3:54 PM

    Hi,
    I encountered this problem since the EA and Vasan (one of the SQL Dev's team) gave the solution for this.
    >
    You can add the following in the sqldeveloper.conf to ensure that the driver doesn't report the column type of a DATE column as TIMESTAMP.
    AddVMOption -Doracle.jdbc.mapDateToTimestamp=false
    >
    As you can see on
    Re: 2.1 EA1: Problems on Date type columns
    I wonder, why this isn't included on the default installation of the product, since it is absolutely wrong to use Timestamp for Date type column on the first place.
    Regards,
    Buntoro

  • Master data infoobject can't handle duplicate records after SP10

    Hi
    I am trying to load master data which happened to contain duplicate records from the source system.  In the DTP of the master data infoobject, I have ticked the 'Handle Duplicate Record Keys' checkbox.  After executing this DTP, the duplicate master data records were trapped in the Error Stack.  I am expecting overwriting of the duplicate master data to take place instead.  I understand that this error was fixed in Note 954661 - Updating master data texts when error in data package which is from SP9.  After applying Support Pack 10, the master data infoobject just can't handle records with duplicate keys.
    Please let me know if you manage to fix this problem.
    Many thanks,
    Anthony

    Found a fix for this problem.  Just applied this OSS note  Note 986196 - Error during duplicate record handling of master data texts.

  • How to export data in fixed lengh format

    I want to take data in 4 tables, and export the data to a fixed lenght flat file format. Then have that file sent to an email address. How can I do this?

    Suzie,
    I never used Toad ;-) But I always use the Oracle SQL Developer instead (which is for free by the way).
    SQL Developer gets installed with every 11g database installation but you could install it also to your 9.2 or 10.x environment - just grab it from otn.oracle:
    http://www.oracle.com/technology/software/products/sql/index.html
    (will take two minutes to install). Then you'd just define your connection to your target database, go into REPORTS (left side) and define a USER DEFINED REPORT. That's pretty easy because for instance you'd just say 'select * from scott.emp;' - once the report gets executed just move your mouse over one of the rows, hit the right mouse buttom and choose the last entry: EXPORT DATA - you'll have the options of TEXT, CSV, SQL LOADER FORMAT, INSERT, XML, HTML and XLS ;-)
    Another possibility within SQL*Plus would be using the column format options (the font of this webpage scrambles the output columns but in SQL*Plus you'll get everything formated in a fixed length format:
    column "info" format A11
    column "P1" format 99999
    column "P2" format 99999
    column "P3" format 99999
    SELECT Info " ", P1, P2, P3
    FROM tr_PortfolioSummary
    ORDER BY SortOrder
    P1 P2 P3
    Nb Bonds 10 40 40
    Nb Warrants 20 50 50
    Nb Shares 30 60 60
    Kind regards
    Mike
    Edited by: Mike Dietrich on Sep 12, 2008 9:41 AM

  • Report script taking too long to export data

    Hello guys,
    I have a report script to export data out of a BSO cube. The cube is 200GB in size. But the exported text file is only 10MB. It takes around 40 mins to export this file.
    I have exported data of this size in less than a minute from other DBs. But this one is taking way too long for me.
    I also have a calc script for the same export but that too is taking 20 mins which is not reasonable for a 10MB export.
    Any idea why a report script could take this long? Is it due to huge size of database? Or is there a way to optimize the report script?
    Any help would be appreciated.
    Thanks

    Thanks for the input guys.
    My DATAEXPORT is taking half the time than my report script export. So yeah it is much faster but still not reasonable(20 mins for one month data) compared to other DBs that are exported very quick.
    In my calc I am just FIXING on level 0 members for most of the dimensions against the specific period, year and scenario. I have checked the conditions for an optimal report script, I think mine is just fine.
    The outline has like 15 dimensions in it and only two of them are dense. Do you think the reason might be the huge size of DB along with too many sparse Dims?
    I appreciate your help on this.
    Thanks

  • Question regarding tranisent analysis and piecewise linear voltage export data

    howdy,
    My work with mulitism focuses on the analysis of equivalent circuits. I approach this through a transient analysis over a custom voltage input which I run through the PWL and export the data through excel for analysis, matching up the time of the transient analysis with the voltage through the circuit. Allowing me to observe the I-V performance.
    The problem is the the exported data is usually an array of hundred of entries. That may or may not align with the discrete values of my custom voltage input and allow for direct analysis between real world data.
    Is there a way to lock the time step (significantly reduce the number of interations) over the transient analysis to allow me to always export the same voltages. I.e only the voltages specified in my custom piecewise power source.
    The reason I am interested in doing this is to generate data which directly matches up with the specified custom voltage input and that matches up with my real world experimental data and would not require me to pick through hundreds of iterations at varied voltages and cherry pick the points that compare to my real life measurements for analysis.
    many thanks. 

    SPICE simulations doesn't have a fixed delta time, if the signal is changing rapidly the SPICE engine will slow down and take more samples and if the signal is constant it will automatically speed and take fewer samples.  You cannot control the SPICE simulation sampling.  One suggestion is to export an .lvm file instead of an Excel file because Multisim will convert the data to have a constant delta t which may be better for you.  You can open the .lvm file with Notepad to view the format.
    Tien P.
    National Instruments
    Attachments:
    LVM export.PNG ‏46 KB

  • How to export data from a spread sheet which has multiple work sheets?

    How to export data from a spread sheet which has multiple work sheets to a single text file with fixed length fields?

    Hello s1,
    saving them as CSV will not give a fixed legth output but, as the naming says, a Character Separated Value file.
    Regards
    Marcus

  • Error while exporting data from HCM to VDS

    Hi,
    When i am trying to export data from the SAP HCM system using the RPLDAP_EXTRACT, the export fails and the log shows that the LDAP_CREATE failed. In the VDS log,
    com.sap.idm.vds.oper.main_listener Message:
    Exception on add: class java.lang.Integer:null incompatible with class
    java.lang.String:null
    I verified the database connectivity from VDS to IDS when seems fine since the LDAP_SEARCH is working fine.
    Cheers !
    Zaheer

    It was a bug in SP2, it is fixed in SP2 Patch 4.
    Cheers !!
    Zaheer

  • Error while exporting data

    i get the following error while exporting data evn after running catexp.sql
    EXP-00056: ORACLE error 31600 encountered
    ORA-31600: invalid input value EMIT_SCHEMA for parameter NAME in function SET_TRANSFORM_PARAM
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 3926
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 4050
    ORA-06512: at "SYS.DBMS_METADATA", line 836
    ORA-06512: at line 1
    can anyone help me on this

    copy past from the notes referred above -
    Symptom(s)
    ~~~~~~~~~~
    You are performing a database export against your 9.2 database. The database
    export abnormally ends with the following errors:
    EXP-00056: ORACLE error 31600 encountered
    ORA-31600: invalid input value EMIT_SCHEMA for parameter NAME in function
    SET_TRANSFORM_PARAM
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 3926
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 4050
    ORA-06512: at "SYS.DBMS_METADATA", line 836
    ORA-06512: at line 1
    EXP-00000: Export terminated unsuccessfully
    Change(s)
    ~~~~~~~~~~
    One of the known causes for the error above is that a patchset was recently
    applied to the database but one or more of the patchset steps were not
    performed.
    Cause
    ~~~~~~~
    View the readme notes for the patchset recently installed. You will see a step
    that should have been performed that directs you to run the script named
    catpatch.sql. Very likely this step was never performed.
    Fix
    ~~~~
    As the sysdba user run the catpatch.sql script found in the directory named
    $ORACLE_HOME/rdbms/admin. Before you submit this script, view it's contents.
    You may see that this script places the database in restricted mode.
    rgds

  • Importing oim 9.1 exported data to oim 11g?

    Hello,
    I just read a few recent threads on upgrading from 9.1 to 11g... It sounds like there isn't a path to 11g at the moment.
    But has anyone tried exporting oim 9.1 data and using oim 11g Deployment Manager to import the 9.1 exported data in?
    We've invested about 4 man years in our current oim 9.1 solution. Just can't affort to throw too many things away!
    Appreciate any feedback.
    Thanks and cheers,
    Luke

    For the time being, stick with what works. Right nows the approval process form 10g does not exist in 11g so it cannot be migrated. There are a few other changes such as event handlers and entity adapters and pre-populate adapters and i'm sure many more. Wait for 11g to mature a bit more and provide the path you need to upgrade. Your implementation will not end with the current state, just don't be in a hurry to move forward to the 11g version. And remember, if it isn't broken, don't fix it. Unless you are in dire need of the functionality that 11g is going to give over 10g, then stay where you are at for the time being.
    -Kevin

  • How to create the Export Data and Import Data using flat file interface

    Hi,
    Request to let me know based on the requirement below on how to export and import data using flat file interface.....
    Please provide the steps involved for the same.......
    BW/BI - Recovery Process for SNP data. 
    For each SNP InfoProvider,
    create:
    1) Export Data:
    1.a)  Create an export data source, InfoPackage, comm structure, etc. necessary to create an ASCII fixed length flat file on the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider. 
    1.b)  All fields in each InfoProvider should be exported and included in the flat file. 
    1.c)  A process chain should be created for each InfoProvider with a start event. 
    1.d)  If the file exists on the target drive it should be overwritten. 
    1.e)  The exported data file name should include the InfoProvider technical name.
    1.f)  Include APO Planning Version, Date of Planning Run, APO Location, Calendar Year/Month, Material and BW Plant as selection criteria.
    2) Import Data:
    2.a) Create a flat file source system InfoPackage, comm structure, etc. necessary to import ASCII fixed length flat files from the XI
    ctnhsappdata\iface\SCPI063\Out folder for each SNP InfoProvider.
    2.b)  All fields for each InfoProvider should be mapped and imported from the flat file.
    2.c)  A process chain should be created for each InfoProvider with a start event. 
    2.d)  The file should be archived in the
    ctnhsappdata\iface\SCPI063\Archive directory.  Each file name should have the date appended in YYYYMMDD format.  Each file should be deleted from the \Out directory after it is archived. 
    Thanks in advance.
    Tyson

    Here's some info on working with plists:
    http://developer.apple.com/documentation/Cocoa/Conceptual/PropertyLists/Introduc tion/chapter1_section1.html
    They can be edited with any text editor. Xcode provides a graphical editor for them - make sure to use the .plist extension so Xcode will recognize it.

  • Warning:The EXPORT data cluster is too large for the application buffer.

    Hi Friends,
    I am getting following warning messages whenever I click on costing in "Accounting" tab.
    1. Costing data may not be up to date  Display Help
    2.  No costing variant exists in controlling scenario CPR0001
    3.  An error occurred in Accounting (system ID3DEV310)
    4. The EXPORT data cluster is too large for the application buffer.
      I can create project automatically from cprojects. PLan costs, budget and actuals maintain in WBS elements can be visible in cproject Object Links.
    Your reply is highly appreciated.
    Regards,
    Aryan

    Hi;
    Please, check the Note --> 1166365
    We are facing the same problem in R3, but apply this no fix it.
    Best regards.
    Mariano

  • Exporting data into text file

    Hi!
    I use this process in apex to export data from table into txt files with fixed length, so without any delimeter
    >
    declare
    v_file_name VARCHAR2 (2000) := 'test.txt';
    id varchar2(9);
    worker varchar2(30);
    address varchar2(26);
    begin
    OWA_UTIL.mime_header ('application/txt', FALSE);
    htp.p('Content-Disposition:attachment;filename="'|| v_file_name|| '"');
    OWA_UTIL.http_header_close;
    FOR x in (select id id from workers where col00 like '1000')
    LOOP
    select col01,col02,col03 into id, worker, addressfrom un_web_prenosi where id = x.id;
    htp.p(id||worker||address);
    END LOOP;
    apex_application.g_unrecoverable_error:=true;
    exception when others then
    null;
    end;
    and I have problem, because if I open file with notepad is everything in one line, but if I open file in notepad++ or I copy content of file in word then I see contents just like it should be (each record in one line). Do you know how can I solve this problem?

    I solved my problem. I forget to put chr(13) into htp.p(id||worker||address); so this line must be like htp.p(id||worker||address||chr(13));

  • Export data in a particular order

    Hi all,
    I want to export data in a particular order of dimension by the DATAEXPORT command.
    I use the below script :
    SET DATAEXPORTOPTIONS
         DataExportLevel "LEVEL0";
    DataExportDimHeader ON ;
    DataExportColFormat ON;
    DataExportOverwriteFile ON;
    FIX ("FY11","Submitted","Budget",@RELATIVE("Entity",0),@CHILDREN ("Period"))
    DATAEXPORT "File" "," "file.txt";
    ENDFIX;
    It always gives the file data in the below order
    "Entity","Allocations","Projects","CostCentres","FinancialYear","Version","Scenario","Account","Period"
    I want in the order of Scenario,Version,Year,Entity,CostCenter,Allocation,Project,Account,Period.
    Is it possible?
    Many thanks.

    The only way to do it would be to change the order of the sparse dimensions in your outline. The export exports in outline order

  • Export data to CSV in SQL Server

    Hi,
    I am trying to write a Java JDBC program exporting data
    from the database into CSV file using SQL Server 2005 DB.
    From that query:
    String="SELECT * into OUTFILE '"+filename+'" FIELDS TERMINATED BY ',' FROM table1";
    getting: "SQLException: Incorrect syntax near 'myfilename.csv'
    Yes, I put filename in single quotes. Same problem without them, but it strips out csv file extension.
    Using JTDS driver 1.2.4 on WinXP.
    Will later need to insert that into a batch process.
    What is a reason for that error? - I saw many many posts on the net, but no good explanations so far.
    So how do I fix that ?
    Is there a better way of doing it in SQL server ?
    I am a Java developer, not a DBA or Unix Sys Admin ;-)
    Please help !
    TIA,
    Oleg.

    Here is one way to do it in Java. You can customize as you like. Do a Google on how to get a DataSource (or DriverManager) connection for SqlServer. If you do not know how to get a PrintWriter, you do something like:
    PrintWriter writer = new PrintWriter(new FileOutputStream('results.csv");This will generate a generic CSV. It is probably not ideal, but should get you started.
        public final void serialize(final DataSource ds, final String sql, final PrintWriter writer)
            throws SQLException {
            Connection conn = null;
            PreparedStatement stmt = null;
            ResultSet results = null;
            try {
                conn = ds.getConnection();
                stmt = conn.prepareStatement(sql);
                results = stmt.executeQuery();
                ResultSetMetaData meta = results.getMetaData();
                int columnCount = meta.getColumnCount();
                while (results.next()) {
                    for (int current = 0; current < columnCount; current++) {
                        Object value = results.getObject(current + 1);
                        if (current > 0) {
                            writer.print(", ");
                        writer.print("\"" + (value == null ? "" : value.toString()) + "\"");
                    writer.println();
                writer.flush();
            finally {
                if (results != null) {
                    try {
                        results.close();
                    catch (SQLException e) {
                        e.printStackTrace();
                if (stmt != null) {
                    try {
                        stmt.close();
                    catch (SQLException e) {
                        e.printStackTrace();
                if (conn != null) {
                    try {
                        conn.close();
                    catch (SQLException e) {
                        e.printStackTrace();
        }My advice is to use BCP to start. The docs for the tool are here: [http://msdn.microsoft.com/en-us/library/aa174646(SQL.80).aspx]. You can manipulate the results in Access or Excel if you need to. Since it seems you are being thrown into an emergency and getting it fast is more important than you spending months learning additional Java API's.
    - Saish

Maybe you are looking for

  • AC 5.3 Risk Analysis to SRM 7.0

    Hi SAP experts. I would like to know if its posible to execute a Risk Analysis through a SRM 7.0 module, and instead of transacctions if it is posible to use the webdynpro links like actions or tcodes. My doubt is based on the fact that in SRM 7.0 th

  • Using the paint() function on a JFrame

    I'm currently enrolled in a graphics course and we've done several programs with line drawing algorithms, polygon filling algorithms, and the like. The only problem is, I have to work on our computers in the lab on campus because of this problem: On

  • Error when creating account aliases

    Hi All, Iam trying to create an new account aliases in INV but iam getting the foll error, This combination of segment values already exists. (ROUTINE=fdfvcc). Tried looking for any aliases been defined but nothing has been set.Can anyone let know wh

  • I need help with an SQL Error: ORA-00904: "TBL_0022"."EUID": invalid identi

    Hello All, I am receiving an error message when I run a relatively simple insert statement. I have verified that all the columns are defined in the tables. Error starting at line 1 in command: INSERT INTO TBL_0022(TBL_0022.OLD_OBSID,TBL_0022.EUID,TBL

  • Mac OS error 200332927

    I just installed Quicktime Pro 7.4.1 to record movies on my mac. I tried to start new recording and got the above message. What happened? How can I get QT pro to record movies from an Canopus ADVC-55? I am currently running OS 10.4.9 on a G5 with QT