Spool to csv file problem--comma included in the data

Friends,
i spooled the table's data to csv file.
when i open it in the excel sheet some of the records are in the different cells.
below is the record....
Colonial Street, Ruwanda
so when i view it in the excel sheet.
Colonial Street is coming in one cell
and
Ruwanda is coming in the next cell.
but "Colonial Street, Ruwanda" is the single entry in the column.
so how can i spool the record which is having comma (,) in between the words to get it in the single cell of excel?
thanks

Usually putting double quotes around your column values should take care of that:
select     '"'||column1
       ||'","'||column2
       ||'","'||column3||'"'
from   your_table

Similar Messages

  • Spool file with comma including quotes

    Hi All,
    I am getting the csv file however location commas are treated as another column in file. I am expecting based on the below SQL only two columns and all the data should be in two columns though we have comma value in location string.
    CREATE TABLE DEPT ( DEPTNO NUMBER,LOCATION VARCHAR2(255));
    INSERT INTO DEPT (DEPTNO,LOCATION) VALUES(100, 'CHICAGO,NEWYORK');
    INSERT INTO DEPT (DEPTNO,LOCATION) VALUES(200, 'ATLANTA,NEWYORK');
    INSERT INTO DEPT (DEPTNO,LOCATION) VALUES(300'OMHA,NEWYORK');
    INSERT INTO DEPT (DEPTNO,LOCATION) VALUES(100, 'NEWJERSEY,NEWYORK');
    COMMIT;
    sql file :
    set pagesize 10000
    set feedback off
    set heading off
    set echo off
    spool r:\dept.csv
    select to_char(deptno)||','||rtrim(location)
      from dept;
    spool off
    set echo on
    set heading on
    set feedback onEdited by: user1758353 on Dec 6, 2011 7:47 AM
    Edited by: user1758353 on Dec 6, 2011 7:47 AM
    Edited by: user1758353 on Dec 6, 2011 7:48 AM

    try to wrap your column location with a double quotes.
    e.g.
    select to_char(deptno)||','||'"'||rtrim(location)||'"'
      from dept;when you opened your csv file on excel it considers the strings enclosed on a double quotes as one column.

  • Spool with .csv file with appropiate line

    Hi All,
    I am spooling .csv file in sqlplus below are the spooling sqlplus command,
    set echo off
    set verify off
    set termout on
    set heading off
    set pages 50000
    set feedback off
    set newpage none
    set linesize 160
    SET TRIMSPOOL on
    spool d:\spool1.csv
    select 'MENU_USAGE_ID,BAC,SRV_CNSLTNT_ID,SRV_CNSLTNT_FNAME,SRV_CNSLTNT_LNAME,PRESENTED_MENUS_COUNT,PRESENTED_SERVICES_COUNT,ACCPT_SERVICES_COUNT,TOTAL_PRSNT_MENUS_VALUE,TOTAL_ACCPT_MENUS_VALUE,MENU_USAGE_DATE,CREATEDBY,UPDATEDBY,CREATED_DATETIME,UPDATED_DATETIME,CUSTOMER_MENU_ID,TOTAL_DECLINED_MENU_PRICE,CHANNEL,MODEL_YEAR,MODEL_DESC,VIN,CURRENT_MILEAGE,RO_NUMBER,GM_SERVICES_ACCEPTED,GM_SERVICES_DECLINED,DEALER_SERVICES_ACCEPTED,DEALER_SERVICES_DECLINED' from dual;
    select
    MENU_USAGE_ID ||','|| BAC||','|| SRV_CNSLTNT_ID||','|| SRV_CNSLTNT_FNAME||','|| SRV_CNSLTNT_LNAME||','|| PRESENTED_MENUS_COUNT||','|| PRESENTED_SERVICES_COUNT||','|| ACCPT_SERVICES_COUNT||','|| TOTAL_PRSNT_MENUS_VALUE||','|| TOTAL_ACCPT_MENUS_VALUE||','|| MENU_USAGE_DATE||','|| CREATEDBY||','|| UPDATEDBY||','|| CREATED_DATETIME||','|| UPDATED_DATETIME||','|| CUSTOMER_MENU_ID||','|| TOTAL_DECLINED_MENU_PRICE||','|| CHANNEL||','|| MODEL_YEAR||','|| MODEL_DESC||','|| VIN||','|| CURRENT_MILEAGE||','|| RO_NUMBER||','|| GM_SERVICES_ACCEPTED||','|| GM_SERVICES_DECLINED||','|| DEALER_SERVICES_ACCEPTED||','|| DEALER_SERVICES_DECLINED from menu_usage;
    spool off;
    But when I am opening .csv file I seeing like
    >
    SQL> select 'MENU_USAGE_ID     BAC     SRV_CNSLTNT_ID     SRV_CNSLTNT_FNAME     SRV_CNSLTNT_LNAME     PRESENTED_MENUS_COUNT     PRESENTED_SERVICES_COUNT     ACCPT_SERVICES_COUNT     TOTAL_PRSNT_MENUS_VALUE     TOTAL_ACCPT_MENUS_VALUE     MENU_USAGE_DATE     CREATEDBY     UPDATEDBY     CREATED_DATETIME     UPDATED_DATETIME     CUSTOMER_MENU_ID     TOTAL_DECLINED_MENU_PRICE     CHANNEL     MODEL_YEAR     MODEL_DESC     VIN     CURRENT_MILEAGE     RO_NUMBER     GM_SERVICES_ACCEPTED     GM_SERVICES_DECLINED     DEALER_SERVICES_ACCEPTED     DEALER_SERVICES_DECLINED' from dual;     
    I have attached the both .csv files .
    My requirement is
    It should come like this in .csv file.
    13706940     partId_001     Mark     James     1     2     2     146.26     146.26     9-Apr-12     tpsdir     tpsdir     9-Apr-12     9-Apr-12     9030845     0     UI     2006     CTS     1G6DP577X60140937     45     RO3     1     0     0     0
    13706949     partId_001     Mark     James     1     6     3     296.87     158.04     ########     tpsdir     tpsdir     ########     ########     9030859     138.83     UI     2010     Avalanche - 4WD     56000          1     3     2     0
    13706948     partId_001     Mark     James     1     3     0     145.84     0     ########     tpsdir     tpsdir     ########     ########     9030858     145.84     UI     2012     CTS     1GASDFJASLDFJLASD     50000          0     2     0     1
    1.     Not required >SQL prompt
    2.     Not required column header
    3. Not required SQL> spool off;
    Can anyone please help on this.
    Thanks,
    Krupa

    Other solution is to avoid using SQL*Plus spooling...
    (From my standard library of examples...)
    As sys user:
    CREATE OR REPLACE DIRECTORY TEST_DIR AS '\tmp\myfiles'
    GRANT READ, WRITE ON DIRECTORY TEST_DIR TO myuser
    /As myuser:
    CREATE OR REPLACE PROCEDURE run_query(p_sql IN VARCHAR2
                                         ,p_dir IN VARCHAR2
                                         ,p_header_file IN VARCHAR2
                                         ,p_data_file IN VARCHAR2 := NULL) IS
      v_finaltxt  VARCHAR2(4000);
      v_v_val     VARCHAR2(4000);
      v_n_val     NUMBER;
      v_d_val     DATE;
      v_ret       NUMBER;
      c           NUMBER;
      d           NUMBER;
      col_cnt     INTEGER;
      f           BOOLEAN;
      rec_tab     DBMS_SQL.DESC_TAB;
      col_num     NUMBER;
      v_fh        UTL_FILE.FILE_TYPE;
      v_samefile  BOOLEAN := (NVL(p_data_file,p_header_file) = p_header_file);
    BEGIN
      c := DBMS_SQL.OPEN_CURSOR;
      DBMS_SQL.PARSE(c, p_sql, DBMS_SQL.NATIVE);
      d := DBMS_SQL.EXECUTE(c);
      DBMS_SQL.DESCRIBE_COLUMNS(c, col_cnt, rec_tab);
      FOR j in 1..col_cnt
      LOOP
        CASE rec_tab(j).col_type
          WHEN 1 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
          WHEN 2 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_n_val);
          WHEN 12 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_d_val);
        ELSE
          DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
        END CASE;
      END LOOP;
      -- This part outputs the HEADER
      v_fh := UTL_FILE.FOPEN(upper(p_dir),p_header_file,'w',32767);
      FOR j in 1..col_cnt
      LOOP
        v_finaltxt := ltrim(v_finaltxt||','||lower(rec_tab(j).col_name),',');
      END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
      UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      IF NOT v_samefile THEN
        UTL_FILE.FCLOSE(v_fh);
      END IF;
      -- This part outputs the DATA
      IF NOT v_samefile THEN
        v_fh := UTL_FILE.FOPEN(upper(p_dir),p_data_file,'w',32767);
      END IF;
      LOOP
        v_ret := DBMS_SQL.FETCH_ROWS(c);
        EXIT WHEN v_ret = 0;
        v_finaltxt := NULL;
        FOR j in 1..col_cnt
        LOOP
          CASE rec_tab(j).col_type
            WHEN 1 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_v_val);
                        v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
            WHEN 2 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_n_val);
                        v_finaltxt := ltrim(v_finaltxt||','||v_n_val,',');
            WHEN 12 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_d_val);
                        v_finaltxt := ltrim(v_finaltxt||','||to_char(v_d_val,'DD/MM/YYYY HH24:MI:SS'),',');
          ELSE
            v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
          END CASE;
        END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
        UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      END LOOP;
      UTL_FILE.FCLOSE(v_fh);
      DBMS_SQL.CLOSE_CURSOR(c);
    END;This allows for the header row and the data to be written to seperate files if required.
    e.g.
    SQL> exec run_query('select * from emp','TEST_DIR','output.txt');
    PL/SQL procedure successfully completed.Output.txt file contains:
    empno,ename,job,mgr,hiredate,sal,comm,deptno
    7369,"SMITH","CLERK",7902,17/12/1980 00:00:00,800,,20
    7499,"ALLEN","SALESMAN",7698,20/02/1981 00:00:00,1600,300,30
    7521,"WARD","SALESMAN",7698,22/02/1981 00:00:00,1250,500,30
    7566,"JONES","MANAGER",7839,02/04/1981 00:00:00,2975,,20
    7654,"MARTIN","SALESMAN",7698,28/09/1981 00:00:00,1250,1400,30
    7698,"BLAKE","MANAGER",7839,01/05/1981 00:00:00,2850,,30
    7782,"CLARK","MANAGER",7839,09/06/1981 00:00:00,2450,,10
    7788,"SCOTT","ANALYST",7566,19/04/1987 00:00:00,3000,,20
    7839,"KING","PRESIDENT",,17/11/1981 00:00:00,5000,,10
    7844,"TURNER","SALESMAN",7698,08/09/1981 00:00:00,1500,0,30
    7876,"ADAMS","CLERK",7788,23/05/1987 00:00:00,1100,,20
    7900,"JAMES","CLERK",7698,03/12/1981 00:00:00,950,,30
    7902,"FORD","ANALYST",7566,03/12/1981 00:00:00,3000,,20
    7934,"MILLER","CLERK",7782,23/01/1982 00:00:00,1300,,10The procedure allows for the header and data to go to seperate files if required. Just specifying the "header" filename will put the header and data in the one file.
    Adapt to output different datatypes and styles are required.

  • Can you open CSV-files with commas?

    It seems to me that we have something of a regression in Numbers 2.0 (part of iWork '09) - it (like Excel before it) no longer opens regular CSV-files.
    When I use the new Numbers to try to make a CSV-file, it'll actually use semicolons instead of commas, and any of my old CSV-files (like this one: http://d.ooh.dk/misc/postnumre.csv ) (with commas) will load all the values (and the commas) in a single column.
    I wonder if it's just me, or perhaps only a problem when using European notation (with comma as the decimal separator)?

    At last, Apple adopted the same behavior than Bento 1.
    When the decimal separator is the period, it works with standard CSV files.
    When the decimal separator is comma, it works with CSV using the semi-colon as item delimiters.
    To open your old CSV, set temporarily your system to a region whose decimal separator is period. Given that, Numbers will open them flawlessly.
    The ability to chose the separator (as it is now in Bento 2) would have been fine .
    Yvan KOENIG (from FRANCE dimanche 11 janvier 2009 20:30:03)

  • Unable to Load CSV file with comma inside the column(Sql Server 2008)

    Hi,
    I am not able load an CSV file with Comma inside a column. 
    Here is sample File:(First row contain the column names)
    _id,qp,c
    "1","[ ""0"", ""0"", ""0"" ]","helloworld"
    "1","[ ""0"", ""0"", ""0"" ]","helloworld"
    When i specify the Text Qualifier as "(Double quotes) it work in SQL Server 2012, where as fail in the SQL Server 2008, complaining with error:
    TITLE: Microsoft Visual Studio
    The preview sample contains embedded text qualifiers ("). The flat file parser does not support embedding text qualifiers in data. Parsing columns that contain data with text qualifiers will fail at run time.
    BUTTONS:
    OK
    I need to do this in sql server 2008 R2 with Service pack 2, my build version is 10.50.1600.1.
    Can someone let me know it is possible in do the same way it is handle in SQL Server 2012?
    Or
    It got resolved in any successive Cumulative update after 10.50.1600.1?
    Regards Harsh

    Hello,
    If you have CSV with double quotes inside double quotes and with SSIS 2008, I suggest:
    in your data flow you use a script transformation component as Source, then define the ouput columns id signed int, Gp unicode string and C unicode string. e.g. Ouput 0 output colmuns
    Id - four-byte signed
    gp - unicode string
    cc - unicode string
    Do not use a flat file connection, but use a user variable in which you store the name of the flat file (this could be inside a for each file loop).
    The user variable is supplied as a a readonly variable argument to the script component in your dataflow.
    In the script component inMain.CreateNewOutputRows use a System.IO.Streamreader with the user variable name to read the file and use the outputbuffer addrow method to add each line to the output of the script component.
    Between the ReadLine instraction and the addrow instruction you have to add your code to extract the 3 column values.
    public override void CreateNewOutputRows()
    Add rows by calling the AddRow method on the member variable named "<Output Name>Buffer".
    For example, call MyOutputBuffer.AddRow() if your output was named "MyOutput".
    string line;
    System.IO.StreamReader file = new System.IO.StreamReader( Variables.CsvFilename);
    while ((line = file.ReadLine()) != null)
    System.Windows.Forms.MessageBox.Show(line);
    if (line.StartsWith("_id,qp,c") != true) //skip header line
    Output0Buffer.AddRow();
    string[] mydata = Yourlineconversionher(line);
    Output0Buffer.Id = Convert.ToInt32(mydata[0]);
    Output0Buffer.gp = mydata[1];
    Output0Buffer.cc = mydata[2];
    file.Close();
    Jan D'Hondt - SQL server BI development

  • How can I EXCLUDE particular folders/files from being included in the Lightroom 5 catalog?

    How can I EXCLUDE particular folders/files from being included in the Lightroom 5 catalog?
    I want to be able to specify paths and/or filename patterns that should NEVER be indexed in the catalog: not that they don't show up under certain circumstances, just that they are NOT THERE (in the catalog) to begin with!!
    I also do NOT want to have to respecify these criteria when "synching" the catalog; criteria should be PERMANENT (or at least 'til they're changed in LR preferences).
    I cannot be the only person in the world with this problem--but it sure seems like it, because I've spent an inordinate amount of time doing online searches & I can't find anything that addresses this issue.
    Can anyone PLEASE HELP me?
    BTW, Adobe employees (if there are any of you out there), I'm simply trying out LR now.  If I can't get this (seemingly fundamental) thing worked out (with relatively little effort), Adobe is going to LOSE A SALE of LR, because it's useless to me without this feature.
    Thanks,
    -t

    Thanks for the info... but how to you TELL LR to EXCLUDE images in a particular directory/directory structure?  The intuitive place to indicate this would be a right-click on the folder itself (or particular images), but no such option exists..  "???"  Thanks again, -pt
    The same way that you avoid buying things in your local convenience store. You don't stick a lot of labels on certain shelves, slap a credit card down on the counter each week, and announce you are buying everything that doesn't have a label on its shelf. You just choose the things you want to buy.
    Your mention of "synching the catalog" suggests that you are expecting the LR library, to mirror the total images in the computer by default. This is not IMO a useful way to think about it. LR contains no images at all, by default: the Catalog starts out empty.
    Then you Add just those selected images that you want to import, among those already in the computer. That's a one-off process which is IMO best achieved controllably in batches - which you can keyword and otherwise organise in LR terms as you go. For new incoming images, ongoing, it is a question of Copying, Moving or Adding these in defined batches as they occur - from the relevant location only - there is no need to re-consider the import of a wider body of images. And one would not want to repeat that whole task anyway: one's prior investment in achieving the right selectivity, has a positive value.
    RP

  • How can we a validate a CSV file in SSIS and log the error details in another CSV File by using Sript Task.

    How can we a  validate a CSV file in SSIS and log the error details in another CSV File by using Sript Task.

    Please see:
    http://www.bidn.com/blogs/DevinKnight/ssis/76/does-file-exist-check-in-ssis
    http://social.msdn.microsoft.com/Forums/en-US/01ce7e4b-5a33-454b-8056-d48341da5eb2/vb-script-to-write-variables-to-text-file

  • .ear file is not including all the required files for deployment

    While deploying from JDeveloper, some of the files (eg. <Project>.cpx file, .class file) are not included in the .ear file.
    Pls help.

    Is the .ear file deployed with the Deploy to EAR file selection?

  • Upload a csv file in application server with the specified codepage(Poland

    Hi,
    i want to upload a csv file in application server with the specified codepage ( for poland ) , please let me know how to use open dataset for that.?
    Anurag

    I would upload the file with transaction CG3Z in binary mode and do the code page translation with a report: read from server an save with different name.

  • Problem in Compresses file.Not able to see the data in file

    Hi All,
    I am downloading file into application server.
    Using:
    OPEN DATASET file FOR OUTPUT FILTER 'compress'.
    OR
    DATA: l_command(100) type c.
       CONCATENATE 'compress ' p_file INTO lw_command SEPARATED BY space.
        ENDIF.
        CALL 'SYSTEM' ID 'COMMAND' FIELD lw_command.
    When I see the file content it contains junk characters.
    If I want to see the data Do I need to uncompress the file.
    con't we see the data without uncompressing the file?.
    Regards,
    vinod

    Hi,
    Uncompress is required to read the content of an Compressed file.
    OPEN DATASET DSN FOR OUTPUT FILTER 'compress'.
    OPEN DATASET DSN FOR INPUT FILTER 'uncompress'.
    The first OPEN statement opens the file so that the displayed data can be read into it in compressed form.
    The second OPEN statement opens the file so that the data is decompressed when it is read from the file.
    http://www.sapnet.ru/abap_docu/ABAPOPEN_DATASET.htm#&ABAP_ADDITION_8@14@

  • Have made a rebbot from Time Machine, and now I lack a lot of pictures. Can see icons, but not enlarge them. Especially pictures where 2009 is included in the date is missing? When I set up my desktop, I can find some of the pictures. Where do They hide,

    Missing Pictures.
    Have made a rebbot from Time Machine, and now I lack a lot of pictures. Can see icons, but not enlarge them. Especially pictures where 2009 is included in the date is missing?
    When I set up my desktop, I can find some of the pictures. Where do They hide, and what shall I do.
    Have made back up, but unfortunately a little too late. What to do?
    MacBook Pro 15 "late 2008, Processor 2.53 Ghz Intel Core 2 Duo, Graphics NVIDIA GeForce 9400M 256 MB, Software Mac OS X 10.7.4 Lion (11E53).

    1. You did not get an error message telling you that your iPhoto library was getting full. You got a message telling you that your HD was getting full, right?
    OS X needs about 10 gigs of hard drive space for normal OS operations - things like virtual memory, temporary files and so on.
    Without this space your Mac will slow down as the OS hunts for space on the disk, files will be fragmented, also slowing things down, apps will crash and the risk of data corruption - that is damage to your files, photos, music - increases exponentially.
    Your first priority is to make more space on that HD. Nothing else can be done until you do.
    Purchase an external HD and move your Photos and Music to it. Both iPhoto and iTunes can run perfectly well with the Library on an external disk.
    Your Library has been damaged from being run on an overfull disk.
    How much free space on it now?

  • Does the data transfer used in the back up in time capsule included in the data allowance from my service provider?

    Is the data transfer used in the backup with airport time capsule included in the data allowance with my service provider?

    gheefromsa wrote:
    I am using Ethernet cable to connect TC to the modem. I am only backing up my mac book air within range of the TC. So I guess it will be OK from your reply. As a trial I did interrupt the back up by turning off the TC, disconnecting the Ethernet cable, turning TC back on. The mac would not connect wirelessly with the TC without the Ethernet connection.
    It is not using ISP .. absolutely for sure. Well 99.999%.. absolute tends to be only in maths.
    But you did the test wrongly.. I said to disconnect the modem.. during the backup.. not turn off the TC.. obviously once interrupted the TC will not work again until everything is plugged in.. but it will work even bridged for a short time.. just unplug the WAN port from the modem.
    The test is moot now.. as I can give you a 99.999% promise it is not using ISP.

  • Ssrs 2012 export to comma delimited (csv) file problem

    In an ssrs 2012 report, I want to be able to export all the data to a csv (comma delimited) file that only contains the detailed row information. I do not want to export any rows that contain header data information.
    Right now the export contains header row information and detail row information. The header row information exists for the PDF exports.
    To have the header row not exist on CSV exports, I have tried the following options on the tablix with the header row information with no success:
    1. = iif(Mid(Globals!RenderFormat.Name,1,5)<>"EXCEL",false,true),
    2. = iif(Mid(Globals!RenderFormat.Name,1,3)="PDF",false,true),
    3. The textboxes that contain the column headers, the dataelelementoutput=auto and
       the text boxes that contain the actual data, I set the dataelelementoutput=output.
    Basically I need the tablix that contains the header information not to appear when the export is for a csv file.
    However when the export is for the detail lines for the csv export, I need that tablix to show up.
    Thus can you tell me and/or show me code on how to solve this problem?

    Hi wendy,
    Based on my research, the expression posted by Ione used to hide items only work well when render format is RPL or MHTML. Because only the two render format’s RenderFormat.IsInteractive is true.
    In your scenario, you want to hide tablix header only when export the report to csv format. While CSV file uses the textbox name in detail section as the column name, not the table header when you view the table. So if we want to hide the header row of the
    tablix, please refer to the steps below to enable the “NoHeader” setting in the RSReportserver.config file:
    Please navigate to RSReportserver.config file: <drive:>\Program Files\Microsoft SQL Server\MSRS1111.MSSQLSERVER\Reporting Services\ReportServer \RSReportserver.config.
    Backup the RSReportserver.config file before we modify it, open the RSReportserver.config file with Notepad format.
    In the <Render> section, add the new code for the CSV extension like this:
        < Extension Name="CSV"   Type="Microsoft.ReportingServices.Rendering.DataRenderer.CsvReport,Microsoft.ReportingServices.DataRendering">
            <Configuration>
                <DeviceInfo>
                   <NoHeader>True</NoHeader>
                </DeviceInfo>
            </Configuration>
        < /Extension>
    Save the RSReportserver.config file.
    For more information about CSV Device Information Settings, please see:
    http://msdn.microsoft.com/en-us/library/ms155365(v=sql.110)
    If you have any other questions, please feel free to ask.
    Thanks,
    Katherine xiong
    If you have any feedback on our support, please click
    here.
    Katherine Xiong
    TechNet Community Support

  • How to include seconds and millisecon​d in the CSV file time stamp generated by the SpreadShee​t Object?

    I'm Using Lookout 6.02, in a XP Pro Windows Machine, Service Pack 3.
    I have an application that collects data at the rate of 20 samples per second. It works just fine, the CSV File contains all the samples without missing a single one.
    The problem is that the time stamp in the data file only shows hours and minutes. With 20 sample per second, there are 1200 samples per minute. It is hard to figure out in which second or sample number, occurred some events collected. These files are open with Excel and analyzed with Excel. There is no easy way to implement a search or a guidance for knowing the sample number or the exact time in Excel.  My problem is to add the seconds and if possible the milliseconds to the time stamp done by the spreadsheet object.
    Any suggestion?
    Other ways to do the same that includes the time stamp all the way down to milliseconds?

    Create another column with the Now() function, then set the calumn format to hh:mm:ss.s
    expression would be now(trigger) for the date and time or now(trigger)-today(trigger) for just time
    The trigger will cause the expression to stay updated, perhaps using the same trigger for logging the spreadsheet.
    Good luck
    Mike
    Message Edited by Mike@DTSI on 01-16-2009 04:12 PM
    Mike Crabtree - Lead Developer
    Destek of Nevada, Inc. / Digital Telemetry Systems, Inc.
    (866) 964-6948 / (760) 247-9512

  • Spool file problem,Can't see the query in output file.

    Hello ,
    I am facing a very old school kind of problem .....about spool file ....
    The scenario -
    I have made a script by name DB_Status_checkup.sql which i want to fire on the database to check the database status. In this script their are many queries regarding the data dictionary views to know about the database. It consist of nearly 25-30 different select queries..
    The problem is i want to make a spool file of the output of that query....i want to see the SQL query & the output below in the spool file, so it will be easy for me to judge the result. But i can't see the SQL query , i can only see the output , & in so many queries it all gets jumbled up....even i can't understand where the next output starts ...
    Sample of my SQL Script ....
    clear buffer
    spool D:\DB_status.txt
    /*To check the database startup time*/
    Select to_char(startup_time, 'HH24:MI DD-MON-YY') "Startup time"
    from v$instance
    .........next query n so on....
    spool off;
    In the output pf the spool file at D:\db_status.txt..
    Startup time
    08:25 16-JUL-10It shows only the output of the query in the spool file not the query,
    What should i do to get the SQL query as well as the output of that query just below it in the spool file ???
    Please suggest if you have anymore ideas , regarding this .....
    ORACLE 10g R2
    Windows Server 2008
    Thanks in advance ...

    Why don't you just use database control, instead of re-inventing the wheel?
    Apart from that, SQL*Plus has it's own reference manual, which you apparently refuse to read.
    The answer to your quiz/doc question is
    set echo on
    Sybrand Bakker
    Senior Oracle DBA

Maybe you are looking for

  • Payment document vs clearing document

    Dear All, What is the difference between payment document and clearing document,in which business scenario's we will use them. Thank you. Regards, D vasanth

  • Word to Pages "Import Error"

    I have a load of files created by drag and drop editing of pdf files into Word tables. I want to use pages to edit these ... it should be easier with services. Most of the files open fine, but a few give me an Import Error. Any ideas? There are no ob

  • Configuring Apache httpd.conf for Multiple Web Servers

    After successfully serving a single web site from my MacMini, I'm having trouble configuring httpd.conf to add an additional site. My httpd.conf file looks something like this: NameVirtualHost *:80 <VirtualHost *:80> ServerName www.domain1.net Docume

  • 10.2.1 finance application problems

    BlackBerry OS 10 with the latest update released for Turkish language support for devices, some apps do not work with them more in the Turkish banking practices and the entire financial sector. on twitter via @ blackberryhelp I do not get the results

  • One SAP for several company

    We are using SAP as our ERP system. And now, we would like to add one more company into the system, Actually, the business of new comapany is totally differ to our existing business, and they are two entity. I would like to know, we use the existing