Problem in csv file generation

Hi ,
I am trying to spool a query in csv format using "col sep ,"
but it is giving problem for the text values like example i have a sql_text column which gives the sql text. in the csv file which was generated the select statement columns are going to next column where the comma is invoked.
is there any way to format while generating csv file that entire sql_text should come in one column
Thanks
Rakesh

set echo OFF pages 50000 lin 32767 feed off heading ON verify off newpage none trimspool on
define datef=&1
define datet=&2
set colsep ','
spool querries.csv
SELECT s.parsing_schema_name,
p.instance_number instance_number,
s.sql_id sql_id,
x.sql_text sql_text,
p.snap_id snap_id,
TO_CHAR (p.begin_interval_time,'mm/dd/yyyy hh24:mi') begin_interval_time,
TO_CHAR (p.end_interval_time,'mm/dd/yyyy hh24:mi') end_interval_time,
s.elapsed_time_delta / DECODE (s.executions_delta, 0, 1, s.executions_delta) / 1000000 elapsed_time_per_exec,
s.elapsed_time_delta / 1000000 elapsed_time,
s.executions_delta executions, s.buffer_gets_delta buffer_gets,
s.buffer_gets_delta / DECODE (s.executions_delta, 0, 1, s.executions_delta) buffer_gets_per_exec,
module module
FROM dba_hist_sqlstat s, dba_hist_snapshot p, dba_hist_sqltext x
WHERE p.snap_id = s.snap_id
AND p.dbid = s.dbid
AND p.instance_number = s.instance_number
AND p.begin_interval_time >
TO_TIMESTAMP ('&datef','yyyymmddhh24miss')
AND p.begin_interval_time <
TO_TIMESTAMP ('&datet','yyyymmddhh24miss')
AND s.dbid = x.dbid
AND s.sql_id = x.sql_id
ORDER BY instance_number, elapsed_time_per_exec DESC ;
SPOOL OFF;
exit;

Similar Messages

  • CSV file generation issue

    Hello All,
    We are facing below issue during CSV file generation :
    Generated file shows field value as 8.73E+11 in output and when we are clicking inside this column result shown is approximate of the correct value like 873684000000. We wish to view the correct value 872684000013.
    Values passes from report program during this file generation are correct.
    Please advice to resolve this issue.
    Thanks in Advance.

    There is nothing wrong in your program, it is the property of excel that if the value in the cell is larger than
    default size it shows in output of that format.

  • Problem reading csv file with special character

    Hai all,
    i have the following problem reading a csv file.
    442050-Operations Tilburg algemeen     Huis in  t Veld, EAM (Lisette)     Gebruikersaccount     461041     Peildatum: 4-5-2010     AA461041                    1     85,92
    when reading this line with FM GUI_UPLOAD this line is split up in two lines in data_tab of the FM,
    it is split up at this character 
    Line 1
    442050-Operations Tilburg algemeen     Huis in
    Line 2
    t Veld, EAM (Lisette)     Gebruikersaccount     461041     Peildatum: 4-5-2010     AA461041                    1     85,92
    Anyone have a idea how to get this in one line in my interbal table??
    Greetz Richard

    Hi Greetz Richard
      Problably character  contains same binary code as line feed + carriage return.
      You can use statement below as workaround.
    OPEN DATASET file FOR INPUT IN TEXT MODE ENCODING UNICODE
    In this case your system must support Unicode encoding
    Kind regards

  • Problem import csv file with SQL*loader and control file

    I have a *csv file looking like this:
    E0100070;EKKJ 1X10/10 1 KV;1;2003-06-16;01C;75
    E0100075;EKKJ 1X10/10 1 KV;500;2003-06-16;01C;67
    E0100440;EKKJ 2X2,5/2,5 1 KV;1;2003-06-16;01C;37,2
    E0100445;EKKJ 2X2,5/2,5 1 KV;500;2003-06-16;01C;33,2
    E0100450;EKKJ 2X4/4 1 KV;1;2003-06-16;01C;53
    E0100455;EKKJ 2X4/4 1 KV;500;2003-06-16;01C;47,1
    I want to import this csv file to this table:
    create table artikel (artnr varchar2(10), namn varchar2(25), fp_storlek number, datum date, mtrlid varchar2(5), pris number);
    My controlfile looks like this:
    LOAD DATA
    INFILE 'e:\test.csv'
    INSERT
    INTO TABLE ARTIKEL
    FIELDS TERMINATED BY ';'
    TRAILING NULLCOLS
    (ARTNR, NAMN, FP_STORLEK char "to_number(:fp_storlek,'99999')", DATUM date 'yyyy-mm-dd', MTRLID, pris char "to_number(:pris,'999999D99')")
    I cant get sql*loader to import the last column(pris) as I want. It ignore my decimal point which in this case is "," and not "." maybe this is the problem. If the decimal point is the problem how can I get oracle to recognize "," as a decimal point??
    the result from the import now, is that a decimal number (37,2) becomes 372 in the table

    Set NLS_NUMERIC_CHARACTERS environment variable at OS level, before running SqlLoader :
    $ cat test.csv
    E0100070;EKKJ 1X10/10 1 KV;1;2003-06-16;01C;75
    E0100075;EKKJ 1X10/10 1 KV;500;2003-06-16;01C;67
    E0100440;EKKJ 2X2,5/2,5 1 KV;1;2003-06-16;01C;37,2
    E0100445;EKKJ 2X2,5/2,5 1 KV;500;2003-06-16;01C;33,2
    E0100450;EKKJ 2X4/4 1 KV;1;2003-06-16;01C;53
    E0100455;EKKJ 2X4/4 1 KV;500;2003-06-16;01C;47,1
    $ cat artikel.ctl
    LOAD DATA
    INFILE 'test.csv'
    replace
    INTO TABLE ARTIKEL
    FIELDS TERMINATED BY ';'
    TRAILING NULLCOLS
    (ARTNR, NAMN, FP_STORLEK char "to_number(:fp_storlek,'99999')", DATUM date 'yyyy-mm-dd', MTRLID, pris char "to_number(:pris,'999999D99')")
    $ sqlldr scott/tiger control=artikel
    SQL*Loader: Release 10.1.0.3.0 - Production on Sat Nov 12 15:10:01 2005
    Copyright (c) 1982, 2004, Oracle.  All rights reserved.
    Commit point reached - logical record count 6
    $ sqlplus scott/tiger
    SQL*Plus: Release 10.1.0.3.0 - Production on Sat Nov 12 15:10:11 2005
    Copyright (c) 1982, 2004, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.1.0.3.0 - Production
    With the Partitioning, OLAP and Data Mining options
    SQL> select * from artikel;
    ARTNR      NAMN                      FP_STORLEK DATUM      MTRLI       PRIS
    E0100070   EKKJ 1X10/10 1 KV                  1 16/06/2003 01C           75
    E0100075   EKKJ 1X10/10 1 KV                500 16/06/2003 01C           67
    E0100440   EKKJ 2X2,5/2,5 1 KV                1 16/06/2003 01C          372
    E0100445   EKKJ 2X2,5/2,5 1 KV              500 16/06/2003 01C          332
    E0100450   EKKJ 2X4/4 1 KV                    1 16/06/2003 01C           53
    E0100455   EKKJ 2X4/4 1 KV                  500 16/06/2003 01C          471
    6 rows selected.
    SQL> exit
    Disconnected from Oracle Database 10g Enterprise Edition Release 10.1.0.3.0 - Production
    With the Partitioning, OLAP and Data Mining options
    $ export NLS_NUMERIC_CHARACTERS=',.'
    $ sqlldr scott/tiger control=artikel
    SQL*Loader: Release 10.1.0.3.0 - Production on Sat Nov 12 15:10:41 2005
    Copyright (c) 1982, 2004, Oracle.  All rights reserved.
    Commit point reached - logical record count 6
    $ sqlplus scott/tiger
    SQL*Plus: Release 10.1.0.3.0 - Production on Sat Nov 12 15:10:45 2005
    Copyright (c) 1982, 2004, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.1.0.3.0 - Production
    With the Partitioning, OLAP and Data Mining options
    SQL> select * from artikel;
    ARTNR      NAMN                      FP_STORLEK DATUM      MTRLI       PRIS
    E0100070   EKKJ 1X10/10 1 KV                  1 16/06/2003 01C           75
    E0100075   EKKJ 1X10/10 1 KV                500 16/06/2003 01C           67
    E0100440   EKKJ 2X2,5/2,5 1 KV                1 16/06/2003 01C         37,2
    E0100445   EKKJ 2X2,5/2,5 1 KV              500 16/06/2003 01C         33,2
    E0100450   EKKJ 2X4/4 1 KV                    1 16/06/2003 01C           53
    E0100455   EKKJ 2X4/4 1 KV                  500 16/06/2003 01C         47,1
    6 rows selected.
    SQL>                                                                            Control file is exactly as yours, I just put replace instead of insert.

  • Comma and quote problem in csv file

    Hi
    My requirement is to append data in an csv file. This is Proxy to File FCC Scenario. for some of the fields from proxy which contains comma(,) and also double-quote("). for these fileds the in the csv file it is spiting in to two columns and appending in to the next column. and the double-quote symbol is not inserting in the csv file.
    1. why the double-quote(") is not inserting in to the csv file columns?
    2. how to over come the comma problem? I want that particular file need to append in one column only.
    Thanks
    Vankadoath

    hi vankadoath,
    Were you able to solve comma issue in CSV file.
    Even i am facing similar issue.....
    wheneever there is a comma in field, it is updating data after comma into next field.
    If anybody has solution for the same.......
    pls suggest the same............
    santosh.
    Edited by: santosh koraddi on Jan 20, 2011 9:44 PM

  • I have received a .xlsx (excel) file.  It is also ziped and I can not open it.  I also have the same problem with .csv files.  I can not open them either.  How do I get them open.  Thanks.

    I have recived via an email attachment a .xlsx (Exel) file which is also zipped.  I can not open it.  I also can never get .csv files open.  How do I overcome these two problems?
    Thanks,
    sorsini

    drag it to the Desktop.  double-click it to unzip it.

  • Problems with CSV File

    I've got a form that collects addresses, and then puts them
    in a .csv file.
    One of the problems I'm having is that some of the addresses
    are longer and pieces of them end up going in different columns.
    Is there a way around this ? Making max lengths, or putting
    them in a DB instead of a csv file?
    Your help is appreciated
    -MK

    Have a read of this:
    http://en.wikipedia.org/wiki/Comma-separated_values
    And this:
    http://www.creativyst.com/Doc/Articles/CSV/CSV01.htm
    They explain how to compose a CSV properly.
    Adam

  • Separator problem in CSV file download

    Hi All,
    I am downloading the internal table data in CSV file format using the function modules SAP_CONVERT_TO_CSV_FORMAT  and GUI_DOWNLOAD.
    In the FM SAP_CONVERT_TO_CSV_FORMAT the default value for field separator is ;. But I want comma.
    I am passing comma as a field separator but I am getting ; in my CSV file.
    Please help.
    Thanks in Advance.
    - Neha.

    Hi
    It seems fm SAP_CONVERT_TO_CSV_FORMAT doesn't use the parameter I_FIELD_SEPERATOR, but a the constant C_FIELD_SEPARATOR with value ;
    So u should replace ; with comma after calling that fm or check a SAP note
    Max

  • CSV file generation with timestamp time details

    CREATE OR REPLACE PROCEDURE extract_proc AS
    l_rows number;
    begin
    l_rows := extract_function('select *
    from table1 where date1 between trunc(sysdate) and sysdate
    'MYDIR1',
    'test.csv');
    end;
    I have a procedure extract_proc shown above which create a CSV in the directory MYDIR1 in the unix.
    I would like to have the timestamp ie ddmmyyyyHH24:MI details in the CSV file name created. ie. to have a file name of test22072011.csv whenever it generates with respective timestamp to be prefixed or suffixed to the file number. Is ther any way to achieve this
    Pls suggest
    Many thanks

    Why don't you set your file name dynamically? Something like:
    AS
         file_name     VARCHAR2(50);
    BEGIN
         file_name := 'test' || TO_CHAR(SYSDATE,'YYYYMMDDHH24MISS') || '.csv';
         extract_function(...,file_name);
    END;
    /

  • Csv file generation

    Hi All,
    Can we have a comma included in the csv generated file?
    Suppose query contains comma
    fieid1 field2
    Pickle This pickle, must be sour.
    sweet More sweets, leads to diabetes.
    My csv should be of two fields not three. But i am getting the csv file in three field because of comma in field2. How Can we resolve it.
    field1 field2 field3
    Pickle This pickle must be sour.
    sweet More sweets leads to diabetes.
    Hi anybody can help on these??
    Thanks

    The same as you put any other character in the file...
    As sys user:
    CREATE OR REPLACE DIRECTORY TEST_DIR AS '\tmp\myfiles'
    GRANT READ, WRITE ON DIRECTORY TEST_DIR TO myuser
    /As myuser:
    CREATE OR REPLACE PROCEDURE run_query(p_sql IN VARCHAR2
                                         ,p_dir IN VARCHAR2
                                         ,p_header_file IN VARCHAR2
                                         ,p_data_file IN VARCHAR2 := NULL) IS
      v_finaltxt  VARCHAR2(4000);
      v_v_val     VARCHAR2(4000);
      v_n_val     NUMBER;
      v_d_val     DATE;
      v_ret       NUMBER;
      c           NUMBER;
      d           NUMBER;
      col_cnt     INTEGER;
      f           BOOLEAN;
      rec_tab     DBMS_SQL.DESC_TAB;
      col_num     NUMBER;
      v_fh        UTL_FILE.FILE_TYPE;
      v_samefile  BOOLEAN := (NVL(p_data_file,p_header_file) = p_header_file);
    BEGIN
      c := DBMS_SQL.OPEN_CURSOR;
      DBMS_SQL.PARSE(c, p_sql, DBMS_SQL.NATIVE);
      d := DBMS_SQL.EXECUTE(c);
      DBMS_SQL.DESCRIBE_COLUMNS(c, col_cnt, rec_tab);
      FOR j in 1..col_cnt
      LOOP
        CASE rec_tab(j).col_type
          WHEN 1 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
          WHEN 2 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_n_val);
          WHEN 12 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_d_val);
        ELSE
          DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
        END CASE;
      END LOOP;
      -- This part outputs the HEADER
      v_fh := UTL_FILE.FOPEN(upper(p_dir),p_header_file,'w',32767);
      FOR j in 1..col_cnt
      LOOP
        v_finaltxt := ltrim(v_finaltxt||','||lower(rec_tab(j).col_name),',');
      END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
      UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      IF NOT v_samefile THEN
        UTL_FILE.FCLOSE(v_fh);
      END IF;
      -- This part outputs the DATA
      IF NOT v_samefile THEN
        v_fh := UTL_FILE.FOPEN(upper(p_dir),p_data_file,'w',32767);
      END IF;
      LOOP
        v_ret := DBMS_SQL.FETCH_ROWS(c);
        EXIT WHEN v_ret = 0;
        v_finaltxt := NULL;
        FOR j in 1..col_cnt
        LOOP
          CASE rec_tab(j).col_type
            WHEN 1 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_v_val);
                        v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
            WHEN 2 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_n_val);
                        v_finaltxt := ltrim(v_finaltxt||','||v_n_val,',');
            WHEN 12 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_d_val);
                        v_finaltxt := ltrim(v_finaltxt||','||to_char(v_d_val,'DD/MM/YYYY HH24:MI:SS'),',');
          ELSE
            v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
          END CASE;
        END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
        UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      END LOOP;
      UTL_FILE.FCLOSE(v_fh);
      DBMS_SQL.CLOSE_CURSOR(c);
    END;This allows for the header row and the data to be written to seperate files if required.
    e.g.
    SQL> exec run_query('select * from emp','TEST_DIR','output.txt');
    PL/SQL procedure successfully completed.Output.txt file contains:
    empno,ename,job,mgr,hiredate,sal,comm,deptno
    7369,"SMITH","CLERK",7902,17/12/1980 00:00:00,800,,20
    7499,"ALLEN","SALESMAN",7698,20/02/1981 00:00:00,1600,300,30
    7521,"WARD","SALESMAN",7698,22/02/1981 00:00:00,1250,500,30
    7566,"JONES","MANAGER",7839,02/04/1981 00:00:00,2975,,20
    7654,"MARTIN","SALESMAN",7698,28/09/1981 00:00:00,1250,1400,30
    7698,"BLAKE","MANAGER",7839,01/05/1981 00:00:00,2850,,30
    7782,"CLARK","MANAGER",7839,09/06/1981 00:00:00,2450,,10
    7788,"SCOTT","ANALYST",7566,19/04/1987 00:00:00,3000,,20
    7839,"KING","PRESIDENT",,17/11/1981 00:00:00,5000,,10
    7844,"TURNER","SALESMAN",7698,08/09/1981 00:00:00,1500,0,30
    7876,"ADAMS","CLERK",7788,23/05/1987 00:00:00,1100,,20
    7900,"JAMES","CLERK",7698,03/12/1981 00:00:00,950,,30
    7902,"FORD","ANALYST",7566,03/12/1981 00:00:00,3000,,20
    7934,"MILLER","CLERK",7782,23/01/1982 00:00:00,1300,,10The procedure allows for the header and data to go to seperate files if required. Just specifying the "header" filename will put the header and data in the one file.
    Adapt to output different datatypes and styles are required.

  • Problem with .csv file

    Hi,
    I need to download the contents of my internal table in .csv format file.
    So, I am separating the contents of the internal table with ',' (comma) and storing it in another internal table t_download.
    DATA: BEGIN OF t_download OCCURS 100,
            rec(800),
          END   OF t_download.
    Now I download the contents using the FM GUI_DOWNLOAD.
    The file should contain 19 records, but it downloads only 13.
    If I separate the columns with '|' and then download, I get all 19 records downloaded.
    Please let me know if I am doing something wrong.
    Regards.

    Hi,
    Pls chk below link..
    Download CSV file
    This shud answer your question.
    Regds,
    Lalit

  • Problem in DME file generation

    Hi experts,
    I am facing a problem in the generation DME file. Required file format for SPAIN is different to that of the standard file that is generated by executing the standard program RFFOES_T. So I copied that standard program into another and made changes for the new file format in the subroutine DTA_SPANIEN. The file is being generated but with systematically 7 white lines (i.e) blank lines between 2 real lines. Cannot find where the problem lies. Can any one help me in this? Thanks in advance.
    Regards,
    Asha

    Hi Calsadillo,
    The below code says how I have passed the different records to the file
    types: begin of t_file,
             record(1000),
           end of t_file,
         begin of t_h1,
           rec(2),
           trans(2),
           field1(10),
           crtd(6),
         end of t_h1,
         begin of t_h2,
           rec(2),
           trans(2),
           paynm(6),
         end of t_h2,
         begin of t_h3,
           rec(2),
           trans(2),
           payadd(35),
           filler(2),
           field2(40),
         end of t_h3,
         begin of t_h4,
           rec(2),
           trans(2),
           matdat(6),
           deltype,
           contype(2),
           filler(7),
         end of t_h4.
    data : lwa_h1 type t_h1,
           lwa_h2 type t_h2,
           lwa_h3 type t_h3,
           lwa_h4 type t_h4,
           lwa_file type t_file.  
    data : li_file type standard table of t_file.
    *Start Filling Record 1
          l_count        = l_count + 1.
          lwa_h1-rec     = lc_03.
          lwa_h1-trans   = lc_56.             
          lwa_h1-field1  = t001z-paval.
          lwa_h1-crtd = '0000000O02XR'.
          lwa_file       = lwa_h1.
          APPEND lwa_file TO li_file.
    *Start Filling Record 2
          l_count       = l_count + 1.
          lwa_h2-rec    = lc_03.
          lwa_h2-trans  = lc_56.
          lwa_h2-paynm  = t001-butxt.
          lwa_file      = lwa_h2.
          APPEND lwa_file TO li_file.
    *Start Filling Record 3
          l_count       = l_count + 1.
          lwa_h3-rec    = lc_03.
          lwa_h3-trans  = lc_56.
          lwa_h3-payadd  = adrc-city1.
        lwa_h3-filler    = lc_06.
          lwa_file      = lwa_h4.
          APPEND lwa_file TO li_file.
    *Start Filling Record 4
          l_count       = l_count + 1.
           lwa_h4-rec    = lc_03.
          lwa_h4-trans  = lc_56.
          lwa_h4-matdat = lc_030.
          lwa_h4-deltype    = lc_004.
          lwa_h4-contype    = lc_003.
          lwa_file      = lwa_h3.
          APPEND lwa_file TO li_file.
      OPEN DATASET par_unix FOR OUTPUT IN LEGACY TEXT MODE
                                    IGNORING CONVERSION ERRORS.
      IF sy-subrc IS INITIAL.
        LOOP AT li_file INTO lwa_file.
          TRANSFER  lwa_file TO  par_unix.
        ENDLOOP.
        CLOSE DATASET par_unix.
      ENDIF.
    where par_unix is the file name entered on the selection screen.
    Now where can i fit the two declarations of code which you have sent? Can you please explain.
    Regards,
    Asha

  • Problem with CSV file and Ostermiller library

    Hello, I'm using Ostermiller library to generate file with csv format.
    The problem is that this library seems to have a bug when it comes to data including the character carriage return.
    For example I try to record a file with the following data:
    Field1 = Text1
    Field2 = This is
    a long text including
    carriage return
    characters
    Theoritically the file should looks like this:
    Text1,"This is
    a long text including
    carriage return
    characters"
    But the library creates the following:
    Text1, This is\ra long text including\rcarriage return\rcharacters
    The \r you can see in not 0x0D but 0x5C 0x72 !!!!!!
    This is not a correct CSV format and then I cannot reopen the file correctly with Excel.
    Did you already see this problem with this library?
    Any idea?

    Even when i use ExcelCSVPrinter the text appears as -
    5,1,First aid course,234,,"Boeing Everett Plant \nRoom No 108",2007-11-26,02:15:00,,04:15:00,Scheduler S.,Resourcing,,hiii abc
    when I open this text in notepad. It formats properly when I open in excel as follows
    5,1,First aid course,234,,"Boeing Everett Plant
    Room No 108",2007-11-26,02:15:00,,04:15:00,Scheduler S.,Resourcing,,hiii abc
    IS there something else that I need to do?

  • Problem in output file generation in AL11 tcode - need quick help

    Hi experts,
    Scenario:- Proxy to File ( file is generated in AL11 directory folder.
    Probelm:
    Last field value is generated with correct values in SXMB_MONI output payload, But when I open a file in AL11 the last field fiew value are getting trimmed. Also when I get the dump of file - file also has the same trimmed value.
    For example: Expected value in MONI is 12345 where as in AL11 i m getting 1234
    I have reduce the length of previous fields and run the scenario again - for this scenario all the values are coming in AL11 as well as in MONI.
    Is somthing to do with screen width length.
    Deepak Jaiswal.
    Edited by: deepak jaiswal on Aug 6, 2009 11:45 AM

    HI ,
          Its a problem with File Transfer tcode sxda_tools and AL11 sometimes. Ask your Basis team to give you file contents from PI Server directly.  This file should have right contents as you expected.
    I also faced the same problem and wasted lot of time to solve it.
    Hence, check the folder where you dump this file on application server and give this path to basis to provide you this file from server.
    Regards,
    Anurag Garg

  • Problem in output file generation in AL11 tcode

    Hi experts,
                     I am doing one proxy to file scenario.file to be generated as a fixed length format .when i run the scenario, i am getting all the values in output payload correctly in Sxmb_moni and file is also created.But when i scroll upto last to the file generated in ALL11 directory. I am not getting whole values that generated in output payload.In my case my last value is generating in output payload but i am getting trimmed value in file generated in al11 directory.when i save this text file also in local system , i am again only getting trimmed values as seen in All11 directory.when i tried with changing any field length value before that last field with small length(size)value.I am able to see the whole output payload value in file also while scrolling in the file generated in al11 directory

    >
    deepak jaiswal wrote:
    > Hi,
    >
    > I summed the fixed length of all fields and it's sum comes 515. But it displays me upto length 514 values in file generated in al11 only.
    >
    > Please help me to reslove this issue.
    >
    > Deepak jaiswal
    Since in AL11 you can have only 256 char in a row, so I hope you will be getting 3 rows in your case. Correct??
    First row 256
    2nd   row 256
    3rd   row   1
    Total     513
    So out will be 513 which is correct because from first 2 rows it has truncated 1 char from both rows. Please check your output length again to confirm.
    So it is not a problem neither your's nor AL11's. When you write your file at legacy system then it will show the correct data or use FTP adapter and write the file on your local machine to see all the char.
    I hope this solves your problem.

Maybe you are looking for

  • "Load report failed" after upgrading to Crystal Report for VS 2010

    I switched to Visual Studio 2010 few months ago. Today, I needed to update my Crystal Report reports, so I installed the Crystal Report for Visual Studio 2010 on my local server. I updated my reports and it works fine on my local server. On my remote

  • RoboHelp HTML - Trying to open any browse or explorer window causes a freeze

    Hi, I am using RoboHelp 9 HTML. Trying to open any Browsing window in the project, like the browse button in Insert Image or even a project, not through Recent, causes the application to freeze. I tried uninstalling and re-installing with no luck. Pl

  • No balance in pla

    hello, The scenario is when i try to post excise doc. in j1ih  other adjust. tab after entering the relevent value when i click post tab the system gives messege that no balance in pla i was posting for PLA. tell me how we can maintained the balance

  • Interactive Report - default Filter

    I'm trying to create a report which would have a default filter (filter users can uncheck) which would be based on logged user :APP_USER in another words if user JIRI logs in and goes to a page, he can see by default all his records (one of my fields

  • Backup 3.1.2

    Does anyone know... I'm currently running Backup, but I can't find any info on whether it backs up for all users on the mac, or do I need to run this on each account? Thanks