Is exporting to Positional Flat File possible?

Post Author: wilsonricharda
CA Forum: Publishing
I would like to know if it's possible to export data generated from a scheduled Webi report to a positional flat file?I know it's possible to export to a CSV but I'm looking for fixed length positional. Any help/suggestions would be appreciated.Thanks,Richard  

Hi Richard,
Could you please let me know how did you schedule a report to send as an CSV attachment. There are only 3 output formats available currently namely:
1) WebIntelligence Document
2) Excel
3) PDF
It is possible to save a report in CSV format. But there is no CSV format available when a scheduled report needs to be exported. Only the above 3 options are available.
Do we need to configure the CMC for the CSV option to appear while specifying the output format of the report during scheduling. Or is anything else required for the CSV option to come up. Could you please provide your inputs on this.
Thanks,
Meghana

Similar Messages

  • Getting error while reading the positional flat files

    Hi,
    I have a requirement to read a Positional Flat file and convert into XML through oracle B2B. I have created a positional flat file .ecs, XSD and parser .ecs file using the blog http://anuj-dwivedi.blogspot.in/2011/10/handling-positionaldelimited-flat-files.html and updated the same in the server.
    I have created a agreemnt for inbound scenario But while testing the inbound scenario i am getting an error in B2B as :
    Error Code     B2B-50083
    Error Description     Machine Info: Description: Document protocol identification error.
    Error Level     ERROR_LEVEL_COLLABORATION
    Error Severity     ERROR
    Error Text     Document protocol identification error.
    I have tried all the possible ways but getting the same error.
    Please guide me to overcome this.
    Thanks,
    Nishanth

    Nishanth,
    Have you followed the step#2 correctly -
    Copying the parser ecs in the installation directory and adding an entry in XERegistry.xml (of all machines if domain is spanned across multiple machines)
    Also make sure that you have provided Identification Value, Identification Start Position and Identification End Position correctly in Document Defintion. Please remember that start position starts from 1 (instead of 0)
    Regards,
    Anuj

  • Size of Infocube Export to a flat file (.CSV)

    First off, I'm new to BW so I apologize if I use the wrong terminology. I was wondering if there was any way to estimate the size of a flat file infocube export. For example, if I look up all the associated DB tables for a cube, let's call it TESTCUBE, would the total size of those tables be about the size of the TESTCUBE export to a flat file. This assumes that the cube is not compressed in any way. So if the the total DB table size is 100 MB then is the flat file about 100 MB? If not, is there a way to estimate this.
    Thanks!

    Can't say for sure how all DBs that BW run on work, but with Oracle, space is consumed in blocks, not by rows/records. 
    Let's say you have a table where the average row width (record length) is 100 bytes.  In a tablespace that has 8K blocks, in theory, you could fit 81 rows in the block.
    However there is usually a certain amount of freespace left in a block when rows are added to accomodate subsequent inserts/updates.  The BW tables I have looked at default this to 10%. So that means youe have over 800 bytes of the block that might not get filled.  There are other things that can result in space being consumed, such as deletions that will leave dead space, and updates can cause a row to migrate to another block, etc.  Also, space is allocated in chunks as a table grows, and you may add one row to a table that causes a 10MB chunk of storage to be allocated, but it is holding only 1 row.
    So as you can begin to see.  A flat export should require less space than the source table (unless you are using database table compression which could completely invalidate what I have said), but by how much depends on all the things I mentioned, and could change from day to day.
    You could certainly use as a rough rule of thumb, that a flat file export will require the same space and you would be safe (except if you are using db compression).

  • Export Table to flat file and upload to FTP server for external vendor

    I have done some extensive searches on the forums and I think I have an idea but really need some help. Basically have a table that needs to be exported to a flat file and uploaded to a FTP server so that our vendor can get the data Monday-Saturday and automated by sysdate. This data is used to produce daily letters that go out to customers.
    After doing some searching I came across the UTL_FILE package and reading more into this right now but I am not sure if it is what I should be using to create the file.
    I am using TOAD right now and can see how to do it manually however really need the file to create itself and send automatically at 6 am Monday - Saturday with the date at the end of the file name making it a unique file for each day.
    Thanks in advance for all of your help.

    As Justin and others have mentioned I wrote [XUTL_FTP|http://www.chrispoole.co.uk/apps/xutlftp.htm] so you wont find it in the Oracle documentation. It's pure PL/SQL and uses UTL_TCP and implements an integrated PL/SQL FTP client / API. One of it's many overloaded procedures FTP's the contents of a refcursor straight to a FTP server of your choice, meaning no intervening file is required. Since it is pure PL/SQL, it is trivial to schedule using DBMS_JOB/DBMS_SCHEDULER.
    Any questions just ask.
    HTH Chris

  • Positional Flat File Handling in B2B

    Problem:
    There is a big problem with handling positional flat files in Oracle B2B. We are trying to handle Positional Flat File, in UTF-8 with BOM encoding, after creating document definition and setting the agreement, we received:
    B2B-51507 Error: The data starting at position 0 is not recognized as a valid data transmission.
    What’s more, we remembered to locate Parser file into Config/schema location, which usually fixes problem.
    Full description:
    We have a sample flat file, with UNIX (LF) endline character, encoded in UTF-8 with BOM:
    11B6Kamil Mazur     Lazurowa       8 12 CityABCD       PL
    Junior Consultant        MyCompanyX
    StreetAAAAAA       00-000 CityABCD      
    First of all, we have created a guideline *.ecs and schema XSD, like that:
    imgur: the simple image sharer
    During XSD creation we have remembered to confirm, that our endline character is in UNIX standard (LF).
    After that, we have created Parser Schema File, and paste it in the schema location in our server:
    /oracle/Middleware/MyDomain/soa/thirdparty/edifecs/XEngine/config/schema,
    after that we paste the schema location in XRegistry file:
    imgur: the simple image sharer
    After that action, we restarted SOA Server (which was running during these changes).
    When SOA Server was in alive again, we’ve created a document definition :
    imgur: the simple image sharer
    Starting and Ending positions are from 2-5, because of BOM character existence. Then, we created Trading Partner and Inbound Agreement, based on this Document Definition:
    imgur: the simple image sharer
    Then we posted out test file via Generic File Listening Channel, with no Java Callout and other properties. Unfortunately, we receive error like this:
    Business Message :
    Id
    AC18017B146145ADDCD00000178AA4F9
    Message Id
    AC18017B146145ADDCB00000178AA4F8-1
    Refer To Message
    Refer To Message
    Sender Type
    Name
    Sender Value
    Szkolenie
    Receiver Type
    Name
    Receiver Value
    MyCompany
    Sender
    Szkolenie
    Receiver
    MyCompany
    Agreement Id
    PersonIn
    Agreement
    PersonIn
    Document Type
    Person
    Document Protocol
    PositionalFlatFile
    Document Version
    Logistics2013
    Message Type
    REQ
    Direction
    INBOUND
    State
    MSG_ERROR
    Acknowledgement Mode
    NONE
    Response Mode
    ASYNC
    Send Time Stamp
    2014-05-19 14:00
    Receive Time Stamp
    2014-05-19 14:00
    Document Retry Interval(Channel)
    0
    Document Remaining Retry(Channel)
    0
    Document Retry Interval(Agreement)
    Document Remaining Retry(Agreement)
    Native Message Size
    141
    Translated Message Size
    Business Action Name
    Business Transaction Name
    Xpath Name1
    Xpath Value1
    Xpath Expression1
    Xpath Name2
    Xpath Value2
    Xpath Expression2
    Xpath Name3
    Xpath Value3
    Xpath Expression3
    Correlation From XPath Name
    Correlation From XPath Value
    Correlation From XPath Expression
    Correlation To XPath Name
    Correlation To XPath Value
    Correlation To XPath Expression
    Wire Message
    Wire Message
    Application Message
    Application Message
    Payload Storage
    Payload Storage
    Attachment
    Attachment
    Label
    soa_b2b_ - Thu May 15 15:04:40 CEST 2014 - 1
    Collaboration Id
    AC18017B146145ADD9200000178AA4F7
    Collaboration Name
    Collaboration Version
    Business Action Name
    Exchange Protocol Name
    Generic File
    Exchange Protocol Version
    1.0
    Interchange Control Number
    Group Control Number
    Transaction Set Control Number
    Error Code
    B2B-51507
    Error Description
    Machine Info: (soatest.corp.prv) The data starting at position 0 is not recognized as a valid data transmission.
    Error Level
    ERROR_LEVEL_COLLABORATION
    Error Severity
    ERROR
    Error Text
    Payload validation error.
    Wire Message:
    Id
    AC18017B146145ADAB400000178AA4F1
    Message Id
    AC18017B146145ADAB400000178AA4F1
    Business Message
    AC18017B146145ADDCD00000178AA4F9
    Packed Message
    Packed Message
    Payload
    Payload
    Protocol Message Id
    test_4value.txt@AC18017B146145ADAE500000178AA4F5
    Refer To Protocol Message Id
    Protocol Collaboration Id
    Protocol Transport Binding
    filename=test_4value.txt filesize=143 ChannelName=SzkolenieChannel file_ext=txt fullpath=/home/oracle/POC/positional/Krzysiek/test_4value.txt timestamp=2014-05-07T10:23:30.000+01:00 tp_profile_name=Szkolenie MSG_RECEIVED_TIME=Mon May 19 14:00:37 CEST 2014
    Message Digest
    Message Digest
    Digest Algorithm
    Transport Protocol
    File
    Transport Protocol Version
    1.0
    Url
    file://localhost//home/oracle/POC/positional/Krzysiek
    security
    Transport Headers
    filename=test_4value.txt filesize=143 ChannelName=SzkolenieChannel file_ext=txt fullpath=/home/oracle/POC/positional/Krzysiek/test_4value.txt timestamp=2014-05-07T10:23:30.000+01:00 tp_profile_name=Szkolenie MSG_RECEIVED_TIME=Mon May 19 14:00:37 CEST 2014
    certificates
    certificates
    State
    ERROR
    Reattempt Count
    Error Code
    B2B-51507
    Error Description
    Machine Info: (soatest.corp.prv) The data starting at position 0 is not recognized as a valid data transmission.
    Error Text
    Payload validation error.
    exchange Retry Interval
    exchange Remaining Retry
    Message Size
    141
    Payload seems to be unchanged. Problem appears, when our B2B works in Cluster (but both configuration parameters b2b.HAInstance  b2b.HAInstanceName are set properly for both soa-servers):
    imgur: the simple image sharer
    There are no additional info in logs :
    [2014-05-19T14:07:47.994+02:00] [soa_server1] [NOTIFICATION] [] [oracle.soa.b2b.engine] [tid: DaemonWorkThread: '20' of WorkManager: 'wm/SOAWorkManager'] [userId: <anonymous>] [ecid: a8bc74c6eb84aa5b:452af1da:146046c88de:-8000-00000000000598a0,0] [APP: soa-infra] Engine: processIncomingMessageImpl: Message id = AC18017B14614616E1A00000178AA510-1 FromParty = Szkolenie Doctype = Person version = Logistics2013
    That’s unfortunately all. Is there any configuration that we missed? How to fix this problem?

    Hi Anuj,
    I have added a transformation in the mediator component and I'm now able to transfer the opaque data from Oracle B2B to composite and then write the contents of flat file to another specified file thorugh file adapter.
    However I still have few more issues in different scenarios:
    I have tried the excercise that was mentioned in your blog:
    http://www.anuj-dwivedi.blogspot.com/2011/10/handling-positionaldelimited-flat-files.html
    I tried to create a flat file with two records or lines as mentioned in the blog. First line contains header information whose length is 57 characters with 5 Fileds in it. Second line contains the actual data with 57 characters and 5 fields in it. In the agreement that was created to handle this flat file, I have checked the Validate & Translate check box.
    During validation, B2B is failing to treat the second line as a new record and it considering it in one record and the validation is failing as nothing is specified from 58th character in the ECS & XSD files.
    Following is the error:
    Error -: B2B-51507: Payload validation error.: Machine Info: (corpdevsoa10) Extra Field was found in the data file as part of Record HEADER. Record HEADER is defined in the guideline at position 1.{br}{br}This error was detected at:{br}{tab}Record Count: 1{br}{tab}Field Count: 6{br}{tab}Characters: 57 through 116
    Please advice me in resolving this issue.
    Thanks,
    Krishna.

  • Export data to flat file with no delimiter

    Hi I've 3 columns in a view that i want to export to a flat file. I don't want the flat file to contain any delimiter such as commas , tabs etc..
    I am using fixed width with the following setting
    Header row delimiter {CR}-{LF}
    Advance Options : 
    InputColumnwith  125
    InputColumnwith 187
    InputColumnwith  7
    When i preview that data it looks correct , divided up into 3 columns
    However when the data is exported to the flat file i get just one long string in the flat file on the first row.
    What do i need to do to get the end of line characters to be identified 
    Thanks In advance 

    Column delimiter is here
    Otherwise header row terminates a file line
    Arthur
    MyBlog
    Twitter

  • Data convertion while exporting data into flat files using export wizard in ssis

    Hi ,
    while exporting data to flat file through export wizard the source table is having NVARCHAR types.
    could you please help me on how to do the data convertion while using the export wizard?
    Thanks.

    Hi Avs sai,
    By default, the columns in the destination flat file will be non-Unicode columns, e.g. the data type of the columns will be DT_STR. If you want to keep the original DT_WSTR data type of the input column when outputting to the destination file, you can check
    the Unicode option on the “Choose a Destination” page of the SQL Server Import and Export Wizard. Then, on the “Configure Flat File Destination” page, you can click the Edit Mappings…“ button to check the data types. Please see the screenshot:
    Regards,
    Mike Yin
    TechNet Community Support

  • Exporting tables to flat files

    Hi,
    Can I export tables to flat files(csv) using pl/sql. I mean suppose I have 2 tables 'student' and 'subject' and I want to create two flat files student.csv and subject.csv Please let me know, if I can do it using pl/sql for both the tables in one go.
    Even if I can do it for one table, than I can write a script to automate for all the tables. Please let me know with your suggestions. Appreciate your help.
    Thanks

    As sys user:
    CREATE OR REPLACE DIRECTORY TEST_DIR AS '\tmp\myfiles'
    GRANT READ, WRITE ON DIRECTORY TEST_DIR TO myuser
    /As myuser:
    CREATE OR REPLACE PROCEDURE run_query(p_sql IN VARCHAR2
                                         ,p_dir IN VARCHAR2
                                         ,p_header_file IN VARCHAR2
                                         ,p_data_file IN VARCHAR2 := NULL) IS
      v_finaltxt  VARCHAR2(4000);
      v_v_val     VARCHAR2(4000);
      v_n_val     NUMBER;
      v_d_val     DATE;
      v_ret       NUMBER;
      c           NUMBER;
      d           NUMBER;
      col_cnt     INTEGER;
      f           BOOLEAN;
      rec_tab     DBMS_SQL.DESC_TAB;
      col_num     NUMBER;
      v_fh        UTL_FILE.FILE_TYPE;
      v_samefile  BOOLEAN := (NVL(p_data_file,p_header_file) = p_header_file);
    BEGIN
      c := DBMS_SQL.OPEN_CURSOR;
      DBMS_SQL.PARSE(c, p_sql, DBMS_SQL.NATIVE);
      d := DBMS_SQL.EXECUTE(c);
      DBMS_SQL.DESCRIBE_COLUMNS(c, col_cnt, rec_tab);
      FOR j in 1..col_cnt
      LOOP
        CASE rec_tab(j).col_type
          WHEN 1 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
          WHEN 2 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_n_val);
          WHEN 12 THEN DBMS_SQL.DEFINE_COLUMN(c,j,v_d_val);
        ELSE
          DBMS_SQL.DEFINE_COLUMN(c,j,v_v_val,2000);
        END CASE;
      END LOOP;
      -- This part outputs the HEADER
      v_fh := UTL_FILE.FOPEN(upper(p_dir),p_header_file,'w',32767);
      FOR j in 1..col_cnt
      LOOP
        v_finaltxt := ltrim(v_finaltxt||','||lower(rec_tab(j).col_name),',');
      END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
      UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      IF NOT v_samefile THEN
        UTL_FILE.FCLOSE(v_fh);
      END IF;
      -- This part outputs the DATA
      IF NOT v_samefile THEN
        v_fh := UTL_FILE.FOPEN(upper(p_dir),p_data_file,'w',32767);
      END IF;
      LOOP
        v_ret := DBMS_SQL.FETCH_ROWS(c);
        EXIT WHEN v_ret = 0;
        v_finaltxt := NULL;
        FOR j in 1..col_cnt
        LOOP
          CASE rec_tab(j).col_type
            WHEN 1 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_v_val);
                        v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
            WHEN 2 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_n_val);
                        v_finaltxt := ltrim(v_finaltxt||','||v_n_val,',');
            WHEN 12 THEN DBMS_SQL.COLUMN_VALUE(c,j,v_d_val);
                        v_finaltxt := ltrim(v_finaltxt||','||to_char(v_d_val,'DD/MM/YYYY HH24:MI:SS'),',');
          ELSE
            v_finaltxt := ltrim(v_finaltxt||',"'||v_v_val||'"',',');
          END CASE;
        END LOOP;
      --  DBMS_OUTPUT.PUT_LINE(v_finaltxt);
        UTL_FILE.PUT_LINE(v_fh, v_finaltxt);
      END LOOP;
      UTL_FILE.FCLOSE(v_fh);
      DBMS_SQL.CLOSE_CURSOR(c);
    END;This allows for the header row and the data to be written to seperate files if required.
    e.g.
    SQL> exec run_query('select * from emp','TEST_DIR','output.txt');
    PL/SQL procedure successfully completed.Output.txt file contains:
    empno,ename,job,mgr,hiredate,sal,comm,deptno
    7369,"SMITH","CLERK",7902,17/12/1980 00:00:00,800,,20
    7499,"ALLEN","SALESMAN",7698,20/02/1981 00:00:00,1600,300,30
    7521,"WARD","SALESMAN",7698,22/02/1981 00:00:00,1250,500,30
    7566,"JONES","MANAGER",7839,02/04/1981 00:00:00,2975,,20
    7654,"MARTIN","SALESMAN",7698,28/09/1981 00:00:00,1250,1400,30
    7698,"BLAKE","MANAGER",7839,01/05/1981 00:00:00,2850,,30
    7782,"CLARK","MANAGER",7839,09/06/1981 00:00:00,2450,,10
    7788,"SCOTT","ANALYST",7566,19/04/1987 00:00:00,3000,,20
    7839,"KING","PRESIDENT",,17/11/1981 00:00:00,5000,,10
    7844,"TURNER","SALESMAN",7698,08/09/1981 00:00:00,1500,0,30
    7876,"ADAMS","CLERK",7788,23/05/1987 00:00:00,1100,,20
    7900,"JAMES","CLERK",7698,03/12/1981 00:00:00,950,,30
    7902,"FORD","ANALYST",7566,03/12/1981 00:00:00,3000,,20
    7934,"MILLER","CLERK",7782,23/01/1982 00:00:00,1300,,10The procedure allows for the header and data to go to seperate files if required. Just specifying the "header" filename will put the header and data in the one file.
    Adapt to output different datatypes and styles are required.

  • Positional flat file schemas for input and output files to be generated with the required usecases

    Hello all,
    I need one help regarding the positional flat file schema which contains multiple headers, body and trailers.
    I have attached the sample input file below. This is a batched input and we have to generate the output which I have given below:
    We are unable to flat file schema which replicates the below input file. Please suggest better approach to achieve this.
    Sample Input FIle:
    010320140211ABC XYZ XYZ ABCD201402110 FSPAFSPA005560080441.02000006557.FSPA.IN AB-CD ABCD/AB-CD BETALKORT
    1020140210AN194107123459.FSPA.IN
    [email protected]
    1020140210AN196202123453.FSPA.IN
    [email protected]
    1020140210AN198103123435.FSPA.IN
    [email protected]
    1020140210AN195907123496.FSPA.IN
    [email protected]
    1020140210AN195903123437.FSPA.IN
    [email protected]
    1020140210AN193909123434.FSPA.IN
    [email protected]
    1020140210AN195607123413.FSPA.IN
    [email protected]
    1020140210AN199205123408.FSPA.IN
    [email protected]
    1020140210AN196206123499.FSPA.IN
    [email protected]
    1020140210AN196709123410.FSPA.IN
    [email protected]
    1020140210AN194708123455.FSPA.IN
    [email protected]
    1020140210AN195710123443.FSPA.IN
    [email protected]
    1020140210AN198311123436.FSPA.IN
    [email protected]
    1020140210AN196712123471.FSPA.IN
    [email protected]
    1020140210AV197005123403.FSPA.IN
    98000000000000014000000000000001000000000000015
    010320140211ABC XYZ XYZ PEDB201402111 FSPAICA 005560080441.02000006557.FSPA.IN AB-CDABCD/ABCDBETALKORT
    1020140210AN195111123491.ICA.IN
    [email protected]
    1020140210AV195309123434.ICA.IN
    98000000000000001000000000000001000000000000002
    Output FIle:
    1020140210AN195607123413.FSPA.IN
    [email protected]
    1020140210AN199205123408.FSPA.IN
    [email protected]
    1020140210AN196206123499.FSPA.IN
    [email protected]
    1020140210AN196709123410.FSPA.IN
    [email protected]
    1020140210AN194708123455.FSPA.IN
    [email protected]
    1020140210AN195710123443.FSPA.IN
    [email protected]
    1020140210AN198311123436.FSPA.IN
    [email protected]
    1020140210AN196712123471.FSPA.IN
    [email protected]
    1020140210AN195111123491.ICA.IN
    110019EPS [email protected]
    1020140210AV197005123403.FSPA.IN
    1020140210AV195309123434.ICA.IN
    98000000000000001000000000000001000000000000002
    99000000000000001
    Note: Header is a single line till BETALKORT and also there is a space before email id. That is not getting pasted properly in the files.
    Thanks and Regards,
    Veena Handadi

    Hi all,
    Header is missed from the output file for the above post:
    Please find the output file:
    010320140211ABC XYZ XYZ ABCD201402110 FSPAFSPA005560080441.02000006557.FSPA.IN AB-CD ABCD/AB-CD BETALKORT
    1020140210AN195607123413.FSPA.IN
    [email protected]
    1020140210AN199205123408.FSPA.IN
    [email protected]
    1020140210AN196206123499.FSPA.IN
    [email protected]
    1020140210AN196709123410.FSPA.IN
    [email protected]
    1020140210AN194708123455.FSPA.IN
    [email protected]
    1020140210AN195710123443.FSPA.IN
    [email protected]
    1020140210AN198311123436.FSPA.IN
    [email protected]
    1020140210AN196712123471.FSPA.IN
    [email protected]
    1020140210AN195111123491.ICA.IN
    110019EPS [email protected]
    1020140210AV197005123403.FSPA.IN
    1020140210AV195309123434.ICA.IN
    98000000000000001000000000000001000000000000002
    99000000000000001

  • B2B integration with EBiz for a custom positional flat file

    I have installed SOA 11g with the latest patch. I am confused with my setup since I could not find any good example for my use case. I get a positional flat file from my trading partner (VDA4905). I have done the following so far:
    Created Document Definition : Custom, version 1.0, type PO with identification type as "flat". I have given the first 3 characters of the file to contain "511" as the identification value.
    I have created trading partners, host listening channel (folder as /tmp/b2b) and agreement. Validated and deployed the agreement.
    I created a dummy file with first 3 characters as 511 and put it in the /tmp/b2b. The file is not getting consumed and I see the pretty standard error "Agreement not found from TP is null".
    My questions are:
    1. When I created the document definition, do I need to specify any xsd or ecs file? Since I am using positional flat file, is this mandatory?
    2. I changed the name of the dummy file in the "FROM TP_HOST_seq.txt", still see the same issue.
    Any help is appreciated.
    Shanthi

    When I changed the file name, it gets consumed. Regarding the file name, I have the following question:
    I get the files from different plants and the file names are different for each plant, however, the processing is the same. If I need to have file name in %From TP%_%SEQ%.txt this format, do I need to create multiple TPs such that the file name maps to TP name? Can I use the same agreement for all these dummy TPs that I create?
    Is there anything I can do, to consume all messages in a specific directory irrespective of the file name?
    Thanks
    Shanthi

  • Exporting data to flat files

    Hi,
    Is it possible to export data from any database to flat files using SQL developer and also automate the process by including some sort of scheduler or script?
    Thanks,
    Nitesh

    If the database is Oracle you can export the contents of table to a flat file by right clicking on the table instance and selecting "Export Data" menu item and then selecting the export format

  • Exporting Hierarchy in Flat File Format

    Hi All,
    I exporting hierarchy in local file from transaction code OB58 (Balance sheet Hierarchy and Profit & Loss Hiearchy) , KSH3 (Cost Center Hierarchy), KCH3 (Profit Center Hierarchy). But the hierarchy is commming as it is displayed at the R3 screen.
    I have to load these hierarchies in BW through flat file. Is there any method other that writing an ABAP program by which we can extract these hierarchies in Flat file loadable format.
    Thanks

    Perhap's it can help you
    [https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/0403a990-0201-0010-38b3-e1fc442848cb?overridelayout=true|https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/0403a990-0201-0010-38b3-e1fc442848cb?overridelayout=true]
    It's an ABAP program but it works well and you can load your flat file easily

  • Export table to flat file and need to insert sysdate in flat file column

    Hi, I created an interface to export oracle table to a csv file. All of the table columns are working well. Then I need to insert the sysdate in csv file column.
    I made the mapping as working in staging area, implementation is to_char(sysdate,'dd/mm/yyyy'). But the result is insert 14 to the column.
    I have tried to create a variable refreshing as select to_char(sysdate,'dd/mm/yyyy') from dual, then mapping that to the column in csv file, but it only insert to 1 row and the format is yyyymmdd.
    I have tried to use  SELECT '<%=odiRef.getSysDate( "yyyyMMdd")%>' from dual for the variable, and it also only insert one row to the flat file.
    I used the same methodology in ODI10g, it works fine.
    So, I am wondering how it can be implemented in 11g.
    Thanks

    The first option you have stated seems like the obvious choice - I don't see any reason why this shouldn't work.  What do you mean by "But the result is insert 14 to the column." Do you mean that it inserted the string "14" (I can't imagine why this would be the case) or that it inserted 14 rows?

  • Export table to flat file programmatically

    Hi all,
    i want to know how to export a table/tables
    programmatically to a flat file
    i know that we can do that by right click on the table
    and choose save as .... etc... (in toad)
    But how do i make it with a script..
    thx

    http://osi.oracle.com/~tkyte/flat/index.html

  • Export Report to flat file with spaces

    I have a report that is several columns wide. It queries data from our SQL based accounting package to create a flat file with spaces(must be accurate). When we go export the file we don't get all of the columns. Any ideas?

    Hi,
    Are you exporting to CSV or Excel File?
    Export to Excel.
    Bashir Awan

Maybe you are looking for

  • How to do EP imports through SDM

    Hello All, We need to import the EP transports into the quality system from the development system. We have High availabity EP system into which the transports are to be imported. I have never done this before through SDM. Also i am not very much awa

  • Finding  required enhancement

    hi , can anybody plz tell how to find an enhancement which is required .... for ex.. i have a req that i have to display a popup mess before saving the sales order while Sales order creation/change... how to find the enhancement... plz tell the steps

  • Skydrive music won't copy to itunes media in windows 8.1

    I am using Itunes 11.1.3.8 and Windows 8.1 This new version of windows skydrive keeps the files on the cloud. I have set itunes to copy music to its own local folder. So, when I add folder to itunes, I thought it would locate the skydrive music and d

  • Editing headshots:  iPhoto '08, PSE 6, or Aperture 2?

    Hi, all: I'm not a big photo guy, but I do have an important project I'd like to tackle... I need to edit a couple of headshot-style photos taken of my daughter. The pics were taken on B&W film; I have the negatives, I had a contact sheet printed, an

  • Mercury Playback Engine on Premiere Pro 8.0

    I have the older version of Premiere Pro CC working well on my PC W8 using my GTX680. I discovered that I had installed the new Premiere Pro (2014) on my PC. When I started it up there was a message about the GPU acceleration and I found that the Mer