Flat File Format Issue

Hi All,
I have to load a flat file of the type Text (Tab Delimited) (*.TXT) in BI. How can i do that.
It is coming from an external system and hence i cannot have the format change. What should i do in
my flat file data source to be able to upload it
Thanks
Rashmi.

Hi Rashmi,
        There can be two formats of flat file uploading ...
one is ascii which means fixed length format .... in which there is no need of delimitation .... according to fixed length each value will go into corresponding field...
the other is the csv (comma seperate value) which the fields are seperated by a commaa...
as u are saying the input file is tab delimited u need to change the file format into any of the above types before u load into bi...
for if u are using a process chain u has to manually convert this file format and then load in the application server .
for if u are loading manaully u can edit the file to this format and load from ur work station.......
Regards
vamsi
Edited by: vamsi talluri on Feb 6, 2009 5:36 AM

Similar Messages

  • Standardize a flat file format w.r.t OWB

    Hi Pat,
    or anyone,
    Can u help me to standardize a flat file format with respect to OWB?
    I have 15 financial source systems. The client agreed in giving the data in csv format. Now I need to standardize the feed file so that it would be easy for me handle it in OWB. Please note depending on my standardization the client would give me the data and in turn it would be easy for me to handle it OWB External Tables. The records size for a single feed file is 1 million. So my questions are
    1.How should I standardize a date field in flat file?. Please note target date field is default date in Oracle.
    2.How should I standardize an address field? I.e. normally here I have 3 address fields. But I want the output in 1 field. What is the best possible methodology to do this? Or can I handle this issue in OWB front.
    3.How should I standardize a feed file name (generic)?
    4.How should I standardize a feed file folder (generic)? Please note the folder will reside in UNIX environment. The concept is, there would be respective folders in UNIX and the feed files would automatically fall into these respective folders in a fortnight basis. So what should I do to achieve this? I know I have to write a shell script for the automatic process. If so how will I do this and schedule it.
    5.And w.r.t above question, how will I tell OWB to go and fetch the files in those folders. I know it is done while mapping. But my question is Is that the only best way I can achieve it?
    6.If I left anything related to feed file, please tell me?
    Please note that based on my standardization, the client would extract the data and give it to me. So if I standardize the feed file which would be easy for me to handle it in OWB, then that would be a great success. And also note this is for a long run. Minimum 10 years.
    Regards,
    rem.

    Good afternoon Rem,
    It's all very theoretical indeed, talking about standards and all. I'll try to answer some of your questions according to my experiences.
    @1, Usually we load flat files into external tables using varchar2-fields only, occasionally number fields; dates however are always loaded into varchar2's and then when reading the external table into the staging area they are converted to Oracle dates using a simple TO_DATE transformation.
    @2, I don't think there's a real standard for that, it should simply meet YOUR needs, so when loading EXT into STG simply perform a transformation concatenating those 3 fields to form one.
    @5, you have to register the location where you have defined your flat files, this should be some Unix location; after that the connector between EXT and STG will know what to do when running the mappings from EXT to STG (thus: no other action is needed to 'tell OWB' where to fetch data from flat files).
    As far as I know there a no standards in naming conventions for files and folders, just go with what seems logical/practical in your situation.
    Regards, Patrick

  • New EDI Implementation - Flat File format for SPSI in eCommerce Gateway

    Hi,
    we are trying to setup EDI with a business partner, goal is to import a DELFOR message into Oracle Order Management. Plans are: translate the inbound message into the SPSI flat file format, load it through eCommerce Gateway / Release Management into Order Management.
    For this purpose I'd like to test to the import of a SPSI flat file into eCommerce Gateway / Release Management, however I have difficulties to get a proper flat file. Is there a way to generate example flat files somehow? I tried to generate my own flat file, based on the Oracle manuals, but how can I be sure that the format is correct (besides of running the import program)? Any advice on this?
    Or is there a better way, e.g. through XML files?
    Thanks in advance,
    David.

    Hi, thanks for the response.
    Yes I use that report already to get the details of the flat file. But what I am missing is for example which records are repeated, e.g. for every 4000 record, do I need one 0010 record, or can I put multiple 4000 records within one 0010 record?
    regards,
    David.

  • Flat file format - importing vs typing it out

    Hi.  I've heard there is a way to import the format (rather that typing it out in connection mgr) of a fixed width flat file.  What r the general steps or where is there a set of instructions online?  I generally have a txt doc with column
    names (lots of them) , widths etc that I would use for import and could easily transfer them to a spreadsheet if necessary.   So far the word import is bringing up a lot of unwanted subject matter in my web searches.   I am spending a lot
    of time typing out flat file formats in conn mgrs and am thinking it may be worth a look at ssis's ability to "import" a format.   We generally land our data in "untyped" staging tables 1st so ssis could assume everything is string.  
    I dont really want to read my records as one long string and parse them later.

    thx.   I suppose I can write and share with the community some general t-sql that generates the following but would like to know a couple of things. 
    First, is any old guid ok in each of the places where a guid is indicated?  I would generate a new one for each guid.
    Second, what is the locale id?
    Of course i would only paste the results in a pkg (dtsx file) that doesnt reference the cols (beyond the connection stuff) yet.
    What gotchtyas am I looking at if i try this with the intention of pasting the generated code over the top of what ssis already generated (inside the dtsx file) in a one column connector created the usual way?
    <DTS:ConnectionManagers>
    <DTS:ConnectionManager
    DTS:refId="Package.ConnectionManagers[Flat File Connection Manager]"
    DTS:CreationName="FLATFILE"
    DTS:Description="ZZZZZZ"
    DTS:DTSID="{some guid}"
    DTS:ObjectName="Flat File Connection Manager">
    <DTS:PropertyExpression
    DTS:Name="ConnectionString">@[User::fileName]</DTS:PropertyExpression>
    <DTS:ObjectData>
    <DTS:ConnectionManager
    DTS:Format="FixedWidth"
    DTS:LocaleID="1033"
    DTS:HeaderRowDelimiter="_x000D__x000A_"
    DTS:RowDelimiter=""
    DTS:TextQualifier="_x003C_none_x003E_"
    DTS:CodePage="1252"
    DTS:ConnectionString="c:...">
    <DTS:FlatFileColumns>
    <DTS:FlatFileColumn
    DTS:ColumnDelimiter=""
    DTS:ColumnWidth="2"
    DTS:MaximumWidth="2"
    DTS:DataType="129"
    DTS:ObjectName="XXXXX"
    DTS:DTSID="{some other guid}"
    DTS:CreationName="" />
    <DTS:FlatFileColumn
    DTS:ColumnDelimiter=""
    DTS:ColumnWidth="2"
    DTS:MaximumWidth="2"
    DTS:DataType="129"
    DTS:ObjectName="YYYYYY"
    DTS:DTSID="{some other guid}"
    DTS:CreationName="" />

  • PO Idoc in Flat File format

    Hi all,
    I am able to generate the PO Idoc in Flat file format using standrad FM IDOC_OUTPUT_ORDERS. but the Flat file appearing is not in a appropriate format.
    I'm looking to generate the Flat file in such a way that there will be only one line for each Idoc Segment.( with Line termination).
    How to Achieve this.
    Regards,
    S Anand

    Resolved.
    we can use CG3Y transaction.
    Regards,
    S Anand

  • Convert amount from flat file format to user format

    hi,
    how to convert amount from flat file format to user specific format.
    input:  1000.00
    output: 1.000,00 (user specific)
    thanks in advance

    move that value to a type WRBTR variable
    and use write statement.
    data v_wrbtr type wrbtr.
    data v_char(20).
    v_wrbtr = 1000.
    write v_wrbtr to v_char.
    v_char will contain the amount in user format.
    Prerequsite, go to SU3 transaction.
    Defaults tab, chose the decimal notation .
    Regards,
    Ravi

  • Flat file formats

    What are the flat files formats accepted for import and export in mdm?
    How MDM recognize the document type/mapping to perform of incoming file (file name, folder) ?
    What are the flat file format for export (delimited, structured, XML) ?
    Does MDM handle Header in a flat file (import and export) ?
    Does MDM handle several line definition in a flat file (import and export) ?

    Hi Joesph,
    Here are answers to your questions:
    <b>Ans 1--></b> As previously told by adhappan ,you can import data using import manager from the following formats:
    Access--Means From  microsoft access
    Excel -- From an excel file
    ODBC--this is generally used to import data from a  flat file.BY flat file i mean to say a tab comma seperated "csv " file or a ".txt "file.
    Port--to import data from a port .In mdm, port actually refers to a directrory
    XML--from an xml file
    XML Schema-- this is used when you try to import data from a file whosr structure you have predefined in console using an xsd file.
    <b>Ans 2--></b> we specify the file format while connecting to source by mentioning it in the <b>type</b> properties.and mapping is performed in the import manager in the map value tab.
    <b>Ans 3--></b>When you import or export a flat file then data in the first  line of file is considered as header.
    <b>Ans 4 --></b> As i previously told ,MDM will handle header.
    <b> Ans 5 --></b>MDM does not handle several line definition of header.
    Hope it will help you.<b> Please remark if it really helped you</b>.
    Thanks,
    <b>Shiv Prashant Dixit</b>

  • HOW MANY TYPES OF FLAT FILE FORMAT

    hi gurus......
    Basically there r 3 types of flat file format
    ASC
    DAT
    BIN
    i am correct or wrong???
    what is the difference between these formats?

    Hi,
    you are right.
    This is from the documentation available in GUI_UPLOAD func.module:
    'BIN'.
    The system reads a binary representation of the internal table from the
    file. The data is neither converted nor interpreted in this case.
    Usually, reasonable results are obtained only if the internal table
    consists of a single column of type 'X'.
    'DAT'
    The components of the internal table are filled from the file. If the
    table contains several columns, the entries in the file must be
    separated by tabs. No conversion exits are performed. eg: .TXT with tab delimited
    The following applies for the different data types:
    I or N or P or F
    The numbers must be formatted according to the decimal representation
    defined in the user settings.
    D
    The date must be formatted according to the date format defined in the
    user settings.
    T
    The time must be formatted as hh:mm:ss.
    'ASC'.
    The components of the internal table are filled from the file. Only
    data types with fixed length are allowed. The data must be contained
    in the file in its full length. eg: Like ·TXT or ·CSV files
    Anirban

  • JCA flat file read issue

    Hi,
    We needs to read a flat file and transform it to destination xml format. Then send it to destination file location.
    Steps we have done:
    1. Create JCA adapter and configure the flat file schema
    2. Create proxy based on the jca
    3. create transformation for the source to target schema (this has no namespace)
    4. Create BS for sending the output
    Everything workins as expected when testing from OSB test console. But then the file is placed in the source folder, the output xml has the namespace xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" to the root node.
    e.g,
    <Root xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" >
    <Child>
    <Child1/>
    <Child2/>
    <Child3/>
    </Child>
    </Root>
    But expected output is
    <Root>
    <Child>
    <Child1/>
    <Child2/>
    <Child3/>
    </Child>
    </Root>
    We tried converting the xml to string then repalcing the xmlns:soap-env="http://schemas.xmlsoap.org/soap/envelope/" value with balnk.
    Also we tried with hadcorded xml using assign action instead of transformation. Even the harded xml also having the namespace to the root node.
    But the end system is failing due to this namespace value.
    Help me to resolve the issue.
    Thanks,
    Vinoth
    Edited by: Vinoth on Jun 8, 2011 10:12 PM

    Ideally your endsystem should not fail if you specify any number of namespace identifiers in the XML unless you are using them in elements or attributes within the XML. They should try to resolve this on their side.
    But to see whats going on in OSB, can you please paste log the $body variable in the before the publish action and paste the content here. Or send the sbconfig of the Proxy and business to me on my email mentioned in the profile if possible.

  • Flat file load issue-Urgent

    Hello
    I'm trying to load a master data infoobject with a flat file and getting errors.
    The infoobject ZCOLLECT_DATE has two fields - ZQUARTER and ZCOLLECTDATE
    ZCOLLECT_DATE is of Data type NUMC, Length 5, conversion routine PERI5 and Output length 6
    I created a flat file in Excel the following way:
    Left the first row blank
    Second column - 1/2008 (for ZQUARTER)
    Thrid column - 01/11/2008 (for ZCOLLECTDATE)
    I saved it as a CSV (save as-> csv).
    In the infopackage, I've selected this file in the External data tab, File type CSV, Data separator ; , Escape sign "
    When I try to load the infoobject, I get the following error:
    "Error in conversion exit CONVERSION_EXIT_PERI5_INPUT"
    I thought this was because the Length in my infoobject was 5, whereas ZQUARTER is of length 6 in my flat file (eg-1/2008) and ZCOLLECTDATE is of format mm/dd/yyyy.
    Any thoughts on what the issue is?
    Thanks
    Alex

    Hi Hemant
    I changed it to 12008, 22008 etc.. but still get the error ""Error in conversion exit CONVERSION_EXIT_PERI5_INPUT"
    Also, when I save the xls file as CSV, I get two pop-ups saying some features are not compatible and only those features that are compatible will be saved. Are these normal warnings, or could they be causing the issue?
    Also, in the infoobject, since it only consists of dates, do I need to specify conversion routine PERI5 in Infoobject properties/maintenance?
    Thanks

  • RFBIBL00 Flat File Creation issue

    Hello All,
    I'm using "RFBIBL00" to post the FI documents, and the input to the FB01 transaction screen is obtained from a flat file from AL11 directory.
    The issue here is regarding the date format of the Document date and Posting date. If the date format is as YYYYMMDD in the flat file, and if I'm using the date parameter (in the default option of one's SAP own data) other than starting with YYYY, the document is not getting posted and it is throwing error saying that date format is not correct. If I have chosen the date format starting with YYYY, then the document is getting posted successfully.
    Any suggestions/ideas to make the date format as independent?
    Thanks,
    Bharath.

    Hi Bharath
    I think you'll have to use the appropriate conversion exit for the same. Choose from the below options which suit you.
    SAP Internal format is 'YYYYMMDD' Example: '19990123'
    1. CONVERT_DATE_TO_EXTERNAL
    2. CONVERT_DATE_TO_INTERNAL
    3. CONVERSION_EXIT_PDATE_INPUT Conversion Exit for Domain GBDAT: DD/MM/YYYY -> YYYYMMDD
    4. CONVERSION_EXIT_PDATE_OUTPUT Conversion Exit for Domain GBDAT: YYYYMMDD -> DD/MM/YYYY
    5. CONVERSION_EXIT_IDATE_INPUT External date INPUT conversion exit
    6. CONVERSION_EXIT_IDATE_OUTPUT External date OUTPUT conversion exit
    7. CONVERSION_EXIT_LDATE_OUTPUT Internal date OUTPUT conversion exit (e.g. YYYYMMDD)
    8. CONVERSION_EXIT_SDATE_INPUT External date (e.g. 01.JAN.1994) INPUT conversion exit
    9. CONVERSION_EXIT_SDATE_OUTPUT Internal date OUTPUT conversion exit (e.g. YYYYMMDD)
    The date coming in from the flat file will certainly tick to one particular format i guess.
    Hope this helps.
    Harsh

  • Flat File  mapping issue

    Hello All,
    I am trying to an extract using flat file method in BI 7.0. I have some 150 fields in my CSV file but i wanted is just about 10 which are scattered around in the CSV file.When i Read Preview Data in the Preview tab i see incorrect data in there.And that too all under one tab, basically all under one field , though in the extraction tab for  Data Seperator i have been using ; and also tried checking the HEX box, for escape i have the ", tried HEX for this aswell.For rows to ignore i have 1.
    One thing i would like to know is how will the BI infoobject know where is the position of the flat file field in the  CSV file and where will it be mapped.i know it can be mapped in the Transformations but that is from the flat file datasource, i am asking about the CSV file.
    Please help me out and tell me what am i doing incorrectly.
    Thanks for your help.Points will be assigned.

    hi,
    use ,and ; as the escape signs.
    system takes care of it when u speicfy the path and name of the file and format as CSV.
    always the system checks the one to one mapping of the falt file fields with the infoobject in the datasource.
    options for u
    1. arrange the neccessary fields in the flat file that exactly maps with infoobjects for mapping. then start loading.
    2. keep as such and load with scattered field and in transformation map the required fields alone.
    second option consumes more memory space unneccessarily.
    For BI 7.0 basic step to laod data from flat (excel files) files for this follow the beloww step-by step directions ....
    Uploading of master data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( master data attributes, text, hierarchies)
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to select Insert Characteristics as info provider
    • Select required info object ( Ex : Employee ID)
    • Under that info object select attributes
    • Right click on attributes and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    4. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. Alternatively monitor icon can be used.
    BW 7.0
    Uploading of Transaction data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    5. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    6. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( Transaction data )
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    7. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to create ODS( Data store object ) or Cube.
    • Specify name fro the ODS or cube and click create
    • From the template window select the required characteristics and key figures and drag and drop it into the DATA FIELD and KEY FIELDS
    • Click Activate.
    • Right click on ODS or Cube and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    8. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. There are two tables in ODS new table and active table to load data from new table to active table you have to activate after selecting the loaded data . Alternatively monitor icon can be used
    Ramesh

  • Flat file truncation issue

    I am attempting to perform a fairly standard operation, extract a table to a flat file.
    I have set all the schemas, models and interfaces, and the file is produced how I want it, apart from one thing. In the source, one field is 100 characters long, and in the output, it needs to be 25.
    I have set the destination model to have a column physical and logical length of 25.
    Looking at the documentation presented at http://docs.oracle.com/cd/E25054_01/integrate.1111/e12644/files.htm - this suggests that setting the file driver up to truncate fields should solve the issue.
    However, building a new file driver using the string 'jdbc:snps:dbfile?TRUNC_FIXED_STRINGS=TRUE&TRUNC_DEL_STRINGS=TRUE' does not appear to truncate the output.
    I noticed a discrepancy in the documentation - the page above notes 'Truncates strings to the field size for fixed files'. The help tooltip in ODI notes 'Truncates the strings from the fixed files to the field size'. Which might explain the observed lack of truncation.
    My question is - what is the way to enforce field sizes in a flat file output?
    I could truncate the fields separately in each of the mapping statements using substr, but that seems counter-intuitive, and losing the benefits of the tool.
    Using ODI Version:
    Standalone Edition Version 11.1.1 - Build ODI_11.1.1.5.0_GENERIC_110422.1001

    bump
    If this is an elementary issue, please let me know what I've missed in the manual.

  • Urgent: Flat file load issue

    Hi Guru's,
    Doing loading in data target ODS via flat file, problem is that flat files shows date field values correct of 8 characters but when we do preview it shows 7 characters and loading is not going through.
    Anyone knows where the problem is why in the preview screen it shows 7 characters of date when flat file has 8 characters.
    Thanks
    MK

    Hi Bhanu,
    How do i check if conversion is specified or not and another thing is it is not just one field we have 6 date fields and all of them are showing 7 characters in PSA after loading where as flat file have 8 characters in all of the 6 date fields.
    In PSA I checked the error message it shows 2 error messages:
    First Error message:
    The value '1# ' from field /BIC/ZACTDAYS is not convertible into the    
    DDIC data type QUAN of the InfoObject in data record 7 . The field      
    content could not be transferred into the communication structure       
    format.                                                                               
    em Response                                                                               
    The data to be loaded has a data error or field /BIC/ZACTDAYS in the    
    transfer structure is not mapped to a suitable InfoObject.                                                                               
    The conversion of the transfer structure to the communication structure 
    was terminated. The processing of data records with errors was continued
    in accordance with the settings in error handling for the InfoPackage   
    (tab page: Update Parameters).                                                                               
    Check the data conformity of your data source in the source system.                                                                               
    On the 'Assignment IOBJ - Field' tab page in the transfer rules, check    
    the InfoObject-field assignment for transfer-structure field              
    /BIC/ZACTDAYS .                                                                               
    If the data was temporarily saved in the PSA, you can also check the      
    received data for consistency in the PSA.                                                                               
    If the data is available and is correct in the source system and in the   
    PSA, you can activate debugging in the transfer rules by using update     
    simulation on the Detail   tab page in the monitor. This enables you to   
    perform an error analysis for the transfer rules. You need experience of  
    ABAP to do this.                                                                               
    2nd Error Message:
    Diagnosis                                                                               
    The transfer program attempted to execute an invalid arithmetic          
        operation.                                                                               
    System Response                                                                               
    Processing was terminated, and indicated as incorrect. You can carry out 
        the update again any time you choose.                                                                               
    Procedure                                                                               
    1.  Check that the data is accurate.                                                                               
    -   In particular, check that the data types, and lengths of the     
                transferred data, agree with the definitions of the InfoSource,  
                and that fixed values and routines defined in the transfer rules                                                                               
    -   The error may have arisen because you didn't specify the      
             existing headers.                                                                               
    2.  If necessary, correct the data error and start processing again.  
    Thanks
    MK

  • Download report into a flat file format

    Hi Friends,
    I would like to know your thoughts on what is the best practice most of you using to download a report into a flat file(NOT csv).
    In my application business users want a report to be downloaded into their local desktop (just like a .csv) but into a flat file and they have their byte specific format.
    My concerns :
    1) Is there any straight method to download a report into a flatfile (like it is now for CSV) from Apex 3.1.2 ?
    2) If I write my own code using UTL_FILE, can I let the user to save them to their desired directory on their desktop ?
    Your thoughts would definitely help me in delivering a right solution to the users in this regard.
    Thanks a lot.
    Raj.

    Raj:
    You could create a 'Before Header' page process to create this 'flat file'. The code for the page process will be something similar to declare
    v_file_name   VARCHAR2 (2000) := 'fixedFieldsFile.txt';
    --- Client Record Format
    f1 char(20);
    f2 char(20);
    f3 char(20);
    ---- End format
    begin
    OWA_UTIL.mime_header ('application/txt', FALSE);
    htp.p('Content-Disposition:attachment;filename="'|| v_file_name|| '"');
    OWA_UTIL.http_header_close;
    for c1 in (select * from emp) loop                           ----- The report query
    f1:= c1.empno;f2:=c1.ename;f3:=c1.sal;
    htp.p(f1||f2||f3);
    end loop;
    apex_application.g_unrecoverable_error:=true;
    exception when others then
    null;
    end; Varad

Maybe you are looking for

  • While "Invoice Cancelling- Accounting entry not effected in Trial balance"

    Dear All, Any one knows this please revert back.this Issue is very urgent (We have to Close the Periods) Issue: While "Invoice Cancelling- Accounting entry not effected in Trial balance" Invoice which is cancelled and accounting events for Invoice &

  • Exception when trying to acess a jsf application

    Im using eclipse 3.3 , tomcat 5.5.2x index.jsp ======== <html> <body> <jsp:forward page="faces/ourJsf/enterAge.jsp" /> </body> </html> ============ enterAge.jsp ========= <%@ page contentType="text/html; charset=Cp1252" %> <%@ taglib uri="http://java

  • XML out of WS

    Hey All, I hav been working on Web Service where in get a WS object say, output from a WS method call this output to my understanding comes as a XML data that we convert to java code to retrieve data out of it through iterations I wonder is there any

  • PO - Sum of quantities .. larger than total quantity

    Does anybody know what this error means: Sum of quantities >...< larger than total quantity We get this error when we use the function BAPI_PO_CHANGE (updating a purchase order). Kind regards, Lieselot

  • Java security error after 8u31 (Timestamped Jarsigned Applet within valid period of Code Signing certificate)

    Hello,   I have an applet running in embeddad systems. This program runs without any problem since 8u31 update! After this update it starts to give java security warning and stops running. Here is the warning message:   "Your security settings have b