Flat file adaptor

Hello,
I am new on golden gate and we search for the mechanism of flat file adapter on golden gate.
Do you know if we can apply the transactional changes from oracle database to an another DBMS, by using flat file adapter?
I mean:
- Are datas extracted to a flat file. If yes, from where can we follow the database changes?
I mean imagine:
There is insert, update and delete in the source database.
-is the transaction order of the operation written to a file? Are also the datas written to a file?
-Is it possible for an another application to understand and apply them to an another DBMS?
Could you please guide me?
Thanks

user12299628 wrote:
I am new on golden gate and we search for the mechanism of flat file adapter on golden gate.Oracle GoldenGate can "capture" transactions on a source database, and writes them out to a binary "trail" file. This trail file can be read by the OGG "flat file adapter", which converts this internal, binary format into plain text files that can be loaded into another DB, ETL tool, etc.
See: (latest release, flat-file adapter 11.1.1) http://docs.oracle.com/cd/E18101_01/index.htm
http://www.oracle.com/us/products/middleware/data-integration/goldengate/application-adapters/overview/index.html
Do you know if we can apply the transactional changes from oracle database to an another DBMS, by using flat file adapter?Yes, but usually there is no reason to... You can capture transactions from one DB (e.g., Oracle or MySQL or SQL Server or DB2) and just apply them normally do another database, even a different vendor on a different OS (Oracle, MySQL, SQL Server, DB2, Teradata, etc).
I mean:
- Are datas extracted to a flat file. If yes, from where can we follow the database changes?Yes, data is captured & written out to flat (text) files. The deltas (inserts/updates/deletes) are in the text files. You can use GG basic stats & reports to keep track of number of operations/tranactions captured.
I mean imagine:
There is insert, update and delete in the source database.
-is the transaction order of the operation written to a file? Are also the datas written to a file?Yes, and yes.
-Is it possible for an another application to understand and apply them to an another DBMS?It's up to you to define the output format written to the file, using properties. For example, you can specify the delimiter, quoting mechanism, escape characters, etc.
Edited by: MikeN on Apr 2, 2012 2:06 PM

Similar Messages

  • FDM Flat file adaptor

    Hi All,
    I have a requirement to validate and customise the data stored on a flat file and need the customised output on another flat file.
    Can this be possible in FDM ?
    I already have the Pull adaptor (PULL G4 E). Does it need a target application? Can we do this without the app?.
    Can anybody tell me the high level steps do this?
    Many Thanks

    The PULL adapter is the adapter that will allow you to do this. It natively reads from a flat file, performs mappings/transformations and outputs to a file. You can further customise the output of your target file by modifying the Export Action script in the adapter.
    Edited by: SH on May 24, 2013 11:27 AM

  • XML File to Flat File scenario: Reciver file adaptor content conversion

    Hello Friends,
    Currently I am working on XML File to Flat File Scenario.
    I used receiver side File adaptor with content conversion.
    My receiver side adaptor is giving error. Error is given below.
    Conversion initialization failed: java.lang.Exception: java.lang.Exception: Error(s) in XML conversion parameters found: Parameter '.fieldFixedLengths' or '.fieldSeparator' is missing
    My content conversion parameters are as:
    row.fieldNames                           Customer_ID,Name,Address,Phone
    row.fieldSeparator                      ,
    row.processConfiguration          FromConfiguration
    row.endSeparator                       'nl'
    Can you please suggest me what kind of error is this?
    Regards,
    Narendra

    Still I am facing below given issue.
    2010-06-08 15:34:25     Error     File adapter receiver channel CC_Customer_FlatFile_Distination is not initialized. Unable to proceed: null
    2010-06-08 15:34:25     Error     Exception caught by adapter framework: Channel has not been correctly initialized and cannot process messages

  • XML to flat file

    Hi all,
    Scenario:
    R/3 -> Proxy -> XI -> JMS -> MQ
    Is it possible to deliver a FLAT FILE to MQ via the JMS adaptor?
    I need to get reed of all the XML tags, since the receiving system is expecting a flat structure.
    Thanks - /Thomas

    hi,
    you can use the following module to remove xml tags and namespaces
    the module is XMLAnonymizerBean
    the url:
    http://help.sap.com/saphelp_nw04/helpdata/en/45/d169186a29570ae10000000a114a6b/frameset.htm
    regards,
    ujjwal

  • File adaptor

    hi, i have a doubt . my requirement is i want to send a flat file as a flat file using file adaptor. in that my structure consists 236 fields, i cant specifiy each and every field in that. it takes more time.
    is there is any alternative for this.

    Hi,
    If both sender and receiver are flat file with content conversion with coma saeparated then there is a way to solve it.
    - first my doubt is , you wat the same file from source to target , no structure modification, is it correct , or you want do some thing in mapping ??
    So-
    create a data type under row as root node with 0..unbound, message type . and create outbound interface and inbound interface with same mt.
    so no need of mapping.if you dont have to do any thing in mapping except one to one.
    and rest of config is same .
    Only you have to give the file cc is
    file record structure ; row
    processing parameters
    ignorerecordset ; true
    row.fieldSeparator : ,
    row.endseparator ; 'nl'
    Regards
    Chilla

  • Refreshing the data in a flat file

    Hi
    I am working on data integration.
    I need to fetch data from Oracle data base and then write it to a flat file.
    It is working fine now,but for the next fetch I don't need the old data to remain there in the file.The data should get refreshed and only the newly fetched data should be present.
    After the data is written to the flat file can i rename the file?
    I need the format of the file to be 'File_yyyyMMDDHHmmss.txt'.
    My final question is how should I FTP this to the target?
    Please help me on this as soon as possible since this is needed in an urgent part of the delivery.

    All you ask is achievable:
    1) The IKM SQL to file has a TRUNCATE option, which will if set to YES, will start from a clean file.
    2) You could rename the file after writing it, but why not just write it with that name? If you set the resource name to be a variable, (e.g. #MyProj.MyFilename), and be sure to set the variable in the package before executing your interface, you should be able to get the file you want. Otherwise, you can use the OdiFileMove tool in your package to rename the file.
    To set the name of the variable you can use a query on the database (if you are using Oracle, something like SELECT 'File_'||TOCHAR(SYSDATE) from DUAL.)
    3) ODI has ftp classes built in- you can find the doc under doc\webhelp\en\ref_jython\jyt_ex_ftp.htm
    Hope this helps
    Message was edited by:
    CTS

  • Loading multiple flat file at a time.

    hi experts,
    I am having 15 flat files with same data structure.so how do i load all the fileswith out creating 15 info packges .
    ( say all the files are on the desktop.)
    I had seen the option that ABAP code an be written in info package in extraction tab.
    Can any one share me the abap code
    Regards
    Laxman.

    You can write dynamic code using ABAP if your number of flat files are not fixed...but if you have to 15 flat files always just creating multiple infopackages.
    For code, you have to store file names/name pattern in some table and reading it at run time. You can use function modules starting with BAPI_IPAK* to create/change/start template IP like BAPI_IPAK_START. You can set IP parameters at run time using FM RSBATCH_MAINTAIN_PAR_SETTINGS.
    Kamaljeet

  • InfoSpoke Flat File Extract to Logical Filename

    I'm trying to extract data from an ODS to a flat file. So far, I've found that the InfoSpoke must write to the application server for large data volume. Also, in order for the InfoSpoke to transport properly, I must use logical filenames. I've attempted to implement the custom class and append structure as defined in the SAP document "How To... Extract Data with OPEN HUB to a Customer Defined Logical Filename". I'm getting an error when attempting to import the included transports (custom class code). It appears to be a syntax error. Has anyone encountered this, and, if so, how did you fix it?

    Hello.
    I'm getting a syntax error also.  I did not import the transport, but applied the notes thru the appendix.  When I modified the method "GET_OBJECT_REF_INT" in class CL_RSB_DEST as below, I get a syntax error on the "create object" statement.
        when rsbo_c_desttype_int-file_applsrv.
    *{   REPLACE        &$&$&$&$                                          1
    *\      data: l_r_file_applsrv type ref to cl_rsb_file_applsrv.
          data: l_r_file_applsrv type ref to zcl_rsb_file_logical.
    *}   REPLACE
          create object l_r_file_applsrv
            exporting i_dest    = n_dest
                      i_objvers = i_objvers
    Class CL_RSB_DEST,Method GET_OBJECT_REF_INT
    The obligatory parameter "I_S_VDEST" had no value assigned to it.

  • Dynamic file name from input payload (RFC 2 flat file)

    Hi,
    I have an RFC to flat file scenario. The output flat file has not an XML structure, it's just a plain text file generated with abap mapping.
    In my source interface (RFC), I have a field called <FILENAME>, I want to use the value of that field to create the target file using dynamic file name. But if  in variable substitution I use payload:ZRFC_NAME,1,FILENAME,1 It doesn't work because the dynamic variable substitution try to access to output payload, not the source one...
    What can I do?

    Hi Marshal,
           You can add a extra node to your target strucutre like
    FileName- Node
    --FileName - Element.
    do the mapping from the field filename of RFC to FileName field in u r target strucure. And use this  field path at Refrence in variable subtituion.
    In the Content converison add the Name & Values as below
    FileName.fileldNames -- FileName
    FileName.fieldFixedLengths -- 0
    FileName.fixedLengthTooShortHandling -- Cut
    So the extra field in u r target structure would not populate in u r target text file.
    Cheers
    Veera

  • Non English Charaters in Flat file(Note pad)

    Dear Friends,
    I have a flat file from presentation server which needs to be uploaded and do some validations on that file based on the data existing.
    I have some Russian language words in that file. E.g: "ОЩИ В ПОДГ". If this word exists in the file i need to get some other data from a database table. But when trying to check for the word, these characters are appearing as junk characters like ####&^^.
    I tried logging into RU language also, but still not working. And also tried changing the file as type Unicode, ASCII and also not worked.
    Please let me know how to get identify these words to see the exact russian words.
    Thanks for your help.
    Srinivas.

    Hello,
    Is your file .txt file.If yes,while saving it gives a pop up for file name.There you have option for encoding, there select encoding as
    Unicode and try to upload.
    It worked for me.I tried with the data provided by you.
    data:begin of it_tab occurs 0,
          val type string,
         end of it_tab.
    CALL FUNCTION 'GUI_UPLOAD'
      EXPORTING
        filename                      = 'C:\Documents and Settings\xxx\Desktop\SDN1.txt'
      FILETYPE                      = 'ASC'
      HAS_FIELD_SEPARATOR           = ' '
      HEADER_LENGTH                 = 0
      READ_BY_LINE                  = 'X'
      DAT_MODE                      = ' '
      CODEPAGE                      = ' '
      IGNORE_CERR                   = ABAP_TRUE
      REPLACEMENT                   = '#'
      CHECK_BOM                     = ' '
      VIRUS_SCAN_PROFILE            = VIRUS_SCAN_PROFILE
      NO_AUTH_CHECK                 = ' '
    IMPORTING
      FILELENGTH                    = FILELENGTH
      HEADER                        = HEADER
      TABLES
        data_tab                      = it_tab
    EXCEPTIONS
       FILE_OPEN_ERROR               = 1
       FILE_READ_ERROR               = 2
       NO_BATCH                      = 3
       GUI_REFUSE_FILETRANSFER       = 4
       INVALID_TYPE                  = 5
       NO_AUTHORITY                  = 6
       UNKNOWN_ERROR                 = 7
       BAD_DATA_FORMAT               = 8
       HEADER_NOT_ALLOWED            = 9
       SEPARATOR_NOT_ALLOWED         = 10
       HEADER_TOO_LONG               = 11
       UNKNOWN_DP_ERROR              = 12
       ACCESS_DENIED                 = 13
       DP_OUT_OF_MEMORY              = 14
       DISK_FULL                     = 15
       DP_TIMEOUT                    = 16
       OTHERS                        = 17
    IF sy-subrc <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    Thanks.
    Ramya.

  • Error while uploading data from a flat file to the hierarchy

    Hi guys,
    after i upload data from a flat file to the hierarchy, i get a error message "Please select a valid info object" am loading data using PSA, having activated all external chars still get the problem..some help on this please..
    regards
    Sri

    there is o relation of infoobject name in flat file and infoobjet name at BW side.
    please check with the object in the BW and their lengths and type of the object and check your flat file weather u have the same type there,
    now check the sequence of the objects in the transfer rules  and activate them.
    there u go.

  • How can we search data in flat files

    I need to search for a particular list of account numbers whose data is stored in a flat file. Each line represents data for one account number. I also need to put all data for the searched account number into a new flat file.
    Can anyone help me on this??
    Thanks in advance....

    I m very new to java programming and dont know what
    to write in the code i.e are there any available
    methods to do so..or do we need to write a search
    logic..That's correct you have to implemetn the logic, but it's pretty easy. Just a loop which reads each line from the file, and looks for the information.
    Kaj

  • How can we append data in existing flat file

    how can we append data to a existing flat file (Text file).

    just fill the itab from data which suppose to append to text file.than read text file from GUI_UPLOAD and loop on itab and with the use of read statements you may append the data.

  • Can we load data in variable screen from flat file?

    Hello all,
    One of my user once asked me she needs to run a report for some employees, i said you can just type that in employee varaible input box.
    She said I need to run the report and I need it for 120 employees, she has the employee numbers in excel sheet.
    What is the best way of doing this if any ?
    Thanks

    This can be done from a flat file, which has list of all values.
    If you are using Bex 3.5, then we have an icon in variable screen bottom right to upload values, where we can specify the path of the file.
    Naveen.a

  • HUGE amount of data in flat file every day to external system

    Hello,
    i have to develop several programs to export all tables data in a flat file for external system ( EG. WEB).
    I have some worries like if is possible by SAP export all KNA1 data that contains a lot of data in a flat file using the extraction:
    SELECT * FROM KNA1 ITO TABLE TB_KNA1.
    I need some advices about these kind of huge extractions.
    I also have to extract from some tables, only the data changes, new record, and deleted records; to do this I thought of developing a program that every day extract all data from MARA and save the extraction in a custom table like MARA; the next day when the programs runs compare all data of the new extraction with the old extraction in the ZMARA table to understand the datachanges, new records or deleted record.. It's a righ approach? Can have problems with performance? Do you now other methods?
    Thanks a lot!
    Bye

    you should not have a problem with this simple approach, transferring each row to the output file rather than reading all data into an internal table first:
    open dataset <file> ...
    select * from kna1 into wa_kna1
      transfer wa_kna1 to <file>
    endselect
    close dataset <file>
    Thomas

Maybe you are looking for