Positional Polling a file with multiple records at single line

Hi
I am trying to poll a file with positional read.It works fine with single record at single line.But how to solve the problem when there are multiple records at single line???How to delimit the records??

i have a file which contains emp records like empID,empName,Dept.And these records of fixed length.I need to poll this file wirh positional read.I used fixed length polling.It works fine when there is only 1 record at one line.But when 2 or more records on same line it takes first record correctly but could not understand the start of second record.
for ex
1 shital mktg
2 priya finance
it works when records are at separate line.
but another case is:
1 shital mktg 2 priya finance
how to delimit these records??

Similar Messages

  • Flat File with multiple record types (OWB 10.2.0.2)

    Hi!
    I`m using OWB 10.2.0.2 and I`m trying to load a flat file with multiple record types, using SQL LOADER.
    In the flat file editor in the Record tab, I`ve set the type values and the corresponding record names like this:
    Type Value Record Name
    ======== ===========
    T TRAILER
    0 DETAILS
    1 DETAILS
    2 DETAILS
    When using this flat file in a mapping to load the data in a staging table, the generated code looks like this:
    INTO TABLE TRAILER
    TRUNCATE
    REENABLE DISABLED_CONSTRAINTS
    WHEN (1:1) = 'T'
    INTO TABLE DETAILS
    APPEND
    REENABLE DISABLED_CONSTRAINTS
    WHEN (1:1) = '0,1,2'
    The above clause (WHEN (1:1) = '0,1,2') is wrong as I expect one "INTO TABLE..." clause for each record type.
    Could this be a bug or am I doing something wrong?
    Thanks a lot for your help,
    Yorgos

    We`re using two target tables, one for the trailer record and the other for the details records.
    We are facing this problem from the moment we upgraded from OWB 10.1 to OWB 10.2.0.2, so we think it must be something with the way the sql loader code is generated in the new version.
    As our data sources are mainly flat files coming from mainframes, this is a huge problem for us. We even asked an expert in DW from Oracle to help us on this, but still haven`t found a solution.
    Is there any workaround for this or should we forget sql loader and go with an external tables + custom PL/SQL code solution?
    Your help is greatly appreciated Jean-Pierre.
    Regards,
    Yorgos

  • Load data from a file with multiple record types to a single table-sqlldr

    We are using two datastores which refer to the same file. The file has 2 types of records header and detail.
    h011234tyre
    d01rey5679jkj5679
    h011235tyrr
    d01rel5678jul5688
    d01reh5698jll5638
    Can someone help in loading these lines from one file with only two data stores(not 2 separate files) using File to Oracle(SQLLDR) Knowledge Module.

    Hi,
    Unfortunately the IKM SQLDR doesn't have the "when" condition to be wrote at ctl file.
    If you wish a simple solution, just add an option (drop me a email if you want a LKM with this)
    The point is:
    With a single option, you will control the when ctl clause and, for instance, can define:
    1) create 2 datastores (1 for each file)
    2) the first position will be a column at each datastore
    3) write the when condition to this first column at the LKM in the interface.
    Does it help you?

  • Loading Hex Files with multiple records with each line starting with : record marker, how do I load this hex file into a front panel table ? thanks in advance, Jeff.

    I have written a routine that loads a Hex file, checks for record marker,length,address etc then loads the data into a table on front panel. However, my routine only loads single record hex files. I need to be able to load multiple record files, I need to identify the ':' colon record marker start for each record in the hex file, I just can't see how to loop my routine so that all records with ':' prefix marker are loaded.

    Jeff,
    I'm not exactly sure of what you are trying to do. Is it possible you could post your code, or a screen shot of the diagram in question, so that I can get a better understanding, and possibly offer some assistance?
    I am familiar with hex files, but not with the : marker separating records. This may be able to be solved with simple text parsing, and as I am not familiar with how you are retrieving and/or displaying the files, I require more information.
    Thanks

  • Get MUTLIPLE output files for SINGLE input file with multiple records

    Hi ,
    I have source xml file  which has multiple records in it and  I want to get multiple seperate  flat files for it.
    this is my source structure
    <root>
    <Header1 fld1=1234 fld2="name1" fld3='SAP1"
    fld4='000" fld5=0 fld6="sdn1"></header>...............no line items here
    <Header2 fld1=123 fld2="name" fld3='SAP"
    fld4='00" fld5=0 fld6="sdn" ><line fld7=value1 fld8=value2 fld9=value3 /line><line fld7=value4 fld8=value5 fld9=value6 /line></header>.....two line items here
    <Header3 fld1=123 fld2="name" fld3='SAP"
    fld4='00" fld5=0 fld6="sdn" ><line fld7=value7 fld8=value8 fld9=value9 /line></header>......one line item here
    Header4 fld1=12345 fld2="name2" fld3='SAP2"
    fld4='0006" fld5=01 fld6="sdn2"</header>.......no line items here
    </root>
    I am trying to get flat files as many time the <b>header</b> values occur that many target files should appear in the target.
    like
    file1:
    1234name1SAP10000~sdn1
    file2:
    123nameSAp000~sdn
    value1value2valu3
    valu4value5value6
    file3:
    123namesap000~sdn
    value7valu8value9
    file4:
    12345name2SAP2000601~SDN2
    here I have 4 headers so i need to get 4 output files ,similarly for 'n' headers I need to get 'n'  output flat files. the above mentioned source xml file is just one file,like that I might get somany sourc xml files which has similar structure in it.
    and also it is important to get the line items,if exists to the header, for the corresponding headers.
    how can I achieve this?do I need to use BPM or any mappings other than Graphical? 
    please suggest me.
    thank you.
    Babu
    sorry,output file structure changed
            Babu

    Babu,
    See the Mapping and the logic.
    Make sure your target message occurrence:http://www.flickr.com/photo_zoom.gne?id=1412748793&size=o
    Check out for source and target structure
    http://www.flickr.com/photo_zoom.gne?id=1412748803&size=o
    http://www.flickr.com/photo_zoom.gne?id=1412748809&size=o
    http://www.flickr.com/photo_zoom.gne?id=1412748815&size=o
    http://www.flickr.com/photo_zoom.gne?id=1412748825&size=o
    Results:http://www.flickr.com/photo_zoom.gne?id=1412748799&size=o
    Note:
    While creating Interface determination choose it as Enhanced/Extended
    The Receiver Content conversion - For field separator and line separator give 'nl' , coz we concatenated all of them using ~ in mapping.
    If you find any discrepancy in the above logic , please let me know.
    raj.

  • VSAM file with multiple record layouts (COBOL REDEFINES)

    Hi. We're using Tuxedo 12c to migrate a COBOL/CICS/VSAM application from MVS to Solaris/Oracle 11gR2.
    We have a few VSAM files with record layouts that start with a record-type field, and based on that, the remainder of the record layout varies based on COBOL copybooks using REDEFINES. We can't be the first to have this issue, and am wondering if there's a best-practice solution. Our initial thought is that each record type should become a separate database table, but if we need to do this manually, we'll also need to modify the post-converted code manually. I'm hoping there's a better way.
    Thanks.

    You are correct; you are not the first person to do this. There are a number of different strategies depending upon what you are trying to accomplish. First, is your objective to leave the COBOL/CICS/VSAM as is? If so, which COBOL compiler do you intend to use? Are you using Tuxedo ART Workbench? Are you aware that you have the option of leaving the VSAM datasets as VSAM on the target platform? If you want to convert VSAM to Oracle DB tables, you can create a separate table for each record type as you suggested or you can treat the entire record as a blob and let the copyfile sort it out once the data is returned to the program. Of course with this approach you will not realize the benefits Oracle DB as an RDBMS. Please provide me some additional context and I will make some best practices suggestions.
    You can get some insight on Tuxedo ART Workbench at http://docs.oracle.com/cd/E18050_01/artwb/docs11gr1/wbuser/FileToOracle.html

  • Splitting of a CSV File with Multiple Records into Multiple XML File

    Dear All,
    <b> I am doing a Scenario of CSV to XML Files. I am using BPM for the same. My incoming CSV File has got multiple records. I want to break this Multiple records into Multiple XML Files having one record each.</b>
    Can someone suggest how can I break this rather Split this into Multiple XML Files.
    Is Multimapping absoltely necesaary for this. Can't we do this without Multimapping. Can we have some workaround in the FCC parameters that we use in the Integration Directory.
    Kindly reply ASAP. Thanks a lot to all in anticipation.
    Pls Help.
    Best Regards
    Chakra and Somnath

    Dear All,
    I am trying to do the Multimapping, and have 0....unbounded also. Someways it is not working.
    <b>
    Smitha please tell me one thing...Assigning the Recordsets per Message to 1, does it mean that it will write multiple XML Files as I want.</b>
    Also I am usinf Set to Read only. So once the File is read it becomes RA from A. Then will it write the other Records.
    I have to use a BPM because there are certain dependencies that are there for the entire Process Flow. I cannot do without a BPM.
    Awaiting a reply. Thanks a lot in anticipation.
    Best Regards
    Chakra and Somnath

  • How to populate a form from an xml file with multiple records

    I want to be able to create a single form that will add pages as needed based upon the number of records being imported from an xml file.
    I have an IRS form that does this exactly, but I do not know how to re-create it.  I entered data into the form, then exported that data to xml.  Then I opened that xml file in excel and added more records (this is just an easy way to manage the xml data in a tabular form).  Then I import the changed xml file into the IRS form and it adds as many pages as I have records.  Perfect!  Unfortunately, I cannot edit the form to examine the technique.  It does not appear to be relying upon any additional local javascript files.  It does use an 'add page' button which may suggest the form is using the template .spawn() technique, but when I try that in my own form, it gives me unique field names for each page which are then not consistent with the xml data.  Hopefully someone can help me understand what I am missing.

    did you want
    <record>0
    <data>"11.97.23.174/32"
    <record>address <data>
    <record>address <data>
    <record>subnet <data>
    <record>ip <data>"10.97.23.174"</data></record>
    </data></record>
    </data></record>
    </data></record>
    </data></record>
    or
    <record>0
    <data>"11.97.23.174/32"
    <record>address <data>
    <record>address <data>
    <record>subnet <data>
    <record>ip <data>"10.97.23.174"</data></record>

  • XML file with multiple records

    I have an XML data file that will look something like this (repeated possibly several hundred times):
    <?xml version='1.0' encoding='utf-8'?>
    <A><V>Wake up to WonderWidgets!</V>
    <C>Why are great</C>
    </A>
    <?xml version='1.0' encoding='utf-8'?>
    <A>Eric</A>
    When I try to parse the code using the Echo03.java sample from the website, I get the following error: END_ELM: </>org.xml.sax.SAXParseException: XML declaration may only begin entities.
         at org.apache.crimson.parser.Parser2.fatal(Parser2.java:3339)
         at org.apache.crimson.parser.Parser2.fatal(Parser2.java:3327)
         at org.apache.crimson.parser.Parser2.maybePI(Parser2.java:1107)
         at org.apache.crimson.parser.Parser2.maybeMisc(Parser2.java:1233)
         at org.apache.crimson.parser.Parser2.parseInternal(Parser2.java:638)
         at org.apache.crimson.parser.Parser2.parse(Parser2.java:333)
         at org.apache.crimson.parser.XMLReaderImpl.parse(XMLReaderImpl.java:448)
         at javax.xml.parsers.SAXParser.parse(SAXParser.java:345)
         at javax.xml.parsers.SAXParser.parse(SAXParser.java:281)
         at Echo03.main(Echo03.java:60)
    Is there any way (short of actually reading the file and looking for the beginning of each record) that I can get the parser to do multiple stard and end documents or at least get it to read and parse all the entries in the file?
    Thanks

    Yout file should look something like this: <
    ?xml version='1.0' encoding='utf-8'?>
    <start>
    <A><V>Wake up to WonderWidgets!</V>
    <C>Why are great</C>
    </A>
    <A>Eric</A>
    </start>
    you can only have one root element ( e.g. <start> )

  • Idoc outbound to file with multiple records

    Hi Friends,
    I want to export multiple materials into one flat file per ale (idoc).
    I know I can create îdocs per "bd10" and download into file per WE19,
    but only several files not many idocs in one file.
    Can somebody help me with this requirement.
    Actually I want to extract material data per Idoc and import it later per LSMW.
    Many Thanks.
    Jimbob

    Your file port defintion (WE21) specifies the function module for creating the pathname (path + filename) for the IDOC file; most likely , it's EDI_PATH_CREATE_CLIENT_DOCNUM.  You can replace this with your own Z-function or one of the other EDI_PATH_CREATE_ functions that will use the same path name over and over.  You would not need to change the write mode, since it's already append mode by default in the system.  Of course, you are correct, there are many other approaches available to get the data moved...

  • Sharing files with multiple users on single iMac

    THIS IS a REPOST from another location - more appropriate "thread"
    After switching to Mac from a PC about a month ago (& not being technologically over-adept), I have finally succeeded in creating separate "User" accounts on my iMac (1 admin, 1 standard, 1 managed) to allow the 3 users in my household to create their own workspace. However, now I would like to allow them to access all files from some applications - iTunes, iPhoto, Address Book, etc. - and some of the files within Numbers, Pages, & Keynote (that were theirs to start with...but they were working on in "my" admin space before establishing separate Users.) As an additional challenge, I'd still like to have access to their files...for monitoring purposes...as we go forward. I've "searched" the Apple support files and found reference to some of this but (1) I can't seem to figure out how to move the files into the Public Folder (I did one, then logged in as another User & it wasn't there...) and (2) If they access files from the Public Folder, will they need to specifically save them back there when done? -- my other users will NEVER remember to put something back where they got it...

    You can give each user read and write permissions to just that folder, or to everything in it. But that will not stick to any new files that are added to the folder hierarchy after that point. When new files are created they are read and write for the creator, and read only for all others. If you want to change that it must be done manually for each new file that you add.
    There is a way to get around this, and make every file/folder added to a folder hierarchy automatically get full read and write permission for the users you specify. It requires setting up ACLs. Do it this way:
    1. Make a new subfolder in /Users/Shared. Call it something like /Users/Shared/sharefolder. Move all the files you want to share into this folder.
    2. Go to System Preferences and unlock the Accounts pref pane. Click the + button. Make a new group (select "group" from the drop down menu). Call it something like "sharegroup". Add the users you want to that group.
    3. Log in to an admin account. Launch Terminal. Paste in all three of these lines at the same time and then press Return:
    sudo chmod -R +a "sharegroup allow delete,chown,list,search,add_file,\
    addsubdirectory,delete_child,file_inherit,directoryinherit" \
    /Users/Shared/sharefolder
    Enter your admin password when prompted and press return again. You are done. The users in the sharegroup can all now read and write everything in the sharefolder. They will also be able to read and write any file that is newly created or copied into the sharefolder from that point on.
    NOTE: After setting this up, and you want to add more files to the sharefolder, be aware that files that already exist outside the sharefolder WON'T inherit the new permissions if they are simply moved in (i.e. drag with Finder). They must be copied in. Hold down the option key when dragging files into the sharefolder. That will ensure that a new copy will be made, and the permissions will inherit properly. You can then delete the old copy of the file from its previous location.
    I hope this helps.

  • File to RFC with multiple records using BPM Scenario Error...!!!

    Hello Guru's,
      I have done the File to RFC with multiple records using BPM scenario as per the Materiel available in the sdn.sap. This involves BAPI (BAPI_MATERIEL_AVAILABILITY). I have done exactly the same what is their in the materiel. SXI_CACHE is also giving return value " 0 ". File is getting deleted from the source directory, but no file in target directory. SXMB_MONI is also showing no error (black Flag). BPM is also error free. Checked the interfaces also.
    Can any one tell me what mistake would i have done.
    Thanks in advance.

    Hi,
    There is one similar discussion I found,
    FTP TO RFC using BPM
    Thanks
    Swarup

  • 1 xml file to multiple xml files with 200 records in each

    Hello Experts,
    I have below scenario.
    Please suggest me which might be the best approch to work on.
    1) XML file to XML file
    I will have 1 pretty huge XML file which i need to break in multiple XML files with 200 records each.
      1) first approch i can have bpm in which i can split the file according to my requirement.
      2) Second approch i can create 2 scenarios in which 1st scenario will pick up XML file and create multiple flat files with File content conversion. Second scenario will pick up all these flat files and create XML files.
    2) XML file to XML file
    Or i can have multiple files with 1 record per file and i need to merge in multiple XML files with 200 records in each.
    So its kind of 1:N or M:N scenarios.
    Please tell me which is might be better performance and design wise.
    Or if you have any idea in any other way i can do this, then please reply as soon as possbile.
    Please tell me if you have OS command for the same or some script to run or anything which i can implement.
    Thanks,
    Hetal

    what is your senario? is it File to File?
    You can use multi mapping concept without BPM. You can handle the 200 records per message logic in the multimapping.
    Regards,
    Praveen Gujjeti.

  • How to load unicode data files with fixed records lengths?

    Hi!
    To load unicode data files with fixed records lengths (in terms of charachters and not of bytes!) using SQL*Loader manually, I found two ways:
    Alternative 1: one record per row
    SQL*Loader control file example (without POSITION, since POSITION always refers to bytes!)<br>
    LOAD DATA
    CHARACTERSET UTF8
    LENGTH SEMANTICS CHAR
    INFILE unicode.dat
    INTO TABLE STG_UNICODE
    TRUNCATE
    A CHAR(2) ,
    B CHAR(6) ,
    C CHAR(2) ,
    D CHAR(1) ,
    E CHAR(4)
    ) Datafile:
    001111112234444
    01NormalDExZWEI
    02ÄÜÖßêÊûÛxöööö
    03ÄÜÖßêÊûÛxöööö
    04üüüüüüÖÄxµôÔµ Alternative2: variable length records
    LOAD DATA
    CHARACTERSET UTF8
    LENGTH SEMANTICS CHAR
    INFILE unicode_var.dat "VAR 4"
    INTO TABLE STG_UNICODE
    TRUNCATE
    A CHAR(2) ,
    B CHAR(6) ,
    C CHAR(2) ,
    D CHAR(1) ,
    E CHAR(4)
    ) Datafile:
    001501NormalDExZWEI002702ÄÜÖßêÊûÛxöööö002604üuüüüüÖÄxµôÔµ Problems
    Implementing these two alternatives in OWB, I encounter the following problems:
    * How to specify LENGTH SEMANTICS CHAR?
    * How to suppress the POSITION definition?
    * How to define a flat file with variable length and how to specify the number of bytes containing the length definition?
    Or is there another way that can be implemented using OWB?
    Any help is appreciated!
    Thanks,
    Carsten.

    Hi Carsten
    If you need to support the LENGTH SEMANTICS CHAR clause in an external table then one option is to use the unbound external table and capture the access parameters manually. To create an unbound external table you can skip the selection of a base file in the external table wizard. Then when the external table is edited you will get an Access Parameters tab where you can define the parameters. In 11gR2 the File to Oracle external table can also add this clause via an option.
    Cheers
    David

  • Reading fixed length file with different record types

    Hi,
    I need to read a fixed-length file with different record types, but the record identifier is in 31st position and not in 1st position.
    But if I give 31 as position in File adpater wizard, BPEL takes whole 1-31 as identifier.
    How we need to read such files.
    Thanks
    Ravdeep

    hi ,
    u cannot use the default wzard for this
    use some thing like this nxsd:lookAhead="30" nxsd:lookFor="S"have a look at the below link it has some examples
    http://download.oracle.com/docs/cd/B31017_01/integrate.1013/b28994/nfb.htm

Maybe you are looking for