Space-delimited fields, convert to table?

Hi,
I have to import some large tables, which I currently have as text files, into a Pages document. The text files have fields delimited by spaces only, and "new line" characters at the end of each line should be interpreted as meaning go to the next row. Is there any chance I can get Pages to recognise this as a table? I tried creating a table with the right number of rows and columns, selecting the whole thing then pasting the contents of my text file, and while the newline characters worked as expected, the spaces did not do anything, and so my data appeared all in the left-most column.
How do I get Pages to interpret the space as delimiting one field from the next?
Thank you!!

I would work backwards.
Create a test table in Pages and copy/paste it into A TextEdit document or export it as a rtf file to see how it is structured, then use that to format the table you want to import.
Do a Search and Replace to insert markers between fields and line ends, then use those to make your table. The search and replace ofcourse has to be on a pattern. I'd say you have a cluster of spaces between your entries, search for those and replace with tabs and get rid of any extra spaces.
*This straight forward method will work*
1. Create a tab delimited set of data with returns at the ends of lines
2. Create a table with equal number of rows and columns
3. Copy all the data text
4. Select all the cells in the table and paste
The delimited text will fall into place in all the cells.
*How to solve problems like this*
I am not omniscient about everything in Pages, but I will experiment and try various options. I have never done this in Pages, but it took me less than a minute to try the above and see that it works.
Peter

Similar Messages

  • Delete L.H.S spaces from field in database table but not for Report

    Dear Experts
    I m facing a problem...actaually i have replaced 999 by blank in a field of database table record.
    Like :     999 abcde
                 999999 hjklmn
                 9999 klplmj
                 99999 hjsfhh
    I have replaced 9 by blank. and using CONDENSE statement i have avoid the left hand side spaces of 9 in the database table.
    Means the data is updated in database table following below using Condense.
    Like:     abcde
                hjklmn
                klplmj
                hjsfhh
    Now,  the problem is that these spaces is showing in the Report. The left hand side spaces is not deleting from the Report. But these spaces is not showing in the databse table.
    Plz tell me wat is the reason and wat is the solution to avoid these space from the Report.
    Note: That field is Character Field.   KNA1-STRAS
    Thanks and Regards,
    Swapnika

    Hello
    Try to use
    SHIFT c LEFT  DELETING LEADING SPACE.
    instead of
    CONDENSE...

  • Export Excel Table in .txt File with space delimited text in UNICODE Format

    Hi all
    I've a big unsolved problem: I would like to convert an Excel table with some laboratory data in it (descriptions as text, numbers, variables with some GREEK LETTERS, ...). The output should be a formatted text with a clear structure. A very good solution is
    given by the converter in Excel "Save As" .prn File. All works fine, the formattation is perfect (it does not matter if some parts are cutted because are too long), but unfortunately the greek letters are converted into "?"!!!
    I've tried to convert my .xlsx File in .txt File with formatting Unicode and the greek letters are still there! But in this case the format is not good, the structure of a table is gone!
    Do you know how to save an Excel file in .prn but with Unicode formatting instead of ANSI or a .txt with space delimited text?
    Thanks a lot to everyone that can help me!
    M.L.C.

    This solution works in Excel/Access 2013.
    Link the Excel table into Access.
    In Access, right-click the linked table in the Navigation Pane, point your mouse cursor to "Export", and then choose "Text File" in the sub-menu.
    Name the file, and then under "Specify export options", check "Export data with formatting and layout".  Click "OK".
    Choose either Unicode or Unicode (UTF-8) encoding.  Click "OK".
    Click "Close" to complete the export operation.

  • Changing the length of a key field in a table

    Hi,
    I want to increase the length of the field from 2 to 4 in a standard SAP table and deliver it to the customers. This field is a key field in table. This field from this table is also used in view and view clusters.
    What is the implication of changing the length to the customers. The customers would have already data in this field and they should not loose any data. Will the existing data for customers remain at length 2 or do they have to do some conversion?
    Regards,
    Srini.
    Edited by: Srinivasa Raghavachar on Feb 7, 2008 12:45 PM

    hi,
    The database table can be adjusted to the changed definition in the ABAP Dictionary in three different
    ways:
    By deleting the database table and creating it again. The table on the database is deleted, the inactive
    table is activated in the ABAP Dictionary, and the table is created again on the database. Data
    existing in the table is lost.
    By changing the database catalog (ALTER TABLE). The definition of the table on the database is
    simply changed. Existing data is retained. However, indexes on the table might have to be built again.
    By converting the table. This is the most time-consuming way to adjust a structure.
    If the table does not contain any data, it is deleted in the database and created again with its new
    structure. If data exists in the table, there is an attempt to adjust the structure with ALTER TABLE. If the
    database system used is not able to do so, the structure is adjusted by converting the table.
    Field 1     Field 2,    Field 3
    NUMC,6  CHAR 8    CHAR, 60
    Field 1    Field 2       Field 3
    NUMC,6 CHAR, 8     CHAR,30
    The following example shows the steps necessary during conversion.
    Starting situation: Table TAB was changed in the ABAP Dictionary. The length of field 3 was reduced
    from 60 to 30 places.
    The ABAP Dictionary therefore has an active (field 3 has a length of 60 places) and an inactive (field 3
    still has 30 places) version of the table.
    The active version of the table was created in the database, which means that field 3 currently has 60
    places in the database. A secondary index with the ID A11, which was also created in the database, is
    defined for the table in the ABAP Dictionary.
    The table already contains data.
    Step 1: The table is locked against further structure changes. If the conversion terminates due to an
    error, the table remains locked. This lock mechanism prevents further structure changes from being
    made before the conversion has been completed correctly. Data could be lost in such a case.
    Step 2: The table in the database is renamed. All the indexes on the table are deleted. The name of the
    new (temporary) table is defined by the prefix QCM and the table name. The name of the temporary
    Step 3: The inactive version of the table is activated in the ABAP Dictionary. The table is created on the
    database with its new structure and with the primary index. The structure of the database table is the
    same as the structure in the ABAP Dictinary after this step. The database table, however, does not
    contain any data.
    The system also tries to set a database lock for the table being converted. If the lock is set, application
    programs cannot write to the table during the conversion.
    The conversion is continued, however, even if the database lock cannot be set. In such a case
    application programs can write to the table. Since in such a case not all of the data might have been
    loaded back into the table, the table data might be inconsistent.
    You should therefore always make sure that no applications access the table being converted
    during the conversion process.
    Step 4: The data is loaded back from the temporary table (QCM table) to the new table (with MOVECORRESPONDING).
    The data exists in the database table and in the temporary table after this step.
    When you reduce the size of fields, for example, the extra places are truncated when you reload the
    data.
    Since the data exists in both the original table and temporary table during the conversion, the storage
    requirements increase during the process. You should therefore verify that sufficient space is available in
    the corresponding tablespace before converting large tables.
    There is a database commit after 16 MB when you copy the data from the QCM table to the original
    table. A conversion process therefore needs 16 MB resources in the rollback segment. The existing
    database lock is released with the Commit and then requested again before the next data area to be
    converted is edited.
    When you reduce the size of keys, only one record can be reloaded if there are several records whose
    key cannot be distinguished. It is not possible to say which record this will be. In such a case you should
    clean up the data of the table before converting.
    Step 5: The secondary indexes defined in the ABAP Dictionary for the table are created again.
    Step 6: The temporary table (QCM table) is deleted.
    Step 7: The lock set at the beginning of the conversion is deleted.
    If the conversion terminates, the table remains locked and a restart log is written.
    Caution: The data of a table is not consistent during conversion. Programs therefore should not access
    the table during conversion. Otherwise a program could for example use incorrect data when reading the
    table since not all the records were copied back from the temporary table. Conversions therefore
    should not run during production! You must at least deactivate all the applications that use tables to
    be converted.
    You must clean up terminated conversions. Programs that access the table might otherwise run
    incorrectly. In this case you must find out why the conversion terminated (for example overflow of the
    corresponding tablespace) and correct it. Then continue the terminated conversion.
    Since the data exists in both the original table and temporary table during conversion, the storage
    requirements increase during conversion. If the tablespace overflows when you reload the data from the
    temporary table, the conversion will terminate. In this case you must extend the tablespace and start the
    conversion in the database utility again.
    If you shorten the key of a table (for example when you remove or shorten the field length of key fields),
    you cannot distinguish between the new keys of existing records of the table. When you reload the data
    from the temporary table, only one of these records can be loaded back into the table. It is not possible
    to say which record this will be. If you want to copy certain records, you have to clean up the table
    before the conversion.
    During a conversion, the data is copied back to the database table from the temporary table with the
    ABAP statement MOVE-CORRESPONDING. Therefore only those type changes that can be executed
    with MOVE-CORRESPONDING are allowed. All other type changes cause the conversion to be
    terminated when the data is loaded back into the original table. In this case you have to recreate the old
    state prior to conversion. Using database tools, you have to delete the table, rename the QCM table to
    its old name, reconstruct the runtime object (in the database utility), set the table structure in the
    Dictionary back to its old state and then activate the table.
    If a conversion terminates, the lock entry for the table set in the first step is retained. The table can no
    longer be edited with the maintenance tools of the ABAP Dictionary (Transaction SE11).
    A terminated conversion can be analyzed with the database utility (Transaction SE14) and then
    resumed. The database utility provides an analysis tool with which you can find the cause of the error
    and the current state of all the tables involved in the conversion.
    You can usually find the precise reason for termination in the object log. If the object log does not
    provide any information about the cause of the error, you have to analyze the syslog or the short dumps.
    If there is a terminated conversion, two options are displayed as pushbuttons in the database utility:
    After correcting the error, you can resume the conversion where it terminated with the Continue
    adjustment option.
    There is also the Unlock table option. This option only deletes the existing lock entry for the table .
    You should never choose Unlock table for a terminated conversion if the data only exists in the
    temporary table, i.e. if the conversion terminated in steps 3 or 4. table for table TAB is therefore QCMTAB.
    Hope this is helpful,Do reward.

  • How to Convert internal table data into text output and send mail in ABAP

    Hi All,
    Good Morning.
    Taking a glance at a code that converts internal table data to an Excel file in ABAP. also checked how to send this excel to mailing list as attachment.
    But thought of doing it without excel.
    I mean, I have an internal table which contains fields of all types (character,integer,date,time). Since it is only around 4 to 5 rows in it (output),why to convert it to excel. not required!!.  Instead I  want to send this output to User's mails as Normal mail body with No attachments.
    Could anybody please suggest me a way as to how to send internal table data as a mail ( not as an excel or PDF etc).
    as of now my findings are, it is quite complex to convert internal table data to email (Text) format. but i believe if there is some way of doing it.
    Best Regards
    Dileep VT

    here's something I have used in the past where we send out information about failed precalculation settings (which are stored in internal table gt_fail)
    notice we use gt_text as "mail body"
    TRY.
    *     -------- create persistent send request ------------------------
           gv_send_request = cl_bcs=>create_persistent( ).
    *     -------- create and set document -------------------------------
    *     create text to be sent
           wa_line = text-001.
           APPEND wa_line TO gt_text.
           CLEAR wa_line.
           APPEND wa_line TO gt_text.
           LOOP AT gt_fail ASSIGNING <fs_fail>.
             MOVE <fs_fail>-retry_count TO gv_count.
             CONCATENATE text-002
                         <fs_fail>-setting_id
                         text-003
                         gv_count
                         INTO wa_line SEPARATED BY space.
             APPEND wa_line TO gt_text.
             CLEAR wa_line.
           ENDLOOP.
           APPEND wa_line TO gt_text.
           wa_line = text-007.
           APPEND wa_line TO gt_text.
    *     create actual document
           gv_document = cl_document_bcs=>create_document(
                           i_type    = 'RAW'
                           i_text    = gt_text
                           i_length  = '12'
                           i_subject = 'Failed Precalculation Settings!' ).
    *     add document to send request
           CALL METHOD gv_send_request->set_document( gv_document ).
    *     --------- set sender -------------------------------------------
           gv_sender = cl_sapuser_bcs=>create( sy-uname ).
           CALL METHOD gv_send_request->set_sender
             EXPORTING
               i_sender = gv_sender.
    *     --------- add recipient (e-mail address) -----------------------
           LOOP AT s_email INTO wa_email.
             MOVE wa_email-low TO gv_email.
             gv_recipient = cl_cam_address_bcs=>create_internet_address(
                                               gv_email ).
             CALL METHOD gv_send_request->add_recipient
               EXPORTING
                 i_recipient = gv_recipient
                 i_express   = 'X'.
           ENDLOOP.
    *     ---------- set to send immediately -----------------------------
           CALL METHOD gv_send_request->set_send_immediately( 'X' ).
    *     ---------- send document ---------------------------------------
           CALL METHOD gv_send_request->send(
             EXPORTING
               i_with_error_screen = 'X'
             RECEIVING
               result              = gv_sent_to_all ).
           IF gv_sent_to_all = 'X'.
             WRITE text-004.
           ENDIF.
           COMMIT WORK.
    *   exception handling
         CATCH cx_bcs INTO gv_bcs_exception.
           WRITE: text-005.
           WRITE: text-006, gv_bcs_exception->error_type.
           EXIT.
       ENDTRY.
    with the following declarations
    * TABLES                                                               *
    TABLES:
       adr6,
       rsr_prec_sett.
    * INTERNAL TABLES & WORK AREAS                                         *
    DATA:
       gt_fail          TYPE SORTED TABLE OF rsr_prec_sett
                             WITH UNIQUE KEY setting_id run_date,
       gt_text          TYPE bcsy_text,
       wa_fail          LIKE LINE OF gt_fail,
       wa_line(90)      TYPE c.
    FIELD-SYMBOLS:
       <fs_fail>        LIKE LINE OF gt_fail.
    * VARIABLES                                                            *
    DATA:
       gv_count(4)      TYPE n,
       gv_send_request  TYPE REF TO cl_bcs,
       gv_document      TYPE REF TO cl_document_bcs,
       gv_sender        TYPE REF TO cl_sapuser_bcs,
       gv_recipient     TYPE REF TO if_recipient_bcs,
       gv_email         TYPE adr6-smtp_addr,
       gv_bcs_exception TYPE REF TO cx_bcs,
       gv_sent_to_all   TYPE os_boolean.
    * SELECTION-SCREEN                                                     *
    SELECT-OPTIONS:
       s_email          FOR adr6-smtp_addr NO INTERVALS MODIF ID sel.
    DATA:
       wa_email         LIKE LINE OF s_email.

  • SQLLDR - tab delimited fields

    Hi:
    I am on 10.2.0.3 on hp-ux B.11.23.
    Need to upload a file with tab-delimited fields. Anybody can tell me what character should be used instead of comma (when the fields are comma-separated) in this statement? TIA.
    load data infile myfile.csv
    replace
    into table PS_MYTABLE
    fields terminated by "," optionally enclosed by '"'

    you can use.. TERMINATED BY WHITESPACE
    FIELDS TERMINATED BY WHITESPACE According to Oracle doc, "If TERMINATED BY WHITESPACE is specified, data is read until the first occurrence of a whitespace character (space, tab, newline). Then the current position is advanced until no more adjacent whitespace characters are found. "..
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/ldr_field_list.htm

  • Activation error while changing the data type of a field of a table

    hi friends,
    i am facing one problem while changing a data type of a field of a table.
    i just created one table(Yqm32) .i have assigned charcter data type to one field(ztotal_count) .now i want to change this charcter data type to numeric data type.
      while changing to NUMC data type activation error is comming as below.
    Table is not yet classified                           
    Field ZTOTAL_COUNT: Type change                       
      ALTER TABLE is not possible                         
    Structure change at field level (convert table YQM32) 
    Check on table YQM32 resulted in errors   
    Table YQM32 could not be activated                       
    (E- Structure change at field level (convert table YQM32)
    plz suggest.i need to change the data type from char to numc.
    Thanks & Regards

    Hi Pabitra......
    From the SE11 change the table as u wanted and then from menubar select
    UTILITIES--> DATABASE UTILITY
    It will open database utility
    there u select the Activate and adjust database button.
    then the database table will get adjusted.
    just try it once.........
    Suresh......

  • Why QUAN_PO_E field in BBP_PDIGP tabl is not reset when PO item is deleted?

    Hi Gurus,
    For some reason, after an item is deleted at PO, its corresponding QUAN_PO_E field in BBP_PDIGP table of SC has the same quantity ( for example, 3 pieces ).
    When the buyer access the SOCO, the item is not displayed because previously the system make a validation: if quantity is equal to quan_po_e ( the difference is zero ), eliminates the item to display and reset the SOURCE_REL_IND to space.
    Somebody knows the reason for which the field QUAN_PO_E is not updated with the value of zero after deleting the items of the purchase order?
    How can I update the QUAN_PO_E field on safe way maintaining the data consistency?
    I have posted new message in order to try separately from thread Re: SOURCE_REL_IND : shopping cart not appearing in the sourcing cockpit

    Hi,
    Could you please check if note  1326380 is implemented ?
    The implementation of this note will not affect the older SC which
    didn't return to sourcing after PO deletion. But the issue should not
    happen for the POs that are deleted after the implementation of the
    note.
    Please, check if field QUAN_PO_E (table BBP_PDIGP) was not cleared after
    PO deletion.
    If not, you can clear this field for the SC item (for which the PO has
    been deleted), then clean_reqreq_up should send the SC to sourcing.
    So a work around would be to manually delete this field, and run
    clean_reqreq_up again.
    Summer

  • How to sum space delimited hex values in a string, and do other stuff?

    Hello
    I am trying to write a Bourne shell script that takes a non-constant-length character string as input, converts each character to hexidecimal value, sums each value, takes the two's complement of the sum, logically ANDS it with FF, and if any of the resultant "nibbles" are a-f, print/save that nibble as A-F, for later display as a two-character-long character string
    For example, assume a variable (as in non-constant) length variable $A
    A="w04:0;2"
    the road that I started down was to define
    B=$( echo $A | od -An -t x1 -N"${#A}" )
    # for this example, $B=77 30 34 3a 30 3b 32
    It's the middle part where I need some magic. What I want to do is add the whitespace-delimited hex values of this string, for this example
    C=77+30+34+3a+30+3b+32
    so for this example, $C=1b2
    then take two's complement of $C; for this example, $TWOS_COMP=ffff ffff ffff ffe4e
    and logically AND it with FF; for this example, almost-final answer is 4e
    then save final result as uppercase character string variable (for this example, variable's contents would be 4E)
    Is what I'm trying to do intuitively obvious to any of you gurus out there, that could rattle off an answer? I'm okay with trashing my approach thus far and trying something totally different, within the constraint that it run inside a Bourne shell script.
    Thanx in advance, if anybody can help!

    I've made progress. I've converted each character of a non-constant-length string $A into a space-delimited string of hex values, added them up, and retained the least significant byte (i.e., the 8 least significant bits). Here's how:
    B=$( echo $A | od -An -t x1 -N"${#A}" )
    D=0
    for C in $B; do
      C=0x$C
      D=$(($D + $C))
    done
    D=$(($D & 16#FF))
    D=`echo "ibase=10;obase=16;$D" | bc`
    The missing piece now is how to take the two's complement of $D. Anybody out there with any ideas on how to handle the two's complement piece inside a Bourne shell?
    Thanx

  • Output required as formatted text (space delimited)

    I need to get the output of a query in a strictly defined space delimited text format. Can anyone advise if that is directly possible in BIP? Currently the method I am using is that I have created an excel template specifying all the formats (column width, number format, etc.), then during runtime, I run the report, open it in excel (with macros disabled), do a "Save As" to save it as a .prn file (Formatted text file), then rename the prn file to .txt
    This works although it is rather cumbersome. Any suggestions for simplifying the process?
    Thanks
    MS

    Thanks Vetsrini, this is precisely what I was looking for. But somehow I am not able to fit things together. I made the rtf format and uploaded into BIP as etext template, but when I view, it doesn't say anything. I saw one of your earlier posts where you had replied to a member's similar query, you had told that there is some problem with the template. Is there any log on the server where the errors are logged?
    And can you give the format for a simple query with just, say, 2 fields - empno and name (6 chars and 25 chars)? It would be great if you give the sample XML and the etext template. The one from the link you sent looks a bit complex.
    Thanking you in anticipation.
    Regards,
    MS

  • SQL Loader, CLOB, delimited fields

    Hello.
    I have to load using SQL Loader data from csv file into table, which one field is CLOB type.
    Here is how ctl file initially looked like:
    UNRECOVERABLE
    LOAD DATA
    INFILE '.\csv_files\TSH_DGRA.csv'
    BADFILE '.\bad_files\TSH_DGRA.bad'
    DISCARDFILE '.\dsc_files\TSH_DGRA.dsc'
    APPEND
    INTO TABLE TSH_DGRA
    FIELDS TERMINATED BY ',' OPTIONALLY ENCLOSED BY '"'
    TRAILING NULLCOLS
    ID_OBJ_TSHD,
    PR_ZOOM_TSHD,
    PR_GRID_TSHD,
    PR_ELMGR_TSHD CHAR(4000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>',
    PR_ALRMGR_TSHD CHAR(4000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>'
    Problems are fields PR_ELMGR_TSHD and PR_ALRMGR_TSHD (CLOBs in table TSH_DGRA). Until data which should be loaded into CLOB fields are under 4000 characters long, it works fine, but what should I do if I want to load data which are longer than 4000 characters?
    If found on Link:[http://download.oracle.com/docs/cd/B14117_01/server.101/b10825/ldr_loading.htm#i1006803] which one sentence said that:
    "SQL*Loader defaults to 255 bytes when moving CLOB data, but a value of up to 2 gigabytes can be specified. For a delimited field, if a length is specified, that length is used as a maximum. If no maximum is specified, it defaults to 255 bytes. For a CHAR field that is delimited and is also greater than 255 bytes, you must specify a maximum length. See CHAR for more information about the CHAR datatype."
    So, my question is, how to specify "up to 2gb" as text said? I can not use CHAR datatype because it is limited to 4000 characters. And I have to load about 60000 characters. I also can not use technique where all data for every CLOB field are in separate files.

    Just specify the maximum expected size:
    PR_ELMGR_TSHD CHAR(100000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>',
    PR_ALRMGR_TSHD CHAR(1000000) OPTIONALLY ENCLOSED BY '<clob>' AND '</clob>'
    The CHAR(1000000) will allow SQLLDR to handle up to 1000000 bytes of input text.

  • HOw to convert Oracle Table data to Excel Sheet

    I need to convert Some of fields from my table to Excell sheet. I dont want to use 3rd party tools. Is there any querry to generate Excell sheet????

    There is no direct path to make Excel file from query, instead you may generate a CSV file (comma-separated-values) or TAB separated values file.
    On SQL*Plus
    SQL> set heading off
    SQL> set newpage none
    SQL> set linesize 255 --could be more
    SQL> spool /path/filename.txt
    SQL> select column1||chr(<ascii for Tab>)||column2||...||columnK
    from table
    SQL> spool off
    Hope it helps

  • Convert the Tables in se14  for cube

    I have the requirement to convert the table for the cube since we have changed the Group and Legal Costs stored in BW as currencies to  6 decimal places.  The two objects  changed to  floating point numbers . Now I am asked to covert the tables for the cube how should i go about this
    first I will go to se14 and give the cube name like /BIC/FZINV_C09 Then what should i do to convert the tables or should i follow some thing else in se14 please clarify me soon. i am asked to do this as per tech document.
    Conversion of all the tables used in Info cubes & ODS objects, in Development Systems
    soniya

    Hi Dinesh and Anil, You both are genius, but please help me what it means by changing the data for legal and group cost directly in data base should i go to se11 and type /bic/fzcobmc01 and select databasetable and now change under "field tab" the data type to floating point and length 16. and also should i change the values by going display table values do i change them manually all 100 values since dinesh pointed the right thing so here is my approach
    soniya
    proceed as below
    1. change the key figures data type and decimal places AND CTIVATE THE KEY FIGURE
    2.

  • Converting Relational Tables to Object Views

    We have extra information stored for each
    field in a table. We call these Vertical
    Tables:
    rowid primary key - ties to normal tab
    column_name "" - ties to column in ""
    text_value the value if text
    reliability rating for level of accuracy...
    date of entry
    Yes there are normal tables that the applications also access if they just need the value of the columns.
    So I want to take this to object views. This way, while oracle is learning how to alter objects, we can still maintain our data and old application while converting to the use of objects.
    I thought the best solution would be Type(s) based on the orignal data type. A generic type, text65_t could be created to hold every column that is less than 65 characters long:
    CREATE TYPE text65_t AS OBJECT
    WITH OBJECT IDENTIFIER(row_id) AS
    row_id VARCHAR2(30),
    value VARCHAR2(65),
    reliability NUMBER(1,0),
    doe DATE
    A view can easily be created of this. The
    mapping is one to one.
    My problem is the next level. How can I build another view on top of text65_t, integer_t, date_t... that ties all of the
    vertical rows back into a normal relational table or object.
    Thanks,
    bill
    null

    Hi migration group,
    I realize this is kind of off topic for
    this discussion group. Where can I ask
    questions like this?
    thanks
    bill

  • Dynamically retreiving fields of a table

    Hi,
    I have a text file in the application server containing data corresponding to the fields of table EPROVAL15.I have declared an internal table and work area of the table type .While I am trying to split the file into respective fields it's throwing an error which is "field should be declared of character type".since fields such as 'VAL0000' or 'VAL0015' are of decimal format,so split statement doesnot work for such data type.Can any1 suggest how to split the file so that data from the text file get appended to the internal table?
    I thought of declaring a structure containing the same number fields as the table EPROFVAL15 with data type as char,but then that would be time consuming process as the table contains many fields.Can any one provide a solution to it?
    I would like to know ,if there is a table containing more than 100 fields and we want to append values from text file to it through an internal table ,then in the split statement do we need to provide every field  name (the usual process that we do) ?i feel it's a time consuming process to mention all the field name manually.is there any way out by which the fields can be called  dynamically for appending the data from the text file into it?

    Hi Brad,
    The text file is delimited by' | '.
    following is the sample code .
    data: gt_IEPROFVAL15  type  IEPROFVAL15,
           gwa_IEPROFVAL15  type  EPROFVAL15.
    split g_temp1 at '|' into gwa_IEPROFVAL15-PROFILE  gwa_IEPROFVAL15-VALUEDAY gwa_IEPROFVAL15-VAL0000
                                      gwa_IEPROFVAL15-VAL0015   gwa_IEPROFVAL15-VAL0030  gwa_IEPROFVAL15-VAL0045.
    The above split statement contains fields up till VAL2345(presently i have shown fields up till VAL0045).The syntactical error is
    ' gwa_IEPROFVAL15-VAL0000 must be a character type data object'.
    Is there any way by which data conversion can take place explictly ?
    If i declare a structure containing the same fields as that of table EPROVAL15(with the data type being char instead of integer) to be used in the split statement  then again it's a tedious process.
    Hope the problem stated is clear to you.

Maybe you are looking for

  • Can someone tell me how to disconnect my credit card from my iTunes account?

    Can someone tell me how to disconnect my credit card from my iTunes account?

  • Docs. for Monitoring Installing Portal Performance steps/scripts etc.......

    Hello can someone please point me to the documentation on how to setup the Portal (10.1.2) for generating and loading performance reports. What files do I need to modify and what scripts to run. I want to generate reports to check the performance of

  • PO Rejection

    Hi All, Is there any Bapi or function module for Purchase Order Rejection. I have created a custom WF template for Purchase Order release. Once the approver cilcks the reject button, then the Purchase Order Status should be changed into Rejected stat

  • In process inspection(QM)

    Hii , In case of Inprocess inspection , we offered particular operation for confirmation to QC. If QC rejects the lot through usuage decision system is allowing confirmation of that operation. I dont want to allow the confirmation of operation if its

  • Starting new task in FM

    Hi, I have FM 'update_table' with asynchrounous call(FM which use starting new task). This FM updates the table and commits the work.But when I try to read the datas from table I don't get the updated datas.This happens only in in real scenario. When