ASCII and UNICODE interoperability

Hi forum,
Do you know if there are problems in a scenario with a ASCII R/3 system as backend for a UNICODE SCM (4.1 or 5.0) ?
Regards.
Ganimede Dignan.

It ends.

Similar Messages

  • ASCII and Unicode problem...

    Would you happen to know if InDesign CS3 still recognizes ASCII characters in an Applescript?
    I did some tweaking of a search & replace script (formerly used in Quark 6.5), to apply in InDesign, but it doesn’t seem to work.
    I read online that when Leopard came out, ASCII became obsolete but InDesign CS3 would be fully compatible with Unicode. So I entered Unicodes in the script and that didn’t work either.
    I also tried saving the script as “text” instead of “script” since some of the default Applescripts in my InDesign folder are “text”. But that didn’t help either.
    Script permissions are set to read/write/execute, so that can’t be the issue.
    Any ideas or suggestions?
    (attached is a portion of the search and replace Applescript, and sample InDesign file I’ve been running the script on)

    InDesign CS3 will accept Unicode and MacRoman text. if you want to enter Unicode characters directly in a script, you need to be using Applescript 2.0, which comes with OS X 10.5 -- before that you generally have to resort to hacky workarounds.
    Shane Stanley <[email protected]>
    AppleScript Pro Sessions <http://scriptingmatters.com/aspro>

  • ASCII and UNICODE

    Hi forum,
    Do you know if there are problems in a scenario with a ASCII R/3 (4.6C) system as backend for a UNICODE SCM (5.0) ?
    Regards.
    Ganimede Dignan.

    for sure - as everything is latin-1 (1100) only, this works )
    Regards
    Volker Gueldenpfennig, consolut.gmbh
    http://www.consolut.de - http://www.4soi.de - http://www.easymarketplace.de

  • What is the diffrence between ASCII and BIN mode

    Hello All,
    What is the diffrence between ASCII and BIN mode
    Regards,
    Lisa.

    'ASC' :
    ASCII format. The table is transferred as text. The conversion exits are
    carried out. The output format additionally depends on the parameters
    CODEPAGE, TRUNC_TRAILING_BLANKS, and TRUNC_TRAILING_BLANKS_EOL.
    'IBM' :
    ASCII format with IBM codepage conversion (DOS). This format correspond
    to the 'ASC' format when using target codepage 1103. This codepage is
    often used for data exchange by disc.
    'DAT' :
    Column-by-column transfer. With this format, the data is transferred as
    with ASC text. However, no conversion exists are carried out and the
    columns are separated by tab characters. This format creates files that
    can be uploaded again with gui_upload or ws_upload.
    'DBF' :
    The data is downloaded in dBase format. Because in this format the file
    types of the individual columns are included, import problems, for
    example, into Microsoft Excel can be avoided, especially when
    interpreting numeric values.
    'WK1' :
    The data is downloaded in Lotus 1-2-3 format.
    'BIN' :
    Binary format. The data is transferred in binary format. There is no
    formatting and no codepage conversion. The data is interpreted row by
    row and not formatted in columns. Specify the length of the data in
    parameter BIN_FILESIZE. The table should consist of a column of type X,
    because especially in Unicode systems the conversion of structured data
    into binary data leads to errors.

  • How to migrate from ascii to unicode (MaxDB 7.5)? loadercli: ERR -25347

    Hi,
    I use MaxDB 7.5.00.26. (Ok, I know that I should switch to 7.6, however, it is not possilble for some customer restriction for now, but should be possible quite soon).
    We'd like to migrate a db from ascii to unicode. Based on the infos in the thread "Error at copying database using dumps via loadercli: error -25364" I tried the following:
    Export sourcedb
    1. Export catalog and data
    C:\> loadercli -d db_asc -u dba,dba
    loadercli> export db catalog outstream file 'C:\tmp1\20080702a_dbAsc.catalog' ddl
    OK
    loadercli> export db data outstream file 'C:\tmp1\20080702b_dbAsc.data' pages
    OK
    loadercli> exit
    Import targetdb
    1. Create a new empty DB with '_UNICODE=yes'
    2. Set 'columncompression' to 'no'
    C:\> dbmcli -d db_uni -u dba,dba param_directput columncompression no
    ERR
    -24979,ERR_XPNOTFOUND: parameter not found
    Couldn't find this parameter e.g. in dbmgui (parameters general, extended and support)
    3. Import catalog and data
    C:\> loadercli -d db_uni -u dba,dba
    loadercli> import db catalog instream file 'C:\tmp1\20080702a_dbAsc.catalog' ddl
    OK
    loadercli> import db data instream file 'C:\tmp1\20080702b_dbAsc.data' pages
    ERR -25347
    Encoding type of source and target database do not match: source = ASCII, target
    = UNICODE.
    loadercli> exit
    What is wrong? Is a migration from ascii to unicode to be done somehow else?
    Can I migrate a db from 7.5.00.26 to 7.6.03.15 in the same way or should it be done in another way.
    It would be greate if you point me to a post etc. where these two migrations are explained in detail.
    Thanks in advance - kind regards
    Michael

    Hi,
    I can neither find "USEUNICODECOLUMNCOMPRESSION" nor "COLUMNCOMPRESSION". Could it be that there do exist from MaxDB version 7.6 on and not in 7.5?
    Kind regards,
    Michael
    The complete parameter list (created by "dbmcli -d db_uni -u dbm,dbm param_directgetall > maxdb_params.txt") is:
    OK
    KERNELVERSION                         KERNEL    7.5.0    BUILD 026-123-094-430
    INSTANCE_TYPE                         OLTP
    MCOD                                  NO
    RESTART_SHUTDOWN                      MANUAL
    _SERVERDB_FOR_SAP                     YES
    _UNICODE                              YES
    DEFAULT_CODE                          ASCII
    DATE_TIME_FORMAT                      INTERNAL
    CONTROLUSERID                         DBM
    CONTROLPASSWORD                       
    MAXLOGVOLUMES                         2
    MAXDATAVOLUMES                        11
    LOG_VOLUME_NAME_001                   LOG_001
    LOG_VOLUME_TYPE_001                   F
    LOG_VOLUME_SIZE_001                   131072
    DATA_VOLUME_NAME_0001                 DAT_0001
    DATA_VOLUME_TYPE_0001                 F
    DATA_VOLUME_SIZE_0001                 262144
    DATA_VOLUME_MODE_0001                 NORMAL
    DATA_VOLUME_GROUPS                    1
    LOG_BACKUP_TO_PIPE                    NO
    MAXBACKUPDEVS                         2
    BACKUP_BLOCK_CNT                      8
    LOG_MIRRORED                          NO
    MAXVOLUMES                            14
    _MULT_IO_BLOCK_CNT                    4
    _DELAY_LOGWRITER                      0
    LOG_IO_QUEUE                          50
    _RESTART_TIME                         600
    MAXCPU                                1
    MAXUSERTASKS                          50
    _TRANS_RGNS                           8
    _TAB_RGNS                             8
    _OMS_REGIONS                          0
    _OMS_RGNS                             25
    OMS_HEAP_LIMIT                        0
    OMS_HEAP_COUNT                        1
    OMS_HEAP_BLOCKSIZE                    10000
    OMS_HEAP_THRESHOLD                    100
    OMS_VERS_THRESHOLD                    2097152
    HEAP_CHECK_LEVEL                      0
    _ROW_RGNS                             8
    _MIN_SERVER_DESC                      16
    MAXSERVERTASKS                        21
    _MAXTRANS                             292
    MAXLOCKS                              2920
    _LOCK_SUPPLY_BLOCK                    100
    DEADLOCK_DETECTION                    4
    SESSION_TIMEOUT                       900
    OMS_STREAM_TIMEOUT                    30
    REQUEST_TIMEOUT                       5000
    _USE_ASYNC_IO                         YES
    _IOPROCS_PER_DEV                      1
    _IOPROCS_FOR_PRIO                     1
    _USE_IOPROCS_ONLY                     NO
    _IOPROCS_SWITCH                       2
    LRU_FOR_SCAN                          NO
    _PAGE_SIZE                            8192
    _PACKET_SIZE                          36864
    _MINREPLY_SIZE                        4096
    _MBLOCK_DATA_SIZE                     32768
    _MBLOCK_QUAL_SIZE                     16384
    _MBLOCK_STACK_SIZE                    16384
    _MBLOCK_STRAT_SIZE                    8192
    _WORKSTACK_SIZE                       8192
    _WORKDATA_SIZE                        8192
    _CAT_CACHE_MINSIZE                    262144
    CAT_CACHE_SUPPLY                      3264
    INIT_ALLOCATORSIZE                    221184
    ALLOW_MULTIPLE_SERVERTASK_UKTS        NO
    _TASKCLUSTER_01                       tw;al;ut;2000*sv,100*bup;10*ev,10*gc;
    _TASKCLUSTER_02                       ti,100*dw;30000*us;
    _TASKCLUSTER_03                       compress
    _MP_RGN_QUEUE                         YES
    _MP_RGN_DIRTY_READ                    NO
    _MP_RGN_BUSY_WAIT                     NO
    _MP_DISP_LOOPS                        1
    _MP_DISP_PRIO                         NO
    XP_MP_RGN_LOOP                        0
    MP_RGN_LOOP                           0
    _MP_RGN_PRIO                          NO
    MAXRGN_REQUEST                        300
    _PRIO_BASE_U2U                        100
    _PRIO_BASE_IOC                        80
    _PRIO_BASE_RAV                        80
    _PRIO_BASE_REX                        40
    _PRIO_BASE_COM                        10
    _PRIO_FACTOR                          80
    _DELAY_COMMIT                         NO
    _SVP_1_CONV_FLUSH                     NO
    _MAXGARBAGE_COLL                      0
    _MAXTASK_STACK                        1024
    MAX_SERVERTASK_STACK                  100
    MAX_SPECIALTASK_STACK                 100
    _DW_IO_AREA_SIZE                      50
    _DW_IO_AREA_FLUSH                     50
    FBM_VOLUME_COMPRESSION                50
    FBM_VOLUME_BALANCE                    10
    _FBM_LOW_IO_RATE                      10
    CACHE_SIZE                            10000
    _DW_LRU_TAIL_FLUSH                    25
    XP_DATA_CACHE_RGNS                    0
    _DATA_CACHE_RGNS                      8
    XP_CONVERTER_REGIONS                  0
    CONVERTER_REGIONS                     8
    XP_MAXPAGER                           0
    MAXPAGER                              11
    SEQUENCE_CACHE                        1
    _IDXFILE_LIST_SIZE                    2048
    _SERVER_DESC_CACHE                    74
    _SERVER_CMD_CACHE                     22
    VOLUMENO_BIT_COUNT                    8
    OPTIM_MAX_MERGE                       500
    OPTIM_INV_ONLY                        YES
    OPTIM_CACHE                           NO
    OPTIM_JOIN_FETCH                      0
    JOIN_SEARCH_LEVEL                     0
    JOIN_MAXTAB_LEVEL4                    16
    JOIN_MAXTAB_LEVEL9                    5
    _READAHEAD_BLOBS                      25
    RUNDIRECTORY                          E:\_mp\u_v_dbs\EVERW_T3
    _KERNELDIAGFILE                       knldiag
    KERNELDIAGSIZE                        800
    _EVENTFILE                            knldiag.evt
    _EVENTSIZE                            0
    _MAXEVENTTASKS                        1
    _MAXEVENTS                            100
    _KERNELTRACEFILE                      knltrace
    TRACE_PAGES_TI                        2
    TRACE_PAGES_GC                        0
    TRACE_PAGES_LW                        5
    TRACE_PAGES_PG                        3
    TRACE_PAGES_US                        10
    TRACE_PAGES_UT                        5
    TRACE_PAGES_SV                        5
    TRACE_PAGES_EV                        2
    TRACE_PAGES_BUP                       0
    KERNELTRACESIZE                       653
    EXTERNAL_DUMP_REQUEST                 NO
    _AK_DUMP_ALLOWED                      YES
    _KERNELDUMPFILE                       knldump
    _RTEDUMPFILE                          rtedump
    _UTILITY_PROTFILE                     dbm.utl
    UTILITY_PROTSIZE                      100
    _BACKUP_HISTFILE                      dbm.knl
    _BACKUP_MED_DEF                       dbm.mdf
    _MAX_MESSAGE_FILES                    0
    _EVENT_ALIVE_CYCLE                    0
    _SHAREDDYNDATA                        10280
    _SHAREDDYNPOOL                        3658
    USE_MEM_ENHANCE                       NO
    MEM_ENHANCE_LIMIT                     0
    __PARAM_CHANGED___                    0
    __PARAM_VERIFIED__                    2008-07-02 21:10:19
    DIAG_HISTORY_NUM                      2
    DIAG_HISTORY_PATH                     E:\_mp\u_v_dbs\EVERW_T3\DIAGHISTORY
    _DIAG_SEM                             1
    SHOW_MAX_STACK_USE                    NO
    LOG_SEGMENT_SIZE                      43690
    SUPPRESS_CORE                         YES
    FORMATTING_MODE                       PARALLEL
    FORMAT_DATAVOLUME                     YES
    HIRES_TIMER_TYPE                      CPU
    LOAD_BALANCING_CHK                    0
    LOAD_BALANCING_DIF                    10
    LOAD_BALANCING_EQ                     5
    HS_STORAGE_DLL                        libhsscopy
    HS_SYNC_INTERVAL                      50
    USE_OPEN_DIRECT                       NO
    SYMBOL_DEMANGLING                     NO
    EXPAND_COM_TRACE                      NO
    OPTIMIZE_OPERATOR_JOIN_COSTFUNC       YES
    OPTIMIZE_JOIN_PARALLEL_SERVERS        0
    OPTIMIZE_JOIN_OPERATOR_SORT           YES
    OPTIMIZE_JOIN_OUTER                   YES
    JOIN_OPERATOR_IMPLEMENTATION          YES
    JOIN_TABLEBUFFER                      128
    OPTIMIZE_FETCH_REVERSE                YES
    SET_VOLUME_LOCK                       YES
    SHAREDSQL                             NO
    SHAREDSQL_EXPECTEDSTATEMENTCOUNT      1500
    SHAREDSQL_COMMANDCACHESIZE            32768
    MEMORY_ALLOCATION_LIMIT               0
    USE_SYSTEM_PAGE_CACHE                 YES
    USE_COROUTINES                        YES
    MIN_RETENTION_TIME                    60
    MAX_RETENTION_TIME                    480
    MAX_SINGLE_HASHTABLE_SIZE             512
    MAX_HASHTABLE_MEMORY                  5120
    HASHED_RESULTSET                      NO
    HASHED_RESULTSET_CACHESIZE            262144
    AUTO_RECREATE_BAD_INDEXES             NO
    LOCAL_REDO_LOG_BUFFER_SIZE            0
    FORBID_LOAD_BALANCING                 NO

  • Unicode in UNICHAR and UNICODE Excel Functions Is Decimal

    CASE # 12 64 08 79 83
    Dear Microsoft Engineers,
    There is a lack of information in  Microsoft Support pages*, where UNICODE and  UNICHAR in Excel 2013 are out of real context, given that unicode is always hexadecimal.
    Please add a notice on these pages about the fact that these Excel functions process decimal unicodes.  You might also add a hint to use HEXDEC and DECHEX conversion when working with UNICHAR and UNICODE.
    Please note that low decimal unicodes equal ASCII.  To highlight the new performance, you'd better take some other examples.
    I need UNICHAR in my Excel developement tool related to a Unicode-using application (MSKLC).
    I contacted Microsoft Support by phone today and got the above case ID.
    Best regards,
    Marcel Schneider
    P.S.:  My first post noted these functions to be ASCII.  So I am sorry not to have considered the system (whether it is hexadecimal, or decimal!).
    * Note:  I was not allowed to post this with hyperlinks to the Support Pages.  Links are the following:
       http://office.microsoft.com/en-us/excel-help/unicode-function-HA102753274.aspx
       http://office.microsoft.com/en-gb/excel-help/unichar-function-HA102753273.aspx

    Hi Marcel,
    Thanks for your feedback, I'll collect the information and submit it with internal ways.
    Have a good time.
    Regards,
    George Zhao
    TechNet Community Support
    It's recommended to download and install
    Configuration Analyzer Tool (OffCAT), which is developed by Microsoft Support teams. Once the tool is installed, you can run it at any time to scan for hundreds of known issues in Office
    programs.

  • What is diff in Open SQl and Unicode

    Hi.
    What is difference in Open SQl and Unicode.
    What are advantages of Unicode ?

    hi osk,
    <b>u cant compare open sql with unicode as both are different..</b>
    just a small explanation..
    <b>Open SQL</b> consists of a set of ABAP statements that perform operations on the central database in the R/3 System. The results of the operations and any error messages are independent of the database system in use. Open SQL thus provides a uniform syntax and semantics for all of the database systems supported by SAP. ABAP programs that only use Open SQL statements will work in any R/3 System, regardless of the database system in use. Open SQL statements can only work with database tables that have been created in the ABAP Dictionary
    <b>unicode</b>
    Data types such as CHAR ASCII and CHAR EBCDIC are mainly suited to English and central European languages. With other character sets, a code attribute is usually used for these data types. This code attribute uses a different presentation code to ASCII and EBCDIC, even for internal storage in the database system. This causes problems if you want to access these database systems using a different character set, or if you want to exchange data between database systems with different character sets.
    You can avoid these problems by using internal character coding in accordance with UNICODE. Internally, the UNICODE data is stored in UTF-16/UCS-2 format. In UTF-16/UCS-2 format, all characters are two bytes long.
    SAP DB is able to display various presentation codes in UNICODE format
    <b>please close the thread after rewarding the appropriate points...</b>
    Message was edited by: Ashok Kumar Prithiviraj

  • How to tell if a kernel is ASCII or Unicode?

    I know this is a simple question but I don't know how to obtain the answer.
    On a 4.70 x110 system, how do you tell if the current kernel is ASCII or Unicode?  We are still trying to upgrade our kernel from 6.20 to 6.40 in order to upgrade the OS to v6.1.
    Thanks in advance,
    Robert

    Hi Robert,
    you can see it in the dtaara KERNEL:
    Datenbereich  . . . . . :   KERNEL      
       Bibliothek  . . . . . :     R3700I0211
    Art . . . . . . . . . . :   *CHAR       
    Länge . . . . . . . . . :   50          
    Text  . . . . . . . . . :               
                Wert                         
    Offset      *...+....1....+....2....+....
         0      '1R7007U01700V5R3M000211     
    You can see it in the U or A at position 7 )
    Regards
    Volker Gueldenpfennig, consolut international ag
    http://www.consolut.net - http://www.4soi.de - http://www.easymarketplace.de

  • Combined Upgrade and Unicode conversion of Sap 4.6C to ECC6.0

    Hello all,
    my project team intends to carry out a combined upgrade and unicode conversion of an SAP ERP 4.6C system with MDMP to ECC6.0 (no enhancement package). The system is running on Oracle 10.2.
    In preparation for this upgrade, I have gone through the SAP notes 928729, 54801.
    We need to get a rough estimate of the entire downtime so as to alert our end users. From the CU&UC documentation in 928729, I read up note 857081. However the program in this note cannot be used to estimate the downtime as my system is < SAP netweaver 6.20.
    Is there any other SAP note or tool or program that I can use to estimate the downtime for the entire CU&UC? Thanks a lot!

    Hi,
    Combined upgrade depend upon number of factors like database size, resources on the server and optimization. In order to get idea of how much downtime, it will take, I would suggest you to do combined upgrade and unicode conversion on sandbox system which should be the replica of your production system. And try to optimize it. From there you can get approx. downtime required.
    Also, please read combined upgrade and unicode conversion guides on  http://service.sap.com/unicode@sap
    Thanks
    Sunny

  • Need suggestion on Multi currency and Unicode character set use in ABAP

    Hi All,
    Need suggestion. In one of the requirement I saw 'multi-currency and Unicode character set experience in FICO'.
    Can you please elaborate me how ABAPers are invlolved in multi currency as I think this is FICO fuctional area.
    And also what is Unicode character set exp.? Please give me some document of you have any.
    Thanks
    Sreedevi
    Moderator message - This isn't the place to prepare for interviews - thread locked
    Edited by: Rob Burbank on Sep 17, 2009 4:45 PM

    Use the default parser.
    By default, WebLogic Server is configured to use the default parser and transformer to parse and transform XML documents. The default parser and transformer are those included in the JDK 5.0.
    The built-in WebLogic Server DOM factory implementation class is com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl.
    The DocumentBuilderFactory.newInstance method returns the built-in parser.
    DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance();

  • 4.7EEx1.10 to ECC6.0 upgrade and Unicode conversion

    Hi Experts,
    We are going to initiate the upgrade from next month onwards. Subsequently i have started preparing the plan and strategy for the same.
    As our current setup is 4.7EEx110/Win 2003 R2-64 bit/Oracle 10.2.0.4.0 (Non unicode). And we have recently migrated on to this setup from WIn2k 32 bit. Also the current hardware is Unicode compatible.
    With respect to strategy for achieving this Upgrade and Unicode conversion, i am planning as follows.
    Step 1) Perform Unicode conversion on the current landscape (Both Export/import on the same servers)
    Step 2) Setup Temporary landscape as part of Dual maintenance strategy and migrate data from the current systems to temporary systems using backup/restore method.
    Step 3) Perform the SAP version upgrade on the current landscape and setup transport routes from temporary to current landscape in order to keep it in sync
    Step 4) after successful upgrade, decommission the temporary landscape
    Please provide your suggestions and valuable advices if there is anything wrong with my strategy and execution plan.
    Regards,
    Dheeraj

    Hi,
    Thanks. As i have already referred these notes as i am seeking advise with respect to my upgrade approach.
    However i have planned to perform in the following manner.
    1) Refresh Sandbox with Prod data and perform Upgrade to ECC6.0 EHP5 & subsequently Unicode conversion on the same server (Since both export & Import has to perform on the same hardware as we have recently migrated on this hardware which is Unicode compatible)
    2) Setup temporary landscape for DEv & QAs and establish transport connection to Production system in order to move urgent changes
    3) Keep a track of the changes which have transported during upgrade phase so that the same can be implemented in the upgraded systems i.e. Dev & QAS
    4) After Sandbox Migration and signoff, we will perform Dev & QAS upgrade & unicode conversion on the same hardware (Note: Since these are running on VMware can we export the data from the upgraded system and import on to a new VM?)
    5) Plan for production cutover and Upgrade the Prod system to ECC6.0 Ehp5 and then Unicode conversion. As i am planning to perform upgrade over the weekend and then Unicode conversion activity in the next weekend (Is it a right way?)
    My Production setup: DB on one Physical host and CI on separate Virtual host
    6) After the stabilization phase, we are planning for OS & DB upgrade as follows:
          a) Windows upgrade from 2003 R2 to Windows 2008 R2
          b) Oracle Upgrade from 10.2 to 11.2
    If anyone thinks that there is anything wrong with my above approach and need changes then please revert.
    I have one more doubt as I am going to upgrade 4.7EEx110 (WAS 620, Basis SP64) to ECC6.0 EHp5.As I presume that I can straight away upgrade from the current version to ECC6.0 Ehp5 without installing EHP. Kindly confirm
    Thanks

  • Windows API ANSI version and UNICODE version

    Windows API have two version:
    ANSI version and UNICODE.
    I want to display chinese, so i use the UNICODE version. but i find it's some mistake.
    and later,  i change to ANSI version, it's correct.
    why the ANSI version can display chinese, and UNICODE can't.
    i use the api function
    FindWindowA
    GetWindowTextA
    who can tell me?

    I understand Vista is a 32 bit.
    IE 8 is what I have but I would consider IE 9 if that would be necessary.
    Just a bit more update. I just tried to download Adobe FP 10 again with the following results, again:
    The Adobe Download Manager windows shows:
    FP 10.3 Status as 100% "Installation Pending".
    FP 10.2 Status as 100% " Instatlling Application"
    and that will go one forever but never really actually installs anything and just keeps cycling through the "Add to Download" e.g. "Fanbase", "Times Reader"
    and "Adobe Air" (none of which I want added to the download), and the download and installation shows as 100%.
    Go figure!

  • Combined Upgrade and Unicode conversion question

    Hello Everyone,
    I will be performing combined Upgrade and Unicode conversion soon. Currently i have run Prepare and do not have any errors.
    I have already run SPUMG consistency check and i do not have any errors there. Since this is Combined Upgrade and unicode conversion according to guide i do not need to do the Nametab conversion right now. But now if i go this place:
    SPUMG -> Status -> Additional Information  , i see a status with red for Unicode nametabs are not consistent or not up-to-date.
    Please let me know if i can ignore this step and do the nametab conversion after upgrade is complete and before unicode conversion.
    Thanks,
    FBK

    Please follow the instructions from the guides.
    The Unicode nametabs will be generated automatically during the upgrade.
    An additional check is integrated into the final preparation steps in the target releases.
    Regards,
    Ronald

  • Need your suggestions - how to display large file in ASCII and HEX

    Hello,
    I want to create an application which can read in a large file and switch between displaying ASCII and HEX (formatted a particular way). There are two problems here that I'm not quite sure how to solve.
    1. How to switch dynamically between ASCII and HEX. Should the HEX formatter be in the document of the JTextArea(or equivalent), or somewhere in the view? If it's in the view then where? I'd rather not read in the file more than once.
    2. How to do some kind of paging scheme for huge files. I'd like to read in part of the file and display it, then when the user scrolls to another area, read in that part of the file.
    Thanks!
    Jeff

    Hello,
    I want to create an application which can read in a
    large file and switch between displaying ASCII and
    HEX (formatted a particular way). There are two
    problems here that I'm not quite sure how to solve.
    1. How to switch dynamically between ASCII and HEX.
    Should the HEX formatter be in the document of the
    e JTextArea(or equivalent), or somewhere in the view?
    If it's in the view then where? I'd rather not read
    d in the file more than once.You can iterate over all the characters in the String using String.charAt, cast the chars to ints, and call Integer.toHexValue(...)
    >
    2. How to do some kind of paging scheme for huge
    files. I'd like to read in part of the file and
    display it, then when the user scrolls to another
    area, read in that part of the file.
    Thanks!
    Jeff

  • Reading a fle format thats part ascii and part binary

    Hey
    I'm trying to write a importer for the PLY filke format. That file format starts of with a header in ascii and then continues with either the data in ascii or in binary. When the data is in ascii I have no problem reading the file but when the data is in binary I'm noth quite sure how I should do it. Right now I do something like this:
    BufferedInputStream stream = new BufferedInputStream(new FileInputStream(file)); // I first create this
    BufferedReader reader = new BufferedReader(new InputStreamReader(stream));  // then this is for reader ascii
    DataInputStream dataInputStream = new DataInputStream(stream); // And this is for reading binaryThen I read the ascii header with the BuffereReader and then wheter I turns out that the data is in ascii I countine reading with the reader, otherwise I start reading with the DataInputStream. But the data from the DataInputStream is not the same as I get from the BufferedReader. I try it on these files that are in diffrent endians and either ascii ot binary
    http://www.cs.ucl.ac.uk/staff/Joao.Oliveira/ply.html in the Nbunny.zip.
    Anyone that can help me with this? Or know about a PLY file reader in java already?
    Discordia
    Message was edited by:
    Discordia

    DataInputStream is NOT the right class to use for general binary input. It is used only for the specific purpose of reading Java primitives. The input obviously needs to be in the right format for that to work, i.e. it needs to be written using a DataOutputStream.
    I would read everything using an InputStream, not a reader. Store the bytes of the header until you reach the end of the header (not sure how to determine that since I don't know the fileformat). Then convert those bytes into a String using the proper charset ("US-ASCII" in this case, if by "ascii" you really mean ASCII).
    Then just continue reading the binary data.

Maybe you are looking for