Java Unicodes and ASCII conversion

Hi,
Does anyone know where I could find the unicodes for the ascii codes from ascii 0 to ascii 32 as when I try printing characters using S.o.p("\n"); it does not seem to work for some reason (to give a new line. So I want to use the unicode values but I cant find them anywhere.
Can anyone help? Thanks in advance.
John Loughran

The Unicode values should be 0-32 also.
If you are having problems writing line breaks, try writing a whole carriage return + line feed instead of just the line feed, like:
write("\r\n");
Some text editors and viewers don't like just plain line feeds. I believe that using a PrintWriter to write the data should translate CRLF's for you.

Similar Messages

  • Unicode and ascii conversion help needed

    I am trying to read passwords from a foxpro .dbf. The encrpytion of the password is crude, it takes the ascii value of each char entered and adds an integer value to it, then stores the complete password to the table. So to decode, just subtract same integer value from each chars retieved from .dbf. pretty simple.
    The problem is that java chars and strings are unicode, so when my java applet retrieves these ascii values from the .dbf they are treated as unicode chars, if the ascii value is over 127 I have problems.
    The question. how can i retrieve these ascii values as ascii values in java?
    Should I use an InputStream like:
    InputStream is=rs.getAsciiStream("password");
    Is there a way to convert from unicode to extended ascii?
    Some examples would be helpful, Thanks in advance.

    version 1
    import java.nio.charset.Charset;
    import java.nio.ByteBuffer;
    import java.nio.CharBuffer;
    class Test {
        static char[] asciiToChar(byte[] b) {
            Charset cs = Charset.forName("ASCII");
            ByteBuffer bbuf = ByteBuffer.wrap(b);
            CharBuffer cbuf = cs.decode(bbuf);
            return cbuf.array();
        static byte[] charToAscii(char[] c) {
            Charset cs = Charset.forName("ASCII");
            CharBuffer cbuf = CharBuffer.wrap(c);
            ByteBuffer bbuf = cs.encode(cbuf);
            return bbuf.array();
    }version 2
    import java.io.*;
    import java.nio.charset.Charset;
    class Test {
        static char[] asciiToChar(byte[] b) throws IOException {
            Charset cs = Charset.forName("ASCII");
            ByteArrayInputStream bis = new ByteArrayInputStream(b);
            InputStreamReader isr = new InputStreamReader(bis, cs);
            char[] c = new char[b.length];
            isr.read(c, 0, c.length);
            return c;
        static byte[] charToAscii(char[] c) throws IOException {
            Charset cs = Charset.forName("ASCII");
            ByteArrayOutputStream bos = new ByteArrayOutputStream();
            OutputStreamWriter osw = new OutputStreamWriter(bos, cs);
            osw.write(c, 0, c.length);
            osw.flush();
            byte[] b = bos.toByteArray();
            return b;
    }

  • 4.6C to ERP6.0 Upgrade - what are the recommendations for Unicode and Java

    Hi All,
    Forgive such a basic question.
    We are just beginning to look at our ERP6.0 upgrade and am unclear whether we should convert to unicode at the same time, and also what are the uses for the Java stack.
    Unicode?
    We are a single language Latin-1 site so don't need Unicode. However will it become mandatory eventually? Are we better, if we have the opportunity now, just to convert. Also I believe the Java stack is only Unicode. Does ASCII ABAP and Java Unicode complicate things?
    Java?
    Will we need it? We are proposing a technical upgrade from our 4.6c Javaless environment. Where does Java come in to play in ERP6.0? Doesn't it increase greatly the iSeries HW requirements? (I'm looking at a 2 way 550 at the moment)
    Look forward to hearing your views.
    Regards
    Steve

    Hi Stephen,
    at the moment, in a latin-1 environment Unicode is for old customers not a requrirement. This might change in 5-10 years ...
    Java: No problem at all with ASCII (in latin-1 environments)
    => there is no need for you, but you could go for it. We did several Unicode conversions already and these Latin-1 conversions are pretty easy and simple especially on iSeries, because of the special InPlace Unicode Conversion.
    ... but this can be done at any time ...
    Regards
    Volker Gueldenpfennig, consolut.gmbh
    http://www.consolut.de - http://www.4soi.de - http://www.easymarketplace.de

  • Java Editor and Unicode Characters

    I'm trying to use unicode characters in strings with the Java Editor.
    String myOption = "\u2666 " + myDirName;
    This works nicely, displaying a diamond in my web application. Source HTML contains a &9830;
    However, if I make changes in the Visual Designer (e.g., add a button), the \u2666 in my Java code changes to a diamond symbol. When that happens, I get a "?" in my web application.
    How do I prevent Java Editor from switching to the display character?

    Thanks for your response.
    After more research, I found that the .java file is corrupted. These means the problem is not related to the browser.
    When I have "\u2666" in the Java Editor window, I see "\u2666" in the .java file in Word Pad. This is properly compiled and the entity ♦ appears in the HTML.
    When the diamond symbol is in the Java Editor window, I see and ascii ? (character #3F). I get a ? in the browser. It is really a question mark (#3F) not an unknown character.
    So, it appears that when the JSP is changed by adding a button, the .java code is rewritten to disk but unicode characters are not translated properly. Or - something like that. If I could keep the ascii string "\u2666" from converting to the diamond, I would be all set.
    I can edit the Java Code, close, save and reopen all I want and the unicode character doesn't translate to the symbol. This ONLY happens when the JSP is modified.
    Any ideas?

  • Time comparision of ASCII conversion VS Unicode conversion

    In general how does the runtime of a Unicode conversion compare to that of the ASCII conversion?  For example, you perform the Unicode conversion on the same hardware on which you completed the ASCII conversion.  The ASCII conversion takes 20 hours.  Does the Unicode conversion take a similiar time?
    thanx
    Mark

    Hi Mark,
    first an answer, that comes only now, because my user was somehow locked or whatever ...
    How long does it take compared to the ASCII CPC ?
    As you are a latin-1 customer I guess, you could make use of the InPlace version in both attempts.
    The InPlace Unicode CPC Export will take 6-7h or the time of the Ascii Export if this was longer. The reload will take about 3h only.
    The major problem is on some other edge:
    The combined upgrade only works fine if you run all the scans and this will take several days ... even when they are "pretty useless" in your case. Therefore, I would recommend an ECC6 upgrade without unicode and then the unicode conversion directly afterwards - in a short special technique, that I used several times already.
    If it is "just" from 4.6C ASCII to ECC 6.0 Unicode, the week will be fully sufficient including all backups - which can sometimes run in parallel. (at least with my ideas)
    I did something like that from 4.6B ASCII to ecc5 Unicode on one weekend from friday evening until sunday afternoon. (for sure, depending on the systemsize)
    Regards
    Volker Gueldenpfennig, consolut.gmbh
    http://www.consolut.de - http://www.4soi.de - http://www.easymarketplace.de

  • Java mapping and no content conversion

    Hi all,
    I have a file to idoc scenario. The file is in flat idoc format and I've made my own java mapping that gets the flat file and converts it to idoc xml format.
    I don't use content conversion in the sender file adapter because I don't wan't the file to be converted to xml before the java mapping processes it. But with this configuration my scenario fails in the interface determination step, as it doesn't find an xml to apply the java mapping.
    Is there a way to make it work as I want? (with my java mapping and not applying content conversion?) If not, is there a way to put all the file in a single xml tag using content conversion?
    I mean something like this:
    <FILE>
    </FILE>
    This way it would be easy for my java mapping to remove the tags and work with the flat file as it does now.
    Thanks in advance.

    Thanks Stephan for your answer.
    In fact I have the defined that interface, as well as a dummy data type and message type that uses it. I still get the error though. My guess is that, as I don't use content conversion in file adapter (message protocol is file) the outbound message is in fact not in xml format, and it cannot identify it in interface determination. That's my guess but perhaps I'm wrong?
    This is the error I'm getting:
    <?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
    - <!--  Interface Determination
      -->
    - <SAP:Error xmlns:SAP="http://sap.com/xi/XI/Message/30" xmlns:SOAP="http://schemas.xmlsoap.org/soap/envelope/" SOAP:mustUnderstand="">
      <SAP:Category>XIServer</SAP:Category>
      <SAP:Code area="IF_DETERMINATION">CX_ID_PLSRV</SAP:Code>
      <SAP:P1>No interface determination found for outbound interface http://********.****/EDELIVERY.SI_OUT_A_0596_RECADV: Error while determining root tag of XML: BOM / charset detection failed</SAP:P1>
      <SAP:P2 />
      <SAP:P3 />
      <SAP:P4 />
      <SAP:AdditionalText />
      <SAP:Stack>Error when determining the inbound interface: No interface determination found for outbound interface http://********.****/EDELIVERY.SI_OUT_A_0596_RECADV: Error while determining root tag of XML: BOM / charset detection failed No interface determination found for outbound interface http://********.****/EDELIVERY.SI_OUT_A_0596_RECADV: Error while determining root tag of XML: BOM / charset detection failed Error while determining root tag of XML: BOM / charset detection failed Error while parsing an XML stream: 'BOM / charset detection failed'.</SAP:Stack>
      <SAP:Retry>M</SAP:Retry>
      </SAP:Error>
    Any idea?

  • Unicode and Java

    Hi
    As we all know Java treat character literals as Unicode characters. I have been studying Unicode and the way they treat characters and I have a doubt which is not specific to Java code but specific to Unicode.
    Unicode states that each character is assigned a number which is unique, this number is called code point.
    The relationship between characters and code points is 1:1.
    Eg: the String *"hello"* (which is sequence of character literals) can be represent by the following Code Points
    *\u0065 \u0048 \u006c \u006c \u006f*
    I also read that a certain character code point must be recognized by a specific encoding or else a question mark (?) is output in place of the character. Not all code points can be recognized by an encoding.
    So, the letter *&#1500;* would not be recognized by all encodings and should be replaced by a question mark (?) right?
    The interesting is that this code point represents a different character and not a *"?"* in other encodings. It should print the same character
    This is the HTML code I used for tests (save it in your hard disk and open using your navigator, then select the following encodings: UTF16, ISO-8859-1)
    <html>
    <body>
    &#1502;&#1506;&#1497;&#1500; &#1488;&#1495;&#1491; &#1489;&#1490;&#1513;&#1501;, &#1500;&#1497;&#1500;&#1492; &#1513;&#1500; &#1488;&#1508;&#1512;&#1497;&#1500;
    &#1504;&#1508;&#1514;&#1495; &#1499;&#1502;&#1493; &#1506;&#1504;&#1503;, &#1493;&#1512;&#1506;&#1501; &#1488;&#1494; &#1502;&#1488;&#1497;&#1512;
    &#1502;&#1506;&#1497;&#1500; &#1488;&#1495;&#1491; &#1489;&#1490;&#1513;&#1501;, &#1500;&#1497;&#1500;&#1492; &#1495;&#1501; &#1493;&#1511;&#1512;
    &#1504;&#1508;&#1512;&#1505; &#1499;&#1502;&#1493; &#1495;&#1493;&#1508;&#1492;, &#1493;&#1502;&#1514;&#1495;&#1514; &#1488;&#1504;&#1497; &#1513;&#1512;
    &#1502;&#1506;&#1497;&#1500; &#1488;&#1495;&#1491; &#1489;&#1490;&#1513;&#1501;, &#1512;&#1496;&#1493;&#1489;, &#1500;&#1502;&#1497; &#1488;&#1499;&#1508;&#1514;
    &#1488;&#1504;&#1497; &#1500;&#1488; &#1506;&#1500; &#1492;&#1488;&#1512;&#1509;, &#1488;&#1497;&#1514;&#1498; &#1500;&#1502;&#1506;&#1500;&#1492; &#1513;&#1496;
    &#1512;&#1493;&#1495; &#1489;&#1508;&#1504;&#1497;&#1501;, &#1496;&#1497;&#1508;&#1493;&#1514; &#1492;&#1490;&#1513;&#1501; &#1492;&#1488;&#1495;&#1512;&#1493;&#1504;&#1493;&#1514;
    &#1504;&#1493;&#1490;&#1506;&#1493;&#1514; &#1489;&#1500;&#1495;&#1497;&#1497;&#1501;, &#1489;&#1508;&#1504;&#1497;&#1497;&#1498; &#1502;&#1513;&#1495;&#1511;&#1493;&#1514;
    &#1488;&#1502;&#1510;&#1506; &#1492;&#1512;&#1495;&#1493;&#1489;, &#1499;&#1493;&#1500;&#1501; &#1499;&#1489;&#1512; &#1497;&#1513;&#1504;&#1497;&#1501;
    &#1492;&#1497;&#1497;&#1514;&#1492; &#1506;&#1491;&#1492; &#1492;&#1512;&#1493;&#1495; &#1493;&#1506;&#1493;&#1491; &#1513;&#1504;&#1497; &#1499;&#1493;&#1499;&#1489;&#1497;&#1501;
    &#1488;&#1502;&#1510;&#1506; &#1492;&#1512;&#1495;&#1493;&#1489;, &#1499;&#1493;&#1500;&#1501; &#1499;&#1489;&#1512; &#1497;&#1513;&#1504;&#1497;&#1501;,
    &#1492;&#1497;&#1497;&#1514;&#1492; &#1506;&#1491;&#1492; &#1492;&#1512;&#1493;&#1495; &#1493;&#1506;&#1493;&#1491; &#1513;&#1504;&#1497; &#1499;&#1493;&#1499;&#1489;&#1497;&#1501;
    &#1512;&#1488;&#1497;&#1514;&#1497; &#1494;&#1493;&#1490; &#1506;&#1497;&#1504;&#1497;&#1497;&#1501;, &#1502;&#1505;&#1512;&#1489;&#1493;&#1514; &#1500;&#1492;&#1497;&#1508;&#1514;&#1495;
    &#1510;&#1493;&#1500;&#1500;&#1514; &#1488;&#1500; &#1506;&#1510;&#1502;&#1498; &#1506;&#1502;&#1493;&#1511; &#1489;&#1497;&#1501; &#1513;&#1500;&#1498;,
    &#1502;&#1491;&#1497; &#1508;&#1506;&#1501; &#1488;&#1514; &#1506;&#1493;&#1500;&#1492;, &#1500;&#1493;&#1511;&#1495;&#1514; &#1511;&#1510;&#1514; &#1488;&#1493;&#1497;&#1512;
    &#1500;&#1488; &#1512;&#1493;&#1510;&#1492; &#1500;&#1492;&#1497;&#1505;&#1495;&#1507;, &#1502;&#1499;&#1497;&#1512;&#1492; &#1488;&#1514; &#1492;&#1502;&#1495;&#1497;&#1512;
    &#1488;&#1502;&#1510;&#1506; &#1492;&#1512;&#1495;&#1493;&#1489;, &#1499;&#1493;&#1500;&#1501; &#1499;&#1489;&#1512; &#1497;&#1513;&#1504;&#1497;&#1501;...
    </body>
    </html>I would appreciate if you correct me in case I am wrong!
    Edited by: charllescuba1008 on Mar 31, 2009 2:08 PM

    charllescuba1008 wrote:
    Unicode states that each character is assigned a number which is unique, this number is called code point. Right.
    The relationship between characters and code points is 1:1.Uhm .... let's assume "yes" for the moment. (Note that the relationship between the Java type char and code point is not 1:1 and there are other exceptions ...)
    Eg: the String *"hello"* (which is sequence of character literals) can be represent by the following Code Points
    *\u0065 \u0048 \u006c \u006c \u006f*Those are the Java String unicode escapes. If you want to talk about Unicode Codepoints, then the correct notation for "Hello" would be
    U+0048 U+0065 U+006C U+006C U+006F
    Note that you swapped the H and e.
    I also read that a certain character code point must be recognized by a specific encoding or else a question mark (?) is output in place of the character.This one is Java specific. If Java tries to translate some unicode character to bytes using some encoding that doesn't support that character then it will output the byte(s) for "?" instead.
    Not all code points can be recognized by an encoding.Some encodings (such as UTF-8) can encode all codepoints, others (such as ISO-8859-*, EBCDIC or UCS-2) can not.
    So, the letter *&#1500;* would not be recognized by all encodings and should be replaced by a question mark (?) right?Only in a very specific case in Java. This is not a genral Unicode-level rule.
    (disclaimer: the HTML code presented was using decimal XML entities to represent the unicode characters).
    What you are seing is possibly the replacement character that your text rendering system uses to represent characters that it knows, but can't display (possibly because the current font has no character for them).

  • Java Internationalization and .properties files

    Hi everyone,
    Isn't it problematic that the way to internationalize your java apps ( .properties files ) doesn't support UTF-8 and you have to user gibberish looking unicode escapes. Some will say that you don't need to write and the converters or editors will handle the ascii conversion but the requirement for such a intermediate process doesn't seem right. Sometime you have to edit some string on the server and it should be human readable using text editors that support UTF-8. I thought loadFromXML and storeToXML methods that came with java 1.5 seemed to solve the problem but than noticed that PropertyResourceBundle doesn't support xml properties files. Is backwards compatibility the reason that properties files aren't utf-8 by default?
    Thanks
    Bilgehan

    Try PRBeditor (http://prbeditor.dev.java.net).

  • Unicode and non-unicode

    WHAT IS DIFFRENTS BETWEEN UNICODE AND NON UNICODE ?
    BRIEFLY EXPLAIN ABOUT UNICODE?
                                                            THANKS IN ADVANCES

    A 16-bit character encoding scheme allowing characters from Western European, Eastern European, Cyrillic, Greek, Arabic, Hebrew, Chinese, Japanese, Korean, Thai, Urdu, Hindi and all other major world languages, living and dead, to be encoded in a single character set. The Unicode specification also includes standard compression schemes and a wide range of typesetting information required for worldwide locale support. Symbian OS fully implements Unicode. A 16-bit code to represent the characters used in most of the world's scripts. UTF-8 is an alternative encoding in which one or more 8-bit bytes represents each Unicode character. A 16-bit character set defined by ISO 10646. A code similar to ASCII, used for representing commonly used symbols in a digital form. Unlike ASCII, however, Unicode uses a 16-bit dataspace, and so can support a wide variety of non-Roman alphabets including Cyrillic, Han Chinese, Japanese, Arabic, Korean, Bengali, and so on. Supporting common non-Roman alphabets is of interest to community networks, which may want to promote multicultural aspects of their systems.
    ABAP Development under Unicode
    Prior to Unicode the length of a character was exactly one byte, allowing implicit typecasts or memory-layout oriented programming. With Unicode this situation has changed: One character is no longer one byte, so that additional specifications have to be added to define the unit of measure for implicit or explicit references to (the length of) characters.
    Character-like data in ABAP are always represented with the UTF-16 - standard (also used in Java or other development tools like Microsoft's Visual Basic); but this format is not related to the encoding of the underlying database.
    A Unicode-enabled ABAP program (UP) is a program in which all Unicode checks are effective. Such a program returns the same results in a non-Unicode system (NUS) as in a Unicode system (US). In order to perform the relevant syntax checks, you must activate the Unicode flag in the screens of the program and class attributes.
    In a US, you can only execute programs for which the Unicode flag is set. In future, the Unicode flag must be set for all SAP programs to enable them to run on a US. If the Unicode flag is set for a program, the syntax is checked and the program executed according to the rules described in this document, regardless of whether the system is a US or a NUS. From now on, the Unicode flag must be set for all new programs and classes that are created.
    If the Unicode flag is not set, a program can only be executed in an NUS. The syntactical and semantic changes described below do not apply to such programs. However, you can use all language extensions that have been introduced in the process of the conversion to Unicode.
    As a result of the modifications and restrictions associated with the Unicode flag, programs are executed in both Unicode and non-Unicode systems with the same semantics to a large degree. In rare cases, however, differences may occur. Programs that are designed to run on both systems therefore need to be tested on both platforms.
    Refer to the below related threads
    Re: Why the select doesn't run?
    what is unicode
    unicode
    unicode
    Regards,
    Santosh

  • Diacritic: unicode to ascii

    I'm looking for an algorithm to convert Unicode ro ASCII characters. The algorithm does not have to cover all unicode characters, but Latin1 and Latin 2 should be covered. If the character cannot be converted (no representation in ascii), then i can do it manually (or use a "?")
    thanx.
    trev

    From your title I assume you want to just drop any diacritics attached to the characters?
    Some time ago I downloaded a file from the Unicode site which contains lines like010C;LATIN CAPITAL LETTER C WITH CARON;Lu;0;L;0043 030C;;;;N;LATIN CAPITAL LETTER C HACEK;;;010D;This contains the Unicode description of the character &#x010c; in particular that it can be decomposed to 0043 (C) and 030C (combining hook above). I don't see that property exposed in the java.lang.Character class, but you could use it to create a hard-coded mapping table.

  • Unicode and non-unicode string data types Issue with 2008 SSIS Package

    Hi All,
    I am converting a 2005 SSIS Package to 2008. I have a task which has SQL Server as the source and Oracle as the destination. I copy the data from a SQL server view with a field nvarchar(10) to a field of a oracle table varchar(10). The package executes fine
    on my local when i use the data transformation task to convert to DT_STR. But when I deploy the dtsx file on the server and try to run from an SQL Job Agent it gives me the unicode and non-unicode string data types error for the field. I have checked the registry
    settings and its the same in my local and the server. Tried both the data conversion task and Derived Column task but with no luck. Pls suggest me what changes are required in my package to run it from the SQL Agent Job.
    Thanks.

    What is Unicode and non Unicode data formats
    Unicode : 
    A Unicode character takes more bytes to store the data in the database. As we all know, many global industries wants to increase their business worldwide and grow at the same time, they would want to widen their business by providing
    services to the customers worldwide by supporting different languages like Chinese, Japanese, Korean and Arabic. Many websites these days are supporting international languages to do their business and to attract more and more customers and that makes life
    easier for both the parties.
    To store the customer data into the database the database must support a mechanism to store the international characters, storing these characters is not easy, and many database vendors have to revised their strategies and come
    up with new mechanisms to support or to store these international characters in the database. Some of the big vendors like Oracle, Microsoft, IBM and other database vendors started providing the international character support so that the data can be stored
    and retrieved accordingly to avoid any hiccups while doing business with the international customers.
    The difference in storing character data between Unicode and non-Unicode depends on whether non-Unicode data is stored by using double-byte character sets. All non-East Asian languages and the Thai language store non-Unicode characters
    in single bytes. Therefore, storing these languages as Unicode uses two times the space that is used specifying a non-Unicode code page. On the other hand, the non-Unicode code pages of many other Asian languages specify character storage in double-byte character
    sets (DBCS). Therefore, for these languages, there is almost no difference in storage between non-Unicode and Unicode.
    Encoding Formats: 
    Some of the common encoding formats for Unicode are UCS-2, UTF-8, UTF-16, UTF-32 have been made available by database vendors to their customers. For SQL Server 7.0 and higher versions Microsoft uses the encoding format UCS-2 to store the UTF-8 data. Under
    this mechanism, all Unicode characters are stored by using 2 bytes.
    Unicode data can be encoded in many different ways. UCS-2 and UTF-8 are two common ways to store bit patterns that represent Unicode characters. Microsoft Windows NT, SQL Server, Java, COM, and the SQL Server ODBC driver and OLEDB
    provider all internally represent Unicode data as UCS-2.
    The options for using SQL Server 7.0 or SQL Server 2000 as a backend server for an application that sends and receives Unicode data that is encoded as UTF-8 include:
    For example, if your business is using a website supporting ASP pages, then this is what happens:
    If your application uses Active Server Pages (ASP) and you are using Internet Information Server (IIS) 5.0 and Microsoft Windows 2000, you can add "<% Session.Codepage=65001 %>" to your server-side ASP script.
    This instructs IIS to convert all dynamically generated strings (example: Response.Write) from UCS-2 to UTF-8 automatically before sending them to the client.
    If you do not want to enable sessions, you can alternatively use the server-side directive "<%@ CodePage=65001 %>".
    Any UTF-8 data sent from the client to the server via GET or POST is also converted to UCS-2 automatically. The Session.Codepage property is the recommended method to handle UTF-8 data within a web application. This Codepage
    setting is not available on IIS 4.0 and Windows NT 4.0.
    Sorting and other operations :
    The effect of Unicode data on performance is complicated by a variety of factors that include the following:
    1. The difference between Unicode sorting rules and non-Unicode sorting rules 
    2. The difference between sorting double-byte and single-byte characters 
    3. Code page conversion between client and server
    Performing operations like >, <, ORDER BY are resource intensive and will be difficult to get correct results if the codepage conversion between client and server is not available.
    Sorting lots of Unicode data can be slower than non-Unicode data, because the data is stored in double bytes. On the other hand, sorting Asian characters in Unicode is faster than sorting Asian DBCS data in a specific code page,
    because DBCS data is actually a mixture of single-byte and double-byte widths, while Unicode characters are fixed-width.
    Non-Unicode :
    Non Unicode is exactly opposite to Unicode. Using non Unicode it is easy to store languages like ‘English’ but not other Asian languages that need more bits to store correctly otherwise truncation will occur.
    Now, let’s see some of the advantages of not storing the data in Unicode format:
    1. It takes less space to store the data in the database hence we will save lot of hard disk space. 
    2. Moving of database files from one server to other takes less time. 
    3. Backup and restore of the database makes huge impact and it is good for DBA’s that it takes less time
    Non-Unicode vs. Unicode Data Types: Comparison Chart
    The primary difference between unicode and non-Unicode data types is the ability of Unicode to easily handle the storage of foreign language characters which also requires more storage space.
    Non-Unicode
    Unicode
    (char, varchar, text)
    (nchar, nvarchar, ntext)
    Stores data in fixed or variable length
    Same as non-Unicode
    char: data is padded with blanks to fill the field size. For example, if a char(10) field contains 5 characters the system will pad it with 5 blanks
    nchar: same as char
    varchar: stores actual value and does not pad with blanks
    nvarchar: same as varchar
    requires 1 byte of storage
    requires 2 bytes of storage
    char and varchar: can store up to 8000 characters
    nchar and nvarchar: can store up to 4000 characters
    Best suited for US English: "One problem with data types that use 1 byte to encode each character is that the data type can only represent 256 different characters. This forces multiple
    encoding specifications (or code pages) for different alphabets such as European alphabets, which are relatively small. It is also impossible to handle systems such as the Japanese Kanji or Korean Hangul alphabets that have thousands of characters."<sup>1</sup>
    Best suited for systems that need to support at least one foreign language: "The Unicode specification defines a single encoding scheme for most characters widely used in businesses around the world.
    All computers consistently translate the bit patterns in Unicode data into characters using the single Unicode specification. This ensures that the same bit pattern is always converted to the same character on all computers. Data can be freely transferred
    from one database or computer to another without concern that the receiving system will translate the bit patterns into characters incorrectly.
    https://irfansworld.wordpress.com/2011/01/25/what-is-unicode-and-non-unicode-data-formats/
    Thanks Shiven:) If Answer is Helpful, Please Vote

  • EBCDIC - ASCII Conversion for XI File Adapters

    All:
    Does XI File Adpater have built-in capability to convert EBCDIC to ASCII and vice versa?  We also need to handle pack decimals.  All mainframe file based integration scenarios require this capability.
    If there's no built-in capability available, are there any Java APIs available from SAP XI bundle some where.
    Thanks and appreciate your feedback.

    Hi,
    Unfortunately, there is no built-in EBCDEC-ASCII conversion in XI.  There are examples when you do a search on GOOGLE.
    The java function can be used in either java mapping or adapter user-module.
    Regards,
    Bill

  • ABAP+Java Kernel and only Java Kernel

    Hi Friends,
    Is there any difference in ABAP+Java Kernel and only Java Kernel?
    Can we use the kernel for above both from the same location as below?
    1#
    Support Packages and Patches> Additional Components> SAP Kernel> SAP KERNEL 64-BIT UNICODE> SAP KERNEL 6.40 64-BIT UNICODE
    Following Files
    DB Dependent  - ORACLE
              SAPEXEDB_175-20000227.SAR
              DBATL640O92_44-20000227.SAR
    Database Independent
              SAPEXE_175-20000226.SAR
    Assuming above SAR files contain R3trans, R3load, tp ,... also.
    2#
    As per my knowledge SAP has given one more path to download J2EE SP's which contains Kernel and J2EE Support Packages.
    http://service.sap.com/patches --> Support Package and Patches -> Entry by Application Group -> SAP NetWeaver -> SAP NetWeaver -> SAP NetWeaver 04 -> Entry by Component -> Application Server Java -> SAP J2EE Engine 6.40 -> J2EERT19_0-10001982.SAR/J2EERT20_0-10001982.SAR (OS Independent) and CTRLDB219_0-20000114.SAR(OS Dendependent)
    What is the difference between above two kernels?
    1# contains R3trans, R3load, tp ...etc which is exclusively used by ABAP Stack related functions. The startsap script used here would check R3trans -d authentication successful before starting database and then only it starts the database from sidadm. I observed OPS$ users available in only ABAP schema SAPSID but not available in SAPSIDDB. If I use the kernel downloaded from 1# for Java Stack only system, startsap would fail as R3trans authentication fail as there is no ABAP schema in Java systems.
    This applies well for ABAP+JAVA or only ABAP based systems.
    If I use only JAVA Stack based system can I use the kernel downloaded from 1# . What is the kernel and SP available in 2#? Is this also contains the same kernel given in 1# or it is different. I also observed in JAVA they are not calling Kernels by Patch No like 154, 175... they call it by SP19. Please correct me if I am wrong.
    Did anyone come accross this type if issue? I would appreciate if anyone can clarify my questions given above.
    My main target is to have clear strategy for ABAP+JAVA, ABAP, JAVA kernels to avoid mess up. Which one to download? What is best practice? Differences?
    Thanks and Regards
    Prasad

    Hi Sacha,
    Thank a lot for your answer I am clear now. Before rewarding points to you.
    I have final question.
    You mean that there is not seperate kernel for Java Stack Systems.. When we apply SP11 or SP19 by default it applies Kernel and also support packages w.r.t the SP. Do we have option to seperate Kernel and Support package while applying SP's for Java System?
    If we take for example SAP NW 2004 (WAS 640) latest kernel is 175 (I mean ABAP+JAVA). If I want to know the latest kernel for Java Stack system, Do I need to say it is at level SP19 or check the sapwebdisp -V (Resepctive patch level)?
    For ABAP system we check disp+work -V to get the source kernel version. How to justify the source kernel version for Java System?
    As per SAP Note: 710914 - Upgrading the J2EE Engine Startup Framework after applying SP for Java System, Do we need to upgrade J2EE Startup Framework?
    Thanks and Regards,
    Prasad

  • Unicode and Non-Unicode Instances in one Transport Landscape

    We have a 4.7 landscape that includes a shared global development system supporting two regional landscapes.  The shared global development system is used for all ABAP/Workbench activity and for global customization used by both regional production systems.  The two regional landscapes include primarily three instances - Regional Configuration, Quality Assurance, and Production.  The transport landscape includes all systems with transport routes for global and regional.
    A conversion to unicode is also being planned for the global development and one regional landscape.  It is possible that we will not convert the other regional landscape due to pending discussions on consolidation.  This means one of the regional landscapes will be receiving global transports from a unicode-based system.  
    All information I've located implies no actual technical constraints.  Make sure you have the right R3trans versions, don't use non-Latin_1 languages, etc.  Basic caveats for a heterogenous environment ....
    Is anyone currently supporting a complete, productive landscape that includes unicode and non-unicode systems?   If so, any issues or problems encountered with transports across the systems?  (insignificant or significant)
    Information on actual experiences will be greatly appreciated ....
    Many thanks in advance.

    Hi Laura,
    Although i do not have the live / practical experience, but this is what i can share.
    I have been working on a Non-Unicode to Unicode conversion project. While we were in the discussion phase there was one such possibility of a scenario that part of the landscapes would remain non-unicode. So based on the research i did by reading and directly interacting with some excellent SPA consultants, i came to know there are absolutely no issues in transporting ABAP programs from a Unicode system to non-unicode system. In a Unicode system the ABAP code has already been checked and rectified for higher syntax checks and these are downward compatible with the ABAP code on lower ABAP versions and non-unicode systems. Hence i beleive there should not be any issues, however as i mentioned this is not from practical experience.
    Thanks.
    Chetan

  • ECC 6.0 - Non-Unicode and BI 7.0 - Unicode any known problems?

    We are determing whether or not to do a unicode conversion during our upgrade from BW 3.5 to BI 7.0.  Our R/3 system is non-unicode and my company does not utilize languages other than English and is not expected to at any time in the near future.  This being said we are trying to determine whether or not we should go ahead and do the conversion prior to being forced to by SAP at some point in the future.  I was just hoping I could get some insight into whether there were known issues with running the two versions in conjuction with each other.
    Thanks in advance,
    John

    We have a customer running this scenario, it should be possible to do so. Just keep in mind to do extensive testing as there might be some notes which are not yet in standard ...
    Regards, Kai

Maybe you are looking for

  • Menu exit-VA01/02/03

    Hi All,   I need to create a menu exit from VA* transactions to call a custom pricing screen. The user would invoke this from item pricing screen and it should call the Z transaction for the particular part number and display the price data. Can this

  • Clickable links in Podcast

    Does anyone know how to put clickable links in a podcast. I've been searching the web ti try nd solve this to no avail, I want to put a website and an email address o my Podcast thanks in advance

  • Don't know how to add songs from Ares to Ipod!

    Can anybody give me detailed instructions on how to get the music i downlaoded from Ares to my Ipod PLEASE!!!

  • ITunes $$ apps on 2 pcs but only one iPod touch- Help!

    Help! Did I screw up and lose $20? My wife has an iPod Touch 5, with an iTunes acct on her pc. I have occasionally downloaded eBooks for her, using my own pc, overdrive, and iTunes. That works fine. Today I thought I'd surprise her with a gift of an

  • Portlet inside Webcenter Framework Application

    Hi, I need to implement portlet functionality of editing and moving around components in my webcenterframework application. i went through the cue cards, and all possible oracle documents. i am not able to figure it out. Able to follow till registeri