Problem with spanish characters

Hello
I am using an existing and proven app written in visual foxpro with a recently installed Oracle11g database (have done this before without problems)
I am using Microsoft ODBC for Oracle and have no possibility of changing this. (always used this without problems)
If I insert a string in in a fixed length char field of the same length of the string (say char(30)) IF THE STRING DOES NOT HAVE SPECIAL CHARACTERS, everything goes as usual. However, if the string contains some spanish language chars (ó, é, á, ñ) the insert fails indicating that the string is too long. I suspect "somewhere" this characters are encoded in double-byte instead of single-byte.
My question is, which is the Oracle setting that allows/prevents this? I need special characters encoded in 1 byte. Language is Spanish.
Some settings that the Oracle DBA passed to me are
NLS_CHARACTERSET = AL32UTF8
NLS_NCHAR_CHARACTERSET= AL16UTF16
Any help will be appreciated, thanks!

In the database character set AL23UTF8 (NLS_CHARACTERSET=AL23UTF8), accented Spanish characters occupy 2 bytes each. This cannot be changed.
Two possible solutions:
-- recreate the database using the character set WE8MSWIN1252
-- redefine the column from CHAR(30), which usually by default means CHAR(30 BYTE), to CHAR(30 CHAR). This means no more than 30 characters but up to 120 bytes (in AL32UTF8).
If you run your application with NLS_LANG=.WE8MSWIN1252 and you do not use byte-oriented SQL functions such as SUBSTRB, INSTRB, or LENGTHB, you will most probably see no functional difference between single-byte storage and multibyte storage of Spanish characters. On the client side, these characters will be single-byte.
-- Sergiusz

Similar Messages

  • Spanish Dictionary and problems with spanish special characters

    I need a spanish dictionary, with all spanish words. I know that your priority are france, spanish, germany and Italy. So, the component has problems with spanish special words, this problem was notified before.
    Have you ready something? A new version?
    Thank you for your time.

    For the special character issue, you can check the reference:
    http://forums.adobe.com/message/2430501

  • Problems with special characters like "u00F6" "u00E4" "u00FC"

    Hi,
    we currently face some major problems with special characters like "ö" "ä" "ü". First I thought it is a SP02 specific problem but it still doesn't work with SP03.
    When I start an import the import manager shows correct converted value. "ö"s "ä"s "ü"s are displayed correctly. But after I've done the import I look into the Data manager and I see instead of these characters charcaters like "̈". (note: the character should show a little square box - but the forum cannot desplay it)
    How can I solve this problem?
    Regards
    Nico

    Hi Nico
      Just try it out the following steps, hope it solves ur issue.
    <b>To set one or more values back to their original values:</b>
    1. In the Import Manager Source Values grid, select the values that you want to set back to their original values.
    2. Right-click on one of the values and choose Set to Original Value from the context menu, or choose Values > Apply Value Conversion Filter > Set to Original Value from the main menu.
    3. MDME sets each value back to its original value, which appears in black in the Source Values grid.
    <b>NOTE : You can use the Set to Original Value filter only at the value-level and not at the field level.</b>
    thanks
    Alexander

  • Problems with special characters in InputStream or XPath

    Hello everyone,
    I am having problems with special characters such as ", ', - and so on, everything becomes ? (questionmark). (Not with special nordic characthers (å,æ,ø) though.)
    The encoding of the XML is UTF-8, and the server that holds the webservice is setting encoding to UTF-8 also, so no conflict there.
    What I have done is something like this:
    {code}
    String url = "http://www.formula1.com/rss/news/latest.rss"; // This is not the feed I use, it is a custom XML document
    InputSource is = new InputSource(url);
    DocumentBuilderFactory fct = DocumentBuilderFactory.newInstance();
    DocumentBuilder builder = fct.newDocumentBuilder();
    Document doc = builder.parse(is);
    Element rootElement = doc.getDocumentElement();
    XPathFactory factory = XPathFactory.newInstance();
    XPath xPath = factory.newXPath();
    out.print(xPath.evaluate("/rss/channel/title", rootElement)); // The xml is not a RSS feed, this is just an example.
    {code}
    Please let me know what I am doing wrong, or give me a hint of how to solve this in another way.
    Best regards,
    G.Hansen
    Edited by: G.Hansen on Mar 24, 2009 2:39 AM

    Hello, thanks for you reply. I had to rewrite your code a little, and I think that your code would not really compile, if you tried yourself :-)
    I ended up with something like this:
    our environment is running in 1.4 mode, so I could not use the for-each loop :-(
    public String printEntities(String s) {
         char[] sArray = s.toCharArray();
         StringBuffer sb = new StringBuffer();
           for (int i = 0; i < sArray.length; i++) {
             if (sArray[i] > 256)
                  sb.append("&#x" + Integer.toHexString(sArray) + ";");
         } else
              sb.append(sArray[i]);
         return sb.toString();

  • How to send Oracle rowid to servlet? | Problem with national characters.

    There is same possibility how to send rowid to servlet?
    I have now definition like this:
    <af:image source="/imageservlet?Par1=#{bindings.Col1.inputValue}"/>
    But If column contents national characters, servlet methods obtained changed these characters.
    My idea is to use not primary key for row, but use oracle rowid. It is simply possible?
    Use something like this:
    <af:image source="/imageservlet?Rowid=#{bindings.Rowid}"/
    Or Do you have ideas how to solve problem with national characters ?
    Thanks
    FiL

    Hi,
    Although your workaround works.
    I think this is a simple encoding problem.
    I simply need to make sure all parameters and pages are encoded with a char set which contains the national characters you mentioned.
    This is a bit dependent on the exact technology your using, but most can be done via the web.xml:
      <jsp-config>
          <jsp-property-group>
              <url-pattern>*.jsp</url-pattern>
              <page-encoding>UTF-8</page-encoding>
          </jsp-property-group>
      </jsp-config>     This forces all JSP pages to be encoded in UTF-8
    Adding the following parameter sometimes helps as well, although I think this one is a bit dated:
    You said your using a servlet so your servlet needs a similar block for its pattern
      <context-param>
        <param-name>PARAMETER_ENCODING</param-name>
        <param-value>UTF-8</param-value>
      </context-param>If you want to be 100% sure the encoding is set right make sure thepages contain:
    <%@ page contentType="text/html;charset=utf-8"%>Or depending on your view technology the syntax can be a bit different
    -Anton

  • Problem with national characters on windows client

    Hello there,
    I'am having problem with national characters on windows client.
    All national data stored in NVARCHAR2 colums, applications (.net) works fine,
    but in sqlplus:
    select city from test_table;
    - everything ok, sqlplus shows national characters
    select dump(N'<national symbols>') from dual
    - returns
    Typ=96 Len=12: 0,191,0,191,0,191,0,191,0,191,0,191
    select * from test_table where city = N'<national symbols> '
    - always returns nothing
    As i understand the problem in
    sql query text (and national literals) convertion
    to servers "WE8ISO8859P1" encoding, Is it possible
    to solve the issue?
    Thanks in advance
    PS.
    Console in right mode (chcp=1251)
    sqlplus shows russian messages well
    Server (oracle 9 on solaris):
    select * from nls_database_parameters
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_SAVED_NCHAR_CS WE8ISO8859P1
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8ISO8859P1
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZH:TZM
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZH:TZM
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_RDBMS_VERSION 9.2.0.6.0
    Client (windows server 2003, oracle client 10):
    NLS_LANG = RUSSIAN_CIS.CL8MSWIN1251

    N'<national symbols>', being part of an SQL statement, will be converted to the database character set (WE8ISO8859P1) before being parsed. Only if the client and the database are both 10.2 or higher, the client can encode the literal appropriately so that it survives this conversion.
    In earlier versions, you can do the encoding yourself. Instead of the N'<national symbols>' literal use the UNISTR function: UNISTR('\xxxx\yyyy\zzzz'), where U+xxxx, U+yyyy, U+zzzz are Unicode code points of your national characters.
    -- Sergiusz

  • Application Hudba (Music) problem with Czech characters

    Hello,
    I have problem with Czech characters for Music files copied into application "Hudba" all Czech letters (š,č,ť...etc.) are translated into Japonies letters (which I could not read at all). Could someone help please.
    Thanks

     You might to consider contacting the app developers http://talk.sonymobile.com/t5/Music/Walkman-Beta-Test/m-p/953659#U953659

  • Problems with Spanish language

    Hello!
    i'm connecting my DB with PHP, my problem comes when i try to capture ñ's or accents (á, é, í, ó, ú ). When i capture these character from my APEX they appear:
    20 fátima 20
    But, when i insert info from PHP the characters don't appear:
    15 f�tima 20
    My NLS_CHARACTERSET value on Oracle is AL32UTF8 and
    select * from nls_session_parameters;
    NLS_LANGUAGE MEXICAN SPANISH
    NLS_TERRITORY MEXICO
    NLS_CURRENCY $
    NLS_ISO_CURRENCY MEXICO
    NLS_NUMERIC_CHARACTERS .,
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD/MM/RR
    NLS_DATE_LANGUAGE MEXICAN SPANISH
    NLS_SORT WEST_EUROPEAN
    NLS_TIME_FORMAT HH24:MI:SSXFF
    NLS_TIMESTAMP_FORMAT DD/MM/RR HH24:MI:SSXFF
    NLS_TIME_TZ_FORMAT HH24:MI:SSXFF TZR
    NLS_TIMESTAMP_TZ_FORMAT DD/MM/RR HH24:MI:SSXFF TZR
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    select * from nls_instance_parameters;
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_SORT -
    NLS_DATE_LANGUAGE -
    NLS_DATE_FORMAT -
    NLS_CURRENCY -
    NLS_NUMERIC_CHARACTERS -
    NLS_ISO_CURRENCY -
    NLS_CALENDAR -
    NLS_TIME_FORMAT -
    NLS_TIMESTAMP_FORMAT -
    NLS_TIME_TZ_FORMAT -
    NLS_TIMESTAMP_TZ_FORMAT -
    NLS_DUAL_CURRENCY -
    NLS_COMP -
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    select * from nls_database_parameters;
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET AL32UTF8
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_RDBMS_VERSION 10.2.0.1.0
    PHP has this sentence but it has ; (comment)
    ;default_charset = "iso-8859-1"
    Is this problem of ORacle or PHP, and what should i change?
    Thank u!

    Hi Rafael,
    I'm sure my friend Paolo is correct (+5 points buddy)
    I ran into this problem with the Swedish and German language files as well :(
    Swedish;
    http://forum.cisco.com/eforum/servlet/NetProf?page=netprof&forum=Unified%20Communications%20and%20Video&topic=Unified%20Communications%20Applications&topicID=.ee835d2&CommCmd=MB%3Fcmd%3Dpass_through%26location%3Doutline%40%5E1%40%40.2cd33d55/0#selected_message
    German;
    http://forum.cisco.com/eforum/servlet/NetProf?page=netprof&forum=Unified%20Communications%20and%20Video&topic=IP%20Telephony&CommCmd=MB%3Fcmd%3Dpass_through%26location%3Doutline%40%5E1%40%40.2cd4047b/0#selected_message
    Cheers!
    Rob

  • EPMA dimension import issue with spanish characters

    We are using special characters for spanish words, like the word Español.
    When this was imported into EPMA using the Import Profile and importing a flat file, EPMA added the alias member as espan then showed a square box and then the l. Then, when deploying the app, no errors recorded in EPMA but planning showed errors since SQL query could not deal with this strange character (the square box symbol) in its query. When I manually change the member in the dimension library to Español, then the app deploys. So my relational database clearly accepts the characters.
    So, the problem is in the Import of the text file when it has the special character ñ or other spanish characters with the accent marks or tildes. Is there something I need to do when importing a file with this character or other spanish language special characters?
    Thanks
    using V11.1.1.3 EPMA with weblogic
    Jeff

    Hi!
    We have the same problem. Please somebody could help us?
    Thanks in advance!
    Analía

  • External Microsoft keyboard with Spanish characters

    I have a Macbook 2,1 (late 2007) white running 10.6.8.  Have been having reoccurring kernelpanic problems (apparently associated with Skype and video) and Bluetooth connectivitybeing dropped intermittently.  Havedone all of the research that I can on both of these problems and have pursuedalmost every one of the suggestions found on the various support groups toaddress these problems.  Havefinally given up.  One of thefeatures that I’m giving up is Bluetooth, so I am now using a wired mouse(Logitech) and just purchased a Microsoft Wired Keyboard 600 here in Seville,Spain, that claims to be Mac-compatible. This keyboard is a Spanish keyboard.  I have installed the driver software for the Microsoftkeyboard for a Mac, available on the Microsoft support sight.  However, I have not been able to getany of the new or relocated keys to work correctly.  My system does not recognize the placement of the variouskeys and symbols provided by the new keyboard and instead, simply inputs thestandard Apple keyboard key symbols. I’ve worked with Language and Text preferences and the keyboardpreferences but to no avail.  Isthere some way that I can make the keys on the Microsoft keyboard workcorrectly?  Thanks in advance forany help that anyone can provide.

    Is one of your problems that the characters typed do not match what is on the keyboard?  In that case you may need to install a special keyboard layout, because PC keyboards are slightly different than the Mac versions.  If you email me I can send you one from Logitech to try (tom at bluesky dot org).

  • Problem with some characters in complex objects

    Hi all,
    I've built a webservice which returns a complex object with several fields inside. All fields are public and accessable via getter and setter methods.
    The problem is, that some of these fields contains numbers or underscores in their names.
    For example:
    public int field_a;
    or
    public String house3of4;
    When I try to import these webservice as a model in a Web Dynpro project, it doesn't work until I remove these characters.
    Is this a known problem or is there any solution for it?
    Thanks
    Thomas

    NLS_LANG in registry is "ARABIC_UNITED ARAB EMIRATES.AR8MSWIN1256"
    I use oracle form 10g for developer
    oracle form 9i for database
    when I build a form in client side and make a text with farsi characters, when I run the form,all characters shows me correct in farsi except four characters(گ چ ژ پ)

  • Problems with Turkish characters not displaying correctly

    A user is having problems with certain Turkish characters in particular fonts not displaying correctly (as shown below).
    The two fonts (Gill Sans and Helvetica LT CondensedLight) are used for English language marketing material.  Ideally, for continuity, the same font would be preferred for our Turkish language marketing material rather than changing it to a similar font.
    Helvetica LT CondensedLight
    Turkish character “ İ ” does not display correctly in either Word or InDesign.
    Gill Sans
    Turkish characters “ ş ” and “ ğ ” display correctly in Word but not InDesign.
    My questions are:
    1. For Gill Sans, why are the characters not displaying in InDesign but are in Word?
    2. I've noticed here: http://www.webtype.com/font/gill-sans-family/#glyphs-tab that the required characters are available within the Gill Sans font.  Is this because the font is newer and includes the additional characters?  Would this work in InDesign?

    One of the big differences between InDesign and Word is that Word will happily substitute glyphs without telling you. So, unless your installs of Word and of InDesign are using different versions of Gill Sans (which is possible, BTW) then your s-with-cedilla and your g-with-macron are actually not in Gill Sans but in some other font. I'm guessing Arial.
    It's possible that you have more than one Gill Sans installed (e.g. one in your Document Fonts folder and another in your system fonts folder) that have different glyph complements. Gill Sans has been around a long time, and the version I have from the 90s has no Turkish support whatsoever.
    As far as the capital-I-with-dot, you can check your glyph coverage in a variety of ways. In both InDesign and Word, you can open up a window that shows you all of the glyphs in a font. In InDesign, find it in Type -> Glyphs. In Word, it's called "Insert Symbol" and you can find it in Word 2010 by going to Insert -> Symbol -> More Symbols. I am guessing that the answer to this question is simply that your cut of Helvetica LT has no cap-I-with-dot. You'd need to pick a font that actually has support for Turkish. On WIndows, I use a very full-featured freebie called BabelMap to check font coverage. They have a Web version here:
    http://www.babelstone.co.uk/Unicode/babelmap.html
    but I personally prefer the downloadable .exe file.

  • File XML Content Conversion: Problem with special characters

    Hello,
    in a file sender cc content conversion is used to transform a flat structure to XML. What we experiencecd is that the message mapping failed due to a character that was not allowed in XML:
    I was assuming that the file content conversion just creates XML messages with allowed characters. Is there any way to configure content conversion to remove control characters which are not allowed in XML? Unfortunately the sender system cannot be modified.
    Thank you.

    Hi Florian,
      Please use this UDF to remove special characters which prevent XML messages to form properly.
    public static String removeSpecialChar(String s)
              try
                   s=s.replaceAll("&","& amp ;");
                   s=s.replaceAll("<"  , "  & lt ;");
                   s=s.replaceAll(">", "& gt ;");
                   s=s.replaceAll("'", "& apos ;");
                   s=s.replaceAll("\"", "& quot ;");
              catch(Exception e)
                   e.printStackTrace();
              return s;
    Please remove spaces between characters within double quotes. I have added them because otherwise you can't see this code properly. Please check this below link , please replace the characters with proper values as the display is causing a problem here   
    http://support.microsoft.com/kb/316063
    regards
    Anupam
    Edited by: anupamsap on Jul 7, 2011 4:22 PM
    Edited by: anupamsap on Jul 7, 2011 4:23 PM

  • Problem with Polish Characters while downloading data using GUI_DOWLOAD

    Hi,
    I was trying to create a file on the presentation server using GUI_DOWLOAD.
    I am facing a problem with the data that contains polish characters.. like Ą , Ł..
    There is  a vendor numer in the system as FRĄCZEK, when i was tring to download, the vendor number is downloaded as FR¥CZEK, even i tried to download from SE16 also, still i am facing the same problem.
    I am facing the above problem with soem other polish characters..
    Is there any solution  to download the data properly.. ?
    Guys please help me.
    Thanks in advacnce.
    Regards,
    Sriram.
    Edited by: Srirama Murthy Maddirala on Oct 27, 2009 8:10 AM

    Hi Sriram,
    <li>Get the front-end code page using NLS_GET_FRONTEND_CP.
    <li>Pass the code page through GUI_DOWNLOAD.
    REPORT  ztest.
    TYPE-POOLS:abap.
    DATA:frontend_codepage TYPE  cpcodepage.
    DATA:codepage          TYPE  abap_encoding.
    "Get the Frontend Code page
    CALL FUNCTION 'NLS_GET_FRONTEND_CP'
      EXPORTING
        langu             = sy-langu
      IMPORTING
        frontend_codepage = frontend_codepage.
    "Pass the codepage through GUI_DOWNLOAD function module
    codepage = frontend_codepage.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
        filename = 'C:\temp\data.xls'
        filetype = 'ASC'
        codepage = codepage
      TABLES
        data_tab = it_data.
    Thanks
    Venkat.O

  • Problems with danish characters in email

    Hi,
    I have created a mailform with danish characters. I have tried to use the following configuration:
    charset=iso-8859-1
    But when i save the mailform, CRM transforms it to:
    charset=utf-8
    And before the header line it inserts:
    <?xml version="1.0" encoding="utf-8"?>
    This is a problem because some of oure customers email clients, can't read this and garbles the danish special characters. The result is non readable emails.
    How do i fix this?
    Kind Regards
    Michael
    Edited by: Michael Knudsen on Oct 16, 2008 11:21 AM

    Hi Michael,
    pls create an OSS for this to allow SAP development to check this for you.
    regards
    martin

Maybe you are looking for