Hexidecimal Non-Allowed Characters in a Unicode System

We have a function module that we've written to replace non-permitted characters with a space in transfer rules.  We see a lot of invisible hexidecimal characters coming in free form text fields.  This work great for English.  However, we have a Unicode system with other languages installed.  We are also getting the hex characters in other character sets. 
Has anyone dealt with this issue and if so what was you solution?
Thanks!
Al

Hello aLaN,
how r u ?
Hey we have faced the problem with Hexadecimal characters, but not the same issue. In our case the problem was in the Source System. In the DB Tables we had some unwanted characters, that was showing some errors while data loading, particularly ERROR 18.
So we resolved it by changing the Source System data.
I have already posted for the hexadecimal issue.... the replies was
I think this is related to Invalid character issues.. or SPaces setting in RSKC..
may be you want to look at eh following post..
Re: invalid characters
/people/siegfried.szameitat/blog/2005/07/18/text-infoobjects-part-1
Example:
let us say..
1. Check in RSKC for allowed characters..
2. Add a code in the update rule to restrict the texts contains..
!"%&''()*+,-./:;<=>?_0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ' are allowed characters in RSKC transaction then other than the above character is 'Invalid' including the smaller case letters and will throw the hex.. error.
since this is from database system even 'NULL' datatype from there is not visble to the eyes and can cause the failure.
Hope this helps
Best Regards....
Sankar Kumar
91 98403 47141

Similar Messages

  • Replacing characters in a Unicode system

    I now need to search all imported texts and replace this `` with a regular double-quote ". Now, I know how to do this in non-Unicode systems, but am totally lost when it comes to UC.
    Could someone please help? TIA - Roman.
    P.S. Never mind, I got it...

    Using convert method from cl_abap_conv_in_ce solves the problem.

  • Converting non-English characters to their unicode representation

    I have series of files/templates where each contains a locale specific language such as Chinese, Japanese or German. I need to find out how do I get their unicode representations so I can send as html formatted email?
    I can already send one for the English template as html formatted email w/out a problem. I was able to find a sample of unicode representation of Japanese and send that as a test. But how do I get the temaplates that I have and convert their contents into unicode?
    Thanks in advance.
    please dis-regard. I figured it out.
    chehrehk

    You need to know what character encoding was used for the template text.
    For example, you could have Japanese text encoded using UTF-8 or
    encoded using ISO-2022-JP and the same Japanese characters would
    be represented as a different sequence of bytes. Without knowing which
    charset was used, you won't be able to convert the byte sequence back
    into Unicode characters (e.g., to store in a Java String).
    If you do know which charset was used, java.io.Reader will convert the
    byte stream into Unicode characters.
    If the charset information is not available, there are heuristics that you
    can use to try to guess the correct charset, but by their nature they're
    going to be wrong sometimes.

  • Printing of chinese characters in ecc6 unicode system

    Hi,
    We are having problem printing out debit memo that have chinese characters. In the preview, the chinese characters are displayed properly.
    but when it's printed out, the chinese characters are displayed as "###". we are using device type "SAPWIN: Rel.4.x/SAPIpd 4.09+".
    When we use device type "CNSAPWIN", the chinese characters are displayed as "ÂíÀ´Î÷ÑÇ".
    Appreciate if anyone can tell me how to solve this problem.

    Hi,
    It is a printer issue. The printer needs to be configured to support the Chinese chanracter set.
    Normally when we are able to view the chinese characters properly in the print preview but not properly displayed on the output it is surely a printer problem.
    This kind of problem we faced many times.
    Thanks
    Ram

  • Pdf conversion from a non unicode system to a unicode system

    Hi,
    I have a BSP application that runs on a unicode system which retrieves data to be output in pdf from a non unicode sap system.
    We get the data from the non unicode system through a RFC call and store data in table otfdata type itcoo, then we perform the conversion through the following:
    perform pdfcnv_convertotf in program saplstxw
    tables otfdata
    pdf
    using
    'T'
    archive_index
    1
    'T'
    pdf_bytecount
    app_tot_xstring
    and then perform the hex conversion, but the printout is full of # chars. The good characters are followed by the # char. It looks like the unicode conversion is not good.
    Thank you in advance for your help.
    Best regards,
    Donatella

    Welcome to SDN
    check out this weblog where Thomas Jung explains unicode conversion for excel, the same thing could be applied in your case as well
    /people/thomas.jung3/blog/2004/08/09/bsp-download-to-excel-in-unicode-format
    Regards
    Raja

  • PDF conversion for chineese characters in Unicode system

    I am facing a problem while converting the SAP Script Output to PDF format for Chinese characters. 
    I am working on ECC (5.0) Unicode system.
    Scenario:
    After saving a Purchase Order an E-mail is sent to the customer - attaching the
    PO output in PDF format.  E-mail was received successfully by the receiver, but while opening the pdf all the chinese characters were displayed in junk characters in the pdf. All the English characters are properly displayed.  I tried to open the pdf file in Acrobat Reader versions 6.0, 7.0, 8.0. but no result.  I used CONVERT_OTF function module for converting the OTF format to PDF format. I tried using the fonts CNSONG also.
    I tried by executing the standard program RSTXPDFT4 for converting to PDF by giving the spool.  In the spool it is showing the Chinese characters perfectly but in the PDF the Chinese characters were were showing as Junk.
    Can you please help and advice to see the Chinese characters in PDF in Unicode systems.
    Thanks in advance.

    >
    Juraj Danko wrote:
    > Hi,
    > I have similar problem than you ... how have you solved it?
    > thanks
    > Juraj
    I found a solution, but I am not sure, if it was for this problem or
    output problem with for example PL in non-unicode systems.
    I created the input for CONVERT_OTF with CALL FUNCTION 'PRINT_TEXT'.
    PRINT_TEXT has to be called with DEVICE = 'PRINTER',
    DEVICE = 'ABAP' uses internally the wrong code page.
    You have also to set otf_options-tdprinter to a valid printer,
    if it is empty, the default printer from user settings is used.
    You can use code example from SAP note 413295.
    Before you call CONVERT_OTF, you can also check entries with 'FC' in OTF input.
    The font (see description of OTF format in SAP help) must be set like described in SAP note 144718.
    /Tibor
    Edited by: Tibor Gerke on Jan 13, 2011 10:29 AM

  • Unable to display Chinese character in non-unicode system properly.

    Hi, all
    My system has been upgraded from 4.6C to 6.0 which is non-unicode system. Chinese characters with long length will be cut and cannot wrap automatically. But it's OK to alphabetes. How could I resolve this?
    Thanks..
    TD. Wei

    The simple and easy way to check
    non-unicode character sets are not supported on unicode system any longer. Am I right?
    Transaction code i18N
    Select
    trouble shooting --> printing  test --> smartforms --> multiple scripts, select your output device and see print preview. it will display all supported characters.
    I guess, above information will be useful for closing the thread.
    Regards,
    SaiRam

  • Non-whitespace characters are not allowed ...

    I'm trying to parse the following XML fragment into a DOM Document object:
    <CNetSpecs><section name="General
    "><spec name="Printer Type
    ">Personal printer - ink-jet - colour
    </spec><spec name="Weight
    ">6.4 kg
    </spec></section><section name="Printer
    "><spec name="Print Speed
    ">Up to 15 ppm - black draft ? Up to 15 ppm - colour draft ? Up to 0.65 min/page - photo best - 101.6 x 152.4 mm ? Up to 1.28 min/page - photo best - 203 x 254 mm
    </spec><spec name="Built-in Devices
    ">Preview screen
    </spec><spec name="Max Resolution ( B&W )
    ">5760 dpi x 1440 dpi
    </spec><spec name="Max Resolution ( Colour )
    ">5760 dpi x 1440 dpi
    </spec></section><section name="Software / System Requirements
    "><spec name="OS Required
    ">Microsoft Windows 98 Second Edition / Windows ME, Apple MacOS X 10.1, Apple MacOS X 10.2, Microsoft Windows 2000 / XP, Apple MacOS 8.6 - 9.2
    </spec></section></CNetSpecs>The fragment is pulled out of a text field in a database.
    The code I'm using to do the parseing is:
    // Parse the CNet XML into a proper doc.
    DOMParser parser = new DOMParser();
    parser.parse(new InputSource(new StringReader(cnetSpecs)));
    org.w3c.dom.Document cnetXml = parser.getDocument();
    xmlUtil.toFile(cnetXml, "/tmp/cnetxml.xml");
    NodeList cnetRoot = cnetXml.getElementsByTagName("CNetSpecs");
    Node CNetStatsNode = xml.importNode(cnetRoot.item(0), true);
    root.appendChild(CNetStatsNode);and I get the following errors:
    [Error] :1:107: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'Personal printer - ink-jet - colour'.
    [Error] :1:150: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw '6.4 kg'.
    [Error] :1:391: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'Up to 15 ppm - black draft ? Up to 15 ppm - colour draft ? Up to 0.65 min/page - photo best - 101.6 x 152.4 mm ? Up to 1.28 min/page - photo best - 203 x 254 mm'.
    [Error] :1:452: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'Preview screen'.
    [Error] :1:528: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw '5760 dpi x 1440 dpi'.
    [Error] :1:603: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw '5760 dpi x 1440 dpi'.
    [Error] :1:847: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'Microsoft Windows 98 Second Edition / Windows ME, Apple MacOS X 10.1, Apple MacOS X 10.2, Microsoft Windows 2000 / XP, Apple MacOS 8.6 - 9.2'.Can anyone point me in the right direction? I've been scratching my head all day.

    I dont think so that it is because of un-escaped "&" character or breaking line sequences.
    As in mine case it is giving for XML as below:
    <?xml version="1.0" encoding="UTF-8"?>
    <publicationManifest date="03/13/2007 14:06:25">
    <contentItem action="add" expireDate="" releaseDate="" revision="1">
    <content>This would go under webstore</content>
    <metadata>This would go for Database</metadata>
    </contentItem>
    </publicationManifest>
    and console output at java IDE is
    [Error] item-manifest_Test.xml:4:46: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'This would go under webstore'.
    [Error] item-manifest_Test.xml:5:45: s4s-elt-character: Non-whitespace characters are not allowed in schema elements other than 'xs:appinfo' and 'xs:documentation'. Saw 'This would go for Database'.
    Therefore , if anyone had found the root cause of this discrepancy please update this topic.

  • Problem publishing database contents from non-unicode to unicode system

    Hello everyone!
    We just set up a new SAP WAS based on Netweaver 2004 as a unicode system. Out problem now is that we have a content management system on our non-unicode system and that we are publishing the contents via rfc to the WAS unicode system to display the contents online. the contents are stored in our own database tables.
    The problem thereby is that many texts pasted from microsoft word contain special characters like bullets, long minus or low-9 quotation marks which are not correctly displayed in the unicode system / on the website. we already found out that it has something to do with the codepage. the sap notes say we should use 1160 instead of 1100 and that the transaction SPUMG would be helpful. but we are not able to select any tables there.
    so now we do not know what to do exactly. do we have to change something in our non-unicode system or do we have to conversion in our unicode system. and what happends if content containing special microsoft word characters is published after the spumg conversion? do we have to to this frequently?
    We would be glad if anyone could help.
    Thanks a lot!

    Hi Martin,
    thanks for your quick answer.
    You got me right. We have a local non-Unicode SAP HCM Netweaver 2004 system running a self-developed web based content management system / wiki. The texts entered in the bsp application are stored in a string field in our database table. Actually we publish the contents to a WAS 6.20 non-Unicode system with the same database tables to provide the content via BSP for the public. Everything is working fine including the special characters.
    Now we want to replace the WAS 6.20 non-Unicode system by a new WAS 7.0/2004 Unicode system. But when publishing the contents via the same RFC function module to the new system the special characters seem to be damaged. We are not able to replace them with abap commands and when they are displayed on the website we only see "boxes".
    If I get you right we have to run SPUMG on our nw 2004 non-unicode productive hcm system, right? but isn't there a danger to damage existing contents?
    Best regards,
    Stefan

  • CP850_BIN2 for non unicode systems??

    Hi All
    This is possibly a really stupid question and I think I know the answer, but I need some one else to say yes!
    We are currently running a SQL upgrade, from 2000 to 2005, for ore ECC6 platform. As part of the upgrade it is recommended that you change the DB collation from SQL_Latin1_General_CP850_BIN to SQL_Latin1_General_CP850_BIN2 which allows unicode data to be sorted properly.
    My question is, if this enhancement is to correct a unicode issue, will it affect my current non-unicode system if applied. I don't want it to cause any issues if I apply it during the upgrade.
    I don't think it will for various reasons, but I am not 100% can anone let me know if I am right, please!
    Thanks in advance!
    Phil

    Hi,
    you will have no issues with your current system when you go to BIN2 now (before you upgrade to SQL 2005). And you have to keep in mind, that you have to have BIN2 for all newer releases of SAP. So if you plan to ever upgrade your system to a newer release in some point in time, you should go to BIN2 now, as otherwiese you have to perform a homogenous systemcopy with R/3 (which is not so much fun) if you miss it now. On SQL 2005 you can't change the collation easily, so if you don't have the correct one, you only can change it when you unload the data, drop the DB, create the DB in the correct collation and reload the data.
    Regards
      Clas

  • Issue with Data flow between Unicode and Non Unicode systems

    Hello,
    I have scenario as below,
    We have  a Unicode – ECC 6.0 and a UTF 7 – Legacy system.
    A message flow between Legacy system to ECC 6.0 system and the data is of 700 KB size.
    Will there be any issue in this as one is Unicode and other is non Unicode?
    Kindly let me know.
    Thanks & Regards
    Vivek

    Hi,
    To add to Mike's post...
    You indicate that your legacy system is non-Unicode and the ERP system is Unicode.  You also said that the data flow is only <i>from</i> the legacy system <i>to</i> the ERP system.  In this case, you should have no data issues, since the Unicode system is the receiving system.  There <b>are</b> data issues when the data flow is in the other direction: <i>from</i> a Unicode system <i>to</i> a non-Unicode system.  Here, the non-Unicode system can only process characters that exist on its codepage and care must be taken from sending systems to ensure that they only send characters that are on the receiving system's codepage (as Mike says above).
    Best Regards,
    Matt

  • Double byte language i.e Japanese or Chinese text in non Unicode System

    Hi,
    I have translated text into Chinese and Japanese in a Unicode system and want to move it into a non Unicode system. Would Chinese/Japanese characters display correctly in a non Unicode system when moved from Unicode system.  I am doing translation in ECC60 or SAP 4.7 Unicode system and moving to SAP 4.7 non Unicode system.
    Thanks
    Balakrishna

    Hi Balakrishna,
    in general the transport between Unicode and Non-Unicode systems is supported.
    However there are restrictions, which are outlined in SAP note 638357.
    In your case it is a prerequisite that the objects to be transported are language dependent (text lang. flag is set on the language key - see SAP note 480671) and the languages are properly setup in the target systems.
    For double byte data there is a specific issue when transferring data from Unicode to Non-Unicode:
    In a Non-Unicode system, one double-byte character needs two bytes, therefore e.g. in a 10 char field, 5 double byte chars are fitting. In a Unicode system, you can insert 10 double-byte chars in a 10 char field.  Hence there is a risk of truncating characters in case of Unicode --> Non-Unicode communication.
    Please also have a look at SAP notes 1322715 and 745030.
    Best regards,
    Nils

  • Error during Upgradation SCM 5.0 to SCM 5.1 Non-Unicode System

    Hi,
    I am Upgrading SCM 5.0 Non-Unicode System to SCM 5.1 Non-unicode on Windows 2003 Server with MSS 2005 DB. During the initial PREPARE (EXTRACTKRN_PRE) phase I am getting an error like
    init
    2010/06/15 07:03:45 PREPARE: START OF PHASE EXTRACTKRN_PRE
    ERROR: Cannot find file 'C:\SCM_2007_Upgrade_Media\51033508_8\NW_7.0_SR3_Kernel_WINDOWS__LNX_X86\KN_WINDOWS_X86_64_AUPG
    KERNEL.TOC'
    SEVERE ERROR(S) OCCURRED IN EXTRACTKRN_PRE !!!
    The file name it is looking for does not exist in any of the Kernel versions I have checked. If I try to skip this step it is asking for the PASSWORD. How can I get this password? (only from SAP Support or any note?).
    What is the Solution for the above ERROR. Please help.
    Thanks,
    Ajay.

    > I am not posting the question in the same group. I am posting in different groups due the confusion of in which group to post > > and of course to get faster response. All people cant check all the groups to answer. Moreover I am not keeping the questions > open. As soon as I get the answer I am proving the answer in every group and immediately closing the question.
    > So what is the problem with that? I do not think this is creating any inconvenience to any one, just increasing the number of questions in the group.
    This is not wanted as written in the "Rules of engagement". This is a public forum, not a "help desk".
    > OK, As you said If I post only in a single group do you guarantee that you will check every time all the groups to answer my question/ find my question?
    Again, this is a forum where the question are answered on a volunteered basis. "Demanding" help is not the way to go, you can ask for help but there's no guarantee that you'll get an answer. If you urgently need help use the official support channels or get yourself a consultant on-site who can answer all your question instantly.
    > And please paste the URL of those rules that this is wrong. Since I didnt read that.
    It's the first entry in every forum, it contains the "Rules of Engagement" which say:
    Please do not Cross post.
    Post your question in the most appropriate forum; not multiple forums. This is bad netiquette and will might only aggravate potential repliers.
    Just imagine if everyone like you would paste questions in several forums.
    Markus
    Edited by: Markus Doehr on Jun 15, 2010 5:29 PM
    Edited by: Markus Doehr on Jun 15, 2010 5:30 PM

  • Unable to Open unix file in UNICODE system which created NON-UNICODE system

    Unable to Open unix file in UNICODE system which created in NON-UNICODE system
    We have two SAP systems both are ECC6.0 but System 1 is NON-Unicode and System2 is Unicode system.
    There is a common unix directory/folder for both system.
    Our requirement is to create one file on unix common folder and write the data to file from system1 .
    In system2 open the same file for appending mode to write the data .
    The file in system 1 created with below sentence.
    OPEN DATASET g_unix_file FOR OUTPUT IN TEXT MODE ENCODING UTF-8.
    Now I have to append the data from system 2 to same file.
    I have tried to used below statement in system 2 to open the file but sy-subrc value comes as '8'.
    1> OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING UTF-8.
    2>OPEN DATASET g_unix_file FOR APPENDING IN legacy TEXT MODE CODE PAGE
    cdp IGNORING CONVERSION ERRORS  .
    3>OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING Default.
    4>OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING NON-UNICODE.
    Tried out all the possibilities as per F1 help given for open dataset , but still there is problem with opn file in appending as well output mode.However the file successfully open in Input mode(Read).
    Please advice suggestion to resolve this issue.
    Thanks.

    Messgae captured as 'Permission Denied". The program gets triggered with system user Id PPID.
    How to check the security access of the User ID.

  • Client Copy between unicode system and non-unicode system

    Hello,
    we have to build up a new system for japan. It is planned to install a new 4.7Ext200 unicode system and then make a client copy from a non-unicode 4.7Ext200 (language= german, english) system to the unicode system. I don't think that this is possible way, but I can't find information regarding client copies from non-unicode to unicode. I would advice the project to convert the existing non-unicode system to unicode, make a system copy for the new system and then install japan languages. Any information, which can help me.
    Regards,
    Alexander

    Hi,
    without conversion client copy is not possible
    look at following
    Re: Client copy between unicode and Non-unicode
    regards,
    kaushal

Maybe you are looking for