Open Dataset - unicode system ECC 6.0,special characters

Hi,
we have upgraded our system from 46b to ecc 6.0 unicode. We face following problem in file download to application server(solaris) using open dataset.
in output file hyphen(-) character is by special character â.
eg. if data in table is  'BANGALORE - 560038' is displayed as 'Bangalore â 560034' in application server file. if we download the same file to windows it shows proper data.
only when we open file in vi editor on application server , system shows this special character.
Kindly help.
Regards,
Sidhesh S

OPEN DATASET G_APFILE FOR INPUT IN TEXT MODE.in 4.6                                                                                                    
    OPEN DATASET G_APFILE IN LEGACY TEXT MODE FOR INPUT in ECC6.0
regards
Giridhar

Similar Messages

  • Open dataset (UNICODE) for english en polish characters

    Hi,
    I have a problem on my multi language project on where I need to manage English language and polish language.
    I have a table (that we can call TABLE) I manage the translation on it. When I am connected in English I can see some squares instead some polishes characters. When I am connected in polish, I have the right characters. Any way I don’t think it’s a problem, cause if I copy/paste the square on Word, I find back the corrects characters.
    My problem is after.
    I have a structure with the collon name and an internal table with data.
    I use the module function DDIF_TABL_GET on my table TABLE to have the collon name corresponding to the language wanted.
    I am building a string with that and some carry return.
    (Here in English I have again some square, but always the good character if I paste it on Words.)
    Now I would like to put this string in a file using the good encoding.
    I have tried lot of ‘opening dataset’ and lot of ‘encoding’.
    I have a file on SAP (transaction FILE)
    (Here in English I have again some square, but always the good character if I paste it on Words.)
    If I download this file on my computer, I have wrong characters.
    Could I have help?
    My wrong code:
      OPEN DATASET lv_filename FOR OUTPUT IN TEXT MODE ENCODING UTF-8.
      TRANSFER u_contenu TO lv_filename.
      CLOSE DATASET lv_filename.

    Hi,
       Use this
      OPEN DATASET File_path  FOR OUTPUT IN TEXT MODE ENCODING NON-UNICODE
      WITH SMART LINEFEED
    it will definitely work.
    Regards,
    Manesh. R

  • Unicode display with system fonts?? (special characters)

    Hi All. We're working on site that supports a couple
    different languages: English, Japanese, and now German. We're using
    dynamic text boxes and Remoting to display internationalized text
    where necessary. Remoting makes a call to the backend, which
    returns a text blurb depending on the country ID that's in use and
    then the result is tossed up into the dynamic text box on screen.
    Simple as pie. Works great for English, and great for Japanese. But
    now we've added German, which I thought would be painless as it's
    another western language.
    Apparently, Flash is having problems displaying special
    characters like the umlaut. In place of the special character, we
    see a black diamond with a question mark. Now normally I'd think
    that the font we're using doesn't have the embedded symbol for the
    umlaut character but the thing is, we're not using embedded fonts
    -- we're using system fonts (verdana) and everything works perfect
    for the Japanese version. What in the world is going on?
    If I dump the results of the remoting call to an html page,
    everything displays as it should so I know the data in the database
    isn't improperly encoded or anything. And Japanese works! WTF? Any
    ideas PLEASE let me know, been banging my head against this for
    waaay too long now...
    Danke
    .

    Alright so if I change the data directly, and make the
    a-umlaut into Ġ everything renders in Flash perfectly.
    Rockin. So I patch my input cms code to run htmlentities() on the
    text area. Instead of translating a-umlaut to Ġ though it
    translates it to ä which Flash's html renderer doesn't
    display correctly... Uh..

  • Weird problem in open dataset

    Hi mate,
    I have strange problem with open dataset. I try to download data to server and hit by shortdump telling me that file is not yet opened. I have checked SY-SUBRC (equals zero) after statement open dataset.
    It just happened today, previously i never encounter these problem before.
    Can someone point out me what the problem is? is it because of some delay in opening dataset or system performance issue? Production is running pretty slow in peak hours around 10-13PM.
    Thanks for your input.
    alia

    Hi Alia,
    Please check SAP OSS Note <b>733626</b>. It seems you don't have enough space in application server when the system tries to generate the input file. Perhaps you may need to archive production files.
    <b>Symptom</b>
    Report RCCLBI03 classifies all objects that are contained in an input file. Incorrect records from the input file are written into an error file.
    The report generates a termination with error code DATASET_CANT_OPEN if an object must be really written into the error file.
    <b>Other terms</b>
    Classification, dump
    <b>Reason and Prerequisites</b>
    The check whether the error file can be opened is insufficient. In the standard case this is RCCLBI03.ERROR on the application server.
    <b>Solution</b>
    If the error file cannot be opened, the report now ends with error message LX105 ("Unable to open file &1"). Then the input file is no longer imported.
    Hoep this will help.
    Regards,
    Ferry Lianto

  • Double byte language i.e Japanese or Chinese text in non Unicode System

    Hi,
    I have translated text into Chinese and Japanese in a Unicode system and want to move it into a non Unicode system. Would Chinese/Japanese characters display correctly in a non Unicode system when moved from Unicode system.  I am doing translation in ECC60 or SAP 4.7 Unicode system and moving to SAP 4.7 non Unicode system.
    Thanks
    Balakrishna

    Hi Balakrishna,
    in general the transport between Unicode and Non-Unicode systems is supported.
    However there are restrictions, which are outlined in SAP note 638357.
    In your case it is a prerequisite that the objects to be transported are language dependent (text lang. flag is set on the language key - see SAP note 480671) and the languages are properly setup in the target systems.
    For double byte data there is a specific issue when transferring data from Unicode to Non-Unicode:
    In a Non-Unicode system, one double-byte character needs two bytes, therefore e.g. in a 10 char field, 5 double byte chars are fitting. In a Unicode system, you can insert 10 double-byte chars in a 10 char field.  Hence there is a risk of truncating characters in case of Unicode --> Non-Unicode communication.
    Please also have a look at SAP notes 1322715 and 745030.
    Best regards,
    Nils

  • Error during data load due to special characters in source data

    Hi Experts,
    We are trying to load Billing data into the BW using the billing item datasource. It seems that there are some special characters in the source data. When the record with these characters is encountered, the request turns red and the package is not loaded even into the PSA. The error we get in the monitor is something like
    'RECORD 5028: Contents from field ****  cannot be converted into type CURR',
    where the field **** is a key figure of type currency. We managed to identify the said record in RSA3 on the source system and found that one of the fields contains some invalid (special) characters that show up as squares in RSA3. The data in the rest of the fields, including the fields mentioned in the error  looks correct.
    Our source system is a non-unicode system wheras the BW system is unicode enabled. I figure that the data in the rest of the fields are getting misaligned due to the presence of the invalid characters in the above field. This was confirmed when we unassigned the field with the special characters from the transfer rules and removed the source field from the transfer structure. After doing this the data was loaded successfully and the request turned green.
    Can anyone suggest a way to either filter out such invalid characters from the source data or make some settings in the BW systems such that the special characters are no longer invalid in the BW system? We cannot write code in the transfer rules because the data package does not even come into the PSA. Is there any other method to solve this problem?
    Regards,
    Ted

    Hi Thad,
    I was wondering, if system is unicode or non unicode that should not matter the amount and currency field . As currencies are defined by SAP and it is in pure English at least a currency code part of 3 Chars. 
    Could this because of some incosistency of data ??
    I would like to know for Which currency  had some special characters it it in that particular record ??
    Hope that helps.
    Regards
    Mr Kapadia

  • File Content Conversion (receiver) and special characters

    Hi all,
    I have a scenario that has a file receiver channel with content conversion. The record structure in the flat file is field-width delimited (hence no field separator) and the parameter 'fieldLengthTooShortHandling' has the value 'Cut' because the receiving system needs only specific widths for the fields. Hence if the field value exceeds the length permitted, the extra characters are clipped.
    I observed that some characters are not handled properly while creating the text file. For example, one of the fields contained a "minus" character (not the hyphen). The flat file was created successfully. I opened the file in notepad and found that the "minus" character has appeared correctly and the column count in that record was as expected. However when the same file was opened in Textpad, the minus character was displayed as â | |  ('a' with caret, bar, bar) So, all the fields after this field were shifted ahead by 2 characters and hence the total column count of the record had gone beyond the actual one.
    All this started due to the error reported by the receiver system which processes the flat file. Due to shift of characters in the flat file, the processing failed. Moreover that system cannot process the special characters (like minus or non-Latin accented characters etc.) So although there is no issue in XI interface as such, I just want to know if anyone has more informtion on why the characters are displayed differently as mentioned above.
    Regards,
    Shankar

    Define data type like
    order_recordset
    order_row 1..unbound
    f1
    f2
    All are same except communication channel configuration,
    Message Protcol : File content select, then below you got additional parameters.
    there you fill
    Document name : Your sender message type.
    Document namespace : Give your scenario namespace
    Recordset name : order_recordset ( mentioned in the data type)
    Recordset structure : order_row, *
    Name Value
    order_recordset.fieldSeparator : 'nl'
    order_row.fieldSeparator : ,
    order_row.endSeparator : 'nl'
    based on your text file you fill the above parameter values.

  • Special Characters Report Painter

    Hi all,
    After SP 2008 implementation on ECC 5.0, special characters as : "( " ; "= "; "+" ; etc... are not recognized for one digit in report painter library.
    After analyse, without changing size colums, if the title contain a special character, it is counting for  2 digits instead of one. The report painter doesn't truncate automatically the title and show a blank zone !
    Is there an OSS note which correct this problem ?
    (Our level of package is SAPKH519)
    Thanks for your help !
    Edited by: Lau bel on Jul 3, 2008 11:00 AM

    Hi Aday,
    Please check this links perhap it may help.
    Report Writer:
    http://help.sap.com/saphelp_erp2005/helpdata/en/5b/d22ba743c611d182b30000e829fbfe/frameset.htm
    Report Painter:
    http://help.sap.com/saphelp_47x200/helpdata/en/66/bc7d2543c211d182b30000e829fbfe/content.htm
    http://www.virtuosollc.com/PDF/Get_Reporter.pdf
    http://www.sap-img.com/zf028.htm
    http://www.sap-img.com/zf027.htm
    Regards,
    Ferry Lianto

  • Open dataset twice once for input and once for output in unicode system

    Hi All,
    In a program
    I used a open dataset to read the data from the file.
    OPEN DATASET cmp_file FOR INPUT IN TEXT MODE
          ENCODING NON-UNICODE.
    Then i closed the file.
    Again later in the program,
    I used a open dataset to transfer the data.
    OPEN DATASET cmp_file FOR OUTPUT IN TEXT MODE
          ENCODING NON-UNICODE.
    But this time I get sy-subrc = 8.
    Unable to open the file and subsequent TRANSFER is leading to the runtime error.
    Note : I am using a unicode system
    I could run the same program well in non-unicode system..
    ->
    Is it that if a file contains data already
    1.I need to delete the data and open it
    or
    2.I need to open in APPENDING mode manadatorily...
    in case of UNICODE system..
    Kindly suggest..

    Hi,
    IF you have write permission al S.O. Level you need to check your DATASET rigths using AUTHORITY_CHECK_DATASET, this validate your rigths with S_DATASET object.
    Example
    TYPE-POOLS SABC.
    CALL FUNCTION 'AUTHORITY_CHECK_DATASET'
    EXPORTING PROGRAM = 'ZDATASET'
    ACTIVITY = SABC_ACT_READ
    FILENAME = '/tmp/sapv01'
    EXCEPTIONS NO_AUTHORITY = 1
    ACTIVITY_UNKNOWN = 2.
    See SABC type pool to know wich activities are aviable.
    Hope this help.
    Regards

  • 'Open Dataset' command - add ASCII code in a unicode system

    Hi,
    I use following comand in order to save a file on an application server. 
    open dataset <path> for output in text mode encoding default.
    Is it possible to convert file to ASCII code instead of the system setting unicode and to ignore conversion errors? If yes, please tell me how.
    Thanks!

    problem exporting text in hebrew
    Similar kind of problem has been solved here.. check that link first
    Use the class CL_ABAP_CONV_OUT_CE and use it to convert file from unicode to ascii.

  • OPEN DATASET...  in a Unicode system

    I'm discussing with a developer @ SAP about the correct use of OPEN DATASET in a Unicode system. I'm not sure I'm correct with my opinion so maybe someone could shed a light on it.
    The source of discussion is the (SAP standard) program RSQUEU01. This program is used to download data from the TemSE; in our case we use it to produce a file that is to be sent to auditors (component FI-AIS).
    Our system is a little endian Unicode system (codpage 4103). If we download the created data using that program we get a dump with "CONVT_CODEPAGE" because a character could not be converted from 4103 to 1100.
    The proposed correction by the developer is changing
    OPEN DATASET EXP_FILENAME FOR OUTPUT IN legacy BINARY MODE.
    to
    OPEN DATASET exp_filename                              
    FOR OUTPUT IN LEGACY BINARY MODE                     
    IGNORING CONVERSION ERRORS.
    I think, that correction is wrong. Since we have 12 languages in the system including some Asian the output might get corrupted.
    When I asked him about that I was told use an application server with a "correct codepage" then - I'm not sure what that means since I can't connect an ASCII application server to a Unicode system.
    I guess the statement should be
    OPEN DATASET EXP_FILENAME FOR OUTPUT IN BINARY MODE ENCODING DEFAULT.
    This makes sure that no data is cut (like doublebyte) and makes sure, the appropriate codepage (LE/BE) is used.
    Are my assumptions right?
    Markus
    (OSS 323320/2010)

    Hi Markus,
    Let's first clarify the difference ways for writing files:
    <ul style="list-style:circle!important;">
    <li>BINARY MODE: Means that we essentially dump a sequence of bytes, which isn't necessarily related to any code page (and characters). I.e. if I'd want to save for example an executable program, the individual bytes have no meaning when interpreted as characters (unless we look at strings stored in the program). Note that legacy binary mode actually allows you to specify a code page though, but in general the recommendation is not to use the legacy option.</li>
    <li>TEXT MODE: Here we have text information that has to be interpreted using a specific code page; thus usually the additional parameter ENCODING should be given, which specifies which code page is used.</li>
    </ul>
    Now, let's clear up a small typo in Ajay's response:
    Yes you are right if you want to have all double byte characters too then you need to use ENCODING DEFAULT which would use 4103 in Unicode system.
    That is incorrect. In a Unicode system [encoding default|http://help.sap.com/abapdocu_70/en/ABAPOPEN_DATASET_ENCODING.htm] corresponds to UTF-8, not UTF-16.
    Back to your problem. Your suggestion doesn't work, because you cannot specify encoding default for a binary output (the legacy binary mode allows you to specify a code page, but that's misleading and I wouldn't use any legacy mode). So when you try to use the syntax you proposed, you'd get a syntax error.
    Generally the recommendation is for Unicode enabled applications to use UTF-8 files with byte order mark, i.e. something like
    open dataset EXP_FILE in text mode encoding utf-8 with byte-order mark.
    However, the real question is what your external audit application expects and it sounds as if it's not Unicode enabled...
    Enough blabber, here's what I'd do: Since you're having issues with a audit-related standard SAP program I'd post the question in forum - other people must have run into that problem. Also, I checked OSS, but couldn't make much sense out of the few notes I've found (and nothing seemed relevant). Check what the expected input format is in the external audit system; possibly post a message to OSS.
    Cheers, harald

  • Open dataset in UTF8. Problems between Unicode and non Unicode

    Hello,
    I am currently testing the file transfer between unicode an non unicode systems.
    I transfered some japanese KNA1 data from non unicode system (Mandt,Name1, Name2,City) to a file with this option:
    set local language pi_langu.
      open dataset pe_file in text mode encoding utf-8 for output with byte-order mark.
    Now I want to read the file from a unicode system. The code looks like this:
    open dataset file in text mode encoding utf-8 for input skipping byte-order mark.
    The characters look fine but they are shifted. name1 is correct but now parts of the city characters are in name2....
    If I open the file in a non unicode system with the same coding the data is ok again!
    Is there a problem with spaces between unicode an non-unicode?!

    Hello again,
    after implementing and testing this method, we saw that the conversion is always taken place in the unicode system.
    For examble: we have a char(35) field in mdmp with several japanese signs..as soon as we transfer the data into the file and have a look at it the binary data the field is only 28 chars long. several spaces are missing...now if we open the file in our unicode system using the mentioned class the size is gaining to 35 characters
    on the other hand if we export data from unicode system using this method the size is shrinking from 35 chars to 28 so the mdmp system can interprete the data.
    as soon as all systems are on unicode this method is obselete/wrong because we don't want to cut off/add the spaces..it's not needed anymore..
    the better way would be to create a "real" UTF-8 file in our MDMP system. The question is, is there a method to add somehow the missing spaces in the mdmp system?
    so it works something like thtat:
          OPEN DATASET p_file FOR OUTPUT IN TEXT MODE ENCODING UTF-8 WITH BYTE-ORDER MARK.
    "MDMP (with ECC 6.0 by the way)
    if charsize = 1.
    *add missing spaces to the structure
    *transfer strucutre to file
    *UNICODE
    else.
    *just transfer struc to file -> no conversion needed anymore
    endif.
    I thought mybe somehow this could work with the class CL_ABAP_CONV_OUT_CE. But until now I had no luck...
    Normally I would think that if I'am creating a UTF-8 file this work is done automatically on the transfer command

  • Unable to Open unix file in UNICODE system which created NON-UNICODE system

    Unable to Open unix file in UNICODE system which created in NON-UNICODE system
    We have two SAP systems both are ECC6.0 but System 1 is NON-Unicode and System2 is Unicode system.
    There is a common unix directory/folder for both system.
    Our requirement is to create one file on unix common folder and write the data to file from system1 .
    In system2 open the same file for appending mode to write the data .
    The file in system 1 created with below sentence.
    OPEN DATASET g_unix_file FOR OUTPUT IN TEXT MODE ENCODING UTF-8.
    Now I have to append the data from system 2 to same file.
    I have tried to used below statement in system 2 to open the file but sy-subrc value comes as '8'.
    1> OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING UTF-8.
    2>OPEN DATASET g_unix_file FOR APPENDING IN legacy TEXT MODE CODE PAGE
    cdp IGNORING CONVERSION ERRORS  .
    3>OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING Default.
    4>OPEN DATASET g_unix_file FOR APPENDING IN TEXT MODE ENCODING NON-UNICODE.
    Tried out all the possibilities as per F1 help given for open dataset , but still there is problem with opn file in appending as well output mode.However the file successfully open in Input mode(Read).
    Please advice suggestion to resolve this issue.
    Thanks.

    Messgae captured as 'Permission Denied". The program gets triggered with system user Id PPID.
    How to check the security access of the User ID.

  • Special characters as keys - Unicode BW system

    Hello together,
    we want to change our SAP BW System 3.5 to Unicode and afterwards we plan the upgrade to BI 7.0.
    We found the OSS note which tells that characteristic keys (e.g. material number, customer number, brand key, country key) must be designed as 7-bit-US-ASCII characters (common characters), but we did not found any note about what happens if we don’t convert these characters into 7-bit-US-ASCII and we’ve got some Infoobjects with special characters.
    Do we really have to convert theses characters? What happens if we don’t change these?
    We used as well the customizing option ALL_CAPITAL. What does ALL_Capital mean at the Unicode-compliant BW system. Does it have the same function as before?
    Thanks and regards,
    Julia
    Message was edited by:
            Julia

    It's a difficult matter that's made extremely easy with Unicode. But you're not there yet so you are stuck with code pages and these can get very messy.
    I dealt with this with Asian languages but it was a long time ago. I do remember there was a white paper from SAP that explained all the possible scenarios and possible solutions. For downloads to Excel etc. you will have to convert to a windows code page or Unicode in your program.
    [This SAP Help page|http://help.sap.com/saphelp_nw04/helpdata/en/9c/70023cfa699508e10000000a11402f/frameset.htm] might contain the mentioned document, I was unable to open it...
    Good luck,
    Gert.

  • Open Dataset for input in BINARY MODE not working after ECC 6.0 upgrade

    Hi All,
    Our requirement is to download an XML file from the application server and there is a customized program to download these files.
    This program uses the statement,  Open dataset ...for input in BINARY MODE
    and it works perfect for 4.7. There were no issues. But, after the upgrade to ECC 6.0 this does not work.
    When the data is read in ECC 6.0 , it is shown in some special characters and it could not be opened with XML editor and the file is not completely downloaded. I read through the forum and tried the following statement as well,
    Open dataset....for input in LEGACY BINARY MODE.
    After this statement, there were no special characters, but there is a blank space introduced before every character.
    Example : TEST(actual)
                      T E S T (After the legacy binary mode)
    Could you please let me know if there is any solution to rectify this problem. Appreciate your help.
    Thanks a million.
    Edited by: Manikd on May 12, 2011 3:52 PM

    But this program was already using BINARY MODE and after upgrade this is not working. I know, it may work in TEXT MODE. However, I cannot change the whole program to TEXT mode now.

Maybe you are looking for

  • UD made but inspection lot quantities in WM are wrong

    On the same day, we received two shipments of the same batch of material from a vendor.  Inspection Lot A  = 119cs.  Inspection Lot B = 13cs.  A total of 132cs. BEFORE materials actually were placed in WM bin locations, our Quality Group instructed t

  • Adding Field Info in Subject Line of Email with Attached PDF

    Had a question regarding the email submit button. I wanted to know if there was a way to put the information put into a specific text field in the email subject line when the PDF is being sent via email as an attached. For Example: Question one on th

  • Help - itunes does not open when my ipod is hooked up to my computer

    i downloaded the newest version of itunes. everytime i plug my ipod into my computer itunes does not open but instead a program saying "new hardware detected" and it asks to search the system and when you do that it detects my ipod as a camera. also

  • How to override the expiration time in SOA Workflow OIM?

    Hello everyone, I am working on SOA Workflows to develop the Approval Process in OIM 11g R2 PS1. I have a requirement as follows: The request should escalate after 3 days of the request submission to the approver's manager. This configuration i have

  • Freeze problem on Lumia 925

    How we switch off phone with non removable battery when it freeze.