Spool ouput in  UTF8 with BOM encoding

I've a requirement to get data from Oracle tables in a flat file.I'm using sqlplus script for the same.But those flat files must be in UTF8 With BOM(Byte Order Mark) format .Can anyone tell me how to spool data to a flat file of UTF8 format by running a sqlplus script in windows environment.Do I need to use convert function for every column in sqlplus script?Im usisng Oracle DB 11g

We could if you were asking how to do this in SQL Developer, e.g. use the Tools > Database Export wizard, but this is not the SQL*Plus or SQL forum.

Similar Messages

  • Encoding Issue: Change UTF8 with BOM char file to UTF16LE without BOM char.

    i am trying to read UTF 8 file with BOM char .. if BOM mark present then i want to remove that BOM char
    and write same file with UTF16LE with out BOM char ..
    Please suggest solution on this .
    FileInputStream fis = new FileInputStream(file);
                   long size = file.length();
                   byte[] b = new byte[(int) size];
                   int bytesRead = fis.read(b, 0, (int) size);
                   if (bytesRead != size) {
                        throw new IOException("cannot read file");
                   byte[] srcBytes = b;
                            int b0 = srcBytes[0] & 0xff;
                   int b1 = srcBytes[1] & 0xff;
                   int b2 = srcBytes[2] & 0xff;
                   int b3 = srcBytes[3] & 0xff;
                   if (b0 == 0xef && b1 == 0xbb && b2 == 0xbf) {
                        System.out.println("Hint: the file starts with a UTF-8 BOM.");
                             String srcStr = new String(b ,"UTF8");
                     String      encoding= "UnicodeLittle";
                            writeFile(filePath, srcStr,encoding);// Here is writing file with UTF16LE
         But files gets written with BOM char .
    how do i remove this .
    Please suggest solution on this

    'uncle_alice' - in the OP's other thread on this topic I posted a decorated InputStream class that will strip of any (well any I could find definitions of) BOM prefix. Using this it is almost trivial for the OP to convert a file from one encoding to another without worrying about the BOM. I showed him the water but I can't make him drink.

  • How to detect encoding file in ANSI, UTF8 and UTF8 without BOM

    Hi all,
    I am having a problem with detecting a .txt/.csv file encoding. I need to detect a file in ANSI, UTF8 and UTF8 without BOM but the problem is the encoding of ANSI and UTF8 without BOM are the same. I checked the function below and saw that ANSI and UTF8
    without BOM have the same encoding. so, How can I detect UTF8 without BOM encoding file? because I need to handle for this case in my code.
    Thanks.
    public Encoding GetFileEncoding(string srcFile)
                // *** Use Default of Encoding.Default (Ansi CodePage)
                Encoding enc = Encoding.Default;
                // *** Detect byte order mark if any - otherwise assume default
                byte[] buffer = new byte[10];
                FileStream file = new FileStream(srcFile, FileMode.Open);
                file.Read(buffer, 0, 10);
                file.Close();
                if (buffer[0] == 0xef && buffer[1] == 0xbb && buffer[2] == 0xbf)
                    enc = Encoding.UTF8;
                else if (buffer[0] == 0xfe && buffer[1] == 0xff)
                    enc = Encoding.Unicode;
                else if (buffer[0] == 0 && buffer[1] == 0 && buffer[2] == 0xfe && buffer[3] == 0xff)
                    enc = Encoding.UTF32;
                else if (buffer[0] == 0x2b && buffer[1] == 0x2f && buffer[2] == 0x76)
                    enc = Encoding.UTF7;
                else if (buffer[0] == 0xFE && buffer[1] == 0xFF)
                    // 1201 unicodeFFFE Unicode (Big-Endian)
                    enc = Encoding.GetEncoding(1201);
                else if (buffer[0] == 0xFF && buffer[1] == 0xFE)
                    // 1200 utf-16 Unicode
                    enc = Encoding.GetEncoding(1200);
                return enc;

    what you want is to get the encoding utf-8 without bom which can only be detected if the file has special characters, so do the following:
    public Encoding GetFileEncoding(string srcFile)
                // *** Use Default of Encoding.Default (Ansi CodePage)
                Encoding enc = Encoding.Default;
                // *** Detect byte order mark if any - otherwise assume default
                byte[] buffer = new byte[10];
                FileStream file = new FileStream(srcFile, FileMode.Open);
                file.Read(buffer, 0, 10);
                file.Close();
                if (buffer[0] == 0xef && buffer[1] == 0xbb && buffer[2]
    == 0xbf)
                    enc = Encoding.UTF8;
                else if (buffer[0] == 0xfe && buffer[1] == 0xff)
                    enc = Encoding.Unicode;
                else if (buffer[0] == 0 && buffer[1] == 0 && buffer[2]
    == 0xfe && buffer[3] == 0xff)
                    enc = Encoding.UTF32;
                else if (buffer[0] == 0x2b && buffer[1] == 0x2f &&
    buffer[2] == 0x76)
                    enc = Encoding.UTF7;
                else if (buffer[0] == 0xFE && buffer[1] == 0xFF)
                    // 1201 unicodeFFFE Unicode (Big-Endian)
                    enc = Encoding.GetEncoding(1201);
                else if (buffer[0] == 0xFF && buffer[1] == 0xFE)
                    // 1200 utf-16 Unicode
                    enc = Encoding.GetEncoding(1200);
               else if (validatUtf8whitBOM(srcFile))
                    enc = UTF8Encoding(false);
                return enc;
    private bool validateUtf8whitBOM(string FileSource)
                bool bReturn = false;
                string TextUTF8 = "", TextANSI = "";
                //lread the file as utf8
               StreamReader srFileWhitBOM  = new StreamReader(FileSource);
               TextUTF8 = srFileWhitBOM .ReadToEnd();
               srFileWhitBOM .Close();
                //lread the file as  ANSI
               StreamReader srFileWhitBOM  = new StreamReader(FileSource,Encoding.Defaul,false);
               TextANSI  = srFileWhitBOM .ReadToEnd();
               srFileWhitBOM .Close();
               // if the file contains special characters is UTF8 text read ansi show signs
                if(TextANSI.Contains("Ã") || TextANSI.Contains("±")
                     bReturn = true;
                return bReturn;

  • UTF-8 files with BOM chrashes DOMParser?

    Hi.
    We are storing XML-documents in an 8i databse with UTF-8 encoding (in CLOBS).
    Problem: If the Unicode XML-document contains a BOM the oracle.xml.parser.v2.DOMParser's
    parse()-method throws an exception.
    I get the following output when using the ParseXMLFromURL.java class supplied in JDeveloper 3.2 samples directory:
    Sample output >>>>
    System Output: XML parse error in file http://localhost/UTF-8_With_BOM.xml
    System Output: at line 1, character 1
    System Output: Start of root element expected.
    <<<<<<< Sample output
    If I change the XML-file not to include a BOM the parser works fine.
    (I set/unset the BOM using EmEditor from http://www.emurasoft.com/ if you'd like to try for yourselves).
    To me it looks like DOMParser interprets the BOM at the start of the XML-file as XML-content instead of a Unicode signature.
    IE 5.5 can handle both formats, shouldn't DOMParse also be able to handle that?
    Any ideas how I can get DOMParse to work with UTF-8(BOM) XML-files?
    Regards,
    Jan-Erik
    Sample XML:
    <?xml version="1.0" encoding='UTF-8'?>
    <newsdoc>
    <news>
    <newstitle>
    Document contains no BOM
    </newstitle>
    <introduction>
    See http://www.unicode.org/unicode/faq/utf_bom.html for info on BOM
    </introduction>
    </news>
    </newsdoc>
    null

    I have the same problem when trying to store UTF-8 encoded XML files with BOM marks in iFS version 1.1.9.0.7.
    The database is 8.1.7.1.1 created with UTF-8 charset.
    I have loaded the XDK for PLSQL 9.0.2.0.0A into the database and replaced the original %ORACLE_HOME%\lib\xmlparserv2.jar with the one distributed in this XDK.
    I get the following error message:
    Wed Aug 01 10:10:06 GMT+02:00 2001: \public\CV-Bank\CV_Patrik_Johansson_intDTD_BOM.xml:
    oracle.ifs.common.IfsException: IFS-12608: Error while pre-parsing with the SAXParser: at line (1), column (1): oracle.xml.parser.v2.XMLParseException: Start of root element expected.
    at oracle.ifs.beans.parsers.IfsXmlParser.preParse(IfsXmlParser.java, Compiled Code)
    at java.lang.Exception.<init>(Exception.java, Compiled Code)
    at oracle.ifs.common.IfsException.<init>(IfsException.java, Compiled Code)
    at oracle.ifs.common.IfsException.<init>(IfsException.java, Compiled Code)
    at oracle.ifs.beans.parsers.IfsXmlParser.preParse(IfsXmlParser.java, Compiled Code)
    at oracle.ifs.beans.parsers.IfsXmlParser.getParserName(IfsXmlParser.java, Compiled Code)
    at oracle.ifs.beans.parsers.IfsXmlParser.parse(IfsXmlParser.java, Compiled Code)
    at oracle.ifs.beans.parsers.IfsXmlParser.parse(IfsXmlParser.java, Compiled Code)
    at oracle.ifs.utils.common.ParserHelper.parseExistingDocument(ParserHelper.java, Compiled Code)
    at oracle.ifs.protocols.ntfs.server.FileProxy.parseFile(FileProxy.java, Compiled Code)
    at oracle.ifs.protocols.ntfs.server.FileProxy.cleanupFile(FileProxy.java, Compiled Code)
    at oracle.ifs.protocols.ntfs.server.FileProxy.runFileProxy(Native Method)
    at oracle.ifs.protocols.ntfs.server.FileProxy.run(FileProxy.java, Compiled Code)
    This is a serious problem since we use an XML editor that adds BOM's.
    Regards
    Patrik Johansson

  • Spool output in utf8 format

    Hi,
    I've a requirement to get data from Oracle tables in a flat file.I'm using sqlplus script for the same.But those flat files must in UTF8 format.Can anyone tell me how to spool data to a flat file of UTF8 format by running a sqlplus script in windows environment.Do I need to use convert function for every column in sqlplus script?
    Thanks,
    Mahender.
    Edited by: user518071 on Sep 9, 2010 1:51 PM

    Hi Srini,
    Here're the steps I followed :
    1) Checked Oracle DB character set
    SELECT * FROM NLS_DATABASE_PARAMETERS and NLS_LANG is AL32UTF8
    2) I'm spooling Oracle Table data with the following script :
    set echo off
    set feedback off
    set pagesize 0
    set linesize 2000
    set heading on
    set trimspool on
    Spool EMP.TXT;
    PROMPT empno#|#ename#|#sal#|#comm#|#Hiredate
    SELECT LTRIM(RTRIM(empno))||'#|#'||
    LTRIM(RTRIM(ename))||'#|#'||
    LTRIM(RTRIM(sal))||'#|#'||
    LTRIM(RTRIM(comm))||'#|#'||
    TO_CHAR(hiredate,'MM/DD/YYYY HH:MI:SS')
    from emp where rownum<=100;
    spool off;
    Thanks,
    Mahender
    Edited by: user518071 on Sep 13, 2010 8:53 AM
    Edited by: user518071 on Sep 14, 2010 1:00 PM

  • Error in Rush Order type with BOM item!!

    Hello,
    I am facing a very strange error while creating a rush Order with BOM item.
    System Error:
    1. No item category available (Table T184 SO ERLA TAQ).
    I except the error. I get next error as
    2. Structure explosion for item 000010 is not possible.
    After this BOM explodes with only first 2 Sub Items ( Item cat for main item is determind as TAQ and Sub Items as TAE) , where as the complete BOM has 7 Sub items.
    I have made the required configs in item cat determination in T184 table for order type SO, as follows.
    SO ERLA - TAQ for Main Item of the BOM
    SO NORM TAQ TAE for the Sub Item of the BOM
    The same item (BOM ) works very well in OR/BV kind of sales order types, by exploding the BOM to full level(i.e all the 7 sub items are determined).
    Has any one come across such an issue. Please let me know what could be the solution?
    Thanks in adv,
    Shripad

    Hi,
    Thanks,
    In vov7 for TAQ Item Cat. the structure scope is maintained as 'A'. And that's the reason for Standard order 'OR' , the BOM is exploding properly.
    It is only with Rush Order , the BOM is not exploding correctly.
    BR
    Shripad

  • Problem in Delivery with BOM

    Dear Experts,
    I have problem with delivery with BOM
    The steps are:
    1. I create Sale Order with BOM A (which consist of material AA 1 PC, AB 1 PC and AC 1 PC) 75 ST
        So Sale order shows:
        Item           Material         Qty    Unit
        10     A     75      ST
        20     AA                75      PC
        30     AB     75      PC
        40     AC                75      PC
    2. I create Delivery reference sale order from 1 so Delivery will be:
        Item           Material         Qty    Unit
        10     A     75      ST
        20     AA                75      PC
        30     AB     75      PC
        40     AC                75      PC
        But I want to delivery only 25 ST so I change quantity in Item 10 from 75 to 25 but the system show:
        Item           Material         Qty           Unit
        10     A     25             ST
        20     AA                25             PC
        30     AB     25.004      PC
        40     AC                25             PC
    What happen with Item 30 why system determine to 25.004. Please help me

    Hi,
    if you want change the item quanty like TAP or TAN plz goto VOV7 select TAP Item category  under bill of material/configuration Maintainstructure scope A and application SD01 put the check mark for *manual alternativs  and same to TAN also
    Your Itemcategory determination should be IN VOV4
    QT-LUMF--TAP
    QT-NORM-TAP-AGN
    If your using BOM item categories is showing gray mode only.
    regards,
    sreenivas

  • Problem with BOM (packaging material) in order change

    Hi Gurus,
    In my b2b shop, i have created an order with BOM material and that order contains one main material and a sub item material after update or order. When i open the same order and click on change, when i click on update, a new sub item is getting added as a main item.
    Details:
    item 10 is 12007492 (main material)
    item 20 is 12007491(sub item for the main material)
    when i click on update,again one more sub item 30 is getting added.
    When i check the logs, all the items guids are getting passed to IN of crm_isa_basket_changeitems functions module. But when check it in runtime using external debuggin mode, the sub item guid is blank.
    Then the main material is again getting exploded and the sub item is coming at item 20 and actual sub item before update is shifting to item 30 as new item.
    Please help me in this regard why it is happening like this.
    Thank you
    Lakshman

    Hello Lakshman,
    I had the same issue, however my backend system was ERP and not CRM. In any case, the logic might be the same.
    When a BOM material is sent to the backend system for simulation/save, only the header needs to be sent. By this, backend system will explode the material and send 2 items back. I guess in your case, ISA is sending both items to SAP and that is the reason you are getting back 3 items.
    Simplest approach is to write some logic in your backend class to avoid sending this subitem to SAP. You can achieve this by manupulating the items table sent to SAP
    Suppose salesDoc is the ISA document and itemsTable is the JCO table that is sent to SAP as a table parameter:
                   ItemList itemList = salesDoc.getItems();
                   for (Iterator iter = itemList.iterator(); iter.hasNext();) {
                        ItemSalesDoc itemSalesDoc = (ItemSalesDoc) iter.next();
                        TechKey parentKey = itemSalesDoc.getParentId();
                        if(null!=parentKey){
                             if (!parentKey.getIdAsString().equalsIgnoreCase("")) {
                                //Here remove it from the tables parameter which is sent to SAP
                                               itemsTable.deleteRow(i);
    I am not too sure how it would work for CRM, but works in ERP.
    Pradeep

  • PO with BOM

    Hi everyone, pls. I need your advice related with PO with BOM.
    I explain the process:
    I always buy a kit containing 350 items, so the PO have only 1 ítem because the manufacturer bills 1 item, but the GR have to be for 350 items. After GR I have to assemble the kit. This process it´s related with motorcycle production.
    I trie to use a PO with item cat. L but before MIGO I have to do ME2O to send materials in subcontracting, so this process is similar to subcontracting but is not the same.
    Thanks for your cooperation.
    Regards.

    Sorry, I think I misunderstood your question. The kit contains not 350 pcs but 350 items (different components).
    In SAP Retail you can use the structured article scenario.
    If you don't use IS Retail you can still have one item on the PO. After GR you can split it into components (e.g. MB1A + 201 for the main item and MB1A+202 for the compoenents - you can of course automize this via development that these technical bookings should be triggered by the GR).
    But sorrowfully standard SAP cannot handle BOm on PO except subcontracting and IS Retail...(as far as I know)
    (More exactly: IS Retail structured article means an additional movement, so it's not really BOM on the PO.)
    Edited by: Csaba Szommer on Jul 7, 2011 9:34 PM

  • Media Encoder CS6 with Media Encoder CC

    Hi,
    I have finally gotten Media Encoder CC installed. I still have Media Encoder CS6 on my computer. Can I uninstall Media Encoder CS6 without having problems with Media Encoder CC or any other CC apps? There is no uninstall optiion for ME CS6.
    Thanks

    Uninstalling Media Encoder CS6 shouldn't cause any problem to any CC apps, including Media Encoder CC.
    There is no stand-alone uninstaller for Media Encoder CS6.  In order to uninstall it, you need to uninstall all CS6 apps on which Media Encoder CS6 is depending.  Those CS6 apps includes, Premiere Pro, After Effects, Flash, Prelude, SpeedGrade, and InDesign.  Media Encoder CS6 is installed as a part of those apps.  If you have any of those CS6 apps, please uninstall all of them.  This also uninstalls Media Encoder CS6.

  • Problem with file.encoding in Linux

    Hello,
    I am currently migrating a java project from Windows to Linux. The project is finally shaping up now, except for some encoding problems.
    All configuration files are saved in ISO-8859-1/Cp1252 format. When reading and displaying these files in Swing (e.g. JTextArea), the special characters ���� are displayed wrong. I have tried to start the VM with -Dfile.encoding=ISO-8859-1 and -Dfile.encoding=Cp1252 without success (this is done in Eclipse under Linux).
    I then tried the opposite. I created some UTF-8 files, started the application under Windows/Eclipse, read the files and displayed them in a JTextArea. Garbage characters were shown instead of ���� (as expected). I then used -Dfile.encoding=UTF-8, and voila, all characters were displayed correctly.
    Why does not -Dfile.encoding work for ISO-8859-1/Linux but -Dfile.encoding work for UTF-8/Windows? Anyone here know?
    The JRE I have been using is 1.4.2_06.
    The Linux is a SuSE 10.0

    Continue the "discussion" here:
    http://forum.java.sun.com/thread.jspa?threadID=737153

  • Premiere CS4 with Media Encoder CS5?

    Hi,
    When I was given my CS4 suite (production premium) years back, it included Premire Pro.  Then, when I told my boss I needed to upgrade to CS5, he purchased the Web Premium Suite.  This was fine, for the most part, as Illustrator, Flash, and Photoshop are the main programs I use, but we just recently got a video project that is giving me some issues.
    We created what turned out being a 4 min + video.  After some trial and error we settled on creating multiple segments in Flash and exporting them as uncompressed MOV files from Flash.  Then we brought all of those MOV files together in a premiere pro project.  The problem seems to occur when I export out an MP4 video from Premire Pro CS4.  Premiere launches Media Encoder CS4 and then creates the MP4.  The MP4 plays on a cellular device like we want it to, but it will not display a thumbnail/preview of the video when viewing the video in the phones gallery.
    We have tried some other tests, and if we export out an MOV or AVI from Flash CS5 then convert that to an MP4 using Media Encoder CS5, there suddenly seems to be a thumbnail on the phone for the video.  After multiple tests, it appears the problem might be related to which version of Media Encoder our final MP4 is produced in.  ME CS4 - No thumbnail. ME CS5 - thumbnail.
    The problem is, we HAVE to use Premiere as our final compiling tool, but I cant seem to get a Premiere Pro CS4 file to work with Media Encoder CS5.
    Any thoughts???
    Thanks,
    Justin

    but to automatically open up AME CS5 ... hmmm.. maybe a hack?
    No, that will not work. Even if you have Audition 4.0 installed (beta), edit in Audition from CS5 will only access Audition 3.0.
    The only way to try it is to open AME CS5, go to File/Add Premiere Pro Sequence and select the CS4 project and then the correct sequence.

  • How to export a PDF to a word document? at the moment it came up with "The encoding (CMap) specified by a font is missing ????? what can I do?  thanks

    How to export a PDF to a word document? at the moment it came up with "The encoding (CMap) specified by a font is missing ????? what can I do?  thanks

    Hi frenchiem,
    This is due to the PDF file cannot display the original fonts. To fix it, creator should create the PDF file with the embed font option enabled. Thus, user can open the file on any platform with the same layout. Or the Acrobat product can display them after you install the appropriate font support package on the client side.
    Regards,
    Rave

  • MRP run for header material with bom usage-3

    Dear All,
    We have header material with bom usage-3.We maintain the stock at child components only ,not at header
    I have created the planned independent requirement for header material.I have created a sales order but sales order is not shown in md04.
    What can be the further setting  I am missing for mrp run of above scenario.
    with regards,
    pradeep

    Dear ,
    Couple of checks points for you :
    1.Check the material master Sales Org Veie-Item Categoery .As you have mentioned Item Categoery -TAQ which is not relevent for requirement planning throgh schedule line category -CT ( CT means there will not be any inventory or good movement and requirement will not be passed for MRP ) .Only CP -line item categoery does requirement passing through MRP and it will show u in MD04. So the summery is if you want to generate an MTO order for the same FG material , you should matian the Sales Line item categoery as MTO in sales Org View and it will pop up in VA01 whiel logging Sales Order .
    Generally , TAQ item categoery are used in Project Billing or Delivery .ok  and they will not be having any availabilit check so it KP ( no check ) .
    But for MTO order , you should have availablty chek 02/01 in MRP3 view and also MTO Sales Item categoery .
    Based on the above information logg an sales order , release credit chek in VKM3, SAVE and check in MD04.
    Create PIR in MD61 and Run MRP on the both cases .I have tested its working .but noit with TAQ item categoery ok
    2.Why do you need to create Sales Order for dependent Item-Y because it will be in X-BOM item  ? Make sure that Y- should have Sales Item categoery in material master -MM02-Sales Org View as NORM item .If you really want to create sale sorder for this item , it should TAN or NORM .
    3.Other PIR detail is fine .
    Try and revert back
    Regards
    JH

  • Issues with Japanese encoding using Mail

    Since recently (I would say since I updated to 10.6), I have an issue with Japanese-encoded (ISO 2022-JP) mails on my English MacOS.
    I have no problem to read, edit and write answers to any mails.
    However with some ISO JP-2022-JP encoded messages (sent with Thunderbird 2.0.0.23 (Windows/20090812) btw) I have the following misbehaviour:
    - if I send the message and let the encoding to "automatic", Mail sends the mail in UTF-8, which I do not want since most of Japanese computer do not understand UTF-8 by default (and the receiver gets panicked: "I can not read your mail T_T !")
    - if I set the encoding to "ISO JP-2022-JP", I can not send nor save the message (see [1] at the end of the post). One should note that the error message when saving is really misleading (and yes my hard-drive has a lot of space left) and it should be fixed by Apple.
    - if I dig a bit deeper, I can in effect find some characters in the original message which prevent Mail to send my mail. It however does not make any sense since:
    - those char were in the original message properly encoded in ISO JP-2022-JP
    - those char are always very common ones
    The only solution I have found so far is to delete the original message in my mail, which is very frustrating...
    A sample of such mail can be found at (I removed personal info. and the mail is about a drinking party):
    - http://files.me.com/trouve.antoine/73w3w9
    Help would be very appreciated.
    Thank you very much.
    Antoine
    [1] I get the following error messages:
    -> try to save:
    *This message can’t be saved to the Drafts mailbox.*
    The message contains one or more attachments that
    are too large to be saved in the Drafts mailbox. Try
    deleting some attachments.
    ->try to send
    *Invalid Text Encoding*
    Some characters in your message could not be
    converted to the “Japanese (ISO 2022-JP)” text
    encoding. Choose a different encoding from the
    “Text Encoding” menu.

    You can find out about the different versions here, for example:
    http://en.wikipedia.org/wiki/ISO/IEC_2022
    Thank you. I feel a bit stupid for not having looked in Wikipedia at first...
    I sometime wonder how could such basic problem like charset not being solved after more than 40 years of computer science...
    Here is a note that addresses that problem, but I don't think it works with 10.6. Might be worth a > try:
    http://discussions.apple.com/thread.jspa?threadID=121808&tstart=60
    Thank for the link.
    It seems to still work: new japanese mails are now sent in "ISO 2022-JP-2".
    However, for messages with the header explicitly specifying "ISO 2022-JP" (which should be "ISO 2022-JP-2" on my mac) it has no influence.
    The only ways I see to solve this issue would be:
    i) to force "ISO 2022-JP-2" for all mails (a bit too extreme)
    ii) to force the use of "ISO 2022-JP-2" instead of "ISO 2022-JP", but I do not think such precise configuration is possible
    This mess appears to be due to Thunderbird which seems to mix "ISO 2022-JP-2" and "ISO 2022-JP", but I do not have any working Thunderbird to test now...

Maybe you are looking for