Converting garbled characters for JAPANESE characters in a custom table

Hi all,
I have a custom table that store Japanese characters.
After my company has upgraded to ECC6.0, this data in the custom table has become garbled and its alot of it garbled.
Is there any SAP tool that can I use to make the correction on those garbled Japanese characters?
Thanks,
William Wilstroth

Hi Nils,
I really really really had a field day reading and testing around UC... To my dissappointment, I do not have the authorization to use SUMG and SCP too as well as a few of the TCODES...
I finally told my higher level technical mgnt. that this table might need some changes...
Has this problem of mine got anything to do with MDMP since its no longer supported in ECC6 and I found one coding that search for MDMP in RSVTPROT...
My colleagues suggest that the data be corrected from table DBTABLOG... which i find, in my opinion, is not the right way...
Thanks,
William

Similar Messages

  • Transaction code for creation of a specific customizing table ?

    Hi,
    What's the transaction code for creation of a specific customizing table ?
    Thanks in advance.

    Carl,
    if you can share with us from which t-code you achieve this , than it would be better for we guys.
    Amit.

  • SmartForms Printing Error for Japanese Characters

    Hi,
    I have an issue with Printing Japanese Characters in the SmartForm. 
    When I Print Preview the smartform I can see the characters in Japanese but when I Print the actual smartform in the printer, all I get is some junk characters like # and !.
    I have the following setting in the smartform.
       Multi-Byte functionality enabled
       Character set used is Shift JIS
       Smartform created in language EN
       Translate to All Languages set to true
       No restriction on Language Control
       Smartstyles used has JPMINCHO fonts only
       Default Para is JPMINCHO
       Characters formatting has explicit JPMINCHO as font
       Login language is JA
       Printer is Local Printer with device type JPSAPWIN
    When I debug the Smartform function module, the OTF Data Table (within the function) shows characters in Japanese (with codepage 8000) but then I look at the actual spool as a result of processing the function, it comes up as garbled.
    Interestingly, I have exactly the same settings on another test smartform to print just 2 columns and it works well.
    Not sure where I am going wrong.
    Appreciate help on this issue.
    Many Thanks
    Krishnan

    hi,
    has anyone solved this problem? even i m facing the problem. me restating the problem as below:
    in VF02 transaction, for billing document we are trying to print smartform. Our logon languagae is EN and the Output Type langugae is EN. In spool, we are getting the CORRECT format ie the texts that are maintained in Japanese are showing as Japanese characters while the whole smartform is in English format.
    Our requirement is to PRINT (from printer) the texts that are maintained in Japanese should get printed in Japanese format. I know that this is problem with printer settings. The printer is printing perfectly fine as other documents in Japanese characters so its ensured that printer is enabled for Japanese.
    I suppose some kind of setting needs to be done at SAP end so that printer recognises double-byte character set of Japanese.
    Waiting for confirmation/ solution.
    Thanks
    Debs

  • Double Byte Characters for Japanese Kanji Language

    Hello All,
    How can you identify the Double Byte characters and how we can check whether they are double byte characters  or not.
    Can you please tell me the double byte characters from Japanese Kanji language for the characters.
    This is very urgent.I will reward you with points.
    Thanks,
    Karan

    *& Report  YTEST_LOGIC
    REPORT  teched_unicode_solution_1               .
    *** Exercise 1: Distinction between byte and character length
    *** after Unicode enabling
    parameter: param type c.
    PERFORM test1 USING param.
    *  FORM test1
    FORM test1 USING text TYPE c.
      DATA: len1 TYPE i,
            len2 TYPE i,
            off TYPE i.
    DESCRIBE FIELD text LENGTH len1 IN BYTE MODE.
    DESCRIBE FIELD text LENGTH len2 IN CHARACTER MODE.
    WRITE:/ LEN1, LEN2.
    ENDFORM.                                                    "test1
    Use the above code to find which characters are double bytes and which are not...
    The double byte characters will have double the length when they are in the byte mode..
    reward points for helpful answers
    Edited by: Rahul Kavuri on Mar 26, 2008 6:35 PM

  • FM for F4 help on a custom table

    Hi ,
    Is there any FM which can show the values of a particular field of a custom table.
    e.g. I have a custom table in which one of the fields in user id.
    Now I wanna show it as a pop up in which the field would be user id. Now if the user clicks on f4 help of this field , I wanna show him all the values of this field from the custom table.
    Please suggest how this can be done.
    Regards,
    Divyanshu

    There are a number of ways to do this.  Here is one.  This example program gets values from T001 for F4 help.  You can use your custom table instead.
    report zrich_0001 .
    tables: t001.
    data: begin of it001 occurs 0,
          bukrs type t001-bukrs,
          butxt type t001-butxt,
          ort01 type t001-ort01,
          land1 type t001-land1,
          end of it001.
    select-options s_bukrs for t001-bukrs.
    initialization.
      select bukrs butxt ort01 land1 into table it001 from t001.
      sort it001 ascending by bukrs.
      delete adjacent duplicates from it001 comparing bukrs.
    at selection-screen on value-request for s_bukrs-low.
      call function 'F4IF_INT_TABLE_VALUE_REQUEST'
           exporting
                retfield    = 'BUKRS'
                dynprofield = 'S_BUKRS'
                dynpprog    = sy-cprog
                dynpnr      = sy-dynnr
                value_org   = 'S'
           tables
                value_tab   = it001.
    start-of-selection.
    Regards,
    Rich Heilman

  • HR ABAP for employee groups with possible custom table...

    Hi Members
    I have a internal table where employee name and employee group are maintained.
    Next, i have other tables where employee name is there and his salary. now I want to add all the salaries of the employees with same groups, how can I do this.
    for Example, I have employee group as ENG, DOC, NUR Etc. (Also I wanted to use some kind of dynamic conditions as emplyee groups might added latter in custom table.
    Thank you.
    Edited by: Julius Bussche on Sep 23, 2008 10:18 PM

    Check out this code:
    *& Report  ZTEST3
    REPORT  ztest3.
    TYPES:
    BEGIN OF x_employee,
      id TYPE n LENGTH 10,
      name TYPE c LENGTH 20,
      group TYPE c LENGTH 4,
    END OF x_employee,
    BEGIN OF x_emp_sal,
      id TYPE n LENGTH 10,
      salary TYPE p LENGTH 16 DECIMALS 2,
    END OF x_emp_sal,
    BEGIN OF x_final,
      group TYPE c LENGTH 4,
      id TYPE n LENGTH 10,
      name TYPE c LENGTH 20,
      salary TYPE p LENGTH 16 DECIMALS 2,
    END OF x_final.
    DATA:
    i_employee TYPE STANDARD TABLE OF x_employee INITIAL SIZE 0,
    i_final TYPE STANDARD TABLE OF x_final INITIAL SIZE 0,
    i_salary TYPE STANDARD TABLE OF x_emp_sal INITIAL SIZE 0,
    wa_employee TYPE x_employee,
    wa_salary TYPE x_emp_sal,
    wa_final TYPE x_final,
    wa_final_temp TYPE x_final.
    DEFINE append_employee.
      wa_employee-id = &1.
      wa_employee-name = &2.
      wa_employee-group = &3.
      append wa_employee to i_employee.
    END-OF-DEFINITION.
    DEFINE append_salary.
      wa_salary-id = &1.
      wa_salary-salary = &2.
      append wa_salary to i_salary.
    END-OF-DEFINITION.
    append_employee:
    1 'John' 'ENG',
    2 'Mary' 'ENG',
    3 'Pooja' 'IT',
    4 'Payal' 'IT',
    5 'Sourav' 'IT'.
    append_salary:
    1 '111.00',
    2 '1111.00',
    3 '11111.00',
    4 '111111.00',
    5 '1111111.00'.
    LOOP AT i_employee INTO wa_employee.
      READ TABLE i_salary INTO wa_salary
       WITH KEY id = wa_employee-id.
      IF sy-subrc = 0.
        wa_final-id = wa_employee-id.
        wa_final-name = wa_employee-name.
        wa_final-group = wa_employee-group.
        wa_final-salary = wa_salary-salary.
        APPEND wa_final TO i_final.
      ENDIF.
    ENDLOOP.
    SORT i_final BY group.
    LOOP AT i_final INTO wa_final_temp.
      wa_final = wa_final_temp.
      AT END OF group.
        SUM.
        WRITE: /1 wa_final_temp-group, 5 wa_final_temp-salary.
      ENDAT.
    ENDLOOP.
    output will be:
    ENG                         1.222,00  
    IT                      1.233.333,00  
    Edited by: Sourav Bhaduri on Sep 24, 2008 1:44 AM

  • Passing Japanese characters to Java program

    Hi,
    I am using Java POI APIs for creating/modifying excel spreadsheets. The platform is Solaris. The data to be populated contains both English as well as Japanese characters. The input data to Java program comes from a perl program. When I populate the data in excel or even print using System.out.println, it prints ???... for Japanese characters. The data is in SJIS format.
    I tried converting it to UNICODE in java but that doesn't seem to be working.
    I also tried using native2ascii for converting the data to UNICODE before it is fed to Java program using the following command
    native2ascii -encoding SJIS <Input File> <Output File>
    Although it converts the charaters into UNICODE correctly in the form \u8a3c\u523... but when it is input to Java, the program prints the string as such.
    But if this string is hardcoded in the program, the Japanese characters are printed correctly.
    Could anybody please throw some light as to the data should actually be passed to the Java program.
    Thanks

    Sekhar
    Ive realised the solution will probably involve HttpServletRequest.setCharacterEncoding.
    Im now upgrading to Tomcat 4 because 3 didnt support this method. Once I'm through the upgrade, I'll try Japanese the chars again.
    I would guess I need to force my web pages to utf-8 and use HttpServletRequest.setCharacterEncoding(utf-8) in the servlet to get it working ?

  • Display Japanese characters on Applet

    Hi All,
    I am able to display Japanese Characters on my Applet locally(Windows 2000 - US) by overwriting the font.properties content with the one in font.properties.ja found in JRE/LIB directory of my Jdeveloper JDK. When I deploy the applet on 9ias server which is configured for UTF8 charset I get squares wherever I have Japanese characters.
    Appreciate if someone could tell the procedures to make the applet to recognize the Japanese/UTF8 characters and display accordingly.

    I have read some documentation on JDBC thin drivers and they have mentioned that whenever JDBC retreives data from a database it converts it to Unicode 1.2 and I am already retreiving data from a UTF 8 database. The reason I think this is true is because I can display japanese characters locally on my windows machine by overwriting the font.properties in the JDevelopers- JDK- JRE - LIB folder by font.properties.ja
    There must be some way to test why it is showing squares for japanese characters when I deploy it on the server. May be not getting the MS Gothic font used to display Japanese?
    Looks like updating all the code is not feasible at this point as the code currently stores the data in Vectors and we can't perform a vector.getBytes(). Any help will be really appreciated.

  • Unicode to Japanese characters

    Hi,
    I need a function to convert unicode to Japanese characters. I have a unicode string in my syncBO and it needs to be converted to the "strange" Japanese characters.
    When I read the unicode String from the MAMText files it is automatically done by the MI framework.
    Unicode string example: \u6A5F\u80FD\u5834\u6240\u4E00\u89A7
    Are there standard MI functions for me which are converting these kind of Strings? Because for the MAMText files it is working via the getString method of the com.sap.ip.me.api.services.MEResourceBundle class.
    I hope someone knows more about this conversion.
    Thanks in advance! Kind regards,
    Bart Elshout

    hi bart,
    it would be nice if you could feedback me with the error problems.
    i know the code might not cater to all of the requirements, you might need to add
    other escaped characters if you need to in the switch scope. i forgot to tell you
    that to use the a2n function you should provide it with the string in its ISO format.
    try running on the MyTest program below with the argument as follows:
    MyTest "u6A5Fu80FD u5834 u6240u4E00u89A7"
    public class MyTest {
      public  static void main(String[] args){
        System.out.println(a2n(args[0]));
    if you use it like this way, it will not work, since the String class will detect this
    and automatically create the native representation of the characters. there's no
    need for the a2n function.
      String s = "u6A5Fu80FD u5834 u6240u4E00u89A7";
      System.out.println(a2n(s));
    if you are reading text from an ascii file, you need to specify that the input stream
    should be  ISO-8859-1 i.e.something like
    new InputStreamReader(inStream, "ISO-8859-1")
    hope this helps.
    jo

  • How to store japanese characters in mysql 5.0

    I want to store japanese characters in mysql 5.0 database through java program and then to retrive the same characters through the program only.Java program means a form containing first name ,last name and address. I am entering corresponding translations in japanese to this fields while inserting to database. In another form i am retrieving those japanese characters.those should display
    in this form.

    How to handle the unicode sir for japanese characters..Pls give me more hints and any reference links to get the answer.

  • Problem with Gui_download using ASC File type - japanese characters

    Hi,
    During upgrade,while downloading data for japanese characters using GUI_DOWNLOAD Function module with file type as 'ASC', the space between 2 fields data getting much wider compared to 4.6C Version ws_download Function module's  data.
    Example: the gap between first field data and second field data in ECC 6.0 is 6 characters length,but in 4.6C it is 2 characters length.
    Is there any possibility to get the results similar to 4.6c version.Please give your valueable suggestions.
    Thanks
    BalaNarasimman

    Hi Sandra
    Please find the detailed information for your questions.
    1.Internal table content before download:During Debugging,it was observed that internal table content was same in both versions.For testing,i used only brand new data(Transaction entry).
    2.Download with code Page conversion:Yes,codepage parameter 4103 was explicitly passed into GUI_DOWNLOAD Function module.Also the front end code page which is used by system is 4110 . No errors occured.
    3.System is an Unicode system only.
    4.Actually this 6 character does not refer the byte value,only the gap between 2 fields data is getting referred in ECC 6.0.Please find the below example.
    Example - File data after Download:
    ECC 6.0: Field1            Field2      (gap - 6 characters space between 2 fields data)  Using GUI_Download
    data       u0152©Ïu201Dԍu2020      EN                               
         4.6C: Field1            Field2       (gap - 2 characters space between 2 fields data) Using WS_Download
         data    u0152©Ïu201Dԍu2020  EN    
    Note:Special characters are Japanese characters:

  • MySQL Japanese characters

    Hello,
    I already searched the froums for a soluton to my problem but couldn't find any for it though I tried some of the things that were proposed. I have a MySQL Database where I store Japanese characters in SJIS. I tested if they are in SJIS and they are.
    I am using mysql 4.1.3b-beta, mysql-connector-java-3.0.14, JBuilderX (jdk 1.4.2), and OS XP.
    When I try to retrieve the data I only get a square, then a some character, then a square, then another character,... but not the actual Kanji/Kana. I can display Japanese characters correctly when I retrieve them from a file, so I don't think it's a problem with the font.
    My code is as follows:
    public void testDB(){
    String result ="";
    Connection con = null;
    Statement st = null;
    try
    Properties prop = new java.util.Properties();
    prop.put("user","");
    prop.put("password","");
    prop.put("useUnicode", "true");
    prop.put("characterEncoding","SJIS");
    String url = "jdbc:mysql://localhost/japanese";
    //Class.forName ("org.gjt.mm.mysql.Driver");
    Class.forName ("com.mysql.jdbc.Driver").newInstance ();
    con = DriverManager.getConnection (url, prop);
    result= "Database connection established";
    st = con.createStatement();
    ResultSet rs = null;
    String querry1 = "SELECT * FROM jp";
    PreparedStatement pstmt = con.prepareStatement(querry1);
    rs = pstmt.executeQuery();
    int i = 0;
    while (rs.next()) {
    result = rs.getString(1);
         jTable1.setFont(new Font("Arial Unicode MS",0,15));
    jTable1.setValueAt(result,i,0);
    i++;
    catch (Exception e)
    // result = "Cannot connect to database server";
    finally
    if (con != null)
    try
    con.close ();
    // result ="Database connection terminated";
    catch (Exception e) { /* ignore close errors */ }
    I hope someone can help me as MySql is my last option after I already discovered that I couldn't use MS Access and Foxpro because the ODBC:JDBC bridge doesn't work correctly for unicode.
    On a side note as I am new to MySql after installing version mysql4.1.3b. I no longer have the mysqld.exe but only mysqld-opt.exe. So I am using this instead. But I don't think that can be responsible for my problem.
    Thaks for any help.
    S

    I see your original question was sent in August so maybe you have found an answer ?.. Would be interesting to see how you fixed it.
    I was having a similar sort of problem but managed to fix it by looking at the fonts available to Java on my machine and setting the font used by the graphics object to one of the japanese fonts. (nb. I see in your code you are setting the font to arial unicode, maybe setting it to a japanaese font name would be enough)
    Hopefully the following will give you some help.
    // this line sticks the name of all fonts in the graphics environment in an array
    String[] nameArray = GraphicsEnvironment
                   .getLocalGraphicsEnvironment()
                   .getAvailableFontFamilyNames();
    // I know that the last two fonts are for japanese characters. I chose the font name second from the end of the array.
    int listfonts=nameArray.length-2 ;
    //then I sent the font for the g2 object.
    Font fontselec ;
    fontselec = new Font(nameArray[listfonts],Font.PLAIN,24) ;
    g2.setFont(fontselec) ;
    best regards
    Graham

  • [Bug Report] CR4E V2: Exported PDF displays Japanese characters incorrectly

    We now plan to transport a legacy application from VB to Java with Crystal Reports for Eclipse. It is required to export report as PDF file, but result PDFs display Japanese characters incorrectly for field with some mostly used Japanese fonts (MS Gothic & Mincho).
    Here is our sample Crystal Reports project:   [download related resources here|http://sites.google.com/site/cr4eexportpdf/example-of-cr4e-export-pdf]
    1. PDFExportSample.rpt located under ..\src contains fields with different Japanese fonts.
    2. Run SampleViewerFrameClient#main(..) to open a Java Report Viewer:
        a) At zoom rate 100%, everything is ok.
        b) Change zoom rate to 200% or 50%, some fields in Japanese font collapse.
        c) Export to PDF file,
             * Fonts "MS Gothic & Mincho": both ASCII & Japanese characters failed.
             * Fonts "Meiryo & HGKyokashotai": everything works well.
             * Open PDF properties, you will see all fonts are embedded with built-in encoding.
             * Interest to note that copy collapsed Japanese characters from Acrobat Reader, then
               paste them into a Notepad window, Notepad will show the correct Japanese characters anyway.
               It seems PDF export in CR4E mistaking to choose right typeface for Japanese characters
               from some TTF file.
    3. Open PDFExportSample.rpt in Crystal Report 2008 Designer (trial version), and export it as PDF.
        The result PDF displays both ASCII & Japanese characters without any problem.
    Test environment as below:
    * Windows XP Professional SP3 (Japanese) with MS Office which including extra fonts (i.e. HGKyokashotai)
    * Font version: MS Gothic, Mincho, Meiryo, all in Version 5.0
        You can download MS Meiryo from Microsoft's Site:
        http://www.microsoft.com/downloads/details.aspx?familyid=F7D758D2-46FF-4C55-92F2-69AE834AC928&displaylang=en)
    * Eclipse 3.5.2
    * Crystal Reports for Eclipse, V2, 12.2.207.r916
    Can this problem be fixed? If yes how long will it take to release a patch?
    We really looking forward to a solution before abandoning CR4E.
    Thanks for any reply.

    I have created a [simple PDF file|http://sites.google.com/site/cr4eexportpdf/inside-the-pdf/simple.pdf?attredirects=0&d=1] exported from CR4E. It is expected to display "漢字" (or in unicode as "\u6F22\u5B57"), but instead being rendered in different ones of "殱塸" (in unicode as "\u6BB1\u5878").
    Look inside into this simple PDF file (you can just open it with your favorite text editor), here is its page content:
    8 0 obj
    <</Filter [ /FlateDecode ] /Length 120>>
    stream ... endstream
    endobj
    Decode this stream, we get:
    /DeviceRGB cs
    /DeviceRGB CS
    q
    1 0 0 1 0 841.7 cm
    13 -13 569.2 -815.7  re W n
    BT
    1 0 0 1 25.75 -105.6 Tm     <-- text position
    0 Tr
    /ttf0 10 Tf                 <-- apply font
    0 0 0 sc
    ( !)Tj                      <-- show glyphs [20, 21], which index is to embedded TrueType font subset
    ET
    Q
    The only embeded font subset is defined as:
    9 0 obj /ttf0 endobj
    10 0 obj /AAAAAA+MSGothic endobj
    11 0 obj
    << /BaseFont /AAAAAA+MSGothic
    /FirstChar 32
    /FontDescriptor 13 0 R
    /LastChar 33
    /Subtype /TrueType
    /ToUnicode 18 0 R                            <-- point to a CMap object
    /Type /Font
    /Widths 17 0 R >>
    endobj
    12 0 obj [ 0 -140 1000 859 ] endobj
    13 0 obj
    << /Ascent 860
    /CapHeight 1001
    /Descent -141
    /Flags 4
    /FontBBox 12 0 R
    /FontFile2 14 0 R                            <-- point to an embedded TrueType font subset
    /FontName /AAAAAA+MSGothic
    /ItalicAngle 0
    /MissingWidth 1000
    /StemV 0
    /Type /FontDescriptor >>
    endobj
    The CMap object after decoded is:
    18 0 obj
    /CIDInit /ProcSet findresource begin 12 dict begin begincmap /CIDSystemInfo <<
    /Registry (AAAAAB+MSGothic) /Ordering (UCS) /Supplement 0 >> def
    /CMapName /AAAAAB+MSGothic def
    1 begincodespacerange <20> <21> endcodespacerange
    2 beginbfrange
    <20> <20> <6f22>                         <-- "u6F22"
    <21> <21> <5b57>                         <-- "u5B57"
    endbfrange
    endcmap CMapName currentdict /CMap defineresource pop end end
    endobj
    I can write out the embedded TrueType font subset (= "14 0 obj") to a file named "[embedded.ttc|http://sites.google.com/site/cr4eexportpdf/inside-the-pdf/embedded.ttf?attredirects=0&d=1]", which is really a tiny TrueType font file containing only the wrong typefaces for "漢" & "字". It seems everything OK except CR4E failed to choose right typefaces from the TrueType file (msgothic.ttc).
    Is it any help? I am looking forward to any solution.

  • Write Japanese characters to CSV

    In my JSP I am doing the following
    <head>
          <script type="text/javascript">
          var newWindow;
          function checkForCSVExport()
                <% 
                      String results = (String) session.getAttribute("EXCEL_FB_DATA") ;
                      String URL = (String) Form.makeUrl(request,"/custom/jsp/search/load_csv.jsp");
                      if(null != results)
                            out.println(
                                        "newWindow=window.open(\""
                                        + URL
                                        + "\", \"export\", \"toolbar=no,menubar=no,width=200,height=200,resizable=no\");");
                %>
    </script>
    </head>the String results contain comma seperated values with Japanese characters in them. We need to write these Japanese characters to CSV. For this we have a JSP page called load_csv.jsp. Here are its contents
    <%@ page contentType="application/csv; charset=UTF-8" %>
    <%@ page import="com.components.search.SearchResultsFeedback"%>
    <%
          String strFBResults = (String)session.getAttribute("EXCEL_FB_DATA");
          String strFBHdr = (String)session.getAttribute("EXCEL_FB_HEADER");
          response.setHeader("Content-Disposition", "attachment;filename=searchresult.csv");
    %>
    <%=strFBHdr%>
    <%=strFBResults%>this works fine for English characters but Japanese characters are displayed as junk characters in CSV. Please help.

    Thanks for replying. I am using UTF-8 throughout.
    Also, there is an interesting thing that I observed. If I right click on the CSV file and select "Open with --> Word" then the word application asks me
    "Select the encoding that makes your document readable". If I select "Unicode (UTF-8)", Japanese characters appear perfectly fine. But why doesn't this happen in excel as that is also UTF-8 compliant since I can type japanese characters into the same excel file.

  • Japanese Characters in Tables

    I have a table which stores name,address and country of emplyees.
    I see that address for most of the employees from Japan are shown as SII¿¿5F
    It looks like its a combination english and japanese characters where japanese characteres are displayed as ¿
    Is there any way to read these records?

    Thanks.
    Changing my question a little.
    Can i view already stored data(japanese characters) with following NLS settings?
    SQL> SELECT * FROM v$nls_parameters;
    PARAMETER VALUE
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_CHARACTERSET UTF8
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY $
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    Note: Db had same settings when data was stored
    Thanks,
    Coolguy

Maybe you are looking for

  • Response message error in queue when using ABAP proxy

    I am using an abap proxy for the first time, and some of the response messages are getting stuck in the sender's (R/3) inbound queue.  To resolve them, I have to either reset the status or execute the LUW.  Sometimes, this doesn't even work.  Here is

  • Third party billing

    Hello Experts, I have created a PR , then a PO and sent to customer this PO. The next step should be for the Wholesaler to bill me and then I bill the customer? What should be configured to allow payment to my wholesaler? Is billing type dependent or

  • DATAPROVIDER JAVA CREATOR 2.1

    Good Afternoon, Well I'm having a problem with Java creator studio, every time I change a person for registration of a person in my system, it changes, but when I change the next person he picks up the name of the person that can be this? ? Below is

  • Problems importing pics to iphoto

    I'm working on a slideshow and I'm having trouble importing some photos to iphoto. I've followed the instructions by going to "add to library" on iphoto, I've selected the pic files but I get an edit that says "Unreadable Files: 17 the following file

  • Air print it doesn't wk

    How do I get air print to wk On my iPad