Inserting double-byte (Japanese) characters

I am trying to insert a < 30 character Japanese string into a
table with a column of type VARCHAR2(30), but I get the
Exception:
ORA-01401: inserted value too large for column
which is fairly self explanatory.......
I'm using the thin driver against Oracle 7.3.4.
We have a C++ application using OCI which does exactly the same
function successfully. So, my question is how do I tell the thin
driver the character set in use in the database? I think it would
be by a driver property, or some NLS_LANG equivalent, but I can't
find any documentation.
Help someone, please. Many Thanks.
Mark Phipps
null

Mark Phipps (guest) wrote:
: I am trying to insert a < 30 character Japanese string into a
: table with a column of type VARCHAR2(30), but I get the
: Exception:
: ORA-01401: inserted value too large for column
: which is fairly self explanatory.......
: I'm using the thin driver against Oracle 7.3.4.
: We have a C++ application using OCI which does exactly the same
: function successfully. So, my question is how do I tell the
thin
: driver the character set in use in the database? I think it
would
: be by a driver property, or some NLS_LANG equivalent, but I
can't
: find any documentation.
: Help someone, please. Many Thanks.
: Mark Phipps
Try different drivers. i.e. classes102.zip or classes111.zip
or check your drivers. Or make sure you're including the right
one in your classpath.
Ciao
null

Similar Messages

  • SQL Developer - displaying double byte (Japanese) characters

    I have installed SQL Developer on an English Windows XP system with Asian regional support installed.
    SQL Developer > preferences > database > set language, territory to Japan/Japanese.
    The strings are displaying as square boxes which means the characters are unable to display, I have verified on a Japanese machine that the characters to display with the same version of SQL Developer.
    Any ideas on how to get these characters to display?

    Hi user625396,
    Square boxes mean font not installed, for example on my Linux box for displaying Chinese:
    Chinese characters in linux see http://isis.poly.edu/~qiming/chinese-debian-mini-howto.html basically:
    1. create a jre/lib/fonts/fallback directory if necessary and
    2. copy or link the fonts files into this directory. (On my Linux system: /usr/share/fonts/truetype/arphic/gbsn00lp.ttf )
    3. I will install Japanese when the business need arises.
    Please confirm this fallback directory works for Japanese and Windows XP, with / changed to \ and the path to a Japanese font replacing /usr/share/fonts/truetype/arphic/gbsn00lp.ttf .
    I had a quick look for a windows XP/Japanese/fallback font page but I could not see one.
    -Turloch

  • Form English Char ( Single Byte  ) TO Double Byte ( Japanese Char )

    Hello EveryOne !!!!
    I need Help !!
    I am new to Java ,.... I got assignment where i need to Check the String , if that string Contains Any Non Japanse Character ( a~z , A~Z , 0 -9 ) then this should be replaced with Double Byte ( Japanese Char )...
    I am using Java 1.2 ..
    Please guide me ...
    thanks and regards
    Maruti Chavan

    hello ..
    as you all asked Detail requirement here i an pasting C code where 'a' is passed as input character ..after process it is giving me Double Byte Japanese "A" .. i want this to be Done ..using this i am able to Convert APLHA-Numeric from singale byte to Doubel Byte ( Japanse ) ...
    Same program i want to Java ... so pleas guide me ..
    #include <stdio.h>
    int main( int argc, char *argv[] )
    char c[2];
    char d[3];
    strcpy( c, "a" ); // a is input char
    d[0] = 0xa3;
    d[1] = c[0] + 0x80;
    printf( ":%s:\n", c ); // Orginal Single byte char
    printf( ":%s:\n", d ); // Converted Double Byte ..
    please ..
    thax and regards
    Maruti Chavan

  • Function module to control printing of double byte chinese characters

    Hi,
    My sapscript printing of GR Slip often overflow to the next line whenever a line item encountered article desciption text in CHINESE.
    The system login as "EN" but we do maintain article description in ENGLISH, ChINESE and mixture of both.
    This result in different field length when printing.
    Is there a way to control it and ensure that it will not overflow to the next line?
    How does SAP standard deals with this sort of printing, single & double byte chars?
    Please assist.

    This is the code that solved our issue.
    Besides we set the InfoObject to have NO master data attributes: it was just used as text attribute in an DSO, not as dimensional attribute in a cube. This solved the issue of the SID value generation error.
    FUNCTION z_bw_replace_sp_char.
    ""Local Interface:
    *"  IMPORTING
    *"     REFERENCE(I_STRING)
    *"  EXPORTING
    *"     REFERENCE(O_STRING)
      FIELD-SYMBOLS: <ic> TYPE x.
    Strings with other un-allowed char -
    DATA:
    ch1(12) TYPE x VALUE
    '410000204200002043000020',
    ch2(12) TYPE x VALUE
    '610000206200002063000020'.
      DATA:
      x8(4) TYPE x,
      x0(2) TYPE x VALUE '0020',
      x0200(2) TYPE x VALUE '0200'.
      DATA: v_len TYPE sy-index,
            v_cnt TYPE sy-index.
      o_string = i_string.
      v_len = STRLEN( o_string ).
    # sign
      IF v_len = 1.
        IF o_string(1) = '#'.
          o_string(1) = ' '.
        ENDIF.
      ENDIF.
    ! sign
      IF o_string(1) = '!'.
        o_string(1) = ' '.
      ENDIF.
      DO v_len TIMES.
        ASSIGN o_string+v_cnt(1) TO <ic> CASTING TYPE x.
        IF <ic> <> x0200. "$$$$$$$$$$$$$$$$$$$$$$
          IF <ic> >= '0000' AND
             <ic> <= '1F00'.  " Remove 0000---001F
            o_string+v_cnt(1) = ' '.
         ELSE.
           CONCATENATE <ic> x0 INTO x8 IN BYTE MODE.
           unassign <ic>.
           SEARCH ch1 FOR x8 IN BYTE MODE.
           IF sy-subrc <> 0.
             SEARCH ch2 FOR x8 IN BYTE MODE.
             IF sy-subrc = 0.
               o_string+v_cnt(1) = ' '.
             ENDIF.
           ELSE.
             o_string+v_cnt(1) = ' '.
           ENDIF.
          ENDIF.
        ENDIF. "$$$$$$$$$$$$$$$$$$$$$$
        v_cnt = v_cnt + 1.
      ENDDO.
    ENDFUNCTION.

  • Double-byte Chinese characters between Flash Media Server and WebLogic 9.2

    i have installed a FMS (Flash Media Server 4.0 developer edition) and a weblogic server on linux. The two servers communicate with each other via WebService (Axis2 on Weblogic 9.2). i can send some XML message from FMS to the Weblogic server and receive message. but the returned message from the Weblogic server is as follows:
    <?xml version="1.0" encoding="utf-8"?>
    <root>
    <response>
    <result>1</result>
    <reports >
    <test>x8054x901Ax6027x6D4Bx8BD5</test>
    </reports>
    </response>
    </root>
    the returned messages between the <test> and </test> tag can be decoded to "测试联通性" in Chinese via utf-8. why not the returned messages just show the result as :
    <test>测试联通性</test>
    Later i made another test: copy the war file in weblogic server to a apache-tomcat server, then the FMS send XML message to apache-tomcat and got the correct message as follows:
    <?xml version="1.0" encoding="utf-8"?>
    <root>
    <response>
    <result>1</result>
    <reports >
    <test>测试联通性</test>
    </reports>
    </response>
    </root>
    then what is the problem with my weblogic? what shall i do to solve the problem? thank you for anyone who gives me some advice.

    i have installed a FMS (Flash Media Server 4.0 developer edition) and a weblogic server on linux. The two servers communicate with each other via WebService (Axis2 on Weblogic 9.2). i can send some XML message from FMS to the Weblogic server and receive message. but the returned message from the Weblogic server is as follows:
    <?xml version="1.0" encoding="utf-8"?>
    <root>
    <response>
    <result>1</result>
    <reports >
    <test>x8054x901Ax6027x6D4Bx8BD5</test>
    </reports>
    </response>
    </root>
    the returned messages between the <test> and </test> tag can be decoded to "测试联通性" in Chinese via utf-8. why not the returned messages just show the result as :
    <test>测试联通性</test>
    Later i made another test: copy the war file in weblogic server to a apache-tomcat server, then the FMS send XML message to apache-tomcat and got the correct message as follows:
    <?xml version="1.0" encoding="utf-8"?>
    <root>
    <response>
    <result>1</result>
    <reports >
    <test>测试联通性</test>
    </reports>
    </response>
    </root>
    then what is the problem with my weblogic? what shall i do to solve the problem? thank you for anyone who gives me some advice.

  • Double byte characters in a String - Help needed

    Hi All,
    Can anyone tell me how to find the presence of a double byte ( Japanese ) character in a String. I need to check whether the String contains a Japanese character or not. Thanx in advance.
    Ramya

    /** returns true if the String s contains any "double-byte" characters */
    public boolean containsDoubleByte(String s) {
      for (int i=0; i<s.length(); i++) {
        if (isDoubleByte(s.charAt(i)) {
          return true;
      return false;
    /** returns true if the char c is a double-byte character */
    public boolean isJapanese(char c) {
      if (c >= '\u0100' && c<='\uffff') return true;
      return false;
    // simpler:  return c>'\u00ff';
    /** returns true if the String s contains any Japanese characters */
    public boolean containsJapanese(String s) {
      for (int i=0; i<s.length(); i++) {
        if (isJapanese(s.charAt(i)) {
          return true;
      return false;
    /** returns true if the char c is a Japanese character. */
    public boolean isJapanese(char c) {
      // katakana:
      if (c >= '\u30a0' && c<='\u30ff') return true;
      // hiragana
      if (c >= '\u3040' && c<='\u309f') return true;
      // CJK Unified Ideographs
      if (c >= '\u4e00' && c<='\u9fff') return true;
      // CJK symbols & punctuation
      if (c >= '\u3000' && c<='\u303f') return true;
      // KangXi (kanji)
      if (c >= '\u2f00' && c<='\u2fdf') return true;
      // KanBun
      if (c >= '\u3190' && c <='\u319f') return true;
      // CJK Unified Ideographs Extension A
      if (c >= '\u3400' && c <='\u4db5') return true;
      // CJK Compatibility Forms
      if (c >= '\ufe30' && c <='\ufe4f') return true;
      // CJK Compatibility
      if (c >= '\u3300' && c <='\u33ff') return true;
      // CJK Radicals Supplement
      if (c >= '\u2e80' && c <='\u2eff') return true;
      // other character..
      return false;
    /* NB CJK Unified Ideographs Extension B not supported with 16-bit unicode. Source: http://www.alanwood.net/unicode */
    }

  • ASCII representations of double-byte characters

    My file contains ASCII representations of double-byte CJK characters (output of native2ascii). How do I restore them back to the original native characters?
    I mean, when I load the file with FileInputStream, what I get are all strings like \uabcd. How do I get the characters represented by these strings?

    My file contains ASCII representations of double-byte
    CJK characters (output of native2ascii. How do
    I restore them back to the original native
    characters?
    I am no expert in unicode so I don't know if this is correct, but I assume that if a String starts with "\u" then there will be 4 more characters that are a hexadecimal representation of the char value. If that's right, then you should be able to parse out the "\uxxxx" and convert it to a char by parsing the hex. For example//the variable unicode is a String like \uabcd
    String hex = unicode.substring(2);
    char result = (char) (Integer.parseInt(hex,16));

  • Implementation of double-byte character support in the Database

    Hi experts,
    There is an oracle 10.2.0.5 standard edition database running on windows platform. The application team needs to add a column of the datatype double byte (chinese characters) to an existing table. By doing so, will we have to change the character set of the database ? Would implementing this cause any impact on the database ?
    Is it possible to change the existing single byte column to the double byte column in a table?
    Edited by: srp on Dec 18, 2010 2:21 PM

    I think you should post your request to Globalization Support forum Globalization Support with following details:
    what is the APIs used to write data ?
    what are the APIs used to read data ?
    what is the client OS ?
    what is the value of NLS_LANG environment variable in client environment
    Note that command line SQL*Plus under Windows does not display Unicode data: try to use instead SQL Developer.
    Example:
    1. run in SQL*Plus:
    SQL> desc t;
    Name                                      Null?    Type
    X                                                  NUMBER(38)
    Y                                                  NVARCHAR2(10)
    SQL> insert into t values(1, unistr('\8349'));
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> select * from t;
             X Y
             1 ■
    SQL> select dump(y, 1016)  from t;
    DUMP(Y,1016)
    Typ=1 Len=2 CharacterSet=AL16UTF16: 83,49Run in SQL Developer:
    select * from t
    X                      Y         
    1                      草       

  • Implementation of double byte character in oracle 10.2.0.5 database

    Hi experts,
    There is an oracle 10.2.0.5 standard edition database running on windows 2003 platform. The application team needs to add a column of the datatype double byte (chinese characters) to an existing table. The database character set is set to WE8ISO8859P1 and the national characterset is AL16UTF16. After going through the Oracle Documentation our DBA team found out that its possible to insert chinese characters into the table with the current character set.
    The client side has the following details:
    APIs used to write data--SQL Developer
    APIs used to read data--SQL Developer
    Client OS--Windows 2003
    The value of NLS_LANG environment variable in client environment is American and the database character set is WE8ISO8859P1 and National Character set is AL16UTF16.
    We have got a problem from the development team saying that they are not able to insert chinese characters into the table of nchar or nvchar column type. The chinese characters that are being inserted into the table are getting interpreted as *?*...
    What could be the workaround for this ??
    Thanks in advance...

    For SQL Developer, see my advices in Re: Oracle 10g - Chinese Charecter issue and Re: insert unicode data into nvarchar2 column in a non-unicode DB
    -- Sergiusz

  • Can Double Byte Codes be used in SharePoint 2013?

    Hello,
    I am looking to create a SharePoint but for Japanese users.  I understand that Double Byte Codes will need to be used when entering information due to the Japanese language.  I have been informed that this can sometimes cause an issue with SharePoint.
    I am looking to test this however I wanted to understand if this can be done / are there any considerations when using DBC's on SharePoint etc and any limitations.
    Any assistance would be greatly appreciated.
    Thanks
    KP

    Hi KP,
    Yes, SharePoint 2013 supports Double Byte Unicode characters:
    If you have non-standard ASCII characters, such as high-ASCII or double-byte Unicode characters, in the SharePoint URL, each of those characters is URL-encoded into two or more ASCII characters when they are passed to the Web browser. Thus, a URL with many
    high-ASCII characters or double-byte Unicode characters can become longer than the original un-encoded URL. The list below gives examples of the multiplication factors:
    High-ASCII characters — for example, (!, ", #, $, %, &, [Space]): multiplication factor = 3
    Double byte Unicode characters — for example, Japanese, Chinese, Korean, Hindi: multiplication factor = 9
    For example, when you translate the names of sites, library, folder, and file in the URL path http://www.contoso.com/sites/marketing/documents/Shared%20Documents/Promotion/Some%20File.xlsx into Japanese, the resulted encoded URL path will become something
    like the following:
    http://www.contoso.com/sites/%E3%83%9E%E3%83%BC%E3%82%B1%E3%83%86%E3%82%A3%E3%83%B3%E3%82%B0/%E6%96%87%E6%9B%B8/DocLib/%E3%83%97%E3%83%AD%E3%83%A2%E3%83%BC%E3%82%B7%E3%83%A7%E3%83%B3/%E3%83%95%E3%82%A1%E3%82%A4%E3%83%AB.xlsx. This path is 224 characters,
    whereas the original URL path is only 94 characters.
    In most cases, one UTF-16 character equals one UTF-16 code unit. However, characters that use Unicode code points greater than U+10000 will equal two UTF-16 code units. These characters include, but are not limited to, Japanese or Chinese surrogate pair
    characters. If your paths include these characters, the URL length will exceed the URL length limitation with fewer than 256 or 260 characters.
    Reference:
    https://technet.microsoft.com/en-us/library/ff919564%28v=office.14%29.aspx?f=255&MSPPError=-2147217396
    Best Regards,
    Eric
    TechNet Community Support
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected].

  • Double Byte Characters for Japanese Kanji Language

    Hello All,
    How can you identify the Double Byte characters and how we can check whether they are double byte characters  or not.
    Can you please tell me the double byte characters from Japanese Kanji language for the characters.
    This is very urgent.I will reward you with points.
    Thanks,
    Karan

    *& Report  YTEST_LOGIC
    REPORT  teched_unicode_solution_1               .
    *** Exercise 1: Distinction between byte and character length
    *** after Unicode enabling
    parameter: param type c.
    PERFORM test1 USING param.
    *  FORM test1
    FORM test1 USING text TYPE c.
      DATA: len1 TYPE i,
            len2 TYPE i,
            off TYPE i.
    DESCRIBE FIELD text LENGTH len1 IN BYTE MODE.
    DESCRIBE FIELD text LENGTH len2 IN CHARACTER MODE.
    WRITE:/ LEN1, LEN2.
    ENDFORM.                                                    "test1
    Use the above code to find which characters are double bytes and which are not...
    The double byte characters will have double the length when they are in the byte mode..
    reward points for helpful answers
    Edited by: Rahul Kavuri on Mar 26, 2008 6:35 PM

  • Japanese double byte characters

    Hi all,
    We have a ECC 6.0 system and we are introducing Japanese markets.
    When we try to see table KNA1: SE16-->KNA1 and we search a word in Japanese language written in single-byte format, there is no problem, the search returns us the entries.
    But when we try to serach the very same word in Japanese written in double-byte format, the search does not return any result.
    How can we activate that the search matches the word entered in double-byte, with single-byte entries in the table??
    Our system is Unicode.
    We have seen some forum results:
    Prerequisites
    u2022 Modify the codepage to level 8000
    u2022 Install Japanese language package for SAP
    u2022 Activate English and Japanese language in Windows
    u2022 Activate Multibyte support in SAP GUI: ALT-F12 -> Options -> I18N tab -> Enable Multibyte support
    u2022 Change SAP font script to u2018Japaneseu2019
    But the thing is that we have SAPGUI 7.20 version, and we do not find the way to activate multibyte...
    Thanks!

    Hi,
    SAP note 1428423 describes the necessary functionality needed.
    However this is not implemented in SAP ERP 6.0 standard.
    The tools described in the note could be used to modify the system. But I think this would be a high effort.
    Best regards,
    Nils Buerckel
    SAP AG

  • Double byte language i.e Japanese or Chinese text in non Unicode System

    Hi,
    I have translated text into Chinese and Japanese in a Unicode system and want to move it into a non Unicode system. Would Chinese/Japanese characters display correctly in a non Unicode system when moved from Unicode system.  I am doing translation in ECC60 or SAP 4.7 Unicode system and moving to SAP 4.7 non Unicode system.
    Thanks
    Balakrishna

    Hi Balakrishna,
    in general the transport between Unicode and Non-Unicode systems is supported.
    However there are restrictions, which are outlined in SAP note 638357.
    In your case it is a prerequisite that the objects to be transported are language dependent (text lang. flag is set on the language key - see SAP note 480671) and the languages are properly setup in the target systems.
    For double byte data there is a specific issue when transferring data from Unicode to Non-Unicode:
    In a Non-Unicode system, one double-byte character needs two bytes, therefore e.g. in a 10 char field, 5 double byte chars are fitting. In a Unicode system, you can insert 10 double-byte chars in a 10 char field.  Hence there is a risk of truncating characters in case of Unicode --> Non-Unicode communication.
    Please also have a look at SAP notes 1322715 and 745030.
    Best regards,
    Nils

  • Using Double Byte Characters in URL For Session Variables

    When I supply the value for a session variable in the URL for an IRPT page where the value contains double byte characters, Japanese in this case, the characters are corrupted by the time they are entered for the session variables.  Does anyone know a solution to this problem or experience in this area?  Currently using xMII 11.5 SR3.

    Hi Bryan,
    I would suspect that under the covers the session variable is of datatype string.  For double byte characters, it would need to be wstring.  There is a better explanation to be found at:
    Link: [Kanji and Java Datatypes|http://www.unix.com.ua/orelly/java-ent/jenut/ch10_04.htm] or you can try google on  Kanji Datatype  OR Kanji Java Datatype
    It could also be a problem with the operating system which I ran into about 10 years ago, but I would hope that Microsoft had moved beyond that by now.
    Maybe some more technical folks could chime in to confirm or deny my explanation.
    Mike
    Edited by: Michael Appleby on Jul 8, 2008 5:23 PM

  • Handling Multi-byte/Unicode (Japanese) characters in Oracle Database

    Hello,
    How do I handle the Japanase characters with Oracle database?
    I have a Java application which retrieves some values from the database; makes some changes to these [ex: change value of status column, add comments to Varchar2 column, etc] and then performs an UPDATE back to the database.
    Everything works fine for the English. But NOT for Japanese language, which uses Multi-byte/Unicode characters. The Japanese characters are garbled after the performing the database UPDATE.
    I verified that Java by default uses UTF16 encoding. So there shouldn't be any problem with Java/JDBC.
    What do I need to change at #1- Oracle (Database) side or #2- at the OS (Linux) side?
    /* I tried changing the NLS_LANG value from OS and NLS_SESSION_PARAMETERS settings in Database and tried 'test' insert from SQL*plus. But SQL*Plus converts all Japanese characters to a question mark (?). So could not test it via SQL*plus on my XP (English) edition.
    Any help will be really appreciated.
    Thanks

    Hello Sergiusz,
    Here are the values before & after Update:
    --BEFORE update:
    select tar_sid, DUMP(col_name, 1016) from table_name where tar_sid in ('6997593.880');
    /* Output copied from SQL-Developer: */
    6997593.88 Typ=1 Len=144 CharacterSet=UTF8: 54,45,53,54,5f,41,42,53,54,52,41,43,54,e3,81,ab,e3,81,a6,4f,52,41,2d,30,31,34,32,32,e7,99,ba,e7,94,9f,29,a,4d,65,74,61,6c,69,6e,6b,20,e3,81,a7,e7,a2,ba,e8,aa,8d,e3,81,84,e3,81,9f,e3,81,97,e3,81,be,e3,81,97,e3,81,9f,e3,81,8c,e3,80,81,52,31,30,2e,32,2e,30,2e,34,20,a,e3,81,a7,e3,81,af,e4,bf,ae,e6,ad,a3,e6,b8,88,e3,81,bf,e3,81,ae,e4,ba,8b,e4,be,8b,e3,81,97,e3,81,8b,e7,a2,ba,e8,aa,8d,e3,81,a7,e3,81,8d,e3,81,be,e3,81,9b,e3,82,93,2a
    --AFTER Update:
    select tar_sid, DUMP(col_name, 1016) from table_name where tar_sid in ('6997593.880');
    /* Output copied from SQL-Developer: */
    6997593.88 Typ=1 Len=144 CharacterSet=UTF8: 54,45,53,54,5f,41,42,53,54,52,41,43,54,e3,81,ab,e3,81,a6,4f,52,41,2d,30,31,34,32,32,e7,99,ba,e7,94,9f,29,a,4d,45,54,41,4c,49,4e,4b,20,e3,81,a7,e7,a2,ba,e8,aa,8d,e3,81,84,e3,81,9f,e3,81,97,e3,81,be,e3,81,97,e3,81,9f,e3,81,8c,e3,80,81,52,31,30,2e,32,2e,30,2e,34,20,a,e3,81,a7,e3,81,af,e4,bf,ae,e6,ad,a3,e6,b8,88,e3,81,bf,e3,81,ae,e4,ba,8b,e4,be,8b,e3,81,97,e3,81,8b,e7,a2,ba,e8,aa,8d,e3,81,a7,e3,81,8d,e3,81,be,e3,81,9b,e3,82,93,2a
    So the values BEFORE & AFTER Update are the same!
    The problem is that sometimes, the Japanese data in VARCHAR2 (abstract) column gets corrupted. What could be the problem here? Any clues?

Maybe you are looking for

  • How to change Profit Center

    Dear PS Gurus, Kindly help me out on this. My Query:- I am trying to create a PO using the T-code ME58. When i am selecting the PR and adopting it i get a message as "Profit Center was set to 9999". However i want to post the data to profit center 13

  • How to import a PSD file into Indesign WITH layers so it can be edited/saved as an InDesign file?

    How to import a Photoshop file into Indesign CS6 so that it's layers show in InDesign and it can then be edited/saved as an InDesign file?

  • HD video

    Hello all, I am new to this forum and haven't been able to find the exact answer to my question. I have a Sony HDR-100 High Def camcorder and I am using Adobe Premiere Elements 7 to try to make movies.  The problem is, the output movies are horrendou

  • Why are my Pictures distorted and not fully filled with color?

    HP ENVY 110 e-All-in-One Printer - D411a.  Why do Printed Pictures come out distorted and and Color not filled in fully?

  • My Rear Camera shows a crack when taking photos

    I have an iPhone 5 running iOS 8.2. When I take photos with the rear camera, there is is crack or line in the upper right corner of every zoomed out photo I take. Can anyone help me out on what could be the problem? I took apart the iPhone 5 and clea