Problem with Arabic characters

Hi:
I don't know if this is the correct place to post the question, but here it goes...
I have an SQL 2005 database, connected via a Linked Server to an Oracle Database.
I have a table in SQL that contains arabic characters, and I need to insert it into an Oracle table.
These characters appear as "????" after being inserted in the oracle table.
I guess I hace some collation/characterset problem, but cannot finf the solution.
The column where the arabic characyers are saved in SQL is defined as:
     [Remarks] [nvarchar](1000) NULL,
And default SQL server collation is: SQL_Latin1_General_CP1_CI_AS
When I do a select on this table, it can see arabic correctly.
ORACLE NLS CONFIGURATION IS:
PARAMETER VALUE
NLS_LANGUAGE AMERICAN
NLS_TERRITORY AMERICA
NLS_CURRENCY $
NLS_ISO_CURRENCY AMERICA
NLS_NUMERIC_CHARACTERS .,
NLS_CHARACTERSET AR8MSWIN1256
NLS_CALENDAR GREGORIAN
NLS_DATE_FORMAT DD-MON-RR
NLS_DATE_LANGUAGE AMERICAN
NLS_SORT BINARY
NLS_TIME_FORMAT HH.MI.SSXFF AM
NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
NLS_DUAL_CURRENCY $
NLS_COMP BINARY
NLS_LENGTH_SEMANTICS BYTE
NLS_NCHAR_CONV_EXCP FALSE
NLS_NCHAR_CHARACTERSET AL16UTF16
NLS_RDBMS_VERSION 11.1.0.6.0
The sql sentence in SQL server that moves the data to oracle is:
INSERT INTO openquery(Oracle_DB, 'select Ticket_NO,Remarks from webportal.webcc_escal_det')
     (Ticket_NO,Remarks )
     SELECT Ticket_NO, Remarks
     FROM Details(nolock)
Thanks in advance!!
Mariana

I think first you should check your character set
SELECT * FROM NLS_DATABASE_PARAMETERS where parameter = 'NLS_CHARACTERSET'
For oracle to display correctly, you need to change your character set that caters to Arabic

Similar Messages

  • Problem with Arabic Character

    Hi all...
    I've a problem and if u please help me.
    My system configuration is :
    Windows 2000 Professional
    Sun JDK 1.3.1
    Oracle 8i
    iplanet web server 4.1
    iplanet app server 6.0 sp3 test drive
    this is my code:
    <%@ page contentType="text/html; charset=ISO-8859-6" language="java" import="java.sql.*, oracle.jdbc.driver.*"%>
    <%
    Driver driverUser = (Driver)Class.forName("oracle.jdbc.driver.OracleDriver").newInstance();
    Connection connConnection = DriverManager.getConnection("jdbc:oracle:thin:@chorcl01:1521:gpa", "sak", "sak");
    String stmSQL = "SELECT company_name FROM transaction_header";
    PreparedStatement pstmStatementUser = connConnection.prepareStatement(stmSQL);
    ResultSet rsResultSet = pstmStatementUser.executeQuery();
    %>
    <html>
    <body>
    <%
    String vararabic = new String(rsResultSet.getString("company_name").getBytes(),"ISO-8859-6");
    %>
    <table>
    <tr>
    <td><%= vararabic %></td>
    </tr>
    </table>
    </body>
    </html>
    When I run this page, I just see some Question Marks (?) insteas of the correct characters(the field filled with Arabic characters).I test this page with utf-8 and wimdows-1256 but nothing happens.
    but still there are question marks.
    The Default encoding of installed JVM is Cp1252, I change it to utf-8 and windows-1256, nothing happens....
    I don't know what's my problem, please HELP...
    Thanks
    Mani

    Hi Mani,
    You are using test drive phase 3 which only allows ezsetup. You can view the Arabic characters by enabling ii8n. This option is avoilabe with custom installation. But in test drive custom installation is not enabled. So you can not view Arabic characters. You can do so with test drive on solaris.
    Thanks
    parsu

  • Problems with special characters like "u00F6" "u00E4" "u00FC"

    Hi,
    we currently face some major problems with special characters like "ö" "ä" "ü". First I thought it is a SP02 specific problem but it still doesn't work with SP03.
    When I start an import the import manager shows correct converted value. "ö"s "ä"s "ü"s are displayed correctly. But after I've done the import I look into the Data manager and I see instead of these characters charcaters like "&#776;". (note: the character should show a little square box - but the forum cannot desplay it)
    How can I solve this problem?
    Regards
    Nico

    Hi Nico
      Just try it out the following steps, hope it solves ur issue.
    <b>To set one or more values back to their original values:</b>
    1. In the Import Manager Source Values grid, select the values that you want to set back to their original values.
    2. Right-click on one of the values and choose Set to Original Value from the context menu, or choose Values > Apply Value Conversion Filter > Set to Original Value from the main menu.
    3. MDME sets each value back to its original value, which appears in black in the Source Values grid.
    <b>NOTE : You can use the Set to Original Value filter only at the value-level and not at the field level.</b>
    thanks
    Alexander

  • Problem with Arabic font: In certain contexts, when writing Arabic with vowel signs (fatha, damma, kasra, sukun) a sequence of sukun   fatha/damma etc. would reverse automatically. Is this a known bug?

    Problem with Arabic font: In certain contexts, when writing Arabic with vowel signs (fatha, damma, kasra, sukun) a sequence of sukun + fatha/damma etc. would reverse automatically. Is this a known bug?
    Example: عَيْنٌ
    would automatically convert to عَيُنْ
    Funnily, it doesn't seem to happen here, but it does when entering text in a web interface (using Firefox, font Bayan) and when using Text Edit.
    Seems to be a problem of a specific font, as e.g. Arial MS Unicode works fine. Any hints?
    Thank you!

    Musaafir wrote:
    I've no idea how i can even start using arabic vowels on Microsoft Word for Apple
    You can't do Arabic on MS Word for Mac.  This app has never supported RTL scripts, so you need to use something else.  Mellel is best, but Pages 5, TextEdit, Nisus Writer, Open/LibreOffice should work OK.
    You switch between languages by using the "flag" menu at the top right of the screen or by using the keyboard shortcuts apple/command plus space.  Go to system prefs/keyboard/shortcuts to make sure that is activated.
    To see which key does what, you use Keyboard  Viewer.
    http://support.apple.com/kb/PH13746
    You place vowels on letters by typing the key for the vowel after the key for the letter.  The vowels are on the option/alt keys, option/alt + a gives you َ

  • Problems with special characters in InputStream or XPath

    Hello everyone,
    I am having problems with special characters such as ", ', - and so on, everything becomes ? (questionmark). (Not with special nordic characthers (å,æ,ø) though.)
    The encoding of the XML is UTF-8, and the server that holds the webservice is setting encoding to UTF-8 also, so no conflict there.
    What I have done is something like this:
    {code}
    String url = "http://www.formula1.com/rss/news/latest.rss"; // This is not the feed I use, it is a custom XML document
    InputSource is = new InputSource(url);
    DocumentBuilderFactory fct = DocumentBuilderFactory.newInstance();
    DocumentBuilder builder = fct.newDocumentBuilder();
    Document doc = builder.parse(is);
    Element rootElement = doc.getDocumentElement();
    XPathFactory factory = XPathFactory.newInstance();
    XPath xPath = factory.newXPath();
    out.print(xPath.evaluate("/rss/channel/title", rootElement)); // The xml is not a RSS feed, this is just an example.
    {code}
    Please let me know what I am doing wrong, or give me a hint of how to solve this in another way.
    Best regards,
    G.Hansen
    Edited by: G.Hansen on Mar 24, 2009 2:39 AM

    Hello, thanks for you reply. I had to rewrite your code a little, and I think that your code would not really compile, if you tried yourself :-)
    I ended up with something like this:
    our environment is running in 1.4 mode, so I could not use the for-each loop :-(
    public String printEntities(String s) {
         char[] sArray = s.toCharArray();
         StringBuffer sb = new StringBuffer();
           for (int i = 0; i < sArray.length; i++) {
             if (sArray[i] > 256)
                  sb.append("&#x" + Integer.toHexString(sArray) + ";");
         } else
              sb.append(sArray[i]);
         return sb.toString();

  • How to send Oracle rowid to servlet? | Problem with national characters.

    There is same possibility how to send rowid to servlet?
    I have now definition like this:
    <af:image source="/imageservlet?Par1=#{bindings.Col1.inputValue}"/>
    But If column contents national characters, servlet methods obtained changed these characters.
    My idea is to use not primary key for row, but use oracle rowid. It is simply possible?
    Use something like this:
    <af:image source="/imageservlet?Rowid=#{bindings.Rowid}"/
    Or Do you have ideas how to solve problem with national characters ?
    Thanks
    FiL

    Hi,
    Although your workaround works.
    I think this is a simple encoding problem.
    I simply need to make sure all parameters and pages are encoded with a char set which contains the national characters you mentioned.
    This is a bit dependent on the exact technology your using, but most can be done via the web.xml:
      <jsp-config>
          <jsp-property-group>
              <url-pattern>*.jsp</url-pattern>
              <page-encoding>UTF-8</page-encoding>
          </jsp-property-group>
      </jsp-config>     This forces all JSP pages to be encoded in UTF-8
    Adding the following parameter sometimes helps as well, although I think this one is a bit dated:
    You said your using a servlet so your servlet needs a similar block for its pattern
      <context-param>
        <param-name>PARAMETER_ENCODING</param-name>
        <param-value>UTF-8</param-value>
      </context-param>If you want to be 100% sure the encoding is set right make sure thepages contain:
    <%@ page contentType="text/html;charset=utf-8"%>Or depending on your view technology the syntax can be a bit different
    -Anton

  • Problem with national characters on windows client

    Hello there,
    I'am having problem with national characters on windows client.
    All national data stored in NVARCHAR2 colums, applications (.net) works fine,
    but in sqlplus:
    select city from test_table;
    - everything ok, sqlplus shows national characters
    select dump(N'<national symbols>') from dual
    - returns
    Typ=96 Len=12: 0,191,0,191,0,191,0,191,0,191,0,191
    select * from test_table where city = N'<national symbols> '
    - always returns nothing
    As i understand the problem in
    sql query text (and national literals) convertion
    to servers "WE8ISO8859P1" encoding, Is it possible
    to solve the issue?
    Thanks in advance
    PS.
    Console in right mode (chcp=1251)
    sqlplus shows russian messages well
    Server (oracle 9 on solaris):
    select * from nls_database_parameters
    NLS_NCHAR_CHARACTERSET AL16UTF16
    NLS_SAVED_NCHAR_CS WE8ISO8859P1
    NLS_LANGUAGE AMERICAN
    NLS_TERRITORY AMERICA
    NLS_CURRENCY $
    NLS_ISO_CURRENCY AMERICA
    NLS_NUMERIC_CHARACTERS .,
    NLS_CHARACTERSET WE8ISO8859P1
    NLS_CALENDAR GREGORIAN
    NLS_DATE_FORMAT DD-MON-RR
    NLS_DATE_LANGUAGE AMERICAN
    NLS_SORT BINARY
    NLS_TIME_FORMAT HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZH:TZM
    NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZH:TZM
    NLS_DUAL_CURRENCY $
    NLS_COMP BINARY
    NLS_LENGTH_SEMANTICS BYTE
    NLS_NCHAR_CONV_EXCP FALSE
    NLS_RDBMS_VERSION 9.2.0.6.0
    Client (windows server 2003, oracle client 10):
    NLS_LANG = RUSSIAN_CIS.CL8MSWIN1251

    N'<national symbols>', being part of an SQL statement, will be converted to the database character set (WE8ISO8859P1) before being parsed. Only if the client and the database are both 10.2 or higher, the client can encode the literal appropriately so that it survives this conversion.
    In earlier versions, you can do the encoding yourself. Instead of the N'<national symbols>' literal use the UNISTR function: UNISTR('\xxxx\yyyy\zzzz'), where U+xxxx, U+yyyy, U+zzzz are Unicode code points of your national characters.
    -- Sergiusz

  • Problem with arabic in 5610xpressmusic

    hii
    i have a problem with arabic language .someone help pleaz how i can read arabic text in my nokia5610xpressmusic

    Take it to a nokia care point and ask them to install the arabic language pack for you.
    Care points/service centres and repair info:
    UK • Europe • Asia-Pacific • USA •
    Canada • Middle East and Africa
    Elsewhere: Click here, select your country, go to the support section, then select repair.

  • Application Hudba (Music) problem with Czech characters

    Hello,
    I have problem with Czech characters for Music files copied into application "Hudba" all Czech letters (š,č,ť...etc.) are translated into Japonies letters (which I could not read at all). Could someone help please.
    Thanks

     You might to consider contacting the app developers http://talk.sonymobile.com/t5/Music/Walkman-Beta-Test/m-p/953659#U953659

  • Problem with some characters in complex objects

    Hi all,
    I've built a webservice which returns a complex object with several fields inside. All fields are public and accessable via getter and setter methods.
    The problem is, that some of these fields contains numbers or underscores in their names.
    For example:
    public int field_a;
    or
    public String house3of4;
    When I try to import these webservice as a model in a Web Dynpro project, it doesn't work until I remove these characters.
    Is this a known problem or is there any solution for it?
    Thanks
    Thomas

    NLS_LANG in registry is "ARABIC_UNITED ARAB EMIRATES.AR8MSWIN1256"
    I use oracle form 10g for developer
    oracle form 9i for database
    when I build a form in client side and make a text with farsi characters, when I run the form,all characters shows me correct in farsi except four characters(گ چ ژ پ)

  • Problem displaying Arabic characters in PDF using Java APIs

    We are experiencing a problem when attempting to display Arabic characters within a PDF document using the Java APIs.
    The relevant Java code is as follows:
    RTFProcessor processor = new RTFProcessor("example.rtf" );
    processor.setOutput( “example.xsl” );
    processor.setExtractXLIFF(true);
    processor.process();
    ByteArrayOutputStream dataOut = new ByteArrayOutputStream();
    DataProcessor dataProcessor = new DataProcessor();
    dataProcessor.setDataTemplate( example.xdt" );
    if( parameterValues != null && parameterValues.length > 0 )
    setReportParameters( dataProcessor, parameterValues ); // method to set any report parameters from the parameterValues list.
    dataProcessor.setConnection( jdbcConnection );
    dataProcessor.setOutput(dataOut);
    dataProcessor.processData();
    FOProcessor foProcessor = new FOProcessor();
    foProcessor.setLocale( locale );
    foProcessor.setData( new ByteArrayInputStream( dataOut.toByteArray() ) );
    foProcessor.setTemplate( “example.xsl” );
    String xliffFileNameAndPath = getXLIFFFile( “example”, locale );
    if( xliffFileNameAndPath != null )
    foProcessor.setXLIFF( xliffFileNameAndPath );
    foProcessor.setOutput( "example.pdf" );
    foProcessor.setOutputFormat(
    outputFormat == PDF_FORMAT ? FOProcessor.FORMAT_PDF : FOProcessor.FORMAT_HTML );
    foProcessor.generate();
    The method getXLIFFFile( ) gets the relevant XLIFF file for the supplied report locale (if it exists) – the three test files that we used were Italian (example_it_IT.xlf), Spanish (example_es_ES.xlf) and Arabic (example_ar_AE.xlf).
    I imported the following JAR files from the XML Publisher release (version 5.6.2) into my Java application: collections.jar, i18nAPI_v3.jar, versioninfo.jar, xdocore.jar and xmlparserv2.jar.
    The output is OK for all three translations in HTML format, using a charset of UTF-8, and for Italian and Spanish in PDF format. However, the Arabic characters display as question marks in PDF format. The same issue occurs if I stream the output as a byte array straight to the HTTP response rather than save within a file.
    Note that the same RTF, XDT and XLIFF files produce the correct output in both HTML and PDF when executed within XMLPublisher.
    Thank you

    Hi
    I had a similar issue with arabic chars. With PDF layout, chars were appearing properly when preveiwd on local m/c, but as I implement file on server, it was displayed as ?????. I had raised a TAR 5798348.993 with oracle about this and they suggested to apply patch 4028294 Oracle Sourcing J Rollup and then patch 4182914. But later my users changed requirement and i cud not apply the patch to test if it works fine.
    But here, you have suggested that installing fonts would do. Is it really that simple? If yes, dont know what is that patch for which I was told to apply.
    Regards
    Varun

  • Problem with Arabic text with flash player 10 32bit under linux platforms

    Arabic encoding characters do not show correct in the right form at all linux distributions .
    note: with windows platforms, it shows arabic characters correctly.
    this video contains captions and subtitles, in the attachement two snapshots,
    First, show the arabic sentences in the right form with "swfdec mozilla" plugins.
    The other, show the arabic words with "adobe flash player 32bit" plugins,
    for example: Arabic font enconding cp-1256
    the character ب
      in the beginning of the word " بـ ".
    in the middle " ـبـ "
    at the end "ـب"
    isolated character "not in word "ب"
    the word
      كتب in the right form
    begin from right with " كـ "
    with adobe flash player
    ب ت ك
    it shows from left to right and use only isolated characters.
    I report to adobe and they response
    Hi Ghost,
    Since you do not have a 64 bit machine please proceed to reinstall the
    32 bit edition.
    For the current release version of Flash player 10, it does support
    Right to left, But you can’t use a text field on stage and have it
    work. When authoring the content in Flash cs4, you would have to use
    the Flash text engine. more information on that can be found here:
    http://help.adobe.com/en_US/ActionScript/3.0_ProgrammingAS3/WS6C0BB8ED-2
    Thank you.
    Regards,
    Thomas S.
    Technical Support Engineer
    Adobe Systems, Inc.
    Get the answers you need, when you need them!
    Knowledge Base: <(><<)>http://www.adobe.com/support>
    Support Plans: <(><<)>http://www.adobe.com/support/programs

    I too am having this problem. On all of my linux distributions, including Ubuntu 8.04, 8.10, knoppix 5,1, fedora 8,9,10 flashplayer 10 is broken in the same way.  It reports that it cannot find webcam, and the settings dialog refuses to recognize mouse clicks.  I've poked around the web quite a bit and have not yet found a solution.  I saw the problem mentioned on an adobe bug reporting site and their response was to close the issue saying it wasn't repeatable on their system.  I tried to re-open it and my post was deleted without comment (I may have posted incorrectly or something)
    Looking here  and elsewhere I see that the problem is very common and I don't see any posted bug-fixes, or even acknowledgement that a bug exists.  Has adobe lost interest in being cross-platform compatible?

  • Problems with Turkish characters not displaying correctly

    A user is having problems with certain Turkish characters in particular fonts not displaying correctly (as shown below).
    The two fonts (Gill Sans and Helvetica LT CondensedLight) are used for English language marketing material.  Ideally, for continuity, the same font would be preferred for our Turkish language marketing material rather than changing it to a similar font.
    Helvetica LT CondensedLight
    Turkish character “ İ ” does not display correctly in either Word or InDesign.
    Gill Sans
    Turkish characters “ ş ” and “ ğ ” display correctly in Word but not InDesign.
    My questions are:
    1. For Gill Sans, why are the characters not displaying in InDesign but are in Word?
    2. I've noticed here: http://www.webtype.com/font/gill-sans-family/#glyphs-tab that the required characters are available within the Gill Sans font.  Is this because the font is newer and includes the additional characters?  Would this work in InDesign?

    One of the big differences between InDesign and Word is that Word will happily substitute glyphs without telling you. So, unless your installs of Word and of InDesign are using different versions of Gill Sans (which is possible, BTW) then your s-with-cedilla and your g-with-macron are actually not in Gill Sans but in some other font. I'm guessing Arial.
    It's possible that you have more than one Gill Sans installed (e.g. one in your Document Fonts folder and another in your system fonts folder) that have different glyph complements. Gill Sans has been around a long time, and the version I have from the 90s has no Turkish support whatsoever.
    As far as the capital-I-with-dot, you can check your glyph coverage in a variety of ways. In both InDesign and Word, you can open up a window that shows you all of the glyphs in a font. In InDesign, find it in Type -> Glyphs. In Word, it's called "Insert Symbol" and you can find it in Word 2010 by going to Insert -> Symbol -> More Symbols. I am guessing that the answer to this question is simply that your cut of Helvetica LT has no cap-I-with-dot. You'd need to pick a font that actually has support for Turkish. On WIndows, I use a very full-featured freebie called BabelMap to check font coverage. They have a Web version here:
    http://www.babelstone.co.uk/Unicode/babelmap.html
    but I personally prefer the downloadable .exe file.

  • File XML Content Conversion: Problem with special characters

    Hello,
    in a file sender cc content conversion is used to transform a flat structure to XML. What we experiencecd is that the message mapping failed due to a character that was not allowed in XML:
    I was assuming that the file content conversion just creates XML messages with allowed characters. Is there any way to configure content conversion to remove control characters which are not allowed in XML? Unfortunately the sender system cannot be modified.
    Thank you.

    Hi Florian,
      Please use this UDF to remove special characters which prevent XML messages to form properly.
    public static String removeSpecialChar(String s)
              try
                   s=s.replaceAll("&","& amp ;");
                   s=s.replaceAll("<"  , "  & lt ;");
                   s=s.replaceAll(">", "& gt ;");
                   s=s.replaceAll("'", "& apos ;");
                   s=s.replaceAll("\"", "& quot ;");
              catch(Exception e)
                   e.printStackTrace();
              return s;
    Please remove spaces between characters within double quotes. I have added them because otherwise you can't see this code properly. Please check this below link , please replace the characters with proper values as the display is causing a problem here   
    http://support.microsoft.com/kb/316063
    regards
    Anupam
    Edited by: anupamsap on Jul 7, 2011 4:22 PM
    Edited by: anupamsap on Jul 7, 2011 4:23 PM

  • Problem with Polish Characters while downloading data using GUI_DOWLOAD

    Hi,
    I was trying to create a file on the presentation server using GUI_DOWLOAD.
    I am facing a problem with the data that contains polish characters.. like Ą , Ł..
    There is  a vendor numer in the system as FRĄCZEK, when i was tring to download, the vendor number is downloaded as FR¥CZEK, even i tried to download from SE16 also, still i am facing the same problem.
    I am facing the above problem with soem other polish characters..
    Is there any solution  to download the data properly.. ?
    Guys please help me.
    Thanks in advacnce.
    Regards,
    Sriram.
    Edited by: Srirama Murthy Maddirala on Oct 27, 2009 8:10 AM

    Hi Sriram,
    <li>Get the front-end code page using NLS_GET_FRONTEND_CP.
    <li>Pass the code page through GUI_DOWNLOAD.
    REPORT  ztest.
    TYPE-POOLS:abap.
    DATA:frontend_codepage TYPE  cpcodepage.
    DATA:codepage          TYPE  abap_encoding.
    "Get the Frontend Code page
    CALL FUNCTION 'NLS_GET_FRONTEND_CP'
      EXPORTING
        langu             = sy-langu
      IMPORTING
        frontend_codepage = frontend_codepage.
    "Pass the codepage through GUI_DOWNLOAD function module
    codepage = frontend_codepage.
    CALL FUNCTION 'GUI_DOWNLOAD'
      EXPORTING
        filename = 'C:\temp\data.xls'
        filetype = 'ASC'
        codepage = codepage
      TABLES
        data_tab = it_data.
    Thanks
    Venkat.O

Maybe you are looking for

  • IMac hard drive failure query

    Hello, my IMac harddrive has failed (bought in 2008). I have managed to copy my photos, music, documents etc onto an external HD (using Linux as my Mac boot disks were not having any effect). My question is, can anyone tell me what other info I shoul

  • Open url with createTempFile fails

    I am deploying a Java application which needs to use a web browser-based application to draw some data and save the data as snapshots. It works fine on my localhost version, but in the deployment environment, it fails. The browser launches on the dep

  • I want to burn a DVD of a slideshow I made using iPhoto.  Can anyone tell me how?

    I have made a slide show from photos in iPhoto.  It works fine and I have published it to my "me gallery".  I now want to make a DVD of this slideshow.  I have tried to use an old version of Toast and I get an error message after the slideshow has be

  • Itunes 11 is crap

    i won't buy anything in the itunes store until apple gives us back the feature of having 2 windows open at the same time. and not only itunes - with every new version of iphoto i got ****** off more. why does apple make everything worse with each new

  • Muvo Micro N200 Random Shut

    Hi I recently purchased a black Muvo Micro N200. It was working great for awhile, but suddenly it just started to shut off. Its random. Then like everyone else, it would shut off when i barely turned it on and see the display muvo n200. Im reformatti