SAP XI 2, JDBC Inbound Adapter. Non Unicode data problem.

Hi All,
As far a I understand the data from inbout JDBC adapter in XI 2 should be in Unicode format. Is it possible to accept non unicode data using this adapter? If answer is yes then how to specify correct data encoding?
Best Regards.
Victor.

Hi,
Yes, tecnicaly XMB is aware of the encoding and not the adapter, but when you read data from file you can tell to the adapter that file is in specific encoding e.g. file.encoding = "ISO-8859-5"
This value is used by adapter to create xml file wich is passed to XMB. By default the passed XML is UTF8 encoded.
So I need similar setting for JDBC adapter and my question is if I can do it?
Best Regards.
Victor.

Similar Messages

  • Encoding Problem: non-Unicode Data to Unicode format of XI

    Hi SDN,
    I have a JDBC sender to SAP BW scenario. The database is MS SQL server. 
    The code page of db CP1CIAS
    Description:SQL Server Sort Order 52 on Code Page 1252 for non-Unicode Data
    Some fields with values like <b>ZAK&#x0;ADY TWORZYW SZTUCZNYCH</b> are failing in XI Mapping with error
    <b>Fatal Error: com.sap.engine.lib.xml.parser.Parser~
    XMLParser : #0 not allowed in Character data sections
    in the trace.</b>
    Please help how should i get over this code page errors. By installing this code page on XI server help?

    There is no such global setting, this is b/c your source has Unicode I trust, and the only one other thing to try would be this:
    Arthur My Blog

  • JDBC sender adapter provides incomplete data

    Hi,
    I have a JDBC sender adapter which reads data from a MySQL table, and passes it to a proxy which writes the data then in a SAP table.
    - Query SQL statement = SELECT Field1, Field2, Field3, Field4 FROM jdbctest WHERE Processed = 'N'
    - Update SQL statement = UPDATE jdbctest SET Processed='Y' WHERE Processed='N'
    I did a stress test with a program that writes data simultaneously to the MySQL table (up to 10000 records), and the JDBC sender adapter which checks the table every 15 seconds.
    At the end, my SAP table contained only 9965 records.
    This means that probably there have been records added between the Query and the Update statement (So they haven't been read, but their status is changed).
    Does anyone have a solution for this?
    Best regards,
    Nicolas De Corte

    Hello Dharamveer,
    I don't really understand your answer.
    My point is that the JDBC adapter only processed 9965 records out of 10000, but it updated all 10000.
    So now in my MySQL table there are 10000 records with status Processed, but only 9965 have been really processed. And I think this is because records have been added between the execution of query and the update statement.

  • Importing non-unicode data into unicode 10gR2 database

    Hi:
    I will have to import non-unicode data into unicode 10gR2 database. The systems the data is coming from are the following: CODA, Timberline, COMMS, CMS, LIMS. These are all RDBMS, sql-enabled systems. We are talking about pretty big amounts of data (a couple hundred GB combined).
    Did anybody go through a similar exersize?
    I know I'll have to setup nls_length_semantics to CHAR.
    What other recommendations could you guys give?
    TIA,
    Greg

    I think "nls_length_semantics" isn't mandatory at this point, and you must extract a little quantity of information from every source and do some probes injecting them into the Oracle10g database.

  • How can i Configure JDBC Receiver Adapter to INSERT data in multiple tables

    Hi ,
    I want to INSERT data in 4 diffrent tables.
    Input Data Format is:
    <H1>
      <d1></d1>
       <d2></d2>
    </H1>
    I want to insert <H1> Header data in particular table & all the details data<D1> in their corresponding table.
    Note that Header Can't repeat but Details records are repeating again & again.
    Is it possible to write Multiple INSERT STATEMENT in single Data Type.??
    I mean to say can we write STATEMENT keyword again & again. if yes HOW ?

    Hi
    see below links.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/cf/406642ea59c753e10000000a1550b0/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/64/ce4e886334ec4ea7c2712e11cc567c/content.htm
    Regards
    Chilla..

  • JDBC Sender Adapter - No Namespace found

    Hello
    I am currently using PI 7.1. The scenario is JDBC to ECC ABAP Client Proxy. When the JDBC Sender Adapter sends the data to the Integration Server, the root element does not contain the xmlns attribute. This root element is constructed from the Document Name of the JDBC Sender Adapter attribute.
    I am getting the following error:
      <?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
    - <!--  Request Message Mapping
      -->
    - <SAP:Error xmlns:SAP="http://sap.com/xi/XI/Message/30" xmlns:SOAP="http://schemas.xmlsoap.org/soap/envelope/" SOAP:mustUnderstand="">
      <SAP:Category>Application</SAP:Category>
      <SAP:Code area="MAPPING">EXCEPTION_DURING_EXECUTE</SAP:Code>
      <SAP:P1>com/sap/xi/tf/_MM_HIS_Material_Consumption_</SAP:P1>
      <SAP:P2>com.sap.aii.mappingtool.tf7.IllegalInstanceExcepti</SAP:P2>
      <SAP:P3>on: Cannot create target element /ns0:MT_Material_</SAP:P3>
      <SAP:P4>Consumption_ECC/Material. Values missing in queue~</SAP:P4>
      <SAP:AdditionalText />
      <SAP:Stack>Runtime exception occurred during application mapping com/sap/xi/tf/_MM_HIS_Material_Consumption_; com.sap.aii.mappingtool.tf7.IllegalInstanceException: Cannot create target element /ns0:MT_Material_Consumption_ECC/Material. Values missing in queue~</SAP:Stack>
      <SAP:Retry>M</SAP:Retry>
      </SAP:Error>
    When I perform a test at RWB, the message gets posted successfully. Once posted, the payload of this message contains the xmlns attribute.
    Here is the root element from both the instances:
    From RWB:  <ns0:MT_Material_Consumption_JDBC xmlns:ns0="http://xxx.com/xxx/finance">
    From JDBC Adapter:  <ns0:MT_Material_Consumption_JDBC>
    Could you please let me know, what has to be done to obtain the xmlns in the sender payload?
    Thank you in advance.
    John Jeyaraj
    Edited by: John jeyaraj on Apr 7, 2011 8:22 AM

    Hi Hiren
    When I test the MM without the xmlns attribute, I get the following error:
    The prefix "ns0" for element "ns0:MT_Material_Consumption_JDBC" is not bound.
    When I add the xmlns attribute, the mapping is successful.
    Thanks
    John

  • Unicode and non-unicode string data types Issue with 2008 SSIS Package

    Hi All,
    I am converting a 2005 SSIS Package to 2008. I have a task which has SQL Server as the source and Oracle as the destination. I copy the data from a SQL server view with a field nvarchar(10) to a field of a oracle table varchar(10). The package executes fine
    on my local when i use the data transformation task to convert to DT_STR. But when I deploy the dtsx file on the server and try to run from an SQL Job Agent it gives me the unicode and non-unicode string data types error for the field. I have checked the registry
    settings and its the same in my local and the server. Tried both the data conversion task and Derived Column task but with no luck. Pls suggest me what changes are required in my package to run it from the SQL Agent Job.
    Thanks.

    What is Unicode and non Unicode data formats
    Unicode : 
    A Unicode character takes more bytes to store the data in the database. As we all know, many global industries wants to increase their business worldwide and grow at the same time, they would want to widen their business by providing
    services to the customers worldwide by supporting different languages like Chinese, Japanese, Korean and Arabic. Many websites these days are supporting international languages to do their business and to attract more and more customers and that makes life
    easier for both the parties.
    To store the customer data into the database the database must support a mechanism to store the international characters, storing these characters is not easy, and many database vendors have to revised their strategies and come
    up with new mechanisms to support or to store these international characters in the database. Some of the big vendors like Oracle, Microsoft, IBM and other database vendors started providing the international character support so that the data can be stored
    and retrieved accordingly to avoid any hiccups while doing business with the international customers.
    The difference in storing character data between Unicode and non-Unicode depends on whether non-Unicode data is stored by using double-byte character sets. All non-East Asian languages and the Thai language store non-Unicode characters
    in single bytes. Therefore, storing these languages as Unicode uses two times the space that is used specifying a non-Unicode code page. On the other hand, the non-Unicode code pages of many other Asian languages specify character storage in double-byte character
    sets (DBCS). Therefore, for these languages, there is almost no difference in storage between non-Unicode and Unicode.
    Encoding Formats: 
    Some of the common encoding formats for Unicode are UCS-2, UTF-8, UTF-16, UTF-32 have been made available by database vendors to their customers. For SQL Server 7.0 and higher versions Microsoft uses the encoding format UCS-2 to store the UTF-8 data. Under
    this mechanism, all Unicode characters are stored by using 2 bytes.
    Unicode data can be encoded in many different ways. UCS-2 and UTF-8 are two common ways to store bit patterns that represent Unicode characters. Microsoft Windows NT, SQL Server, Java, COM, and the SQL Server ODBC driver and OLEDB
    provider all internally represent Unicode data as UCS-2.
    The options for using SQL Server 7.0 or SQL Server 2000 as a backend server for an application that sends and receives Unicode data that is encoded as UTF-8 include:
    For example, if your business is using a website supporting ASP pages, then this is what happens:
    If your application uses Active Server Pages (ASP) and you are using Internet Information Server (IIS) 5.0 and Microsoft Windows 2000, you can add "<% Session.Codepage=65001 %>" to your server-side ASP script.
    This instructs IIS to convert all dynamically generated strings (example: Response.Write) from UCS-2 to UTF-8 automatically before sending them to the client.
    If you do not want to enable sessions, you can alternatively use the server-side directive "<%@ CodePage=65001 %>".
    Any UTF-8 data sent from the client to the server via GET or POST is also converted to UCS-2 automatically. The Session.Codepage property is the recommended method to handle UTF-8 data within a web application. This Codepage
    setting is not available on IIS 4.0 and Windows NT 4.0.
    Sorting and other operations :
    The effect of Unicode data on performance is complicated by a variety of factors that include the following:
    1. The difference between Unicode sorting rules and non-Unicode sorting rules 
    2. The difference between sorting double-byte and single-byte characters 
    3. Code page conversion between client and server
    Performing operations like >, <, ORDER BY are resource intensive and will be difficult to get correct results if the codepage conversion between client and server is not available.
    Sorting lots of Unicode data can be slower than non-Unicode data, because the data is stored in double bytes. On the other hand, sorting Asian characters in Unicode is faster than sorting Asian DBCS data in a specific code page,
    because DBCS data is actually a mixture of single-byte and double-byte widths, while Unicode characters are fixed-width.
    Non-Unicode :
    Non Unicode is exactly opposite to Unicode. Using non Unicode it is easy to store languages like ‘English’ but not other Asian languages that need more bits to store correctly otherwise truncation will occur.
    Now, let’s see some of the advantages of not storing the data in Unicode format:
    1. It takes less space to store the data in the database hence we will save lot of hard disk space. 
    2. Moving of database files from one server to other takes less time. 
    3. Backup and restore of the database makes huge impact and it is good for DBA’s that it takes less time
    Non-Unicode vs. Unicode Data Types: Comparison Chart
    The primary difference between unicode and non-Unicode data types is the ability of Unicode to easily handle the storage of foreign language characters which also requires more storage space.
    Non-Unicode
    Unicode
    (char, varchar, text)
    (nchar, nvarchar, ntext)
    Stores data in fixed or variable length
    Same as non-Unicode
    char: data is padded with blanks to fill the field size. For example, if a char(10) field contains 5 characters the system will pad it with 5 blanks
    nchar: same as char
    varchar: stores actual value and does not pad with blanks
    nvarchar: same as varchar
    requires 1 byte of storage
    requires 2 bytes of storage
    char and varchar: can store up to 8000 characters
    nchar and nvarchar: can store up to 4000 characters
    Best suited for US English: "One problem with data types that use 1 byte to encode each character is that the data type can only represent 256 different characters. This forces multiple
    encoding specifications (or code pages) for different alphabets such as European alphabets, which are relatively small. It is also impossible to handle systems such as the Japanese Kanji or Korean Hangul alphabets that have thousands of characters."<sup>1</sup>
    Best suited for systems that need to support at least one foreign language: "The Unicode specification defines a single encoding scheme for most characters widely used in businesses around the world.
    All computers consistently translate the bit patterns in Unicode data into characters using the single Unicode specification. This ensures that the same bit pattern is always converted to the same character on all computers. Data can be freely transferred
    from one database or computer to another without concern that the receiving system will translate the bit patterns into characters incorrectly.
    https://irfansworld.wordpress.com/2011/01/25/what-is-unicode-and-non-unicode-data-formats/
    Thanks Shiven:) If Answer is Helpful, Please Vote

  • BW Extractors for NON-BW Data Warehouse

    Hi,
    I am working on a client who wishes to use a custom developed Data Warehouse. Is there any way to use SAP's standard extractors to extract data to these non-sap dw systems (or download the data in flat file format from extractor after run)? We are looking for mainly LO Cockpit and Finance extractors.
    Best regards,
    Nikhil

    One option I know of is using Informatica you can extract data from SAP R/3 and load a non-BW data warehouse via SAP BCI adapters. BCI adapter makes use of delivered extractors.
    Check out the webex replay of InformaticaWorld conference
    http://www11.informatica.com/replays/IW06_BOsession_Kato.wrf

  • Unicode issue jdbc receiver adapter

    Hi guys,
    I'm facing an unicode issue while trying to put some data into a MS database.
    The sender system is a SAP ERP 6.0 system.
    When we are trying to transfer some text in polish or korean or something else, this textes are not transferred correctly.
    I already checked the SAP Note 831162 FAQ: XI 3.0 / PI 7.0 / PI 7.1 JDBC Adapter
    and used the url parameter sendStringParametersAsUnicode=true
    Furthermore I changed my mapping to get the sql statements, so that I add the attribute hasQuot="no"
    and a leading N in front of the data string as described in the Note and in some threads in SDN
    Unicode problem in JDBC receiver adapter
    JDBC adapter , problem in inserting Korean Charaters in MS SQL
    but nothing solved my problem.
    any ideas?
    Do all the languages have to be installed on the XI system?
    Kind regards
    Jochen

    Hi Sarvesh,
    thanks for helping.
    Unfortunately I´m not sure what I should exactly do.
    If I get you wright I have to test the mapping using the test functionality in message mapping (test tab).
    I did this already
    The result message looks like this
    <?xml version="1.0" encoding="UTF-8"?>
    <ns0:IDOC_CDB_IMPORT_insert xmlns:ns0="http://endress.com/xi/sbo/cdb">
       <ns0:StatementName1>
          <ns0:dbTableName action="insert">
             <ns0:table>CFG.DBO.IDOC_CDB_IMPORT</ns0:table>
             <ns0:access>
                <ns0:IDOC>CHRMAS</ns0:IDOC>
                <ns0:SEGMENT>E1CABTM</ns0:SEGMENT>
                <ns0:ATNAM></ns0:ATNAM>
                <ns0:SPRAS>PL</ns0:SPRAS>
                <ns0:ATBEZ hasQuot="no">N&apos;D?ugo?? ca?kowita Tl; L:&apos;</ns0:ATBEZ>
             </ns0:access>
          </ns0:dbTableName>
       </ns0:StatementName1>
    </ns0:IDOC_CDB_IMPORT_insert>
    How to proceed?
    copy this message to notepad or ultra edit and then doing what?
    please specify clearly.
    this message will be send to the jbdc adapter and this will insert it to database.
    <ns0:ATBEZ hasQuot="no">N&apos;D?ugo?? ca?kowita Tl; L:&apos;</ns0:ATBEZ>
    will end up in a table entry like D?ugo?? ca?kowita Tl; L:
    has anybody an idea how I will get the wright characters in database, when I test my UDF in eclipse, the variable during debugging has the correct text.
    When output this variable via system.ou.print(text); the same issue occurs, the special characters are shown as ?.
    I appreciate for your help, unfortunately my polish colleagues hustle a little bit.
    kind regards
    Jochen

  • Oracle Receiver JDBC Adapter - Handling Unicode Characters

    We have an IDOC to JDBC scenario.
    In this IDOC is sending data like -  10/14u2019/P7 After 4 there is special character coming from SAP ( It is not single quote).
    Mapping is going through OK and data is getting saved in Oracle Database as 10/14&#x19;/P7 with & # x 19;
    I came across following solution in forums and SAP Note.
    I am not sure how to modify Oracle JDBC URL to handle Unicode characters properly.
    Or is there any other approach we can follow to achieve this..
    Any input is really appreciated
    Q: I am inserting Unicode data into a database table or selecting Unicode data from a table. However, the data inserted into or retrieved from the table appears garbled. Why doesn't the JDBC Adapter handle Unicode correctly?
    A: While the JDBC Adapter is Unicode-aware, many JDBC drivers and/or database management systems aren't by default and need a codepage or Unicode-awareness to be configured explicitly. For the respective JDBC drivers, this codepage setting is often configured via the driver URL. For details, refer to the documentation of your JDBC driver or database management system.

    Hi Simona,
    1.To start the visual admin, execute "go" file:
    On Windows: Run \usr\sap\<SAPSID>\JC<xx>\j2ee\admin\go.bat
    On UNIX: Run /usr/sap/<SAPSID>/JC<xx>/j2ee/admin/go
    2.supply the credentials to login into visual admin
    3.under "cluster" tab select "server node"
    4.you will find "log viewer" under "services"
    Since you are new, I recommend you to take help from your BASIS team.
    Hope it helps !
    Hi Alwin,
    Just a quick clarification.
    I used the URL you have mentioned, when we were on SP5. After that we upgraded to SP9.
    From SP9, if you try to use the URL http://XISERVER:50000/AdapterFramework then it automatically redirects to a new webpage with the link to the URL i have mentioned.
    Regards,
    Sridhar

  • Gettin data from SAP R/3 4.7 unicode sys to SAP ERP non-unicode sys via RFC

    We are trying to get data from SAP R/3 4.7 unicode sys to SAP ERP non-unicode sys via RFC. We are reading field from the database via FM and this field got converted by SAP standard Method (CL_ABAP_CONTAINER_UTILITIES=>fill_CONTAINER_C) from structure. When reading this field from ERP sys and converted back to structure using SAP standard Method the value is un-accurate, and different(blank) giving values.
    How, if the data originally was a char or numeric it works fine, but not for decimal.
    best regards,
    Abu-Qaht

    I have a range fields with different type of ranges, am saving these to the database by converting them to a string via the same method. I got the same problem so I tried to convert the string to xstring using function “SCMS_STRING_TO_XSTRING”, before send it to the other system, but the function raised an exception: CX_SY_CONVERSION_CODEPAGE

  • How to check for Unicode/Non-Unicode SAP?

    Hi All,
    Please let me know how can I check whether an SAP system is a Unicode/Non-Unicode from SAP GUI?
    Thanks in advance,
    Adarsh

    Hi,
    You can check under menu :
    System -> Status -> SAP System Data -> Unicode System -> Yes/ No

  • Sending special characters to non-unicode non-sap system: user exit

    Hello All,
    We are sending data from a SAP unicode system to a non-sap non-unicode system via IDOC. The idoc is standard idoc GLMAST which contains gl account information.
    Some text fields contained in this idoc can contain special Polish characters. If they are sent unchanged to the non-unicode system, they give problems in the destination system.
    What we would like to do is to build a user exit to convert these special characters, for example : 'é' becomes 'e', ....
    We used enhancement object ALE00001, function EXIT_SAPLBD11_001 and implemented it, but it seems this exit is not called. Can this user exit be used for this functionality?
    We also tried to change in SM59 the type of system of the destination to non-unicode, so that SAP replaces those special characters by #. But, than the error 02. Codepage not found is given in the idoc. Note: the link to the external system is not set up yet, so no actual connection is possible. Is this why we receive this error, and will this functionality work in the end with a non-sap system?
    Thanks for helping.
    Kind Regards,
    Bart Pelsmaekers

    I faced this problem in many project i implement below logic.
      DATA: c_splchar(2) VALUE '90',
            c_defaultchar(1) type c VALUE '#'.
    You have to move one by one character to this function module
              CALL FUNCTION 'URL_ASCII_CODE_GET'
                   EXPORTING
                        trans_char = spl_char
                   IMPORTING
                        char_code  = spl_code.
    All non unicode(Better you check) are always greater than 90.
             if spl_code is gt c_splchar.
                move c_defaultchar to c_splchar.
             endif.       
    "Reward points if usefull"
    Thanks,
    Narayan

  • How to read a Unicode text file in a non-unicode SAP system

    I'm currently has a text file which is saved as UTF-8 format. And inside this file has Thai characters.
    However, my SAP system is a non-unicode system.
    When I use the code
    open dataset DSN for output in TEXT MODE encoding UTF-8
    the thai characters become some funny characters.
    How can I solve this in the ABAP program without changing the system from non-unicode to unicode system?

    Hi Eswar,
       How can I check whether the code page for THAI already installed or not in SAP system?
       Here is the code I use to test to read a file that saved as UTF-8.
    DATA: l_filepath         TYPE string.
    DATA : BEGIN OF l_filepath_str OCCURS 0,
                 comm(1000)  TYPE c,
               END OF l_filepath_str.
    DATA: l_datasetsucc LIKE rlgrap-filename.
    l_filepath = '/BAAC/Files/APP_.txt'.
    OPEN DATASET l_filepath FOR INPUT IN TEXT MODE ENCODING UTF-8.
    IF sy-subrc EQ 0.
      DO.
        CLEAR l_filepath_str.
        READ DATASET l_filepath INTO l_filepath_str.
        IF sy-subrc EQ 0.
          APPEND l_filepath_str.
        ELSE.
          EXIT.
        ENDIF.
      ENDDO.
    ELSE.
      WRITE:/ 'Error reading from file:', sy-subrc.
    ENDIF.
    CLOSE DATASET l_filepath.
    LOOP AT l_filepath_str.
      WRITE: / l_filepath_str-comm.
    ENDLOOP.
        It returns me a runtime error. Any idea?

  • Data extraction from Non-Unicode ECC6 to Unicode SAP BI system

    Hi,
    We have an existing non-unicode ECC6 system. Currently we are installing SAP BI unicode system. Can anyone tell me are there any issues in data extraction for SAP BI from a non-unicode ECC6 system to an Unicode SAP BI system ?
    Please also note that our data consists of Asian (korean, Japanese & Chinese) fonts.
    Regards,
    Anirban

    Hi Des Gallagher,
    Thank you for your reply.
    I have gone through the notes suggested by you, but they suggest issues related to BW 3.x versions. We are currently on SAP_BW 700 - SP16. Also, among other notes i found note 510882 which might be helpful for custom developments.
    But i am still wondering whether we are going to face major issues related to data extraction from non-unicode ECC6 system to unicode SAP_BW 700 system.
    Incase you have any further details, please let me know.
    Thanks in advance.
    Regards,
    Anirban Kundu

Maybe you are looking for

  • OID Authorization issue

    Hi all, I need some help to setup LDAP Authentication. I configured the AuthProvider, I can see the OID Users and Groups within myrealm->Users and Groups tab. I can also see the groups defined within the ADF Application which is deployed to the manag

  • Reset Identity Column Value in SQL Server depending on whats being inserted

    I have a SQL Server 2008 database with a table that has an ID field that has an identity increment and seed of 1. This work great as each new record created automatically gets a new and distinct number. Now when I insert certain new records I need to

  • MM: Delivery and material document

    Hi Gurus, We have PP-PI industry and we have configured SAP in all areas like MM,SD,PP, WMQM with handling unit. Now for some materials when we receive or issue we get a delivery and for some materials we receive material document. Can you please tel

  • Http Reciver Adapter Header fields

    Hi all, I have a requirement where the receiver is http adapter. I have some fields/structure to be added to the http request. One way i know is to generate entire envelope in the map and then send as request having header and body both. Is there any

  • Is my temperature monitor working correctly?

    i've downloaded the marcel bresink temperature monitor for my macbook. before the firmware update, it ran at 67 degrees celsius. after the firmware update, it runs at 40's. when i restarted my computer few hours later, and loaded up the temperature m