Can we read an oracle non-unicode database to an sap unicode dataset????

Hi
Can we read an oracle non-unicode database to an sap unicode environtment using open dataset.,transfer....using a connection??
Regards

Hello Jacques,
please check sapnote #808505 (Secondary connection to Oracle DB w/ different character set).
Regards
Stefan

Similar Messages

  • HI,Can anybody pls explain me, while extracting database table from sap-r/3

    HI,Can anybody pls explain me, while extracting database table from sap-r/3 to sap-bw using GENERIC DATA SOURCE it will ask us Name of the APPLICATION COMPONENT what does it means?

    Application Component is a collcetion of tightly coupled S/W component. You can think of it is like folder, where all the related S/W will be put in. Like MM will have all the DS related to MM.
    Thanks..
    Shambhu

  • Can SCAN be used for non 11gR2 databases ?

    I am in the processes of database consolidation phase, where in I am consolidating 15 server to a 6 node single Grid for RAC , non -RAC & RAC one node.
    I have non -RAC standalone database on NFS mount , visible on all the 6 nodes , idea is to be able to start the db from any node I like for maintenance which I can do with RAC & RAC ONE NODE.
    I am planning to run SCAN on 3 node , but the question is, can I use SCAN listener for all the databases ?.
    Currently started with a 2 node cluster where I have 11gR2 and couple of RAC one nodes and a 10g rac .
    I am able to point to scan LISTENER and able to connect to 10g db , to me SCAN looks very pomissing for our kind of environment where db's can run on any node with out impacting client changes.
    Before I present this to the management which I am planning to do This Thursday ( big day for me as I am projecting to save quite a bit of money ).
    I don't want to push some thing that's not going to work , even thoug I can prove it now that it works.
    And more importantly I have not see a document in google or ORACLE support that says I can nor a sign that any one using for non 11gR2.
    How is this SCAN compared to OID or SUN OPEN DS ?

    Hi,
    generally, SCAN can be used for 10g databases and you discovered the first half: for 10g databases you will have to modify the REMOTE_LISTENER entry for each 10g database instance to point to the SCAN listeners (as opposed to pointing to the remote local listeners, which is the default in 10g). You could even have the databases registers themselves with SCAN and the remote listeners, if you wanted to... It's more or less a matter of configuration. But for simplification reasons, I will stick to the case where you have your 10g databases register with the SCAN listeners only.
    Now the other half is the client and the client configuration. An 11g Rel. 2 client configured for RAC would have a TNSNAMES entry that has only one address line for the RAC databases. The host entry in this one address line should point to the SCAN (the SCAN name is ideally resolved in DNS). A 10g client configured for RAC would have as many address lines in the TNSNAMES as you have nodes in the cluster.
    The 10g client SCAN configuration would then be in the middle so to speak: You would have 3 address lines in your TNSNAMES, in which each host entry would resolve to one SCAN address (I assume you will use the recommended default of 3 SCAN IPs). If you choose, you can have a name resolution for each of your SCAN IPs, but this would not be required. Now, why would you do it this way? Because this configuration will always work and does not make you dependent on certain functionality that your DNS server may or may not offer.
    For the remaining questions: SCAN is a DNS entry resolving one name to more than one (typically 3) IP addresses. OID is short for Oracle Internet Directory, which is a complete LDAP server. And you are right that there is no document how to configure 10g clients for SCAN from Oracle yet. However, there is a quite good document on SCAN on otn.oracle.com/rac, but I am sure you are aware of it already.
    Hope that helps. Thanks,
    Markus

  • Transaction Isolation Level to Read UnCommited in Non OLTP Database

    HI,
    We are having a database which for NOT OLTP process. That is OLAP DB. Operation on that DB is only Select and (Incremental Insert - FOR DWH ) not Update/Delete and we are performing ROLAP operations in that DB.
    By Default SQL Server DB isolation Level is READ COMMITTED.AS Our DB IS OLAP SQL Server DB we need to change the isolation level toRead Uncommited. We google it down but We can achive in
    Transaction level only by SET isoaltion Level TO Read UNCOMMITED
    or ALLOW_SNAPSHOT_ISOLATION ON or READ_COMMITTED_SNAPSHOT
    Is there any other way if we can Change the Database isolation level to READ uncommitedfor Entire Database?, insteads of achiving in Transaction Level or With out enabling SET ALLOW_SNAPSHOT_ISOLATION ON or READ_COMMITTED_SNAPSHOT
    Please use Marked as Answer if my post solved your problem and use Vote As Helpful if a post was useful.
    Please use Marked as Answer if my post solved your problem and use Vote As Helpful if a post was useful.

    Hi,
    My first question would be why do you want to change Isolation level to read uncommitted, are you aware about the repercussions you will get dirty data, a wrong data.
    Isolation level is basically associated with the connection so is define in connection.
    >> Transaction level only by SET isoaltion Level TO Read UNCOMMITED or ALLOW_SNAPSHOT_ISOLATION ON or READ_COMMITTED_SNAPSHOT
    Be cautious Read UNCOMMITED  and Snapshot isolation level are not same.The former is pessimistic Isolation level and later is Optimistic.Snapshot isolation levels are totally different from read uncommitted as snapshot Isolation level
    uses row versioning.I guess you wont require snapshot isolation level in O:AP DB.
    Please read below blog about setting Isolation level Server wide
    http://blogs.msdn.com/b/ialonso/archive/2012/11/26/how-to-set-the-default-transaction-isolation-level-server-wide.aspx
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers
    My TechNet Wiki Articles

  • Can I join the oracle and mssql database together?

    Hi,
    Did someone know how to join the table from two different platform database? I need to join oracle table with Mssql table in the select statement. Is possible to do so or any hints out there?
    Thanks
    Daron

    Hi,
    Definitly yes, Oracle and MSSQL are two different products and there are unsupported datatypes exists in MSSQL server. Except few instances, you can go ahead and join in one single query provided if you have database link.
    Cheers!!!!
    Mahesh Ragineni

  • I have a problem about language and fonts. in internet explorer if you can't read a font you just rigth click and use unicode but in firefox i do'nt know what shoud i do.? please help me.

    how can i use unicode or somthings like that in mozila firefox.

    and now, when i tried to check again my network preference below the airport tab is showing a message that
    "my airport does not have an IP address and cannot connect to the internet!"
    then after seconds it will goes back to the first message that
    "airport has the self assigned IP address etc....."
    i don't know whats going on now!
    airport tab is not showing green color anyway! i'ts always in yellow!
    i tried turning it on and off but nothing happen! HEEELLLLPP!!!!

  • Errors while performing non-Unicode database export

    Hi,
    I am exporting of a non-unicode database (to perform a unicode conversion). The export has completed without any problems, but I frequently got the following messages in logfile.
    Error 1- UMGCOMCHAR read check skip, no data found; probably old SPUMG
    Error 2- enviornment variable I18N_POOL_WIDTH is not set. Checks are active
    Error 3- I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    My query is that
    1- Are these above 3 errors an issue.
    2- When I import already exported files into a unicode database, will it cause any problem or loss of data.
    3- What is the fix to this issue.
    Points to be awarded for any kind of small help.
    Thanks

    Depending on the data that you have in your non-unicode database, potentially data expansions or data loss may occur.
    Data expansions
    For example, a 1 byte character in a VARCHAR2(1) column may expand to 2 bytes or 3 bytes in a Unicode (UTF8) database; hence you may need to re-define you scheme prior to importing the data into your new Unicode database.
    Data Loss
    This happens only if you have invalid characters inside your non-unicode database. For example, you may have some 8-bit non-ASCII characters inside your US7ASCII database, during export these characters will be converted to some replacement characters (?).
    However you can use the character set scanner (csscan) to scan your source database to detect both of the above scenarios. Please visit the Globalization Support section of OTN for more info - http://technet.oracle.com/tech/globalization/content.html
    Regards
    Nat

  • DMV quesiton. Can not run the DMV query in non system database.

    As the title .I can run following query when I use  'use master' , But don't work at the DB which be created my myself.
    --Error massage
    Msg 208, Level 16, State 1, Line 1
    Invalid object name 'SYS.DM_EXEC_QUERY_STATS'.
    --SQL Version : 2005
    --compatibility level : 90
    --The query
    SELECT TOP 10
    QT.TEXT AS STATEMENT_TEXT,
    QP.QUERY_PLAN,
    QS.TOTAL_WORKER_TIME AS CPU_TIME
    FROM SYS.DM_EXEC_QUERY_STATS QS
    CROSS APPLY SYS.DM_EXEC_SQL_TEXT (QS.SQL_HANDLE) AS QT
    CROSS APPLY SYS.DM_EXEC_QUERY_PLAN (QS.PLAN_HANDLE) AS QP
    ORDER BY TOTAL_WORKER_TIME DESC

    Yes, you can run the query under none system database.... Make sure that  your db is set 90.....
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • Connecting to oracle 12c multitenant database

    My question, using coldfusion 11 enterprise datasource gui,
    how do you connect to oracle 12c multitenant database,
    when you need to log into a pluggable database, beneath
    the container database?
    What's the connection string, or is it even possible to connect(?) because earlier versions of oracle
    didn't use multitenant architecture, you just make a user/password,
    grant connect privileges, and connect.
    Using Oracle Linux 6.5 64bit.
    This is assuming I can connect to the oracle CDB (container database) and PDB (pluggable database) from the shell using sqlplus.
    <!------------------------ I can tsnping the pluggable database name pdb1 ---------------------------------------------->
    [oracle@localhost ~]$ tnsping pdb1
    TNS Ping Utility for Linux: Version 12.1.0.1.0 - Production on 27-JUN-2014 12:13:12
    Copyright (c) 1997, 2013, Oracle.  All rights reserved.
    Used parameter files:
    /u01/app/oracle/product/12.1.0/dbhome_1/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS = (PROTOCOL = TCP)(HOST = localhost.localdomain)(PORT = 1521)) (CONNECT_DATA = (SERVER = DEDICATED) (SERVICE_NAME = pdb1.localhost.localdomain)))
    OK (0 msec)
    <!------------------------------ I can connect to pdb1 using sqlplus shell ---------------------------------------------------->
    [oracle@localhost ~]$ sqlplus user1/password20@pdb1
    SQL*Plus: Release 12.1.0.1.0 Production on Fri Jun 27 12:21:06 2014
    Copyright (c) 1982, 2013, Oracle.  All rights reserved.
    Last Successful login time: Fri Jun 27 2014 09:45:05 -07:00
    Connected to:
    Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Advanced Analytics and Real Application Testing options
    SQL>

    Having the same issue, can't get the connection string right with the PDB?
    Anyone out there? Kurrttt, did you ever get this figured out?

  • Oracle In-Memory Database Cache

    Hi,
    I was reading about Oracle in Memory database cache and i am wondering is this option available on 10g, from what i read it is only on 11g and it is extra option and have to pay for it.
    Any more info, will be great.
    thanks

    From here
    The In-Memory Database Cache option of Oracle Database Enterprise Edition is based on Oracle TimesTen In-Memory Database.TimesTen is also available for 10g.

  • How can I read my iPhone backup files?

    I would like to have a readable version of my text message backup files. I have found the file but I can't read it in its current database format. Please help!

    Why not read them on your phone? Or if the screen's too small for that to be comfortable, email them to yourself and read them on the Mac.

  • Peoplesoft convert Oracle non-unicode database to unicode database

    I am following doc 1437384.1 to convert a Peoplesoft database from non-unicode database to unicode database
    I use the following export statement (as user PS)
    SET NO TRACE;
    SET OUTPUT output_file.dat;
    SET NO DATA;
    EXPORT *;
    And the following import statement (as user sysadm)
    SET NO TRACE;
    SET NO DATA;
    SET INPUT output_file;
    SET LOG log_file;
    SET UNICODE ON;
    SET STATISTICS OFF;
    SET ENABLED_DATATYPE 9.0;
    IMPORT *;
    Before I do the datapump import, I am comparing the objects
    SQL> select object_type, count(*) from dba_objects where owner = 'SYSADM' group by object_type order by 1 asc;
    OBJECT_TYPE COUNT(*)
    INDEX 33797
    LOB 2775
    TABLE 28829
    TRIGGER 9
    VIEW 21208
    on oracsc63 (targetdb):
    SQL> select object_type, count(*) from dba_objects where owner = 'SYSADM' group by object_type order by 1 asc;
    OBJECT_TYPE COUNT(*)
    INDEX 23748
    LOB 2170
    TABLE 19727
    I don't have the same number of object. When I do the import this means that around 10000 tables will not have the UTF-8 format.
    Any ideas how I can solve this? How has experience with this peoplesoft conversions?

    Hello Jacques,
    please check sapnote #808505 (Secondary connection to Oracle DB w/ different character set).
    Regards
    Stefan

  • Can't run non-UNICODE-DbSl against UNICODE database

    Hi everyone. I get this error during the update of Solution Manager 7.1 SPS4 (dual stack) to SPS10. It is a green install. This happens during the
    phase PREP_EXTRACT/PREIMP! Of the ABAP stack. Step 5.2. I do not see a lot of people with this problem, so I assume it is something I am doing wrong, as it is my first time using the tool SUM.
    Details of the system:
    Solution Manager 7.1 SPS4
    Windows 2008 R2 64 bit
    Microsoft SQL 2008 R2 64 bit SP1
    Central installation type with all components on
    the same host
    Kernel level 600
    Here is what I have done so far:
    Downloaded the media as described in the installation
    guide.
    Worked my way through the install using the
    Software Provision Manager 1.0
    Updated the Kernel from level 401 to 600, by
    copying the files from the download folder to the F:\usr\sap\SM1\SYS\exe\uc\NTAMD64
    folder. This is the folder that is assigned to the DIR_CT_RUN parameter.
    I start the STARTUP.BAT file in the SUM folder.
    I then start the SUM GUI and run through the
    steps and in step 3 I map the Stack configuration file (XML) to one listed in
    the folder “51047130\DATA” called “SPS10_stack.xml”
    Then I
    get to 5.2 where the system stops with this error:
    Severe error(s) occurred in phase PREP_EXTRACT/PREIMP!
    Last error code set: Single errors (code &lt;= 8) found in logfile 'PREIMP.ELG'
    ERROR: Detected the following errors:
    # F:\usr\SUM\abap\log\R710VPE.<DB>:
          4 ETW000 TRACE-INFO: 19:  [    dev trc,00000]  Driver: sqlncli10.dll Driver release: 10.50.2769                                3938  0.039733
          4 ETW000 TRACE-INFO: 20:  [    dev trc,00000]  GetDbRelease: 10.50.2769.00                                                      1603  0.041336
          4 ETW000 TRACE-INFO: 21:  [    dev trc,00000]  GetDbRelease: Got DB release numbers (10,50,2769,0)                                21  0.041357
          4 ETW000 TRACE-INFO: 22:  [    dev trc,00000]  Can't run non-UNICODE-DbSl against UNICODE database                              2647  0.044004
          4 ETW000 TRACE-INFO: 23:  [    dev trc,00000]  CheckCodepageType failed. Connect terminated.                                      13  0.044017
          2EETW169 no connect possible: "DBMS = MSSQL                            ---  SERVER = '<SERVER>' DBNAME = '<DB>'"
    Can
    anyone please help?
    PS:
    The admin user that I use do have access to the database. It is listed with
    sysadmin role and the default DB is the Solution Manager db.

    ERROR: Detected the following errors:
    # F:\usr\SUM\abap\log\R710VPE.<DB>:
          4 ETW000 TRACE-INFO: 19:  [    dev trc,00000]  Driver: sqlncli10.dll Driver release: 10.50.2769                                3938  0.039733
          4 ETW000 TRACE-INFO: 20:  [    dev trc,00000]  GetDbRelease: 10.50.2769.00                                                      1603  0.041336
          4 ETW000 TRACE-INFO: 21:  [    dev trc,00000]  GetDbRelease: Got DB release numbers (10,50,2769,0)                                21  0.041357
          4 ETW000 TRACE-INFO: 22:  [    dev trc,00000]  Can't run non-UNICODE-DbSl against UNICODE database                              2647  0.044004
          4 ETW000 TRACE-INFO: 23:  [    dev trc,00000]  CheckCodepageType failed.
    As per error, you have used wrong kernel. Please make sure that you use unicode kernel as your system is also unicode.
    Thanks,
    Sunny

  • Connect PostgreSQL to Oracle 10g through ODBC: can't read all_tables

    Oracle 10g is not able to read the database table names (all_tables and all_objects, es. select * from all_tables@hspostgresodbc) from PostgreSQL using ODBC (v. 08.01.0200). This functionality is essential to be able to use Oracle Warehouse Builder and import external database definitions (heterogeneous services).
    However, Oracle (SQL*Plus) is able to retrieve records when the table name is stated in the SQL query (es. select * from “my_table”).
    The problem seems to be specifically with the PostgreSQL ODBC because Oracle is able to read schemes from other databases through ODBC. I’ve done some tests with Access dbs just to be sure that the problem is not with my system (Windows XP SP2).
    The problem appears both with ANSI and UNICODE PostgreSQL drivers.
    Any suggestion of how workaround this problem?
    thanks in advance,
    Raoni

    Dear all,
    I’ve tried different tracing levels with out success. The log file does not seem to be created by Oracle.
    Looking closer to the debug and comm logs from the driver 8.01.0200, I’ve found out that the problem was generated after a table wasn’t found (the driver seems to search for the all_tables table, and just afterwards to query the scheme pg_catalog). In this case the all the following queries returned with the error “ERROR: current transaction is aborted, commands ignored until end of transaction block”. Maybe this problem is related with the bug [#1000641] http://pgfoundry.org/tracker/index.php?func=detail&aid=1000641&group_id=1000125&atid=538
    Following the Postgres odbc developers suggestion, I’ve installed the psqlODBC v 08.02.0002 and I’ve got a completely different result.
    As before, the simple query “select * from my_table@hsodbc” returns the result using SQL*Plus clearly faster than before.
    Doing the query “select * from all_tables@hsodbc” the result is OK (the name of 56 tables) but the process is very slow and burdensome. The file paging memory in the machine running Oracle and the SQL*Plus jumps from the normal 700 MB to 1.7 GB, and the query takes about 20 minutes to complete.
    Also the 1.7 GB remains allocated and the query process halted in PostGreSQL even after the on screen result of the query. The memory is freed and the process terminated after the logging out SQL*Plus.
    Probably as result of this memory explosion, the system runs out of memory and Oracle Warehouse Builder selecting data base link -> schema, after some minutes get an error similar to: “can’t read from memory address XXXXX”
    All those test where done with PostGreSQL 8.0.1 (on Linux Suse 9.0) and Oracle 10g (Windows XP SP2).
    To be sure that the problem was not related to the PostGreSQL version, we updated the server to the 8.1.3 (running in Linux Suse 9.0 with psqlODBC v 08.02.0002 ) and OWB couldn’t see any table as before and doing Select * from all_tables@hsodbc we’ve got the message:
    ERROR:
    ORA-00604: error occurred at recursive SQL level 1
    ORA-28500: connection from ORACLE to a non-Oracle system returned this message:
    [Generic Connectivity Using ODBC]
    ORA-02063: preceding 2 lines from HSPG
    no rows selected
    On the other hand, with the same configuration as before (Oracle 10g under Windows XP SP1, psqlODBC v 08.02.0002) but with PostGreSQL 8.1.3 under Windows XP SP1, OWB can read table names (same tables imported from Linux PostgreSQL) but not all_tables (a minor problem, because what is important for us is to make OWB work with PostgreSQL).
    The encoding for all DBs is UTF8.
    Any ideas of where could be the problem?
    regards,
    R.

  • Oracle Database Migration Assistant for Unicode (DMU) is now available!

    Oracle Database Migration Assistant for Unicode (DMU) is a next-generation GUI migration tool to help you migrate your databases to the Unicode character set. It is free for customers with database support contracts. The DMU is built on the same GUI platform as SQL Developer and JDeveloper. It uses dedicated RDBMS functionality to scan and convert a database to AL32UTF8 (or to the deprecated UTF8, if needed for some reasons). For existing AL32UTF8 and UTF8 databases, it provides a validation mode to check if data is really encoded in UTF-8. Learn more about the tool on its OTN pages.
    There is a new Database Migration Assistant for Unicode. We encourage you to post all questions related to the tool and to the database character set migration process in general to that forum.
    Thanks,
    The DMU Development Team

    HI there!
    7.6.03 ? Why do you use outdated software for your migration.
    At least use 7.6.06 or 7.7.07 !
    About the performance topic - well, you've to figure out what the database is waiting for.
    Activate time measurement, activate the DBanalyzer with a short snapshot interval (say 120 or 60 seconds) and check what warnings you get.
    Also you should use the parameter check to make sure that you don't run into any setup-induced bottlenecks.
    Apart from these very basic prerequisites for the analysis of this issue, you may want to check
    SAP Note 1464560 FAQ: R3load on MaxDB
    Maybe you can use some of the performance features available in the current R3load versions.
    regards,
    Lars
    p.s.
    open a support message if you're not able to do the performance analysis yourself.

Maybe you are looking for