Migration of data using ETL tools.

Hi,
We want to tranfer data from Sap to Oracle.Could you please suggest which ETL tool we should choose
for fulfilling our requirement.
Would somebody will be able to access it from outside SAP? Is it possible to extract data at certain frequency e.g. every week, every month using ETL tools.
Regards,
Divya.

Hi,
A feature in BI called Open Hub Service is used for transporting SAP data to non-SAP datamarts, etc. This data can be transferred to database tables or flat files. To know the detials, please refer:
http://help.sap.com/saphelp_nw70/helpdata/en/bf/50453c01f4f75fe10000000a11402f/frameset.htm
Rgds,
Sonal

Similar Messages

  • Scope for data warehousing ETL Tool

    Hi all
    can anybody explain scope for data warehousing ETL Tool
    for oracle developer in future this is ok or..
    regards
    Message was edited by:
    174313

    What exactly is your question?
    The scope of using an ETL tool would be setting up and maintaining datawarehouses and building ETL processes to populate these datawarehouses.
    A tool is generally preferred over hand coding because tools allow better maintenance, shorter development cycles etc.
    Oracle has a pretty good ETL tool, called Oracle Warehouse Builder. It's not the best tool available, but if you compare price/ functionality I would say in most cases it will do.
    If your question is if it's a wise pick to master OWB or any ETL tool my answer would be a clear YES! Datawarehouses and BI are becoming more important every day, their use gets broader everyday.
    If your question is if an investment in OWB is a wise investment for the future I would answer a clear YES again. It's increddilbe to see what progress Oracle made with the tool, coming from a 'laughing stock' postition, regarded completely immature good for nothing tool, to where they are now, with 10.2. Regarded as one of the leaders by Gartner.
    Oracle has recognized (a long while ago) that ETL is the bread and butter of the future and invests in a good quality tool for accomplishing this.
    I hope this answes your question, if not, please try to specify more clearly.
    Regards,
    Toin.

  • Issue Migrating Character Data Using a Full Export and Import

    Hi There;
    I have a database in my local machine that doesn't support Turkish characters. My NLS_CHARACTERSET is WE8ISO8859P1, It must be changed to WE8ISO8859P9 , since it supports full Turkish characters. I would like to migrate character data using a full export and import and my strategy is as follows:
    1- create a full export to a location in network,
    2- create a new database in local machine that it's NLS_CHARACTERSET is WE8ISO8859P9 (I would like to change NLS_LANGUAGE and NLS_TERRITORY by the way)
    3- and implement full import to newly created database.
    I 've implemented first step, but I couldn't implement the second step. I 've created the second step by using toad editor by clicking Create -> New Database but I can not connect the new database. I must connect new database in order to perform full import. How can I do this?
    Thanks in advance.
    Technical Details
    NLS_LANGUAGE.....................AMERICAN
    NLS_TERRITORY.....................AMERICA
    NLS_CURRENCY.....................$
    NLS_ISO_CURRENCY..............AMERICA
    NLS_NUMERIC_CHARACTERS    .,
    NLS_CHARACTERSET.............WE8ISO8859P1
    NLS_CALENDAR.....................GREGORIAN
    NLS_DATE_FORMAT................DD-MON-RR
    NLS_DATE_LANGUAGE...........AMERICAN
    NLS_SORT...............................BINARY
    NLS_TIME_FORMAT.................HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT......DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT............HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT..DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY............ $
    NLS_COMP...............................BINARY
    NLS_LENGTH_SEMANTICS.......BYTE
    NLS_NCHAR_CONV_EXCP........FALSE
    NLS_NCHAR_CHARACTERSET...AL16UTF16
    NLS_RDBMS_VERSION............10.2.0.1.0

    First, if your applications run on Windows, do not use WE8ISO8859P9. Use TR8MSWIN1254.
    Second, if you create a new database, the database is not necessarily immediately accessible to outer world. I do not know Toad and I have no idea if it performs all necessary steps required for the new database to be visible.  For example, in the Toad itself, I assume you should create a new connection that references the new SID of the newly created database and use this new connection to connect. However, connections without a connection string use the ORACLE_SID setting in Registry to tell connecting applications which instance (database) to use.  To change the database accessed with an empty connection string you need to modify Registry (unless Toad has an option to do this for you). If you want to connect without changing Registry, you need a connect string. This requires setting up Oracle Listener to serve the new database (unless default configuration is used and the database registers itself with the default listener). It also requires changing tnsnames.ora file to create an alias for the new database. Net Manager and/or Net Configuration Assistant can help you with this.
    I wonder if Database Configuration Assistant would not be a better tool to create new Oracle databases.
    Thanks,
    Sergiusz

  • Am unable to unload correct japanese data using ODISqlUnloader tool

    Hi,
    I am unable to unload correct japanese data using ODISqlUnloader tool to my
    local system.
    Database LNS_LANG setting is AL32UTF8 and I can easily see those data in my
    SQLPLUS window but when I unloading those getting error or JUNK characters
    for japanese scripts. which charset should I use for unloading japanese
    characterset properly
    regards,
    Palash

    Hi all,
    Our problem with SqlUnloader where it is unable to unload japanese
    data, So we built a new open tool for unloading data .
    but in our case open tool ojdbc14.jar that we had in driver folder of ODI
    installation jdbc Implementation- Version: "Oracle JDBC Driver version -
    10.1.0.5.0".
    While we changed the same with Specification-Version: "Oracle JDBC Driver
    version - 9.0.2.0.0" it's working well.
    but for this ojdbc14.jar SQLUNLOADER showing java.sql.SQLException: Invalid
    character encountered in: failAL32UTF8Conv as my data base characterset is
    AL32 UTF8.
    Can you people confirm that is this a bug that we can't unload japanese data
    using OdiSqlUnloader or I am doing somthing wrong.
    Note : My opentool is working fine with ojdbc14.jar of version - 9.0.2.0.0 but
    fails for Driver version - 10.1.0.5.0 . to excute the the same I changed the
    ojdbc14.jar present in driver folder of odi installation directory . So will this affect
    any other part of odi excution . My ODI version is 10.1.3.2.0.

  • Getting error while fetching Log data using B1TE Tools

    Hi,
    i am using B1TE Tools to test an add on which generating an error while fetching the Log File using DI Logs Reader. The error is below:
    "Wrong Xml file format. "The data at the root level is invalid, Line 1, position 1.".
    please report it on B1TE blog"
    Please help..
    Regards,
    Sunita

    > Do you want to add and/or update the data in the already existing tables or do you want to replace the content completely?
    >
    > so in that way :
    > bot the options are fine what ever take less time.
    Sorry mate, but YOU have to know what you want here.
    I gave you an easy to follow set of steps.
    As you don't seam to mind the outcome, just might just use them...
    > I wanted to know weathe i can use the  loadercli for thie export import or not? if yes then is there any new steps to do before i do the export import?
    We had this discussion before...
    >
    > For that the easiest option would be just to drop the tables of SAPR3 and run the import again.
    >
    > For ease of use you could also just do:
    > - logon as superdba
    > - drop user SAPR3
    > - create user SAPR3 password SOMEPW not exclusive dba
    >
    > After these steps you can easily pump the data into the database again.
    >
    > So here in th above given steps , i am creating a new SAPR3 user and why it is not exclusive dba ?
    >  i already have that user SAPR3 can i use the same.
    Yes, you do have the SAPR3 user.
    But you don't seem to like to read documentation or learn about how the tools work or anything like that.
    Therefore I gave you s simple way to reach your goal.
    Of course it's possible to reuse the user.
    But then you would have to deal with already existing tables, already existing data etc.
    You don't seem to be able to do that. So, the easy steps might be better suited for your needs.
    regards,
    Lars

  • Upload file in to km folder using ETL tools

    Hi Experts,
    I have requirement, to upload html file in to km repository using some ETL tool. Is it possible? If not...
    How i can bring a file from a different server in to portal km repository?
    Regards
    Baby

    Hi all,
    I am closing this question... since i got a better way to proceed...
    In our portal we have a file system mountpoint (etc). We are able to FTP html files in to this folder through ETL scheduling... 
    I think this is better way, also have discussion with some of my friends... any comments..
    Regards
    Baby
    Edited by: Baby on May 25, 2009 4:11 PM

  • Data Warehouse and ETL tools for data verification ?

    Data Warehouse and ETL tools for data verification ?
    How need to to data verification using ETL tool ? Also how to relate this thing to datawaehouse ?
    Thanks in Advance

    Hi  Shyamal Kumar,
    1)  BW it self  facilitates to do the ETL (Extraction Transformation Loading)  steps:
         example:
                     Extraction  - from SAP or other data bases
                     Transformation - using transfer rules, Updates rules
                     Loading  -  Loading into ODS, Cube, master data
    2) Typically used ETL tools in the industry are:
         a)   datastage from Ascential (owned by IBM)
         b)   Informatica
         c)   Mercator
    Regards, BB

  • I can't see my external disk source when trying to migrate data using the Migration Assistance.

    I have an external hard disk (Lacie) with all my backed up data. After doing a clean install of Mavericks I now want to migrate my data using the Migration Asistance but it is not able to find the source.
    Can anyone help please?
    Thanks.

    Is the layer's visibility check box turned on or off?  What is the layer's Blend Mode?  What is the format of the original file?

  • Difference between uploading data using BDC,CATT

    Dear experts,
    what is the difference between uploading the data using BDC and CATT tool?
    is it only the coding or something else.
    and can you explain the situations where we prefer to upload the data using CATT tool.
    thanks in advance.

    Check this thread -
    Differences between  BDC , LSMW, BAPI,CATT
    Regards,
    Amit

  • Low-Cost data migration and ETL tools

    Hi, Im building a database for my company. We are a rather small size book company with a lot of references and still growing.
    We have a Mysql database here and are trying to find some good tools to use it at its best. Basically we are just starting up the database after dealing with Excel: we had a size problemu2026 So im trying to find a program that will allow us to do two different things: the migration of our data from the old system to the new one and a specialized software to perform ETL (Extract, transform and load) on our database.
    About the price of the tools, if we were one year ago, the accounting department would have been pretty relaxed about this. But today, we have some budget restrictions and therefore need a low cost tool. So could you give me some advice on a good data migration and etl tool for a low cost?
    Thanks for your help.

    Hi,
    Some companies have budget problems and are having to reduce their cost of operation.
    Have you ever heard of open source tools? They can do the job easily and less expensive than proprietary solutions that cost a lot of money.
    You can have a look at a good ETL open source program called Talend Open Studio: it is user-friendly but also has advanced features intended for technical users (java debugger, code injectionu2026). It can perform data migration and ETL as you wrote in your first post.
    The website is [here|http://www.talend.com/solutions-data-integration/data-migration.php] to download the open source program. They have a forum and documentation you can read. Tell us what you think about the software.
    For an ETL benchmark: [ETL Benchmark|http://blogs.sun.com/aja/entry/talend_s_new_data_processing]

  • Can We use FDM as ETL tool between SQL and Oracle

    I want to use FDM as ETL tool between SQL and Oracle. Can it be possible. I didn,t found any target adapter for oracle database.My source system is SQL and Target system is Oracle database.
    Rahul
    Edited by: user12190125 on Nov 9, 2009 4:23 AM

    Rahul,
    I believe this is possible to do, but not an easy one and there are a few considerations:
    How much data are you processing? FDM has a lot of features which support the business process. While this is great for users and audit trail etc. it slows down performance if you want to process a lot of data. It also depends on the type of mappings you use (Like mappings are slower than explicit mappings).
    How familiar are you with VBScript? There is no explicit target adapter for Oracle, but there is a data mart adapter which can be used for anything. You have to implement everything yourself though, mainly the Export and Load actions. In there you will also have to handle the the connections to the MSSQL and Oracle databases.
    Check the data mart adapter and see if you feel comfortable with defining the vb code in there. There are reasons for and against this approach. ODI would probably be the better choice unless you really need to have FDM's process support.
    Regards,
    Matt

  • Is it possible to use EBS as a ETL tool ?

    Is it possible to use EBS as a ETL tool ?

    Hi,
    Are you referring to "Extract, Transform, Load"? It is possible to download/upload the data from/to Oracle Apps and there are many APIs and tools (provided by Oracle or 3rd party) which should be helpful -- Please see these links for details:
    Oracle Integration Repository
    http://irep.oracle.com
    3rd Party/ISV Generic Installation Advice
    3rd Party/ISV Generic Installation Advice
    Note: 951604.1 - Implementing Oracle EBS Service on a Separate Instance
    However, you still need to use such tools with caution (make sure it is supported) as you do not want to end up with referential integrity issues and data corruption problems.
    Regards,
    Hussein

  • Loading a Z* custom transparent  table using BW ETL tools

    Hi,
    In BW is it possible to load a Z* custom transparent  table(instead of DSO)  directly  using BW ETL tools?.  Thanks.
    Thanks.
    Raj.

    With BI 7.0 a typical tale is DSO[ write protected]...which should solve all your requirements.
    You can have  a start/end routine to add records to a z-table while processing data stagging [ datasrc>psa>dtp>dso ]..and delete data in target BI-datatarget[ dso]...and still keep data in ur ztable.
    Edited by: Gopi Eriki on Aug 13, 2009 5:37 PM

  • Not able to extract performance data from .ETL file using xperf commands. getting error "Events were lost in this trace. Data may be unreliable ..."

    Not able to extract  performance data from .ETL file using xperf commands.
    Xperf Commands:
    xperf –i C:\TempFolder\Test.etl -o C:\TempFolder\BootData.csv  –a process
    Getting following error after executing above command:
    "33288636 Events were lost
    in this trace. 
    Data may be unreliable
    This is usually caused
    by insufficient disk bandwidth for ETW lo
    gging.
    Please try increasing the minimum
    and maximum number of buffers
    and/or
                    the buffer size. 
    Doubling these values would be a good first at
    tempt.
    Please note, though, that
    this action increases the amount of me
    mory
                    reserved
    for ETW buffers, increasing memory pressure on your sce
    nario.
    See "xperf -help start"
    for the associated command line options."
    I changed page size file but its does not work for me.
    Any one have idea, how to solve this problem and extract ETL file data.

    I want to mention one point here. I have total 4 machines out of these 3 machines above
    commands working properly. Only one machine has this problem.<o:p></o:p>
    Hi,
    I consider that you can try to use xperf to collect the trace etl file and see if it can be extracted on this computer:
    Refer to following articles:
    start
    http://msdn.microsoft.com/en-us/library/windows/hardware/hh162977.aspx
    Using Xperf to take a Trace (updated)
    http://blogs.msdn.com/b/pigscanfly/archive/2008/02/16/using-xperf-to-take-a-trace.aspx
    Kate Li
    TechNet Community Support

  • Working on ETL tools interoperability using Common Warehouse Model (CWM)

    Hi All,
    Its just a piece of information and not a question.
    I have been working on proving the ETL tools interoperability using Common Warehouse Metamodel(CWM), an OMG standard. The whole concept is to take out the metadata from an ETL tool say OWB and put it into CWM Metadata Repository and this metadata can be used for building the same project in any other tool say Informatica or may be in the same ETL tool.
    The main thing in this process is to map each ETL tool with the CWM concepts and then using Model to Model Tranformations(Technologies like Xtend) one can set up a communication between different ETL tools.
    Till now I have worked with OWB only. I, with my team, have extracted all information from an OWB project (which has medium level of complexity like two oracle modules(schemas) and few tables, views and mappings with various operators) and put it in CWM repository and extracted it back from CWM MDR to OWB itself. We haven't worked with any other ETL tool because of the unavailablity of any other ETL tool with us. We will be working with Pentaho kettle in near future and try to prove the whole process as two way communication.
    The whole process can be described in steps as below :
    1. Creation of a manual OWB Ecore model(model representaion in Eclipse Modelling Framework) which gives all dependencies and reletionships in OWB objects like Project,OracleModule etc.
    2. Creation of CWM ecore model from Rational Rose mdl which has been provided by OMG on their site.
    3. Generation of Java code(Gen Model) from the above mentioned ecore model(It is needed to create an object from OWB).
    4. Extraction of project from OWB using public views which has been exposed by OWB itself. You can refer to [http://download.oracle.com/docs/cd/B31080_01/doc/owb.102/b28225/toc.htm|http://download.oracle.com/docs/cd/B31080_01/doc/owb.102/b28225/toc.htm] link for OWB public views and other APIs.
    5. (Actually Step 4 is a part of this step only )Writing a Java code which has JDBC connection for accessing OWB public views and Ecore model as imported java files(Step 3 has been done for this part only). This java code will return an OWB project object(instance of the Ecore model) which will be used in further steps.
    6. Writing an Xtend code to do a model to model tranformation from OWB to CWM.
    7. Writing an Open Architecture Workflow to combine all the steps in one step which will take the output of java code(step 5) and put it into Xtend Code(step 6) and then will take the output of Xtend code and give it to the XMIWriter(an OAW component) to write an XMI which is actually a CWM Ecore Model instance.
    8. Saving above XMI(CWM model instance) to the CWM MDR using Hibernate and Teneo.
    In the same way we can extract metadata from CWM MDR and put it into OWB. But the only problem with OWB is that we cannot persist OWB object in OWB repositories as OWB tables are very cryptic and tough to understand. So for that we have used TCL scripts(OMB Plus scripts) to create a project in OWB using OWB ecore instance. You can refer to the above oracle documentation link for TCL scripts.
    Let me know if I can assist you if you are working on the same.
    You can mail me for any queries. My email id is [email protected].
    Thanks,
    Deepak

    Hi
    1. Why do we need to install another standalone HTTP server in a separate home? Where do we use that server?
    DA: The separate HTTP server is for the Workflow Monitor, which is not necessary (it has some use cases mind you).
    2. To make the OWB work correctly while using ETL features, do we always need to run Workflow Configuration Assistant, because I wasn't able to generate code from OWB editor after building a mapping while Workflow Configuration Assistant wasn't running.
    DA: Not necessary, what error did you get? Mappings can be designed, deployed and executed without Workflow. Workflow can be used for orchestrating the the mappings (ie .running a bunch of them in a specific order with others tasks).
    3. Whenever I try to save my work in OWB, I get an error , message : Preference.properties (Access is denied). Though it saves my work but I don't understand why I am getting this error. It looks like OWB is trying to access some Property from the Preferences (Tools menu) but can't access.
    DA: It sounds like the directory where you have installed OWB does not have permissions for the OS user you are executing it. Is the install user different from the execution user? Either run using the installed user, or change the permissions of the directories (grant the executing user write permissions under all directories under owb).
    4. I also get error while closing the Mapping Editor :-
    DA. same issue as 3.
    Cheers
    David

Maybe you are looking for