Best way to import data to multiple tables in oracle d.b from sql server

HI All am newbie to Oracle,
What is the Best way to import data to multiple tables in Oracle Data base from sql server?
1)linked server?
2)ssis ?
If possible share me the query to done this task using Linked server?
Regards,
KoteRavindra.

check:
http://www.mssqltips.com/sqlservertip/2011/export-sql-server-data-to-oracle-using-ssis/
      koteravindra     
Handle:      koteravindra 
Status Level:      Newbie
Registered:      Jan 9, 2013
Total Posts:      4
Total Questions:      3 (3 unresolved)
why so many unresolved questions? Remember to close your threads marking them as answered.

Similar Messages

  • Implementing "Temporary Tables" in Oracle after migrating from SQL Server

    Hello,
    Some stored procedures referencing temporary tables in SQL Server were migrated to Oracle. In the Oracle DB, these stored procedures are invalid. In a bid to make them valid, I created the referenced temporary tables in my Oracle DB, via SQL*PLUS having logged into the database as the appropriate user. And when I query the user_tables view, I see the temporary tables that I created. Thereafter, when I compile the stored procedures, I still get the message: PL/SQL: ORA-00942: table or view does not exist. And the stored procedures remain invalid. Pls,what do I do ?
    Thanks for your cooperation

    Chux,
    The Migration Workbench should have created all the necessary temporary tables as part of the process to convert the stored procedures. Can you post the example part of the T/SQL stored procedure and what we converted it to?
    Our support for Temporary Tables is discussed in our reference guide.
    http://download-west.oracle.com/docs/html/B10254_01/ch5.htm#1026115
    Donal

  • Best way to compare data of 2 tables present on 2 different servers

    Hi,
    We are doing data migration and I wil like to compare data between 2 tables which are present on 2 different server. I know to find the difference i can go for minus or full outer join and by creating the database link.
    But my problem is the volume of the data. The tables under consideration has approximately 40-60 columns and number of rows in each tables are around 60-70 million. Also both the tables are on 2 diffferent servers.
    I would like to know
    1] What will be the best way to compare the data and print the difference from performance perepective? I know that if I am going for DB links then its will definitely impact the performance as the tables are across 2 different servers.
    2] Is it advisable to go for using SQL - PL/SQL for this kind of senario or dump the data in flat files and use C or C++ code to find the difference.
    Regards,
    Amol

    Check this at asktom.oracle.com. Search for "Marco Stefanetti" and follow the few posts between Marco and Tom. As far as your tables being on separate servers, you could consider dumping the data to file and using external table or using CTAS ( create table as select ) to get both tables locally.
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:2151582681236

  • Best way to import data only from sql server 2k to ora9.2

    I have a situation where I don't care about indexes, procedures. The DDLs from sql server 2k and ora9.2
    are (mostly) teh same, I need to port the data from sqlserver to my ora database.
    what is the best way to do it?
    thanks.

    Hello Zack,
    The Oracle Migration Workbench can help you to migrate your data from your Oracle 9.2 database to SQL Server 2000. Please refer to Chapter 6, 'Migrating the Database' in the Oracle Migration Workbench User's Guide for more information about how to migrate your data using the Oracle Migration Workbench.
    You can find this document at the following location on the Oracle Technology Network (OTN) Web site:
    http://download.oracle.com/docs/html/B15857_01/migrate_db.htm#sthref282
    I hope this helps.
    Regards,
    Jocelyn (Oracle Migration Workbench Team)

  • Best way of partitioning the huge size table in oracle 11gr2

    hi,
    OS: linux
    DB: oracle 11gR2
    table size is about 2T and single table space with multiple dbf files in ASM,
    table partition with dbms.redefinition is running past 2 days (range partition).. its running but not that fast..
    note: exchange partition is too slow.. hence not used..
    what is the best way of doing the huge size table partition. please suggest, thanks.

    >
    what is the best way of doing the huge size table partition
    >
    A few questions
    1. Is that an OLTP or OLAP table?
    2. Is all of the data still needed online or is some of it archivable?
    3. What type of partitiioning are you doing? RANGE? LIST? COMPOSITE?
    4. Why do you you say 'exchange partition is too slow' - did you do a test?
    For example if data will be partitioned by create date then one stragety is to initially put all existing data into the root partition of a range partitioned table. Then you can start using new partitions for new incoming data and take your time to partition the current data.

  • Best way to Import Data from a SQL Server Table?

    Hi,
    Firstly thanks for looking at this question.
    We need to import data into SAP BPC 5.1 on a twice daily basis and have chosen not to export to a .CSV file but instead to hold all data in a SQL table and import it directly from there.  As part of the import we wish to run the default logic, however only over the data which is imported as opposed to having to running default logic over the entire database after every import.
    We did some research on this topic and the only thing we could find that would work as described above is using the "Import SQL" package.  However we keep experiencing problems with it and have not yet been able to run it successfully; the errors it gives are not consistent from run to run which makes it difficult to start a thread, though we are getting help from the helpdesk at the moment.
    My question here though is - is there another way that someone knows of to import data into SAP BPC from a SQL table, and being able to run default logic over just the data being imported, or is our only hope getting the "Import SQL" package working?
    Any help much appreciated.
    Regards,
    Iain
    Forgot to mention details of our environment:
    SAP BPC v5.0.495, 2 server environment -
    Server 1 (DB/AS/SSIS/File server) = 64bit Windows 2k3 server with 64bit SQL Server Enterprise Edition
    Server 2 (IIS/App server) = 32bit Windows 2k3 server
    Edited by: Iain Hambleton on Jun 17, 2008 3:25 PM

    Lain,
    I recently created SSIS packages that need to work with a staging table that does all kind of manipulations of that data before it is loading into SAP BPC, because it also has to load the data in the drillthrough table within the same package. From one point in my package the data is also in a SQL table so basically the same as in your situation. This also works with an export to CSV within the DTSX file like Alwin said, because in this case you can use the standard transformation and conversion stuff during the load.
    I also needed to limit the data region for logic to the data region that is in the load for currency conversion purposes. This is not very much of a problem. I had a situation where I have an Accounts receivable cube containing a daily time dimension for keyduedates and a datasrc dimension containing weeks. Every week has a complete overview of the open AR items in that week and need to be converted for that week (datasrc) only and not for the whole database every time we load. But by default data would be converted based on the Keyduedates dimension while I wanted the Week dimension to be used as the data region. I solved it by using these rows in the logic:
    *Scope_by=version,Weeks
    *xdim_memberset weeks=%weeks_set%
    *xdim_memberset version = %version_set%
    I can send you the SSIS package and logic if you want. Just send me your details then.
    -Joost
    Edited by: Joost Hoppenbrouwers on Jun 17, 2008 4:25 PM

  • Best way to import data using a Java exe

    We're trying to pull a bunch of tables from an Informix SE 7.25 database, truncate some tables on an Oracle schema, and import the Informix tables into the Oracle system. We don't need to modify the data -- just pull from one, truncate, and import into the other system. What's the most efficient (and simple!) way to accomplish this via a Java executable? Thank you.

    You can do anything from using JDBC bulk inserts or generating your own insert statements and executing an Oracle native application to import them, ala SQL*Plus. Whether it is an executable or not is really simply a deployment decision. Take a look at the various JDBC batch/bulk operations. Also take a lot at what formats Oracle's imp tool supports and see if generating it that way is simpler. I'm not even sure you need Java to accomplish this task. A native shell script can probably do it too.
    - Saish

  • Best way to show data of multiple lists - List Dashboard

    Hi,
    I am using SP 2013. I have around 10 custom lists, in which users add items. When a item is added a custom Visual studio approval workflow is triggerred.
    I am trying to create a visual webpart (VS 2012) , which will show the items that the logged-in user has created under each of the list. I want to show the data in a list-by-list manner (kind of a dashboard). Each of the list has different kind of columns.
    Which would be ideal way to achieve this? Should it be a dynamic table or a dynamic SPGridView or is there any other way to achieve this?
    There would be multiple entries the user would have added under each list. So a paging would help.
    Thanks

    Hi,
    According to your post, my understanding is that you wanted to display all the items that the logged-in users had created under each of the list in the site.
    I don’t think it’s a good idea to display all the items that the logged-in users had created under each of the list in the site in a visual web part.
    As we all known, different list has different fields, it will be different to display different list fields in a visual web part.
    If you still want to use the visual web part, I recommend you use the common fields that all the lists contained, suach as the Title, Createdy by, etc.
    Then query all the lists in the site by the created by field to compare with the current user.
    As a workaround, I recommend you use a page to display all the items that the logged-in users had created in the site.
    You can create a custom view based on current user for the lists, then add the lists web part in a page with the custom view.
    Now the page would only show the current users items.
    http://go.limeleap.com/community/bid/297846/Custom-List-View-Based-on-Current-User-Using-SharePoint-Designer
    http://social.technet.microsoft.com/Forums/sharepoint/en-US/5e347dea-475e-4b95-8905-0f6e11bab7bf/sharepoint-list-filtering-by-current-user-or-group?forum=sharepointgeneralprevious
    What’s more, you can also use the target audience to achieve it.
    http://lixuan0125.wordpress.com/2012/06/18/audience-targeting-sharepoint-2010/
    Thanks,
    Jason
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jason Guo
    TechNet Community Support

  • Import Data into multiple tables

    Hi folks,
    I already did some research in this forum but I cannot find a solution. As far as I understood I need to import the data from the flatfile into a staging table and then distribute it to the different tables by running a SQL script/statement.
    Do you have any examples of such a SQL statement/script?
    Thanks in advance!
    Regards,
    Tino
    Message was edited by:
    tino.albrecht

    repeat for each table:
    insert /*+ APPEND */ into <table 1>
    select <source columns for table 1>
    from <table where the flat file was imported>
    where <conditions that identify records for table 1>;
    OR, alternatively, with a single statement
    insert
    when <conditions that identify records for table1>
    then into <table 1>(<columns of table 1>) values(<source columns for table 1>)
    when <conditions that identify records for table2>
    then into <table 2>(<columns of table 2>) values(<source columns for table 2>)
    select * from <table where the flat file was imported>;

  • Best way to import data?

    Hi everyone,
    I haven't worked with In Design before so I have no idea where to start. I have a client who outputs data into an access database which is then put into word, which is then moved into In Design and formatted manually. This is for a large A4 sized brochure book.
    Is there anyway to design sections of  the document as a report so that you can set sections and somehow automatically import the data. Similar to a report in Access where you set a repeating section and tell it which part of the information you want to be displayed.
    The section of the brochure is the same for every page so I don't see any reason for the process to be so manual.
    Does anyone know if there is an easy way to do what I want to achieve? Would In Design take a XML file, or have the ability to read from an access database?
    I'm a developer not a designer. I'm going to be outputting a file for the designer to use to create the brochure.
    Thanks in advance!
    Regards,
    Rebecca

    Hi,
    It's quite likely that CatBase could do this job for you - it would take the data from a database, spreadsheet, or whatever, apply tags to fomrat it, and then create a tagged text file which you'd import into your InDesign layout. Everything would automatically be formatted as required ... Here's a short video that shows a Yellow Pages-style directory being published - probably not exactly what you're after but it gives you an idea of how the process works.
    For more info about CatBase see www.catbase.com or contact me! ([email protected])
    Pat

  • Transfer tables from sql server 2008 to oracle 11g

    Hi ,
    I have to transfer 2-3 tables (5 Lacs records each) from sql server 2008 to oracle Ent. 11g (platform Linux)
    i have to upload data with some conditions.
    not allowed to install oracle client on sql server for link server. i can get data in csv file
    which will be the best possible way
    1) using Db link on oracle server
    2) get data into csv file and load using SQL loader
    3) get data into csv and load using external table
    4) using tool - sql developer
    Thanx,

    user10188639 wrote:
    Hi ,
    I have to transfer 2-3 tables (5 Lacs records each) from sql server 2008 to oracle Ent. 11g (platform Linux)
    i have to upload data with some conditions.
    not allowed to install oracle client on sql server for link server. i can get data in csv file
    which will be the best possible way
    1) using Db link on oracle server
    2) get data into csv file and load using SQL loader
    3) get data into csv and load using external table
    4) using tool - sql developer
    Thanx,
    >which will be the best possible way
    which metric measures best? is higher or lower value better?
    How will you, I or anyone recognize which option is best solution?

  • Best ways to view data, total records of an application table ie VBAK

    Hi all,
    What is the best way to view data of an application table in the source system?
    I know about SE16....but are there other ways to know details ie the total no of records and different field information about a
    application table ie VBAK in source R3?
    Also, using SE16 when i checked for VBAK and clicked on the "number of enteries" then it showed 0...however
    when i directly checked from the sqlplus then i found about 5000 records in there in VBAK. I am not sure why
    via SE16 it showed 0. Does anybody have any idea what i missed here?
    Thanks...will give points for ur input.
    ak

    I tried "number of enteried" on se16 and it shows 0 enteries without any selection criterion...i cheked by putting relevant time range as well but it shows 0...
    As i told that when i checked VBAK separately via logging to database directly then i did find 5678 rows there.
    Please note that this is a new demo version....so i thought that i first need to activate the table which i did using tcode SE11. Now the VBAK table is active but still via SE16 shows 0 nuber of enteris....
    Can anybody please advise here..
    Thx
    ak

  • Tabular Modeling. What is the best practice for importing data into VS to limit the records in the designer?

    Should I wrap the queries in a procedure with a @StartDate and @EndDate and create a test partition to pass a small date range? 
    Or can i use the Table properties screen to put the command there and will it run and not be affected or affect the partitions?This would be nice if this SQL statement on this screen was independent of the partitions and I could just leave it with the the
    command text = EXEC TransactionDetail '2014-01-01', '2014-05-31' Especially since if you have many tables that load based on a date range. i would not want to jump in and change that query on all of them.
    Is there a a way to have a parameter in the project so all tables would get the same @startDate and @EndDate so I could change it in one place?
    And I am not stuck to these questions\options, If there is a better way to mass change the queries to run a subset of data for the designer I'd like to hear it.
    Thank You,
    Phil

    Hi Phil,
    According to your description, you are looking for the best way to control the rows that are loaded into a table, right?
    When importing data to a table of tabular model, we can apply filters to control the rows that are loaded into a table. After you have imported the data, you cannot delete individual rows. However, you can apply custom filters to control the way that
    rows are displayed. Rows that do not meet the filtering criteria are hidden. For the detail information about it, please refer to the link below.Filter Data in a Table (SSAS Tabular)
    If I have anything misunderstand, please point it out.
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here
    Charlie Liao
    TechNet Community Support

  • Is there a way to import data from excel -when one of the columns in excel is hyperlink column?

     Is there a way to import data from excel  - so if a column is hyperlink - the whole data will move to the list (text + link of the hyperlink column)?
    keren tsur

    Hi,
    According to your description, you want to export excel which contains a hyperlink column to SharePoint list.
    Refer to the following steps:
    Open the Excel, insert/create the table. 
    Now click on any cell of table and go to the ‘Design Tools’.
    Click on the Export and then ‘Export table to SharePoint List’.
    You will see a popup where you need to provide the URL of SharePoint site, list name and description.
    Then click on next, On the next screen you will see columns with data types which are going to create in SharePoint list.
    Now click finish and wait until the operation gets finished. You will see that list gets created in SharePoint site with the records.
    Here are two links, you can use as a reference:
    http://sharepointrhapsody.com/2013/03/25/how-to-create-a-connected-excel-file-to-sharepoint-list/
    http://social.technet.microsoft.com/wiki/contents/articles/18705.sharepoint-2013-how-to-export-excel-sheet-to-sharepoint-list.aspx
    Best Regards,
    Lisa Chen
    Lisa Chen
    TechNet Community Support

  • Best Way to Load Data in Hash Partition

    Hi,
    I have partitioning by Hash on a Large Table of 5 TB. We have to load Data say more than 500GB daily on that table from ETL.
    What is the best way to Load data into that Big Table which has hash Partition .
    Regards
    Sahil Soni

    Do you have any specific requirements to match records to lookup tables or it just a straight load - that is an insert?
    Do you have any specific performance requirements?
    The easiest and fastest way to load data into Oracle is via external file and parallel query/parallel insert. Remember that parallel DML is not enabled by default and you have to do so via alter session command. You can leverage multiple CPU cores and direct path operation to perform the load.
    Assuming your database is on a linux/unix server - you could NFS load the file if it is on a remote system, but then you will be most likely limited by network transfer speed.

Maybe you are looking for