Moving source database

The source databse was moved from Oracle to Microsoft SQL server. What is the easiest way to fix all mappings and process flows to work on the new database?
Tx, Kets

Hi Kets,
you need to perform these steps
1) create in OWB new module for MS SQL server and import all tables/views which you need
2) identify mappings which reference to tables/views from old Oracle source (with Impact Analysis OWB feature, OMB*Plus or via design repository public view)
3) update mapping table/view operators (identified on previous steps) with new reference to MSSQL module (manually - with Inbound Synchonization function or via OMB*Plus with OMBSYNCHRONZIE command - for idea with OMB*Plus look at Re: best practice on 'from dev to test'move. in this script you need remove backslah before square brackets)
4) deploy all modified mappings (again - manualy via Control Cenetr or via OMB*Plus with OMBDEPLOY command)
If number of modified mappings is quite small (say no more than 20 mappings) - maybe it is simpler perform all steps manually.
In case of large updates - it is possible to design single OMB*Plus scripts for steps 2-4
Don't see any reason why you want to update processflow...
Regards,
Oleg

Similar Messages

  • Moving CMS database from one server to another

    Using the following:
    Crystal Enterprise 10, installed on W2K3 Server R2 (not making any changes to this server);
    CMS database, installed on W2K Server, SQL Server 2000, SP3
    New database server: W2K3 Server R2, SQL Server 2005, SP2
    I found what I think are the steps to moving the database from one place to another in the CE10 Administrator's guide (pages 291-300, but I want to make sure there aren't any "gotchas" when moving the database location and/or upgrading the database version.  I'm primarily a DBA, so I already know about the Schema security change going from SQL 2000 to SQL 2005.
    In a nutshell, here's my plan:
    1. Take fresh backup of the existing CMS database
    2. Stop Crystal Management Server
    3. Click "Specify CMS Data Source" on the toolbar.
    4. ETC--following the steps in the admin guide
    If anyone has done this and has run into any problems (what I'm expecting) or had it go smoothly (I would be surprised), please let me know what to expect.
    Thanks!

    All you really need to do is stop the crystal services.
    Backup the CMS and Auditing database
    Detach the CMS and Auditing database
    Copy CMS and Auditing databases to new DB server
    Attach the CMS and Auditing databases
    Set the compatibility mode
    On the servers/s that host the CMS service, modify the ODBC connections to point to the new database server
    Start the CMS service, fire up the Crystal services.
    One thing to note, I don't think SQL 2005 is a supported platform for CE10 - this doesn't mean it won't work and since the support lifecycle for the product is long gone it doesn't really make much difference however you may want to run some thorough testing prior to the move.
    Edited by: James Pretorius (CCLTD) on Aug 5, 2009 9:20 PM

  • Exchange 2010 - 2013 Public Folder Migration target database gets much bigger than source database

    Hi,
    i have a question about the pf migration from Exchange 2010 to Exchange 2013. The source database ist about 370 GB. The migration process is still running. At the moment the target database is already 491GB big. Is that a normal behaviour during the migration
    process.
    Thanks

    Hi,
    After you have finished the public folder migration process, you can compare the item count of public folders to make sure all public folders have been moved to Exchange 2013.
    You can check this by comparing the snapshots before and after the migration. You can refer to the "How do I know this worked?" section in the following article.
    Migrate Public Folders to Exchange 2013 From Previous Versions
    http://technet.microsoft.com/en-us/library/jj150486(v=exchg.150).aspx
    Best regards,
    Belinda
    Belinda Ma
    TechNet Community Support

  • Locks made against source database (SQLServer 6.5 to 9.0.1.2 migration)

    I was wondering if there is any information about the types of locks that OMWB will take out on the source SQLServer database in a migration. From what I can tell these locks occur on the source database while the table data is loaded in to the target Oracle database. Are they taken at any other time during the migration process? How does OMWB determine what locks to take?
    I need to migrate several production SQLServer 6.5 databases and I need to understand how OMWB will perform locks so I don't bottleneck our production SQLServer databases while they are converted to Oracle. Most of the migration work will be performed off hours, but I still would like some insight into the locking. In a related question, how does OMWB maintain consistency between tables during the migration process?
    Thanks in advance for your help.
    -Reid

    Hi,
    You will be able to migrate from 6.5 to 9.3.1, if the 6.5 version is on a seperate server from the planned 9.3.1 version then you can use the migration wizard to do so.
    If you want to migrate users and groups make sure the 9.3.1 version is still in native security mode, otherwise you will need to use esscmd/maxl to extract user information and then import.
    The migration wizard will not take the data so you should really export that from your source.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Source database in NOARCHIVELOG temporarily for maintenance

    Hi all,
    We have a big maintenance to run next week end. There will be mostly index rebuild, statistics gathered and so. We are using 10.2.0.3. Source database is a RAC one.
    We would like to put our database in NOARCHIVELOG temporarily during this maintenance windows. This may sound like an easy task, but i rather consider it as a risky one - I need a 100% safe procedure!!
    I wonderer, if there is arcticles in Oracle documentation or Metalink that could help.
    I think I know several of the steps required:
    - Stop the capture
    - Put the database in NOARCHIVELOG mode
    - Re-instantiate the tables using dbms_capture_adm.prepare_table_instantiation on the source database
    - Ré-instantiate the tables using dbms_apply_adm.set_table_instantiation_scn on the destination database
    - I think I will have to advance the SCN on the capture but I'm not sure if I have to change the FIRST_SCN or the START_SCN or both?
    - Put the database back in ARCHIVELOG mode
    - Start the capture
    I'm not sure in wich order i have to do these tasks.
    Thank you for any advice/link on that subjet!
    Jocelyn

    Thanks a lot Rijesh for your response,
    In the meantime, I do my own search and came to the same conclusion.
    I also experimented with note 471695.1 on Metalink (Required Steps to Recreate a Capture Process). It "almost" works! It recreates the capture correctly, but it does not carry the existing table rules... (the content of DBA_STREAMS_TABLE_RULES before dropping the capture).
    So it's probably going to be much easier to recreate the whole thing from scratch (we have an existing procedure that is very mature).
    Regards,
    Jocelyn

  • Why am I getting  'Unable to signon to the Source Database.' just prior to 'Running the DBTSFIX Report' step.

    We are upgrading from Tools 8.53 to Tools 8.54 in HCM. I am receiving logging event 'Unable to signon to the Source Database.' just prior to 'Running the DBTSFIX Report' step. I found case 1545297.1 where it said to choose initial move as opposed to move

    For Tools only upgrades, while creating the jobs, you should choose 'Initial' and not 'Move to production'
    An excerpt from http://docs.oracle.com/cd/E56736_01/psft/acrobat/Getting_Started_on_Your_PeopleTools_Upgrade_08_2014.pdf is given below
    MTP indicates that this step runs only in Move to Production upgrade passes. The step is not applicable to the initial pass. This is not applicable for PeopleSoft PeopleTools-only upgrades.

  • The 'Third Party Connections' drop down box on the 'Source Database' step in the Migration Wizard is empty.

    Here are the environmental details:  
    SQLDeveloper 32-bit Version 3.2.20.09 (Build MAIN-09.87)
    Oracle Database:  12C single instance (no RAC)
    Oracle client:  12C 32 bit
    Windows 64 bit desktop
    I am attempting to migrate an MS Access application to Oracle.  I pick the ‘Migrate’ selection from the ‘Tools’ Menu SQL Developer.  This puts me into the Migration Wizard.  When I get to the ‘Source Database’ step of the migration wizard, the ‘Choose the Third Party database from which you are migrating’  drop-down box is not being populated with the current available connections. It is empty.  If I type a valid connection name into the box, SQL Developer takes it on this screen but does not keep it for the next screen/step.  It has no persistence.  I even tried to make sure that each connection was open/connected before beginning the migration process.
    Due to the fact that I am attempting to migrate an MSAccess application, I am mixing 32-bit and 64 bit software.  This may be the issue as to why the drop down is losing addressability.  I have not applied any patches to this version of SQL Developer.

    Hi,
    If you need easy way to migrate Microsoft Access to Oracle probably it is more reasonable to review some tools specially designed to this purpose. For example:
    Access to Oracle converter by Intelligent Converters
    Access to Oracle converter by SpectralCore
    Access to Oracle converter by Interface Computers
    Hope, this will help.
    Sincerely,
    Vlad

  • Multiple source databases

    Hello,
    I need to load data into dimensions and cube from multiple source databases. We have separate transaction databases for each country, but data must be loaded into common target schema. At the moment we have created mappings which loads data from one source database.
    What is the best way to populate, for example customers dimension from customers transaction tables (located in multiple source databases) if business identifiers can overlap?

    I would break this up into two pieces: the load from the source systems into the staging area, and then the load from the staging area into the target table, which you mentioned was a dimension.
    There is no reason you would have to use views or synonyms if you don't want to. You could approach this either with a SET operator, in this case a union, that brought all the sources together into the staging table. However, I would likely create multiple mappings for each source system. I would create a mapping for each table that you have to pull data from, and then create a separate process flow that controls all the tables being pulled from a particular source database. This allows you to schedule the extractions from the source systems independently of each other in case they need to run at different times. The end result of all of this will be a series of staging tables that are loaded from source tables from all the different locations. For each of the singular mappings, I would create a constant that defined the source of the data for this mapping, and load that constant into a column called SOURCE_SYSTEM, or perhaps COUNTRY, as you mentioned. Then, this SOURCE_SYSTEM/COUNTRY column + SOURCE_SYSTEM_ID would serve as the natural key.
    You can choose to persist these tables (as an ODS of sorts) or not... that is up to you and your requirements.
    Finally, I would create a single mapping that loads the final target table from the single staging table.
    Let me know if this doesn't make sense.
    Regards,
    Stewart Bryson

  • ADF (View Object) with open source database

    Hi all,
    Recently, I am evaluating JDeveloper 9.0.5.2 ADF with an open source database (Firebird).
    I follow the instruction to:
    1. Create an entity object
    2. Create a view object for the entity object in 1
    3. Create an application module for the view object in 2
    4. Right click on the application module and test -> all worked fine
    5. I created a jsp and drag the view (in 2) to it from Data Control Palette
    6. Test it in OC4J -> an error occurred (not because of jdbc config)
    JBO-27122: SQL error during statement preparation. Statement: SELECT * FROM (SELECT Department.DEPT_NO, Department.DEPARTMENT AS DEPARTMENT1, Department.HEAD_DEPT, Department.MNGR_NO, Department.BUDGET, Department.LOCATION, Department.PHONE_NO FROM DEPARTMENT Department) QRSLT ORDER BY DEPT_NO
    GDS Exception. 335544569. Dynamic SQL Error SQL error code = -104 Token unknown - line 1, char 16 SELECT
    To my knowledge, the error should be due to Firebird doesn't support the query syntax " select * (select col1, col2... from table) QRSLT where... order by... "
    I am quite headache as this query is automatically generated by JDeveloper. I am figuring out whether I can change it. It seems that I have to play with the ViewObjectImpl class.
    I do think ADF is an amazing technology. However, is it only so amazing with the Oracle technologies? Please correct me if I am wrong.
    Hons

    Dear Shay Shmeltzer,
    Thank you for your reply. After hours of trying on Firebird, I still get the error:
    JBO-27122: SQL error during statement preparation. Statement: SELECT * FROM (SELECT Department.DEPT_NO, Department.DEPARTMENT AS DEPARTMENT1, Department.HEAD_DEPT, Department.MNGR_NO, Department.BUDGET, Department.LOCATION, Department.PHONE_NO FROM DEPARTMENT Department) QRSLT ORDER BY DEPT_NO
    GDS Exception. 335544569. Dynamic SQL Error SQL error code = -104 Token unknown - line 1, char 16 SELECT
    I then downloaded the MySQL 4.0 and install on my system to do the exact procedures. No error! Thing works fine! Oh man...
    For the Firebird case, a strange thing to note is that the data table (I dragged) contains data, but the error message is on top of it.
    Actually, I don't have a particular database preference but I do hope ADF can work equally for these open-source databases. I think the problem may also be contributed by Firebird. Hope that I can figure it out this week.
    Regards,
    Hons

  • SQL Server 2000 to Oracle 9.2 problem [b]"Source database not captured"[/b]

    hi, first-timer here working with the latest omwb version.
    anyway
    i'm trying to do a offline migration of the "Northwind" demo database as an example, they are no problems with the process, everything works like a charm ... until is time to press the "finish" button in the wizard-summary screen.
    after that the output seems normal and there's no error, the completion pop-up message throws a 0 warnings 0 errors, supposedly ending the loading of the model, but the capture is labeled as aborted and the source database is not captured at all.
    i have done everything in the manual, creating the privileged user "omwb_user", seeing that the corresponding tablespace was created, giving enough space everywhere.
    so, what's the problem here?
    i will appreciate any help or hint
    thank in advance

    i thought that mi domain standard of "john.doe" was the problem, i take out all the users, and now there is only the "sa" and "guest" accounts.
    tried the online capture and it worked completely, tried the offline, the same problem:
    ** Started : Fri Oct 22 09:26:41 COT 2004
    ** Workbench Repository : Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Repository Connection URL: jdbc:oracle:thin:@localhost:1521:orcl
    ** The following plugins are installed:
    ** Microsoft SQLServer 2000 Plugin, Production Release 10.1.0.2
    ** Microsoft SQLServer 6.5 Plugin, Production Release 10.1.0.2
    ** Active Plugin : SQLServer2K
    EXCEPTION :SQLServer2KDisconnSourceModelLoad.loadSourceModel(): oracle.mtg.migration.MigrationStopException: java.lang.IndexOutOfBoundsException: Index: 15, Size: 15
    ** Shutdown : Fri Oct 22 09:27:14 COT 2004
    any ideas?
    thanks

  • Streams Replication:Source database Physical or Logical Standby DB

    Can the source database in streams replication be a physical or logical standby database ? If so, is the process of configuring streams the same as a regular database ? Are there any best practices or different configuration if the source is Logical or Physical standby DB ?
    Thanks in advance.

    Never done it, but I don't see any reason why it should not work.
    Streams, at capture site, is only a data dictionary game and in a logical standby your data dictionary is open read write.
    Streams, at capture site, never touch the source tables, in fact they may even not exists from Streams point of view,
    as it deals only with the redo that are generated.
    So Streams horizon is limited to the data dictionary, the log buffer, the archives and, in SYSAUX tablespace, all the LOGMNR_% tables. All these structures are read write in the logical standby. However, for the capture/propagation you may have to set to true the 'include_tagged_lcr' parameters.

  • Moving Publisher database to 2008R2 but subscriber 2005 will remain same. How to reconfigure this?

    We are planning to upgrade(not in place, moving to a new machine) the publisher(2005 EE) to 2008R2 EE but the subscriber(2005 EE) will be the same.
    Replication Type: Transactional replication
    Number of replicated databases: 1
    Existing
    SQL Server 2005 SP4 (Publisher & Distributer) with 8 publications
    SQL Server 2005 SP4 (Subscriber)
    New
    SQL Server 2005 SP4 (Publisher & Distributer) this will be upgrading to SQL Server 2008R2 SP2
    SQL Server 2005 SP4 (Subscriber) this will be same.
    After moving the database to new SQL Server 2008R2, how to re-configure the replication with the existing 2005 subscriber?
    There are hundreds of articles (objects) involving in the replication so it is not easy to reconfigure in GUI mode. So looking for plan to script the existing replication and plan to run those scripts on new server to re-configure is this possible?

    The best option is to script and recreate.  There is no real way to backup a replicated database and restore it on a new server and expect the publication to be intact on the new distributor and for it to pick up where it left off and continue.
    You can use the sync with backup option and log ship to the new server restore the final log ship with the keep replication switch and then rename the old server with the old publisher name. I have not tried this as part of an upgrade.
    Another option is to script the publication, drop the subscriptions, run the script on the new publisher and then do no-sync subscriptions.
    Using a script is the best way to do it, trying to do this via the gui is madness.
    looking for a book on SQL Server 2008 Administration?
    http://www.amazon.com/Microsoft-Server-2008-Management-Administration/dp/067233044X looking for a book on SQL Server 2008 Full-Text Search?
    http://www.amazon.com/Pro-Full-Text-Search-Server-2008/dp/1430215941

  • Columns mismatch between Source database and Source in Informatica

    Hi,
    I'm trying to create few columns into the datawarehosue tables. I can see the columns in my source system (Siebel CRM) in toad. When I access the source table via Informatica Powercenter Designer. I don't see that column in the source tabel. I have a doubt that whether I'm connecting to the right source database from Informatica. How do I check this and change this? The column which I'm looking for in the source system is not custom column. Its Siebel out of the box column. Could any one provide more information on this.
    Thanks in advance,
    Ravi

    You may have to reimport the table from the Source Analyzer workspace. When you import, make sure you are pulling from the correct source database and schema.
    Informatica does not automatically update the table definition if the source changes.

  • Changes in Source Database.

    Hi All,
    OWB Config Details is as follows.
    OWB ver 9.2 with Windows NT
    Oracle 9i.
    In near future we are expecting changes in source database i.e. source database from existing will change to a new database which may be same or different.
    The level of changes identified are as follows.
    Table Name Changes
    Column Name Changes
    Business Logic Changes
    View name changes
    I guess there will be changes in other parameter like ip address,sid, dblink ..etc.
    I knew in OWB 9.0.2.56 column level changes can be incorporated by re-importing the table & adding,modifying or deleting the column & by selecting in-bound / out-bound reconcillation option in mapping editor. I would like to know is it the same way we need to do in OWB 9.2 version or is there any better way or utility which can ease this task?
    As per my knowledge we will need to do it for each & every mapping & in case of business logic change, we will need to recreate the mapping.
    Can someone help me in this?
    Thanks in Advance.
    Regards,
    Vidyanand

    Hi Igor,
    You are talking about simplifying matters when there have been changes in the source, and that that is going to be in the release of Q2-2004.
    Which release might that be and - since it is already mid september - is it already 'on the market'?
    What I'm looking for is an 'easy' way to incorporate changes in source objects in all mapping objects that are not database objects.
    So far this has always been a manual job, since only DB-objects can be reconciled. If you have a very big mapping with a lot of objects between source and target DB-object, this has proven to be extremely time-consuming, since all objects have to be opened and the attribute properties of altered columns manually adjusted to the new situation.
    Is this wat you are talking about that is going to be greatly improved?
    If not, what would be the best approach for these sorts of operations?
    Cheers, Patrick
    ps Using version 9.2.0.2.8 and waiting for DBA to patch to 9.2.0.4

  • Is mysql a free open source database or do you need a license for it in prd

    is mysql a free open source database or do you need a license for it in production.

    Hi,
    It should be free under GPL (General Public license).Usually it will comes with your operating system distribution (eg:RHEL)
    If you download it from Oracle site, it should have the commercial license tied to it, and you need to pay for that if it is in use for production.
    http://www.mysql.com/about/legal/licensing/index.html
    regards,
    mrak

Maybe you are looking for

  • IPhoto quits unexpectedly using VectorKit plug-in

    I am running OS X 10.9.5 and iPhoto 9.5.1. Holding down Opt + Ctrl when opening iPhoto doesn't help. Removing iPhoto and its pkg files and reinstalling iPhoto from App Store Purchases did not help. I did download iPhoto Library Manager.app and made a

  • Need help in troubleshooting communication channel using soap adapter

    Hello friends I have a simple scenario. I have created 2 communication channels, one for searching a record and another for updating it with a new value Both uses SOAP as sender and RFC as receiver with a Standard BAPI. Everything was working ok, but

  • Packaged applications NOT COMPATIBLE

    Hy guys, I tried two apps with my newly installed oracle apex running in ubuntu server, no one running, i tried the customer tracker app and others but no one runnig. Any hint? Thanks a lot in advance! Maurizio Oracle Database 10g Express Edition Rel

  • Publishing pl/sql as web services using axis

    Hello, I am a newbie to web services. I want to publish pl/sql packages as web services using axis and tomcat. Can you please tell me where to start - I will appreciate any information, links to tutorials and docs. Can you please tell me what I need

  • N97 mass memory dataloss

    My friend has N97 phone with some memory problems. This morning she had appr. 1200 emails (mostly headers only) and appr. 900 textmessages in her phone. That all was in E: mass memory. It seemed like the amount of these messages slowered the phone. I