Database Upgrade in Data Gaurd

Hi all,
i have an active data guard running
i wont to upgrade my database what should i do?? should i stop my dataguard
shutdown both primary database and secondary database
upgrade primary first then secondary database, then i'll make again dataguard  ??
or any other best recommendations ??

Hi CKPT,
Can you help me to find the best method how to upgrade 10.2.0.4 to 11.2.0.3 with physical standby database, in my opinion :
1. Stop applying redo log
2. Stop all standby service
3. Upgrade standby
4. Switchover to upgrade standby
5. Upgrade old primary if no error in step 4
6. Switchback
Please give any suggestion to best method and minimal downtime, and note i have more 4TB data, thank's.

Similar Messages

  • Database Upgrade using Data Pump

    Hi,
    I am moving my database from a Windows 2003 server to a Windows 2007 server. At the same time I am upgrading this database from 10g to 11gR2(11.2.0.3).
    therefore I am using the export / import method of upgrade ( via Data Pump not the old exp/imp ).
    I have successfully exported by source database and have created the empty shell database ready to take the import. However I have a couple of queries
    Q1. regarding all the SYSTEM objects from the source database. How will they import given that the new target database already has a SYSTEM tablespace
    I am guessing I need to use the TABLE_EXISTS_ACTION option for the import. However should I set this to APPEND, SKIP, REPLACE or TRUNCATE - which is best ?
    Q2. I am planning to slightly change the directory structure on the new database server - would it therefore be better to pre-create the tablespaces or leave this to the import but use the REMAP DATAFILE option - what is everyone's experience as to which is the better way to go ? Again if I pre-create the tablespaces, how do I inform the import to ignore the creation of the tablespaces
    Q3. these 2 databases are on the same network, so in theorey instead of a manual export, copy of the dump file to the new server and then the import, I could use a Network Link for Import. I was just wondering where there any con's of this method over using the explicit export dump file ?
    thanks,
    Jim

    Jim,
    Q1. regarding all the SYSTEM objects from the source database. How will they import given that the new target database already has a SYSTEM tablespace
    I am guessing I need to use the TABLE_EXISTS_ACTION option for the import. However should I set this to APPEND, SKIP, REPLACE or TRUNCATE - which is best ?If all you have is the base database and nothing created, then you can do the full=y. In fact, this is probably what you want. The system tablespace will be there so when Data Pump tries to create it , it will just fail that create statement. Nothing else will fail. In most cases, your system tables will already be there, and this is ok too. If you do schema mode imports, you will miss out on some of the other stuff.
    Q2. I am planning to slightly change the directory structure on the new database server - would it therefore be better to pre-create the tablespaces or leave this to the import but use the REMAP >DATAFILE option - what is everyone's experience as to which is the better way to go ? Again if I pre-create the tablespaces, how do I inform the import to ignore the creation of the tablespacesIf the directory structure is different (which they usually are) then there is no easier way. You can run impdp but with sqlfile and you can say - include=tablespace. This will give you all of the create tablespace commands in a txt file and you can edit the text file to change what ever you want to change. You can tell datapump to skip the tablespace creation by using - exclude=tablespace
    Q3. these 2 databases are on the same network, so in theorey instead of a manual export, copy of the dump file to the new server and then the import, I could use a Network Link for Import. I >was just wondering where there any con's of this method over using the explicit export dump file ?The only con could be if you have a slow network. This will make it slower, but if you have to copy the dumpfile over the same network, then you will still see the same basic traffic. The pros are that you don't have to have extra disk space. Here is how I look at it.
    1. you need XX GB for the source database
    2. you need YY GB for the source dumpfile
    3. you need YY GB for the target dumpfile that you copy
    4. you need XX GB for the target databse.
    By doing network you get rid if 2*YY GB for the dumpfiles.
    Dean

  • EM console HA page says "Data Gaurd not configured for this database"

    I am attempting to use EM console's "High Availability Console" to create a standby database, but, the option is not available. Under "Data Gaurd Summary" the page shows "Oracle DataGaurd is not configured on this database."
    How do I configure this option so that I can use the console to create Standy databases? I have looked everywhere I can think of and cannot find any instructions!

    Click on your database name, and you'll be lead to a page which shows you all the statistics of your Db
    Click on the "Availability" Sub Tab.
    You'll see the option to "Add Standby Database". You can take it from there. (you'll be prompted to log in to your database -- log in with sys)
    Now click on "Add Standby Database"
    The process I normally follow is this
    1. RMAN your primary
    2. Create standby control file from primary.
    3. Get your init file from primary
    4. Copy everything over.
    5. Make the init file changes (fal client, fal server on primary, any kind of directory mapping, point the init file to the standby control file on the standby side)
    6. Bring up the standby db (doing all the other changes that are required)
    Then you can import that Db into OEM,
    or you could just use the Wizard from OEM :)

  • Oracle Database upgrade to 11G (Host_Command Issues)

    Sorry I'm posting this again. Original post in Database - Upgrade. I'm not getting a lot of clicks in that category
    We recently upgraded to Oracle 11g (Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production). Our PL/SQL release: (PL/SQL Release 11.2.0.2.0 - Production). We use a lot of host commands inside our PL/SQL stored procedures. One of these included fetching data from an external server and pushing the data into Oracle Tables.
    Most of this process is running fine, except the execution never completes. There is no issue data wise. It's just that the Host_command doesn't receive the completion status and seems to be running even though there's nothing going on in the background.
    The Oracle e-business suite's concurrent program based on this stored proc also never completes, though it doesn't do anything in the background. For testing purpose, our DBA used 2 insert statements one before and one after the host_command. The insert before the host_command worked but not the one after. Please help! This is happening after we upgraded to 11g 3 weeks ago!
    Thanks, Naveen Gagadam.

    Pl do not post duplicates - Upgrade to 11g, Host Command Inside Stored Proc doesn't end.
    I have moved your duplicate post to the EBS forums
    Srini

  • Database Upgradation from oracle 9.2.0.8 to oracle11gR2

    Hi
    we are using AIX 5.3 TL 10 SP4 and we are upgrading DB from oracle9.2.0.8 to oracle 11.2.0.2
    i am having some doubt from below pre-upgrade scripts out put.
    oratst@ebsdevdb on /tmp # sqlplus "/as sysdba"
    SQL*Plus: Release 9.2.0.8.0 - Production on Mon Jul 4 10:11:48 2011
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.8.0 - Production
    SQL> spool 11202_upgrade.txt
    SQL> @utlu112i.sql
    Oracle Database 11.2 Pre-Upgrade Information Tool 07-04-2011 10:17:06
    Script Version: 11.2.0.2.0 Build: 001
    Database:
    --> name: EBSTEST
    --> version: 9.2.0.8.0
    --> compatible: 9.2.0
    --> blocksize: 8192
    --> timezone file: V1
    Logfiles: [make adjustments in the current environment]
    --> The existing log files are adequate. No changes are required.
    Tablespaces: [make adjustments in the current environment]
    --> SYSTEM tablespace is adequate for the upgrade.
    .... minimum required size: 8546 MB
    --> TOOLS tablespace is adequate for the upgrade.
    .... minimum required size: 8 MB
    --> TEMP tablespace is adequate for the upgrade.
    .... minimum required size: 61 MB
    --> AR_TABLE tablespace is adequate for the upgrade.
    .... minimum required size: 33497 MB
    --> FND_TABLE tablespace is adequate for the upgrade.
    .... minimum required size: 45447 MB
    --> AMVD tablespace is adequate for the upgrade.
    .... minimum required size: 7 MB
    --> ASOD tablespace is adequate for the upgrade.
    .... minimum required size: 7 MB
    --> CCTD tablespace is adequate for the upgrade.
    .... minimum required size: 7 MB
    --> IBUD tablespace is adequate for the upgrade.
    .... minimum required size: 1 MB
    --> IEMD tablespace is adequate for the upgrade.
    .... minimum required size: 5 MB
    --> IEOD tablespace is adequate for the upgrade.
    .... minimum required size: 4 MB
    --> JTFD tablespace is adequate for the upgrade.
    .... minimum required size: 135 MB
    --> OKCD tablespace is adequate for the upgrade.
    .... minimum required size: 34 MB
    --> XDPD tablespace is adequate for the upgrade.
    .... minimum required size: 7 MB
    --> XNPD tablespace is adequate for the upgrade.
    .... minimum required size: 7 MB
    --> APPS_UNDOTS1 tablespace is adequate for the upgrade.
    .... minimum required size: 849 MB
    Update Parameters: [Update Oracle Database 11.2 init.ora or spfile]
    Note: Pre-upgrade tool was run on a lower version 64-bit database.
    --> If Target Oracle is 32-Bit, refer here for Update Parameters:
    WARNING: --> "compatible" must be set to at least 10.1.0
    --> If Target Oracle is 64-Bit, refer here for Update Parameters:
    WARNING: --> "compatible" must be set to at least 10.1.0
    WARNING: --> "shared_pool_size" needs to be increased to at least 532 MB
    WARNING: --> "java_pool_size" needs to be increased to at least 128 MB
    Renamed Parameters: [Update Oracle Database 11.2 init.ora or spfile]
    -- No renamed parameters found. No changes are required.
    Obsolete/Deprecated Parameters: [Update Oracle Database 11.2 init.ora or spfile]
    --> optimizer_max_permutations 10.1 OBSOLETE
    --> row_locking 10.1 OBSOLETE
    --> undo_suppress_errors 10.1 OBSOLETE
    --> max_enabled_roles 10.1 DEPRECATED
    --> enqueue_resources 10.2 OBSOLETE
    --> sql_trace 10.2 DEPRECATED
    --> background_dump_dest 11.1 DEPRECATED replaced by "diagnostic_dest"
    --> user_dump_dest 11.1 DEPRECATED replaced by "diagnostic_dest"
    Components: [The following database components will be upgraded or installed]
    --> Oracle Catalog Views [upgrade] VALID
    --> Oracle Packages and Types [upgrade] VALID
    --> JServer JAVA Virtual Machine [upgrade] VALID
    --> Oracle XDK for Java [upgrade] VALID
    --> Real Application Clusters [upgrade] INVALID
    --> Oracle Text [upgrade] VALID
    --> Oracle XML Database [install]
    --> Oracle Java Packages [upgrade] VALID
    --> Oracle interMedia [upgrade] VALID
    --> Spatial [upgrade] VALID
    Miscellaneous Warnings
    WARNING: --> Passwords exist in some database links.
    .... Passwords will be encrypted during the upgrade.
    .... Downgrade of database links with passwords is not supported.
    WARNING: --> Deprecated CONNECT role granted to some user/roles.
    .... CONNECT role after upgrade has only CREATE SESSION privilege.
    WARNING: --> Database is using a timezone file older than version 14.
    .... After the release migration, it is recommended that DBMS_DST package
    .... be used to upgrade the 9.2.0.8.0 database timezone version
    .... to the latest version which comes with the new release.
    WARNING: --> Database contains INVALID objects prior to upgrade.
    .... The list of invalid SYS/SYSTEM objects was written to
    .... registry$sys_inv_objs.
    .... The list of non-SYS/SYSTEM objects was written to
    .... registry$nonsys_inv_objs.
    .... Use utluiobj.sql after the upgrade to identify any new invalid
    .... objects due to the upgrade.
    .... USER ADWANIAH has 1 INVALID objects.
    .... USER APPS has 34 INVALID objects.
    .... USER CHAKKACHANSA has 1 INVALID objects.
    .... USER FA_USER has 9 INVALID objects.
    .... USER SYSSEL has 2 INVALID objects.
    .... USER SYSTEM has 30 INVALID objects.
    .... USER XXITDFINAD has 7 INVALID objects.
    WARNING: --> SYSTEM schema default tablespace has been altered.
    .... The SYSTEM schema default tablespace is currently set to TOOLS.
    .... Prior to upgrading your database please reset the
    .... SYSTEM schema default tablespace to SYSTEM using the command:
    .... ALTER USER SYSTEM DEFAULT TABLESPACE SYSTEM;
    WARNING: --> Database contains schemas with objects dependent on DBMS_LDAP package.
    .... Refer to the 11g Upgrade Guide for instructions to configure Network ACLs.
    .... USER APPS has dependent objects.
    Recommendations
    Oracle recommends gathering dictionary statistics prior to
    upgrading the database.
    To gather dictionary statistics execute the following commands
    while connected as SYSDBA:
    EXECUTE dbms_stats.gather_schema_stats('CTXSYS',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('DBSNMP',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('MDSYS',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('ORDPLUGINS',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('ORDSYS',options=>'GATHER'
    estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('OUTLN',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('SYS',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    EXECUTE dbms_stats.gather_schema_stats('SYSTEM',options=>'GATHER'
    ,estimate_percent=>DBMS_STATS.AUTO_SAMPLE_SIZE
    ,method_opt=>'FOR ALL COLUMNS SIZE AUTO'
    ,cascade=>TRUE);
    Oracle recommends removing all hidden parameters prior to upgrading.
    To view existing hidden parameters execute the following command
    while connected AS SYSDBA:
    SELECT name,description from SYS.V$PARAMETER WHERE name
    LIKE '\_%' ESCAPE '\'
    Changes will need to be made in the init.ora or spfile.
    Oracle recommends reviewing any defined events prior to upgrading.
    Changes will need to be made in the init.ora or spfile.
    Oracle recommends reviewing any defined events prior to upgrading.
    To view existing non-default events execute the following commands
    while connected AS SYSDBA:
    Events:
    SELECT (translate(value,chr(13)||chr(10),' ')) FROM sys.v$parameter2
    WHERE UPPER(name) ='EVENT' AND isdefault='FALSE'
    Trace Events:
    SELECT (translate(value,chr(13)||chr(10),' ')) from sys.v$parameter2
    WHERE UPPER(name) = '_TRACE_EVENTS' AND isdefault='FALSE'
    Changes will need to be made in the init.ora or spfile.
    SYSAUX Tablespace:
    [Create tablespace in the Oracle Database 11.2 environment]
    --> New "SYSAUX" tablespace
    .... minimum required size for database upgrade: 500 MB
    From above output script
    1)SYSTEM tablespace is adequate for the upgrade.
    .... minimum required size: 8546 MB
    From system tablespace minimum required size means free space from system tablespace or total size of system tablespace?
    i am going to use dbua utility for upgrade means does it alter tablespace as per output of scripts or we need to manually alter tablespace size as per requirement?
    2)WARNING: --> Database contains schemas with objects dependent on DBMS_LDAP package.
    .... Refer to the 11g Upgrade Guide for instructions to configure Network ACLs.
    .... USER APPS has dependent objects.
    what is above warning and how can i configure ACL's for APPS User?
    Thanks
    With Regards
    A-Z

    Hi
    oratst@ebsdevdb on /ebdbh/11g/data/cfgtoollogs/dbua/ebstest/upgrade1 # more Upgrade_Directive.log
    Connected.
    ODMA_DIRECTIVE:VERSION:9.2.0.8
    ODMA_DIRECTIVE:MIGRATE_SID:
    ODMA_DIRECTIVE:ORA:IGNORE:06512:
    ODMA_DIRECTIVE:ORA:FATAL:00600:
    ODMA_DIRECTIVE:ORA:FATAL:01012:
    ODMA_DIRECTIVE:ORA:FATAL:01031:
    ODMA_DIRECTIVE:ORA:FATAL:01034:
    ODMA_DIRECTIVE:ORA:FATAL:01078:
    ODMA_DIRECTIVE:ORA:FATAL:01092:
    ODMA_DIRECTIVE:ORA:FATAL:01109:
    ODMA_DIRECTIVE:ORA:FATAL:01119:
    ODMA_DIRECTIVE:ORA:FATAL:01507:
    ODMA_DIRECTIVE:ORA:FATAL:01722:
    ODMA_DIRECTIVE:ORA:FATAL:03113:
    ODMA_DIRECTIVE:ORA:FATAL:03114:
    ODMA_DIRECTIVE:ORA:FATAL:07445:
    ODMA_DIRECTIVE:ORA:FATAL:12560:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01650:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01651:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01652:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01653:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01654:
    ODMA_DIRECTIVE:ORA:RECOVER_TBS:01655:
    ODMA_DIRECTIVE:ORA:RECOVER_ROLL:01562:
    ODMA_DIRECTIVE:ORA:RECOVER_INIT:04031:
    ODMA_DIRECTIVE:SCRIPT:UPGRADE:rdbms/admin/catupgrd.sql:
    ODMA_DIRECTIVE:BOUNCE_DATABASE:UPGRADE:
    ODMA_DIRECTIVE:SCRIPT:UPGRADE:rdbms/admin/catuppst.sql:
    ODMA_DIRECTIVE:SCRIPT:UPGRADE:sqlplus/admin/help/hlpbld.sql helpus.sql:
    Thanks
    With Regards
    A-Z

  • How can I determine what sites are being referenced within Central Admin Upgrade and Migration Manage Databases Upgrade Status?

    When I go to Central Admin > Upgrade and Migration  > Manage Databases Upgrade Status, I have 2 content databases which have the status:
    Database is up to date, but some sites are not completely upgraded.
    How can I determine which sites are not completely upgraded?

    Manage Databases Upgrade Status will provide you all active and offline DB details, you can get same result
    using below PowerShell cmdlet.
    Get-SPDatabase and Get-SPContentDatabase will provide all active database/Content DB in Farm which include Service application db, central admin DB.
    Get-SPDatabase | Format-Table Name, ID
    Coming back to your question, if you find that there are some site are not completely upgraded then run below command and understand the cause if issue on specific DB.
    Test-SPContentDatabase WSS_ContentDB_Name
    If you find any missing file issue in DB then resolve these issue to upgrade content database.
    (verify all customizations referenced within the content database are also installed in the web application. This cmdlet can be issued
    against a content database currently attached to the farm, or a content database that is not connected to the farm )
    Use the Upgrade-SPContentDatabase cmdlet
    to resume a failed database upgrade or begin a build-to-build database upgrade against a SharePoint content database
    Upgrade-SPContentDatabase WSS_Content
    reference:
    http://technet.microsoft.com/en-us/library/ff607813(v=office.15).aspx
    http://technet.microsoft.com/en-us/library/ff607941(v=office.15).aspx
    If my contribution helps you, please click Mark As Answer on that post and
    Vote as Helpful
    Thanks, ShankarSingh(MCP)

  • Directory ORACLE_HOME doesn't exist. in Database Upgrade 11.2.0.2

    Hi all,
    I am following the below document for database upgrade on EBS 12.1.3 database from 11.1.0.7 to 11.2.0.2.
    "Interoperability Notes EBS R12 with Database 11gR2 [ID 1058763.1]"
    We have completed upto
    6.     Install Oracle Database 11g Products from the 11g Examples CD
    but not "3.Update application tier context file with new database listener port number (conditional) "
    as it was conditional
    Old home:-F:\oracle\PROD\db\tech_st\11.1.0
    New Home:-F:\app\mocppd\product\11.2.0\dbhome_1
    set ORACLE_HOME=F:\app\mocppd\product\11.2.0\dbhome_1
    set ORACLE_BASE=F:\app\mocppd
    set LD_LIBRARY_PATH=F:\app\mocppd\product\11.2.0\dbhome_1\LIB
    set PERL5LIB=F:\app\mocppd\product\11.2.0\dbhome_1\perl\libcd F:\app\mocppd\product\11.2.0\dbhome_1\nls\data\old
    perl cr9idata.pl
    Got error on
    "7.Create nls/data/9idata directory" as
    perl cr9idata.pl
    F:\app\mocppd\product\11.2.0\dbhome_1\nls\data\old>perl cr9idata.pl
    cygwin warning:
      MS-DOS style path detected: F:/app/mocppd/product/11.2.0/dbhome_1
      Preferred POSIX equivalent is: /cygdrive/f/app/mocppd/product/11.2.0/dbhome_1
      CYGWIN environment variable option "nodosfilewarning" turns off this warning.
      Consult the user's guide for more details about POSIX paths:
        http://cygwin.com/cygwin-ug-net/using.html#using-pathnames
    Directory /F:/app/mocppd/product/11.2.0/dbhome_1/nls/data doesn't exist. Can't p
    roceed.I didnot understand why the "Directory /F:/app/mocppd/product/11.2.0/dbhome_1/nls/data doesn't exist" has '/'
    in starting of the path
    Thanks
    Nawneet

    but why it taking /usr/bin/perl. we are in windows .This is an expected output since you use Cygwin.
    Please set perl to the one you have under RDBMS %ORACLE_HOME% and try again (%ORACLE_HOME%\appsutil\perl).
    Thanks,
    Hussein

  • Zen 4 - Zen 6.5 Inventory database upgrade fails

    Hi all,
    Have installed Zen 6.5 over a Zen 4 server (nw65) running inventory.
    Database is on a seperate oracle server.
    When I start inventory, it starts the database upgrade and then fails
    with a java exception part way through, and then says "stopping the
    upgrade service".
    Following is the log. Anyone know which bit is falling over and why?
    There's a few errors in the log and I don't know which are cosmetic/non
    critical. Either way, I've researched each and can't find any fixes
    anyway.
    cheers
    Dave
    [12/8/05 14:58:40.307] ZENInv - Server Config: The severconfig service
    start is pending...
    [12/8/05 14:58:40.464] ZENInv - Server Config: Getting Cascade INit
    time
    [12/8/05 14:58:41.462] ZENInv - Server Config: Database Location
    Policy DN is CN=Server Package - DAN02:General:ZENworks
    Database.OU=POLICIES.OU=APPLICATIONS.O=CHISHOLM
    [12/8/05 14:58:43.573] ZENInv - Server Config: The severconfig service
    is started...
    [12/8/05 14:58:43.707] ZenInv - Upgrade Service: Info : Read
    zenInvComponentStatus = 6.5 12 0
    [12/8/05 14:58:43.707] ZenInv - Upgrade Service: Inventory Server
    zenInvComponentStatus = 6.5 12 0
    [12/8/05 14:58:43.720] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks.desktop.inventory.storer.Stor erServiceInit'
    [12/8/05 14:58:43.740] ZENInv - Storer: Start with
    'ServiceStatusChangeListener' invoked
    [12/8/05 14:58:44.824] ZenInv - Upgrade Service: Unzipping the file
    migrate.zip
    [12/8/05 14:58:46.718] ZenInv - Upgrade Service: Transforming the
    Migration SQL files...
    [12/8/05 14:58:51.768] ZENInv - Storer: zenworks.isClosed =true
    [12/8/05 14:58:53.699] ZENInv - Storer: The current NDS tree=CIT
    [12/8/05 14:58:53.724] ZENInv - Storer: Storer from dblocation user:
    MW_DBA driver: oracle.jdbc.driver.OracleDriver protocol: jdbc:oracle:
    subname: thin:@ port: 1522
    [12/8/05 14:58:53.728] ZENInv - Storer: url[From DBObject]:
    jdbc:oracle:thin:@ora01a.chisholm.vic.edu.au:1522: dw
    [12/8/05 14:58:53.756] ZENInv - Storer: wsdn:
    CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLMi = 0
    [12/8/05 14:58:53.973] ZENInv - Storer: Loading Storer test properties
    file
    [12/8/05 14:58:54.053] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks.desktop.inventory.strconverte r.STRConverterServiceInit'
    [12/8/05 14:58:54.059] ZENInv - STRConverter: Start with
    'ServiceStatusChangeListener' invoked
    [12/8/05 14:58:54.755] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:54.914] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:55.012] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:55.729] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:55.780] ZENInv - Storer: Database is not
    initialized..will retry after 300 seconds
    [12/8/05 14:58:55.787] ZENInv - Storer: Storer: Database not
    initialized as it is being upgraded ...will retry after 300 seconds.
    [12/8/05 14:58:55.788] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:55.826] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:55.860] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:58:57.439] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:04.100] ZENInv - STRConverter: Loading STRConverter test
    properties file
    [12/8/05 14:59:04.201] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks.desktop.inventory.senderrecei ver.control.ReceiverServiceInit'
    [12/8/05 14:59:15.618] ZENInv - SendRec Common: entPushDir =
    SYS:\ZENworks\Inv\ScanDir\EntPushDir
    [12/8/05 14:59:15.618] ZENInv - SendRec Common: entMergeDirD =
    SYS:\ZENworks\Inv\ScanDir\EntMergeDir
    [12/8/05 14:59:15.618] ZENInv - SendRec Common: dbDirD =
    SYS:\ZENworks\Inv\ScanDir\DbDir
    [12/8/05 14:59:15.625] ZENInv - SendRec Common: serverName = DAN02
    [12/8/05 14:59:15.626] ZENInv - SendRec Common: serviceDN =
    CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:15.626] ZENInv - SendRec Common: treeName = CIT
    [12/8/05 14:59:15.626] ZENInv - SendRec Common: hasSSD = false
    [12/8/05 14:59:15.626] ZENInv - SendRec Common: hasISD = false
    [12/8/05 14:59:15.626] ZENInv - SendRec Common: hasESD = true
    [12/8/05 14:59:15.632] ZENInv - SendRec Common: hasDB = true
    [12/8/05 14:59:15.633] ZENInv - SendRec Common: securityDir =
    SYS:\PUBLIC\ZENWORKS\WMINV\PROPERTIES
    [12/8/05 14:59:15.661] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks3x.desktop.inventory.senderrec eiver.control.ReceiverServiceInit'
    [12/8/05 14:59:15.882] ZENInv - CascadedBaseTime Server:
    zenInvCascadeBaseTimeService: CBTServiceController: Startup Properties:
    {basetime=Thu Jan 01 10:00:00 EST 1970,
    servicename=zenInvCascadeBaseTimeService, portnumber=0}
    [12/8/05 14:59:16.367] ZENInv - IFS Server: zeninvReceiverService:
    FileServiceController: Startup Properties: {chunksize=4096,
    transfers=100,
    rootdirectory=SYS:\ZENworks\Inv\ScanDir\EntPushDir \ZipDir,
    timeout=60000, servicename=zeninvReceiverService, portnumber=0}
    [12/8/05 14:59:25.771] ZENInv - TCPReceiver: cascadingBaseTime = 0
    [12/8/05 14:59:25.771] ZENInv - TCPReceiver: entPushDir =
    SYS:\ZENworks\Inv\ScanDir\EntPushDir
    [12/8/05 14:59:25.777] ZENInv - TCPReceiver: serverName = DAN02
    [12/8/05 14:59:25.777] ZENInv - TCPReceiver: serviceDN =
    CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:25.777] ZENInv - TCPReceiver: treeName = CIT
    [12/8/05 14:59:25.778] ZENInv - TCPReceiver: hasDB = true
    [12/8/05 14:59:25.971] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks.common.inventory.dictionaryup date.provider.DictProvider'
    [12/8/05 14:59:26.005] ZENInv - TCPReceiver: Receiver Started without
    CLUSTER
    [12/8/05 14:59:26.006] ZENInv - TCPReceiver: Receiver Binds to Port
    Number : 65432
    [12/8/05 14:59:26.073] ZENInv - IFS Server: zenInvDictProvider:
    FileServiceController: Startup Properties: {chunksize=4096,
    transfers=100, rootdirectory=SYS:\ZENworks\Inv\server\DictDir,
    timeout=60000, servicename=zenInvDictProvider, portnumber=0}
    [12/8/05 14:59:26.183] Service Manager: start(ServiceDataAccessor,
    String[]) not found in
    'com.novell.zenworks.common.inventory.dictionaryup date.consumer.DictConsumer'
    [12/8/05 14:59:28.918] ZENInv - Dictionary Consumer:
    DictConsumerUtility::getUpdatePolicyDN: getDictionaryUpdatePolicy
    returned attribs.returnValue = 0
    [12/8/05 14:59:28.919] ZENInv - Dictionary Consumer:
    DictConsumerService::DictDownloadThread::run:
    UpdatePolicyNotFoundException.
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.UpdatePolicyNotFoundException
    at
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.DictConsumerUtility.getUpdatePolicyDN (DictConsumerUtility.java:237)
    at
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.DictConsumerService$DictDownloadThrea d.setUpdatePolicyAttribs(DictConsumerService.java: 688)
    at
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.DictConsumerService$DictDownloadThrea d.getFileClientProperties(DictConsumerService.java :616)
    at
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.DictConsumerService$DictDownloadThrea d.transferFiles(DictConsumerService.java:429)
    at
    com.novell.zenworks.common.inventory.dictionaryupd ate.consumer.DictConsumerService$DictDownloadThrea d.run(DictConsumerService.java:211)
    [12/8/05 14:59:28.921] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:28.953] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:28.986] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:29.326] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:29.326] ZENInv - Dictionary Consumer:
    DictConsumerService::FileDownloadListener::downloa dFailed.
    [12/8/05 14:59:59.680] ZenInv - Upgrade Service: Starting Schema
    Migration
    [12/8/05 14:59:59.751] ZenInv - Upgrade Service: mw_dba->alter
    tablespace cim3 default storage (maxextents unlimited) is being
    executed
    [12/8/05 14:59:59.752] ZenInv - Upgrade Service: Could not do the
    operation. details:ORA-25143: default storage clause is not compatible
    with allocation policy
    ErrorCode:25143
    [12/8/05 14:59:59.752] ZenInv - Upgrade Service: Totaltime taken for
    Schema Migration : 38
    [12/8/05 14:59:59.851] ZENInv - Storer: cim.isClosed =true
    [12/8/05 14:59:59.937] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 14:59:59.970] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.133] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.277] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.719] ZenInv - Upgrade Service: migrateDbToMidas(true)
    failed in AlterProcedure
    [12/8/05 15:00:00.728] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.761] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.795] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.964] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:00.966] ZenInv - Upgrade Service: Upgrade database to
    ZENworks 6.5 failed. Reinstall ZENworks 6.5 and start the inventory
    server again
    [12/8/05 15:00:00.967] ZENInv - Status Reporting: Messages are written
    into XML file for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:01.000] ZENInv - Status Reporting: Number of records to
    add are: 1 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:01.038] ZENInv - Status Reporting: Adding record 0 for
    DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:01.188] ZENInv - Status Reporting: Number of modified
    records are: 0 for DN=CN=DAN02_ZenInvService.OU=DAN.O=CHISHOLM
    [12/8/05 15:00:01.237] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables, Running EnumModifier in integrated mode with
    upgradeService
    [12/8/05 15:00:01.237] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables, dbDriver:oracle.jdbc.driver.OracleDriver
    [12/8/05 15:00:01.237] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables,
    dbURL:jdbc:oracle:thin:@ora01a.chisholm.vic.edu.au :1522:dw
    [12/8/05 15:00:01.239] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables, dbType:1
    [12/8/05 15:00:01.239] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables, dbUser:MW_DBA
    [12/8/05 15:00:01.239] ZenInv - Upgrade Service: [EnumModifier]
    modifyEnumTables, dbPasswd:novell
    [12/8/05 15:00:01.248] ZenInv - Upgrade Service: [EnumModifier] Product
    Version: ZFD 4
    [12/8/05 15:00:01.298] ZenInv - Upgrade Service: [EnumModifier]
    checkConditions(), Executing check 'Statements':5
    [12/8/05 15:00:01.298] ZenInv - Upgrade Service: [EnumModifier]
    DatabaseEnumsModifier::checkResult(), query: insert into
    cim.upgrademethod_en_US Values (13, 'Socket 423')
    [12/8/05 15:00:01.311] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, errorcode:1
    [12/8/05 15:00:01.311] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, Non-fatal error:1
    [12/8/05 15:00:01.312] ZenInv - Upgrade Service: [EnumModifier]
    DatabaseEnumsModifier::checkResult(), query: insert into
    cim.upgrademethod_en_US Values (14, 'Socket A (Socket 462)')
    [12/8/05 15:00:01.316] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, errorcode:1
    [12/8/05 15:00:01.316] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, Non-fatal error:1
    [12/8/05 15:00:01.317] ZenInv - Upgrade Service: [EnumModifier]
    DatabaseEnumsModifier::checkResult(), query: insert into
    cim.upgrademethod_en_US Values (15, 'Socket 478')
    [12/8/05 15:00:01.321] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, errorcode:1
    [12/8/05 15:00:01.322] ZenInv - Upgrade Service: [EnumModifier]
    checkResult, Non-fatal error:1
    [12/8/05 15:00:01.322] ZenInv - Upgrade Service: [EnumModifier] Da

    ok putting inv into "ALL" debug mode, it told me it was running
    init_sl_0.sql
    I found this file contained in migrate.zip under the wminv\properties
    directory.
    So I extracted the migrate.zip file, edited init_sl_0.sql and removed
    the alter tablespace lines. Why they're wanting to do alter tablespace
    I have no idea - applications shouldn't expect to be able to do that...
    it's up to the organisation and the DBA to manage the tablespaces, not
    the application.
    I then re-zipped all the files using Winzip and named the file
    migrate.zip and placed it in the properties directory.
    Now, when the upgrade runs I get a java exception when it tries to
    unzip migrate.zip and the upgrade bombs.
    I also tried just opening the migrate.zip file with Winzip and directly
    editing the file from within Winzip, which does allow for this. However
    it appears the files inside the archive are flagged read-only, so
    Winzip can't rewrite the edited file back into the same folder location
    within the archive.
    Here's the java exception:
    ***12/21/05 16:24:40.789] ZenInv - Upgrade Service: Unzipping the file
    migrate.zip
    [12/21/05 16:24:40.827] ZenInv - Upgrade Service:
    [12/21/05 16:24:40.827] ZenInv - Upgrade Service: Exception while
    unzipping the migration zip file
    java.lang.NullPointerException
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.MigUnzip.unZipFile(AlterProcedure.java:4557)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.MigUnzip.unzip(AlterProcedure.java:4522)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.AlterProcedure.unzipMigrateZipFile(AlterProcedu re.java:1983)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.AlterProcedure.unzipMigrateZipFile(AlterProcedu re.java:1975)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.UpgradeServiceController.aInitMigration(Upgrade ServiceController.java:218)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Nativ e Method)
    at
    sun.reflect.NativeMethodAccessorImpl.invoke(Native MethodAccessorImpl.java:39)
    at
    sun.reflect.DelegatingMethodAccessorImpl.invoke(De legatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:324)
    at
    com.novell.zenworks.desktop.inventory.senderreceiv er.statemachine.StateMachine.execute(StateMachine. java:68)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.UpgradeServiceController.init(UpgradeServiceCon troller.java:113)
    at
    com.novell.zenworks.desktop.inventory.upgradeServi ce.UpgradeServiceThread.run(UpgradeServiceThread.j ava:81)
    [12/21/05 16:24:40.828] ZenInv - Upgrade Service: unzipMigrateZipFile()
    failed in AlterProcedure
    [12/21/05 16:24:40.829] ZenInv - Upgrade Service: in aLogError() method
    [12/21/05 16:24:40.830] ZenInv - Upgrade Service: Upgrade database to
    ZENworks 6.5 failed. Reinstall ZENworks 6.5 and start the inventory
    server again
    Ok.. what am I doing wrong ? :) It shouldn't be this hard....
    cheers
    Dave

  • Oracle 11.2.0.3.0 Database Upgrade changes DATA_PUMP_DIR

    Along with existing RMAN backups we do Exports - of our DB using and OS User and Oracle Wallet.
    Of the DB's we have upgraded the Data Pump Directory
    Select * from dba_directories; (there are other commands to get this info as well).
    I captured screens from the DBUA upgrades, but did not see an option to change this information.
    Is there a way to feed this information to the install moving forward. IE, ./DBUA -silent ?
    Also, anyone tracked the percentage of storage increase from 10.2/11.1 to 11.2.
    All the ones we did grew, so be aware.
    Thanks in advance!

    Are you stating that the upgrade process is changing or resetting data in DBA_DIRECTORIES ? It should not - if you have documented this experience, pl report it to Support - it may be a known bug or a new one.
    The increase in size of the database should only be restricted to SYSTEM and SYSAUX tablespaces (and possibly TEMP and UNDO), but user tablespaces should see no increase.
    HTH
    Srini

  • Is this possible in Oracle Data Gaurd 11g?

    Dear All,
    I was looking for replication scenerio while this thing came in my mind.
    We have setup 2 database servers on 2 different location, thousands of miles apart. Both servers have oracle 11gR1 and are on linux operating system.
    We have been facing connectivity problems as users are accessing our application connected to one of these servers from all over the world and when there is connectivity problems applications are no more accessible.
    I want to setup a stand by database server, like if there is connectivity problem on one server, or its down due to any reason applications are automatically shift to the second available server. The primary server is obvioulsy the first server and in case of no connectivity or any other problem that should bein use.
    In this case we have to keep an exact copy of the primary server on second server. I want to know if Oracle data Gaurd will provide us such a functionality when the servers are on remote locations? How the data will be copied to secondry server?
    Kindly let me know if there is any document available to setup such a scenerio. I will be thankful to you.
    Regards,
    Imran Baig

    DataGuard Physical standby will keep a copy of your database on a standby server. Utilizing fast start failover with an observer (3rd party) that sees both the primary and standby server will give you the capability to automatically failover to the standby server should connectivity to the primary be interrupted. Fast start failover is configured with a threshold that will determine the amount of time it waits for lost connectivity prior to actually doing the failover allowing you to configure the threshold prior to failover. Your application accessing the database will need to reconnect to the new primary after failover.

  • Database Upgrade 10.1.0.4 to 10.2.0.4

    Newbie here.
    So I am starting to use OEM on Windows XP workstation and have upgraded it to 10.2.0.4 and also wanted to upgrade the repository to 10.2.0.4 as well.
    So this is what I have done so far.
    I have installed 10.2.0.1 into a new home. I have patch it to 10.2.0.4. I am using DBUA to upgrade the database and have set my ORACLE_HOME to the new home and started DBUA. Everything is working just fine until the new OracleService is created and is trying to start. I received a message that it is unable to start service. If i click on continue I received the ORA-12560 TNS error. I have tried multiple times to get this to work and it always fails on this part..
    Also to let you know that when I connect to SQLPlus before upgrade I have to connect using sysplus sysman/xxxxx@emrep with out emrep it won't connect. I have also set TNS_ADMIN to the original homes network/admin directory.
    I have research the documentation on upgrading, but maybe I am missing a step or someone has run into the same issue.
    Any help you can lend me would be appreciated.

    The database files won't be copied over. As general practice, databases aren't stored under Oracle Homes. You could always fix that by moving the database, backup the control file to trace, and edit it with new file locations. I wouldn't worry about that right now.
    Did you run the prerequisite script before trying to upgrade?
    <10.2 HOME>/rdbms/admin/utlu102i.sql
    Once a database is opened upgrade, you won't be able to open it with the old binaries.
    Was the listener up when you tried to connect to it with 10.1? Were there any errors?
    The manual process of upgrading the database is to export the 10.2 env, open the database upgrade, upgrade the data dictionary <10.2HOME>/rdbms/admin/catupgrd.sql, startup the database normally, and Compile all the invalid objects <10.2HOME>/rdbms/admin/utlrp.sql. I would make sure to go over the sql that executes prerequisite checks and read through the upgrade docs. I would also take a backup before doing anything.

  • Database Upgrade Agent not working for upgrading Oracle 9i DB to Oracle 10i

    Hi All,
    Presently I have completed my Oracle 10i installation on Solaris Server but I am facing an issue during the upgrade process of Oracle 9i to 10g. Whenever I try to upgrade the 9i databases to 10g using DBUA(Database Upgrade assistant) I am facing an issue that Data base not running from oracle home or pfile not found. The exact error message says:-
    The upgrade assistant failed in bringing up the database XIPCOTS. Oracle Home /opt/apps/home/oracle9i/product/releases/92/92 obtained from file /var/opt/oracle/oratab was used to connect to the database. Either the database is not running from Oracle Home /opt/apps/home/oracle9i/product/releases/92/92 or the correct Initialisation parameter file (pfile) was not found. To start the database , Oracle needs the pfile. Provide the full path name of the init file for the database XIPCOTS.
    In reality the XIPCOTS database is running fine in Oracle 9i and we can connect it using PL/SQL developer in windows.I have also provided the path of pfile but it is still not accepting the same.
    Path of pfile I provided :- /XIP/ofaroot/XIPCOTS/pfile/initXIPCOTS.ora
    Any help or suggestions would be greately appreciated.
    Our both Oracle 9i and 10g are installed on same Solaris server where solaris version is 9
    Installation path for 9i :- /opt/apps/home/oracle9i/product/releases/92/92
    Installation path for 10g :- /opt/apps/home/oracle/product/releases/10201
    Oracle 9i version :- 9.2.0.7
    Oracle 10g version :- 10.2.01
    Thanks in advance.
    Best Regards,
    Gourav Banerjee

    Can you please mention the specific folder name under oracle home where we need to put this pfile because there are two folder dbs and ds I have allready tried putting it inside dbs its not working. And my home path is correct I hope and is set for oracle 9i directories in the .oraenv file do we need to set it for oracle 10g before the DBUA upgrade.

  • Database Upgrade to 10.2.0.4

    Using Windows 2003 64-bit, 10.2.0.3 Standard Edition.
    I clone the database to a spare server, shut it down immediate, and stopped all Oracle services. I installed the 10.2.0.4 patchset into the original Oracle Home without a problem. I then ran the Database Upgrade Assistant and it ran through to completion. However, it did declare a series of errors when upgrading the 'Data Mining' component:
    Error ignored: ORA-29831: operator binding not found ORA-06553: PLS-306: wrong number or types of arguments in call to 'SDO_FILTER'
    ...and others like it. The report at the end of the process is saying the Oracle Server component is 'Successful *' -and the asterisk links to a list of the errors that I clicked 'ignore' to during the process. The one for 'data mining' says 'Skipped'.
    I don't think I mind the errors about Data Mining, because as far as I know, we don't use that (are we even allowed to on an SE license?) When I look in V$OPTION, for example, we get:
    OLAP Window Functions TRUE
    Block Media Recovery FALSE
    Fine-grained Auditing FALSE
    Application Role FALSE
    Enterprise User Security FALSE
    Oracle Data Guard FALSE
    Oracle Label Security FALSE
    OLAP FALSE
    Table compression FALSE
    Join index FALSE
    Trial Recovery FALSE
    Data Mining                                                      FALSE
    Online Redefinition FALSE
    Streams Capture FALSE
    File Mapping FALSE
    So my question is: did the upgrade work properly or not?! Should I mind that the errors appeared when upgrading a component that doesn't even exist, as far as I know? Is there some way of knowing that all is as it should be?
    I have logged onto the database after the upgrade and all looks fine, with queries working as expected and so on. But I just have a nagging doubt...

    SQL> SELECT owner, object_type, COUNT(*)
      2  FROM dba_objects
      3  WHERE status = 'INVALID'
      4  GROUP BY owner, object_type;
    OWNER                          OBJECT_TYPE           COUNT(*)
    MDSYS                          PACKAGE BODY                 1
    ACC_GLOBAL                     PACKAGE BODY                 1The ACC_GLOBAL invalid is a known issue: I don't have the tnsnames.ora with the right alias for the resolution of a database link that package contains. MDSYS is not a schema we use, I think, and the package that's invalid is SDO_3GL which is the package I expected to be wrong as a result of the error messages received during the upgrade.
    As for your second query:
    COMP_NAME                                VERSION                        STATUS
    Oracle interMedia                        10.2.0.4.0                     VALID
    Oracle XML Database                      10.2.0.4.0                     VALID
    Oracle Text                              10.2.0.4.0                     VALID
    Oracle Data Mining                       10.2.0.3.0                     OPTION OFF
    Oracle Expression Filter                 10.2.0.4.0                     VALID
    Oracle Rule Manager                      10.2.0.4.0                     VALID
    Oracle Workspace Manager                 10.2.0.4.3                     VALID
    Oracle Database Catalog Views            10.2.0.4.0                     VALID
    Oracle Database Packages and Types       10.2.0.4.0                     VALID
    JServer JAVA Virtual Machine             10.2.0.4.0                     VALID
    Oracle XDK                               10.2.0.4.0                     VALID
    Oracle Database Java Packages            10.2.0.4.0                     VALIDDoesn't look a miserable failure to me, really. Data Mining doesn't look right, but we knew that already because of those error messages during the upgrade. Fact remains, we don't use it, it's not an option we've installed, etc etc.

  • Impact of database upgrade in Bw delta queue

    Hi Gurus,
    We are going to upgrade the R/3 oracle database. I suppose that it have an impact in the BW delta queue, and that before the upgrade the logistics queues should be uploaded to BW.
    I'm right? exits some note or some checklist about wihch activities should be performed in BW due to the database upgrade?
    Thanks in advance for your help.

    Hi,
    This speaks about support pack upgrade. But i think this also applies for DB upgrade too.
    Its always better to drain the delta queues before an upgrade.
    As a standard practice we drain the delta queues by running the IP/ chain multiple times.
    As a prerequiste we cancel/reschedule the V3 jobs to a future date during this activity.
    The V3 extraction delta queues must be emptied prior to the upgrade to avoid any possible data loss.
    V3 collector jobs should be suspended for the duration of the upgrade.
    They can be rescheduled after re-activation of the source systems upon completion of the upgrade.
    See SAP Notes 506694 and 658992 for more details.
    Page 17
    Load and Empty all Data mart Delta Queues in SAP BW. (e.g. for all export DataSources)
    The SAP BW Service SAPI, which is used for internal and ‘BW to BW’ data mart extraction, is
    upgraded during the SAP BW upgrade. Therefore, the delta queues must be emptied prior to the
    upgrade to avoid any possibility of data loss.
    upgrade preparation and postupgrade checklist
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/472443f2-0c01-0010-20ab-fbd380d45881
    /message/3221895#3221895
    OSS notes 328181 and 762951 as a prerequisites.
    Failure to follow the instructions in those notes may probably result in data loss.
    https://websmp207.sap-ag.de/~form/sapnet?_FRAME=CONTAINER&_OBJECT=011000358700002662832005E
    /thread/804820
    Effect on BW of R/3 Upgrade   
    How To Tackle Upgrades to SAP ERP 6.0
    /people/community.user/blog/2008/03/20/how-to-tackle-upgrades-to-sap-erp-60
    thanks,
    JituK

  • Migrate SQL 2008 Analysis database to 2012 AS database along with data level security defined in current production cube

    I want to migrate Analysis Services 2008 database to 2012 AS database along with data level security defined in current production cube
    Note: Only Production environment have security, while no security is defined in development environment
    Potential Approach:
    1 - Using Synchronization Wizard: Gives me error : "The OLAP element at line1 can not appear under envelope......" and this is because Synchrinzation works only for same version
    and in my case, there are different versions of SQL (SQL 2008 and 2012)
    2 - Using Visual studio conversion wizard - Convert SQL 2008 AS project to 2012 and then process cube, so I can get the cube working but then how can I get data level security since 100's of data level security is defined in production Cube, so how can I
    migrate that across
    3 - Script out XMLA and deploy cube - But then again having issues with how can i script SSAS security
    4 - Would taking backup of SSAS 2008 database and restore to SSAS 2012 will help ?
    Any suggestions would be appreciated
    Thanks,
    Mihir

    Hi Mihir,
    According to your description, you want to migrate the SQL Server Analysis Services (SSAS) 2008 database which have some security setting with it to SSAS 2012, right? We can migrate existing Analysis Services databases either during Setup, by upgrading an
    existing instance of Analysis Services, or after Setup, by running the Migration Wizard. Generally, when migrating a database to another server, all the setting will be migrated. So in your scenario, you can refer to the steps on the links below to migrate
    your SSAS database.
    How to: Migrate Analysis Services Databases
    Migrating Existing Analysis Services Databases
    Regards,
    Charlie Liao
    TechNet Community Support

Maybe you are looking for

  • Remove Header and Footer from a report

    Hi, I use the report S_ALR_87012104. But I want to export this report removing the header and footer. Is it possible to export the report excluding header and footer. That is line items only. Regards Mahbub

  • Parsing Excel's CSV file in PL/SQL

    Someone may need this... This procedure parses a line of Excel's CSV file into a PL/SQL table's elements. FUNCTION func_split_csv (p_string IN VARCHAR2) RETURN gt_tbltyp_strings IS * NAME: func_split_csv * PURPOSE: Split the passed Excel's CSV string

  • Why does the stage layout change when native menu added

    I have a program that was fine till I added a native menu to it. After adding the menu I have a space on both the left and the right of the stage. If I draw a shape at x=0, y=0 it appears about 40 pixels to the right of the window border. If I remove

  • Icloud quit working; cannot load application

    Icloud has been working fine on my PC until recently.  Now I cannot get to contacts or mail.  Instead I get a message that says "Can't load calendar.  There was a problem loading the application." The full icloud trouble report (which I forwarded sev

  • No surround (5.1) sound on x-fi mu

    +++ This message was also posted on speaker systems, but because it may be also relate to this soundcard, I post it too +++?Hi, I have new x-fi music soundcard and my old Cambridge Soundworks DTT2500 System. When I set the speakermode of the x-fi to