Downgrading Oracle database 12.1.0 to 11R2
Hi everyone,
I'm very new to the oracle products and databases as well, so obviuosly my question might look like a bit simple for someone, but I do really need an answer to it.
I'm using OTN Developer Day VM, which comes with the preinstalled Oracle DB 12cR1. I'm wondering if its possible to dowgrade it to 11R2? Maybe in order to do the downgrade it must have been previously upgraded from 11R2 to 12cR1 ?
Regards
I'm very new to the oracle products and databases as well, so obviuosly my question might look like a bit simple for someone, but I do really need an answer to it.
I'm using OTN Developer Day VM, which comes with the preinstalled Oracle DB 12cR1. I'm wondering if its possible to dowgrade it to 11R2? Maybe in order to do the downgrade it must have been previously upgraded from 11R2 to 12cR1 ?
Your 'guess' is correct. You can only 'downgrade' IF you have previously 'upgraded' AND you have not modified the COMPATIBLE parameter.
See The Database Upgrade Guide
Downgrading Oracle Database to an Earlier Release
For supported releases of Oracle Database, you can downgrade a database to the release from which you last upgraded. For example, if you recently upgraded from release 11.2.0.4 to Oracle Database 12c, and you did not change the compatible initialization parameter to 12.1 or higher, then you can downgrade to release 11.2.0.4. If your Oracle Database 12c is release 12.1.0.2, and you did not change the compatible initialization parameter to 12.1.0.2, then you can downgrade to release 12.1.0.1 and so forth.
Similar Messages
-
Hi All,
I am working as development Dbs. I had upgraded Oracle DB 9.2.0.1.0 to version 9.2.0.8.0 for some testing purpose in Windows XP(32bit). I didnt take the backup before applying patch. Now they want to downgrade the db to 9.2.0.6.0. How can i do that? Is there any patch for that? or Is there any patch for downgrading to 9.2.0.1.0? Please let me know a solution.
Thanks,
ManikandanI meant if the COMPATIBLE parameter has been set to something greater than 9.2.0.1 or 9.2.0.6 (depending on which you want to downgrade to). If it is still 9.2.0.0 that is fine.
Check the patch documentation for rolling back the 9.2.0.8 patch. -
Hi,
I would like to test downgrade process through the scripts (@catdwgrd.sql). Oracle documentation http://docs.oracle.com/cd/E18283_01/server.112/e17222/downgrade.htm
says
"2.If you previously installed a recent version of the time zone file and used the DBMS_DST PL/SQL package to upgrade TIMESTAMP WITH TIME ZONE data to that version, then you must install the same version of the time zone file in the release to which you are downgrading. For example, the latest time zone files that are supplied with Oracle Database 11g Release 2 (11.2) are version 14 as of this printing. If after the database upgrade, you had used DBMS_DST to upgrade the TIMESTAMP WITH TIME ZONE data to version 14, then install the version 14 time zone file in the release to which you are downgrading. This ensures that your TIMESTAMP WITH TIME ZONE data is not logically corrupted during retrieval. To find which version your database is using, query V$TIMEZONE_FILE"
1) In my case, I used DBMS_DST and now the version if the time zone file is 14. If I have to go back to 10.2.0.4, Should i copy the files to $ORACLE_HOME/ORACORE/ZONEINFO (where $ORACLE_HOME is 10.2.0.4)?
2) If my Oracle 10.2.0.4 is shared with multiple instances, will i introduce risk for other databases running on the same oracle home when I copy the files?
3) What other steps should I do other than copying the version 14 files into this directory?
Thank You
Sarayuuser13312943 wrote:
Hi,
I would like to test downgrade process through the scripts (@catdwgrd.sql). Oracle documentation http://docs.oracle.com/cd/E18283_01/server.112/e17222/downgrade.htm
says
"2.If you previously installed a recent version of the time zone file and used the DBMS_DST PL/SQL package to upgrade TIMESTAMP WITH TIME ZONE data to that version, then you must install the same version of the time zone file in the release to which you are downgrading. For example, the latest time zone files that are supplied with Oracle Database 11g Release 2 (11.2) are version 14 as of this printing. If after the database upgrade, you had used DBMS_DST to upgrade the TIMESTAMP WITH TIME ZONE data to version 14, then install the version 14 time zone file in the release to which you are downgrading. This ensures that your TIMESTAMP WITH TIME ZONE data is not logically corrupted during retrieval. To find which version your database is using, query V$TIMEZONE_FILE"
1) In my case, I used DBMS_DST and now the version if the time zone file is 14. If I have to go back to 10.2.0.4, Should i copy the files to $ORACLE_HOME/ORACORE/ZONEINFO (where $ORACLE_HOME is 10.2.0.4)?
No - you should not copy files from the 11gR2 HOME to the 10.2.0.4 HOME - instead apply patch 9742718 to the 10.2.0.4 HOME (which corresponds to v14 of the DST patch)
Updated DST transitions and new Time Zones in Oracle Time Zone File patches [ID 412160.1]
2) If my Oracle 10.2.0.4 is shared with multiple instances, will i introduce risk for other databases running on the same oracle home when I copy the files?
You will need to follow steps in 9742718 for all of the databases running out of the 10.2.0.4 HOME
3) What other steps should I do other than copying the version 14 files into this directory?
See above
Thank You
SarayuHTH
Srini -
ORA-600 [17069] error while running catrelod.sql to downgrade 11g database
Hi,
We are downgrading our 11.2.0.2 database to 10.2.0.4. We have successfullly run catdwgrd.sql without any errors in 11g env. While running catrelod.sql in 10g env we are facing the following error:
SQL> @?/rdbms/admin/catrelod.sql
TIMESTAMP
COMP_TIMESTAMP RELOD__BGN 2013-04-24 20:15:39 2456407 72939
DOC>#######################################################################
DOC>#######################################################################
DOC> The following statement will cause an "ORA-01722: invalid number"
DOC> error if the database server version is not 10.0.0.
DOC> Shutdown ABORT and use a different script or a different server.
DOC>#######################################################################
DOC>#######################################################################
DOC>#
no rows selected
DOC>#######################################################################
DOC>#######################################################################
DOC> The following statement will cause an "ORA-01722: invalid number"
DOC> error if the database has not been opened for MIGRATE.
DOC>
DOC> Perform a "SHUTDOWN ABORT" and
DOC> restart using MIGRATE.
DOC>#######################################################################
DOC>#######################################################################
DOC>#
no rows selected
Session altered.
Session altered.
no rows selected
DECLARE
ERROR at line 1:
ORA-00600: internal error code, arguments: [17069], [0x170FB8CA8], [], [], [],
[], [], []I have checked the trace file in udump but it dint give any readable information.
Please find below environment details.
OS : RHEL 5 64 bit
11g Database : 11.2.0.2
10g Database : 10.2.0.4
Kindly assist.KR10822864 wrote:
Whenever an ORA-600 error is raised a trace file is generated and an entry written to the alert.log with details of the trace file location. Starting with Oracle Database 11g Release 1, the diagnosability infrastructure was introduced which places the trace and core files into a location controlled by the DIAGNOSTIC_DEST initialization parameter when an incident, such as an ORA-600 is created. For earlier versions, the trace file will be written to either USER_DUMP_DEST (if the error was caught in a user process) or BACKGROUND_DUMP_DEST (if the error was caught in a background process like PMON or SMON). The trace file contains vital information about what led to the error condition
please post 40 lines of alert log info .
MOS Note:ORA-600 [17069] "Failed to pin a library cache object after 50 attempts" [ID 39616.1]
"Look in the trace file for the text 'LIBRARY OBJECT HANDLE: handle=170fb8ca8'"
if not found any massages like above please raise [email protected].
Hi KR,
Thanks for your help. As suggested I have checked the trace file to search and found the below info. I suppose it is some kind of lock but not sure about it. Please let me know if the following makes sense:
SO: 0x21158a6e0, type: 3, owner: 0x211006f28, flag: INIT/-/-/0x00
(call) sess: cur 211572570, rec 211572570, usr 211572570; depth: 0
SO: 0x21158a9b8, type: 3, owner: 0x21158a6e0, flag: INIT/-/-/0x00
(call) sess: cur 211572570, rec 0, usr 211572570; depth: 1
SO: 0x1727a2618, type: 54, owner: 0x21158a9b8, flag: INIT/-/-/0x00
LIBRARY OBJECT PIN: pin=1727a2618 handle=170fb8ca8 mode=S lock=171243e98
user=211572570 session=211572570 count=1 mask=001d savepoint=0x43 flags=[00]
SO: 0x171243e98, type: 53, owner: 0x21158a9b8, flag: INIT/-/-/0x00
LIBRARY OBJECT LOCK: lock=171243e98 handle=170fb8ca8 mode=S
call pin=0x1727a2618 session pin=(nil) hpc=0000 hlc=0000
htl=0x171243f18[0x17278a310,0x17278a310] htb=0x17278a310 ssga=0x172789928
user=211572570 session=211572570 count=1 flags=PNC/[0400] savepoint=0x43
LIBRARY OBJECT HANDLE: handle=170fb8ca8 mtx=0x170fb8dd8(0) cdp=0
name=SYS.STANDARD
hash=51570e225ed8a9a803b7318f191e0a8d timestamp=04-18-2006 00:00:00
namespace=TABL flags=KGHP/TIM/SML/[02000000]
kkkk-dddd-llll=0000-001d-001d lock=S pin=S latch#=3 hpc=0004 hlc=0004
lwt=0x170fb8d50[0x170fb8d50,0x170fb8d50] ltm=0x170fb8d60[0x170fb8d60,0x170fb8d60]
pwt=0x170fb8d18[0x170fb8d18,0x170fb8d18] ptm=0x170fb8d28[0x170fb8d28,0x170fb8d28]
ref=0x170fb8d80[0x170fb8d80,0x170fb8d80] lnd=0x170fb8d98[0x170f05b50,0x170fa6858]
LIBRARY OBJECT: object=170e82de8
type=PCKG flags=EXS/LOC[0005] pflags=NST/IVR[0201] status=VALD load=0
DATA BLOCKS:
data# heap pointer status pins change whr
0 170e831f8 170e82fc0 I/P/A/-/- 0 NONE 00
SO: 0x21158e470, type: 5, owner: 0x21158a6e0, flag: INIT/-/-/0x00
(enqueue) CU-70E88BC0-00000001 DID: 0001-001F-00000004
lv: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 res_flag: 0x2
res: 0x17677adc8, mode: X, lock_flag: 0x0
own: 0x211572570, sess: 0x211572570, proc: 0x211006f28, prv: 0x17677add8
SO: 0x177f82c28, type: 59, owner: 0x21158a6e0, flag: INIT/-/-/0x00
cursor enqueue
child: 170e89348, flag: 53, number: 0
parent: 170e89738 -
Cloning Oracle database 8.1.7.0
Hi,
We have upgraded our database from Oracle 8.1.7.0 to Oracle 8.1.7.4 , applied terminal patch set(Oracle 8.1.7.4) , as we have to upgrade it further to 10g rel2.
After upgrading to 8.1.7.4, we required to apply the database fixes against the mentioned bugs.
One of the Serious Server Issues is mentioned below after upgradation,
Changing the Database Character Set can corrupt LOBs ( Note 118242.1)
we have gone through the note mentioned above,
In that one of the point regarding cloning is as below :
If the database version is 8.1.7.3 or 8.1.7.4:
- install non-patched Oracle 8.1.7.0.0 from the original CD-Pack
into a separate ORACLE_HOME
- take a full backup of the database
- create a clone database from this backup using the 8.1.7.0.0
version of the RDBMS software -- contact Oracle Support
if you need help; the clone should open with the older
non-patched software without any downgrade steps
- export all non-system tables with non-NULL CLOB and/or NCLOB columns
from the clone; do not forget to set NLS_LANG environment
variable to the database character set for the export session;
you can use the 8.1.7.0.0 version of the Export utility
- truncate the exported tables but in the original database
- reimport the exported tables into the original database;
do not forget to set NLS_LANG environment variable to
the database character set for the import session;
use the Import utility from the original ORACLE_HOME
- drop the clone database - do not use the clone for other
purposes as it is not supported to run a patched database
with the non-patched software.
How to Clone the Oracle database (8.1.7.0) on windows OS ?
From where i would get the Cloning steps to accomplish the above steps on Windows OS ?
With RegardsHi,
Please, refer to Steve Rea's Web for DB Cloning examples:
For Cloning using Datafile copy:
http://uaex.edu/srea/Cloning_a_Database_using_Datafile_Copy.htm
For Cloning using Export/Import:
http://uaex.edu/srea/Cloning_a_Database_using_Export_Import.htm
Cheers:
Francisco Munoz Alvarez
www.oraclenz.com -
Looking for the best solution for downgrading a database from 11.2 to 11.1
we have an install of 11.2 grid and have made several attempts at adding 11.1 binaries and using the 11.2 infrastructure. we patched, etc... but all of our attempts failed in getting 11.1 to work across the cluster.
we have decided to make all of the installs version 11.1.
I am looking for the most straight forward approach to migrate our 11.2 down to 11.1 with the least amount of downtime.
thanks, tomHi Tom;
Please check below note
Master Note For Oracle Database Downgrade [ID 1151427.1]
How To Downgrade From Database 11.2 To Previous Release [ID 883335.1]
I belive it will answer all your question.
Please also check below links:
http://blogs.oracle.com/db/2010/09/master_note_for_oracle_database_downgrade_doc_id_11514271.html
Regard
Helios -
Downgrade oracle version 11g to 10g
Hi
while connecting from 11g client to 8i it throwing error version not supported
how to downgrade oracle 11g client to 10 g client.
Thanks
RegardsHi Mirza,
thanks for ur reply, but is there any way to downgrade the oracle client from 11g to 10gon server there more database so that we cannot unistall.
I believe you Install 11g client software on client machine, if its so you can safely uninstall it and Install 10g again.
If this 11g Installation configured on server with a database its a different scenario.
You question is not clear. Please elaborate it...
Also i dont think there is no such option for downgrading client.
thanks,
X A H E E R -
How to load a java script in oracle database
Is it possible to load d java script in Oracle database. while the object type is
java resource ..RENUJP wrote:
I meant to load a javascript to oracle database not to oracle appliocation.
like loadjava....
I can load a javascript to oracle database.. but i can't call it...Please re-read the comments above, especially the part about this not being a javascript nor oracle forum. Exactly what part about this information don't you understand? -
Query in timesten taking more time than query in oracle database
Hi,
Can anyone please explain me why query in timesten taking more time
than query in oracle database.
I am mentioning in detail what are my settings and what have I done
step by step.........
1.This is the table I created in Oracle datababase
(Oracle Database 10g Enterprise Edition Release 10.2.0.1.0)...
CREATE TABLE student (
id NUMBER(9) primary keY ,
first_name VARCHAR2(10),
last_name VARCHAR2(10)
2.THIS IS THE ANONYMOUS BLOCK I USE TO
POPULATE THE STUDENT TABLE(TOTAL 2599999 ROWS)...
declare
firstname varchar2(12);
lastname varchar2(12);
catt number(9);
begin
for cntr in 1..2599999 loop
firstname:=(cntr+8)||'f';
lastname:=(cntr+2)||'l';
if cntr like '%9999' then
dbms_output.put_line(cntr);
end if;
insert into student values(cntr,firstname, lastname);
end loop;
end;
3. MY DSN IS SET THE FOLLWING WAY..
DATA STORE PATH- G:\dipesh3repo\db
LOG DIRECTORY- G:\dipesh3repo\log
PERM DATA SIZE-1000
TEMP DATA SIZE-1000
MY TIMESTEN VERSION-
C:\Documents and Settings\dipesh>ttversion
TimesTen Release 7.0.3.0.0 (32 bit NT) (tt70_32:17000) 2007-09-19T16:04:16Z
Instance admin: dipesh
Instance home directory: G:\TimestTen\TT70_32
Daemon home directory: G:\TimestTen\TT70_32\srv\info
THEN I CONNECT TO THE TIMESTEN DATABASE
C:\Documents and Settings\dipesh> ttisql
command>connect "dsn=dipesh3;oraclepwd=tiger";
4. THEN I START THE AGENT
call ttCacheUidPwdSet('SCOTT','TIGER');
Command> CALL ttCacheStart();
5.THEN I CREATE THE READ ONLY CACHE GROUP AND LOAD IT
create readonly cache group rc_student autorefresh
interval 5 seconds from student
(id int not null primary key, first_name varchar2(10), last_name varchar2(10));
load cache group rc_student commit every 100 rows;
6.NOW I CAN ACCESS THE TABLES FROM TIMESTEN AND PERFORM THE QUERY
I SET THE TIMING..
command>TIMING 1;
consider this query now..
Command> select * from student where first_name='2155666f';
< 2155658, 2155666f, 2155660l >
1 row found.
Execution time (SQLExecute + Fetch Loop) = 0.668822 seconds.
another query-
Command> SELECT * FROM STUDENTS WHERE FIRST_NAME='2340009f';
2206: Table SCOTT.STUDENTS not found
Execution time (SQLPrepare) = 0.074964 seconds.
The command failed.
Command> SELECT * FROM STUDENT where first_name='2093434f';
< 2093426, 2093434f, 2093428l >
1 row found.
Execution time (SQLExecute + Fetch Loop) = 0.585897 seconds.
Command>
7.NOW I PERFORM THE SIMILAR QUERIES FROM SQLPLUS...
SQL> SELECT * FROM STUDENT WHERE FIRST_NAME='1498671f';
ID FIRST_NAME LAST_NAME
1498663 1498671f 1498665l
Elapsed: 00:00:00.15
Can anyone please explain me why query in timesten taking more time
that query in oracle database.
Message was edited by: Dipesh Majumdar
user542575
Message was edited by:
user542575TimesTen
Hardware: Windows Server 2003 R2 Enterprise x64; 8 x Dual-core AMD 8216 2.41GHz processors; 32 GB RAM
Version: 7.0.4.0.0 64 bit
Schema:
create usermanaged cache group factCache from
MV_US_DATAMART
ORDER_DATE DATE,
IF_SYSTEM VARCHAR2(32) NOT NULL,
GROUPING_ID TT_BIGINT,
TIME_DIM_ID TT_INTEGER NOT NULL,
BUSINESS_DIM_ID TT_INTEGER NOT NULL,
ACCOUNT_DIM_ID TT_INTEGER NOT NULL,
ORDERTYPE_DIM_ID TT_INTEGER NOT NULL,
INSTR_DIM_ID TT_INTEGER NOT NULL,
EXECUTION_DIM_ID TT_INTEGER NOT NULL,
EXEC_EXCHANGE_DIM_ID TT_INTEGER NOT NULL,
NO_ORDERS TT_BIGINT,
FILLED_QUANTITY TT_BIGINT,
CNT_FILLED_QUANTITY TT_BIGINT,
QUANTITY TT_BIGINT,
CNT_QUANTITY TT_BIGINT,
COMMISSION BINARY_FLOAT,
CNT_COMMISSION TT_BIGINT,
FILLS_NUMBER TT_BIGINT,
CNT_FILLS_NUMBER TT_BIGINT,
AGGRESSIVE_FILLS TT_BIGINT,
CNT_AGGRESSIVE_FILLS TT_BIGINT,
NOTIONAL BINARY_FLOAT,
CNT_NOTIONAL TT_BIGINT,
TOTAL_PRICE BINARY_FLOAT,
CNT_TOTAL_PRICE TT_BIGINT,
CANCELLED_ORDERS_COUNT TT_BIGINT,
CNT_CANCELLED_ORDERS_COUNT TT_BIGINT,
ROUTED_ORDERS_NO TT_BIGINT,
CNT_ROUTED_ORDERS_NO TT_BIGINT,
ROUTED_LIQUIDITY_QTY TT_BIGINT,
CNT_ROUTED_LIQUIDITY_QTY TT_BIGINT,
REMOVED_LIQUIDITY_QTY TT_BIGINT,
CNT_REMOVED_LIQUIDITY_QTY TT_BIGINT,
ADDED_LIQUIDITY_QTY TT_BIGINT,
CNT_ADDED_LIQUIDITY_QTY TT_BIGINT,
AGENT_CHARGES BINARY_FLOAT,
CNT_AGENT_CHARGES TT_BIGINT,
CLEARING_CHARGES BINARY_FLOAT,
CNT_CLEARING_CHARGES TT_BIGINT,
EXECUTION_CHARGES BINARY_FLOAT,
CNT_EXECUTION_CHARGES TT_BIGINT,
TRANSACTION_CHARGES BINARY_FLOAT,
CNT_TRANSACTION_CHARGES TT_BIGINT,
ORDER_MANAGEMENT BINARY_FLOAT,
CNT_ORDER_MANAGEMENT TT_BIGINT,
SETTLEMENT_CHARGES BINARY_FLOAT,
CNT_SETTLEMENT_CHARGES TT_BIGINT,
RECOVERED_AGENT BINARY_FLOAT,
CNT_RECOVERED_AGENT TT_BIGINT,
RECOVERED_CLEARING BINARY_FLOAT,
CNT_RECOVERED_CLEARING TT_BIGINT,
RECOVERED_EXECUTION BINARY_FLOAT,
CNT_RECOVERED_EXECUTION TT_BIGINT,
RECOVERED_TRANSACTION BINARY_FLOAT,
CNT_RECOVERED_TRANSACTION TT_BIGINT,
RECOVERED_ORD_MGT BINARY_FLOAT,
CNT_RECOVERED_ORD_MGT TT_BIGINT,
RECOVERED_SETTLEMENT BINARY_FLOAT,
CNT_RECOVERED_SETTLEMENT TT_BIGINT,
CLIENT_AGENT BINARY_FLOAT,
CNT_CLIENT_AGENT TT_BIGINT,
CLIENT_ORDER_MGT BINARY_FLOAT,
CNT_CLIENT_ORDER_MGT TT_BIGINT,
CLIENT_EXEC BINARY_FLOAT,
CNT_CLIENT_EXEC TT_BIGINT,
CLIENT_TRANS BINARY_FLOAT,
CNT_CLIENT_TRANS TT_BIGINT,
CLIENT_CLEARING BINARY_FLOAT,
CNT_CLIENT_CLEARING TT_BIGINT,
CLIENT_SETTLE BINARY_FLOAT,
CNT_CLIENT_SETTLE TT_BIGINT,
CHARGEABLE_TAXES BINARY_FLOAT,
CNT_CHARGEABLE_TAXES TT_BIGINT,
VENDOR_CHARGE BINARY_FLOAT,
CNT_VENDOR_CHARGE TT_BIGINT,
ROUTING_CHARGES BINARY_FLOAT,
CNT_ROUTING_CHARGES TT_BIGINT,
RECOVERED_ROUTING BINARY_FLOAT,
CNT_RECOVERED_ROUTING TT_BIGINT,
CLIENT_ROUTING BINARY_FLOAT,
CNT_CLIENT_ROUTING TT_BIGINT,
TICKET_CHARGES BINARY_FLOAT,
CNT_TICKET_CHARGES TT_BIGINT,
RECOVERED_TICKET_CHARGES BINARY_FLOAT,
CNT_RECOVERED_TICKET_CHARGES TT_BIGINT,
PRIMARY KEY(ORDER_DATE, TIME_DIM_ID, BUSINESS_DIM_ID, ACCOUNT_DIM_ID, ORDERTYPE_DIM_ID, INSTR_DIM_ID, EXECUTION_DIM_ID,EXEC_EXCHANGE_DIM_ID),
READONLY);
No of rows: 2228558
Config:
< CkptFrequency, 600 >
< CkptLogVolume, 0 >
< CkptRate, 0 >
< ConnectionCharacterSet, US7ASCII >
< ConnectionName, tt_us_dma >
< Connections, 64 >
< DataBaseCharacterSet, AL32UTF8 >
< DataStore, e:\andrew\datacache\usDMA >
< DurableCommits, 0 >
< GroupRestrict, <NULL> >
< LockLevel, 0 >
< LockWait, 10 >
< LogBuffSize, 65536 >
< LogDir, e:\andrew\datacache\ >
< LogFileSize, 64 >
< LogFlushMethod, 1 >
< LogPurge, 0 >
< Logging, 1 >
< MemoryLock, 0 >
< NLS_LENGTH_SEMANTICS, BYTE >
< NLS_NCHAR_CONV_EXCP, 0 >
< NLS_SORT, BINARY >
< OracleID, NYCATP1 >
< PassThrough, 0 >
< PermSize, 4000 >
< PermWarnThreshold, 90 >
< PrivateCommands, 0 >
< Preallocate, 0 >
< QueryThreshold, 0 >
< RACCallback, 0 >
< SQLQueryTimeout, 0 >
< TempSize, 514 >
< TempWarnThreshold, 90 >
< Temporary, 1 >
< TransparentLoad, 0 >
< TypeMode, 0 >
< UID, OS_OWNER >
ORACLE:
Hardware: Sunos 5.10; 24x1.8Ghz (unsure of type); 82 GB RAM
Version 10g Enterprise Edition Release 10.2.0.3.0 - 64bit Production
Schema:
CREATE MATERIALIZED VIEW OS_OWNER.MV_US_DATAMART
TABLESPACE TS_OS
PARTITION BY RANGE (ORDER_DATE)
PARTITION MV_US_DATAMART_MINVAL VALUES LESS THAN (TO_DATE(' 2007-11-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D1 VALUES LESS THAN (TO_DATE(' 2007-11-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D2 VALUES LESS THAN (TO_DATE(' 2007-11-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D3 VALUES LESS THAN (TO_DATE(' 2007-12-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D1 VALUES LESS THAN (TO_DATE(' 2007-12-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D2 VALUES LESS THAN (TO_DATE(' 2007-12-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D3 VALUES LESS THAN (TO_DATE(' 2008-01-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D1 VALUES LESS THAN (TO_DATE(' 2008-01-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D2 VALUES LESS THAN (TO_DATE(' 2008-01-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D3 VALUES LESS THAN (TO_DATE(' 2008-02-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_MAXVAL VALUES LESS THAN (MAXVALUE)
LOGGING
NOCOMPRESS
TABLESPACE TS_OS
NOCACHE
NOCOMPRESS
NOPARALLEL
BUILD DEFERRED
USING INDEX
TABLESPACE TS_OS_INDEX
REFRESH FAST ON DEMAND
WITH PRIMARY KEY
ENABLE QUERY REWRITE
AS
SELECT order_date, if_system,
GROUPING_ID (order_date,
if_system,
business_dim_id,
time_dim_id,
account_dim_id,
ordertype_dim_id,
instr_dim_id,
execution_dim_id,
exec_exchange_dim_id
) GROUPING_ID,
/* ============ DIMENSIONS ============ */
time_dim_id, business_dim_id, account_dim_id, ordertype_dim_id,
instr_dim_id, execution_dim_id, exec_exchange_dim_id,
/* ============ MEASURES ============ */
-- o.FX_RATE /* FX_RATE */,
COUNT (*) no_orders,
-- SUM(NO_ORDERS) NO_ORDERS,
-- COUNT(NO_ORDERS) CNT_NO_ORDERS,
SUM (filled_quantity) filled_quantity,
COUNT (filled_quantity) cnt_filled_quantity, SUM (quantity) quantity,
COUNT (quantity) cnt_quantity, SUM (commission) commission,
COUNT (commission) cnt_commission, SUM (fills_number) fills_number,
COUNT (fills_number) cnt_fills_number,
SUM (aggressive_fills) aggressive_fills,
COUNT (aggressive_fills) cnt_aggressive_fills,
SUM (fx_rate * filled_quantity * average_price) notional,
COUNT (fx_rate * filled_quantity * average_price) cnt_notional,
SUM (fx_rate * fills_number * average_price) total_price,
COUNT (fx_rate * fills_number * average_price) cnt_total_price,
SUM (CASE
WHEN order_status = 'C'
THEN 1
ELSE 0
END) cancelled_orders_count,
COUNT (CASE
WHEN order_status = 'C'
THEN 1
ELSE 0
END
) cnt_cancelled_orders_count,
-- SUM(t.FX_RATE*t.NO_FILLS*t.AVG_PRICE) AVERAGE_PRICE,
-- SUM(FILLS_NUMBER*AVERAGE_PRICE) STAGING_AVERAGE_PRICE,
-- COUNT(FILLS_NUMBER*AVERAGE_PRICE) CNT_STAGING_AVERAGE_PRICE,
SUM (routed_orders_no) routed_orders_no,
COUNT (routed_orders_no) cnt_routed_orders_no,
SUM (routed_liquidity_qty) routed_liquidity_qty,
COUNT (routed_liquidity_qty) cnt_routed_liquidity_qty,
SUM (removed_liquidity_qty) removed_liquidity_qty,
COUNT (removed_liquidity_qty) cnt_removed_liquidity_qty,
SUM (added_liquidity_qty) added_liquidity_qty,
COUNT (added_liquidity_qty) cnt_added_liquidity_qty,
SUM (agent_charges) agent_charges,
COUNT (agent_charges) cnt_agent_charges,
SUM (clearing_charges) clearing_charges,
COUNT (clearing_charges) cnt_clearing_charges,
SUM (execution_charges) execution_charges,
COUNT (execution_charges) cnt_execution_charges,
SUM (transaction_charges) transaction_charges,
COUNT (transaction_charges) cnt_transaction_charges,
SUM (order_management) order_management,
COUNT (order_management) cnt_order_management,
SUM (settlement_charges) settlement_charges,
COUNT (settlement_charges) cnt_settlement_charges,
SUM (recovered_agent) recovered_agent,
COUNT (recovered_agent) cnt_recovered_agent,
SUM (recovered_clearing) recovered_clearing,
COUNT (recovered_clearing) cnt_recovered_clearing,
SUM (recovered_execution) recovered_execution,
COUNT (recovered_execution) cnt_recovered_execution,
SUM (recovered_transaction) recovered_transaction,
COUNT (recovered_transaction) cnt_recovered_transaction,
SUM (recovered_ord_mgt) recovered_ord_mgt,
COUNT (recovered_ord_mgt) cnt_recovered_ord_mgt,
SUM (recovered_settlement) recovered_settlement,
COUNT (recovered_settlement) cnt_recovered_settlement,
SUM (client_agent) client_agent,
COUNT (client_agent) cnt_client_agent,
SUM (client_order_mgt) client_order_mgt,
COUNT (client_order_mgt) cnt_client_order_mgt,
SUM (client_exec) client_exec, COUNT (client_exec) cnt_client_exec,
SUM (client_trans) client_trans,
COUNT (client_trans) cnt_client_trans,
SUM (client_clearing) client_clearing,
COUNT (client_clearing) cnt_client_clearing,
SUM (client_settle) client_settle,
COUNT (client_settle) cnt_client_settle,
SUM (chargeable_taxes) chargeable_taxes,
COUNT (chargeable_taxes) cnt_chargeable_taxes,
SUM (vendor_charge) vendor_charge,
COUNT (vendor_charge) cnt_vendor_charge,
SUM (routing_charges) routing_charges,
COUNT (routing_charges) cnt_routing_charges,
SUM (recovered_routing) recovered_routing,
COUNT (recovered_routing) cnt_recovered_routing,
SUM (client_routing) client_routing,
COUNT (client_routing) cnt_client_routing,
SUM (ticket_charges) ticket_charges,
COUNT (ticket_charges) cnt_ticket_charges,
SUM (recovered_ticket_charges) recovered_ticket_charges,
COUNT (recovered_ticket_charges) cnt_recovered_ticket_charges
FROM us_datamart_raw
GROUP BY order_date,
if_system,
business_dim_id,
time_dim_id,
account_dim_id,
ordertype_dim_id,
instr_dim_id,
execution_dim_id,
exec_exchange_dim_id;
-- Note: Index I_SNAP$_MV_US_DATAMART will be created automatically
-- by Oracle with the associated materialized view.
CREATE UNIQUE INDEX OS_OWNER.MV_US_DATAMART_UDX ON OS_OWNER.MV_US_DATAMART
(ORDER_DATE, TIME_DIM_ID, BUSINESS_DIM_ID, ACCOUNT_DIM_ID, ORDERTYPE_DIM_ID,
INSTR_DIM_ID, EXECUTION_DIM_ID, EXEC_EXCHANGE_DIM_ID)
NOLOGGING
NOPARALLEL
COMPRESS 7;
No of rows: 2228558
The query (taken Mondrian) I run against each of them is:
select sum("MV_US_DATAMART"."NOTIONAL") as "m0"
--, sum("MV_US_DATAMART"."FILLED_QUANTITY") as "m1"
--, sum("MV_US_DATAMART"."AGENT_CHARGES") as "m2"
--, sum("MV_US_DATAMART"."CLEARING_CHARGES") as "m3"
--, sum("MV_US_DATAMART"."EXECUTION_CHARGES") as "m4"
--, sum("MV_US_DATAMART"."TRANSACTION_CHARGES") as "m5"
--, sum("MV_US_DATAMART"."ROUTING_CHARGES") as "m6"
--, sum("MV_US_DATAMART"."ORDER_MANAGEMENT") as "m7"
--, sum("MV_US_DATAMART"."SETTLEMENT_CHARGES") as "m8"
--, sum("MV_US_DATAMART"."COMMISSION") as "m9"
--, sum("MV_US_DATAMART"."RECOVERED_AGENT") as "m10"
--, sum("MV_US_DATAMART"."RECOVERED_CLEARING") as "m11"
--,sum("MV_US_DATAMART"."RECOVERED_EXECUTION") as "m12"
--,sum("MV_US_DATAMART"."RECOVERED_TRANSACTION") as "m13"
--, sum("MV_US_DATAMART"."RECOVERED_ROUTING") as "m14"
--, sum("MV_US_DATAMART"."RECOVERED_ORD_MGT") as "m15"
--, sum("MV_US_DATAMART"."RECOVERED_SETTLEMENT") as "m16"
--, sum("MV_US_DATAMART"."RECOVERED_TICKET_CHARGES") as "m17"
--,sum("MV_US_DATAMART"."TICKET_CHARGES") as "m18"
--, sum("MV_US_DATAMART"."VENDOR_CHARGE") as "m19"
from "OS_OWNER"."MV_US_DATAMART" "MV_US_DATAMART"
where I uncomment a column at a time and rerun. I improved the TimesTen results since my first post, by retyping the NUMBER columns to BINARY_FLOAT. The results I got were:
No Columns ORACLE TimesTen
1 1.05 0.94
2 1.07 1.47
3 2.04 1.8
4 2.06 2.08
5 2.09 2.4
6 3.01 2.67
7 4.02 3.06
8 4.03 3.37
9 4.04 3.62
10 4.06 4.02
11 4.08 4.31
12 4.09 4.61
13 5.01 4.76
14 5.02 5.06
15 5.04 5.25
16 5.05 5.48
17 5.08 5.84
18 6 6.21
19 6.02 6.34
20 6.04 6.75 -
Using Oracle ODBC Gateway connecting to a remote Oracle database
Oracle 11gR2
RHEL 6.4
Has anyone use the Oracle ODBC Gateway to connect to another Oracle database? Any issues with that configuration? Where do I get the ODBC drivers for Linux?
(I know, "why not use a dblink?" -- well that would be against company security policies)Hi,
From the Oracle point of view we support using DG4ODBC for Oracle to Oracle connections. However, we have not actually tested it as DG4ODBC is primarily designed for access to non-Oracle databases.
How DG4ODBC will work between Oracle databases depends on the ODBC driver and what that supports. You will need an ODBC Oracle driver which you can get from various suppliers including Oracle but also vendors such as DataDirect, Easysoft etc. You could try a Google search.
You say you do not want to use database links but that is how DG4ODBC is used, You cannot do -
sqlplus user/password@dg4odbc_oracle
Once Dg4ODBC is setup and configured as in this note - if you are using Linux 64-bit -
How to Configure DG4ODBC on 64bit Unix OS (Linux, Solaris, AIX, HP-UX Itanium) to Connect to Non-Oracle Databases Post Install (Doc ID 561033.1)
then in the Oracle database you create a database link and select from tables in the other Oracle database -
select * from table@dg4odbc_db_link ;
Regards,
Mike -
Problem with httpd and oracle database
hi
i have oracle linux with oracle database 11gR2
i install php with oci as the following link :
http://oss.oracle.com/projects/php/dist/documentation/installation.html
but when i try phpinfo() i found just only
*" /etc/php.d/oci8.ini"* in the result
and when i test connection i have this error
*" Fatal error: Call to undefined function oci_connect() "*
so what is the best way to install oci on oracle linux ?What version of the OS are you using? Any clues in the syslog and httpd log? Are you using SELinux, which is often an overlooked issue?
Can you provide more details?
so what is the best way to install oci on oracle linux ?According to the instructions, if support is needed: http://www.oracle.com/technetwork/topics/php/zend-server-096314.html -
Import error on Oracle Database Express 10.2.0.1.0
Hi,
I try to import data from oracle V10.01.02 running on SUSE10 Linux to
oracle 10g (10.1.0.2.0) running on Windows Server 2003.
I am able to import the big part from my ata, but not all data.
The begin of my log file is:
Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0
- Production
With the Partitioning, OLAP and Data Mining options
Export file created by EXPORT:V10.01.00 via conventional path
And after some time I receive:
IMP-00019: row rejected due to ORACLE error 12899
IMP-00003: ORACLE error 12899 encountered
ORA-12899: value too large for column "SAFCI"."EMP"."NAME"
(actual: 65, maximum: 64)
Column 1 1000005025
Column 2 ??????? ???????
Column 3 19-SEP-2002:00:00:00
Column 4 9089
Column 5 ??. ?????
Column 6 1.9828
Column 7 377.77
Column 8 75.55
Column 9 ???????????? ???????? ? ??? ???? ? 32 ??.
Column 10 19-SEP-2002:00:00:00
Column 11 ?????? ?????, ?.?. 172747675, ???. 23.11.200?
Column 12 ????????? ?????????
Column 13 ? ????
Column 14 T
Column 15
Column 16 F
IMP-00019: row rejected due to ORACLE error 12899
IMP-00003: ORACLE error 12899 encountered
ORA-12899: value too large for column "SAFCI"."EMP"."NAME"
(actual: 65, maximum: 64)
Column 1 1000006408
Column 2 ??????? ???????
Column 3 05-NOV-2002:00:00:00
Column 4 9089
Column 5 ??. ?????
Column 6 1.939
Column 7 82
Column 8 16.4
Column 9 ?????????? ? ???? ???? ? 40 ??.
Column 10 05-NOV-2002:00:00:00
Column 11 ?????? ?????, ?.?. 172747675, ???. 23.11.200?
Column 12 ????????? ?????????
Column 13 ? ????
Column 14 T
Column 15
Column 16 F 36943 rows imported
I can not understand this problem, because I exported the hole user and
also try to import the hole user in my new system.
Pls., can some one point me to some paper about this problem or help me
to solve the problem.
Thanks
configuration Oracle10g (EXP source)
SQLWKS> select * from nls_database_parameters
2>
PARAMETER VALUE
NLS_LANGUAGE BRAZILIAN PORTUGUESE
NLS_TERRITORY BRAZIL
NLS_CURRENCY R$
NLS_ISO_CURRENCY BRAZIL
NLS_NUMERIC_CHARACTERS ,.
NLS_CHARACTERSET WE8ISO8859P1
NLS_CALENDAR GREGORIAN
NLS_DATE_FORMAT DD/MM/RR
NLS_DATE_LANGUAGE BRAZILIAN PORTUGUESE
NLS_SORT WEST_EUROPEAN
NLS_TIME_FORMAT HH24:MI:SSXFF
NLS_TIMESTAMP_FORMAT DD/MM/RR HH24:MI:SSXFF
NLS_TIME_TZ_FORMAT HH24:MI:SSXFF TZR
NLS_TIMESTAMP_TZ_FORMAT DD/MM/RR HH24:MI:SSXFF TZR
NLS_DUAL_CURRENCY Cr$
NLS_COMP BINARY
NLS_LENGTH_SEMANTICS BYTE
NLS_NCHAR_CONV_EXCP FALSE
NLS_NCHAR_CHARACTERSET AL16UTF16
NLS_RDBMS_VERSION 10.1.0.2.0
configuration Oracle10g Express (IMP destination)
SQL> select * from nls_database_parameters;
PARAMETER VALUE
NLS_LANGUAGE AMERICAN
NLS_TERRITORY AMERICA
NLS_CURRENCY $
NLS_ISO_CURRENCY AMERICA
NLS_NUMERIC_CHARACTERS .,
NLS_CHARACTERSET AL32UTF8
NLS_CALENDAR GREGORIAN
NLS_DATE_FORMAT DD-MON-RR
NLS_DATE_LANGUAGE AMERICAN
NLS_SORT BINARY
NLS_TIME_FORMAT HH.MI.SSXFF AM
NLS_TIMESTAMP_FORMAT DD-MON-RR HH.MI.SSXFF AM
NLS_TIME_TZ_FORMAT HH.MI.SSXFF AM TZR
NLS_TIMESTAMP_TZ_FORMAT DD-MON-RR HH.MI.SSXFF AM TZR
NLS_DUAL_CURRENCY $
NLS_COMP BINARY
NLS_LENGTH_SEMANTICS BYTE
NLS_NCHAR_CONV_EXCP FALSE
NLS_NCHAR_CHARACTERSET AL16UTF16
NLS_RDBMS_VERSION 10.2.0.1.0Hi
Your import database use a multibyte characterset, your export db a singlebyte cs.
This means, a char can need more than 1 byte.
Try this before import (and before create tables!!):
alter system set nls_length_semantics=char;Greetings
Sven -
How to take Input from the callers telephone keypad into oracle database
My client requires to automate a registration task for his customers by telephone. The requirement is the Oracle database should take numeric input from the callers telephone keypad eg 1 or 2 or 3 and the oracle must store this input inside its database and do some required processing with such input and generate a unique numeric id (sequence) and send back this unique numeric ID to the caller who is on hold. Please Note. Is there any electronic device to be installed at the client site to solve the above task eg. any telephone line to be connected to the CPU and etc?
I completely have zero knowledge. Please can you outline/summarise the instructions how to construct and install to meet my client requirement.
If this is not possible in Oracle I Can migrate to any database. Please suggest.
please help me. if you cant help i appretiate if you give me some cluesYou need to start with your telephone system. Normally this situation would involve some sort of PBX or other internal switching equipment. Talk to your vendor. There's likely to be software, etc. available to perform the interaction with the end user.
It should have programmable "hooks" to interface with various databases. To Oracle it just looks like another program.
Ken -
Help needed to loadjava apache poi jars into oracle database.
Help needed to loadjava apache poi jars into oracle database. Many classes left unresolved. (Poi 3.7, database 11.1.0.7). Please share your experience!
Hi,
The first 3 steps are just perfect.
But with
loadjava.bat -user=user/pw@connstr -force -resolve geronimo-stax-api_1.0_spec-1.0.jar
the results are rather unexpected. Here is a part of the log file:
arguments: '-user' 'ccc/***@bisera7-db.dev.srv' '-fileout' 'c:\temp\load4.log' '-force' '-resolve' '-jarsasdbobjects' '-v' 'geronimo-stax-api_1.0_spec-1.0.jar'
The following operations failed
resource META-INF/MANIFEST.MF: creation (createFailed)
class javax/xml/stream/EventFilter: resolution
class javax/xml/stream/events/Attribute: resolution
class javax/xml/stream/events/Characters: resolution
class javax/xml/stream/events/Comment: resolution
class javax/xml/stream/events/DTD: resolution
class javax/xml/stream/events/EndDocument: resolution
class javax/xml/stream/events/EndElement: resolution
class javax/xml/stream/events/EntityDeclaration: resolution
class javax/xml/stream/events/EntityReference: resolution
class javax/xml/stream/events/Namespace: resolution
class javax/xml/stream/events/NotationDeclaration: resolution
class javax/xml/stream/events/ProcessingInstruction: resolution
class javax/xml/stream/events/StartDocument: resolution
class javax/xml/stream/events/StartElement: resolution
class javax/xml/stream/events/XMLEvent: resolution
class javax/xml/stream/StreamFilter: resolution
class javax/xml/stream/util/EventReaderDelegate: resolution
class javax/xml/stream/util/StreamReaderDelegate: resolution
class javax/xml/stream/util/XMLEventAllocator: resolution
class javax/xml/stream/util/XMLEventConsumer: resolution
class javax/xml/stream/XMLEventFactory: resolution
class javax/xml/stream/XMLEventReader: resolution
class javax/xml/stream/XMLEventWriter: resolution
class javax/xml/stream/XMLInputFactory: resolution
class javax/xml/stream/XMLOutputFactory: resolution
class javax/xml/stream/XMLStreamReader: resolution
resource META-INF/LICENSE.txt: creation (createFailed)
resource META-INF/NOTICE.txt: creation (createFailed)
It seems to me that the root of the problem is the error:
ORA-29521: referenced name javax/xml/namespace/QName could not be found
This class exists in the SYS schema though and is valid. If SYS should be included as a resolver? How to solve this problem? -
Difference between upgrdae and migration about oracle database
Difference between upgrdae and migration about oracle database
please give the commentsWell, the question is almost philosophic...<br>
In 9i, there is a Migration Guide whereas in 10g there is a Upgrade Guide.<br>
Furthermore, in 9i, there is the command line startup migrate whereas in 10g that's startup upgrade.
Somebody think upgrade when go to new release, and migration when go to new version.<br>
Others think upgrade when new version replace database in place, and migration when new version include a move of database.<br>
Another point of view is : upgrade is for technical, and migration for application/data.<br>
<br>
Well, after these explanations, your upgrade/migratation notion will not be more clear, but I think that is not very important, only a terminology game. The most important is to know what you need : new version or new release.<br>
<br>
Nicolas.
Maybe you are looking for
-
New problem printing from InDesign CS4 in Lion
After having great success with Aisolve's earlier fix on printing from InDesign CS4 after upgrading to Lion, I'm having a new problems after installing the OS X 10.7.2 upgrade last week. When I try to print from InDesign CS4 and Illustrator CS4 to m
-
trying to rent a movie on apple tv for the first time and have to authorize my computer but can't find the itunes menu that allows me to do this. any thoughts out there?
-
Concatenate out put of 2 or more columns in BEX Report in to one column
Hi, I am working on service order report in BEX, and i would like to Concatenate out put of 2 or more columns in BEX Report in to one column. Please share me if any scenarios. Thank Neni
-
Cancel a submit in a javascript
I have a tableview model which provides a list of order numbers. I have made it so this column of order numbers is a clickable link. I have set the table column to encode(false) so that my column shows up as an actual link. Here is the actual c
-
hi all ! I want to change the color of my rows depending on the result I get from a request. Something like <c:forEach var="fila" items="${requestScope.citas_diarias}"> <c:set var="index" value="${index+1}"/> <td width="10" class="IndexColumn" ><span