Excessive Redo Generation After Upgrading on Oracle 10gR2

We had our production database hosted on Oracle 9.2.0. Few months back we have migrated it to Oracle 10.2.0.4.0.
After Migration I have noticed that redo generation has become very very high. In earlier case no. of log files generating in production hours were around 20 where as after migration it become around 200 files per day. I have run statspack report on this database. Statspack report is also saying that log file switch wait is become very high. Parameter timed_statistics has also been set to FALSE. Workload on the database is same before & after upgrade. Queries running in the sessions are also same before & after upgrade. All the parameters & memory structures are same after upgrade. Satatpack report is saying that db block change & disk write is become very high. I had used import export for upgrading the databases. Please provide a solution for this problem.
Thanks In advance for all your favours....

Hi;
Please check below notes which could be helpful for your issue:
Diagnosing excessive redo generation [ID 199298.1]
Excessive Archives / Redo Logs Generation Troubleshooting [ID 832504.1]
Troubleshooting High Redo Generation Issues [ID 782935.1]
How to Disable (Temporary) Generation of Archive Redo Log Files [ID 177218.1]
Regard
Helios

Similar Messages

  • FORM NOT RESPONDING AFTER UPGRADING TO ORACLE 10GR2 DATABASE

    Hi All
    We are upgrading from Oracle 9i (single AIX Server) to Oracle 10gR2 on HACMP of 2 AIX 5.3 Server.
    Our application is in Forms 9.0.4 deployed into iAS 10.1.2.0.2.
    There is a custom Login Form that checks for UserID and Password and validate
    to Authentication Table.
    This form works fine in Oracle 9i but doesn't work in Oracle 10g (After
    Upgrading).
    On the Login Form, this happen :
    After user enter UserID the forms query the DB and display UserName succesfully,
    but after enter Password and press OK, it does not responding,
    takes forever to complete..
    Additional information :
    Before :          After :
    Oracle 9i          Oracle 10gR2
    SIngle AIX Server     Two AIX Server in HACMP
    Charset : AL16UTF16     WE8ISO....
    Given the information above, What could possibly cause this problem ?
    Thank you for your help,
    xtanto.

    xtanto,
    not sure what causes the problem, but my suggestion is to check the password query.
    To give you another hint: Your way of authenticating users is not hack safe. As you describe that providing a user ID will show the name first before I have to provide the password, it allows me to fish for valid user accounts. The recommended way of using login screens is to have the user enetring userid and password in teh same screen and use both information to authenticate the user in a single query.
    Frank

  • Invalid SQL command after upgrading to Oracle 9i

    After upgrading to Oracle 9i, the SQL queries in some of our existing reports are no longer valid.  The Oracle error returned is "old style outer join () can not be used with ANSI joins". It appears that the SQL query generated by Crystal Reports is correct in that it uses "LEFT OUTER JOIN" instead of just naming the tables separated by a comma, but it also includes the "()" qualifer on the matching columns which is no longer allowed.  Is there any way to overcome this error short of rewriting the reports based on a command?

    Hello,
    A couple suggestions come to mind.
    Business Objects tend to take problematic SQL generation very seriously at least on the SQL and Oracle platforms (due to their widespread use) and I would first grab the latest database/engine hotfixes to be certain it hasn't been tracked and corrected already. If you can clearly reproduce this issue (sounds like you can) then it should be easy to convince the support team at BOBJ to track the issue. However...
    Since their hotfix release cycle is rather slow especially for items that have workarounds (even sometimes painful workarounds) you may want to consider using a Stored Procedure instead of a Command.
    Commands have a notorious history of issues/limitations that tend to burn you later - especially when you have to update many reports. A stored procedure would likely render the + qualifier a moot point.
    Cheers, Dylan
    <p><strong>Dylan Lopez - </strong>604.628.4467 </p><p>Previously: Microsoft .NET Resource in Developer Support @ Business Objects</p><p>Currently: Principal @ Inno Software Inc / Independent Contractor Group</p>

  • Password problem in oracle 10g after upgrading from oracle 9i

    Hi to all,
    i'm migrating oracle 9i to oracle 10g. in oracle 9i username/password are same character(eg kareem/kareem)
    after upgrading to oracle 10g will it create any problem. Because in oracle 10g the condition is when creating user is username and password should not contain same character (kareem/kareem not allowed).
    Thanks in advance

    in oracle 10g document, they have mention below condition to create user. See the second condition.
    • Passwords must be between 4 and 30 characters long.
    • Passwords cannot be the same as the user name.
    • Passwords must be from the database character set and can include the underscore (_), dollar ($), and pound sign (#) characters.
    • Passwords cannot be Oracle reserved words.
    • Don't start the password with a numeral.
    • Don's use change_on_install or manager as passwords, since they are used by Oracle.
    As i'm going to upgrade oracle 9i to 10g. if it give problem because username and password are same then i have to request my development team to change the password in their application.Can you please verify and tell me.
    Edited by: user7071421 on Sep 25, 2008 4:33 AM

  • Any performance gains after upgrade to Oracle 10 ?

    Hello,
    we have been using EBS 11.5.9 (with Oracle 9). Does anybody see any improvements from performance point of view after upgrade to Oracle 10r2 (but without upgrade of EBS to R12) ?
    Regards

    Somethings have been faster for us and somethings have been slower. In general it seems the same or better. When we get a chance to look at the jobs that are performing poorly I'm sure we will be able to make them perform. Make sure you find and apply all of the performance patches for the modules you have implemented, we are missing one right now and it is hurting our Configurator developers instance. I have actually set the the optimizer_features_enable back to 9i. This is something I would not recommend doing in a production environment, and we are only doing it until we can get the patch that fixes the optimizer in 10g applied.
    One surprise for us was the increase in memory required, in our test and development instances we have found a 1GB SGA is as small as you can go with 10g and still have reasonable performance. We had a number of 9i ebus databases around 300-400MB SGA that performed just fine. This limited the number of environments we could squeeze on a single server (not really a bid deal if you only have a few dev, test and training environments).
    In our production environment in order to have the same size datacache more memory was consumed. Were we could get away with a shared pool of 4GB under 9i, it grows to over 6GB in 10g. Looks like more PGA is also being used. 10g along with RUP5 and upgrading the java from 1.3 to 1.5 has increased our memory consumption. We increased memory from 64GB to 96GB and we are using a large portion of the additional 32GB. We were right at the limit with 64GB in our 9i production instance but we had never experienced the paging that we did with 10g. Now at 96GB we nolonger see any paging.

  • Excessive log switching after upgrading to RAC + DG

    Hi everyone, I'm not sure if this should go in the RAC or the Data Guard section, seeing how I'm not sure which one is the potential culprit, but here goes:
    I recently assisted in upgrading a database from Solaris 9i to HP-UX 10gR2 (10.2.0.4) using RAC and Data Guard (2 nodes). However after upgrading we noticed that there is now an increase in the redo log switches. We currently have 4 logfile groups for each thread sized at 100MB. Previously when the database was just a single instance, such a configuration was sufficient as we were only seeing 4 log switches per hour, max 8 at peak times. After the migration, this spiked up to around 40-60 log switches per hour, sometimes hitting 80.
    The userbase/load of the database hasn't changed since the upgrade, we just migrated the database. Is this normal? Should I have to increase the redo logs after migrating to RAC + DG? I've implemented DG before as well as RAC, but not together so I'm not sure if this would be a factor, however I'm not totally sure if that line of thinking is correct. Any help would be much appreciated.

    Previously when the database was just a single instance, such a configuration was sufficient as we were only seeing 4 log switches per hour, max 8 at peak times. After the migration, this spiked up to around 40-60 log switches per hour, sometimes hitting 80. Are all the ArchiveLogs actually 100MB each ? If Oracle has to switch logs more frequently, it doesn't necessarily mean that more Redo is being generated. Check the volum of Redo. (e.g. from v$sysstat, from StatsPack, from AWR, from V$ARCHIVED_LOG or by just listing the ArchiveLog file sizes)
    Hemant K Chitale

  • Excess archive generation after migration

    hi
    Am facing large size of archive generation after my migration to 11.2.0.1 from 10.2.0.1 in windows 32 bit.
    While I was using 10g my total archive size in a day was only 10GB. now after the migration its coming al most 50gb.
    Guys.. can anybody let me know what i need to do to reduce my archive generation?
    Am using the same application and queries as in 10g.

    Disconnect everyone from the database.
    Or learn how to read documentation and how to administer Oracle.
    Or read up on the differences between 11gR2 and 10gR2.
    Sybrand Bakker
    Senior Oracle DBA

  • Error while executing SSIS Package after upgrading the Oracle Client

    Hello,
    I am running to the below error after we upgraded the Oracle client to 11g from 10g. The source of the Data is Oracle database . Looked online but each error is pointing towards different route, Confused here any help is very much appreciated. Have to say
    I am not a pro with SSIS.
    SSIS package "Bentek_ISO_Package (2).dtsx" starting.
    Information: 0x4004300A at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Validation phase is beginning.
    Information: 0x4004300A at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Validation phase is beginning.
    Warning: 0x80049304 at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Warning: Could not open global shared memory to communicate with performance DLL; data flow performance counters are not available.  To resolve, run this package as an administrator,
    or on the system's console.
    Information: 0x40043006 at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Prepare for Execute phase is beginning.
    Information: 0x40043007 at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Pre-Execute phase is beginning.
    Error: 0xC0202009 at Temp_tbl Injection Data Flow Task, OLE DB Source [1]: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80040E14.
    An OLE DB record is available.  Source: "OraOLEDB"  Hresult: 0x80040E14  Description: "ORA-12801: error signaled in parallel query server P183, instance *****:qrmgmt2 (2)
    ORA-01861: literal does not match format string".
    An OLE DB record is available.  Source: "OraOLEDB"  Hresult: 0x80004005  Description: "ORA-12801: error signaled in parallel query server P183, instance ****:qrmgmt2 (2)
    ORA-01861: literal does not match format string".
    Error: 0xC004701A at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: component "OLE DB Source" (1) failed the pre-execute phase and returned error code 0xC0202009.
    Information: 0x4004300B at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: "component "OLE DB Destination" (52)" wrote 0 rows.
    Information: 0x40043009 at Temp_tbl Injection Data Flow Task, SSIS.Pipeline: Cleanup phase is beginning.
    Task failed: Temp_tbl Injection Data Flow Task
    Warning: 0x80019002 at *******: SSIS Warning Code DTS_W_MAXIMUMERRORCOUNTREACHED.  The Execution method succeeded, but the number of errors raised (2) reached the maximum allowed (1); resulting in failure. This occurs when the number of errors reaches
    the number specified in MaximumErrorCount. Change the MaximumErrorCount or fix the errors.
    SSIS package "****** (2).dtsx" finished: Failure.

    Do you have any Date conversion functions in your Oracle datetime fields? I did run into a similar issue when we upgraded the oracle client and the provider started to decipher the metadata of the date columns. The SSIS metadata conversion for the date
    columns yielded dates that were not valid. I had to update my queries to use the TO_CHAR conversion to bring it out as a string and then save it to the date field.
    TO_CHAR (DATE, 'MM/DD/YYYY HH24:MI:SS')
    Regards, Dinesh

  • SQL query was failed in my report after migrating to Oracle 10gR2

    We have a aplication running under Oracle 9.2.0.4 DB. we migrated our DB to Oracle 10gR2. While running the report, we got an error:
    ORA-00904: "A3"."FACILITY_SYSTEM_ID": invalid identifier
    ORA-06512: at "CEAS_MK_RPT.GET_LIST", line 12
    The SQL query which causes this error is:
    INSERT INTO gtt_facility_seq_generator
    (date_range)
    SELECT get_list
    (CURSOR (SELECT auf.facility_system_id
    FROM authorized_facility auf
    WHERE auf.facility_system_id = afs.facility_system_id
    AND auf.credit_application_system_id = :p_cred_appln_id
    FROM authorized_facility afs,
    (SELECT af.facility_system_id facility_system_id
    FROM facility_obligor fo, authorized_facility af, party p
    WHERE fo.facility_system_id = af.facility_system_id
    AND fo.party_system_id = p.party_system_id
    AND NOT EXISTS (
    SELECT NULL
    FROM facility_third_party_subst iftps
    WHERE iftps.facility_system_id =
    fo.facility_system_id)
    ) tab1
    WHERE afs.credit_application_system_id = :p_cred_appln_id
    AND tab1.facility_system_id = afs.facility_system_id
    ORDER BY afs.creation_date;
    The content of function get_list() is:
    CREATE OR REPLACE FUNCTION CEAS_MK_RPT.Get_List
    p_cursor IN sys_refcursor
    RETURN VARCHAR2
    IS
    l_sep VARCHAR2(4);
    l_text VARCHAR2(30000);
    l_text_return VARCHAR2(30000);
    BEGIN
    LOOP
    FETCH p_cursor INTO l_text;
    EXIT WHEN p_cursor%NOTFOUND;
    l_text_return := l_text_return || l_sep || l_text;
    l_sep := CHR(10);
    END LOOP;
    CLOSE p_cursor;
    RETURN l_text_return;
    dbms_output.put_line ('a');
    END Get_List;
    The same report was executed perfectly in Oracle 9i. Kindly help us to short out this issue. Thanks in advance.
    Regards,
    Sengol S

    Hi Nirav,
    Many thanks for your response. I verified the same, but the same query was executed successfully after removing the INSERT INTO part(executed the SELECT part only). What might be the cause for this error?
    For your kind information :
    I have 2 schema in my DB( ceas_mk_app and ceas_mk_rpt ).
    All the source tables are present in ceas_mk_app schema - These tables are used by ceas_mk_rpt schema(using synonym) to generate the report.
    I'm executing the above said query from ceas_mk_rpt to insert the necessary data in to report temp. table.
    Thanks,
    Sengol S
    Edited by: seng1256 on Dec 22, 2008 12:31 PM

  • Issues after Upgrade from Oracle 10.1.0.5 to 10.2.0.5 in Windows Server2003

    Good Morning Good People,
    I have run in to some issues after upgrading my 10.1.0.5 database to 10.2.0.5. Initially after the firtst upgrade from 10.1.0.5 to 10.2.0.0 one has to run Critical Patch set 4 to get to 10.2.0.5. Well after initial up grade everything worked fine. I could connect to the database from within the server as well as remotely through SQL*PLUS client. However, after completeing the patch I'm still able to connect to the database from within the server, but remotely i get the following error:
    ORA-12514 : TNS Listener does not currently know of service requested in connect descriptor.
    I also get Error ORA-12505: TNS:listener does not currently know of SID given in connect descriptor The Connection descriptor used by the client was: (description=(address=(host=mcquery2)(protocol=tcp)(port=1521))(connect_data=(sid=mcquery)(server=dedicated)))
    I've tried serval different things to trouble shoot the issue but nothing seems to work.
    If anyone has run into s similar issue please help. This is ruining my weekend! LOL!
    Edited by: user10835577 on Aug 29, 2010 9:15 AM

    From the client or the server?
    SERVER:
    tnsnames.ora:
    MCQUERY=
    (DESCRIPTION=
    (ADDRESS=
    (PROTOCOL=TCP)
    (HOST=10.36.27.184)
    (PORT=1521)
    (CONNECT_DATA=
    (SERVICE_NAME=MCQUERY)
    listener.ora
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = D:\oracle\product\10.1.0\Db_1)
    (PROGRAM = extproc)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = TCP)(HOST = 10.36.27.184)(PORT = 1521))
    FROM REMOTE SQLPLUS Client:
    tnsnames.ora
    MCQUERY=
    (DESCRIPTION=
    (ADDRESS=
    (PROTOCOL=TCP)
    (HOST=10.36.27.184)
    (PORT=1521)
    (CONNECT_DATA=
    (SERVICE_NAME=MCQUERY)
    listener.ora
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (SID_NAME = PLSExtProc)
    (ORACLE_HOME = D:\oracle\product\10.1.0\Client_3)
    (PROGRAM = extproc)
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS_LIST =
    (ADDRESS = (PROTOCOL = TCP)(HOST = cotsdot370.cotsdot.dom)(PORT = 1521))
    (ADDRESS = (PROTOCOL = TCP)(HOST = 10.36.27.184)(PORT = 1521))
    Thanks,
    Malcolm
    Edited by: user10835577 on Aug 29, 2010 12:01 PM

  • Issue  after upgrade from oracle 10.2.0.4 to 11g (11.1.0.6.0)

    hello
    After upgrade to 11g we have problem with dbms_workload_replay package it has error Error(1): PLS-00753: malformed or corrupted wrapped unit in body and can't be compiled . It is wrapped so we can't find error. We run utlrp.sql but no results.
    Any idea ??
    THX

    Then it looks similar to bug 6651232
    May I suggest you get in touch with Oracle support for this?

  • Performance Problem After upgrade to oracle 10g

    Hi
    I have upgrade one of my datawarehouse database from oracle 9.2.0.8 to oracle 10.2.0.4 running on solaris 9
    After the upgrade jobs which were running in the database is taking hell lot of time.
    The jobs are accessing the views which is used to get the monthly report data from the database.
    what could be the solution and where to start from to get the RCA to resolve this performance issue
    Please let me know if you require any other information
    database is currently running in the automatic shared memory management mode ie SGA_MAX and SGA_TARGET parameters are defined for that

    There are a lot of differences between 10g and 9i in this regard, among these are:
    - There is a default job that gathers statistics every night which is not there in 9i. You might have totally different statistics as in 9i due to that job, depending on how and if at all you used to collect statistics in 9i
    - The 10g DBMS_STATS package collects histograms on some columns by default (parameter METHOD_OPT=>'FOR ALL COLUMNS SIZE AUTO' default in 10g whereas 'FOR ALL COLUMNS SIZE 1' in 9i) which can have a significant effect on the execution plans
    - The 10g optimizer has CPU costing enabled by default which can make significant changes to your execution plans due to different costing of table scans and order of predicate evaluation. In addition it uses NOWORKLOAD system statistics if system statistics have not been gathered explicitly
    - 10g checks the min and max values stored for columns in the data dictionary. If your predicates are way off compared to these values then 10g begins to adjust the calculated selectivity of the predicate which can again significantly affect your execution plans
    - 10g introduces the "Cost Based Query Transformation (CBQT)" feature which means that rather than applying heuristic transformation rules transformations are costed and potentially discarded whereas 9i applies transformations unconditionally whenever possible
    Check also the following note resp. white paper:
    http://optimizermagic.blogspot.com/2008/02/upgrading-from-oracle-database-9i-to.html
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • Data Protector backup integration fails after upgrade to Oracle 11.2.0.2

    After upgrading a system from Oracle 10.2.0.2 to 11.2.0.2, the Data Protector backup integration fails with error
    [Normal] From: BSM@<dp_cell_server>.<domain> "<sap_host>_<SID>_Online"  Time: 28.04.2011 09:10:21
         Backup session 2011/04/28-74 started.
    [Normal] From: BSM@<dp_cell_server>.<domain> "<sap_host>_<SID>_Online"  Time: 28.04.2011 09:10:21
         OB2BAR application on "<sap_host>.<domain>" successfully started.
    /usr/sap/<SID>/SYS/exe/run/brbackup: error while loading shared libraries: libclntsh.so.10.1: cannot open shared object file: No such file or directory
    [Major] From: OB2BAR_OMNISAP@pervs<SID>.<domain> "OMNISAP"  Time: 04/28/11 09:10:21
         BRBACKUP /usr/sap/<SID>/SYS/exe/run/brbackup -t online -d util_file_online -c -m all -u system/******** returned 127
    /usr/sap/<SID>/SYS/exe/run/brarchive: error while loading shared libraries: libclntsh.so.10.1: cannot open shared object file: No such file or directory
    [Major] From: OB2BAR_OMNISAP@pervs<SID>.<domain> "OMNISAP"  Time: 04/28/11 09:10:21
         BRARCHIVE /usr/sap/<SID>/SYS/exe/run/brarchive -d util_file -s -c -u system/******** returned 127
    [Normal] From: BSM@<dp_cell_server>.<domain> "<sap_host>_<SID>_Online"  Time: 28.04.2011 09:10:22
         OB2BAR application on "<sap_host>.<domain>" disconnected.
    [Critical] From: BSM@<dp_cell_server>.<domain> "<sap_host>_<SID>_Online"  Time: 28.04.2011 09:10:22
         None of the Disk Agents completed successfully.
         Session has failed.
    We checked the configuration, tried linking /oracle/client/11x_64 to 10x_64, removing the link and renewing the 10x_64 oracle client, but still Data Protector can't find the library libclntsh.so.10.1. A new system which was installed with Oracle 11.2.0.2 doesn't have any problems with online backups.
    Any help is appreciated!

    Hi,
    did you create the other link according to the upgrade guide (point 5.5) ?
    You need to create this link so that BR*Tools
    (which is linked to the Oracle 10.2 client)
    can use the newer Oracle 11.2 clients.
    After the software installation has finished,
    create a symbolic link in $ORACLE_HOME/lib as follows:
    cd $ORACLE_HOME/lib
    ln u2013s libnnz11.so libnnz10.so
    or (HP-UX)
    ln u2013s libnnz11.sl libnnz10.sl
    Volker
    Second part: For the DP job, there can be set different environment settings inside the DP job.
    Did you check if there are still relicts from the previous release ?
    You could integrate a small pre-exec script into the job and let it plot the output of
    id
    env
    set
    to a logfile, just to see which values the backup is using.
    Edited by: Volker Borowski on Apr 29, 2011 10:12 AM

  • Upgrade to Oracle 10gR2 from 8i

    Hello, i'm going to install Oracle 10gR2 on Sun Solaris SPARC64 and i have to migrate user schema and related application software in Pro*C language.
    Which are the issues to consider in order to avoid compilation errors and/or run time errors.
    Thanks for your help.

    Direct upgrade to 10g is possible from the latest 8i version (8.1.7.4). In case you are below this patch level, first you must upgrade to this and then upgrade to 10g.
    In case 32 bit conversions are involved, you must consider the utlip.sql script which must be run during the upgrade phase.
    I suggest you to read Metalink Note 316889.1 where a Complete checklist for manual upgrades to 10gR2 is detailed.
    ~ Madrid
    key{ upg  10g }

  • Problem after upgradation from Oracle 8.1.7 to Oracle 9i

    Hello ,
    I currently installed Oracle 9i, and i m using Mapx 4.5,
    and in the executable I am adding serverlayer
    (a layer formed by querying the geoloc column from oracle
    table), which was working fine with Oracle 8.1.7, but after
    installation of Oracle 9i , i encounter the following error
    while adding a serverlayer
    ORA-13373 Element of type Extent is not supported for Geodetic
    data
    Cause: Element type extent for a polygon geometry is not
    supported for
    geodetic data.
    Action: Convert the extent type polygon to a regular 5 point
    polygon
    and set
    etype accordingly.
    ORA-06512 at string line string
    Cause: Backtrace message as the stack is unwound by unhandled
    exceptions.
    Action: Fix the problem causing the exception or write an
    exception
    handler
    for this condition. Or you may need to contact your application
    administrator or database administrator
    Can anybody tell me what's going wrong ?
    As I urgently got to fix it.
    Thanks in advance
    Binoy.

    Hi,
    Oracle9i has a geodetic data model. Your data has a geodetic
    srid, which means that the data will need to conform to the rules
    associated with geodetic data. These rules include:
    No optimized rectangles are supported (my guess is the app is
    using an optimized rectangle as a query window, but the data
    could be the issue also).
    No arcs or circle types are supported.
    A single line segment cannot be equal to or greater than 1/2 the
    great circle distance.
    A polygon cannot have an area equal to or greater than 1/2 the
    surface area of the earth.
    You should try to find what version of Mapx supports Oracle9i.
    Regards,
    dan

Maybe you are looking for

  • Access Another Class' Property Value That is Defined in the faces-config?

    I have a private property "dataTransferType" with public getter and setter in the DataFile class. The initial value of the "dataTransferType" is defined in the faces-config.xml file. I want to get the "initial value" of the "dataTransferType" in anot

  • Set  filter value by other filter prompt value in Webi on top of BW

    Dear All,    In webi report which on top of BW cube, I create a filter with prompt for YearMonth selection.     Is there any way to create another filter for Year which based on the select Year month?   For example , user input YM = 200903 --> Year =

  • How to restrict the Cost Center Search ?

    Hi everybody, Im a little bit worried... In my Org Model, I defined in the CNT user's attribute some Cost Centers.  I was thinking that this action would restrict the Cost Centers that the end-user can use.  Obviously not.. it seems that it is just a

  • Select Options for Field

    Hi all, I have a requirement to make a select options low field disable and high field editable. Can anyone guide me on this.......!!! Awiting u r reply. Regards, Rashmi

  • Am i having Kernal Panics?

    Every so often an application stops responding and must be force quit. After it is force quit a window pops up saying ______ quit unexpectedly. This is normal. What isn't normal is when this happens to OS X. The screen suddenly goes black and the fan