Logical standby and Primary keys

Hi All,
Why primary keys are essential for creating logical standby database? I have created a logical standby database on testing basis without having primary keys on most of the tables and it's working fine. I have not event put my main DB in force logging mode.

I have not event put my main DB in force logging mode. This is because, redo log files or standby redo logfiles transforms into set of sql statements to update logical standby.
Have you done any DML operations with nologging options and do you notice any errors in the alert.log? I just curious to know.
But I wanted to know that, while system tablespace in hot backup mode,In the absence of both a primary key and a nonnull unique constraint/index, all columns of bounded size are logged as part of the UPDATE statement to identify the modified row. In other words, all columns except those with the following types are logged: LONG, LOB, LONG RAW, object type, and collections.
Jaffar

Similar Messages

  • Doubt regarding switchover from logical standby to primary

    I am using Oracle 10g databases on Linux environment
    I was trying creating a data guard configuration and once I was successfully done I created a logical standby database from the physical standby . Then I was trying to do a switchover .
    By mistake I gave the prepare switchover command in the logical standby before executing it in the primary . And then when I gave it in the primary both the status turned to be "PREPARING SWITCHOVER"
    In logical standby
    SQL> ALTER DATABASE PREPARE TO SWITCHOVER TO PRIMARY;
    Database altered.
    SQL>  SELECT SWITCHOVER_STATUS FROM V$DATABASE;
    SWITCHOVER_STATUS
    PREPARING SWITCHOVER
    In primary
    SQL> ALTER DATABASE PREPARE TO SWITCHOVER TO LOGICAL STANDBY;
    Database altered.
    SQL> select switchover_status from v$database;
    SWITCHOVER_STATUS
    PREPARING SWITCHOVERNow I am not able to commit any of them to primary or logical standby
    In primary
    SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO LOGICAL STANDBY;
    ALTER DATABASE COMMIT TO SWITCHOVER TO LOGICAL STANDBY
    ERROR at line 1:
    ORA-16217: prepare to switchover has not completed
    ORA-06512: at line 1
    In logical standby
    SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO PRIMARY;
    ALTER DATABASE commit  TO SWITCHOVER TO PRIMARY
    ERROR at line 1:
    ORA-16109: failed to apply log data from previous primaryLet me know what should be done now ? Is there anyway to solve this issue ?

    Can you try using the following first on Primary and then on Standby?
    ALTER DATABASE PREPARE TO SWITCHOVER CANCEL;

  • Local logical standby and remote physical standby

    Dear All,
    My database version is 10.2.0.1.0(production) and Operating System Linux x86-64 OS Version Red Hat Enterprise 4
    Below is my configuration details:
    Rprimary Database:Two Nodes RAC
    StandBy Database LOCAL:Standalone database
    Standby Database Remote:Two Node RAC
    So,my question is, i want my local physical DR to convert into logical standby and
    Remote DR will operate as physical standby as it IS.
    sql>EXECUTE DBMS_LOGSTDBY.BUILD
    After executing the above statement in my production primary database will there be any performance HIT.
    And another important is that my above statement shouldn't affect my remote physical DR database.
    Just the local DR should get converted into logical standby database
    Thanks and Regards
    Monoj Das

    monoj wrote:
    Dear All,
    My database version is 10.2.0.1.0(production) and Operating System Linux x86-64 OS Version Red Hat Enterprise 4
    Below is my configuration details:
    Rprimary Database:Two Nodes RAC
    StandBy Database LOCAL:Standalone database
    Standby Database Remote:Two Node RAC
    So,my question is, i want my local physical DR to convert into logical standby and
    Remote DR will operate as physical standby as it IS.
    sql>EXECUTE DBMS_LOGSTDBY.BUILD
    After executing the above statement in my production primary database will there be any performance HIT.
    And another important is that my above statement shouldn't affect my remote physical DR database.
    Just the local DR should get converted into logical standby database
    Thanks and Regards
    Monoj DasIn any Data Guard environment, you can have either physical or logical or even mixed[physical + logical]
    There will be no performance issue if you run the package for logical standby DBMS_LOGSTDBY.BUILD , This package provides you to manage even logical standby database.

  • Difference between Unique key and Primary key(other than normal difference)

    Hello,
    1).Can any one tell me any other difference between Unique key and Primary key other than it having NULLs.
    2). What is the difference the words 'DISTINCT' and 'UNIQUE' in a sql query.
    Thanks in advance.

    Hi
    If you don't believe me than see the documentation in
    OTN.
    Ott Karesz
    http://www.trendo-kft.hu
    SQL> create table scott.tbl_clob
      2  (sss CLOB)
      3  /
    Tabelle wurde angelegt.
    SQL> insert into scott.tbl_clob values('wrwrwrw')
      2  /
    1 Zeile wurde erstellt.
    SQL> insert into scott.tbl_clob values('wrwrwrw')
      2  /
    1 Zeile wurde erstellt.
    SQL> select distinct sss from scott.tbl_clob
      2  /
    select distinct sss from scott.tbl_clob
    FEHLER in Zeile 1:
    ORA-00932: nicht übereinstimmende Datentypen
    SQL> select unique sss from scott.tbl_clob
      2  /
    select unique sss from scott.tbl_clob
    FEHLER in Zeile 1:
    ORA-00932: nicht übereinstimmende Datentypen
    SQL> select distinct to_char(sss) from scott.tbl_clob
      2  /
    TO_CHAR(SSS)
    wrwrwrw
    SQL> select unique to_char(sss) from scott.tbl_clob
      2  /
    TO_CHAR(SSS)
    wrwrwrw
    SQL>

  • Dynamic SQL Joining between tables and Primary keys being configured within master tables

    Team , Thanks for your help in advance !
    I'm looking out to code a dynamic SQL which should refer Master tables for table names and Primary keys and then Join for insertion into target tables .
    EG:
    INSERT INTO HUB.dbo.lp_order
    SELECT *
    FROM del.dbo.lp_order t1
    where not exists ( select *
    from hub.dbo.lp_order tw
    where t1.order_id = t2.order_id )
    SET @rows = @@ROWCOUNT
    PRINT 'Table: lp_order; Inserted Records: '+ Cast(@rows AS VARCHAR)
    -- Please note Databse names are going to remain the same but table names and join conditions on keys
    -- should vary for each table(s) being configured in master tables
    Sample of Master configuration tables with table info and PK Info :
    Table Info         
    Table_info_ID    Table_Name    
    1        lp_order    
    7        lp__transition_record    
    Table_PK_Info        
    Table_PK_Info_ID    Table_info_ID    PK_Column_Name
    2                1    order_id
    8                7    transition_record_id
    There can be more than one join condition for each table
    Thanks you !
    Rajkumar Yelugu

    Hi Rajkumar,
    It is glad to hear that you figured the question out by yourself.
    There's a flaw with your while loop in your sample code, just in case you hadn't noticed that, please see below.
    --In this case, it goes to infinite loop
    DECLARE @T TABLE(ID INT)
    INSERT INTO @T VALUES(1),(3),(2)
    DECLARE @ID INT
    SELECT @ID = MIN(ID) FROM @T
    WHILE @ID IS NOT NULL
    PRINT @ID
    SELECT @ID =ID FROM @T WHERE ID > @ID
    So a cursor would be the appropriate option in your case, please reference below.
    DECLARE @Table_Info TABLE
    Table_info_ID INT,
    Table_Name VARCHAR(99)
    INSERT INTO @Table_Info VALUES(1,'lp_order'),(7,'lp__transition_record');
    DECLARE @Table_PK_Info TABLE
    Table_PK_Info_ID INT,
    Table_info_ID INT,
    PK_Column_Name VARCHAR(99)
    INSERT INTO @Table_PK_Info VALUES(2,1,'order_id'),(8,7,'transition_record_id'),(3,1,'order_id2')
    DECLARE @SQL NVarchar(MAX),
    @ID INT,
    @Table_Name VARCHAR(20),
    @whereCondition VARCHAR(99)
    DECLARE cur_Tabel_Info CURSOR
    FOR SELECT Table_info_ID,Table_Name FROM @Table_Info
    OPEN cur_Tabel_Info
    FETCH NEXT FROM cur_Tabel_Info
    INTO @ID, @Table_Name
    WHILE @@FETCH_STATUS = 0
    BEGIN
    SELECT @whereCondition =ISNULL(@whereCondition+' AND ','') +'t1.'+PK_Column_Name+'='+'t2.'+PK_Column_Name FROM @Table_PK_Info WHERE Table_info_ID=@ID
    SET @SQL = 'INSERT INTO hub.dbo.'+@Table_Name+'
    SELECT * FROM del.dbo.'+@Table_Name+' AS T1
    WHERE NOT EXISTS (
    SELECT *
    FROM hub.dbo.'+@Table_Name+' AS T2
    WHERE '+@whereCondition+')'
    SELECT @SQL
    --EXEC(@SQL)
    SET @whereCondition = NULL
    FETCH NEXT FROM cur_Tabel_Info
    INTO @ID, @Table_Name
    END
    Supposing you had noticed and fixed the flaw, your answer sharing is always welcome.
    If you have any question, feel free to let me know.
    Eric Zhang
    TechNet Community Support

  • Unique and primary key

    column with unique constraint + not null constraint = primary key! (to some extent) Is it correct?
    I invite your ideas

    http://www.techonthenet.com/oracle/unique.php
    http://www.allapplabs.com/interview_questions/db_interview_questions.htm#q13
    Difference between Unique key and Primary key(other than normal difference)

  • Table and Primary Key Case Sensitive in Automated Row Fetch

    Why are the table and primary key fields case sensitive in the Automated Row Fetch process? I'm debugging an error in this process and I'm not sure what the case should be.

    Russ - It's a defect of sorts. Use upper case in the process definition. I don't think these processes will work with tables or columns whose names are not known to the database as upper-case values, i.e., defined using double-quoted non-upper case.
    Scott

  • Logical Standby and Streams

    Hi,
    I am considering different replication scenarios for our future system and have a (maybe stupid:) question. Is it technically possible to create a logical standby database (Data Guard) and than replicate it further using Streams?
    Regards,
    Tim

    Yes , This is possible with logical standby but you need to have mechanisms to take care of role transitioning in case of fail over. See Oracle document http://download.oracle.com/docs/cd/E14072_01/server.112/e10700/whatsnew.htm (Key word search "streams")
    Thanks
    http://swervedba.wordpress.com/
    Edited by: swervedba on May 30, 2011 8:48 PM

  • Fast Start Failover has broken my Logical Standby and Flashback has not helped

    I have a primary ( PRIMARYP ), a physical standby ( STANDBYP ) and a logical standby ( STANDBYL ).
    I configured Fast Start Failover ( with STANDBYP as the target ) and started an Observer. I did a shutdown abort on my primary and the Fast Start Failover did make STANDBYP the primary. I was able to reinstate PRIMARYP as a physical standby.
    However when I checked my broker config it gave for STANDBYL
    ORA-16795: the standby database needs to be recreated
    Strange I though, since I would have expected STANDBYL to survive ok given that the new primary, STANDBYP was a physical standby of PRIMARYP.
    Not to worry I thought, I simply will flashback STANDBYL to the beginning of the day and SQL APPLY will reapply all the redo.
    This is where the problem begins
    I mount STANDBYL
    RMAN> connect target /
    connected to target database: STANDBYL (DBID=440422466, not open)
    RMAN> flashback database to time "TO_DATE('05/12/13 09:30:00','DD-MM-YY HH24:MI:SS')";
    Starting flashback at 05-DEC-13
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=11 device type=DISK
    starting media recovery
    archived log for thread 1 with sequence 24 is already on disk as file /home/app/oracle/fast_recovery_area/STANDBYL/archivelog/2013_12_04/o1_mf_1_24_99yqwog0_.arc
    archived log for thread 1 with sequence 25 is already on disk as file /home/app/oracle/fast_recovery_area/STANDBYL/archivelog/2013_12_05/o1_mf_1_25_9b0mlyvr_.arc
    media recovery complete, elapsed time: 00:00:04
    Finished flashback at 05-DEC-13
    RMAN> quit
    Recovery Manager complete.
    oraclel[STANDBYL]>dgmgrl
    DGMGRL for Linux: Version 11.2.0.3.0 - Production
    Copyright (c) 2000, 2009, Oracle. All rights reserved.
    Welcome to DGMGRL, type "help" for information.
    DGMGRL> connect sys/manager
    Connected.
    DGMGRL> shutdown;
    ORA-01109: database not open
    Database dismounted.
    ORACLE instance shut down.
    DGMGRL>
    DGMGRL>
    DGMGRL> startup;
    ORACLE instance started.
    Database mounted.
    ORA-01589: must use RESETLOGS or NORESETLOGS option for database open
    ORA-01589 - that a nasty surprise ! I didn't think I would need to do anything with the log sequences after a flashback
    Any ideas of what is going on ?
    thanks,
    Jim

    Hello;
    This pdf might meet your needs :
    http://blogs.oracle.com/AlejandroVargas/resource/DATAGUARD-FSFO-IMPLEMENTATION.pdf
    Additional info :
    http://oracle-tech.blogspot.com/2008/08/fast-start-failover-it-is-reliable.html
    http://jakub.wartak.pl/blog/?p=80
    http://www.databasejournal.com/features/oracle/article.php/3849106/Fast-Start-Failover-in-Oracle-11g-Data-Guard.htm
    Best Regards
    mseberg

  • Interesting issue with Logical Standby and database triggers

    We have a Logical Standby that each month we export (expdp) a schema (CSPAN) that is being maintained by SQL Apply and import (impdp)it to a 'frozen copy' (eg CSPAN201104) using REMAP_SCHEMA.
    This works fine although we've noticed that because triggers on the original schema being exported have the original schema (CSPAN) hard-referenced in the definition are imported into and owned by the new 'frozen' schema but are still 'attached' to the original schema's tables.
    This is currently causing the issue where the frozen schema trigger is INVALID and causing the SQL Apply to fail. This is the error:
    'CSPAN201104.AUD_R_TRG_PEOPLE' is
    invalid and failed re-validation
    Failed SQL update "CSPAN"."PEOPLE" set "ORG_ID" = 2, "ACTIVE_IND" = 'Y', "CREATE_DT" = TO_DATE('22-JUL-08','DD-MON-RR'),"CREATOR_NM" = 'LC', "FIRST_NM" = 'Test', "LAST_PERSON" ='log'...
    Note: this trigger references the CSPAN schema (...AFTER INSERT ON CSPAN.PEOPLE...)
    I suspect that triggers on a SQL Apply Maintained schema in a Logical Standby do not need to be valid (since they do not fire) but what if they reference a SQL Apply schema but are 'owned' by a non-SQL Apply schema? This trigger references a SQL Apply table so it should not fire
    This is 10gR2 (10.2.0.4) on 64 bit Windows.
    Regards
    Graeme King

    OK, I've finally got around to actually test this and it looks like you are not quite correct Larry in this statement...
    'Since this trigger belongs to a new schema that is not controlled by SQL Apply (CSPAN201105) it will fire. But the trigger references a schema that is controlled by SQL Apply (CSPAN) so it will fail because it has to be validated.'
    My testing concludes that even though the trigger belongs to a schema CSPAN201105 (not controlled by SQL Apply) and references a schema controlled by SQL Apply - it does not fire. However it DOES need to be valid or it breaks SQL Apply.
    My testing was as follows:
    Primary DB
    Create new EMP table in CSPAN schema on Primary
    Create new table TRIGGER_LOG in CSPAN schema on Primary
    Create AFTER INSERT/UPDATE trigger on CSPAN.EMP table (that inserts into TRIGGER_LOG table)
    **All of the above replicates to Standby**
    Standby DB
    Create new table TRIGGER_LOG_STNDBY in CSPAN201105 schema on Primary
    Create new trigger in CSPAN201105 schema that fires on INSERT/UPDATE on CSPAN.EMP but that inserts into CSPAN201105.TRIGGER_LOG_STNDBY table)
    Primary DB
    Insert 4 rows into CSPAN.EMP
    Update 2 rows in CSPAN.EMP
    TRIGGER_LOG table has 6 rows as expected
    Standby DB
    TRIGGER_LOG table has 6 rows as expected
    TRIGGER_LOG_STNDBY table has **0 rows**
    Re-create trigger in CSPAN201105 schema that fires on INSERT/UPDATE on CSPAN.EMP but that inserts into CSPAN201105.TRIGGER_LOG_STNDBY table) **but with syntax error**
    Primary DB
    Update 1 row in CSPAN.EMP
    TRIGGER_LOG table has 7 rows as expected
    Standby DB
    SQL Apply is broken - ORA-04098: trigger 'CSPAN201105.TEST_TRIGGER_TRG' is invalid and failed re-validation

  • NEED HELP!!!!  Logical Standby and RMAN.

    Hello,
    I have a Data Guard env setup on win 2000 with 9.2.0.4.
    I now have primary and standby. I was wondering if I create a logical standby to achieve transparent application failover. What I am saying is if the primary node dies, then in tnsnames I want to point to the logical standby, since it is an open database, the users can still perform without knowing the db went down. Then if we indeed need a failover we can failover to the physical standby. This would ensure availability without having to implement RAC or AQ, Right? Also, I read that you should create rman schema (recovery catalog) on standby db. Any thoughts?

    Were you able to figure this one out?
    I have a similar configuration as you - the same motherboard and an eVga Geforce 3 6600 fanless.  With the latest BIOS (3.8) and this video card, the system will not go into Standby completely.  I hear the HDD head park and the video goes black, but the LEDs on the front of the PC chassis and the case/PS fans do not stop.  Pressing the button to bring the PC out of Standby, rather, whatever state it is in, works.

  • Logical standby and truncate partition

    Hi,
    I'm evaluation whether a logical standby database would meet our needs.
    We have a live database and want a reporting database that is identical to the live one, just minutes behind in time to the live one, plus we want to create other summary tables etc on the reporting db.
    Logical standby seems to meet our needs but I have one query.
    - On the live db the most of the tables are organized by date partitions and only 5 days are kept with new partitions being created every night for the forthcoming days and the oldest date partitions being truncated.
    On the reporting database we want to keep 30 days partitions.
    Can we have all the DDL and DML from live being applied to the standby APART from the specific truncate partition statements?
    Many Thanks,
    Kailas

    Addendum.
    If you can , it is better to truncate the partition but not drop it, after exporting the partition for archive purposes. If you ever need to bring the data back, the data will go into the correct partition . If the table structure has been changed by adding or lengthening columns, this will still work OK where column names match.
    Test ease of restoration of archived data to existing and non-existing partitions with altered structure (dropped columns,added columns, renamed columns) for yourself though.
    Regards, Vin.
    PS. When splitting maxdata partition to create new 'highest' partition, use a small initial extent then the 'next' being the expected real size needed. When the partition is truncated, the only space that should remain un-claimable will be that allocated for the initial extent.

  • What is the diffrence between Row id and primary key ?

    dear all
    my question is about creating materialized views parameters (With Rowid and
    With Primary kry)
    my master table contains a primary key
    and i created my materialized view as follow:
    CREATE MATERIALIZED VIEW LV_BULLETIN_MV
    TABLESPACE USERS
    NOCACHE
    LOGGING
    NOCOMPRESS
    NOPARALLEL
    REFRESH FAST ON DEMAND
    WITH PRIMARY KEY
    AS
    SELECT
    BCODE ID, BTYPE BTYPE_ID,
    BDATE THE_DATE,SYMBOL_CODE STOCK_CODE,
    BHEAD DESC_E, BHEADARB DESC_A,
    BMSG TEXT_E, BMSGARB TEXT_A,
    BURL URL, BTIME THE_TIME
    FROM BULLETIN@egid_sefit;
    I need to know is there a diffrence between using (with row id) and (with primary key) on the performance of the query?

    Hi again,
    fast refreshing complex views based on rowids, according to the previous subject.
    (You're example shows that) are not possible.
    Complex remote (replication) snapshots cannot be based on Rowid too.
    for 10.1
    http://download-west.oracle.com/docs/cd/B14117_01/server.101/b10759/statements_6002.htm#sthref5054
    for 10.2
    http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14200/statements_6002.htm#sthref6873
    So I guess (didn't check it) that this applies ONLY to replication snapshots.
    This is not documented clearly though (documentation bug ?!)
    Documentation states that the following is generally not possible with Rowid MVIEWS:
    Distinct or aggregate functions
    GROUP BY or CONNECT BY clauses
    Subqueries
    Joins
    Set operations
    Rowid materialized views are not eligible for fast refresh after a master table reorganization until a complete refresh has been performed.
    The main purpose of my statements was to try to give a few tips how to avoid common problems with this complex subject, like for example: being able to CREATE an MVIEW with fast refresh clause does not really guarantee that it will refresh fast in the long run (reorganisation, partition changes) if ROWID based, further the rowid mviews have limitations according to the documentation (no group by, no connect by, link see above) plus fast refresh means only to use filter columnns of the mview logs, plus for aggregates you need additional count (*) pseudo columns.
    kind regards
    Karsten

  • Problem with type conversion and primary key during row fetch

    [Note, this error occurs in Oracle XE, APEX 2.1.0.00.39]
    I have been having a problem with Automatic Row Fetch:
    ORA-01460: unimplemented or unreasonable conversion requested
    So in an effort to resolve this, I created a PL/SQL process with the query below and was able to isolate the same error. The problem seems to come from one of the primary keys being a "number" type (APP_ID). The error occurs on the line:
    where u.app_id=:P5_APP_ID
    I have tried the following variations on this line in an effort to resolve this, but all generate the same error:
    1) where to_char(u.app_id) = :P5_APP_ID
    2) where u.app_id = to_number(:P5_APP_ID)
    3) where to_char(u.app_id) = to_char(:P5_APP_ID)
    I've also tried the laternate syntax "&__." and "#__#", but these don't function in the Source field and show up as syntax errors.
    Any suggestions are welcome.
    begin
    for r in (
    select app_name, apptype, appcreator, appurl
    from application_users u, application_info i
    where u.app_id=:P5_APP_ID
    and i.app_id=u.app_id
    and u.username=:P5_USERNAME)
    loop
    begin
    :P5_APP_NAME := r.app_name;
    :P5_APPURL := r.appurl;
    exception
    when others then
    raise_application_error(-20000,'In Loop Failure',true);
    end;
    end loop;
    exception
    when others then
    raise_application_error(-20000,'Out of Loop Failure',true);
    end;
    Thanks in advance,
    Barney

    I found a prior post referencing a similar issue and it was solved by using the "v(__)" syntax. This did resolve my issue, however, I have a quick follow-on which I'm hoping someone can answer quickly...
    Since the "v(__)" syntax won't work for the Automatic Row Fetch (at least to the best of my knowledge), I have to do a manual process. However, the manual query as shown above doesn't actually populate any of the form values through the bind variables. They all remain at their cached values from prior data entry on the form.
    Is using the bind variables for assignment incorrect, or is there something that must be done to get the updates to show up in the form (ordering of processes, etc.).
    My manual process is running in the Load: Before Header state so I would have expected it to update all of the fields.
    Thanks in advance,
    Barney

  • Null in Composite Primary Key and "Primary keys must not contain null"

    Hello all.
    I'm a newbie concerning to JPA/EJB3, but I was wondering if toplinks doesn't support composite primary keys with null in some field (something perfectly right in any RDBMS).
    I used JDeveloper (I'm using Oracle 10g database and JDeveloper 10.1.3.2.) wizards to generate JPA classes and I checked out generated files (with annotations), so they should be right (by the way, other O-R mappings for my model are working right, but this one).
    I'm getting the next error:
    Exception Description: The primary key read from the row [DatabaseRecord(
         TSUBGRUPOSLDI.CD_GRUP => 01
         TSUBGRUPOSLDI.CD_SUBGRUP => null
         TSUBGRUPOSLDI.CG_POBL => 058
         TSUBGRUPOSLDI.CG_PROV => 28
         TSUBGRUPOSLDI.DSCR => Sanidad)] during the execution of the query was detected to be null. Primary keys must not contain null.
    Compound primary key is (CD_GRUP, CD_SUBGRUP). No foreign keys, no joins (only a NamedQuery: "select o from ..."). It's the simplest case!
    I checked out that everything runs ok if there's no "null" value in CD_SUBGRUP.
    After some research (this and other forums) I'm beginning to believe that it's not supported, but not sure.
    Am I doing sth wrong? If not, what is the reason to not support this? Will it be supported in the future?
    Thanks in advance.

    Null is a special value and in many databases is not comparable to another null value (hence the isNull operator) and may pose problems when used to uniquely identify an Entity. TopLink does not support null values within a composite PK. As the nullable column is most likely not designated as a PK within your database table (many databases do not allow this) I recommend updating the Entity PKs to match that of the database.
    --Gordon                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

Maybe you are looking for

  • Calling BI Publisher report  using FND_SUBMIT.SUBMIT_REQUEST---

    Hi Experts, We have a requirement in which BI Publisher report will be called using a PL/SQL procedure. i am successfully able to call BI publisher report using FND_SUBMIT.SUBMIT_REQUEST function. Business requirement is that the BI publisher report

  • IPad Switzerland and iTunes store restrictions

    Hello, I have just bought myself a iPad 32g wifi and I'm a bit disappointed. I can't seem to even access the area to rent a movie or tv show. Does Switzerland restrict all movies? Is there any other way I can rent movies or do I have to live in the U

  • Account Expired

    We are using the Adobe Connect Host service and payment is by credit card. Before, we did not receive any payment problem e-mail from Adobe. Today, suddently, when we want to login and start up a meeting, it shows "Your Account Has Expired"!!! We can

  • Does SAP ECC 6.0  run on Oracle 11.2 in Windows 2008 server ( win NT 6.1) ?

    HI all, Does SAP ECC 6.0  run on Oracle 11.2 in Windows 2008 server ( win NT 6.1) ? Thanks in Advance Sapsanthose

  • 11gr2 and 10gr2 in a single machine

    Hi, Already i have installed 10g release2 database in my system in OEL5 Now i planned to install 11g release2 also in the same system In future i also need to upgrade my 10g database to 11g If so, What are the changes or configuration that i should m