PS Datamove fails on one table (pspcmprog)

Thanks for taking my question!
We are trying to install peopletools 8.49 on demo database 10.2.0.3 10g 64bit server and for some reason one table will not load. The job fails at just about the same place - after inserting about 6000. The error is below. Any help or suggestion would be great!!
Thanks, Kathie
Started: Tue Jul 31 13:08:27 2007
Data Mover Release: 8.49.03
Database: CSDMO (ENG)
Indexing export file P:\csdemo\data\hcengs.db
Input file: P:\csdemo\data\hcengs.db (ENG)
Importing PSPCMPROG
Rows inserted into PSPCMPROG
50, 100, 150, 200, 250, 300, 350, 400, 450, 500, 550, 600, 650, 700, 750,
800, 850, 900, 950, 1000, 1050, 1100, 1150, 1200, 1250, 1300, 1350, 1400,
1450, 1500, 1550, 1600, 1650, 1700, 1750, 1800, 1850, 1900, 1950, 2000,
2050, 2100, 2150, 2200, 2250, 2300, 2350, 2400, 2450, 2500, 2550, 2600,
2650, 2700, 2750, 2800, 2850, 2900, 2950, 3000, 3050, 3100, 3150, 3200,
3250, 3300, 3350, 3400, 3450, 3500, 3550, 3600, 3650, 3700, 3750, 3800,
3850, 3900, 3950, 4000, 4050, 4100, 4150, 4200, 4250, 4300, 4350, 4400,
4450, 4500, 4550, 4600, 4650, 4700, 4750, 4800, 4850, 4900, 4950, 5000,
5050, 5100, 5150, 5200, 5250, 5300, 5350, 5400, 5450, 5500, 5550, 5600,
5650, 5700, 5750, 5800, 5850, 5900, 5950, 6000, 6050, 6100, 6150, 6200,
6250, 6300, 6350, 6400, 6450, 6500, 6550, 6600, 6650, 6700, 6750, 6800,
6850, 6900,File: Data MoverSQL error. Stmt #: 0 Error Position: 0 Return: 3135 - ORA-03135: connection lost contact
Failed SQL stmt:INSERT INTO PSPCMPROG (OBJECTID1, OBJECTVALUE1, OBJECTID2, OBJECTVALUE2, OBJECTID3, OBJECTVALUE3, OBJECTID4, OBJECTVALUE4, OBJECTID5, OBJECTVALUE5, OBJECTID6, OBJECTVALUE6, OBJECTID7, OBJECTVALUE7, PROGSEQ, VERSION, NAMECOUNT, PROGLEN, PROGRUNLOC, PROGFLAGS, LICENSE_CODE, LASTUPDDTTM, LASTUPDOPRID, PROGEXTENDS, PROGTXT) VALUES (:1, :2, :3, :4, :5, :6, :7, :8, :9, :10, :11, :12, :13, :14, :15, :16, :17, :18, :19, :20, :21, TO_TIMESTAMP(:22,'YYYY-MM-DD-HH24.MI.SS.FF'), :23, :24, :25)
Error: Unable to insert row 37
Error: SQL execute error for PSPCMPROG
Ended: Tue Jul 31 13:11:35 2007
Unsuccessful completion

03135, 00000, "connection lost contact"
// *Cause:  1) Server unexpectedly terminated or was forced to terminate.
//          2) Server timed out the connection.
// *Action: 1) Check if the server session was terminated.
//          2) Check if the timeout parameters are set properly in sqlnet.ora.It looks like either someone, or something is killing your session. Are you running this across a network? If so, you may be either dropping the network connection or possibly it is really slow and is timing out.
John

Similar Messages

  • Failed to upgrade more than one table at same time

    Hi
    In Deployment Manager, I failed to upgrade more than one table at same time.
    I tried to hightlight 4 tables and set the default action as Upgrade, and click File/Generate Deploy. It passed the code generation step then I click Deploy, they are all failed with no error message.
    But they are all successful when I upgrade them one by one. Any one has any idea about this?
    For the known reason, we have no choice to do the deployment with action of 'upgrade' through OMBPlus, instead, can only do that interactively through OWB Client. I can't imagine to ask our Production side DBA to upgrade 80 tables one by one. Or I have to use the generated scripts to do the upgrade, which will resulted in 'no deploy status updated' in OWB. Any help will be very appricated.
    The version I'm using is OWB 10g.
    Thanks,
    Daming

    Hi
    First of all, Patrick's solution doesn't work for me. I didn't do cloning and there is no problem when check the WB tables.
    Second, I think your solution is only good for the developing environment just to get tables upgrade via deployment manager. In most cases, when you do a new release on PROD environment you just exp/imp the MDL file from DEV to PROD and any developing is not recommended on PROD. But your approach is trying to manually the DB, and then EDIT the mapping to do the Reconcile and then deploy. Surely you can do that if you have full control on your PROD side. However, in my situation, I have no access to the PROD for the security reason and an DBA Operator of the Client is responsible to implement my Migration process on PROD by himself.

  • Import for just one table fails....

    Hi,
    On production server i have declared a db user , let's call him PRD_USER, with default tablespace PRD_TBL.
    On development server i have declared a db user , let's call him PRD_USER, with default tablespace DEV_TBL.
    On production server, i use the exp db utility in order to do the import as:
    imp system/manager from user=PRD_USER touser=PRD_USER ignore=Y file_name ='....' log='....'.
    The import succeeds for about 25 tables and indexes and constraints but it fails for just one table with error : {i do not remember the ORA- error, and i have not access to it right now} tablespace DEV_TBL does not exist.
    Of course this tablespace does not exist on production env. but how does this problem arise since the default tablespace for this user is not DEV_TBL but PRD_TBL...????
    Do you have the slightest idea what may be the cause and how can i overcome this problem during import...???{Note: I gave a temporary solution... taking the sql creation script of the table leaving out the reference of tablespace 'DEV_TBL'}.
    Both servers run exactly the same version of DB..
    Note : I use DB 10g v.2
    Thank you,
    Sim

    HI,
    "If the table has Partitions, the import is attempting to create the Partitions (in the CREATE TABLE statement) on the original tablespace."
    The table is not partitioned.....
    I'll check for the second characteristic...(LOB-column on Monday)
    Thank you,
    Sim

  • Loading data from one table to another table

    I need to load the data (data conversion) from one table to 3 different table.
    I have to load the data from Source table to following 3 target tables.
    So Consumer table has email pk and it has relationship with CONSUMER_RCV table and CONSUMER_RCV
    table has relationship with CONSUMER_ATTR table. I am doing right now with pl/sql but I am looking
    to use MULTI INSERT or SQL*LOADER or another efficient way.
    I tried with MultiInsert but it getting failed as it's violating Unique Constraints as source table might contains
    Duplicate emails.
    Could you please show me how I can load the data efficiently?
    Also I need to load in CONSUMER_ATTR table from CONTACT table which has ATTR1, ATTR2 columns.
    so My mapping will be ...
    Contact.email, first_name, last_name will go into Consumer table
    then in CONSUMER_RCV, I generate the id column value by sequence, load email and DATE_CREATEd from contact table
    And rest of the value is hard coded
    Then in CONSUMER_ATTR table I will generate the CONSUMER_ATTR_id through sequence, for ID, will use the same
    id column value which I have generated by sequence in CONSUMER_RCV, then put the attr_type = 'ATTR1' and for
    attr_value column I will insert the value from contact table's ATTR1 value,
    Then generate another CONSUMER_ATTR_ID, will use the same id column value which I have generated by sequence
    in CONSUMER_RCV, then put the attr_type = 'ATTR2' and for
    attr_value column I will insert the value from contact table's ATTR2 value,
    Please let me know if you need further explanation.
    I am using Oracle 9i R2.
    Please consider CONTACT is Source table, CONSUMER is a Target1, CONSUMER_RCV is a Target2 and CONSUMER_ATTR is a Target3
    to simplify the problem.
    Following is the Table structure
    =========================
    CREATE TABLE CONTACT (
    ID VARCHAR2 (40) NOT NULL, -- will go into Target1
    EMAIL VARCHAR2 (100) ,      -- might be duplicate and will go into Target1 and Target2
    FIRST_NAME VARCHAR2 (100) NOT NULL, -- will go into Target1
    LAST_NAME VARCHAR2 (100) NOT NULL, -- will go into Target1
    COUNTRY VARCHAR2 (40) NOT NULL, -- will go into Target1
    PHONE VARCHAR2 (100), -- will go into Target1
    NOTIFY VARCHAR2 (3), -- will go into Target2 table's RECEIVE_EMAIL column
    CREATE_DATE DATE ,               -- will go into Target1 and target2
    ATTR1 VARCHAR2 (400),           -- will go into Target3 as attr_type= ATTR1 and will load actual value in ATTR_VALUE
    ATTR2 VARCHAR2(100),          -- will go into Target3 as attr_type= ATTR2 and will load actual value in ATTR_VALUE
    ATTR2 VARCHAR2(100),          -- will go into Target3 as attr_type= ATTR1 and will load actual value in ATTR_VALUE
    CONSTRAINT CONTACT_PK
    PRIMARY KEY ( USER_ID ) ) ;
    CREATE TABLE CONSUMER(
    EMAIL VARCHAR2 (100) NOT NULL, -- PK
    TITLE VARCHAR2 (40),
    FIRST_NAME VARCHAR2 (40),
    LAST_NAME VARCHAR2 (40),
    ADDRESS1 VARCHAR2 (40),
    ADDRESS2 VARCHAR2 (40),
    CITY VARCHAR2 (30),
    STATE          VARCHAR2 (30),
    ZIP          VARCHAR2 (10),
    COUNTRY VARCHAR2 (40),
    PHONE VARCHAR2 (15),
    DATE_CREATED DATE NOT NULL,
    CONSTRAINT CONSUMER_PK
    PRIMARY KEY ( EMAIL ) ) ;
    CREATE TABLE CONSUMER_RCV (
    ID                VARCHAR2 (40) NOT NULL,-- PK
    EMAIL VARCHAR2 (100) NOT NULL,-- FK reference to Consumer
    SITE VARCHAR2 (100) NOT NULL, -- default website
    RECEIVE_EMAIL VARCHAR2 (1),
    CREATE_DATE DATE NOT NULL,
    CONSTRAINT CONSUMER_RCV_PK
    PRIMARY KEY (ID) ) ;
    ALTER TABLE CONSUMER_RCV ADD CONSTRAINT CONSUMER_RCV_FK
    FOREIGN KEY (EMAIL)
    REFERENCES CONSUMER (EMAIL) ;
    CREATE TABLE CONSUMER_ATTR (
    CONSUMER_ATTR_ID           VARCHAR2 (40) NOT NULL, -- PK
    ID                     VARCHAR2 (40) NOT NULL, -- FK reference to COnsumer_RCV
    ATTR_TYPE           VARCHAR2 (100) NOT NULL,
    ATTR_VALUE           VARCHAR2 (4000) NOT NULL,
    CONSTRAINT CONSUMER_ATTR_PK
    PRIMARY KEY ( CONSUMER_ATTR_ID ) ) ;
    ALTER TABLE CONSUMER_ATTR ADD CONSTRAINT CONSUMER_ATTR_FK
    FOREIGN KEY (CONSUMER_ATTR_ID)
    REFERENCES CONSUMER_RCV (ID) ;

    HI Hema,
      How are the entries related.. is it like for one entry in y table there are more than one entry in x table then you have to use loop with in loop , here are the both the conditions.
    1) For each entry in Y there are more than one entries in X
      sort y by quota trpid.
      sort x by quota trpid.
      loop at y.
      loop at x where quota eq y-quota
                            and trpid eq y-trpid.
      z-value = x-value.
      append z.
      endloop.
      endloop.
    2) For each y there is one entry in x table.
    sort y by quota trpid.
      sort x by quota trpid.
      loop at y.
      read table x with key quota = y-quota
                                trpid = y-trpid binary search.
    if sy-subrc eq 0.
      z-value = x-value.
      append z.
    Endif.
      endloop.
    Mahesh

  • Error while data shifting from one table to another table

    i have two tables.
    in one table i dont have constraints.in anothe rtable i have constraints.
    my requiremnt is i have to shift data from unconstrained table to constrained table
    while shifting data any constaint violation is there system hasto raise error by specifying the error
    in below example it should be ename,esal values not null.
    i tried with save exception but it gives only error.i want error by mentiong that column names also ie ename,esal is not null
    if ename value only null then it should be give ename is null
    create table tablea (eno number(10), ename varchar2(20) not null, esal number(5) not null)
    create table tabl (eno number(10), ename varchar2(20) , esal number(5) )
    DECLARE
    TYPE tabl_tab IS TABLE OF tabl%rowTYPE;
    tab tabl_tab;
    -- create an exception handler for ORA-24381
    ERRORS NUMBER;
    dml_errors EXCEPTION;
    PRAGMA EXCEPTION_INIT (dml_errors, -24381);
    BEGIN
    SELECT *
    BULK COLLECT INTO tab
    FROM tabl
    WHERE rownum<=2;
    -- add '_SR' to the job_id of the most senior employees
    FORALL i IN tab.FIRST .. tab.LAST SAVE EXCEPTIONS
    insert into tablea values tab(i);
    EXCEPTION
    WHEN dml_errors
    THEN
    rollback ; --roll backs all the good records updation
    -- Now we figure out what failed and why.
    ERRORS := SQL%BULK_EXCEPTIONS.COUNT;
    DBMS_OUTPUT.put_line ('Number of statements that failed: ' || ERRORS);
    FOR i IN 1 .. ERRORS
    LOOP
    DBMS_OUTPUT.put_line ( 'Error #'
    || i
    || ' occurred during '
    || 'iteration #'
    || SQL%BULK_EXCEPTIONS (i).ERROR_INDEX
    DBMS_OUTPUT.put_line ( 'Error message is '
    || SQLERRM
    (-SQL%BULK_EXCEPTIONS (i).ERROR_CODE)
    END LOOP;
    END;
    how to implement that.

    it should give error as ename column is not null

  • Mapping a CMP bean to more than one table

    Hello,
    I like to map a CMP bean to more than one table. I am using wlserver7.0 SP2. I
    have a VendorBean with a CMP field "comments" mapped to VENDOR_COMMENTS table.
    All other CMP fields in VendorBean are mapped to VENDOR table. Below is the snippet
    of weblogic-cmp-rdbms-jar.xml:
    <weblogic-rdbms-bean>
    <ejb-name>VendorBean</ejb-name>
    <data-source-name>sweepDataSource</data-source-name>
    <table-map>
    <table-name>VENDOR</table-name>
    <field-map>
    <cmp-field>expIndicator</cmp-field>
    <dbms-column>VENDOR_EXP_IND</dbms-column>
    </field-map>
    <field-map>
    <cmp-field>vendorId</cmp-field>
    <dbms-column>VENDOR_SEQ_KEY</dbms-column>
    </field-map>
    </table-map>
    <table-map>
         <table-name>VENDOR_COMMENTS</table-name>
         <field-map>
    <cmp-field>vendorId</cmp-field>
    <dbms-column>VENDOR_SEQ_KEY</dbms-column>
    </field-map>
         <field-map>
    <cmp-field>comments</cmp-field>
    <dbms-column>VENDOR_COMMENT</dbms-column>
    </field-map>
    </table-map>
    As you can see, "vendorId" is the primary key cmp field mapped to VENDOR_SEQ_KEY
    column in both tables. The relationship between the table rows are 1 to 1. When
    I run ejbc, I get the exception below:
    [ejbc] weblogic.utils.AssertionError: ***** ASSERTION FAILED *****[ Asserti
    on violated ]
    [ejbc] at weblogic.utils.Debug.assertion(Debug.java:74)
    [ejbc] at weblogic.ejb20.cmp.rdbms.RDBMSBean.processRole(RDBMSBean.java
    :1389)
    [ejbc] at weblogic.ejb20.cmp.rdbms.RDBMSBean.processDescriptors(RDBMSBe
    an.java:1105)
    [ejbc] at weblogic.ejb20.cmp.rdbms.Deployer.readTypeSpecificData(Deploy
    er.java:299)
    [ejbc] at weblogic.ejb20.persistence.PersistenceType.setTypeSpecificFil
    e(PersistenceType.java:490)
    [ejbc] at weblogic.ejb20.persistence.PersistenceType.setupDeployer(Pers
    istenceType.java:407)
    [ejbc] at weblogic.ejb20.deployer.CMPInfoImpl.setup(CMPInfoImpl.java:11
    4)
    [ejbc] at weblogic.ejb20.ejbc.EJB20CMPCompiler.generatePersistenceSourc
    es(EJB20CMPCompiler.java:64)
    [ejbc] at weblogic.ejb20.ejbc.EJBCompiler.doCompile(EJBCompiler.java:22
    3)
    [ejbc] at weblogic.ejb20.ejbc.EJBCompiler.compileEJB(EJBCompiler.java:3
    44)
    [ejbc] at weblogic.ejbc20.runBody(ejbc20.java:470)
    [ejbc] at weblogic.utils.compiler.Tool.run(Tool.java:126)
    [ejbc] at weblogic.ejbc20.main(ejbc20.java:711)
    But if I put the "comments" <field-map> inside the VENDOR <table-map> and get
    rid off the <table-map> for VENDOR_COMMENTS, I was able to run ejbc. Of course,
    I get the run time exception during deployment because, there is no column called
    "VENDOR_COMMENT" in VENDOR table. Please help me. Thanks a million.

    Wrong newsgroup, needs to go to one of the ejb newsgroups.
    mbg
    "Sai S Prasad" <[email protected]> wrote in message
    news:[email protected]...
    >
    Hello,
    I like to map a CMP bean to more than one table. I am using wlserver7.0SP2. I
    have a VendorBean with a CMP field "comments" mapped to VENDOR_COMMENTStable.
    All other CMP fields in VendorBean are mapped to VENDOR table. Below isthe snippet
    of weblogic-cmp-rdbms-jar.xml:
    <weblogic-rdbms-bean>
    <ejb-name>VendorBean</ejb-name>
    <data-source-name>sweepDataSource</data-source-name>
    <table-map>
    <table-name>VENDOR</table-name>
    <field-map>
    <cmp-field>expIndicator</cmp-field>
    <dbms-column>VENDOR_EXP_IND</dbms-column>
    </field-map>
    <field-map>
    <cmp-field>vendorId</cmp-field>
    <dbms-column>VENDOR_SEQ_KEY</dbms-column>
    </field-map>
    </table-map>
    <table-map>
    <table-name>VENDOR_COMMENTS</table-name>
    <field-map>
    <cmp-field>vendorId</cmp-field>
    <dbms-column>VENDOR_SEQ_KEY</dbms-column>
    </field-map>
    <field-map>
    <cmp-field>comments</cmp-field>
    <dbms-column>VENDOR_COMMENT</dbms-column>
    </field-map>
    </table-map>
    As you can see, "vendorId" is the primary key cmp field mapped toVENDOR_SEQ_KEY
    column in both tables. The relationship between the table rows are 1 to 1.When
    I run ejbc, I get the exception below:
    [ejbc] weblogic.utils.AssertionError: ***** ASSERTION FAILED *****[Asserti
    on violated ]
    [ejbc] at weblogic.utils.Debug.assertion(Debug.java:74)
    [ejbc] atweblogic.ejb20.cmp.rdbms.RDBMSBean.processRole(RDBMSBean.java
    :1389)
    [ejbc] atweblogic.ejb20.cmp.rdbms.RDBMSBean.processDescriptors(RDBMSBe
    an.java:1105)
    [ejbc] atweblogic.ejb20.cmp.rdbms.Deployer.readTypeSpecificData(Deploy
    er.java:299)
    [ejbc] atweblogic.ejb20.persistence.PersistenceType.setTypeSpecificFil
    e(PersistenceType.java:490)
    [ejbc] atweblogic.ejb20.persistence.PersistenceType.setupDeployer(Pers
    istenceType.java:407)
    [ejbc] atweblogic.ejb20.deployer.CMPInfoImpl.setup(CMPInfoImpl.java:11
    4)
    [ejbc] atweblogic.ejb20.ejbc.EJB20CMPCompiler.generatePersistenceSourc
    es(EJB20CMPCompiler.java:64)
    [ejbc] atweblogic.ejb20.ejbc.EJBCompiler.doCompile(EJBCompiler.java:22
    3)
    [ejbc] atweblogic.ejb20.ejbc.EJBCompiler.compileEJB(EJBCompiler.java:3
    44)
    [ejbc] at weblogic.ejbc20.runBody(ejbc20.java:470)
    [ejbc] at weblogic.utils.compiler.Tool.run(Tool.java:126)
    [ejbc] at weblogic.ejbc20.main(ejbc20.java:711)
    But if I put the "comments" <field-map> inside the VENDOR <table-map> andget
    rid off the <table-map> for VENDOR_COMMENTS, I was able to run ejbc. Ofcourse,
    I get the run time exception during deployment because, there is no columncalled
    "VENDOR_COMMENT" in VENDOR table. Please help me. Thanks a million.

  • [AS] How to test the presence of at least one table?

    Hello everyone,
    I would like to test for the presence of at least one table in a document before starting a process (on edge strokes). 
    I found this, but I do not know if this is really effective:
                                  set CountOfTables1 to count of tables of every story
                                  set CountOfTables2 to every table of every story
    The first gives me a list of the number of table in each story; the second gives me the objects reference of every table.
    Is there another way?
    TIA.
    Oli.

    Marc
    The test I did for nested tables stank (table pasted in rectangle and that rectangle pasted in a table ).  It does not work for nested tables
    I tested .isValid and it's a lot slower.
    Uwe
    Yes, I noticed that difference after posting my eeakk comment.
    Using slice(0) after the getElements can make a big difference but still the simple length going to be quicker in this case.
    Also the getElements without the slice(0) will return "too many elements" if there are to many elements (between 1500 - 15,000).
    On the other hand for repeated access of the variable (looping) it's as know normally quicker to use the getElements().slice(0).length than just length
    In summary
    1) Your anyItem() method is going to be very quick on a document which has a high ratio of stories containing tables.
    2) Although your method could and possibly probably be 100 times quicker than mine I would definitely use my method in terms of typing time and space verses the 1/2 second it might save on execution time per 10,000 tables.
    3) The only accurate method (in this thread) for counting the tables including nested and footnotes is Marc's Grep method.
    So I guess the 3 of us can share first place.
    I just wonder if using the same technique as Marc used in our discussion sometime back on nested buttons might get to a quick count than using the Grep method here.
    Clearly needs to be repeated on different types of document setups one can try the below
    if (!app.properties.activeDocument) {alert ("Really!?"); exit()};
    var scriptCount = 1;
    // Script 1 table.length
    var doc = app.activeDocument,
          start = Date.now(),
          t = doc.stories.everyItem().tables.length,
          finish = Date.now();
    $.writeln ("\rtable.length Script (" + scriptCount++ + ") took " + (finish - start) + "ms\r" + ((t) ? t + " Table" + ((t>1) ? "s" : "") : "Diddlysquat"));
    // Script 2 getElements
    var doc = app.activeDocument;
    var start = Date.now();
    var t = doc.stories.everyItem().tables.everyItem().getElements().length;
    var finish = Date.now();
    $.writeln ("\rgetElements Script (" + scriptCount++ + ") took " + (finish - start) + "ms\r" + ((t) ? t + " Table" + ((t>1) ? "s" : "") : "Diddlysquat"));
    // Script 3 getElements.slice(0)
    var doc = app.activeDocument;
    var start = Date.now();
    var t = doc.stories.everyItem().tables.everyItem().getElements().slice(0).length;
    var finish = Date.now();
    $.writeln ("\rgetElements.slice(0) Script (" + scriptCount++ + ") took " + (finish - start) + "ms\r" + ((t) ? t + " Table" + ((t>1) ? "s" : "") : "Diddlysquat"));
    // Script 4      isValid
    var start = Date.now();
    var t = doc.stories.everyItem().tables.everyItem().isValid;
    var finish = Date.now();
    $.writeln ("\risValid Script (" + scriptCount++ + ") took " + (finish - start) + "ms\rThe document contains " + ((t) ? "tables" : "no tables"));
    // Script 5   Marc's Grep
    var start = Date.now();
    var t = countTables();
    var finish = Date.now();
    $.writeln ("\rMarc's Grep Script as said only accurate one but slow (" + scriptCount++ + ") \rtook " + (finish - start) + "ms\r" + ((t) ? t + " Table" + ((t>1) ? "s" : "") : "Diddlysquat"));
    // Script 6 very lot of anyItem
    var start = Date.now();
    var myResult = atLeastOneTableInDoc(app.documents[0]);
    var finish = Date.now();
    $.writeln ("\rUwes Anyone for Bingo Script (" + scriptCount++ + ") took " + (finish - start) + "ms\rThe document contains " + ((myResult) ? "tables" : "no tables"));
    // Script 7 anyItem length
    var start = Date.now();
    var myResult = detectATable();
    var finish = Date.now();
    $.writeln ("\ranyItem length Script (" + scriptCount++ + ") took " + (finish - start) + "ms\rThe document contains " + ((myResult) ? "tables" : "no tables"));
    // Script 8 anyItem elements length
    var start = Date.now();
    var myResult = detectATable2();
    var finish = Date.now();
    $.writeln ("\ranyItem elements length Script (" + scriptCount++ + ") took " + (finish - start) + "ms\rThe document contains " + ((myResult) ? "tables" : "no tables"));
    //FUNCTION USING anyItem() EXTENSIVELY:
    function atLeastOneTableInDoc(myDoc){
        var myResult = 0;
        if(myDoc.stories.length === 0){return myResult};
        var myStories = myDoc.stories;
        //LOOP length == length of all Story objects
        //using anyItem() for checking length of Table objects
        for(var n=0;n<myStories.length;n++){
            if(anyStory = myStories.anyItem().tables.length > 0){
                myResult = 1;
                return myResult;
        //FALL-BACK, if anyItem() will fail:
    //EDIT:    if(!myResult){
            for(var n=0;n<myStories.length;n++){
                if(myStories[n].tables.length > 0){
                    myResult = 2;
                    return myResult;
    //EDIT:       };
        return myResult;
    }; //END function atLeastOneTableInDoc(myDoc)
    function detectATable(){
        s=app.documents[0].stories;
         if(s.anyItem().tables.length){
            return true; // Bingo
       for(var n=0;n<s.length;n++){
            if(s[n].tables.length) return true
        return false
    function detectATable2(){
        s=app.documents[0].stories;
         if(s.anyItem().tables.length){
            return true; // Bingo
        var sl = app.documents[0].stories.everyItem().getElements().slice(0);
       for(var n=0;n<s.length;n++){
            if(s[n].tables.length) return true
        return false
    function countTables()
        app.findTextPreferences = null;
        app.findTextPreferences.findWhat = "\x16";
        var t = app.properties.activeDocument||0;
        return t&&t.findText().length;
    P.s.
    Marc,
    A bit of homework. I didn't test it on the above but I found the your highRes function trimmer seems to have a large favoritism to the first function to compare.
    I was testing 2 prototypes and swapping the order swapped the result. One can make the prototypes the same and see the time difference.

  • Import command for one table core dumps

    Hi,
    The following import command is causing a core dump;
    $ imp system/***** file=/upload/dbdumps/06-09-19_11.00.01-db.dmp fromuser=ioeadmin tables=(utils) 2>&1 |tee imp.log
    But this one does not:
    $ cat imp.params
    FULL=y
    FILE=/upload/dbdumps/06-09-19_11.00.01-db.dmp
    GRANTS=y
    INDEXES=y
    $ imp system/*****@odinprd parfile=imp.params 2>&1 |tee imp.log
    More info:
    SunOS 5.8 / Oracle 8.1.5
    Thank you for any help on this

    I don't have access to Metalink. What would it say -- "imp tables=" doesn't work in 8i?
    So I have tried the following:
    1. drop the table I want to restore.
    2. run this command to restore all tables and let it fail on those tables that already exist and recreate the one i dropped.
    imp system/code4202@odinprd full=y show=y file=/upload/dbdumps/06-09-27_00.00.00-db.dmp 2>&1 |tee imp.log
    But it doesn't. It says this about the table in question:
    "ALTER SCHEMA = "IOEADMIN""
    "CREATE TABLE "PARTNER" ("PARTNER_ID" VARCHAR2(16) NOT NULL ENABLE, "PARTNER"
    "_TYPE" VARCHAR2(16), "ADDRESS_ID" NUMBER(10, 0), "PAYMENT_ID" NUMBER(10, 0)"
    ", "PRICE_LIST_ID" NUMBER(10, 0), "DEFAULT_CARRIER" VARCHAR2(32), "DEFAULT_M"
    "ETHOD" VARCHAR2(32), "CARRIER_ACCOUNT_NUMBER" VARCHAR2(32), "BILLTYPE" VARC"
    "HAR2(10), "TMB_TYPE" NUMBER) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255"
    " LOGGING STORAGE(INITIAL 10240) TABLESPACE "APP_DATA""
    . . skipping table "PARTNER"
    why is it skip this guy? It does not exist:
    SQL> desc partner;
    ERROR:
    ORA-04043: object partner does not exist
    SQL> desc ioeadmin.partner;
    ERROR:
    ORA-04043: object ioeadmin.partner does not exist

  • Follwing query fails on one database, but the same works on other Database

    Hi,
    The following SQL fails on one database with ORA-01722 but the same works on other database.
    SQL> SELECT i.version, d.name, i.instance_name, i.host_name
    FROM v$instance i, v$database d, v$parameter p 2
    3 WHERE p.name = 'instance_number'
    4 AND DECODE(p.value, 0, 1, p.value) = i.instance_number;
    AND DECODE(p.value, 0, 1, p.value) = i.instance_number
    ERROR at line 4:
    ORA-01722: invalid number
    both databases are at same version (10.2.0.1) both are on same machine.
    Initailly I suspected NLS settings, I am surprised to see NLS settings are also same on both the databases.
    Any thoughts why is it happening

    Thanks for the reply
    Yes, i know that i am comparing varchar against number.
    bu there is no way that the database where the query is working has correct types on both sides as the tables are standard oracle dictionary tables.
    Yes, i tried this earlier, it works
    AND DECODE(p.value, '0', '1', p.value) = to_char(i.instance_number)
    it is very weired to see that error, i feel it something else that causing the trouble.

  • How to use left outer join on more than one table (source)

    Hi all,
    In our project we are converting the Unix shell & SQL scripts into OWB mappings. We are facing problem converting left outer join in OWB. here is one of the example. i have just pasted the FROM and where condition.
    FROM ym_scr t1, branch_finmonth t3
    LEFT OUTER JOIN item_image t2
    ON (t1.branch_no = t2.branch_no
    AND t1.item_no = t2.item_no
    AND t3.to_date between t2.tran_dt and t2.to_dt
    AND t2.to_dt >= '$start_images' ) <<<========= '$start_images' THIS COMES FROM THE UNIX VARAIBLE ,
    We converts the same when we are putting it in OWB join Operator
    INGRP1.branch_no = INGRP2.branch_no(+)
    AND INGRP1.item_no = INGRP2.item_no(+)
    AND INGRP3.to_date between t2.tran_dt and t2.to_dt
    AND INGRP2.to_dt >= INGRP2.start_images
    But as you see in the OWB opreator we can put left join with INGRP1 with INGRP2.
    We can not make same join with INGRP3 & INGRP4 , becoz it failed saying "you cannot left outer join on more than one table/source".
    so overall this OWB code makes incomplete left outer join as per the above ANSI SQL join in Oracle.
    Bcoz of the problem we are getting less number of rows...........
    SO please please help me on this
    Regards
    Ashok

    Hi.
    I know this topic is here for a while now, but I had the same problem today, searched for help and didn´t find anything.
    Later I figured out how to do this.
    You just have to put the (+) one time, and the OWB converts in an LEFT OUTER JOIN statement.
    For this case, all you have to do is:
    ( INGRP1.branch_no = INGRP2.branch_no(+) AND INGRP1.item_no = INGRP2.item_no AND INGRP2.to_dt >= INGRP2.start_images )
    Regards,
    Godoy

  • HR Execution plan failing because target table needs to be extended?

    I'm going a fresh install of OBIEE 10.1.3.4.1 with Dac Build AN 10.1.3.4.1.20090415.0146. This is first one so is it common to have this issue or is something else going on that I just can't see? Should I just enable auto extend on this one table or could there be corruption causing this error?
    Thanks
    In DAC I can validate Informatic and physical servers all connect and are OK and iunder execute, I've ran analyze repository tables and it reporting OK, ran create repository report and no errors or mission objects reports. When I run Execution plan for Human Resources - Oracle R1211 - Flexfield, it failed 5 of the 6 phases wherne 3 are stopped (never ran) and two failed (SDE_ORA_Flx_EBSSegDataTmpLoad, SDE_ORA_Flx_EBSValidationTableDataTmpLoad).
    When I review the logs on the unix server under the /u02/BIAPPS_SUITE/INFR/server/infa_shared/SessLogs/ORA_R1211.DATAWAREHOUSE.SDE_ORAR1211_Adaptor.SDE_ORA_Flx_EBSSegDataTmpLoad.log, I see that it is failing because it has a problem with the target table W_ORA_FLX_EBS_SEG_DATA_TMP
    READER_1_1_1> RR_4049 SQL Query issued to database : (Fri May 28 23:04:22 2010)
    READER_1_1_1> RR_4050 First row returned from database to reader : (Fri May 28 23:04:22 2010)
    WRITER_1_*_1> WRT_8167 Start loading table [W_ORA_FLX_EBS_SEG_DATA_TMP] at: Fri May 28 23:04:21 2010
    WRITER_1_*_1> CMN_1761 Timestamp Event: [Fri May 28 23:04:22 2010]
    WRITER_1_*_1> WRT_8229 Database errors occurred:
    **************ORA-01652: unable to extend temp segment by 128 in tablespace BIA_DW_TBS*
    When I review the ORA_R1211.DATAWAREHOUSE.SDE_ORAR1211_Adaptor.SDE_ORA_Flx_EBSValidationTableDataTmpLoad.log i found the following errors:
    INSERT INTO
    W_ORA_FLX_EBS_VALID_TAB_TMP(FLEX_VALUE_SET_ID,APPLICATION_TABLE_NAME,VALUE_COLUMN_NAME,ID_COLUMN_NAME,VALUE_SET_WHERE_CLAUSE, DATASOURCE_NUM_ID,VALUE_SET_SQL) VALUES ( ?, ?, ?, ?, ?, ?, ?) WRITER_1_*_1> WRT_8020
    ************ No column marked as primary key for table[W_ORA_FLX_EBS_VALID_TAB_TMP]. UPDATEs Not Supported.
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_ORA_FLX_EBS_VALID_TAB_TMP]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158 And then again the can't extend error WRT_8036 Target: W_ORA_FLX_EBS_VALID_TAB_TMP (Instance Name: [W_ORA_FLX_EBS_VALID_TAB_TMP])
    WRT_8038 Inserted rows - Requested: 10080 Applied: 10080 Rejected: 0 Affected: 10080
    READER_1_1_1> BLKR_16019 Read [13892] rows, read [0] error rows for source table [FND_FLEX_VALIDATION_TABLES]
    instance name [FND_FLEX_VALIDATION_TABLES]
    READER_1_1_1> BLKR_16008 Reader run completed.
    WRITER_1_*_1> CMN_1761 Timestamp Event: [Fri May 28 23:04:21 2010]
    WRITER_1_*_1> WRT_8229 Database errors occurred:
    ***** ORA-01652: unable to extend temp segment by 128 in tablespace BIA_DW_TBS

    Worked with DBA and though source tables were not huge, we extended manually and proccess completed once we re-ran.
    Thanks all.

  • Copy records of one table into another using Update statement

    Supposing I have 2 tables test and test1.
    Test has columns col1 and col2. and test1 has only one table col1.
    select * from test returns:
    COL1 COL2
    a
    b
    c
    d
    e
    select * from test1 returns
    COL1
    p
    q
    r
    s
    t
    Suppose i want to copy values of test1.col1 to test.col2 so tht final result of
    select * from test should be
    COL1 COL2
    a p
    b q
    c r
    d s
    e t
    I found a query in the OCP introduction book:
    update test set col2=(select col1 from test11)
    this works fine only when we have a single record in test1 table but fails otherwise.
    how can this be achieved using the update statement ?

    SQL> desc test
    Name Null? Type
    COL1 VARCHAR2(10)
    COL2 VARCHAR2(30)
    SQL> desc test1
    Name Null? Type
    COL1 VARCHAR2(10)
    SQL> insert into test values ('a','');
    1 row created.
    SQL> insert into test values ('b','');
    1 row created.
    SQL> insert into test values ('c','');
    1 row created.
    SQL> insert into test values ('d','');
    1 row created.
    SQL> insert into test1 values ('e');
    1 row created.
    SQL> insert into test1 values ('f');
    1 row created.
    SQL> insert into test1 values ('g');
    1 row created.
    SQL> insert into test1 values ('h');
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> ed
    Wrote file afiedt.buf
    1 select a.col1, b.col1, a.rn from
    2 (select row_number() over (order by col1) rn, col1 from test) a,
    3 (select row_number() over (order by col1) rn, col1 from test1) b
    4* where a.rn = b.rn
    SQL> /
    COL1 COL1 RN
    a e 1
    b f 2
    c g 3
    d h 4
    SQL> ed
    Wrote file afiedt.buf
    1 update test set col1 =
    2 (
    3 select forupd from (
    4 select a.col1, b.col1 test1col, a.col1||' '||b.col1 forupd, a.rn from
    5 (select row_number() over (order by col1) rn, col1 from test) a,
    6 (select row_number() over (order by col1) rn, col1 from test1) b
    7 where a.rn = b.rn
    8* ) z where z.col1 = test.col1)
    SQL> /
    4 rows updated.
    SQL> commit;
    Commit complete.
    SQL> select * from test;
    COL1 COL2
    a e
    b f
    c g
    d h
    SQL>
    This will work only if you have the same number of lines in COL1 in both tables. If you have different number of lines, then you need to code more at join (outer, inner).
    But for complicated cases, please post sample data.
    Michael
    PS: And supossing that the values are distinct in TEST table. Else update will fail, because more rows will be retrived.
    If more values in TEST table in COL1, then you need to assign a row number on TEST also, and in WHERE of query you need to place one more join condition:
    something like:
    8* ) z where z.col1 = test.col1
    and z.rn = test.rn)
    SQL> /
    But as i said, get more specifications.
    Message was edited by:
    kjt
    Added the PS.

  • Quantity conversion check failed for one item; all qty lines deleted

    Hi IS Oil Experts,
    Iam using the BAPI "BAPI_GOODSMVT_CREATE_OIL" for posting goods movement. I filled the additional quantities and unit of measures also in the BAPI table, But when tried to execute, Iam getting the error, "Quantity conversion check failed for one item; all qty lines deleted"
    Please anybody tell me what is the error from end for resulting this message.
    Thanks in advance
    Prathib

    Dear you!
    I also get the issue that you meet. Can you help me to process it, Please/
    Best regards, Huy.

  • How to delete the double records connected to one or more than one tables in SQL 2008?

    Hi
    Can anyone please help me with the SQL query. I Im having a table called People with columns names: personno., lastname, firstname and so on. The personno. is having duplicate records,so all the duplicate records i have written with "double" in
    the beginning of the numbers. I tried deleting these double records but they are linked to one or more than one tables. I have to find out, all the tables blocking the deleting of double person. And then create select statements which creates update statements
    in order to replace the current id of double person with substitute id. (The personno. is in the form of id's in the database)
    Thanks

    You should not append "double" in the personno. When we append it will not be able to join or relate to other table. Keep the id as it is and use another field(STATUS) to mark as duplicate. Also we will require another field(PRIMARYID) against
    those duplicate rows i.e the main or the primary personno.
    SELECT * FROM OtherTable a INNER JOIN
    (SELECT personno, status, primaryid FROM PEOPLE WHERE status = 'Duplicate') b
    ON a.personno = b.personno
    UPDATE OtherTable SET personno = b.primaryid
    FROM OtherTable a INNER JOIN
    (SELECT personno, status, primaryid FROM PEOPLE WHERE status = 'Duplicate') b
    ON a.personno = b.personno
    NOTE: Please take backup before applying the query. This is not tested.
    Regards, RSingh

  • Adding Data From One Table to Another

    Now, this doesn't strike me as a particularly complex problem, but I've either strayed outside the domain of Numbers or I'm just not looking at the problem from the right angle. In any case, I'm sure you guys can offer some insight.
    What I'm trying to do is, essentially, move data from one table to another. One table is a calendar, a simple two column 'date/task to be completed' affair, the other is a schedule of jogging workouts, i.e, times, distances. Basically, I'm trying to create a formula that copies data from the second table onto the first but only for odd days of the week, excepting Sundays (and assuming Monday as the start of the week). Now, this isn't the hard part, I can do that. The problem comes when I replicate the formula down the calendar. Even on the days when the 'if' statement identifies it as an 'even day', the cell reference to the appropriate workout on the second table is incremented, so when it comes to the next 'odd day', it has skipped a workout.
    I can't seem to see any way of getting it to specifically copy the NEXT line in the second table, and not the corresponding line.
    This began as a distraction to try and organise my running so I could see at a glance what I had to do that day and track my progress, but now it's turned into an obsession. SURELY there's a solution?
    Cheers.

    Hi Sealatron,
    Welcome to Apple Discussions and the Numbers '09 forum.
    Several possible ways to move the data occur to me, but the devil's in the details of how the data is currently arranged.
    Is it
    • a list of three workouts, one for each of Monday, Wednesday and Friday, then the same three repeated the following week?
    • an open-ended list that does not repeat?
    • something else?
    Regards,
    Barry

Maybe you are looking for