Replicated Partition: One to many?

I am hoping to set up a replicated partition, which will take exchange rate data from source to multiple targets.The issue is that I have multiple homes in any given target database for (eg) USD rate. ie I have tried using UDA's, but one USD tagged member in source, has multiple USD tagged members in target, hence partition doesnt verify.In short, can I map one to many, source to target in a replicated partition?Thanks in Advance

Are you mapping from one outline to many outlines by using the same member in the source outline to couple different target outlines? If your answer is yes, then yes you can do it. ---Message Posted by kmooney ??3/19/02 07:16---I am hoping to set up a replicated partition, which will take exchange rate data from source to multiple targets.The issue is that I have multiple homes in any given target database for (eg) USD rate. ie I have tried using UDA's, but one USD tagged member in source, has multiple USD tagged members in target, hence partition doesnt verify.In short, can I map one to many, source to target in a replicated partition?Thanks in Advance

Similar Messages

  • Split Partition - One to Many

    I have to split a single monthly partition (for month-end date) into daily partition for each day of the month. I am currently doing it one at a time, that is, by splitting one partition into two like 31st january into 31st January and 1st January. Then 31st January into 31st and 2nd January.( It is all in a loop). My question is can I one partition into more than 2 partition in one command so that they can happen in parallel.
    My current process is taking 10 hours to complete ( One month data is about 30 million which is splitting into approximate 1 million rows each day).
    I need to bring this time down to about couple of hours.
    Ant suggestion is welcome.

    Hello,
    h5. Best and fast option is to create any empty table with all the required daily partition and just move data and use direct path insert.
    CREATE TABLE T1
      D  DATE                                           NULL
    TABLESPACE TEST_DATA
    PARTITION BY RANGE (D)
      PARTITION P_MAR92009 VALUES LESS THAN (TO_DATE(' 2009-03-10 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
        LOGGING
        NOCOMPRESS
        TABLESPACE TEST_DATA,
        PARTITION P_MAR102009 VALUES LESS THAN (TO_DATE(' 2009-03-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
        LOGGING
        NOCOMPRESS
        TABLESPACE TEST_DATA
      insert /*+ APPEND */ into daily_partition_table select * from monthly_partition_table;Regards

  • Looking for a One to Many script to extend the system partition for Windows 7 machines

    Looking for a One to Many script to extend the system partition for Windows 7 machines

    Pre-written scripts can be found in the repository:
    http://gallery.technet.microsoft.com/scriptcenter
    If you can't find what you need, you can request a script (no idea if anyone ever bothers to fulfill these requests though, I know I don't):
    http://gallery.technet.microsoft.com/scriptcenter/site/requests
    Let us know if you have any specific questions.
    Don't retire TechNet! -
    (Don't give up yet - 12,950+ strong and growing)

  • How just return one row of a one to many join..

    So I have a one to many join where the SMOPERATOR table has data I need however it has a couple of rows that match the JOIN condition in there. I just need to return one row. I think this can be accomplished with a subquery in the join however have not been able to come up with the right syntax to do so.
    So:
    SELECT "NUMBER" as danumber,
    NAME,
    SMINCREQ.ASSIGNMENT,
    SMOPERATOR.PRIMARY_ASSIGNMENT_GROUP,
    SMOPERATOR.WDMANAGERNAME,
    SMINCREQ.owner_manager_name,
    SMINCREQ.subcategory, TO_DATE('01-'||TO_CHAR(open_time,'MM-YYYY'),'DD-MM-YYYY')MONTHSORT,
    (CASE WHEN bc_request='f' THEN 'IAIO'
    WHEN (bc_request='t' and substr(assignment,1,3)<>'MTS') THEN 'RARO'
    WHEN (bc_request='t' and substr(assignment,1,3)='MTS') THEN 'M'
    ELSE 'U' end) as type
    from SMINCREQ
    left outer join SMOPERATOR on SMINCREQ.assignment=SMOPERATOR.primary_assignment_group
    WHERE SMINCREQ.owner_manager_name=:P170_SELECTION and SMOPERATOR.wdmanagername=:P170_SELECTION
    AND open_time BETWEEN to_date(:P170_SDATEB,'DD-MON-YYYY') AND to_date(:P170_EDATEB,'DD-MON-YYYY')
    AND
    (bc_request='f' and subcategory='ACTIVATION' and related_record<>'t')
    OR
    (bc_request='f' and subcategory<>'ACTIVATION')
    OR
    (bc_request='t' and substr(assignment,1,3)<>'MTS')
    order by OPEN_TIMe

    Hi,
    This sounds like a Top-N Query , where you pick N items (N=1 in this case) off the top of an orderded list. I think you want a separate ordered list for each assignment; the analytic ROW_NUMBER function does that easily.
    Since you didn't post CREATE TABLE and INSERT statements for your sample data, I'll use tables from the scott schema to show how this is done.
    Say you have a query like this:
    SELECT       d.dname
    ,       e.empno, e.ename, e.job, e.sal
    FROM       scott.dept  d
    JOIN       scott.emp   e  ON   d.deptno = e.deptno
    ORDER BY  dname
    ;which produces this output:
    DNAME               EMPNO ENAME      JOB              SAL
    ACCOUNTING           7934 MILLER     CLERK           1300
    ACCOUNTING           7839 KING       PRESIDENT       5000
    ACCOUNTING           7782 CLARK      MANAGER         2450
    RESEARCH             7876 ADAMS      CLERK           1100
    RESEARCH             7902 FORD       ANALYST         3000
    RESEARCH             7566 JONES      MANAGER         2975
    RESEARCH             7369 SMITH      CLERK            800
    RESEARCH             7788 SCOTT      ANALYST         3000
    SALES                7521 WARD       SALESMAN        1250
    SALES                7844 TURNER     SALESMAN        1500
    SALES                7499 ALLEN      SALESMAN        1600
    SALES                7900 JAMES      CLERK            950
    SALES                7698 BLAKE      MANAGER         2850
    SALES                7654 MARTIN     SALESMAN        1250Now say you want to change the query so that it only returns one row per department, like this:
    DNAME               EMPNO ENAME      JOB              SAL
    ACCOUNTING           7782 CLARK      MANAGER         2450
    RESEARCH             7876 ADAMS      CLERK           1100
    SALES                7499 ALLEN      SALESMAN        1600where the empno, ename, job and sal columns on each row of output are all taken from the same row of scott.emp, though it doesn't really matter which row that is.
    One way to do it is to use the analytic ROW_NUMBER function to assign a sequence of unique numbers (1, 2, 3, ...) to all the rows in each department. Since each sequence startw with 1, and the numbers are unique within a department, there will be exactly one row per departement that was assigned the numebr 1, and we''ll display that row.
    Here's how to code that:
    WITH     got_r_num     AS
         SELECT     d.dname
         ,     e.empno, e.ename, e.job, e.sal
         ,     ROW_NUMBER () OVER ( PARTITION BY  d.dname
                                   ORDER BY          e.ename
                           )         AS r_num
         FROM     scott.dept  d
         JOIN     scott.emp   e  ON   d.deptno = e.deptno
    SELECT       dname
    ,       empno, ename, job, sal
    FROM       got_r_num
    WHERE       r_num     = 1
    ORDER BY  dname
    ;Notice that he sub-query got_r_num is almost the same as the original query; only it has one additional column, r_num, in the SELECT clause, and the sub-qeury does not have an ORDER BY clause. (Sub-queries almost never have an ORDER BY clause.)
    The ROW_NUMBER function must have an ORDER BY clause. In this example, I used "ORDER BY ename", meaning that, within each department, the row with the first ename (in sort order) will get r_num=1. You can use any column, or expression, or expressions in the ORDER BY clause. You muight as well use something consistent and predictable, like ename, but if you really wanted arbitrary numbering you could use a constant in the analytic ORDER BY clause, e.g. "ORDER BY NULL".

  • Replicated Partition-Cell Count not matching, what to do with Mappings tab

    Hi All
    I am creating a Replicated partition between 2 cubes on 9.3.1. The source cube has 8 dimensions(Cube A) and the target cube has 7 dimensions(Cube B).
    2 Dimensions in cube A dont exist in Cube B and 1 Dimension in Cube B doesnt exist in Cube A.
    6 of the dimensions map directly.
    How do i deal with the 2 Dimensions in cube A that dont match cube B? At the moment I have mapped them to (void) in the mappings tab, by just selecting the dimension name and putting void in the target.
    Cube B has dimension called SBU. All data from Cube A needs to go against member SBU Input in Cube B. I am stuck as to how to do this?
    Cheers
    Simon
    Edited by: SimonWatts on Oct 16, 2008 4:23 PM

    I believe the only issue that is likely to be wrong is the need to select a single member in the dimensions that you are eliminating (through a mapping to void). The 2 dims that exist in cube A should be selected as a single member for each, and mapped to void. If you selected the entire dimension in the partition, and only mapped the root to void, you would have a mismatched cell count.
    So, change the partition spec to use a single member in those dimensions (if needed), and make sure that the member you select is mapped to void, which is not necessarilly the dimension itself.
    Then, you do a similar void mapping to the SBU Input member, "Void = "SBU Input"". That handles the new cell count aspect if the other dimensions line up.
    In total, there are three mappings identified from this. Two map source members to void, and one to map a target member to void.

  • Replicated partition is taking a long time

    I have been using a replicated partition to move some data from one Database to another and it has been working working fine.
    All of a sudden the replicated partition is taking a long time ( does not complete) when I run a maxl script to refresh the replicated partition. I can't seem to figure out what caused the issue. I made sure the there are no virtual members defined in the sparse dimensions and the mapping are correct.
    Any idea of why the process will take so long all of a sudden?

    I have narrowed down to the problem on part of the hierarchy of the dimension. The parent member has about 2500 children under it and any time we include this part of the tree in the partition the partition takes a long time( the total number of members in the Cost center is 18K)

  • Problem with replicated partitioning between two BSO db's. version 9.3.0.1

    Hi there,
    I have a problem with partitioning between two BSO db’s.
    I created the partition, it validates and I made an export for the partion definiton(xml file).
    Everything is fine if I set this partition up as transparent partition.
    This means I can see the data from the src in the trg db as ecpected.
    But when I change over to replicated (what in fact I want to do) nothing happens for the trg site - there are still no data. (Yes I have choosen Update all cells.)
    I got no error in any way neither AAS nor in the app logs of src and trg.
    (Mayby this it important: one src-dimension is mapped to void for the trg.)
    Can anyone give me a hint or even confirm this behaviour ?
    Thank you in advance
    Andre
    Below the log infos:
    log of src db:
    [Fri Dec 04 15:29:03 2009]Local/AAA///Info(1013210)
    User [hyperion] set active on database [db]
    [Fri Dec 04 15:29:03 2009]Local/AAA/db/hyperion/Info(1013091)
    Received Command [PutReplicatedCells] from user [hyperion]
    [Fri Dec 04 15:29:03 2009]Local/AAA///Info(1013210)
    User [hyperion] set active on database [db]
    [Fri Dec 04 15:29:03 2009]Local/AAA/db/hyperion/Info(1013091)
    Received Command [XferReplicatedCells] from user [hyperion]
    [Fri Dec 04 15:29:03 2009]Local/AAA///Info(1013214)
    Clear Active on User [hyperion] Instance [1]
    [Fri Dec 04 15:29:03 2009]Local/AAA/db/hyperion/Info(1023005)
    Update of replicated partition elapsed time : [0] seconds
    [Fri Dec 04 15:29:03 2009]Local/AAA/db/hyperion/Info(1243010)
    Partition refreshed
    log of trg db:
    [Fri Dec 04 15:29:03 2009]Local/AAA1///Info(1013210)
    User [hyperion] set active on database [db]
    [Fri Dec 04 15:29:03 2009]Local/AAA1/db/hyperion/Info(1013091)
    Received Command [ReqXferReplicatedCells] from user [hyperion]
    [Fri Dec 04 15:29:03 2009]Local/AAA1///Info(1013210)
    User [hyperion] set active on database [db]
    [Fri Dec 04 15:29:03 2009]Local/AAA1/db/hyperion/Info(1013091)
    Received Command [ProtocolXchg] from user [hyperion]
    [Fri Dec 04 15:29:03 2009]Local/AAA1/db/hyperion/Info(1023044)
    Processing distributed request from [P4B5A856A, AAA, db]
    [Fri Dec 04 15:29:03 2009]Local/AAA1/db/hyperion/Info(1013091)
    Received Command [PutLastUpdatedTime] from user [hyperion]
    [Fri Dec 04 15:29:03 2009]Local/AAA1///Info(1013214)
    Clear Active on User [hyperion] Instance [1]
    [Fri Dec 04 15:29:03 2009]Local/AAA1///Info(1013214)
    Clear Active on User [hyperion] Instance [1]

    It sounds like you are doing it all correctly. One thing you might try just to be sure you are not getting any data in the target database is the following:
    1) clear the target database
    2) Run the replication for all cells
    3) Run an export of all data on the target database
    See if you get anything in the export file, maybe you are getting some of the data or possibly it is a rollup issue. Just a thought.
    Jeff

  • Performance tips on transparent and replicated partitions?

    I am quite new to this partitioning scene. Does anybody has tips on how to improve the performance and create the most feasible tranparent and replicated partitions? Thanks.

    Our design is triangle shaped structure. There is a main application getting fed from two sub applications and these two sub applications are connected also by a transparent partition. One of the sub applications is the same as the main application. Same number of dimensions with the same dimensions. The second sub application is a smaller subset of the main application with less dimensions. All the transparent partitions defined as "(@IDESCENDENTS("Managerial Statement of Operations") @IDESCENDENTS("Income Statement Ratios") @IDESCENDENTS("Operating Ratios") @IDESCENDENTS("Statistics")@IDESCENDENTS("System Administration")) @IDESCENDENTS("Time") @IDESCENDENTS("Income Statement Type") @IDESCENDENTS("Funding Type") @IDESCENDENTS("Network Indicator") @IDESCENDENTS("Product") @IDESCENDENTS("Market Segment") @IDESCENDENTS("State") @IDESCENDENTS("Scenario") @IDESCENDENTS("Trading Partner") @IDESCENDENTS("Organization")" at the source and "(@IDESCENDENTS("Managerial Statement of Operations") @IDESCENDENTS("Income Statement Ratios") @IDESCENDENTS("Operating Ratios") @IDESCENDENTS("Statistics")@DESCENDENTS("System Admin - PL")) @IDESCENDENTS("Time") @IDESCENDENTS("Income Statement Type") @IDESCENDENTS("Funding Type") @IDESCENDENTS("Network Indicator") @IDESCENDENTS("Product") @IDESCENDENTS("Market Segment") @IDESCENDENTS("State") @IDESCENDENTS("Scenario") @IDESCENDENTS("Trading Partner") @IDESCENDENTS("Organization")" at the target. The dimensions are "Account", "Time", "Scenario", "Market Segment", "Income Statement Type", "Funding Type", "Product", "State", "Organization", "Trading Partner", and "Network Indicator". Thanks.

  • One to many relationship question.

    Let's say we have table1 and table2
    Table1 has ID, and let's say Name for fields.
    Table2 has ID, and let's say SSN for fields.
    This is not so, but illustrates my points well.
    Table2 has many records with the same ID, who just happen to have the same SSN. That would usually be a logical conclusion.
    Table1 has only 1 of each ID in it. So you might say that ID is a foreign key in table2, except one problem. I don't think they know what primary keys are here, much less foreign keys.
    Anyway, I have to write a query where I get the ID from table1 and need the SSN from table2. That's the only place in this scenario that SSN is.
    My question is this - since this is a one to many relationship, yet I only need 1 result, how would I write this using the (+) way of notation? I am using 8i.
    Right now, I am doing something like this, using a DISTINCT:
    Select DISTINCT t1.ID, t1.Name, t2.SSN
    From table1, table2,
    Where t1.ID = t2.ID (+);
    First of all, is this the correct way of writing it?
    And second, I am wondering if, perchance for some error, there were 2 or more SSNs in table2, what would happen then? Would I get a too many rows error?
    And if so, how can I trap it and possibly get just the first occurrence of it?
    Is that even a sound way of doing it?
    Thanks for any thoughts here.

    You would need to turn your table2 into an in-line view in order to outer join to it properly. Minimally, something like:
    SELECT t1.id, t1.name, t2.ssn
    FROM table1 t1,
         (SELECT DISTINCT id, ssn
          FROM table2) t2
    WHERE t1.id = t2.id(+)If there may be multiple ssn in t2 for a single id, you will need to determine which one you want to keep and use something like:
    SELECT t1.id, t1.name, t2.ssn
    FROM table1 t1,
         (SELECT id, ssn
          FROM (SELECT id, ssn,
                       ROW_NUMBER() OVER (PARTITION BY id,
                                          ORDER BY column_that_makes_it_first) rn
                FROM table2)
          WHERE rn = 1) t2
    WHERE t1.id = t2.id(+)This should work on 8.1.7.4 Enterprise Edition. If you have a different version, then the analytic function may not work so you would need something more like:
    SELECT t1.id, t1.name, t2.ssn
    FROM table1 t1,
         (SELECT id, ssn
          FROM (SELECT id, ssn, rownum rn
                FROM (SELECT id, ssn,
                      FROM table2
                      ORDER BY id, column_that_makes_it_first))
          WHERE rn = 1) t2HTH
    John

  • Difference between transparent and replicated partitions

    whats the difference between transparent and replicated partitions in EAL. currently we have created essbase cube via transparent partition in eal. performance of obiee reports on essbase cube is very bad. now somebody suggested to create essbase cube via replicated partition in EAL. what difference would it make. one guy said replicated partition would make it more slow. i am confused. please explain.

    Transparent Partition does not occupy or create extra memory Space in the Destination Application , it just shows the Data from Source to the Destination cube . This partition is completely dependent on Existence of Source Data . While , in replicated Partition you can use a copy of source data in the Destination cube.
    FYI : Hyperion Essbase 11.1.1.0 &amp;#8211; Partitioning and its use cases &amp;#8211; Transparent, Replicated and Linked part…

  • Replicated Partition Hangs

    <p>I have created a replicated partition and everything validatescorrectly.  When I attempt to replicate across the partitionit seems to be working but its just hanging.  There is an .indfile created immediately in the new db but no page files. After a few minutes I kill the request and a page filesuddenly appears and the data in the cube exists like I wouldexpect it to.  Anyone have any idea whats going on?</p>

    I looked up the error message codes and found that 1130203 seems to be the key.<BR><blockquote>quote:<br><hr><b>1130203 </b><BR>Possible Problems<BR>Analytic Services is unable to allocate memory. <BR><BR>Possible Solutions<BR>Try any of the following suggestions to fix the problem. Once you fix the problem, check to see if the database is corrupt. <BR><BR>Check the physical memory on the server computer. In a Windows environment, 64 MB is the suggested minimum for one database. In a UNIX environment, 128 MB is the suggested minimum for one database. If the error keeps occuring, add more memory to the server computer. <BR>If you are on a UNIX computer, check the user limit profile. <BR>Check the block size of the database. If necessary, reduce the block size. <BR>Check the data cache and data file cache setting. If necessary, decrease the data cache and data file cache sizes. <BR>Make sure that the Analytic Services computer has enough resources. Consult the Analytic Services Installation Guide for a list of system requirements. If a resource-intensive application, such as a relational database, is running on the same computer, the resource-intensive application may be using the resources that Analytic Services needs. <BR>See Also<BR>1006006<BR>1006010<BR>1006015<BR>1006023<BR><hr></blockquote><BR><BR>It looks like you may be requiring more memory than might be available. This is the same sort of error that you see when you find dynamic calcs needing more calc lock block space, a problem normally resolved by increasing the data cache.<BR><BR>It may be that the sparse-dense arrangements of your source and target databases are organized in such a way that you need to have multiple blocks cached for each during the replication. Without seeing the outlines and partition definitions, this is a best guess. I don't know whther the order of dimensions in the partition definition makes any difference, but different organizations can cause memory allocation problems, especially if the server is running at full load. <BR><BR>You might see if there is any difference when replicating during a slow use time. Check to see that all inactive applications are stopped. That may help memory allocation. If you have a separate test server, you may want to play a bit with settings. Otherwise, try monitoring memory use on the server when doing the partition.

  • Replicated Partition Question

    With a replicated partition is it possible to have the source area be the same and send it to 2 different targets in the same database.
    here is what i am trying to accomplish
    Source A
    @DESCENDANTS("GM",0),@DESCENDANTS("SECURED",0),"OCT","NOV","DEC",@DESCENDANTS("CUSTOM1",0),@DESCENDANTS("CUSTOM2",0),@DESCENDANTS("CUSTOM3",0),@DESCENDANTS("ICP",0),"2014","Version4","OUTLOOK","Local",@DESCENDANTS("SVCNA",0)
    Target A
    @DESCENDANTS("GM",0),@DESCENDANTS("SECURED",0),"OCT","NOV","DEC",@DESCENDANTS("CUSTOM1",0),@DESCENDANTS("CUSTOM2",0),@DESCENDANTS("CUSTOM3",0),@DESCENDANTS("ICP",0),"2014","MONTH1","OUTLOOK","Local",@DESCENDANTS("SVCNA",0)
    MAP A
    "VERSION4" to "MONTH1"
    Source B (same as A)
    @DESCENDANTS("GM",0),@DESCENDANTS("SECURED",0),"OCT","NOV","DEC",@DESCENDANTS("CUSTOM1",0),@DESCENDANTS("CUSTOM2",0),@DESCENDANTS("CUSTOM3",0),@DESCENDANTS("ICP",0),"2014","Version4","OUTLOOK","Local",@DESCENDANTS("SVCNA",0)
    Target B
    @DESCENDANTS("GM",0),@DESCENDANTS("SECURED",0),"OCT","NOV","DEC",@DESCENDANTS("CUSTOM1",0),@DESCENDANTS("CUSTOM2",0),@DESCENDANTS("CUSTOM3",0),@DESCENDANTS("ICP",0),"2014","MONTH1","ROLLING3MTH","Local",@DESCENDANTS("SVCNA",0)
    MAP B
    "VERSION4" to "MONTH1"
    "OUTLOOK" to "ROLLING3MTH"
    Tried setting this up and it doesnt work
    Thinking I need a maxl to create partition, replicate drop it create a second partition, replicate, and drop it.
    Any thoughts
    Thanks

    They are.
    Version Dimension is Version1-5 in one app and Month1, Month2, and Month3 in the other app.
    Scenario Dimension is Outlook in one app and Outlook AND ROLLING3MTH in the other.
    Another thing is I can get the replicated partition to refresh the data.  But if I have the map for OUTLOOK to ROLLING3MTH it only pushes to ROLLING3MTH.  If I leave out the map it only goes to OUTLOOK in Target.

  • Replicated partition Refresh

    Hi All,
    We have replicated partition from Cube03 to Cube01.. We refresh the replicated partition data daily using a MAXL Script.
    In Cube03 the data in INtersection01 is copied to Intersection02 at the end of budget cycle.......and the data in INtersectiono1 is deleted...and We refresh the replicated partition to get data to Cube01 for these new intersections....
    Since we have been refreshing the data in INtersectiono1 to Cube01....Is it stored in Cube01...Do we need to delete this data in Cube01 for INtersectiono1 ...so that ONLY the new intersection data Intersection02 is reflected in Cube01
    Appreciate your respone........

    Hi,
    The partition exists but it has been created within EAS where the server added is the cluster as opposed to the actual server name essbase is on - EssbaseCluster-1 - So this is what it has for the Essbase Server for the Data Source and Data Target. When we try and create a partition with using the actual server name it orphans (I have another thread about that one)
    Here is the output from the MAXL that I executed when trying the server name:
    MAXL> refresh outline on replicated partition finance.sscstf to finance.bud
    2> at 'http://HYPESSBASE-DEV.***.co.uk'
    3> apply all;
    ERROR - 1040042 - Unable to get the cluster name from the url [http://HYPESSBASE-DEV.***.co.uk].
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    ERROR - 1040042 - Unable to get the cluster name from the url [http://HYPESSBASE-DEV.***.co.uk].
    ERROR - 1241128 - Partition does not exist - 'can't be found in local partition definition file'.
    Here is the output when trying the server IP address
    MAXL> refresh outline on replicated partition finance.sscstf to finance.bud
    2> at '**.**.**.**'
    3> apply all;
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    WARNING - 1040048 - HTTP communication failed with error [Couldn't connect to server].
    ERROR - 1241128 - Partition does not exist - 'can't be found in local partition definition file'.

  • Replicated Partitions

    I have strange behavior with my replicated partitions. When my application
    starts up I always get N+1 replicants. For example, if I specify 3
    replicants, Econsole starts up 4 active partitions. Even more bizzare is
    that it also starts up 2 active router partitions... If I manually
    shutdown the extra partitions my application DOES NOT get DEGRADED
    indicating that it truly knows the number of replicants that SHOULD be
    there.
    Anyone have any similar experiences, solutions or ideas for this problem???
    Thanks,
    Rich Boswell
    The World Bank
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

    try turning off autostart temporarily and see if the behavior
    persists
    is it possible that one of your other partitions makes a call
    to your replicated partition prior to the replicated partition's
    startup?
    [email protected] wrote:
    >
    I have strange behavior with my replicated partitions. When my application
    starts up I always get N+1 replicants. For example, if I specify 3
    replicants, Econsole starts up 4 active partitions. Even more bizzare is
    that it also starts up 2 active router partitions... If I manually
    shutdown the extra partitions my application DOES NOT get DEGRADED
    indicating that it truly knows the number of replicants that SHOULD be
    there.
    Anyone have any similar experiences, solutions or ideas for this problem???
    Thanks,
    Rich Boswell
    The World Bank
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>--
    John Jamison [email protected]
    Vice President and Chief Technology Officer
    Sage IT Partners, Inc.
    Voice: 415 392-7243 x 306
    Fax: 415 391-3899
    Internet Enabled Business Change
    http://www.sageit.com
    -----------------------------------------------------

  • Unable to read one-to-many relations using Hibernate

    Hi,
    I am trying with a very simple one-to-many relationship. When I am storing the objects, there are no problems. But when I am trying to read out the collection, it says invalid descriptor index. Please help.
    Regards,
    Hibernate version:
    hibernate-3.1rc2
    Mapping documents:
    <?xml version="1.0"?>
    <!DOCTYPE hibernate-mapping PUBLIC
    "-//Hibernate/Hibernate Mapping DTD 3.0//EN"
    "http://hibernate.sourceforge.net/hibernate-mapping-3.0.dtd">
    <hibernate-mapping>
         <class name="Parent">
              <id name="id">
                   <generator class="identity"/>
              </id>
              <set name="children">
                   <key column="parent_id"/>
                   <one-to-many class="Child"/>
              </set>
         </class>
         <class name="Child">
              <id name="id">
                   <generator class="identity"/>
              </id>
              <property name="name"/>
         </class>
    </hibernate-mapping>
    Code between sessionFactory.openSession() and session.close():
    The Parent class:
    public class Parent
         private Long id ;     
         private Set children;
         Parent(){}
         public Long getId()
              return id;
         public void setId(Long id)
              this.id=id;
         public Set getChildren()
              return children;
         public void setChildren(Set children)
              this.children=children;
    The Child class:
    public class Child
         private Long id;
         private String name;
         Child(){}
         public Long getId()
              return id;
         private void setId(Long id)
              this.id=id;
         public String getName()
              return name;
         public void setName(String name)
              this.name=name;
    The Main class:
    public class PCManager
         public static void main(String[] args)
              PCManager mgr = new PCManager();
              List lt = null;
              if (args[0].equals("store"))
                   mgr.createAndStoreParent(new HashSet(3));
              else if (args[0].equals("list"))
                   mgr.listEvents();
              HibernateUtil.getSessionFactory().close();
         private void createAndStoreParent(HashSet s)
              HibernateUtil.getSessionFactory().getCurrentSession().beginTransaction();
              Parent p1 = new Parent();
              int size = 3;
              for (int i=size; i>0; i--)
                   Child c = new Child();
                   c.setName("Child"+i);
                   s.add(c);
              p1.setChildren (s);
              Iterator elems = s.iterator();
              do {     
                   Child ch = (Child) elems.next();
                   HibernateUtil.getSessionFactory().getCurrentSession().save(ch);
              }while(elems.hasNext());
              HibernateUtil.getSessionFactory().getCurrentSession().save(p1);
              HibernateUtil.getSessionFactory().getCurrentSession().getTransaction().commit();
         private void listEvents()
              HibernateUtil.getSessionFactory().getCurrentSession().beginTransaction();
              Parent result = (Parent) HibernateUtil.getSessionFactory().getCurrentSession().load(Parent.class, new Long(1));
              System.out.println("Id is :"+ result.getId());
              Set children = result.getChildren();
              Iterator elems = children.iterator();
              do {     
                   Child ch = (Child) elems.next();
                   System.out.println("Child Name"+ ch.getName());
              }while(elems.hasNext());
              HibernateUtil.getSessionFactory().getCurrentSession().getTransaction().commit();          
    Full stack trace of any exception that occurs:
    When I run with "hbm2ddl.auto" property as validate and trying to list the contents, I get the following out put.
    C:\Documents and Settings\mirza\Desktop\hibernate-3.1rc2\MyHibernate>ant run -Da
    ction=list
    Buildfile: build.xml
    clean:
    [delete] Deleting directory C:\Documents and Settings\mirza\Desktop\hibernate
    -3.1rc2\MyHibernate\bin
    [mkdir] Created dir: C:\Documents and Settings\mirza\Desktop\hibernate-3.1rc
    2\MyHibernate\bin
    copy-resources:
    [copy] Copying 4 files to C:\Documents and Settings\mirza\Desktop\hibernate
    -3.1rc2\MyHibernate\bin
    compile:
    [javac] Compiling 5 source files to C:\Documents and Settings\mirza\Desktop\
    hibernate-3.1rc2\MyHibernate\bin
    run:
    [java] 09:09:23,433 INFO Environment:474 - Hibernate 3.1 rc2
    [java] 09:09:23,449 INFO Environment:489 - loaded properties from resource
    hibernate.properties: {hibernate.cglib.use_reflection_optimizer=true, hibernate
    .cache.provider_class=org.hibernate.cache.HashtableCacheProvider, hibernate.dial
    ect=org.hibernate.dialect.SQLServerDialect, hibernate.max_fetch_depth=1, hiberna
    te.jdbc.use_streams_for_binary=true, hibernate.format_sql=true, hibernate.query.
    substitutions=yes 'Y', no 'N', hibernate.proxool.pool_alias=pool1, hibernate.cac
    he.region_prefix=hibernate.test, hibernate.jdbc.batch_versioned_data=true, hiber
    nate.connection.pool_size=1}
    [java] 09:09:23,465 INFO Environment:519 - using java.io streams to persis
    t binary types
    [java] 09:09:23,465 INFO Environment:520 - using CGLIB reflection optimize
    r
    [java] 09:09:23,481 INFO Environment:550 - using JDK 1.4 java.sql.Timestam
    p handling
    [java] 09:09:23,559 INFO Configuration:1257 - configuring from resource: /
    hibernate.cfg.xml
    [java] 09:09:23,559 INFO Configuration:1234 - Configuration resource: /hib
    ernate.cfg.xml
    [java] 09:09:23,872 INFO Configuration:460 - Reading mappings from resourc
    e: PCMapping.hbm.xml
    [java] 09:09:24,013 INFO HbmBinder:266 - Mapping class: Parent -> Parent
    [java] 09:09:24,045 INFO HbmBinder:266 - Mapping class: Child -> Child
    [java] 09:09:24,045 INFO Configuration:1368 - Configured SessionFactory: n
    ull
    [java] 09:09:24,061 INFO Configuration:1014 - processing extends queue
    [java] 09:09:24,061 INFO Configuration:1018 - processing collection mappin
    gs
    [java] 09:09:24,061 INFO HbmBinder:2233 - Mapping collection: Parent.child
    ren -> Child
    [java] 09:09:24,076 INFO Configuration:1027 - processing association prope
    rty references
    [java] 09:09:24,076 INFO Configuration:1049 - processing foreign key const
    raints
    [java] 09:09:24,155 INFO DriverManagerConnectionProvider:41 - Using Hibern
    ate built-in connection pool (not for production use!)
    [java] 09:09:24,170 INFO DriverManagerConnectionProvider:42 - Hibernate co
    nnection pool size: 1
    [java] 09:09:24,170 INFO DriverManagerConnectionProvider:45 - autocommit m
    ode: false
    [java] 09:09:24,170 INFO DriverManagerConnectionProvider:80 - using driver
    : sun.jdbc.odbc.JdbcOdbcDriver at URL: jdbc:odbc:MySQL
    [java] 09:09:24,170 INFO DriverManagerConnectionProvider:86 - connection p
    roperties: {}
    [java] 09:09:24,264 INFO SettingsFactory:77 - RDBMS: Microsoft SQL Server,
    version: 08.00.0194
    [java] 09:09:24,264 INFO SettingsFactory:78 - JDBC driver: JDBC-ODBC Bridg
    e (SQLSRV32.DLL), version: 2.0001 (03.85.1117)
    [java] 09:09:24,296 INFO Dialect:100 - Using dialect: org.hibernate.dialec
    t.SQLServerDialect
    [java] 09:09:24,311 INFO TransactionFactoryFactory:31 - Using default tran
    saction strategy (direct JDBC transactions)
    [java] 09:09:24,327 INFO TransactionManagerLookupFactory:33 - No Transacti
    onManagerLookup configured (in JTA environment, use of read-write or transaction
    al second-level cache is not recommended)
    [java] 09:09:24,327 INFO SettingsFactory:125 - Automatic flush during befo
    reCompletion(): disabled
    [java] 09:09:24,327 INFO SettingsFactory:129 - Automatic session close at
    end of transaction: disabled
    [java] 09:09:24,343 INFO SettingsFactory:144 - Scrollable result sets: ena
    bled
    [java] 09:09:24,343 INFO SettingsFactory:152 - JDBC3 getGeneratedKeys(): d
    isabled
    [java] 09:09:24,343 INFO SettingsFactory:160 - Connection release mode: au
    to
    [java] 09:09:24,358 INFO SettingsFactory:184 - Maximum outer join fetch de
    pth: 1
    [java] 09:09:24,358 INFO SettingsFactory:187 - Default batch fetch size: 1
    [java] 09:09:24,358 INFO SettingsFactory:191 - Generate SQL with comments:
    disabled
    [java] 09:09:24,358 INFO SettingsFactory:195 - Order SQL updates by primar
    y key: disabled
    [java] 09:09:24,358 INFO SettingsFactory:338 - Query translator: org.hiber
    nate.hql.ast.ASTQueryTranslatorFactory
    [java] 09:09:24,374 INFO ASTQueryTranslatorFactory:21 - Using ASTQueryTran
    slatorFactory
    [java] 09:09:24,374 INFO SettingsFactory:203 - Query language substitution
    s: {no='N', yes='Y'}
    [java] 09:09:24,374 INFO SettingsFactory:209 - Second-level cache: enabled
    [java] 09:09:24,374 INFO SettingsFactory:213 - Query cache: disabled
    [java] 09:09:24,374 INFO SettingsFactory:325 - Cache provider: org.hiberna
    te.cache.HashtableCacheProvider
    [java] 09:09:24,374 INFO SettingsFactory:228 - Optimize cache for minimal
    puts: disabled
    [java] 09:09:24,374 INFO SettingsFactory:233 - Cache region prefix: hibern
    ate.test
    [java] 09:09:24,405 INFO SettingsFactory:237 - Structured second-level cac
    he entries: disabled
    [java] 09:09:24,437 INFO SettingsFactory:257 - Echoing all SQL to stdout
    [java] 09:09:24,452 INFO SettingsFactory:264 - Statistics: disabled
    [java] 09:09:24,452 INFO SettingsFactory:268 - Deleted entity synthetic id
    entifier rollback: disabled
    [java] 09:09:24,452 INFO SettingsFactory:283 - Default entity-mode: POJO
    [java] 09:09:24,593 INFO SessionFactoryImpl:155 - building session factory
    [java] 09:09:24,938 INFO SessionFactoryObjectFactory:82 - Not binding fact
    ory to JNDI, no JNDI name configured
    [java] 09:09:24,954 INFO SchemaValidator:99 - Running schema validator
    [java] 09:09:24,954 INFO SchemaValidator:107 - fetching database metadata
    [java] 09:09:24,954 INFO Configuration:1014 - processing extends queue
    [java] 09:09:24,954 INFO Configuration:1018 - processing collection mappin
    gs
    [java] 09:09:24,954 INFO Configuration:1027 - processing association prope
    rty references
    [java] 09:09:24,954 INFO Configuration:1049 - processing foreign key const
    raints
    [java] 09:09:24,985 WARN JDBCExceptionReporter:71 - SQL Error: 0, SQLState
    : S1002
    [java] 09:09:24,985 ERROR JDBCExceptionReporter:72 - [Microsoft][ODBC SQL S
    erver Driver]Invalid Descriptor Index
    [java] 09:09:25,001 ERROR SchemaValidator:129 - Error closing connection
    [java] java.sql.SQLException: [Microsoft][ODBC SQL Server Driver]Invalid transaction state
    [java] Initial SessionFactory creation failed.org.hibernate.exception.Gener
    icJDBCException: could not get table metadata: Child
    [java] java.lang.ExceptionInInitializerError
    [java] at HibernateUtil.<clinit>(Unknown Source)
    [java] at PCManager.listEvents(Unknown Source)
    [java] at PCManager.main(Unknown Source)
    [java] Caused by: org.hibernate.exception.GenericJDBCException: could not g
    et table metadata: Child
    [java] at org.hibernate.exception.SQLStateConverter.handledNonSpecificE
    xception(SQLStateConverter.java:91)
    [java] at sun.jdbc.odbc.JdbcOdbc.createSQLException(JdbcOdbc.java:6879)
    [java] at sun.jdbc.odbc.JdbcOdbc.standardError(JdbcOdbc.java:7036)
    [java] at sun.jdbc.odbc.JdbcOdbc.SQLDisconnect(JdbcOdbc.java:2988)
    [java] at sun.jdbc.odbc.JdbcOdbcDriver.disconnect(JdbcOdbcDriver.java:9
    80)
    [java] at sun.jdbc.odbc.JdbcOdbcConnection.close(JdbcOdbcConnection.jav
    a:739)
    [java] at org.hibernate.tool.hbm2ddl.SchemaValidator.validate(SchemaVal
    idator.java:125)
    [java] at org.hibernate.impl.SessionFactoryImpl.<init>(SessionFactoryIm
    pl.java:299)
    [java] at org.hibernate.cfg.Configuration.buildSessionFactory(Configura
    tion.java:1145)
    [java] at HibernateUtil.<clinit>(Unknown Source)
    [java] at org.hibernate.exception.SQLStateConverter.convert(SQLStateCon
    verter.java:79)
    [java] at org.hibernate.exception.JDBCExceptionHelper.convert(JDBCExcep
    tionHelper.java:43)
    [java] at org.hibernate.exception.JDBCExceptionHelper.convert(JDBCExcep
    tionHelper.java:29)
    [java] at org.hibernate.tool.hbm2ddl.DatabaseMetadata.getTableMetadata(
    DatabaseMetadata.java:100)
    [java] at org.hibernate.cfg.Configuration.validateSchema(Configuration.
    java:946)
    [java] at org.hibernate.tool.hbm2ddl.SchemaValidator.validate(SchemaVal
    idator.java:116)
    [java] at org.hibernate.impl.SessionFactoryImpl.<init>(SessionFactoryIm
    pl.java:299)
    [java] at org.hibernate.cfg.Configuration.buildSessionFactory(Configura
    tion.java:1145)
    [java] ... 3 more
    [java] Caused by: java.sql.SQLException: [Microsoft][ODBC SQL Server Driver
    ]Invalid Descriptor Index
    [java] at sun.jdbc.odbc.JdbcOdbc.createSQLException(JdbcOdbc.java:6879)
    [java] at sun.jdbc.odbc.JdbcOdbc.standardError(JdbcOdbc.java:7036)
    [java] at sun.jdbc.odbc.JdbcOdbc.SQLGetDataString(JdbcOdbc.java:3862)
    [java] at sun.jdbc.odbc.JdbcOdbcResultSet.getDataString(JdbcOdbcResultS
    et.java:5561)
    [java] at sun.jdbc.odbc.JdbcOdbcResultSet.getString(JdbcOdbcResultSet.j
    ava:338)
    [java] at sun.jdbc.odbc.JdbcOdbcResultSet.getString(JdbcOdbcResultSet.j
    ava:395)
    [java] at PCManager.listEvents(Unknown Source)
    [java] at PCManager.main(Unknown Source)
    [java] at org.hibernate.tool.hbm2ddl.TableMetadata.<init>(TableMetadata
    .java:30)
    [java] at org.hibernate.tool.hbm2ddl.DatabaseMetadata.getTableMetadata(
    DatabaseMetadata.java:85)
    [java] ... 7 more
    [java] Exception in thread "main"
    [java] Java Result: 1
    BUILD SUCCESSFUL
    Total time: 4 seconds
    Name and version of the database you are using:
    Microsoft SQLServer2000
    The generated SQL (show_sql=true):
    When the program is run with "hbm2ddl.auto" property as create, I get the following output.
    C:\Documents and Settings\mirza\Desktop\hibernate-3.1rc2\MyHibernate>ant run -Daction=store
    Buildfile: build.xml
    clean:
    [delete] Deleting directory C:\Documents and Settings\mirza\Desktop\hibernate
    -3.1rc2\MyHibernate\bin
    [mkdir] Created dir: C:\Documents and Settings\mirza\Desktop\hibernate-3.1rc
    2\MyHibernate\bin
    copy-resources:
    [copy] Copying 4 files to C:\Documents and Settings\mirza\Desktop\hibernate
    -3.1rc2\MyHibernate\bin
    compile:
    [javac] Compiling 5 source files to C:\Documents and Settings\mirza\Desktop\
    hibernate-3.1rc2\MyHibernate\bin
    run:
    [java] 09:12:54,820 INFO Environment:474 - Hibernate 3.1 rc2
    [java] 09:12:54,836 INFO Environment:489 - loaded properties from resource
    hibernate.properties: {hibernate.cglib.use_reflection_optimizer=true, hibernate
    .cache.provider_class=org.hibernate.cache.HashtableCacheProvider, hibernate.dial
    ect=org.hibernate.dialect.SQLServerDialect, hibernate.max_fetch_depth=1, hiberna
    te.jdbc.use_streams_for_binary=true, hibernate.format_sql=true, hibernate.query.
    substitutions=yes 'Y', no 'N', hibernate.proxool.pool_alias=pool1, hibernate.cac
    he.region_prefix=hibernate.test, hibernate.jdbc.batch_versioned_data=true, hiber
    nate.connection.pool_size=1}
    [java] 09:12:54,852 INFO Environment:519 - using java.io streams to persis
    t binary types
    [java] 09:12:54,852 INFO Environment:520 - using CGLIB reflection optimize
    r
    [java] 09:12:54,867 INFO Environment:550 - using JDK 1.4 java.sql.Timestam
    p handling
    [java] 09:12:54,946 INFO Configuration:1257 - configuring from resource: /
    hibernate.cfg.xml
    [java] 09:12:54,946 INFO Configuration:1234 - Configuration resource: /hib
    ernate.cfg.xml
    [java] 09:12:55,259 INFO Configuration:460 - Reading mappings from resourc
    e: PCMapping.hbm.xml
    [java] 09:12:55,400 INFO HbmBinder:266 - Mapping class: Parent -> Parent
    [java] 09:12:55,447 INFO HbmBinder:266 - Mapping class: Child -> Child
    [java] 09:12:55,447 INFO Configuration:1368 - Configured SessionFactory: n
    ull
    [java] 09:12:55,447 INFO Configuration:1014 - processing extends queue
    [java] 09:12:55,447 INFO Configuration:1018 - processing collection mappin
    gs
    [java] 09:12:55,447 INFO HbmBinder:2233 - Mapping collection: Parent.child
    ren -> Child
    [java] 09:12:55,463 INFO Configuration:1027 - processing association prope
    rty references
    [java] 09:12:55,479 INFO Configuration:1049 - processing foreign key const
    raints
    [java] 09:12:55,557 INFO DriverManagerConnectionProvider:41 - Using Hibern
    ate built-in connection pool (not for production use!)
    [java] 09:12:55,557 INFO DriverManagerConnectionProvider:42 - Hibernate co
    nnection pool size: 1
    [java] 09:12:55,557 INFO DriverManagerConnectionProvider:45 - autocommit m
    ode: false
    [java] 09:12:55,573 INFO DriverManagerConnectionProvider:80 - using driver
    : sun.jdbc.odbc.JdbcOdbcDriver at URL: jdbc:odbc:MySQL
    [java] 09:12:55,573 INFO DriverManagerConnectionProvider:86 - connection p
    roperties: {}
    [java] 09:12:55,651 INFO SettingsFactory:77 - RDBMS: Microsoft SQL Server,
    version: 08.00.0194
    [java] 09:12:55,667 INFO SettingsFactory:78 - JDBC driver: JDBC-ODBC Bridg
    e (SQLSRV32.DLL), version: 2.0001 (03.85.1117)
    [java] 09:12:55,682 INFO Dialect:100 - Using dialect: org.hibernate.dialec
    t.SQLServerDialect
    [java] 09:12:55,698 INFO TransactionFactoryFactory:31 - Using default tran
    saction strategy (direct JDBC transactions)
    [java] 09:12:55,714 INFO TransactionManagerLookupFactory:33 - No Transacti
    onManagerLookup configured (in JTA environment, use of read-write or transaction
    al second-level cache is not recommended)
    [java] 09:12:55,714 INFO SettingsFactory:125 - Automatic flush during befo
    reCompletion(): disabled
    [java] 09:12:55,714 INFO SettingsFactory:129 - Automatic session close at
    end of transaction: disabled
    [java] 09:12:55,729 INFO SettingsFactory:144 - Scrollable result sets: ena
    bled
    [java] 09:12:55,729 INFO SettingsFactory:152 - JDBC3 getGeneratedKeys(): d
    isabled
    [java] 09:12:55,745 INFO SettingsFactory:160 - Connection release mode: au
    to
    [java] 09:12:55,745 INFO SettingsFactory:184 - Maximum outer join fetch de
    pth: 1
    [java] 09:12:55,745 INFO SettingsFactory:187 - Default batch fetch size: 1
    [java] 09:12:55,745 INFO SettingsFactory:191 - Generate SQL with comments:
    disabled
    [java] 09:12:55,745 INFO SettingsFactory:195 - Order SQL updates by primar
    y key: disabled
    [java] 09:12:55,745 INFO SettingsFactory:338 - Query translator: org.hiber
    nate.hql.ast.ASTQueryTranslatorFactory
    [java] 09:12:55,777 INFO ASTQueryTranslatorFactory:21 - Using ASTQueryTran
    slatorFactory
    [java] 09:12:55,792 INFO SettingsFactory:203 - Query language substitution
    s: {no='N', yes='Y'}
    [java] 09:12:55,792 INFO SettingsFactory:209 - Second-level cache: enabled
    [java] 09:12:55,792 INFO SettingsFactory:213 - Query cache: disabled
    [java] 09:12:55,792 INFO SettingsFactory:325 - Cache provider: org.hiberna
    te.cache.HashtableCacheProvider
    [java] 09:12:55,808 INFO SettingsFactory:228 - Optimize cache for minimal
    puts: disabled
    [java] 09:12:55,808 INFO SettingsFactory:233 - Cache region prefix: hibern
    ate.test
    [java] 09:12:55,808 INFO SettingsFactory:237 - Structured second-level cac
    he entries: disabled
    [java] 09:12:55,839 INFO SettingsFactory:257 - Echoing all SQL to stdout
    [java] 09:12:55,839 INFO SettingsFactory:264 - Statistics: disabled
    [java] 09:12:55,839 INFO SettingsFactory:268 - Deleted entity synthetic id
    entifier rollback: disabled
    [java] 09:12:55,839 INFO SettingsFactory:283 - Default entity-mode: POJO
    [java] 09:12:55,980 INFO SessionFactoryImpl:155 - building session factory
    [java] 09:12:56,325 INFO SessionFactoryObjectFactory:82 - Not binding fact
    ory to JNDI, no JNDI name configured
    [java] 09:12:56,341 INFO Configuration:1014 - processing extends queue
    [java] 09:12:56,341 INFO Configuration:1018 - processing collection mappin
    gs
    [java] 09:12:56,341 INFO Configuration:1027 - processing association prope
    rty references
    [java] 09:12:56,341 INFO Configuration:1049 - processing foreign key const
    raints
    [java] 09:12:56,356 INFO Configuration:1014 - processing extends queue
    [java] 09:12:56,356 INFO Configuration:1018 - processing collection mappin
    gs
    [java] 09:12:56,356 INFO Configuration:1027 - processing association prope
    rty references
    [java] 09:12:56,372 INFO Configuration:1049 - processing foreign key const
    raints
    [java] 09:12:56,372 INFO SchemaExport:153 - Running hbm2ddl schema export
    [java] 09:12:56,388 DEBUG SchemaExport:171 - import file not found: /import
    .sql
    [java] 09:12:56,388 INFO SchemaExport:180 - exporting generated schema to
    database
    [java] 09:12:56,403 DEBUG SchemaExport:283 -
    [java] alter table Child
    [java] drop constraint FK3E104FC976A59A
    [java] 09:12:56,466 DEBUG SchemaExport:283 -
    [java] drop table Child
    [java] 09:12:56,544 DEBUG SchemaExport:283 -
    [java] drop table Parent
    [java] 09:12:56,654 DEBUG SchemaExport:283 -
    [java] create table Child (
    [java] id numeric(19,0) identity not null,
    [java] name varchar(255) null,
    [java] parent_id numeric(19,0) null,
    [java] primary key (id)
    [java] )
    [java] 09:12:56,779 DEBUG SchemaExport:283 -
    [java] create table Parent (
    [java] id numeric(19,0) identity not null,
    [java] primary key (id)
    [java] )
    [java] 09:12:56,873 DEBUG SchemaExport:283 -
    [java] alter table Child
    [java] add constraint FK3E104FC976A59A
    [java] foreign key (parent_id)
    [java] references Parent
    [java] 09:12:56,952 INFO SchemaExport:200 - schema export complete
    [java] 09:12:56,952 WARN JDBCExceptionReporter:48 - SQL Warning: 5701, SQL
    State: 01000
    [java] 09:12:56,952 WARN JDBCExceptionReporter:49 - [Microsoft][ODBC SQL S
    erver Driver][SQL Server]Changed database context to 'master'.
    [java] 09:12:56,952 WARN JDBCExceptionReporter:48 - SQL Warning: 5703, SQL
    State: 01000
    [java] 09:12:56,952 WARN JDBCExceptionReporter:49 - [Microsoft][ODBC SQL S
    erver Driver][SQL Server]Changed language setting to us_english.
    [java] 09:12:56,983 INFO SessionFactoryImpl:432 - Checking 0 named queries
    [java] Hibernate:
    [java] insert
    [java] into
    [java] Child
    [java] (name)
    [java] values
    [java] (?) select
    [java] scope_identity()
    [java] Hibernate:
    [java] insert
    [java] into
    [java] Child
    [java] (name)
    [java] values
    [java] (?) select
    [java] scope_identity()
    [java] Hibernate:
    [java] insert
    [java] into
    [java] Child
    [java] (name)
    [java] values
    [java] (?) select
    [java] scope_identity()
    [java] Hibernate:
    [java] insert
    [java] into
    [java] Parent
    [java] default
    [java] values
    [java] select
    [java] scope_identity()
    [java] Hibernate:
    [java] update
    [java] Child
    [java] set
    [java] parent_id=?
    [java] where
    [java] id=?
    [java] Hibernate:
    [java] update
    [java] Child
    [java] set
    [java] parent_id=?
    [java] where
    [java] id=?
    [java] Hibernate:
    [java] update
    [java] Child
    [java] set
    [java] parent_id=?
    [java] where
    [java] id=?
    [java] 09:12:57,390 INFO SessionFactoryImpl:831 - closing
    [java] 09:12:57,390 INFO DriverManagerConnectionProvider:147 - cleaning up
    connection pool: jdbc:odbc:MySQL
    BUILD SUCCESSFUL
    Total time: 5 seconds
    Debug level Hibernate log excerpt:
    Included in the above description.

    That's not the right mapping for the 1:m relationship in Hibernate.
    First of all, I believe the recommendation is to have a separate .hbm.xml file for each class, so you should have one for Parent and Child.
    Second, you'll find the proper syntax for a one-to-many relationship here:
    http://www.hibernate.org/hib_docs/v3/reference/en/html/tutorial.html#tutorial-associations
    See if those help.
    The tutorial docs for Hibernate are quite good. I'd recommend going through them carefully.
    %

Maybe you are looking for

  • Addition of new field to customer master data

    Hi Is it possible to create a new field in customer master table screen  and to have the same field in sales order creation screen. can we do it in Ideas version of ECC. please guide me step by step. Regards Jai

  • BT FON How do i get rid of this C##p

    You will have to excuse me if i sound annoyed, thats because i am! I have this BT FON thing screwing up my connection and i cant turn it off. Firstly I didnt ask for this, and i dont want it. the notion that i have to opt out of this rather than into

  • Hyperlink in PDF form

    Hi all - I have some very simple web pages that I created in some word documents - I imported these docs into liveCycle to create PDF's.....works perfect except my hyperlinks don't work in the PDF.....any idea how I can get these to work?  Do I need

  • Cannot Access EAC/ECP on exchange server 2013 after installation

    Hi, I can not Cannot Access EAC/ECP on exchange server 2013 after installation . Error 404.4  code 0x80070002 Help! Thanks,

  • How can I stop folders on my desktop from automatically organizing?

    I am trying to figure out how to stop my folders from automatically organizing alphabetically as soon as I make a new folder on my desktop.