Toplink with mysql: problem with blob size

I'm using toplink with a mysql database. I want to store some data in a blob.
In mysql there exist different sizes of blobs (in my case I need a mediumblob).
But if I create the scheme for the database with jpa/toplink it alway creates only a column with type blob.
I can explicitly tell the database to use a mediumblob by this:
@Column("MEDIUMBLOB")
But by doing this I limit my program to mysql of course, as this data type is not known to other databases.
Does anybody know a more elegant solution to setting the blob size?
for example with hibernate it can be done this way:
@Column(length=666666)

Looks like you are using JPA, and in JPA you would set the columnDefinition to the type that you want, e.g.
@Lob
@Column(name="BLOBCOL", columnDefinition="MEDIUMBLOB")
byte[] myByteData;
As you mentioned, this does introduce a dependency on the database. However, you can always either put the Column metadata in XML, or override it with something else later in XML.
The length attribute was intended and specified for use with Strings but I guess there is no reason why it couldn't be used for BLOBs as well (please enter an enhancement request, or submit the code to add the feature to EclipseLink). Be aware, though, that doing so at this stage is going to be introducing a dependency on the provider to support a non-spec defined feature.

Similar Messages

  • Blob sizes in mySql

    MySQL supports different blob sizes. I'm told that kodo does not support
    blobs in MySQL however I can override this by extending the MySQLDictionary
    class.
    This is what I was told.....
    "package com.xyz;
    import java.sql.*;
    import kodo.jdbc.schema.*;
    import kodo.jdbc.sql.*;
    public class CustomMySQLDictionary
    extends MySQLDictionary
    protected String appendSize (Column col, String typeName)
    if (col.getType () == Types.BLOB && col.getSize () > 0)
    return <sized blob string>
    return super.appendSize (col, typeName);
    Plug your dictionary into Kodo with:
    kodo.jdbc.DBDictionary: com.xyz.CustomMySQLDictionary
    In your metadata, set the size of your field with:
    <field name="blobField">
    <extension vendor-name="kodo" key="jdbc-size" value="xxx"/>
    </field>
    I have done this with a couple of minor changes. It almost does what I
    need it to do.
    Basically I want to create an ANT script which will enhance the necessary
    files, create my database, and then dump the schema to a file(so that I
    can have a sql script to run later). I have everything working with the
    exception of the of creating the database. The database gets created
    however the column where I specify blob, still gets created as a blob. I
    want to use a different blob size.
    The above code helped me with dumping the schema to a file and dumps the
    appropriate data as I would expect it.
    So how do I get the database created properly? Also, is there a way to
    automatically generate the schema without actually creating the database
    first?

    Mike Krell wrote:
    MySQL supports different blob sizes. I'm told that kodo does not support
    blobs in MySQL however I can override this by extending the MySQLDictionary
    class.
    This is what I was told.....
    "package com.xyz;
    import java.sql.*;
    import kodo.jdbc.schema.*;
    import kodo.jdbc.sql.*;
    public class CustomMySQLDictionary
    extends MySQLDictionary
    protected String appendSize (Column col, String typeName)
    if (col.getType () == Types.BLOB && col.getSize () > 0)
    return <sized blob string>
    return super.appendSize (col, typeName);
    Plug your dictionary into Kodo with:
    kodo.jdbc.DBDictionary: com.xyz.CustomMySQLDictionary
    In your metadata, set the size of your field with:
    <field name="blobField">
    <extension vendor-name="kodo" key="jdbc-size" value="xxx"/>
    </field>
    I have done this with a couple of minor changes. It almost does what I
    need it to do.
    Basically I want to create an ANT script which will enhance the necessary
    files, create my database, and then dump the schema to a file(so that I
    can have a sql script to run later). I have everything working with the
    exception of the of creating the database. The database gets created
    however the column where I specify blob, still gets created as a blob. I
    want to use a different blob size.
    The above code helped me with dumping the schema to a file and dumps the
    appropriate data as I would expect it.
    So how do I get the database created properly? Also, is there a way to
    automatically generate the schema without actually creating the database
    first?After doing some further investigation, the call "col.getType ()" returns
    a Types.VARBINARY and not a Types.BLOB. So I changed my code to test for
    this instead and also test for the column size to indicate the appropriate
    BLOB size. Is this correct?
    It appears that a LONGBLOB is coming back as a "LONG VARBINARY", even
    though I am testing for this, I cannot get a LONGBLOB. Why?
    Also, prior posts indicate that BLOBs are not supported in Kodo and MySQL.
    I'm confused as to what this means because in the src files that you
    delivered with your product, I see references to BLOB, MEDIUMBLOB, etc.
    The file I'm referring to is kodo.jdbc.sql.MySQLDictionary. So tell me
    again why blobs aren't supported?

  • Problem with Cell size in Excel output of XML report

    Dear all,
    I am facing a problem with cell size when i run my XML report in Excel output. I found that it imitates the cell size of whatever i gave in the RTF. I cannot increase the cell size in RTF as my report contains 60 columns and max width of MS Word table is 22 inches.
    Can any one suggest a way of doing this which shows full data in Excel sheet depending on the column data size with out any word wrap.
    Thanks
    RAJ

    Hi ,
    You can try with
    <xsl:attribute xdofo:ctx="block" name="wrap-option">no-wrap</xsl:attribute>
    may be helpful to you
    Thanks,
    Ananth
    http://bintelligencegroup.wordpress.com/

  • I can't download the apps with big size like need for spees, but in apps with normal size(less than 100KB)there is no problem,is there any body who can help me?

    I can't download the apps with big size like need for spees

    Have a little patience. 4 minutes is only a short time for someone to find your post and reply to it. Also clicking on "This Solved My Problem" marks your problem as solved so many folks that may otherwise help will pass the question by since it is "solved."
    I suggest you start a new question that gives a good title to a specific problem. Then in the body of the thread describe the problem in detail, tell us what steps you have tried so far to correct the problem, what happened, and what error messages you received.

  • Problem with font size, not showing up as expected.

    I've been wrestling with this for a few weeks now, but surely must be getting closer to having it resolved. It's just a problem with font size on just this one page.
    http://www.kgngroup.net/logos_and_packaging.html
    I want the body text to be 10px, but after two weeks, have played with every conceivable way, altering body font size in CSS and in the document itself. I've viewing it in Safari but others on Windows machines tell me that it is out of whack also.
    My main problem is that I can change the font size to extremes but it still shows up as the same size in the browser (about 12px.) I know changes can be made because I can modify the wording and that does change online. It's just a problem with the font size. Oh, and in the third column, the one to the right, in Dreamweaver, the font looks like it is miniscule, cannot even be read, but in the browser, since the style11 is applied, all three columns look equal size.
    Not sure where to go from here. I just want to get the body font down to 10 pixels.
    TIA,
    Ken

    John Waller wrote:
    I guess this a follow up to your earlier thread
    http://forums.adobe.com/message/3627681
    What would really help you in diagnosing issues such as this is to install the Firebug add-in for Firefox (as suggested in the earlier thread).
    Firebug shows you visually what CSS rules apply to whatever element you click on the screen.
    In this case, you have a confusion of CSS font sizes cascading on top of each other and that's confused again by mixing units for no real reason. .style11 is unnecessarily applied to lots of individual <p> tags. Remove those classes and delete the rule.
    e.g.
    The tiny font in the RHS column is affected first by
    .style8 {
    font-size: 0.5em;
    affecting the <td> then by
    .style11 {font-size: 75%}
    affecting the <p> tag.
    End result = tiny text (approx 2px).
    In these cases of mass confusion, my approach is to strip out all font sizing from the CSS and start over.
    Avoid %. Use px.
    Apply font sizing rules logically starting from the <body> tag downwards.
    That'll give you a solid baseline without weird effects.
    Ah, I find your answer worth repeating.
    Very cool that you have diagnosed so well and spelled it out here.
    I will get the Firefox add-on you mention, plus, print out this response and follow it like a recipe.
    It makes perfect sense and you have saved me from the 'clutter' of being wound up in too many directions with this because I could not make all of the connections that you have done here. I appreciate your help very much.
    Thank You,
    Ken

  • Problem using Toplink with JUnit

    Hi,
    I have a problem using Toplink with JUnit. Method under test is very simple: it use a static EntityManager object to open a transaction and persists data to db. When I invoke the method from a test method, it gives me the exception:
    java.lang.AssertionError
         at oracle.toplink.essentials.ejb.cmp3.persistence.PersistenceUnitProcessor.computePURootURL(PersistenceUnitProcessor.java:248)
         at oracle.toplink.essentials.ejb.cmp3.persistence.PersistenceUnitProcessor.findPersistenceArchives(PersistenceUnitProcessor.java:232)
         at oracle.toplink.essentials.ejb.cmp3.persistence.PersistenceUnitProcessor.findPersistenceArchives(PersistenceUnitProcessor.java:216)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.initialize(JavaSECMPInitializer.java:239)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.initializeFromMain(JavaSECMPInitializer.java:278)
         at oracle.toplink.essentials.internal.ejb.cmp3.JavaSECMPInitializer.getJavaSECMPInitializer(JavaSECMPInitializer.java:81)
         at oracle.toplink.essentials.ejb.cmp3.EntityManagerFactoryProvider.createEntityManagerFactory(EntityManagerFactoryProvider.java:119)
         at javax.persistence.Persistence.createEntityManagerFactory(Persistence.java:83)
         at javax.persistence.Persistence.createEntityManagerFactory(Persistence.java:60)
         at it.valerio.electromanager.model.EntityFacade.<clinit>(EntityFacade.java:12)
         at it.valerio.electromanager.business.ClienteBiz.insertIntoDatabase(ClienteBiz.java:36)
         at it.valerio.electromanager.test.model.ClienteTest.insertDBTest(ClienteTest.java:30)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
         at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
         at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
         at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
         at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
         at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:66)
         at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
         at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
         at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
         at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
         at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:38)
         at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
         at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
         at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
         at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
         at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
    Where is the problem???
    Regards,
    Valerio

    EntityFacade class is very simple and it uses a static EntityManager object. Here the code:
    public class EntityFacade {
         private static EntityManager em = Persistence.createEntityManagerFactory("ElectroManager").createEntityManager();
         private static Logger logger=Logger.getLogger(EntityFacade.class);
         public static void insertCliente(Cliente c)
              logger.debug("Inserisco cliente nel db: " + c);
              em.getTransaction().begin();
              c.setId(getNextIdForTable("Cliente"));
              em.persist(c);
              em.getTransaction().commit();
    If I call the method from inside a main it works well, so I think the problem is not the classpath neither the URL in the persistence.xml. However the URL is:
    <property name="toplink.jdbc.url" value="jdbc:derby:c:/programmi/ElectroManager/db/electroManager"/>
    I use the latest build version of TopLink.
    Thanks.

  • Do you see problems with text size when importing from Photoshop?

    Hi all -
    We're working on improving some things in the Photoshop-to-Reflow workflow, and right
    now specifically investigating trouble with imported text.
    Aside from the desire to have support for local fonts, are there any problems you run into
    when importing text from Photoshop into Reflow? We've heard some issues with text size and position,
    but want to make sure we understand all of the cases.
    If you have problems in this area, please share them in this thread, and I'd also love to get some
    sample PSDs that demonstrate the problems. (You can send them directly to me; vmitnick at adobe dot com .)
    thanks,
    Vic

    I'm using CS4.. and I'm using the standard "Save As" .pdf features....
    and I'm reading it with adobe acrobat pro 9.1...
    i think it has something to do with compression. but i can't figure out how to solve it. converting the text to symbols maybe?

  • Problem with export size @ Premiere CS3

    Hello, I have a problem that until recently did not bother me so much and that the final export @ the premierCS3 ( also @ CS4 and CS5 )
    The films I make are 5-10 minutes long, so usualy the final size is big, but still not THAT BIG! and i can work with this size.
    Lately  I've edited a feature film and the final export is 25GB and it simply  intolerable that even after encodings using simple software I got to 7Gb and quality went down significantly.
    How do I solve this trouble? I do not think that movie that long (1h:45m) should be bigger than 1GB!!
    Hope you can help me
    Wishing you a good week,
    ROEi.

    Go to the PPro FAQ List http://forums.adobe.com/community/premiere/faq_list and read about export options
    A DV AVI file is about 13Gig per hour... export size is determined by the CODEC and compression used, plus the data rate
    Some CODECS have better compression than others, and a higher data rate will create a larger file
    You need to tell what CODEC you are using for export, and what data rate

  • Problem with model size in Visual Composer

    Hello madam/sir,
    We are using ECC 5.0 and VC 700.8.0.1. We got error message during
    deployment "Error in compiling Flex application. Branch between 19217
    and 228774 around line 0 exceeds 32K". When we delete several blocks in
    the model, it works good. So tried to delete different blocks so we
    think that the problem is not with specific block but with model size.
    Our model is in one iView. We cannot use additional nested
    iViews because we have HTML View blocks that cannot be in
    nested iView (according to VC documentation). Is there any solution for
    this error? What does it mean?
    Best Regards,
    Sergey Kozyrev

    I agree with Ashutosh, you have to remodeling your model. You can also try to restart VC and deploy your model again, this might also help sometimes (I had sometimes the issue, that I removed any components and then I get the 32K error, after restarting the error disappeared).
    You can also check if you have any global entry list, which you do not use in your model, this also could help to reduce you model size.
    Best Regards,
    Marcel

  • Problem with BLOB data type and acute characters

    Hi all, I have an issue related with BLOBS cols:
    I have table A with a BLOB column and I need to insert that data into another table B, the thing is I use dbms_lob.read() that returns data into a RAW variable and then I convert it to a varchar2(with utl_raw.cast_to_varchar2) and it works ok if I have no "acute characters" on it, e.g Québec.
    If I have acute characters, the "é" is inserted wrongly with an "¿" character on the table B. I tested the BLOB on the table A to check the data and the BLOB is showing correctly the "é" character, it seems that the dbms_lob.read() funct cannot handle accute characters and replace them with "¿".
    How can I replace acute characters when I use the dbms_lob.read functionality?
    Thank you!

    what's the OS of the db server? is this code an anonymous pl/sql block, or stored pl/sql? does it run on the server or a client? if it's on the client, what's the o/s of the client?
    you might want to look at the following, plus the first 3 notes in it's reference section
    https://metalink.oracle.com/metalink/plsql/f?p=130:14:3038746456038321058::::p14_database_id,p14_docid,p14_show_header,p14_show_help,p14_black_frame,p14_font:NOT,264157.1,1,1,1,helvetica

  • Weird Problem with BLOB CMP field

    Hi,
    I am trying to deploy an EJB2.0 CMP Bean in Weblogic6.1 .The Bean has one of the CMP Field as a serializable object mapped to a BLOB Datatype in Oracle8i
    When i try to creae this Entity Bean
    I get "java.io.IOException: ORA-22920: row containing the LOB value is not locked" as a nested exception.
    When i increased the isolation level . I got "java.io.IOException: ORA-01002 fetch out of sequence" as the nested exception
    And one more horrible thing that is happening is the other arguments to the create method are getting updated in the DB except the Blob field(Supposed to rollback? when there is an EJB Exception ??)
    Any thoughts? Comments?
    Regards
    Sathya

    Hi,
    I seem to recall that there are multiple issues with BLOBS & WLS. I'll describe my setup, and you can see if your missing anything:
    - In weblogic-cmp-rdbms-jar.xml you need to set the column type as a Blob:
    <field-map>
    <cmp-field>binary</cmp-field>
    <dbms-column>binary</dbms-column>
    <dbms-column-type>OracleBlob</dbms-column-type>
    </field-map>
    - In weblogic-ejb-jar.xml you may need to set the isolation level of your set method:
    <transaction-isolation>
    <isolation-level>TRANSACTION_READ_COMMITTED_FOR_UPDATE</isolation-level>
    <method>
         <ejb-name>TerminalSoftwareKernelBDO</ejb-name>
         <method-name>setBinary</method-name>
    </method>
    </transaction-isolation>
    - The bean get/set methods should look like this:
    public abstract void setBinary(java.lang.Object binary);
    public abstract java.lang.Object getBinary();
    - In WLS 6.1 up to Service Pack 2 (I don't know about SP3 yet), BLOBs are incorrectly implemented. For some (odd) reason you can only put serialized objects into the DB... I think this means that if you have a BLOB put in by anything but WLS, WLS get very unhappy. So empty out your table with the blobs first.
    Hope that helps.
    Daniel.

  • Problem while importing table with blob datatype

    hi i am having a database 9i on windows xp and dev database 9i on AIX 5.2
    while i am taking export of normal tables and trying to import i am successful.but when i am trying to import a table with blob datatype it is throwing "tablespace <tablespace_name> doesn't exist" error
    here how i followed.
    SQL*Plus: Release 9.2.0.1.0 - Production on Mon Oct 8 14:08:29 2007
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Enter user-name: test@test
    Enter password: ****
    Connected to:
    Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    SQL> create table x(photo blob);
    Table created.
    exporting:
    D:\>exp file=x.dmp log=x.log tables='TEST.X'
    Export: Release 9.2.0.1.0 - Production on Mon Oct 8 14:09:40 2007
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Username: pavan@test
    Password:
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    About to export specified tables via Conventional Path ...
    Current user changed to TEST
    . . exporting table X 0 rows exported
    Export terminated successfully without warnings.
    importing:
    D:\>imp file=x.dmp log=ximp.log fromuser='TEST' touser='IBT' tables='X'
    Import: Release 9.2.0.1.0 - Production on Mon Oct 8 14:10:42 2007
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Username: system@mch
    Password:
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.6.0 - 64bit Production
    With the Partitioning, Real Application Clusters, OLAP and Oracle Data Mining op
    tions
    JServer Release 9.2.0.6.0 - Production
    Export file created by EXPORT:V09.02.00 via conventional path
    Warning: the objects were exported by PAVAN, not by you
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    import server uses US7ASCII character set (possible charset conversion)
    . importing TEST's objects into IBT
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE TABLE "X" ("PHOTO" BLOB) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS "
    "255 STORAGE(INITIAL 65536 FREELISTS 1 FREELIST GROUPS 1) TABLESPACE "TESTTB"
    "S" LOGGING NOCOMPRESS LOB ("PHOTO") STORE AS (TABLESPACE "TESTTBS" ENABLE "
    "STORAGE IN ROW CHUNK 8192 PCTVERSION 10 NOCACHE STORAGE(INITIAL 65536 FREE"
    "LISTS 1 FREELIST GROUPS 1))"
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'TESTTBS' does not exist
    Import terminated successfully with warnings.
    why it is happening for this table alone?plz help me
    thanks in advance

    Here is exerpt from {
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:378418239571}
    =============================================
    Hi Tom,
    I have a dump file containing blob datatypes, when i import the dump file in a schema it gives an
    error stating that the tablespace for Blob datatype does not exists. My question is how do i import
    the dump file in the default tablespace of the importing user.
    Followup March 2, 2004 - 7am US/Eastern:
    You'll have to precreate the table.
    do this:
    imp userid=u/p tables=that_table indexfile=that_table.sql
    edit that_table.sql, fix up the tablespace references to be whatever you want to be, run that sql.
    then imp with ignore=y
    for any MULTI-SEGMENT object (iot's with overflows, tables with lobs, partitioned tables, for
    example), you have to do this -- imp will not rewrite ALL of the tablespaces in that
    multi-tablespace create -- hence you either need to have the same tablespaces in place or precreate
    the object with the proper tablespaces.
    Only for single tablespace segments will imp rewrite the create to go into the default if the
    requested tablespace does not exist.
    ===================================================
    To summarize: precreate target table when importing multi-segment tables

  • Sample insert into table with BLOB column

    This is my first opportunity to work with BLOB columns. I was wondering if anyone had some sample code that shows the loading of a blob column and inserted into the table with a blob column.
    I have to produce a report (including crlf) and place it into a blob column. The user will download the report at a later time.
    Any suggestions / code samples are greatly appreciated!!!!

    You can enable string binding in TopLink.
    login.useStringBinding(int size);
    or you could enable binding in general,
    login.bindAllParameters();

  • Need  URGENT help with BLOB!!!!

    Hi,
    I am using a certain java server that has the capability of persisting objects to permanent storage (Oracle 8.1.5.0 in my case).
    The server automatically creates all necessary tables, and some of them have columns of BLOB datatypes in them.
    Then server retrieves a row of data and attempts to figure out how big is the size of the single BLOB record.
    All these queries are in xml configuration files and the server's java code uses them to access the database. I can not modify the datatypes during table creation without modifying actual server code, which I don't want to do for obvious reasons. I can't use stored procedures to figure out the size of the BLOB either for the same reason.
    Now here is the problem:
    Some of the logic of the code depends on queries that are supposed to return the size of BLOB data. Queries look like this:
    SELECT DATALENGTH(col_a)
    FROM table_A
    WHERE col_b = 'something'
    and of course col_a is in table_A and its datatype is BLOB (The sql above works on MS SQL).
    IS THERE AN EQUIVALENT OF datalength() IN ORACLE?
    I have unsuccessfully tried LENGTH(blob), which throws ORA-00932 inconsistent datatypes error. Also tried octet-length(), BLOB SIZE() with same results (different error: ORA-00923 FROM keyword not found where expected).
    Any ideas?
    Your help is greatly appreciated,
    Gugo M

    Thank you so much!
    Looks very simple (probably is for a dba).
    Thanks again.
    Gugo

  • NoSuchElementException when using sequence table in Toplink with Oracle 9i

    Hi,
    UnitOfWork(11608452)--Connection(15933498)--UPDATE SEQUENCE SET SEQ_COUNT = SEQ_COUNT + 1000 WHERE SEQ_NAME = 'ABN_NOTIF_SEQ'
    UnitOfWork(11608452)--Connection(15933498)--SELECT SEQ_COUNT FROM SEQUENCE WHERE SEQ_NAME = 'ABN_NOTIF_SEQ'
    ClientSession(17794660)--Connection(15933498)--commit transaction
    java.util.NoSuchElementException
         at java.lang.Throwable.fillInStackTrace(Native Method)
         at java.lang.Throwable.fillInStackTrace(Compiled Code)
         at java.lang.Throwable.<init>(Compiled Code)
         at java.lang.Exception.<init>(Compiled Code)
         at java.lang.RuntimeException.<init>(Compiled Code)
         at java.util.NoSuchElementException.<init>(NoSuchElementException.java:35)
         at java.util.Vector.firstElement(Compiled Code)
         at TOPLink.Private.DatabaseAccess.DatabasePlatform.sequenceTableGetSequenceNumberNamed(Compiled Code)
         at TOPLink.Private.DatabaseAccess.DatabasePlatform.getSequenceNumberNamed(Compiled Code)
         at TOPLink.Public.PublicInterface.Session.getNextSequenceNumberValue(Compiled Code)
         at TOPLink.Public.PublicInterface.UnitOfWork.getNextSequenceNumberValue(Compiled Code)
         at TOPLink.Private.Descriptors.ObjectBuilder.assignSequenceNumber(Compiled Code)
    This is the error I am getting occasionally. I am using Toplink with custom sequence table and pre-allocation size is 1000.
    Can you Please advise me on this problem.
    Thanks in Advance,
    Durga

    hello,
    i m using msync for synchronization on Pocket PC 2003 with eVB. I dont know about msyncCom.
    i have mentioned a part of my code where the problem is found.
    rsmastbl.Open tblnm, conn, adOpenKeyset, adLockOptimistic
    MsgBox tblnm & " opened"
    'get the key fields to be compared in the query to get required record
    MsgBox rsmastbl.Fields(1).name
    keyfldnm1 = rsmastbl.Fields(1).name
    MsgBox "keyfldnm got value = " & keyfldnm1
    rsmastbl.Close
    MsgBox "keyfldnm got value = " & keyfldnm1
    Now here the value of keyfldnm1 before closing the record set is found to be correct.
    But the message box after closing the record set shows no value in the keyfldnm1.
    its very urgent....
    if anybody can help me
    thanx

Maybe you are looking for