BPC 7 Logfile too large on database drive

I have an LDF file that has grown quite large on my database server running SQL Server 2008 and BPC 7M SP5
F:\Microsoft SQL Server\MSSQL10.MSSQLSERVER\MSSQL\DATA\%APPSET%_log.ldf
What is the best method to reduce the size of this file?
Does anyone know if this logfile is associated with the logging configured on Management console, or is this purely a SQL log that "Shrinking" the database should reduce?
Best regards,
Greg

Hi,
With SQL Server you can set up your appset DB with Simple or Full Recovery mode (edit DB properties, it' in one of the properties pane).
When you are on Full recovery mode (wich is your case), SQL server will record all changes on your database (for exemple each time there is a DELETE or UPDATE SQL transaction) in the LDF file.
It's why the LDF will grow.
When you configure your database with full recovery mode (wich is the default configuration when you create a new appset), you must backup and shrink them as often as it's needed (each hours).
Beside that, the real question is "do you need to configure your DB with a full recovery mode ?"
=> Full recovery mode is only needed if you want to restore the most up to date data in case of db corruption.
If your business can deal with the last full backup (wich you may create each night), then you don't really need to be in Full Recovery mode.
So :
Simple recovery mode : You can only restore the last full backup. Meaning that you could lost around 1 day of data.
Full recovery mode : you can restore all your data at the time of the last LDF backup but this need extra administration and space on your hard drives.

Similar Messages

  • I get an error message File too large for formatted drive when trying to copy a large video file (~50GB) to a new 64 GB USB drive.  Any thoughts?

    I am copying from a MAC to the new drive - not sure the format of the USB drive, whatever the SANDISK Cruzer is default formatted to..

    You state that Cruzer is formatted which is proably FAT32. FAT32 has limits on file size of 4 GB.
    Since you are copying larger files then you will need to reformat.
    Allan

  • I have a 500 GB hard drive and a 1TB Time Capsule running on a MacBook Pro.  It was all working well until the MacBook went in for a repair a week or so ago.  Since then, TC will not perform a backup;  instead, it says the backup is too large for the disk

    Since having my MacBook Pro repaired (for a video problem) Time Capsule returns the following message:  "This backup is too large for the backup disk. The backup requires 428.08 GB but only 192.14 GB are available."
    I notice that there is also a new sparse bundle.
    Since TC has my ONLY backup (going back about 4 years) I am reluctant to wipe it and start over fresh as I am afraid of losing files. 
    Is there a way of dealing with this?
    I am using Snow Leopard 10.6.8

    The repair shop likely replaced a major circuit board on your MacBook Pro, so Time Machine thinks that you have a "new" computer and it wants to make a new complete backup of your Mac.
    You are going to have to make a decision to either add another new Time Capsule....or USB drive to your existing Time Capsule....and in effect start over with a new backup of your Mac and then move forward again.
    For "most" users, I think this is probably the best plan because you preserve all your old backups in case you need them at some point, and you start over again with a new Time Capsule so you have plenty of room for years of new backups.
    Or, as you have mentioned, you have the option of erasing the Time Capsule drive and starting all over again. The upside is that you start over and have plenty of room for new backups. The downside is that you lose years of backups.
    Another option....trying to manually delete old backups individually....is tricky business....and very time consuming. To get an idea of what is involved here, study this FAQ by Pondini, our resident Time Capsule and Time Machine expert on the Community Support area. In particular, study the pink box.
    http://web.me.com/pondini/Time_Machine/12.html
    Once you look through this, I think you may agree that this type of surgery is not for the faint of heart.  I would suggest that you consider this only if one of the other options just cannot work for you.

  • How can I put my recently purchased iTunes movies on to my thumb drive and WD My Passport external hard drive. My Mac keeps saying "can't be copy because it is too large for the volumes format

    How can I put my recently purchased iTunes movies onto my thumb drive and WD my passport external hard drive.My Mac is saying that can't be copied because it is too large for volumes format

    [file] is too large for the volumes format
    That is usually an indication that the Volume was never Erased or Partitioned for GUID partition table and Mac OS, and is still in a Windows format.
    Mac OS X Extended (journaled) has no such file-size limitations

  • HT1229 My iPhoto Library (version 8.1.2) is 280GB (greater than 50% of my 500GB total storage memory on my iMac.  It was too large for me to drag it to a new hard drive so the Apple geniuses did it for me.  However they did not delete the Library from my

    My iPhoto Library (version 8.1.2) is 280GB (greater than 50% of my 500GB total storage memory on my iMac.  It was too large for me to drag it to a new hard drive so the Apple geniuses did it for me.  However they did not delete the Library from my iMac (that's my responsibility).  I dragged it to Trash and when it started to move I clicked over to the new hard drive to confirm it had indeed been copied.  I became nervous when I didn't see among the few files on this otherwise empty new hard drive anything that resembled a 280GB Library so I cancelled the migration to trash.
    How can I be sure that my iphoto has been copied and that all my "metadata" survived in tact?

    the new backup drive
    I thought the new drive would be your data drive to host the iPhoto library. Do you also use it for TimeMachine backups?
    I am unable to search either in email as well as Finder.  I AM able to search within iPhoto though, thankfully
    Spotlight may still be busy rebuildig its index.
    You could try to rebuild the Spotlight Index, if you do see no progress:
    Spotlight: How to re-index folders or volumes
    I hope, other frequent posters will drop in. I have not used iPhoto 8.x in long time.

  • I am having trouble with a video.  I have tried to copy it to a flash drive because it is too big to fit on a DVD, it says it is too large for the volumes format.  Please advise.

    I have done many DVD's for seniors for graduation so I am familier with Imovie.  I have just finished a projec that is a combination of video, music and stills.  I shared it through idvd.  Now I can't burn it to a dvd because it is too large.  It is too big even for a two sided disk.  I have researched it and i find that blue ray would be big enough.  Does apple offer  blue ray burner?  if so, what is the cost?  I tried another route and took my external hard drive to copy it to a PC.  there computer wouldn't recongize my drive.  I am assuming it is because it is formated for apple.  I tired copying it to a flash drive and and it says that it can't be copied because it is too large for the volume's format.  Suggestions?   PLEASE!!!!!

    Could the video be too big because of using too high a resolution?   480p is sufficient for a DVD.  the thread at:
    https://discussions.apple.com/thread/5868958?tstart=30
    may be of help.
    Apple does not support Blue ray - you have to use a thrid party application like Toast.
    To transfer the video to a PC you need a FAT-formatted hard drive (or of course a sufficiently big memory stick).
    You could also go via a network, via the cloud (eg Dropbox) or with 'Wetransfer" or similar.
    Geoff.

  • My time Machine keeps saying, "Time Machine could not complete the backup. This backup is too large for the backup disk. The backup requires 345.74 GB but only 289.80 are available." I have already excluded files. I have a 1tb external drive. HELP!!!

    For over two weeks now I have been frustated and not having my TIme Machine back up to my 1tb external drive. I dont understand why now its a problem.  It keeps saying
    "This backup is too large for the backup disk. The backup requires 345.74GB but only 289.80GB are avialable.  Time Machine needs work space on the bakup disk, in addition to the space required to store backups. Open Time Machine preferences to select a large backup disk or make the bakup smaller by excluding files." So I have already excluded almost all of my files, and even deleted the backup disk yet, that quote still keeps popping up. I am truly at a wall with this. I have a Mac OS X version 10.7.5. CAN SOMEONE HELP ME PLEASE????

    If you have more than one user account, these instructions must be carried out as an administrator.
    Launch the Console application in any of the following ways:
    ☞ Enter the first few letters of its name into a Spotlight search. Select it in the results (it should be at the top.)
    ☞ In the Finder, select Go ▹ Utilities from the menu bar, or press the key combination shift-command-U. The application is in the folder that opens.
    ☞ Open LaunchPad. Click Utilities, then Console in the icon grid.
    Make sure the title of the Console window is All Messages. If it isn't, select All Messages from the SYSTEM LOG QUERIES menu on the left. If you don't see that menu, select
    View ▹ Show Log List
    from the menu bar.
    Enter the word "Starting" (without the quotes) in the String Matching text field. You should now see log messages with the words "Starting * backup," where * represents any of the words "automatic," "manual," or "standard." Note the timestamp of the last such message. Clear the text field and scroll back in the log to that time. Select the messages timestamped from then until the end of the backup, or the end of the log if that's not clear. Copy them (command-C) to the Clipboard. Paste (command-V) into a reply to this message.
    If there are runs of repeated messages, post only one example of each. Don't post many repetitions of the same message.
    When posting a log extract, be selective. Don't post more than is requested.
    Please do not indiscriminately dump thousands of lines from the log into this discussion.
    Some personal information, such as the names of your files, may be included — anonymize before posting.

  • Alter mount database failing: Intel SVR4 UNIX Error: 79: Value too large for defined data type

    Hi there,
    I am having a kind of weird issues with my oracle enterprise db which was perfectly working since 2009. After having had some trouble with my network switch (replaced the switch) the all network came back and all subnet devices are functioning perfect.
    This is an NFS for oracle db backup and the oracle is not starting in mount/alter etc.
    Here the details of my server:
    - SunOS 5.10 Generic_141445-09 i86pc i386 i86pc
    - Oracle Database 10g Enterprise Edition Release 10.2.0.2.0
    - 38TB disk space (plenty free)
    - 4GB RAM
    And when I attempt to start the db, here the logs:
    Starting up ORACLE RDBMS Version: 10.2.0.2.0.
    System parameters with non-default values:
      processes                = 150
      shared_pool_size         = 209715200
      control_files            = /opt/oracle/oradata/CATL/control01.ctl, /opt/oracle/oradata/CATL/control02.ctl, /opt/oracle/oradata/CATL/control03.ctl
      db_cache_size            = 104857600
      compatible               = 10.2.0
      log_archive_dest         = /opt/oracle/oradata/CATL/archive
      log_buffer               = 2867200
      db_files                 = 80
      db_file_multiblock_read_count= 32
      undo_management          = AUTO
      global_names             = TRUE
      instance_name            = CATL
      parallel_max_servers     = 5
      background_dump_dest     = /opt/oracle/admin/CATL/bdump
      user_dump_dest           = /opt/oracle/admin/CATL/udump
      max_dump_file_size       = 10240
      core_dump_dest           = /opt/oracle/admin/CATL/cdump
      db_name                  = CATL
      open_cursors             = 300
    PMON started with pid=2, OS id=10751
    PSP0 started with pid=3, OS id=10753
    MMAN started with pid=4, OS id=10755
    DBW0 started with pid=5, OS id=10757
    LGWR started with pid=6, OS id=10759
    CKPT started with pid=7, OS id=10761
    SMON started with pid=8, OS id=10763
    RECO started with pid=9, OS id=10765
    MMON started with pid=10, OS id=10767
    MMNL started with pid=11, OS id=10769
    Thu Nov 28 05:49:02 2013
    ALTER DATABASE   MOUNT
    Thu Nov 28 05:49:02 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Trying to start db without mount it starts without issues:
    SQL> startup nomount
    ORACLE instance started.
    Total System Global Area  343932928 bytes
    Fixed Size                  1280132 bytes
    Variable Size             234882940 bytes
    Database Buffers          104857600 bytes
    Redo Buffers                2912256 bytes
    SQL>
    But when I try to mount or alter db:
    SQL> alter database mount;
    alter database mount
    ERROR at line 1:
    ORA-00205: error in identifying control file, check alert log for more info
    SQL>
    From the logs again:
    alter database mount
    Thu Nov 28 06:00:20 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Thu Nov 28 06:00:20 2013
    ORA-205 signalled during: alter database mount
    We have already checked in everywhere in the system, got oracle support as well without success. The control files are in the place and checked with strings, they are correct.
    Can somebody give a clue please?
    Maybe somebody had similar issue here....
    Thanks in advance.

    Did the touch to update the date, but no joy either....
    These are further logs, so maybe can give a clue:
    Wed Nov 20 05:58:27 2013
    Errors in file /opt/oracle/admin/CATL/bdump/catl_j000_7304.trc:
    ORA-12012: error on auto execute of job 5324
    ORA-27468: "SYS.PURGE_LOG" is locked by another process
    Sun Nov 24 20:13:40 2013
    Starting ORACLE instance (normal)
    control_files = /opt/oracle/oradata/CATL/control01.ctl, /opt/oracle/oradata/CATL/control02.ctl, /opt/oracle/oradata/CATL/control03.ctl
    Sun Nov 24 20:15:42 2013
    alter database mount
    Sun Nov 24 20:15:42 2013
    ORA-00202: control file: '/opt/oracle/oradata/CATL/control01.ctl'
    ORA-27037: unable to obtain file status
    Intel SVR4 UNIX Error: 79: Value too large for defined data type
    Additional information: 45
    Sun Nov 24 20:15:42 2013
    ORA-205 signalled during: alter database mount

  • Is the final cut pro application too large for my 13" mac pro or should I install it on an external drive?

    Is the final cut pro application too large for my mac book pro laptop? Should I install it on a hard drive?

    4 GB hard drive!?!  It's a memory size not hard drive size   Your hard drive is much bigger than 4 GB . If you have 2011 model then hard drive size is 500GB or 750 GB HDD. You can install final cut pro and many other programs .

  • BPC 7M SP4 EVDRE missing rows - Error is  "1004-Selection is too large.}

    Hello,
    On a customer who installed BPC 7 Ms SP4 I have on client Exception log the error:
    ===================[System Error Tracing]=====================
    [System Name]   : BPC_ExcelAddin
    [Job Name]         : clsExpand::applyDataRangeFormula
    [DateTime]          : 2009-07-17 09:44:13
    [Exception]
           Detail<sg     : {1004-Selection is too large}
    ===================[System Error Tracing End ]=====================
    When I see this error there is an EVDRE input schedule expanding which have some of the rows missing rows descriptions.
    This means there is a gap of missing row header formulas stating from second row to somewhere in the middle of the report.
    If I reduce the number of the resulting rows for the same EVDRE input schedule results are ok.
    Do you know some setting to fix this?
    I tried increasing the Maximum Expansion Limit for rows and columns in Workbook Options without success.
    Thank you.

    Hi all,
    I have the same problem with 7.0MS SP07. With 59 expanded members it works. With 60 expanded members it fails.
    Mihaela, could you explain us what is the purpose of the parameters you talk about?
    thanks,
    Romuald

  • I placed my aperture library on a 1 TB external drive (the file is 300 GB) and Aperture churns and churns but will not open the file. I don't know if the file is too large or if it is the external drive or what? any ideas? Smaller files will open.

    I placed my aperture library on a 1 TB external drive (the file is 300 GB) and Aperture churns and churns but will not open the file. I don't know if the file is too large or if it is the external drive or what? any ideas? Smaller files will open. This is very annoying.

    If this is an external USB drive, you must understand that USB is really, really slow. There may be nothing wrong except for that.
    In my opinion, USB external drives are only acceptable for backups, and to get "near Hard Drive" performance, you need FireWire.

  • My audit database getting too large

    Post Author: amr_foci
    CA Forum: Administration
    my audit database getting too large, how i reset it?

    Post Author: jsanzone
    CA Forum: Administration
    Amr,
    The best that I can determine, there is no official documentation from BusinessObjects regarding a method to "trim" the Auditor database.  Based on previous disucssions, I seem to remember that you are on XI R2, but if I'm wrong, then these notes will not apply to you.  Here is the scoop:
    There are six tables used by Auditor: 1) APPLICATION_TYPE (initialized w/ 13 rows, does not "grow") 2) AUDIT_DETAIL (tracks activity at a granular level, grows) 3) AUDIT_EVENT (tracks activity at a granular level, grows) 4) DETAIL_TYPE (initialized w/ 28 rows, does not "grow") 5) EVENT_TYPE (initialized w/ 41 rows, does not "grow") 6) SERVER_PROCESS ( (initialized w/ 11 rows, does not "grow")
    If you simply want to remove all audit data and start over, then truncate AUDIT_EVENT and AUDIT_DETAIL.
    If you want to only remove rows based on a period, then consider that the two tables, AUDIT_DETAIL and AUDIT_EVENT, are transactional in nature, however, AUDIT_DETAIL is a child to the parent table AUDIT_EVENT, thus you will want to remove rows from AUDIT_DETAIL based on its link to AUDIT_EVENT before removing rows from AUDIT_EVENT first.  Otherwise, rows in AUDIT_DETAIL will get "orphaned" and never be of any use to you, and worse, you will not readily know how to ever delete these rows again.
    Here is the SQL statements:delete from AUDIT_DETAILwhere event_id =(select Event_ID from AUDIT_EVENT                  where Start_Timestamp between '1/1/2006' and '12/31/2006')godelete from AUDIT_EVENT                  where Start_Timestamp between '1/1/2006' and '12/31/2006'go
    One word of caution is to down you BOE application before doing this maintenance work, otherwise there is a possibility that Auditor will be busy trying to bring new rows to your database while you're busy delete rows and you might encounter an unwanted table lock, either on the work you're doing or the work that BOE is trying to perform.
    Good luck!

  • Insane "backup is too large" message, 4x total space available on drive.

    I've read a large number of the "backup is too large" messages in this forum, but none of them seem to explain the message I am getting:
    "This backup is too large for the backup volume...requires 2460.3 GB but only 731.5 GB are available."
    Under Options, the only disk included in the backup is the internal hard disk (250GB) and the "Total Included" is 161.3 GB as displayed in this screen.
    Since I can backup my "entire" disk 4 times in the space remaining, this doesn't make any sense.
    Suggestions of blowing away and starting over would just be completely counterproductive to the idea of having a backup system and I really need to fix this issue and not lose the archived history which I reference (regularly).
    The time machine backup is stored on a Mirrored 2x1TB system, so only 270GB is currently being taken up by backups.
    Thanks in advance for any thoughts,
    Barney

    All of the stuff I've read seems to indicate that Time Machine should be able to sort this issue out normally.
    Yes, but you should see the long "preparing" and "deep traversal" only for a good reason -- if there's no crash, the incremental backups should be pretty quick.
    In this case there was a crash, which likely is in the vicinity of when the last backup succeeded, so I do confirm this as the likely cause.
    I'm wondering if one of those crashes managed to corrupt the File System Event Store. That might explain what's going on, as TM estimates how much space it's going to need from there.
    I run drive repair fairly regularly and have just before and just after this crash...its amazing how often permissions get scrambled even when there are no crashes or strange events...but that's another topic.
    The other thing that repeated crashes might do is, corrupt your disk(s). I'd recommend a +Repair Disk+ on your TM volume and a +Verify Disk+ on your internal HD. If the internal needs repair, you'll have to boot from your Leopard disc and use it's copy of Disk Utility.
    I've never had anything other than Permissions issues with the internal drive. I've had external drive failures, but recently replaced all of my old drives (hence the raid sets) so this should not be a major issue in the future.
    Let me ask this: you say it crashes a lot, but are these OS crashes or programs that get into tight loops that can't be force-quit, so you do a force power-off?
    A lot is probably an extreme statement, but I'm contrasting it to how stable Mac traditionally is (I develop in all three of the major universes). I have to reboot a few times a day due to instability, but sudden stop/force shutdown is more like once every couple of weeks on average.
    Also, rarely is it my programming that directly contributes to the crash or force quite situation. Its normally strictly an OS crash [grey quit unexpectedly box] or other software hang [can't force quite or kill]. I just attribute this to the heavy load that the computer is constantly under and the influx of relatively unstable development tools (XCode alone crashes a few times a day in the middle of work, and its a pretty stable piece of software).
    I do hate to say it, but if you can't reduce the frequent crashes, you may want to use a different backup app, such as CarbonCopyCloner (which I use in addition to TM), SuperDuper!, or the like. Their incremental backups run much longer than a "normal" TM incremental backup, as they don't use the internal log, but also won't get confused by a corrupt FSEventStore.
    The Time Machine has really been a dream come true for development. I use source control and offline storage and full backups in addition to TM, but TM brings with it an incredible ease of browsing and analying file history and progression. I could see it being better, but I've found few things as easy to use and readily accessible as TM.
    I'll run the tests from your previous posts since that has the added benefit of not losing the backup data. If its possible to just reset the system when something freaks out then I'm okay with that. I'd just hate to constantly reset all the history with it as well.
    Thanks again for all of the feedback and I'll report back soon on success or failure.
    Barney

  • TIme Machine  backup grows too large during backup process

    I have been using Time Machine without a problem for several months, backing up my imac - 500GB drive with 350g used. Recently TM failed because the backups had finally filled the external drive - 500GB USB. Since I did not need the older backups, I reformatted the external drive to start from scratch. Now TM tries to do an initial full backup but the size keeps growing as it is backing up, eventually becoming too large for the external drive and TM fails. It will report, say, 200G to back up, then it reaches that point and the "Backing up XXXGB of XXXGB" just keeps getting larger. I have tried excluding more than 100GB of files to get the backup set very small, but it still grows during the backup process. I have deleted plist and cache files as some discussions have suggested, but the same issue occurs each time. What is going on???

    Michael Birtel wrote:
    Here is the log for the last failure. As you see it indicates there is enough room 345g needed, 464G available, but then it fails. I can watch the backup progress, it reaches 345G and then keeps growing till it give out of disk space error. I don't know what "Event store UUIDs don't match for volume: Macintosh HD" implies, maybe this is a clue?
    No. It's sort of a warning, indicating that TM isn't sure what's changed on your internal HD since the previous backup, usually as a result of an abnormal shutdown. But since you just erased your TM disk, it's perfectly normal.
    Starting standard backup
    Backing up to: /Volumes/Time Machine Backups/Backups.backupdb
    Ownership is disabled on the backup destination volume. Enabling.
    2009-07-08 19:37:53.659 FindSystemFiles[254:713] Querying receipt database for system packages
    2009-07-08 19:37:55.582 FindSystemFiles[254:713] Using system path cache.
    Event store UUIDs don't match for volume: Macintosh HD
    Backup content size: 309.5 GB excluded items size: 22.3 GB for volume Macintosh HD
    No pre-backup thinning needed: 345.01 GB requested (including padding), 464.53 GB available
    This is a completely normal start to a backup. Just after that last message is when the actual copying begins. Apparently whatever's happening, no messages are being sent to the log, so this may not be an easy one to figure out.
    First, let's use Disk Utility to confirm that the disk really is set up properly.
    First, select the second line for your internal HD (usually named "Macintosh HD"). Towards the bottom, the Format should be +Mac OS Extended (Journaled),+ although it might be +Mac OS Extended (Case-sensitive, Journaled).+
    Next, select the line for your TM partition (indented, with the name). Towards the bottom, the Format must be the same as your internal HD (above). If it isn't, you must erase the partition (not necessarily the whole drive) and reformat it with Disk Utility.
    Sometimes when TM formats a drive for you automatically, it sets it to +Mac OS Extended (Case-sensitive, Journaled).+ Do not use this unless your internal HD is also case-sensitive. All drives being backed-up, and your TM volume, should be the same. TM may do backups this way, but you could be in for major problems trying to restore to a mis-matched drive.
    Last, select the top line of the TM drive (with the make and size). Towards the bottom, the *Partition Map Scheme* should be GUID (preferred) or +Apple Partition Map+ for an Intel Mac. It must be +Apple Partition Map+ for a PPC Mac.
    If any of this is incorrect, that's likely the source of the problem. See item #5 of the Frequently Asked Questions post at the top of this forum for instructions, then try again.
    If it's all correct, perhaps there's something else in your logs.
    Use the Console app (in your Applications/Utilities folder).
    When it starts, click +Show Log List+ in the toolbar, then navigate in the sidebar that opens up to your system.log and select it. Navigate to the +Starting standard backup+ message that you noted above, then see what follows that might indicate some sort of error, failure, termination, exit, etc. (many of the messages there are info for developers, etc.). If in doubt post (a reasonable amount of) the log here.

  • Value too large for column "OIMDB"."UPA_FIELDS"."FIELD_NEW_VALUE"

    I am running OIM 9.1.0.1849.0 build 1849.0 on Windows Server 2003
    I see the following stack trace repeatedly in c:\jboss-4.0.3SP1\server\default\log\server.log
    I am hoping someone might be able help me resolve this issue.
    Thanks in advance
    ...Lyall
    java.sql.SQLException: ORA-12899: value too large for column "OIMDB"."UPA_FIELDS"."FIELD_NEW_VALUE" (actual: 2461, maximum: 2000)
         at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:112)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:331)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:288)
         at oracle.jdbc.driver.T4C8Oall.receive(T4C8Oall.java:745)
         at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:216)
         at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:966)
         at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1170)
         at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3339)
         at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3423)
         at org.jboss.resource.adapter.jdbc.WrappedPreparedStatement.executeUpdate(WrappedPreparedStatement.java:227)
         at com.thortech.xl.dataaccess.tcDataBase.writePreparedStatement(Unknown Source)
         at com.thortech.xl.dataobj.PreparedStatementUtil.executeUpdate(Unknown Source)
         at com.thortech.xl.audit.auditdataprocessors.UserProfileRDGenerator.insertUserProfileChangedAttributes(Unknown Source)
         at com.thortech.xl.audit.auditdataprocessors.UserProfileRDGenerator.processUserProfileChanges(Unknown Source)
         at com.thortech.xl.audit.auditdataprocessors.UserProfileRDGenerator.processAuditData(Unknown Source)
         at com.thortech.xl.audit.genericauditor.GenericAuditor.processAuditMessage(Unknown Source)
         at com.thortech.xl.audit.engine.AuditEngine.processSingleAudJmsEntry(Unknown Source)
         at com.thortech.xl.audit.engine.AuditEngine.processOfflineNew(Unknown Source)
         at com.thortech.xl.audit.engine.jms.XLAuditMessageHandler.execute(Unknown Source)
         at com.thortech.xl.schedule.jms.messagehandler.MessageProcessUtil.processMessage(Unknown Source)
         at com.thortech.xl.schedule.jms.messagehandler.AuditMessageHandlerMDB.onMessage(Unknown Source)
         at sun.reflect.GeneratedMethodAccessor127.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:324)
         at org.jboss.invocation.Invocation.performCall(Invocation.java:345)
         at org.jboss.ejb.MessageDrivenContainer$ContainerInterceptor.invoke(MessageDrivenContainer.java:475)
         at org.jboss.resource.connectionmanager.CachedConnectionInterceptor.invoke(CachedConnectionInterceptor.java:149)
         at org.jboss.ejb.plugins.MessageDrivenInstanceInterceptor.invoke(MessageDrivenInstanceInterceptor.java:101)
         at org.jboss.ejb.plugins.CallValidationInterceptor.invoke(CallValidationInterceptor.java:48)
         at org.jboss.ejb.plugins.AbstractTxInterceptor.invokeNext(AbstractTxInterceptor.java:106)
         at org.jboss.ejb.plugins.TxInterceptorCMT.runWithTransactions(TxInterceptorCMT.java:335)
         at org.jboss.ejb.plugins.TxInterceptorCMT.invoke(TxInterceptorCMT.java:166)
         at org.jboss.ejb.plugins.RunAsSecurityInterceptor.invoke(RunAsSecurityInterceptor.java:94)
         at org.jboss.ejb.plugins.LogInterceptor.invoke(LogInterceptor.java:192)
         at org.jboss.ejb.plugins.ProxyFactoryFinderInterceptor.invoke(ProxyFactoryFinderInterceptor.java:122)
         at org.jboss.ejb.MessageDrivenContainer.internalInvoke(MessageDrivenContainer.java:389)
         at org.jboss.ejb.Container.invoke(Container.java:873)
         at org.jboss.ejb.plugins.jms.JMSContainerInvoker.invoke(JMSContainerInvoker.java:1077)
         at org.jboss.ejb.plugins.jms.JMSContainerInvoker$MessageListenerImpl.onMessage(JMSContainerInvoker.java:1379)
         at org.jboss.jms.asf.StdServerSession.onMessage(StdServerSession.java:256)
         at org.jboss.mq.SpyMessageConsumer.sessionConsumerProcessMessage(SpyMessageConsumer.java:904)
         at org.jboss.mq.SpyMessageConsumer.addMessage(SpyMessageConsumer.java:160)
         at org.jboss.mq.SpySession.run(SpySession.java:333)
         at org.jboss.jms.asf.StdServerSession.run(StdServerSession.java:180)
         at EDU.oswego.cs.dl.util.concurrent.PooledExecutor$Worker.run(PooledExecutor.java:748)
         at java.lang.Thread.run(Thread.java:534)
    2008-09-03 14:32:43,281 ERROR [XELLERATE.AUDITOR] Class/Method: UserProfileRDGenerator/insertUserProfileChangedAttributes encounter some problems: Failed to insert change record in table UPA_FIELDS

    Thankyou,
    Being the OIM noob that I am, had no idea where to look.
    We do indeed have some user defined fields of 4000 characters.
    I am now wondering if I can disable auditing, or maybe increase the size of the auditing database column?
    Also, I guess I should raise a defect in OIM as the User Interface should not allow the creation of a user field for which auditing is unable to cope.
    I also wonder if the audit failures (other than causing lots of stack traces) causes any transaction failures due to transaction rollbacks?
    Edited by: lyallp on Sep 3, 2008 4:01 PM

Maybe you are looking for