Berkeley DB Storage related questions

Hi, I am new to BDB.
We recently migrated our internal DB into BDB (we are using the C++ version). Here are some of the details:
Data is saved using hash table (key has a fixed size of 16 bytes, but data size varies - from 20 bytes to a few kbs)
We have an index on each record (the index has fixed size of 4 bytes). Index using btree.
We do about 200 read, 200 write (update or insert), 200 delete every second. Update will always change the index.
Over time, the number of records in the DB only increases slightly (total is about 2 million records).
In the last month or two we have observed the following storage change:
When the internal DB was converted into BDB, the data was about 8GB, index about 1GB. The actual data (key, data, index) we put into the BDB counts about 50% of the actual storage (the other 50% is BDB's overhead/other data structure, etc). Now, the data becomes about 9GB (slightly more than our data grows rate), but the index db size has grown to 9GB as well. Since the number of records are pretty much the same, the index DB has a lot of waste.
My questions:
Are there any parameters (eg fill factors) that could affect the storage usage? We just used the defaults.
Over time, how to defragment the DB without taking the DB off line?
Thanks in advance for your help.

Hi John,
So you are using the C++ API. What is the BDB release you are using (e.g. 5.1.19, 5.0.26 etc), on what OS, filesystem, and what is the underlying filesystem's I/O block size? (to get the block size, use something like "dumpe2fs /dev/hdN | grep 'Block size'" on Linux, "fsutil fsinfo ntfsinfo D:" on Windows -- look for "Bytes Per Cluster", "df -g /part_mount_point" on Solaris)
Note that Btree and Hash databases are grow-only, that is, space freed by deleting items from a Btree or Hash database is never returned to the filesystem, although it is reused where possible; see the Disk space requirements section in the Berkeley DB Reference Guide.
You can use compaction to reclaim disk space and shrink the databases' sizes; see the Db::compact() method. If you need clarification on the compact() method let me know; also, there are a lot of useful threads on the forum dealing with compaction, just do a search on "compact database".
Data is saved using hash table (key has a fixed size of 16 bytes, but data size varies - from 20 bytes to a few kbs)More precisely, to how many "few" Kbs?
the data becomes about 9GB (slightly more than our data grows rate), but the index db size has grown to 9GB as wellCould you check that the secondary key creator callback function is correct? It's a little bit surprising to see the secondary index database growing from 1GB to 9GB.
Are there any parameters (eg fill factors) that could affect the storage usage? We just used the defaults.Yes, the most important is the database page size which will influence the fill factor; see Selecting a page size.
As your data item size varies to a few Kbs your database probably has a lot of overflow pages (these pages store overflow items which are too large to be stored entirely onto a normal leaf page - an overflow item will be stored onto multiple overflow pages; retrieving such items is more expensive as the entire set of pages needs to be read into cache).
Also, for specific configuration for each of the two access methods see Btree access method specific configuration and Hash access method specific configuration.
Use the db_stat utility with the -d option and provide the database statistics for each of the databases (primary and secondary).
Over time, how to defragment the DB without taking the DB off line?Use compaction as initially mentioned; compaction is done online, that is, you do not have to close the database (take it offline).
Regards,
Andrei

Similar Messages

  • Storage related question for RAC

    Hi Gurus,
    I have a question here about the storage design. My company has 2 node Oracle RAC already setup on Linux with 11g for clusterware, ASM and 10g for database(for application support we are using still 10g for database). This RAC is dedicated for one database and shared by several applications.
    Now, we decided to add two more nodes and additional storage to it.
    The existing storage uses EMC SAN DMX 4574 (raw partitions) disks. And the additional storage disks provided by our storage team are EMC SAN DMX 3752.
    Is it possible to combine these two different DISK (different series from same vendor) into one single storage sub system visible to entire RAC Database? All the 4 nodes will be used for the same database and application. Its a production system so we wanted to confirm and proceed further as we dont have a test scenario.
    Shall we connect these different disks to the switches so that ASM can see all the disks at once and can be used for database?
    Please advise.
    Regards,
    vasu
    Edited by: vasu77 on Feb 23, 2010 10:12 AM

    Hi Vasu,
    The existing storage uses EMC SAN DMX 4574 (raw partitions) disks. And the additional storage disks provided by our storage team are EMC SAN DMX 3752.
    Is it possible to combine these two different DISK (different series from same vendor) into one single storage sub >system visible to entire RAC Database? All the 4 nodes will be used for the same database and application. Its a >production system so we wanted to confirm and proceed further as we dont have a test scenario.First of all, You have to ask the OS support team in order to see if there is any kind of problem mixing it. I don't think it would be a problem, but as you don't have any test environemnt, You'd better ask them.
    For the ASM side, it's not a problem, You can work with this kind of configuration. In fact I have a couple of environemnt using this configuration. I'm using a expensive storage with a better performance to critical tablespaces, and a lower cost storage to put historical data and other kind of data that are not accessed every time.
    Shall we connect these different disks to the switches so that ASM can see all the disks at once and can be used for database? Well,
    As I told before, I don't think it would be a problem, but again... You should ask the OS support team. For the ASM side, You shouldn't mix disks with diferent behaviour in the same diskgroup, You should use disks with the same size and similar/equal performance characteristics. As ASM balances the I/O over the disks in a ASM diskgroup, it's important to configure in this way in order to avoid performance problems.
    Hope it helps,
    Cerreia

  • I want to check all functions of PCI 6534.I have read the user manual..I have some memory related questions.​Please help me for that.

    I want to check all functions of PCI 6534.I have read the user manual..I have some memory related questions.Please help me for that.
    1.)If i am using the continuous output mode.and the size of generated data is less than 32 MB.If i want to preload the memory,what should i do?I want that first of all i load all my data to onboard memory & then i want to make start the transfer between 6534 & peripheral.Is it possible?As per me it should be.Plz tell me how should i do this?I think that in normal procedure the transfer between 6534-peripheral & outputting data from pc buffer to onboard memory works parallely.But i don't want this.Is it poss
    ible?
    (2).Similarly in finite input operation(pattern I/O) is it possible to preload the memory and then i read it?Because i think that the PC memory will be loaded automatically when 6534 acquires the data and then when we use DIO read vi the pc buffer data will be transferred to application buffer.If this is true,i do not want this.Is it possible?
    (3) One more question is there if i am using normal operation onboard memory will be used bydefault right?Now if i want to use DMA and if i have data of 512 bytes to acquire.How will it work and how should i do it?Please tell me the sequence of operations.As per my knowledge in normal DMA operation we have 32 Bytes FIFO is there so after acquisition of 32 bytes only i can read it.How it will known to me that 32 bytes acquisition is complete?Next,If i want to acquire each byte separately using DMA interrupts what should i do?Provide me the name of sourse from which i can get details about onboard memory & DMA process of 6534 specifically
    (4).In 6534 pattern Input mode,if i want to but only 10 bits of data.and i don't want to waste any data line what should i do?

    Hi Vishal,
    I'll try to answer your questions as best I can.
    1) It is definitely possible to preload data to the 32MB memory (per group) and start the acquisition after you have preloaded the memory. There are example programs on ni.com/support under Example Code for pattern generation and the 6534 that demonstrate which functions to use for this. Also, if your PC memory buffer is less than 32MB, it will automatically be loaded to the card. If you are in continuous mode however, you can choose to loop using the on-board memory or you can constantly be reading the PC memory buffer as you update it with your application environment.
    2) Yes, your data will automatically be loaded into the card's onboard memory. It will however be transferred as quickly as possible to the DMA FIFO on the card and then transferred to the PC memory buffer through DMA. It is not going to wait until the whole onboard memory is filled before it transfers. It will transfer throughout the acquisition process.
    3) Vishal, searching the example programs will give you many of the details of programming this type of application. I don't know you application software so I can't give you the exact functions but it is easiest to look at the examples on the net (or the shipping examples with your software). Now if you are acquiring 512 bytes of data, you will start to fill your onboard memory and at the same time, data will be sent to the DMA FIFO. When the FIFO is ready to send data to the PC memory buffer, it will (the exact algorithm is dependent on many things regarding how large the DMA packet is etc.).
    4) If I understand you correctly, you want to know if you waste the other 6 bits if you only need to acquire on 10 lines. The answer to this is Yes. Although you are only acquiring 10 bits, it is acquired as a complete word (16bits) and packed and sent using DMA. You application software (NI-DAQ driver) will filter out the last 6 bits of non-data.
    Hope that answers your questions. Once again, the example code on the NI site is a great place to start this type of project. Have a good day.
    Ron

  • In which Board we need to post cRIO related Questions?

    sir,
         In which board do i need to post cRIO related Questions and doubts regarding my application with cRIO .Please help me previously i have posted in  motion control drives as my application is related to motor control using cRIO.bu i am not getting any reply .i couldnt find any cRIO related board in select board option.
    Solved!
    Go to Solution.

    Hi illa,
    You can post your query under the Real Time Measurement and Control
    http://forums.ni.com/t5/Real-Time-Measurement-and/bd-p/280

  • Two related questions:  ColdFusion 10/Java applications and J2EE supported servers

    I have two related questions:
    1.  CF10 and integration with Java Web applications
    We have a couple of Java applications running on JRun and interfacing with CF9 applications.  The JRun clusters were created through the JRun Admin and, apart from lack of Axis 2.0 support, have served us well for years now.  And, as would be the case, the ColdFusion9/Java/Flash application is a critical public-facing application that the business uses for bidding on projects.
    It appears that with ColdFusion 10 on Tomcat, we will not be able to run those Java applications on a Tomcat-CF10 JVM cluster.  Is this correct?  IF so, what are our options? 
    2.  J2EE Application Servers supported by Adobe for CF10
    Which of these is correct?
    A.  This URL (http://www.adobe.com/products/coldfusion-enterprise/faq.html) states "ColdFusion 10 supports IBM® WebSphere, Oracle® WebLogic, Adobe JRun, Apache Tomcat, and JBoss."
    B.  This URL (http://wwwimages.adobe.com/www.adobe.com/content/dam/Adobe/en/products/coldfusion/pdfs/cf1 0/coldfusion10-support-matrix.pdf) states:
    "J2EE application servers: WebLogic Server 10.3, 11.1, WebSphere Application Server 7, ND 7 JBoss 5.1, 6.0, 7.1.0"
    I *think* "A" above is wrong re. support for Adobe JRun.  It does not specify a version of Apache Tomcat unless it is simply referring to the custom version the comes with CF10.
    Option "B" above shows no support of Adobe JRun or 'standard' Apache Tomcat.
    Thanks,
    Scott

    Question 1 above was answered:  "No support for Java web applications under CF10's custom version of Tomcat"
    Question 2:  No answer yet:  Is Apache Tomcat (NOT Adobe's customized version) supported for CF10 J2EE deployment?  I do not see any installation instructions on how to install CF10 on Apache Tomcat 6 or 7.
    Is anybody using Apache Tomcat as their J2EE app servers and, again, NOT Adobe's customized/limited version? 
    Thanks,
    Scott

  • Hyperion 9.2 move storage relational DB

    Hi,
    Can anyone suggest what is the procedure to move storage relation DB server for Hyperion 9.2 with shared service, essbase and BI components?
    Is it feasible as following steps:
    1. backup and restore all databases from old server to new server.
    2. reconfigure the new IP address, user name and password through configurator.
    3. confirm all "reuse existing data.." buttons.
    Any risk and checking steps?
    Regards,
    Kenneth

    1) Bi store report in RDB and file system - best way migrate - export reports to xml files and import
    2) Shared Service store in OpenLdap with BerkleyDB storage - best way migrate - export to xml :)
    3) Essbase work with native database and best way to migrate use export to flat files

  • Deafult search engine on a new tab-this thread was closed but I have a related question. Once you put the new url in, what then? Close? When tried, did not work

    "deafult search engine on a new tab" - this thread was closed but I have a related question. [the word “default” is spelled as it on the thread] Once you put the new url in, what then? Close it? When tried, it did not work. The closest answer is noted here. Instructing to go to about:fig. Specifically, I installed the accuweather add=on and it took over my tab home page - the one that comes up when you click for a new tab. I didn't like it so I clicked the button to remove it and the page went blank. I searched to see out to restore the tab page, tired the reply at this page :
    http://support.mozilla.org/en-US/questions/753256?s=install+search+box+on+new+tabs&r=3&as=s
    But it did not work. Can some on help me? Just want to know how to make google or some page open on a new tab.
    Many thanks

    Thanks so much for your support. As I said I am not a professional, so I by chance saw a friend who is and this is what he suggested and it worked.
    We googled for an answer and found this site: http://www.guidingtech.com/8004/have-new-tab-page-open-specific-website-firefox-chrome/
    It says to Download and install Download and install''' NewTabURL''' and gives instructions on how to. Now when I open the tab it goes to my preferred search page!

  • PM/EAM related questions posted in PLM forum

    Hi,
    I was going through some of the threads available in PLM forum, but to my surprise, most of the threads posted are related to PM, which should have been posted in EAM forum. It was clearly mentioned by the moderator in one of the sticky thread [PLM-Forumsplit - Please ensure you are using the right forum|PLM-Forumsplit -  Please ensure you are using the right forum;
    Questions regarding Plant Maintenance (PM) or Customer Service (CS) are discussed in the Enterprise Asset Management (EAM) forum
    forums - posts will be moved to the right forum.
    Please ensure your are opening a new thread in the right forum
    Despite the clear instructions given my the moderator, PM related questions are still posted to PLM forum and some of the active participant of EAM forum answers these questions in PLM forum. It would be nice if those threads are moved to appropriate forum, so that the relevant informations are available under one forum, also, PLM forum should be monitored regularly and inappropriate threads are to be moved to their respective forum.
    Few PM related threads posted recently in PLM forum:
    [Change PR for External order|Change PR for External order;
    [BAPI or Function Module to Complete Maintenance calls in IP10|BAPI or Function Module to Complete Maintenance calls in IP10;
    [delete material from PM order|delete material from PM order;
    [Message no. IW056 Work center change --> default data redefined|Message no. IW056 Work center change --> default data redefined;
    [Component in Task list not reflecting in maintenance order|Component in Task list not reflecting in maintenance order;
    [Deleted functional location/equipments|Deleted functional location/equipments;
    [PM orders not listing components from material BOM|PM orders not listing components from material BOM;
    [Completion date in Maintenance Plans in IP10|Completion date in Maintenance Plans in IP10;
    I had used abuse reports for some, but numbers of such posts are many.

    I hate to say this, but I often see the same thing, and I've never been able to figure out exactly what triggers selectability vs. non-selectability, but I suspect some of it has to do with tranparency flattening in particular, and possibly text wrap in cases where that comes into play. ID optimizes the PDF it creates, I think, and that might cause text, or even images, to be broken into chunks in funny ways. It occurs to me that turning off optimzing for fast web view might help in this regard.

  • Please post documentation-related questions or comments on the

    Please post documentation-related questions or comments on the Performance Management Applications {forum:id=1318} forum     
    http://forums.oracle.com/forums/forum.jspa?forumID=1318

    I just got a call today that I reached my 300GB limit for the month.  I called and got a pretty rude response from the security and data usage department.  The guy told me in so many words that if I do not like or agree with the policy that I should feel free to find another service provider.!!! I tried to explain that we watch Netflix and XFinity on-demand alot and I was told that that can not be anywhere close to the data usage. I checked my router and watching a "super HD, dolby 5.1" TV show on Netflix will average about 5-6 GB per hour (1.6MB/s) ... sp this means that I can only watch no more than 1-2 Super HD TV shows a day via Netflix before I run out of my data usage.    This seems a bit redicilous doesn't it? Maybe the TV ads about the higher speed than the competition should be accompanied with "as long as you don't use it too often"   Not a good experience ... 

  • Where do I  post my CRM related questions

    Where do I post my CRM related questions.
    thanks
    siva
    Message was edited by:
    sivapara

    Hi Siva,
    You can select the module of CRM from the category and post your question there. You can even search for CRM and Post your question.
    Regards,
    [email protected]

  • Online Help Related Questions

    Hi
    FYI there is no longer a dedicated place on Microsoft.com for online help related questions.  
    We are told that VS General (here) is the best place for help questions.
    Note that these sites are a good place to get help on help (Microsoft and MVP presence):
    http://groups.yahoo.com/group/hatt/ -- for WinHelp & HTML Help & general help discussion
    http://tech.groups.yahoo.com/group/MSHelpViewer/ -- For HelpViewer help (VS 2010/2012 help; Windows 8 help)
    http://tech.groups.yahoo.com/group/MSHelp2/  -- For MS Help 2.x help (VS 2002/2003/2005/2008 help)
    Feel free to ask questions on this thread if you need to. At least then we can find your post. 
    You can also email the VS/MSDN help team directly -- [email protected]
    Please let them know how you feel about the lack of support for Online Help in the forums.
    Rob
    www.helpmvp.com -- Various Help resources
    Rob Chandler Help MVP www.helpwareGroup.com | mshcmigrate.helpmvp.com | hv2.helpmvp.com

    Actually, we cannot (at least anymore) directly email the team. I tried and got a reject from the mail server.  So, I'm going to post here so I can feel better knowing I said something and then I can get back to programming :-)
    Ok, I’ve been developing Windows apps since Win 3.1 (yeah, a long time) and have always admired the Microsoft attitude towards Developers and it’s desire to educate and empower R&D.
    So, I’m wondering what’s going on?  Why am I being asked to have an account to develop anything for MY own PC?  Why am I being asked for my Microsoft account password, repeatedly, when I’m just looking up documentation on UserControl or C3 keywords…
    repeatedly?  This is nothing short of annoying and it gives me pause and concern about the ecosystem.  Is the plan to lock down programmer’s capabilities, knowledge and tools?
    I’m involved in the TAP program for Windows 10 and have voiced concerns in the same category with regard to installation and deployment of Universal Apps.  Security is not a reasonable excuse.  Some of us love to program and actually
    control their PCs.  We are not planning to sell everything we write on the Windows Store.  What happened to programming something cool for your home network or just for friends?  We need, and used to have, Microsoft’s support in doing
    this sort of thing.  Documentation and tools are more important now than ever.  I’m saddened to see the direction things are heading…
    If there is a sane reason I’m being badgered to ‘log in’ to read .NET documentation every hour, please share.  If not, please consider lobbying to get things back on track.
    Thanks

  • Button, or other panel related question

    so, if any of you have read my previous topic about painting rectangles in a panel, that worked great, but now i need to mkae those rectangles components so that when someone hovers over it or clicks on it it displays information at a certain area. I know how to make listeners and how to add buttons in a panel, BUT how do I add buttons at a specific coordinate. This program graphs data on a specific scale and therefore adding a button in the approximate area using a grid layout will not be good enough. Is there anyway for me to place the the button/component/object at the extact location that I want?
    thanks in advance

    In future GUI related questions should be posted in the appropriate GUI forum (likely Swing).
    so, if any of you have read my previous topic about
    painting rectangles in a panel, that worked great,
    but now i need to mkae those rectangles components so
    that when someone hovers over it or clicks on it it
    displays information at a certain area. What?

  • Storage / Backup Question

    As this relates to how I work with my files whilst using FCE, I suspect this is probably the right board to post on....here's my question....
    I'm starting to produce more and more videos for work - I capture with a small Sony HD camera, import the files to my external hard drive....and then start work on them....
    This works fine for now...but as my first Terabyte starts to be used up, I'm also starting to consider issues such as backing up...
    I started to read about RAID storage options.....RAID 1 appears to be the way to go in terms of storing my master files, and ensuring they are safe....my questions are....
    - Does this sound like the best strategy?
    - Will my Mac need a seperate RAID card installed to handle the extra processor/RAM work?
    -Are there any particular setups / brands anybody could recommend?
    Many thanks in advance
    Jonathan

    When I was recently organising new backup facilities for my sctatch disk I got advice from someone on Macfixit against using a RAID system for backups, either a hard or software RAID system, for the reasons that Martin mentioned - lack of sufficiently reliability. The reason for re-equipping was that the internal power supply circuitry of the (failed) external Scratch Disk caused the failure. If that happened in a RAID system (which I believe shares a common internal power circuitry) the whole RAID would fail. I got a rescue firm to read and copy the Capture Scratch on to another HD but many files did not survive.
    I went for two 500 GB external LaCie Firewire 800 (actually triple inteface) drives and use one for my scratch disk and copy everything across to the other now and again. As a test, by resetting FCE's Scratch settings to the second drive, (to simulate a failure of the first) FCE worked normally.
    Loosing the scrastch disk (as happened to me recently) creates lots of work in re-capturing. So I am relying on two backup systems for my Capture Scratch - the second HD and the Mini DV tapes, all of which I keep as they are quite cheap.
    Message was edited by: SteveKir

  • CRM Related Questions for Configuring Service Desk in Solution Manager

    Hi All,
    We have 3 requirements while configuring service desk in solution manager:
    1) Is there any possibility that we can restrict view of messages in solman for different teams. E.g: Finance Support team should not view message raised for Material Management Team ?
    2) Is there any possibility in service desk that we can assign message on the basis of location e.g: message raised for finance from location A should go to different support team of finance and for location B, it should go to different finance team ?
    3) Is there any possibility that we can make some question necessary while user is raising very high priority message from satellite system e.g: when we raise Very high priority message on service market place, we need to answer few mandatory questions ?
    I tried to search these functionality in standard scenario but I am not able to find something relevant. Please suggest.
    Thanks
    Sunny

    Hi Rama,
    I Need to configure the feature to generate a ticket to SAP directly in VERY HIGH priority.
    Could you please help me out into this. If you have any doc related to that Please send me.
    Thanks in Advance!
    Regards,
    Prabhat Singh

  • 'Genius' related question for anyone with over 1.2TB of music?

    Anyone out there with a iTunes library with over 1.2TB of music?
    Reason I ask, is that I've not been able to solve a problem in updating Genuis - I'm getting the error - 'Genius results can't be updated right now. There is not enough memory available ~ Please try again later.' or the 'unknown 4002 error'.  The steps I've taken to try and correct sometime back can be found in https://discussions.apple.com/message/12584815#12584815 and also shown here in sequence below. 
    I keep thinking this will be fixed with the next version iTunes but so far with the latest, things still the same.  So looking to see if anyone with a similar size library is having similar problems and it just a limitation of the program or it's something specific on my end that I can fix.
    Appreciate...
    Up to this weekend I've had not issues with Genius. Worked beautifully and updated last Sunday as I have been doing every week since Apple added Genius. Yesterday I added a few albums and went to update and now getting the error of 'Genius results can't be updated right now. There is not enough memory available ~ Please try again later.' Now my library is well over >200,000 songs and only added 6 albums this time. Music sits on a external 2TB drive ~100GB free and my MacBook Pro has over 60GB open. Running latest Mac OSX and latest iTunes. Nothing new installed or changed since last Genius update.
    To the points raised in this thread.
    1. I've been updated weekly with a Library with over 200K songs till now with no problem. Majority is what I've ripped from my albums & CDs.
    2. Deleted the albums recently added since last successful update - same result.
    3. Connected directly to the network bypassing router as suggested here - same result.
    4. Reinstalled iTunes program - same result.
    5. Repaired permissions - same result.
    Problem remains. Welcome any additional thoughts.
    Question for anyone with a music library
    https://discussions.apple.com/message/12584815#12584815
    Ok, tried all the suggestions here yet the problem remains.
    In addition what I posted above -
    1) Dumped the cookies in user/library/cookies folder
    2) Deleted the apple cookies found under Safari Preferences "Show Cookies"
    3) Repaired disk permissions
    4) Deleted iTunes Library Genius.itdb file
    5) Reinstalled iTunes (with the new 10.1 version)
    6) Restarted computer
    7) Rebuilt iTunes Library folder http://support.apple.com/kb/ht1451
    8) Deleted all iTunes dead tracks
    http://dougscripts.com/itunes/scripts/ss.php?sp=removedeadsuper
    9) Increased internet access speed
    10) Removed all open "[" and '}' brackets from songs, albums & artists
    10) Rebuilt my entire library from scratch!
    Is there anyone out there with over 200K songs that can update their Genius? One thing I noted is the genuis.itdb file is just over 2GB. I'm wondering if there is a maximum size on this based on the error I'm getting. Any thought or know info on this?
    Ok, think I may have found something here with this. I backed out all the music added prior to this problem occurring and ran the Genius update. All worked fine. I then started adding the music back little by little (running the update each time) until the problem returned. When the problem came back I ruled out the specific music added by backing that bit out, verified all was ok and added a different batch back in.
    Bottom-line it appears that when my library exceeds exactly 1.208TB the problem appears. Interestingly enough it does not appear to be related to the number of songs, the size of the genius.itdb file (in fact when it was working the file was at 2.25GB - larger than what I reported before). It specifically seems to be related to the size of the overall library - when I have it for example at 1.2079 all is ok, 1.2081 it fails with this error.
    I did try moving the library and related files to the external drive but same result. abavetta, I think your suggestion of this script may be a workaround for now - although from what I read, is this not the same that you can do anyway with iTunes by holding the alt/option key open start up? In fact I do have two libraries now - one with the 2TB drive which I use at home and another for 2x500GB disks I carry with me when i travel. I just reported the same back to Apple but doubt if any immediate solution will come for this. Too bad as one of the great things about having such a large library is the mixes you can generate via Genius. Anyway, such is life and will just live with this for now. I also just found this program called Tangerine! http://www.potionfactory.com/tangerine/ which creates playlists like Genius - although through a different technique I might also take a look at.

    Just to let all know that may be having a similar problem, what I found by accident to fix this problem.
    Reset all my Home Folder permissions using the resetpassword method from the startup utility in Lion.  Then turned off/on to reset Genius and cleaned out the cache folder for good measure.  Finally after more than a year without, have Genius up and running again.

Maybe you are looking for