Dynamic JCombobox for a very big database resulset

I want to use a JCombobox or similar for selecting values from a big database resultset. I'm using an editable one with SwingX autocomplete decorator. My strategy is:
* show only first Xs regs.
* let the user to enter some text in the combobox and refine the search reloading the model.
Someone have some sample code or know some components that do that.
Or can point me to some implementation details?�
A lot of thanks in advance,
PD:
I need something efficient that don't query database to much.

what is the size of the table and how many lines are you going to delete?
I would recommend you to delete only up to 5000 or 10000 records in one step.
do 100 times.
select *
          from
          into table itab.
          where
          up to 10.000 records.
if ( itab is initial )
  exit.
endif.
delete ... from table itab.
commit work.
If this is still too slow, than you should create a secondary index with zone.
You can drop the index after the deletion is finished.
Siegfried

Similar Messages

  • Optimize delete in a very big database table

    Hi,
    For delete entries in database table i use instruction:
    Delete from <table> where <zone> = 'X'.
    The delete take seven hours (the table is very big and  <zone> isn't an index)
    How can i optimize for reduce the delete time.
    Thanks in advance for your response.
    Regards.

    what is the size of the table and how many lines are you going to delete?
    I would recommend you to delete only up to 5000 or 10000 records in one step.
    do 100 times.
    select *
              from
              into table itab.
              where
              up to 10.000 records.
    if ( itab is initial )
      exit.
    endif.
    delete ... from table itab.
    commit work.
    If this is still too slow, than you should create a secondary index with zone.
    You can drop the index after the deletion is finished.
    Siegfried

  • Export very big database.

    Hello,
    I have 9i database on Linux. This database is very big in tera bytes.I want to shift the database to other server.
    export backup is taking too much time. Can you please suggest me how can i shift my database in few hours?
    Thanks in advance.
    Anand.

    Tricky. Especially since you don't say if the new server is running Linux, too. And you also don't say (which makes a big difference) if the new server will be running 10g.
    But you might be able to do a transportable tablespace migration. That involves exporting only the contents of your existing data dictionary (a matter of a few minutes at most, usually); copying the DBF files to the new server; and then plugging them in by importing your data dictionary export. The major time factor in that lot is the physical act of copying the datafiles between servers. But at least you're not extracting terabytes of data and then trying to re-insert the same terabytes!
    If your new server is not running Linux, forget it, basically, because cross-platform tablespaces are only do-able in 10g and with lots of restrictions and caveats (but you might get lucky... you'd have to read tahiti.oracle.com to find out if you could get away with it).
    If your new server is running 10g, you're also going to be in for tricky times, though it's not impossible to transport between 9i and 10g. Easiest thing, if possible, is to create your 10g database with COMPATIBLE set to 9.x.x, do the transport and then increase your compatible parameter afterwards.

  • Very simple database require

    Hi
    I'm looking for a very simple database solution. I have some very large .csv files that I need to query against before importing to Excel. Filemaker, etc. is over the top for what I need. Any ideas?
    Thanks.
    PowerPC G5   Mac OS X (10.4.6)  

    Welcome to Apple Discussions!
    You could use a script that converts CSV to Appleworks:
    http://www.tandb.com.au/appleworks/import/
    And then export from Appleworks to Excel.
    I know the PowerMac G5 doesn't come with Appleworks, but it is a quarter of the price of Filemaker Pro.
    Maybe the authors of the script could help you.

  • Database Log File becomes very big, What's the best practice to handle it?

    The log of my production Database is getting very big, and the harddisk is almost full, I am pretty new to SAP, but familiar with SQL Server, if anybody can give me advice on what's the best practice to handle this issue.
    Should I Shrink the Database?
    I know increase hard disk is need for long term .
    Thanks in advance.

    Hi Finke,
    Usually the log file fills up and grow huge, due to not having regular transaction log backups. If you database is in FULL recovery mode, every transaction is logged in Transaction file, and it gets cleared when you take a log backup. If it is a production system and if you don't have regular transaction log backups, the problem is just sitting there to explode, when you need a point in time restore. Please check you backup/restore strategy.
    Follow these steps to get transactional file back in normal shape:
    1.) Take a transactional backup.
    2.) shrink log file. ( DBCC shrinkfile('logfilename',10240)
          The above command will shrink the file to 10 GB.(recommended size for high transactional systems)
    >
    Finke Xie wrote:
    > Should I Shrink the Database? .
    "NEVER SHRINK DATA FILES", shrink only log file
    3.) Schedule log backups every 15 minutes.
    Thanks
    Mush

  • How can we suggest a new DBA OCE certification for very large databases?

    How can we suggest a new DBA OCE certification for very large databases?
    What web site, or what phone number can we call to suggest creating a VLDB OCE certification.
    The largest databases that I have ever worked with barely over 1 Trillion Bytes.
    Some people told me that the results of being a DBA totally change when you have a VERY LARGE DATABASE.
    I could guess that maybe some of the following topics of how to configure might be on it,
    * Partitioning
    * parallel
    * bigger block size - DSS vs OLTP
    * etc
    Where could I send in a recommendation?
    Thanks Roger

    I wish there were some details about the OCE data warehousing.
    Look at the topics for 1Z0-515. Assume that the 'lightweight' topics will go (like Best Practices) and that there will be more technical topics added.
    Oracle Database 11g Data Warehousing Essentials | Oracle Certification Exam
    Overview of Data Warehousing
      Describe the benefits of a data warehouse
      Describe the technical characteristics of a data warehouse
      Describe the Oracle Database structures used primarily by a data warehouse
      Explain the use of materialized views
      Implement Database Resource Manager to control resource usage
      Identify and explain the benefits provided by standard Oracle Database 11g enhancements for a data warehouse
    Parallelism
      Explain how the Oracle optimizer determines the degree of parallelism
      Configure parallelism
      Explain how parallelism and partitioning work together
    Partitioning
      Describe types of partitioning
      Describe the benefits of partitioning
      Implement partition-wise joins
    Result Cache
      Describe how the SQL Result Cache operates
      Identify the scenarios which benefit the most from Result Set Caching
    OLAP
      Explain how Oracle OLAP delivers high performance
      Describe how applications can access data stored in Oracle OLAP cubes
    Advanced Compression
      Explain the benefits provided by Advanced Compression
      Explain how Advanced Compression operates
      Describe how Advanced Compression interacts with other Oracle options and utilities
    Data integration
      Explain Oracle's overall approach to data integration
      Describe the benefits provided by ODI
      Differentiate the components of ODI
      Create integration data flows with ODI
      Ensure data quality with OWB
      Explain the concept and use of real-time data integration
      Describe the architecture of Oracle's data integration solutions
    Data mining and analysis
      Describe the components of Oracle's Data Mining option
      Describe the analytical functions provided by Oracle Data Mining
      Identify use cases that can benefit from Oracle Data Mining
      Identify which Oracle products use Oracle Data Mining
    Sizing
      Properly size all resources to be used in a data warehouse configuration
    Exadata
      Describe the architecture of the Sun Oracle Database Machine
      Describe configuration options for an Exadata Storage Server
      Explain the advantages provided by the Exadata Storage Server
    Best practices for performance
      Employ best practices to load incremental data into a data warehouse
      Employ best practices for using Oracle features to implement high performance data warehouses

  • How to extend dynamic selection for Logical Database PNPCE

    Hi All,
    I need to extend dynamic selection for Logical Database PNPCE. Can someone able to help with steps?
    Thanks
    Ranjith

    Hi All,
    I need to extend dynamic selection for Logical Database PNPCE. Can someone able to help with steps?
    Thanks
    Ranjith

  • Is this beta? The image for Yosemite is a very big rock, appropriate as this OS sends your mac back to the stone-age. iTunes crashes on open, iPhoto won't download, movies won't play, Safari won't show graphics, wifi down to speed of crippled snail.

    Is this beta. The image for Yosemite is a very big rock, appropriate as this OS sends your mac back to the stone-age. iTunes crashes on open, iPhoto won't update, Safari shows text without graphics, can't get Java 8 'cause Safari "can't find server". Wifi slower than a crippled snail. Does anyone know how to uninstall this rubbish?

    travellingbirder wrote:
    Is this beta ?
    You tell us Finding the OS X version and build information on your Mac.

  • Very big problem with JSF about FORM and "id=" for HTML form's elements and

    I have discovered a very big problem with JSF about FORM and "id=" for HTML form's elements and java instruction "request.getParameterNames()".
    Suppose you have something like this, to render some datas form a Java Beans :
    <h:dataTable value="#{TablesDb2Bean.myDataDb2ListSelection}" var="current" border="2" width="50%" cellpadding="2" cellspacing="2" style="text-align: center">
    <h:column>
    <f:facet name="header">
    <h:outputText value="Name"/>
    </f:facet>
    <h:outputText id="nameTableDb2" value="#{current.db2_name_table}"/>
    </h:column>
    </h:dataTable>
    Everything works fine...
    Suppose you want to get the name/value pairs for id="nameTableDb2" and #{current.db2_name_table} to process them in a servlet. Here is the HTML generated :
    <td><span <span class="attribute-name">id=<span class="attribute-value">"j_id_jsp_1715189495_22:0:nameTableDb2">my-table-db2-xxxxx</span></td>
    You think you can use the java instructions :
    Enumeration NamesParam = request.getParameterNames();
    while (NomsParam.hasMoreElements()) {
    String NameParam = (String) NamesParam.nextElement();
    out.println("<h4>"++NameParam+ "+</h4>);
    YOU ARE WRONG : request.getParameterNames() wants the syntax *name="nameTableDb2" but JSF must use id="nameTableDb2" for "<h:outputText"... So, you can't process datas in a FORM generated with JSF in a Servlet ! Perhaps I have made an error, but really, I wonder which ?
    Edited by: ungars on Jul 18, 2010 12:43 AM
    Edited by: ungars on Jul 18, 2010 12:45 AM

    While I certainly appreciate ejb's helpful responses, this thread shows up a difference in perspective between how I read the forum and how others do. Author ejb is correct in advising you to stay inside JSF for form processing if form processing is what you want to do.
    However, I detect another aspect to this post which reminds me of something Marc Andreesen once said when he was trying to get Netscape off the ground: "there's no such thing as bad HTML."
    In this case, I interpret ungar's request as a new feature request. Can I phrase it like this?
    "Wouldn't it be nice if I could render my nice form with JSF but, in certain cases, when I REALLY know what I'm doing" just post out to a separate servlet? I know that in this case I'll be missing out on all the nice validation, conversion, l10n, i18n, ajax, portlet and other features provided by JSF".
    If this is the case, because it really misses the point of JSF, we don't allow it, but we do have an issue filed for it
    https://javaserverfaces-spec-public.dev.java.net/issues/show_bug.cgi?id=127
    If you can't wait for it to be fixed, you could decorate the FormRenderer to fix what you want.
    I have an example in my JSF book that shows how to do this decoration. http://bit.ly/edburnsjsf2
    Ed

  • Why does not Adobe flash player for iPad ,This is a very big problem

    Why does not Adobe flash player for iPad ,This is a very big problem ,I do not know why Apple is so Popular

    but Adobe is, in fact, continuing to update Flash for Android.
    Only essential bug and security fixes. No new feature updates have been released for 2 years.
    Of course we have those who wish fervently that the lack of Flash supprt will damage Apple severely and obsess about it all the time.
    I thought they'd realised it was a lost cause ages ago... they mostly seem to have disappeared from this forum, at least...

  • Iphone fell down and there is a very big dent on the left side. the screen split on both side. can i send it to apple and the repair it for like 160€ or how does it work?

    hi my iphone fell down and there is a very big dent on the left side. the screen split on both side. can i send it to apple and the repair/change it for like 160€ or how does it work?

    Service Answer Center - iPhone - http://support.apple.com/kb/index?page=servicefaq&geo=US&product=iphone  <-- enter correct country in drag-down menu once on page.

  • How do I show multiple rollover images on a page inserted dynamically (pulled out of a sql database

    How do I show multiple rollover images on a page inserted dynamically (pulled out of a sql database table) using Dreamweaver’s Repeat Region. Example: I have different products each one associated (through their productID) with two images (one that’s showing in the page and one for the rollover effect) that are pulled from a database using Dreamweaver’s Recordset. So I want to end up with a page containing row after row of images(one for every product).When moused over each image will reveal the second (rollover) image for the same product which in turn can be a link(the image itself ) that when clicked leads to a detailed page with more information about the product the image is associated with. To show row after row with images for the different products in the database table I am using Dreamweaver’s Insert Rollover Image command and then the  Repeat Region – I have no problem to complete the steps, to insert the image and the rollover one at once and set the paths so they are pulled dynamically depending on the productID they are associated with .I have also no problem to apply the Repeat Region so not only the first image associated with the first product in the table is shown but all of them-a routine and standard procedure when using the Repeat Region to dynamically generate and display multiple rows of data. However, when I preview the page the rollover effect is lost –the images are shown but when moused over the second  image does not  show. The problem is caused when the Repeat Region is applied-once again I am allowed to apply it but the rollover stops working, a kind of interference. ANY SOLLUTION PLEASE, ANY WORK AROUND.

    I gotta tell you, using multiple images for rollover effects is going to be a big challenge in your dynamic scenario. 
    If this were my product page, I would use thumbnails with a bit of CSS opacity or filters to desaturate and make them full opacity/color on mouse over.  Nice effect with much less bandwidth. Easily done globally with CSS code.
    Two examples:
    http://alt-web.com/GALLERY/GalleryTest.php
    http://alt-web.com/TEMPLATES/CSS-Semi-liq-photo-sheet.shtml
    Nancy O.

  • Hyperlinks (a href) have been outdated for a very long time

    I have felt that "a href" hyprlinks have been outdated for a very long time.
    so much so that they should no longer be known to most developers (like frames and html based hyperlink maps)
    the only thing they can do is cause a new page to load.
    The event which causes this has to be a click (on an image, string of text or other compatable html object).
    and the event itself has to be the loading of a new page.
    by comparason by using javascript you can use ANY html attribute (Onclick, OnMouseover, OnMouseout ect) to trigger to event
    or you can use ANY event listener to trigger the event.
    You could also set these to only trigger under specific circumstances. (for example an onclick event followed by a mouse over at least 3 seconds later)
    and the event which occurs can be anything
    it could be a simple page change
    or it could bea page change with a modificaton (such as a preloader displaying to hide the fact that the new page is loading)
    or it could be a page change, or an animation, or anything else you can think of determined by an IF statemen
    furthermore by using AJAX or JQUERY you can have an event which alters a database or SQL file or, in simple cases plain text file based on the users actions - and then determines the action it performs based on the content of the database or SQL file or text file
    This coveres everything from very simple actions
    to very complex dynamic actions which have an infininte range of possibilites
    so
    why are is a href still used?

    I don't think any serious web developer can afford to make sweeping assumptions like this.  Anchor tags are stable and 100% reliable on all platforms & web devices.  It might surprise you to know that many people disable client-side scripts for security reasons.  How will these people interact with your site?
    https://addons.mozilla.org/en-US/firefox/addon/noscript/
    Nancy O.

  • Is it ok to set "auto shrink" ON for reportserver and reportserverTempDB database?

    Is it ok to set "auto shrink" ON for the reportserver and reportserverTempDB databases?
    I have read multiple 'best practice' advice articles to not se 'auto shrink' to true, and I understand the potential issue with index fragmentation, but I percieve this advice is for user created application databases and not for the ReportServer and
    reportserverTempDB databases.

    Hello Tom,
    Could you put it on - sure. Should you put it on - (My personal opinion is) never.
    There is more to it than just fragmenting indexes.
    About every half an hour or so the internal information kept about the database will be used to figure out if it should and can shrink and to about what size. This might be great for overnight when no one is using it, but the first time new extents are needed
    (since it just shrunk very small from no/light usage of reporting services) all requests will be halted until the data and log files grow. Basically you'll create a cycle of start->grow->stop->shrink->grow->shrink->grow->ad nausea. If
    you don't want to waste the resources of constantly growing and shrinking (plus all of the stopping when it happens) then don't enable it. If you don't care or it's not a big deal, go ahead and enable it but I'm not quite sure why a few MB would make a big
    difference?
    The reporting server databases are still "user" databases, they are not special system databases.
    Hope this helps.
    Sean Gallardy | Blog |
    Twitter

  • Very Big Table (36 Indexes, 1000000 Records)

    Hi
    I have a very big table (76 columns, 1000000 records), these 76 columns include 36 foreign key columns , each FK has an index on the table, and only one of these FK columns has a value at the same time while all other FK have NULL value. All these FK columns are of type NUMBER(20,0).
    I am facing performance problem which I want to resolve taking in consideration that this table is used with DML (Insert,Update,Delete) along with Query (Select) operations, all these operations and queries are done daily. I want to improve this table performance , and I am facing these scenarios:
    1- Replace all these 36 FK columns with 2 columns (ID, TABLE_NAME) (ID for master table ID value, and TABLE_NAME for master table name) and create only one index on these 2 columns.
    2- partition the table using its YEAR column, keep all FK columns but drop all indexes on these columns.
    3- partition the table using its YEAR column, and drop all FK columns, create (ID,TABLE_NAME) columns, and create index on (TABLE_NAME,YEAR) columns.
    Which way has more efficiency?
    Do I have to take "master-detail" relations in mind when building Forms on this table?
    Are there any other suggestions?
    I am using Oracle 8.1.7 database.
    Please Help.

    Hi everybody
    I would like to thank you for your cooperation and I will try to answer your questions, but please note that I am a developer in the first place and I am new to oracle database administration, so please forgive me if I did any mistakes.
    Q: Have you gathered statistics on the tables in your database?
    A: No I did not. And if I must do it, must I do it for all database tables or only for this big table?
    Q:Actually tracing the session with 10046 level 8 will give some clear idea on where your query is waiting.
    A: Actually I do not know what you mean by "10046 level 8".
    Q: what OS and what kind of server (hardware) are you using
    A: I am using Windows2000 Server operating system, my server has 2 Intel XEON 500MHz + 2.5GB RAM + 4 * 36GB Hard Disks(on RAID 5 controller).
    Q: how many concurrent user do you have an how many transactions per hour
    A: I have 40 concurrent users, and an average 100 transaction per hour, but the peak can goes to 1000 transaction per hour.
    Q: How fast should your queries be executed
    A: I want the queries be executed in about 10 to 15 seconds, or else every body here will complain. Please note that because of this table is highly used, there is a very good chance to 2 or more transaction to exist at the same time, one of them perform query, and the other perform DML operation. Some of these queries are used in reports, and it can be long query(ex. retrieve the summary of 50000 records).
    Q:please show use the explain plan of these queries
    A: If I understand your question, you ask me to show you the explain plan of those queries, well, first, I do not know how , an second, I think it is a big question because I can not collect all kind of queries that have been written on this table (some of them exist in server packages, and the others performed by Forms or Reports).

Maybe you are looking for

  • How to disable only inactive firewall notification in the action center by GPO

    Hi I need to disable only inactive notifications firewall in the action center PCs through GPO. I have seen several posts and it seems possible disable by GPO only all notifications in the action center and not only firewall notifications. I have a D

  • Oracle.jbo.JboException: JBO-3300 on Jdevelooper 9.0.2.8

    Hi All, When i tried to run my test.jpr project, project get compiled successfuly and run first index.jsp but when i click any link on index.jsp it gives an error oracle.jbo.JboException, msg=JBO-33001: Cannot find the configuration file /org/act/epr

  • AM Activation/Passivation Problem

    Dear All, Jdeveloper 11.1.1.5.0 + ADF faces + ADF BC I have a procedure in database that just inserts the data in a table but doesnot commits.. Below is the procedure Procedure p_insert_emp_test Is Begin Insert Into emp_r (Emp_Id,f_Name,l_Name,Dept_I

  • Quadro or ati?

    I'm planning on buying a mac pro and i can't decide wether i should get a quadro or ati graphics card. I will mainly be using this to do work in maya, photoshop, etc. but i would like to know if the quadro would deliver the same gaming performance as

  • Porting a Workshop app to another J2EE server?

    Hello all. I'm new to WL Workshop and have been reading abou the required Workshop Runtime. Is this something that could be ported to another J2EE compliant server? I noticed the PageFlow Portability Kit for the page flows... but what about the Runti