Insert and Updateusing Bulk Operations

Hi,
We are on 10g (10.2.0.5.0).
We have a batch that will insert/update ~30mm records in each load. The logic is usual if a record exists in target table update it if not insert.
Considering the amount of data to be loaded simple For loop will be very performance intensive. I am looking for a way to do this upsert using bulk operations if possible and want to check if someone can share their experience with similar situations.
Off course all 30mm records will not be loaded in one go, these will be divided into small chunks say 100k.
Thanks
AT

MERGE should be more efficient than rolling your own merge process using loops and collections. I'm not sure what "poor performance" you're referring to.
You could, of course, create a cursor on the source table, loop through the cursor doing a BULK COLLECT with a LIMIT into a local collection, do a FORALL with a SAVE EXCEPTIONS to try to insert the data and then go through the exceptions to issue an update. You could also iterate over the collection in a FOR loop, issue the UPDATE, check how many rows were modified, and do an INSERT if there were no rows updated. Either option should be substantially slower than a single MERGE but it would let you use bulk operations. Which of the two is more efficient will depend on your ratio of inserts to updates.
Justin

Similar Messages

  • Using Bulk operations for INSERT into destination table and delete from src

    Hi,
    Is there any way to expediate the process of data movement?
    I have a source set of tables (with its pk-fk relations) and a destination set of tables.
    Currently my code as of now, is pickin up the single record in cursor from the parentmost table, and then moving the data from other respecitve tables. But this is happening one by one... Is there any way I can make this take less time?
    If I use bulk insert and collections, i will not be able to use the DELETE in the same block for same source record.
    Thanks
    Regards
    Abhivyakti

    Abhivyakti
    I'm not 100% sure how your code flows from what you've stated, but generally you should try and avoid cursor FOR LOOPS and possibly BULK COLLECTING.
    I always follow the sequence in terms of design:
    1. Attempt to use bulk INSERTS, UPDATES and/or DELETES first and foremost. (include MERGE as well!)
    2. If one cannot possibly do the above then USE BULK COLLECTIONS using a combination of RETURNING INTO's and
    FORALL's.
    However, before you follow this method and if you relatively new to Oracle PL/SQL,
    share the reason you cannot follow the first method on this forum, and you're bound to find some
    help with sticking to method one!
    3. If method two is impossible, and there would have to be a seriously good reason for this, then follow the cursor FOR LOOP
    method.
    You can combine BULK COLLECTS with UPDATES and DELETES, but not with INSERTS
    bulk collect into after insert ?
    Another simple example of BULK COLLECTING
    Re: Reading multiple table type objects returned
    P;

  • ODBC, bulk inserts and dynamic SQL

    I am writing an application running on Windows NT 4 and using the oracle ODBC driver (8.01.05.00, that inserts many rows at a time (10000+) into an oracle 8i database.
    At present, I am using a stored procedure to insert each row into the database. The stored procedure uses dynamic SQL because I can only determine the table and field names at run time.
    Due to the large number of records, it tends to take a while to perform all the inserts. I have tried a number of solutions such as using batches of SQL statements (e.g. "INSERT...;INSERT...;INSERT..."), but the oracle ODBC driver only seems act on the first statement in the batch.
    I have also considered using the FOR ALL statement and SQL*Loader utility.
    My problem with FOR ALL is that I'm not sure it works on dynamic SQL statements and even if it did, how do I pass an array of statements to the stored procedure.
    I ruled out SQL* Loader because I could not find a way to invoke it it from an ODBC statement. Secondly, it requires the spawining of a new process.
    What I am really after is something similar the the SQL Server (forgive me!) BULK INSERT statement where you can simply create an input file with all the records you want to insert, and pass it along in an ODBC statement such as "BULK INSERT <filename>".
    Any ideas??
    null

    Hi,
    I faced this same situation years ago (Oracle 7.2!) and had the following alternatives.
    1) Use a 3rd party tool such as Sagent or CA Info pump (very pricey $$$)
    2) Use VisualC++ and OCI to hook into the array insert routines (there are examples of these in the Oracle Home).
    3) Use SQL*Loader (the best performance, but no real control of what's happening).
    I ended up using (2) and used the Rouge Wave dbtools.h++ library to speed up the development.
    These days, I would also suggest you take a look at Perl on NT (www.activestate.com) and the DBlib modules at www.perl.org. I believe they will also do bulk loading.
    Your problem is that your program is using Oracle ODBC, when you should be using Oracle OCI for best performance.
    null

  • Bulk inserts and dynamic SQL

    I am writing an application running on Windows NT 4 and using the oracle ODBC driver (8.01.05.00, that inserts many rows at a time (10000+) into an oracle 8i database.
    At present, I am using a stored procedure to insert each row into the database. The stored procedure uses dynamic SQL because I can only determine the table and field names at run time.
    Due to the large number of records, it tends to take a while to perform all the inserts. I have tried a number of solutions such as using batches of SQL statements (e.g. "INSERT...;INSERT...;INSERT..."), but the oracle ODBC driver only seems act on the first statement in the batch.
    I have also considered using the FOR ALL statement and SQL*Loader utility.
    My problem with FOR ALL is that I'm not sure it works on dynamic SQL statements and even if it did, how do I pass an array of statements to the stored procedure.
    I ruled out SQL* Loader because I could not find a way to invoke it it from an ODBC statement. Secondly, it requires the spawining of a new process.
    What I am really after is something similar the the SQL Server (forgive me!) BULK INSERT statement where you can simply create an input file with all the records you want to insert, and pass it along in an ODBC statement such as "BULK INSERT <filename>".
    Any ideas??
    null

    Hi,
    I faced this same situation years ago (Oracle 7.2!) and had the following alternatives.
    1) Use a 3rd party tool such as Sagent or CA Info pump (very pricey $$$)
    2) Use VisualC++ and OCI to hook into the array insert routines (there are examples of these in the Oracle Home).
    3) Use SQL*Loader (the best performance, but no real control of what's happening).
    I ended up using (2) and used the Rouge Wave dbtools.h++ library to speed up the development.
    These days, I would also suggest you take a look at Perl on NT (www.activestate.com) and the DBlib modules at www.perl.org. I believe they will also do bulk loading.
    Your problem is that your program is using Oracle ODBC, when you should be using Oracle OCI for best performance.
    null

  • Can we use both INSERT and UPDATE at the same time in JDBC Receiver

    Hi All,
    I would like to know is it possible to use both INSERT and UPDATE at the same time in one interface because I have a requirement in which I have to perform both the task.
    user send the file which contains both new and old record and I need to save those in MS SQL database.
    If the record exist then use UPDATE otherwise use INSERT.
    I looked on sdn but didn't find any blog which perform both the things at the same time.
    Interface Requirement
    FILE -
    > PI -
    > JDBC(INSERT & UPDATE)
    I am thinking to use JDBC Lookup but not sure if it good to use for bulk record.
    Can somebody please suggest me something or send me the link of any blog or anything to solve this problem.
    Thanks,

    Hi ,
              If I have understood properly the scenario properly,you are not performing insert and update together. As you posted
    "If the record exist then use UPDATE otherwise use INSERT."
    Thus you are performing either an insert or an update which depends on outcome of a search if the records already exist in database or not. Obviously to search the tables you need " select * from ...  where ...." query. If your query returns some results you proceed with update since this means there are some old records already in database. If your query returns no rows  you proceed with "insert into tablename....." since there are no old records present in database.
      Now perhaps the best method to do the searching, taking a decision to insert or update, and finally insert or update operation is to be done by a stored procedure in MS SQL database.  A stored procedure is a subroutine available to applications accessing a relational database system. Here the application is PI server.   If you need further help on how to write and call stored procedure in MS SQL you can look into these links
    http://www.daniweb.com/web-development/databases/ms-sql/threads/146829
    http://www.sqlteam.com/article/stored-procedures-parameters-inserts-and-updates
    [ This part you can ignore, Since its not sure that you will face this situation
        Still you might face some problems while your scenario runs. Lets consider this scenario, after the stored procedure searches the database it found no rows. Thus you proceed with an insert operation. If your database table is being accessed by multiple applications (or users) other than yours then it is very well possible that after the search operation completed with a null result, an insert/update operation has been performed by some other application with the same primary key. Now when you are trying to insert another row with same primary key you get an error message like "duplicate entry not possible for same primary key value". Thus you need to be careful in this respect. MS SQL has a feature called "exclusive locks ". Look into these links for more details on the subject
    http://msdn.microsoft.com/en-us/library/aa213039(v=sql.80).aspx
    http://www.mssqlcity.com/Articles/Adm/SQL70Locks.htm
    http://www.faqs.org/docs/ppbook/r27479.htm
    http://msdn.microsoft.com/en-US/library/ms187373.aspx
    http://msdn.microsoft.com/en-US/library/ms173763.aspx
    http://msdn.microsoft.com/en-us/library/e7z8d5hf(v=vs.80).aspx
    http://mssqlserver.wordpress.com/2006/11/08/locks-in-sql/
    http://www.mollerus.net/tom/blog/2008/03/using_mssqls_nolock_for_faster_queries.html
        There must be other methods to avoid this problem. But the point is you need to be sure that all access to database for insert/update operations are isolated.
    regards
    Anupam

  • Import/Export Wizard is a bulk operation?

    If I use import/export wizard to load data from a file to a table will it do it as a bulk operation and will be minimally logged when using bulk logged recovery mode?
    http://technet.microsoft.com/en-us/library/ms191244(v=sql.105).aspx  

    Yes. The Import Export wizard creates a SSIS package on the backend which will have default transfer option as below
    This is same as BULK INSERT and so it would be minimally logged in case of BULK recovery mode.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Modifiy number of inserts and number of errors, etc variables

    Hi!
    i need your help...
    when you execute an 'normal' interface (just copy table to another table) and then, you go to the Metadata Navigator (or operator), in each step of the ikm (and other), is a parameter that indicates de number of error, number of inserts and number of updates.
    i want to set the value of this parameter manually.... knows anyone how can i do this?
    i tried to add <%odiRef.setNbInsert(10)%> in one step on the IKM (after do the insert action), also alfter the step for commit action... but it doesn't work.
    did i something wrong?
    Thanks!

    haaah!!! I never realized that. For some reason, when i first started in Logic Pro 9, i had like 14 inserts already on the channel strips, and I've been working with that same template since first opening LP9 (which i've only had it since Aug 15).
    So I see now, it's the friendly PluginsertSendslotfairy. :O)
    Thanks!

  • Quick questions (bulk operations, WAN, provisioning, split brain)

    Thank you for your assistance with the following questions..
    1. Can I use the "disk overflow" feature when using the "partitioned cache" topology? I need to just do puts and gets.
    2. Are bulk operations (i.e. bulk put, bulk get, bulk erase) available in all cache topologies? I am interested in bulk put/get when using a "partitioned cache" and when using your "WAN capability" to connect geographical dispersed clusters (Coherence*Extend). Do you guarantee data and event delivery over the WAN?
    3. Does the WAN capability (Coherence*Extend) allow multiple clusters to talk with each other (N:N) or does it only support a "star topology" (1 main cluster connected to N other "peripheral" clusters)?
    4. As a result of a load bottleneck (i.e. a cache server(s) running too hot) in a "partitioned region" topology, can I "on-the-fly" add new cache servers/boxes to load balancing my "client load"? What client load balancing policies are typically recommended (i.e. sticky, random)? How does dynamic provisioning of new servers (e.g., adding more RAM to my distributed cache) affect the hashing of my map regions and my network bandwidth? What kind of performance would I expect if I add one or more new cache members when I was close to have an OOM on one or more of my already running cache servers? Even if I was not close to an OOM, what kind of network spike would I expect and what kind of client operation latency should I expect?
    5. Could I have more information to understand your "data location transparency"?
    6. How do you solve "split brain" problems (i.e. sets of members get separated and later try to rejoin cluster)? Do you have or is it possible to implement a "quorum-based policy" to decide who should live and who should die? We have seen "split brain" problems with messaging retransmission storms in the past and are looking for options.

    Hi user614602,
    1 - 3. Yes to all.
    4 - 6. Those are quite advanced and open-ended questions (and by no means "quick"). Please contact your sales rep to schedule a conference call to discuss your requirements and how Coherence addresses the issues you rise.
    Regards,
    Gene

  • Goldengate Extracts reads slow during Table Data Archiving and Index Rebuilding Operations.

    We have configured OGG on a  near-DR server. The extracts are configured to work in ALO Mode.
    During the day, extracts work as expected and are in sync. But during any dialy maintenance task, the extracts starts lagging, and read the same archives very slow.
    This usually happens during Table Data Archiving (DELETE from prod tables, INSERT into history tables) and during Index Rebuilding on those tables.
    Points to be noted:
    1) The Tables on which Archiving is done and whose Indexes are rebuilt are not captured by GoldenGate Extract.
    2) The extracts are configured to capture DML opeartions. Only INSERT and UPDATE operations are captured, DELETES are ignored by the extracts. Also DDL extraction is not configured.
    3) There is no connection to PROD or DR Database
    4) System functions normally all the time, but just during table data archiving and index rebuild it starts lagging.
    Q 1. As mentioned above, even though the tables are not a part of capture, the extracts lags ? What are the possible reasons for the lag ?
    Q 2. I understand that Index Rebuild is a DDL operation, then too it induces a lag into the system. how ?
    Q 3. We have been trying to find a way to overcome the lag, which ideally shouldn't have arised. Is there any extract parameter or some work around for this situation ?

    Hi Nick.W,
    The amount of redo logs generated is huge. Approximately 200-250 GB in 45-60 minutes.
    I agree that the extract has to parse the extra object-id's. During the day, there is a redo switch every 2-3 minutes. The source is a 3-Node RAC. So approximately, 80-90 archives generated in an hour.
    The reason to mention this was, that while reading these archives also, the extract would be parsing extra Object ID's, as we are capturing data only for 3 tables. The effect of parsing extract object id's should have been seen during the day also. The reason being archive size is same, amount of data is same, the number of records to be scanned is same.
    The extract slows down and read at half the speed. If normally it would take 45-50 secs to read an archive log of normal day functioning, then it would take approx 90-100 secs to read the archives of the mentioned activities.
    Regarding the 3rd point,
    a. The extract is a classic extract, the archived logs are on local file system. No ASM, NO SAN/NAS.
    b. We have added  "TRANLOGOPTIONS BUFSIZE" parameter in our extract. We'll update as soon as we see any kind of improvements.

  • Is there some way I can insert and play these additional audio files, accessible via separate button

    I am converting a CD-ROM course to an online format with that software.
    In the majority of slides created I have used the Import Audio function to insert a primary audio file containing important curriculum information.
    This file typically activates the moment you open the slide.
    So far so good.
    However, in many slides I also need to insert some secondary audio files.
    While the Import Audio function permits me to splice several files into one, this isn't what I want.
    Instead, I need to insert one or more additional audio files that can be activated independently via separate buttons on a given slide.
    I can in fact do this with the Insert Button function, but in the process a Windows Media Player file opens when I convert the project into HTML.
    My question:
    Is there some way I can insert and play these additional audio files, accessible via separate buttons, without creating and opening a Windows Media Player file?
    Thanks

    It is hard to assess the situation without sitting at your computer and having a look myself.  First, if you're seeing space savings I doubt all your files are a compressed format.  Frequently if you try compressing a compressed format file such as AAC (don't confuse with AIFF) you will actually find the file gets larger because there is overhead to compression and if it cannot compress any more then you end up just adding overhead on top of no savings.  Second, normally if you zip the iTunes library it would basically make it inaccessible to iTunes and iTunes would make you a brand new library or at least flag all your tracks as missing (as you are seeing ain part).
    I don't know what computer you are running, which operating system, which version of iTunes.  Unless they have done something radical in Lion in terms of letting you zip your iTunes folder and still have it functional, you either didn't zip the whole thing, or... I don't know.  One scenario is that somehow your media have been split up and you have them in several locations.  One isn't really part of your library except for 80 tracks, so when you compressed it (and I still am puzzled by the compression unless it was AIFF files you were no longer using), it didn't really make much difference to the playability.  Then you have an untouched library elsewhere.
    Use the Reveal In Finder command to see where some of those problem files are located vs. where the playable files are located.

  • How change state when card is inserted and pulled out

    It is not Java Card question, but I think it is better place to post question like this. I have running Java Card applet on card, the communication between card and host application works, but could you anybody tell me how handle inserting and pulling card out? My idea is, I have running Java 6 SE application:
    1. and I will insert the card into a reader and application will automatically find out that card was inserted and write out it.
    2. and I will pull out the card from a reader and application will automatically find out that card was pulled out and write out it.
    Could you anybody help me, please?
    Thank you very much for any answer.
    Martin

    mFilo wrote:
    It is not Java Card question, but I think it is better place to post question like this. I have running Java Card applet on card, the communication between card and host application works, but could you anybody tell me how handle inserting and pulling card out? My idea is, I have running Java 6 SE application:
    1. and I will insert the card into a reader and application will automatically find out that card was inserted and write out it.
    2. and I will pull out the card from a reader and application will automatically find out that card was pulled out and write out it.I think you can address only terminals. A way to identify a card is with additional info like ATR, CPLC data or specific applet's file control information (response to select command). I found in javax.smartcardio.CardTerminals following methods that might help:
    waitForChange
    public void waitForChange()
                       throws CardException
        Waits for card insertion or removal in any of the terminals of this object.
        This call is equivalent to calling waitForChange(0).
        Throws:
            IllegalStateException - if this CardTerminals object does not contain any terminals
            CardException - if the card operation failed
    waitForChange
    public abstract boolean waitForChange(long timeout)
                                   throws CardException
        Waits for card insertion or removal in any of the terminals of this object or until the timeout expires.
        This method examines each CardTerminal of this object. If a card was inserted into or removed from a CardTerminal since the previous call to waitForChange(), it returns immediately. Otherwise, or if this is the first call to waitForChange() on this object, it blocks until a card is inserted into or removed from a CardTerminal.
        If timeout is greater than 0, the method returns after timeout milliseconds even if there is no change in state. In that case, this method returns false; otherwise it returns true.
        This method is often used in a loop in combination with list(State.CARD_INSERTION), for example:
          TerminalFactory factory = ...;
          CardTerminals terminals = factory.terminals();
          while (true) {
              for (CardTerminal terminal : terminals.list(CARD_INSERTION)) {
                  // examine Card in terminal, return if it matches
              terminals.waitForChange();
        Parameters:
            timeout - if positive, block for up to timeout milliseconds; if zero, block indefinitely; must not be negative
        Returns:
            false if the method returns due to an expired timeout, true otherwise.
        Throws:
            IllegalStateException - if this CardTerminals object does not contain any terminals
            IllegalArgumentException - if timeout is negative
            CardException - if the card operation failed

  • What is difference of truncate/insert and delete/insert ?

    hi all:
    what is difference of truncate/insert and delete/insert ?

    Hi,
    Truncate will truncate the table, which means there is a DDL operation to empty the table. DDL operations cannot be rolled back, but truncate is much faster than delete (because of that reason). If you do not require recovery of the deleted records in case of a failure, then truncate/insert is more optimal than delete/insert. Notice that for truncate to work, you cannot have enabled foreign keys pointing to the table. Truncate will never perform cascaded deletes.
    Hope this explains.
    Mark.

  • Since installing LTR 5.4, which I've now upgraded to 5.6, I've encountered repeated slowness and malfunctions in operations, especially when using the Compare View function and the Tab key to open and close the right and left side panels.  Such problems n

    Since installing LTR 5.4, which I've now upgraded to 5.6, I've encountered repeated slowness and malfunctions in operations, especially when using the Compare View function and the Tab key to open and close the right and left side panels.  Such problems never arose during two years of using LTR-4 and nothing else has changed on my computer.  I have a pretty simple system with only a few plug-ins, which are usually not in operation.  I have 12GB of RAM in my Windows 7 PC.  I could illustrate these problems with screen shots if you would tell me how to submit screen shots.  Otherwise I will try to describe the problems in words.
    The problem is clearly cumulative, growing worse as usage time passes.  Compare View feature gradually slows down and eventually seems to choke as my work session proceeds. If I Exit LTR and re-enter and start all over, things will work normally for maybe 30 minutes, but then the Compare View feature begins to become very slow to respond.   In a recent example with my screen full of thumbnails in Library mode I highlighted two images to compare. LTR started to open the Compare View screen by first having the top row of thumbnails disappear to be replaced by the "SELECT" and "CANDIDATE" words in their spaces  (but no images), but Compare View never succeeded in gaining control of the screen. After some seconds the top row of thumbnails reasserted its position and the Compare View windows disappeared. But LTR kept trying to bring them back. Again the top row of thumbnails would go away, Select and candidate would reappear, try again, and give up. This went on for at least 2-3 minutes before I tried to choose File and Exit, but even that did not initially want to respond. It doesn't like to accept other commands when it's trying to open Compare View. Finally it allowed me to exit.
    To experiment I created a new catalog of 1100 images.  After 30-40 minutes, the Compare View function began to operate very slowly. With left and right side panels visible and two thumbnails highlighted, hitting Compare View can take half a minute before the two mid-size  images open in their respective SELECT and CANDIDATE windows. When the side panels are open and two images are in the Select/Candidate spaces, hitting the Tab button to close the side panels produces a very delayed response--25-30 seconds to close them, a few more seconds to enlarge the two images to full size. To reverse the process (i.e., to recall the two side panels), hitting Tab would make the two sides of the screen go black for up to a minute, with no words visible. Eventually the info fields in the panels would open up.
    I also created a new user account and imported a folder of 160 images. After half an hour Compare View began mis-placing data.  (I have a screen shot to show this.)  CANDIDATE appears on the left side of SELECT, whereas it should be on the right. The accompanying camera exposure data appears almost entirely to the left of the mid-screen dividing line. Although the Candidate and Select headings were transposed, the image exposure data was not, but the data for the image on the right was almost entirely to the left of the line dividing the screen in two.
    Gurus in The Lightroom Forum have examined Task Manager data showing Processes running and Performance indicators and they see nothing wrong.  I could also send screen shots of this data.
    At this point, the only way I can process my images is to work 30-40 minutes and then shut down everything, exit, and re-start LTR.  This is not normal.  I hope you can find the cause, and then the solution.  If you would like to see my screen shots, tell me how to submit them.
    Ollie
    [email protected]

    Since installing LTR 5.4, which I've now upgraded to 5.6, I've encountered repeated slowness and malfunctions in operations, especially when using the Compare View function and the Tab key to open and close the right and left side panels.  Such problems never arose during two years of using LTR-4 and nothing else has changed on my computer.  I have a pretty simple system with only a few plug-ins, which are usually not in operation.  I have 12GB of RAM in my Windows 7 PC.  I could illustrate these problems with screen shots if you would tell me how to submit screen shots.  Otherwise I will try to describe the problems in words.
    The problem is clearly cumulative, growing worse as usage time passes.  Compare View feature gradually slows down and eventually seems to choke as my work session proceeds. If I Exit LTR and re-enter and start all over, things will work normally for maybe 30 minutes, but then the Compare View feature begins to become very slow to respond.   In a recent example with my screen full of thumbnails in Library mode I highlighted two images to compare. LTR started to open the Compare View screen by first having the top row of thumbnails disappear to be replaced by the "SELECT" and "CANDIDATE" words in their spaces  (but no images), but Compare View never succeeded in gaining control of the screen. After some seconds the top row of thumbnails reasserted its position and the Compare View windows disappeared. But LTR kept trying to bring them back. Again the top row of thumbnails would go away, Select and candidate would reappear, try again, and give up. This went on for at least 2-3 minutes before I tried to choose File and Exit, but even that did not initially want to respond. It doesn't like to accept other commands when it's trying to open Compare View. Finally it allowed me to exit.
    To experiment I created a new catalog of 1100 images.  After 30-40 minutes, the Compare View function began to operate very slowly. With left and right side panels visible and two thumbnails highlighted, hitting Compare View can take half a minute before the two mid-size  images open in their respective SELECT and CANDIDATE windows. When the side panels are open and two images are in the Select/Candidate spaces, hitting the Tab button to close the side panels produces a very delayed response--25-30 seconds to close them, a few more seconds to enlarge the two images to full size. To reverse the process (i.e., to recall the two side panels), hitting Tab would make the two sides of the screen go black for up to a minute, with no words visible. Eventually the info fields in the panels would open up.
    I also created a new user account and imported a folder of 160 images. After half an hour Compare View began mis-placing data.  (I have a screen shot to show this.)  CANDIDATE appears on the left side of SELECT, whereas it should be on the right. The accompanying camera exposure data appears almost entirely to the left of the mid-screen dividing line. Although the Candidate and Select headings were transposed, the image exposure data was not, but the data for the image on the right was almost entirely to the left of the line dividing the screen in two.
    Gurus in The Lightroom Forum have examined Task Manager data showing Processes running and Performance indicators and they see nothing wrong.  I could also send screen shots of this data.
    At this point, the only way I can process my images is to work 30-40 minutes and then shut down everything, exit, and re-start LTR.  This is not normal.  I hope you can find the cause, and then the solution.  If you would like to see my screen shots, tell me how to submit them.
    Ollie
    [email protected]

  • Missing Format Text, Insert, and Edit Source, etc Options on Some Pages

    I’m having trouble understanding why certain pages have more edit options than others. I’m new to Sharepoint Workspace 2010 and I’ve been reading/ researching for answers for 2 days now. I tried to HyperSnap in some pictures however it will not allow me
    to until my account is verified and it's not sending a verification email, sorry!
    Issue: 
    On some pages, I have the edit pencil with lots of great edit options such as Format Text, Insert and Edit Source. (This is good!)
    On some pages I have no edit pencil, but when I go to Page
    à Edit Page à I get few edit options, not including Format Text, Insert or Edit Source. (This is not good)
    After reading other somewhat similar posts, I noticed that on the pages where I get lots of edit options, they show SitePages in the URL. Where I get few edit options, SitePages is not in the URL.
    Is there a fix so that I can have the expanded edit options in all of my pages? Thank you for your help!

    Hi,
    In SharePoint we don’t have edit page options [ribbon toolbar page edit formatting options] for all types of page for application page we cannot edit application page like site pages
    again the missing edit button might be due to many causes:  
    Can you verify that it's not a permissions issue? Try navigating to the page in edit mode by appending to the url: your-site/Pages/YourPage.aspx?ControlMode=Edit&DisplayMode=Design
    Or you can try this
    your-site/Pages/YourPage.aspx? ?ToolPaneView=2
    1. You can check  that on some pages the ribbon is hidden by default. You need to make it visible by clicking on "Show Ribbon" on the "Site Actions" drop-down menu.
    2. You haven't been doing Visual Upgrade (which actually replaces your master page with the v4.master)
    3. If you have different master page and may be your master page is not 100% compatibile with SP2010, nor does it use the elements from v4.master (collaboration) or the example nightandday.master
    To add the button in your page you could use the <SharePoint:PageStateActionButton id="PageStateActionButton" runat="server" Visible="false" /> (if you've seen already that would generate the button on the left side of
    the ribbon)!
    Please can you ping the URL of the page and how you crated that page to further assist, we would be happy to help.
    Krishana Kumar http://www.mosstechnet-kk.com

  • How to insert and edit equations with Math Type in IBA??

    Hi,
    I want to add  fractions in IBA with Math Type.
    Here is what i found:
    And here is the General Preferences
    Like you see, i can't select Insert and edit equations with Math Type, it's grey.
    How can i use Math Type?
    Tx

    We have an article that describes not only how to use MathType with iBooks Author (iBA), but also using LaTeX and MathML. http://www.dessci.com/en/support/mathtype/works_with.asp#!target=ibooks_author_m ac
    It's important to note that no matter how you get equations into iBA, all equations are represented as MathML in the published iBook. (except, obviously I hope, equations that are simply images that you insert)
    If you need more help, feel free to ask here.
    Bob Mathews
    Design Science

Maybe you are looking for