Taking a long time to export

Why is it taking so long to exort a sile show with with a 185 photos to idvd?its been over 5min and its only half way done. Thanks

When you use the Share to iDVD export mode iPhoto has to create a Quicktime movie of the slideshow prior to exporting it into iDVD. If you've got lots of KB effects, fancy transitions, etc. along with those 185 photos it's going to take a long live. The 99 slide limit refers to creating a slideshow in iDVD with imported stills.

Similar Messages

  • Report script taking very long time to export in ASO

    Hi All,
    My report script is taking very long time to execute and finally a message appears as timed out.
    I'm working on ASO Cubes and there are 14 dimensions for which i need to export all data for all the dimensions for only one version.
    The data is very huge and the member count in each dimension is also huge, so which is making me difficult to export the data.
    Any suggestions??
    Thanks

    Here is a link that addresses several ways to optimize your report script. I utilize report scripts for Level 0 exports in an ASO environment as well, however the majority of our dimemsions are attribute dimensions.
    These are the most effective solutions we have implemented to improve our exports via report scripts:
    1. Make sure your report script is written in the order of how the Report Extractor retrieves data.
    2. Supressing Zero and Missing Data
    3. We use the LINK command within reports for some dimensions that are really big and pull at Level 0
    4. Using Symmetric reports.
    5. Breakout the exports in multiple reports.
    However, you may also consider some additional solutions outlined in this link:
    1. The MDX optimizing commands
    2. Back end system settings
    http://download.oracle.com/docs/cd/E12825_01/epm.111/esb_dbag/drpoptim.htm
    I hope this helps. Maybe posting your report script would also help users to provide feedback.
    Thanks
    Edited by: ronnie on Jul 14, 2011 9:25 AM
    Edited by: ronnie on Jul 14, 2011 9:53 AM

  • Taking a long time to export to Quicktime

    Hello,
    I'm trying to export my Flash "file" to Quicktime. The issue I'm having is that it's taking a really long time to "record flash content". My file is pretty long (about 5-6min long) so maybe this is normal. Or is it?

    When you use the Share to iDVD export mode iPhoto has to create a Quicktime movie of the slideshow prior to exporting it into iDVD. If you've got lots of KB effects, fancy transitions, etc. along with those 185 photos it's going to take a long live. The 99 slide limit refers to creating a slideshow in iDVD with imported stills.

  • Taking a long time to export a report to PDF in web app

    Hi,
    I downloaded CR for Eclipse 2.0 to integrate a "Create PDF Report" option into our web application. The original .rpt files were created using CR 2008. I used the sample online code to do this:
    ReportClientDocument reportClientDoc = new ReportClientDocument();
    reportClientDoc.open(REPORT_NAME, 0);   
    // .. set report params here..
    ByteArrayInputStream byteArrayInputStream = (ByteArrayInputStream)reportClientDoc.getPrintOutputController().export(ReportExportFormat.PDF);
    I was able to successfully run this and generate a PDF output file, however the report takes over 2 minutes to complete, whether there were only 2 records or 75 records!
    From CR 2008 on a Windows machine, the export to PDF option takes about 2 - 3 seconds for the exact same report.
    Our web servers are very powerful Linux blade servers from Dell.
    Why is there such a delay? Does it have to do with the licensed version versus the free version? Or is this just a limitation of being in a web app?
    -Mike

    We created a version of the same test report that only generates two PDF pages.It has cut the time down, but it still takes it about 16 seconds to open the .rpt file and another 6 seconds to export to PDF and write it out to a file. [Total of about 22 seconds.]
    When we run the same report on CR2008 on a Windows laptop, it takes less than 1 second to load the .rpt file, under 1 second to generate the report and another 1 second to generate the PDF. [Total under 3 seconds.]
    Our question still remains: Why is there a considerable time difference between CR2008 and CR4Eclipse [3 secs vs 22 secs] and is there anything we can do on the web app side to decrease the report output time?
    Here are the "highlights" from the verbose logs (the full logs were about 32K). I tried showing where there were obvious time gaps. If there is a big time gap between statements separated by ".... " it was just doing it's regular work (evaluating functions, etc). If you provide me an email address I can send the complete logs as an attachment.
    2009-09-04 16:18:32,212 DEBUGaction.TestReportsAction (?:?) - begin hub/TestReportsAction
    2009-09-04 16:18:32,213 DEBUGaction.TestReportsAction (?:?) - 1-calling new ReportClientDocument();
    2009-09-04 16:18:32,215 DEBUGaction.TestReportsAction (?:?) - 2-calling .open(REPORT_NAME
    2009-09-04 16:18:33,193 INFO reportloading.saveddata (?:?) - FieldManager
    2009-09-04 16:18:33,194 INFO reportloading.saveddata (?:?) - NDatabaseFields: 63
    2009-09-04 16:18:33,195 INFO reportloading.saveddata (?:?) - Field: {INVOICE_HEADER_VIEW.INVOICE}, NBytesInField: 8, ValueType: number
    2009-09-04 16:18:33,288 INFO reportloading.saveddata (?:?) - Field: {%Unit_Price}, NBytesInField: 131070, ValueType: string
    2009-09-04 16:18:33,289 INFO reportloading.saveddata (?:?) - NOlapDimensionFields: 0
    2009-09-04 16:18:33,289 INFO reportloading.saveddata (?:?) - NOlapGenerationFields: 0
    2009-09-04 16:18:33,290 INFO reportloading.saveddata (?:?) - NULL Field
    2009-09-04 16:18:37,141 INFO reportloading.saveddata (?:?) - FieldManager
    2009-09-04 16:18:37,142 INFO reportloading.saveddata (?:?) - NDatabaseFields: 8
    2009-09-04 16:18:37,177 INFO reportloading.saveddata (?:?) - NOlapGenerationFields: 0
    2009-09-04 16:18:37,177 INFO reportloading.saveddata (?:?) - NULL Field
    2009-09-04 16:18:39,333 INFO formulas.Compiler (?:?) - Formula {@Record_Selection} compiled.
    2009-09-04 16:18:39,340 INFO formulas.Compiler (?:?) - Formula {@Saved_Data_Selection} compiled.
    2009-09-04 16:18:47,868 INFO ReportCommand.SetReportDateCommand (?:?) - SetReportDateCommand,24667963,Exit perform,<document=ReportDocument:<report=main><reportPath=/webdata/webapps/tcmIS/reports/invoice_header_en.rpt>>;
    2009-09-04 16:18:48,042 INFO formulas.Compiler (?:?) - Formula {@Invtest} compiled.
    2009-09-04 16:18:48,045 INFO formulas.Compiler (?:?) - Formula {@Load column names} compiled.
    2009-09-04 16:18:48,701 DEBUGaction.TestReportsAction (?:?) - opened report: /reports/invoice_header_en.rpt
    2009-09-04 16:18:48,702 DEBUGaction.TestReportsAction (?:?) - 3-calling getParameterFieldController() - setting params
    2009-09-04 16:18:48,792 DEBUGaction.TestReportsAction (?:?) - 4-calling export(ReportExportFormat.PDF);
    2009-09-04 16:18:48,794 INFO reports.queryengine (?:?) - Connection logon:Database DLL=s(crdb_oracle.dll);PreQEServerName=s(tcmprod);Server=s(tcmprod);Tru
    sted_Connection=b(false);Password=;User ID=s(sskidmore)
    2009-09-04 16:18:48,812 INFO driverImpl.jdbc (?:?) - Connection Opened null
    2009-09-04 16:18:48,829 INFO reports.queryengine (?:?) - Connection logon successfully
    2009-09-04 16:18:48,835 INFO reports.dataengine (?:?) - Create new report state
    2009-09-04 16:18:48,838 INFO reports.dataengine (?:?) - Need to recompile formulas, the number of formulas needed to recompile is 1
    2009-09-04 16:18:55,202 WARN Simplifier.Fields (?:?) - Field {INVOICE_HEADER_VIEW.BILLING_ENTITY_ADDRESS_LINE_1} is null. Interrupting formula evaluation.
    2009-09-04 16:18:55,203 INFO formulas.Evaluator (?:?) - Formula {@-suppressSection-13} evaluated to: null
    2009-09-04 16:18:55,302 INFO formatter.paginator (?:?) - Page 2 formatted.
    2009-09-04 16:18:55,303 INFO lightmodel.FCMFormattedContentModeller (?:?) - FCM: modelling contents
    2009-09-04 16:18:55,344 INFO lightmodel.FCMFormattedContentModeller (?:?) - FCM: modelling contents
    2009-09-04 16:18:55,981 INFO formatter.paginator (?:?) - Moving to Page 3
    2009-09-04 16:18:56,015 WARN Simplifier.Fields (?:?) - Field {INVOICE_HEADER_VIEW.BILLING_ENTITY_ADDRESS_LINE_1} is null. Interrupting formula evaluation.
    2009-09-04 16:18:56,016 INFO formulas.Evaluator (?:?) - Formula {@-suppressSection-12} evaluated to: null
    2009-09-04 16:18:56,019 WARN Simplifier.Fields (?:?) - Field {INVOICE_HEADER_VIEW.BILLING_ENTITY_ADDRESS_LINE_1} is null. Interrupting formula evaluation.
    2009-09-04 16:18:56,020 INFO formulas.Evaluator (?:?) - Formula {@-suppressSection-13} evaluated to: null
    2009-09-04 16:18:56,024 INFO formulas.Evaluator (?:?) - Formula {@-suppressSection-11} evaluated to: b(true)
    2009-09-04 16:18:56,087 DEBUGaction.TestReportsAction (?:?) - exported report to stream
    2009-09-04 16:18:56,089 INFO reports.basicdataengine (?:?) - Invalidating controller
    2009-09-04 16:18:56,090 INFO reports.basicdataengine (?:?) - Invalidating controller
    2009-09-04 16:18:56,091 INFO reports.basicdataengine (?:?) - Invalidating controller
    2009-09-04 16:18:56,094 INFO reports.queryengine (?:?) - Connection logoffServer:
    2009-09-04 16:18:56,095 INFO driverImpl.jdbc (?:?) - Connection Closed null
    2009-09-04 16:18:56,185 DEBUGaction.TestReportsAction (?:?) - Successfully exported report to /webdata/html/reports/invoice_PGE_465.pdf
    Edited by: Mike Najera on Sep 5, 2009 12:13 AM

  • Importing Assets from ALER to export to OER is taking so long time

    Guys,
    I am importing Service Bus Projects from ALSB ant trying to export it to Oracle Enterprise Repository 11g. But when I try to export the configuration jar to Oracle Enterprise Repository using the Harvester tool it is taking so long time.
    I have more than 1500 WSDL and XSD (with some circular reference), it is running for 48 hours without finish it.
    What can I do to Harvester and Oracle ER be faster ?
    Do we have some configuration to turn off validation or something like this ?
    Thanks in advanced.
    Edited by: rrocha on 18/04/2011 14:11

    Welcome to the Apple Community.
    Itunes library or iTunes store.

  • FDM 11.1.2.1 taking too much time to Export (Multi-load)

    Hi...
    There is something very odd here...
    I'm using FDQM 11.1.2.1 to multi-load a file. (300mb)
    So, the steps: Import (about 5min), Validade (about 5 min) - I have some explicit maps (about 2300).
    Export is my problem.... its taking more than 1 hour.
    I made a test using a Essbase rule to load this data using SQL with Join on FDM mapping table - TDATAMAP (because of the maps). So, in this way i spend about 3 minutes to load everything...
    The question is: why using FDM web client is taking to much time to Export data?
    Target: Planning App
    Thanks in advance
    -Cheers,
    Rafael

    Yes, you are exactly right. That is the correct order for the outline. Sorry, it's been a long time since I've used Essbase and my memory isn't what it used to be. Still, what keeps data load times to a minimum is to have the sparse fields in sort order so that the same data blocks don't have to be pulled into memory multiple times. The sparse fields should be sorted and in the same order as the outline. The best way to ensure that this is occurring is to have the sparse fields first, dense fields last.
    With dense dimensions sorted 1st, you can see how the sparse fields can get out of sort order, causing blocks to be revisited:
    D1A -> D2A -> S1A -> S2A -> S3A -> Data
    D1A -> D2A -> S1A -> S2A -> S3B -> Data
    D1B -> D2A -> S1A -> S2A -> S3A -> Data
    The 3rd record forces Essbase to re-visit the datablock that was already loaded by the 1st record.
    On the other hand, if sparse fields are first and sorted left to right, the first two records are loaded at the same time to the same block:
    S1A -> S2A -> S3A -> D1A -> D2A -> Data
    S1A -> S2A -> S3A -> D1B -> D2A -> Data
    S1A -> S2A -> S3B -> D1A -> D2A -> Data
    Sorry for the confusion. Hope this clarifies field sort order vs. outline order.

  • Non-self-contained movie huge & takes a very very long time to export

    This is for FCE 4.
    I have a 1.5 hour movie, and when I export it to quicktime (Not quicktime conversion) it takes WAY too long.
    I have "make self-contained movie" unchecked, so I thought creating a reference movie would be very quick. Why is it taking so long and why is the resulting file huge?
    My source files is DV, and only minor editing and effects were used. However the entire movie was cropped & resized. Is that why?

    Ah I think I figured it out.
    I did another test project with the same source DV file, but didn't do ANY editing to it.
    It took 2 minutes to export a 70mb reference mov file from a 7 minute DV sequence.
    Then I cropped & resized the sequence by changing parameters in the "motion" tab.
    Then it took 20 minutes to export a much larger reference mov file from the same 7 minute DV sequence.
    Then I clicked "render all" after selecting every render option in the drop-down menu.
    Then it took less than 1 minute to export a 70mb reference mov file from the same 7 minute DV sequence.
    So I guess I was having problems because I didn't render "everything", because I didn't select "FULL" on the drop-down menu for rendering. I thought only RED segments in the timeline needs to be rendered prior to export, but apparently the "FULL" render must be selected as well so that the entire timeline is nice and purple. (or some shade of blue)
    Also, a cropped and resized video is very large in size (GB) if it is exported as a reference file only without prior rendering. * Can someone else confirm that this is the normal behavior in FCE? *
    Oh well, lesson learned.. This movie is going to take a very long time to export, and result in a very large file. But I can't use it without cropping & resizing it!
    Message was edited by: Yongwon Lee

  • Discoverer report taking too long time to open.

    HI,
    Discovere reports are taking too long time to open. Please help to resolve this.
    Regards,
    Bhatia

    What is the Dicoverer and the Application release?
    Please refer to the following links (For both Discoverer 4i and 10g). Please note that some Discoverer 4i notes also apply to Discoverer 10g.
    Note: 362851.1 - Guidelines to setup the JVM in Apps Ebusiness Suite 11i and R12
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=362851.1
    Note: 68100.1 - Discoverer Performance When Running On Oracle Applications
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=68100.1
    Note: 465234.1 - Recommended Client Java Plug-in (JVM/JRE) For Discoverer Plus 10g (10.1.2)
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=465234.1
    Note: 329674.1 - Slow Performance When Opening Plus Workbooks from Oracle 11.5.10 Applications Home Page
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=329674.1
    Note: 190326.1 - Ideas for Improving Discoverer 4i Performance in an Applications 11i Environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=190326.1
    Note: 331435.1 - Slow Perfomance Using Disco 4.1 Admin/Desktop in Oracle Applications Mode EUL
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=331435.1
    Note: 217669.1 - Refreshing Folders and opening workbooks is slow in Apps 11i environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=217669.1

  • Taking too long time to get LOV

    HI,
    I have created a customer folder in which the query retuns 0.5 million records.
    I have created a item class in airline_name column which is being used in the worksheet as parameter.
    The problem is it is taking too long time near about 2 min to get LOV when the user wants to search the exact name.
    Thanks,
    Himanshu Tiwari

    Hi,
    Usually, you should not use the folder that the report is based on to define the LOV. You should use a separate folder to define the LOV that is optimised to return the content of the LOV.
    Rod West

  • Discoverer reports taking a long time!!!

    Hi all,
    One of our clients is complaining that the discoverer reports are taking a long time to run for the last few days, the report used to take 30 minutes before but now is running for hours!!
    I have checked the SGA and I have killed the idle sessions but still there was no improvement in the performance.
    The version of BI discoverer is 10 and database also is 10g and the platform is win server 2003.
    I have checked the forums and they talk about explain plan and tkprof and other commands, but my problem is that i am unable to find the query that discoverer is running i mean once the report is clicked the query runs and gives the estimate time it would take. can some one tell me where this query is stored so that i can check this query,
    Also there were no changes made in the query or to the database.
    The temp space fills up 100%, i increased the size of temp space but still it goes to 100% also i noticed that the CPU utilisation goes to 100%
    i also increased the SGA but still no go.
    can someone kindly help me as to what could be causing this problem
    also kindly guide me to some good documents for tuning the discoverer.
    thanks in advance,
    regards,
    Edited by: user10243788 on Jan 4, 2010 12:47 AM

    Hi,
    The fact that the report used to work fast and now not can be related to many things but my guess is that the database statistics were changed and so the explain plan has changed.
    This can be done due to change in the volume of the data that crossed a level were oracle optimizer change the behavior but it can be other things as well.
    Anyway it is not relevant since it will be easier to tune the SQL than to find what have changed.
    In order to find whether the problem is with the discoverer or in the SQL extract the SQL as described above and run it in SQL tool (SQL Plus, TOAD, SQL Developer and so on).
    The best way to get to the problem is run a trace on your session and then use the TKPROF command to translate it to a text file you can analyze - you can assist your DBA team they should have no problem doing that.
    By doing that you will get the problematic statements/ functions/ procedures that the report uses.
    From there you can start working on improving the performance.
    Performance is expertise for itself so i'm sorry i don't know to tell you where to start from, I guess the start will be from understanding the meaning of the explain plan.
    Hope I helped a little although I wish Ii had a magic answer for you
    BTW, until you resolve that problem you can use the discoverer scheduler to run the reports in the background and so the users will get the data.
    Tamir

  • Time Machine Taking a Long Time to Index / Back Up

    Hi everyone.
    I noticed this on the latest backup to my Time Capsule device.  I'm backing up over a WiFi connection to a time capsule device.  It is taking a long time to index the backup.  Here's the code from the latest backup.
    10/17/12 9:43:03.799 PM com.apple.backupd: Starting standard backup
    10/17/12 9:43:04.091 PM com.apple.backupd: Attempting to mount network destination URL: afp://Michael%20Payne@Extreme%20Pleasantville._afpovertcp._tcp.local/Time%20Mac hine%20Backups
    10/17/12 9:43:12.657 PM com.apple.backupd: Mounted network destination at mountpoint: /Volumes/Time Machine Backups using URL: afp://Michael%20Payne@Extreme%20Pleasantville._afpovertcp._tcp.local/Time%20Mac hine%20Backups
    10/17/12 9:43:34.015 PM com.apple.backupd: QUICKCHECK ONLY; FILESYSTEM CLEAN
    10/17/12 9:43:37.440 PM com.apple.backupd: Disk image /Volumes/Time Machine Backups/Michael’s MacBook Pro.sparsebundle mounted at: /Volumes/Time Machine Backups 1
    10/17/12 9:43:37.459 PM com.apple.backupd: Backing up to: /Volumes/Time Machine Backups 1/Backups.backupdb
    10/17/12 9:45:23.296 PM com.apple.backupd: 500.1 MB required (including padding), 3.27 GB available
    10/17/12 9:46:48.844 PM com.apple.backupd: Copied 1436 files (3.4 MB) from volume Macintosh HD.
    10/17/12 9:46:51.897 PM com.apple.backupd: 448.1 MB required (including padding), 3.27 GB available
    10/17/12 9:46:51.898 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:47:53.681 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:48:55.642 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:49:56.511 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:50:56.564 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:51:57.643 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:52:57.903 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:53:59.341 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:55:01.281 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:56:03.171 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:57:03.966 PM com.apple.backupd: Waiting for index to be ready (100)
    10/17/12 9:58:33.083 PM com.apple.backupd: Copied 824 files (1.2 MB) from volume Macintosh HD.
    10/17/12 9:58:40.359 PM com.apple.backupd: Starting post-backup thinning
    10/17/12 9:59:51.851 PM com.apple.backupd: Deleted /Volumes/Time Machine Backups 1/Backups.backupdb/Michael’s MacBook Pro/2012-10-16-213620 (5.6 MB)
    10/17/12 10:00:17.440 PM com.apple.backupd: Deleted /Volumes/Time Machine Backups 1/Backups.backupdb/Michael’s MacBook Pro/2012-10-16-113505 (1.6 MB)
    10/17/12 10:00:30.046 PM com.apple.backupd: Deleted /Volumes/Time Machine Backups 1/Backups.backupdb/Michael’s MacBook Pro/2012-10-16-103509 (1.6 MB)
    10/17/12 10:00:43.781 PM com.apple.backupd: Deleted /Volumes/Time Machine Backups 1/Backups.backupdb/Michael’s MacBook Pro/2012-10-16-093513 (1.3 MB)
    10/17/12 10:00:43.781 PM com.apple.backupd: Post-back up thinning complete: 4 expired backups removed
    10/17/12 10:00:45.207 PM com.apple.backupd: Backup completed successfully.
    10/17/12 10:00:56.988 PM com.apple.backupd: Ejected Time Machine disk image.
    10/17/12 10:00:57.550 PM com.apple.backupd: Ejected Time Machine network volume.
    I am just curious as to why this is, and what I might do to fix it.  I did see one suggestion in another website suggesting diabling spolight indexing of the backup image with sudo mdutil -i off /Volumes/Time Machine Backups, but don't know where to type that in.  It then states to reenable it, but don't know where to type that in, either.  Any advice would be helpful.  I did run a verification not too long ago, and here's that code.
    10/14/12 3:44:14.289 PM com.apple.backupd: Backup verification requested by user.
    10/14/12 3:44:41.503 PM com.apple.backupd: Running backup verification
    10/14/12 4:04:31.602 PM com.apple.backupd: Backup verification passed!
    If any of you have or had a similar problem, I'd appreciate your help.  For the time being, I may just disable time machine during the evening, and only have it run during the day when I'm at work.

    Is your sig still current (10.7.3)? Or did you update to 10.7.5? Because 10.7.5 had a Spotlight/Time Machine bug that slowed down a lot of backups. If that's what it is, there's now a 10.7.5 Supplemental Update that fixes the slow Time Machine problem.
    If it isn't 10.7.5, I don't know what the problem might be.

  • My external library 100 GB on a 2TB drive is taking a long time to open.

    Hey there,
    I recently moved my library to an external drive.  When I tried to load the external drive, it said it needed to repair and I let it happen overnight because it was taking a long time.
    Now it seems to just freeze when I try to load the external library.  Can someone let me know if there is something I can do to make this process speed up?  Should I move some of my projects to another library and keep it off-line?
    Thanks,
    Austin

    Could it possibly be due to videos stored in the aperture library?  I think one of the folders has multiple GB of video...  Just a thought.
    If Aperture is freezing on opening, this can be caused by corrupted or incompatible media, but it would not explain the original error message your reported - "permissions".
    When the repair finishes, try to launch Aperture with the Shift-key held down. This will defer the generation of previews. If media cannot beprocessed, Aperture will hang, when it tries to generate the previews for those images or videos. If you can get Aperture to launch after deferring the preview generation, you could open the Activity window (Window > Show Activity) and then selectively try to generate previews for your last imported images and videos  (Photos > update preview). The Activity window will show, what is currently processed, and perhaps you can spot, which of the files is causing the hang.

  • My email all of sudden is taking a long time to open

    Since last week my msn.com email is taking a long time to open I haven't changed a thing and I have had this email since 1998

    How large is your hard drive and how much hard drive space do you have left?
    Which Macbook Pro do you have?
    Which os & version are you using?

  • My mac pro is taking a long time to open files

    When I rey and open applications like Finder, it is taking a long time to open or it hangs how can I fix this
    I have OS10.7.2 and 300GB left out of 500GB.
    Mac book peo 17
    Message was edited by: waldoamt

    How large is your hard drive and how much hard drive space do you have left?
    Which Macbook Pro do you have?
    Which os & version are you using?

  • Hyperion Report taking a long time to process local query

    Hi All,
    I am trying to run a report on Hyperion IR 9.3.1. I am facing a performance issue with this report. I am joining 13 tables using full outer join. Each table is having data about 900 rows and the final output i am getting from the local query is about 11000 rows. This local query is taking a long time to get process about 3 - 5 minutes. I suppose it should run with in 30 sec as number of rows are very few. Can anyone tell me what is the problem with this local query and how the performance of the report can be increased?
    Thanks in advance.
    Regards
    Ujjawal

    Be aware that XP takes approx 1gb of your RAM leaving you with 1gb for whatever else is running. MS Outlook is also a memory hog.
    To check Virtual Memory Settings:
    Control Panel -> System
    System Properties -> Advanced Tab -> Performance Settings
    Performance Options -> Adavanced Tab - Virtual Memory section
    Virtual Memory -
    what are
    * Initial Size
    * Maximum Size
    In a presentation at one of the Hyperion conferences years ago, Mark Ostroff suggested that the initial be set to the same as Max. (Max is typically 2x physical RAM)
    These changes may provide some improvement.

Maybe you are looking for

  • Installing a windows 7 (64 bit) printer driver to existing windows xp network

    Dear-Users......... i have connected 5 computers through a switch and ( hp laserjet 2014) printer is directly connected to one of them and shared on the network.  3 computers running on windows xp and one i added last time it was windows 7 (32 bit) i

  • Once per session, connected anyway

    I have an authentication scheme that is set to once per session. The users are defined in oid and have to logon using sso. So far so good. But in following scenario it goes wrong. 1. User enters the htmldb url 2. User gets the logon screen from sso 3

  • Purchase Orders Forms

    Hi folks Is there a manner to use the same window call by "Purchase Orders" function PO_POXPOEPO in read-only? Otherwise, is there a function other then PO_POXPOEPO which can call the "Manage Tax: PO" function PO_MANAGE_TAX? Tx

  • JDeveloper 10.1.3 embeded OC4J can't host Apache CXF Service.

    I use JDeveloper 10.1.3 to created a test web application, which uses Spring and Apache CXF 2.0.3 hosting a sample web service, but when the application start the following error occurs: [Starting OC4J using the following ports: HTTP=8988, RMI=23891,

  • Validation "Error Message" problem

    In APEX 4.2 i have under "Page Processing"-->"Validations" my valdidation rule created with name "10 validate report". My APEX Validation attributes are: Type="Function Returning boolean" Validation Expression 1 = "return rpt_utl_pkg.validate('my_db_