Features missing for working with large spreadsheets

Have worked with some large spreadsheets,
mostly at work, and numbers is missing some key elements.
or tell me if I am wrong
In no particular order:
1) there is no stop automatic recalculation.
With a large spreadsheet you want to turn autocalculate off
and be able to push a button when you want it to calculate.
2) no split window.
you have to be able to split the window
have the first row on the top split, the last row on the bottom split
add the formula to top cell in column, select cell,
select bottom cell in column (in lower split)
fill down and hit calculate.
I am sure there are others but these are biggies for me.

_Go to "Provide Numbers Feedback" in the "Numbers" menu_, describe what you wish.
Then, cross your fingers, and wait _at least_ for iWork'09
Yvan KOENIG (from FRANCE samedi 19 juillet 2008 16:58:13)

Similar Messages

  • Best practices for working with large placed bitmap images?

    Hey all,
    I need some advice on the best way to approach building these files. I've been working on some banners that are very large: 3 x 7 feet.
    Each banner has a simple vector graphic treatment at the top and bottom (rectangle with a different colored rule on top, and vector logo) and a small amount of text, just a URL and a headline.The headline is type (not converted to outlines) and usually has some other effect applied to it, say a drop shadow or outer glow. Under these graphics is a full bleed image. The placed images need to be 150ppi at actual size, so they're honking big, sometimes up to 2GB. Once the layouts are approved, they have to go to a vendor for output.
    The Illustrator docs are really large, and I've read in other threads how to combat that (PDF compatibility, raster settings). But even still, does anyone have any insight into the best way to deal with these things? The dimensions are large, and then the images are large, and it just makes for lots of looking at the spinning ball of death...
    If it were me, I'd build them in InDe, but the vector graphics need to be edited for each one, and I so don't like to do that in InDe unless forced. To me, it's still ultimately a page layout app, not a drawing app. (Old school here.)
    FYI, our machines are all MBPs with 8G ram and the latest Intel Core 2 Duo chips, 2.66 and 2.8GHz. If we keep the files local (as opposed to working on the server) it should be fairly zippy... No?
    Any advice is appreciated, thanks!

    You can get into memory trouble with very large placed pdf files. Tiffs too.
    This has to do with the preview, which contains much more information than you need for working with.
    On the other hand if you place EPSs and take care not to turn on overprint preview you can get away with huge files.
    If you do turn on overprint preview your machine will slow down a lot and the file may become totally unmanageable.
    Compare this with to InDesign where you can control the quality of the preview. A hi-res preview will slow you down and most often you don't need it anyway.
    I was working (in Illie) the other day on much larger files than you mention – displays for whole walls – and had some considerable trouble until I reverted to the old EPS format. They say it's dying but it ain't dead yet .

  • Software available for working with large video files?

    Hello,
    I'm working in PP CS6. I was wondering if there are any workarounds or 3rd party plugins/software that
    make working with really large video files easier and faster?
    Thanks.
    Mark

    Hi Jeff,
    Thanks for helping. This is the first time I shot video with my Nikon D5200. It was only a 3 minute test clip
    set at the highest resolution, 1920x1080-60i. I saw the red line above the clip in PP CS6 and hit the enter
    key to render the clip.
    It took almost 18 minutes or so to render the clip. This is probably normal but I was wondering if there is
    a way to reduce the file size so it doesn't take quite as long to render. I just remember a few years back
    that when the Red camera was out, guys were working with really huge files and there was a program
    from Cine something that they used to reduce the file size and make it more manageable when editing.
    I could be mistaken. I've been out of the editing look for a few years and just getting back into it.
    Thanks.
    Mark
    Here's my PC's components list you asked for:
    VisionDAW 4U 8-Core Xeon Workstation
      2 Intel QUAD-Core Xeon 5365-3.0GHz, 8MB, 1333MHz Processors
      16GB 667MHz Fully Buffered Server Memory Modules (2x2GB)
      Microsoft® Windows® Windows 7 Ultimate (x64)
      WDC 250GB, Ultra ATA100, 7200 rpm, 8MG Buffer Main OS HD
      2 WWDC 750GB, SATA II, 7200 RPM, 16MB Buffer HD (Raid 0)
      2 WDC 750GB, SATA II, 7200 rpm, 16MG Buffer HD (Samples)
      2 WDC 1TB Enterprise Class, SATA II, 7200 RPM, 32MB Buffer Hard Drive
      MOTU 24 I/O (main) / MOTU 2408mk3 (slave)
      Plexor PX-800A 18X Dbl. Layer DVD+/-RW Optical Drive
      Buffalo BuRay Drive (External) BR-816SU2 
      Front Panel USB Acess
      Integrated FireWire (1394a) interface
      Thermaltake Toughpower 850W Power Supply
      3xUAD1 Universal Audio Cards
      NVIDIA QUADRO FX 1800 / Memory 768 MB GDDR3
      CUDA Parallel Processor Cores / 64
      Maximum Display Resolution Digital @60Hz = 2560x1600
      Memory Interface 192Bit
      Memory Bandwidth (GB/sec) / 38.4 GB/sec
      PCI-Express, DUAL-Link DVI 1
      Digital Outputs 3 (2 out of 3 active at a time)
      Dual 25.5" Samsung 2693HM LCD HD Monitors

  • Advice for working with large clip files

    A few years ago I made a movie using iMovie2. At the time I was working with clips recoded on one minidv disc. I am now ready to begin a project (just bought iLife06) that is much larger. I have roughly ten 60 min minidv discs. I am getting nervous about the amount of memory needed to import and work with this many clips.
    Is it possible to import the clips to my external fire wire HD? Can I still work in iMovie while the project lives on the ext firewire drive? Can anyone tell me roughly how much memory is needed for larger projects like this? (I expect the final project will be 30-40 minutes).
    Since the footage all comes from our European concert tour, it easily divides into 3 separate sections - would it be easier to create 3 different iMovie projects? Is it possible to combine them in the end to create one film?
    Thank you so much for your help with this.
    Sincerely,
    Bob Linscott

    Is it possible to import the clips to my external
    fire wire HD? Can I still work in iMovie while the
    project lives on the ext firewire drive?
    Should be fine. I've been editing 4 hours worth of footage down to about 50 minutes, on a project stored on a FireWire 800 drive. FireWire 400 (if that's what you're using) is half the speed, but should still be fine, I think.
    So, create a new project, save it to your external drive, and then import your footage into the project.
    Can anyone
    tell me roughly how much memory is needed for larger
    projects like this? (I expect the final project will
    be 30-40 minutes).
    If you mean hard drive space for the project, I think you're talking about 11 GB per hour: so, in short, a lot. I think you could import your footage, edit a bit, and empty your trash to free up space, but iMovie does seem to keep your originals fairly relentlessly, so I'm not entirely sure.
    Watch out for extracting audio from clips too: I've been doing this in order to do cross-fades, but it increases the size of your movie. I ended up buying a second, bus-powered 80 GB external hard drive just to hold the iMovie project
    Since the footage all comes from our European concert
    tour, it easily divides into 3 separate sections -
    would it be easier to create 3 different iMovie
    projects? Is it possible to combine them in the end
    to create one film?
    This could well be a good idea. Once your projects gets over a certain length, you may experience the "herky jerky" pkayback issue (search for "herky jerky" on these forums), making editing difficult. So three separate projects might be easier.
    To combine them all at the end, you'll want to export the 2nd and 3rd projects as full quality DV (Share > QuickTime > Compress Movie For: Full Quality), then import them (File > Import) into your first project, and add them to the timeline.

  • Working with large spreadsheets

    I have recently started using Numbers for my spreadsheets after years of using Excel and have a query.
    If you need to input data in a cell which is a long way down the spreadsheet or far off to the right the column or row titles can be lost off the screen. In excel I could drag a bar across or down to lock the titles in position so I could still see if I was inputting the information in the correct cell.
    Does Numbers have this functionality?
    Thanks
    Rob

    _Go to "Provide Numbers Feedback" in the "Numbers" menu_, describe what you wish.
    Then, cross your fingers, and wait _at least_ for iWork'09
    Yvan KOENIG (from FRANCE samedi 19 juillet 2008 16:58:13)

  • Mac Pro RAID block size recommendations for working with audio in Logic Pro

    I have recently ordered a Mac Pro and plan to do a RAID configuration across 3 HDD's
    The RAID type i am going to do is a RAID 0 striped.
    The computer is going to be used primarily for audio post production and working with 20+ 24-Bit audio files at any one time within a Logic project.
    I want to know what is the best block size i should use when configuring the RAID.
    I understand that using a higher block size is best for working with large files but do i need to do this in my case or will the default 32k block size be enough?
    Thanks in advance

    Use 64k. Things like databases like having 32k blocks because of all the small files. Audio files are pretty small even at 24-bit 192KHz. Go to 128k if all you are doing is streaming and no samples. But 20+ 24-bit is really not too large anyway considering most modern HDD's can stream 100MB/s off one spindle. You'll probably be fine regardless of the block size you choose. But most audio pro's choose 64k.

  • Speed up Illustrator CC when working with large vector files

    Raster (mainly) files up to 350 Mb. run fast in Illustrator CC, while vector files of 10 Mb. are a pain in the *blieb* (eg. zooming & panning). When reading the file it seems to freeze around 95 % for a few minutes. Memory usage goes up to 6 Gb. Processor usage 30 - 50 %.
    Are there ways to speed things up while working with large vector files in Illustrator CC?
    System:
    64 bit Windows 7 enterprise
    Memory: 16 Gb
    Processor: Intel Xeon 3,7 GHz (8 threads)
    Graphics: nVidia Geforce K4000

    Files with large amounts vector points will put a strain on the fastest of computers. But any type of speed increase we can get you can save you lots of time.
    Delete any unwanted stray points using  Select >> Object >> stray points
    Optimize performance | Windows
    Did you draw this yourself, is the file as clean as can be? Are there any repeated paths underneath your art which do not need to be there from live tracing or stock art sites?
    Check the control panel >> programs and features and sort by installed recently and uninstall anything suspicious.
    Sorry there will be no short or single answer to this, as per the previous poster using layers effectively, and working in outline mode when possible might the best you can do.

  • Tips for working with USB floppy drives

    I was just wondering if anyone had any tips on working with USB floppy drives in Archlinux.
    One thing I've noticed is that trying to use mkfs.msdos on my USB floppy drive directly doesn't seem to result in a usable disk (it's mountable, but reports that I don't have enough space available to copy files), but if I use the command on an image file the same size as a floppy disk and then use dd to copy the blank image to my USB drive I get a disk that's perfectly usable.  Anybody know what the deal is with that or how I might be able to successfully format a disk in fewer steps (other than keeping the blank images around and just using dd to format the floppies, haha)?
    Also, I'd like to hear any other tips or tricks any of you have for working with USB floppy drives.  They just seem so much more usable under Windows as opposed to Linux, and that disappoints me a little.

    jmetal88 wrote:Okay, thanks guys.  I haven't figured out the issue yet, but you're helping me narrow it down.  I decided to run a df -h command after each try of formatting the disk this time.  I tried the low-level ufiformat command followed by a mkfs.msdos on the device, using a 720k disk, and df -h showed 700k free on the file system.  It's just not making the file system large enough (I believe the file I was trying to copy over was 710k or so).  When I run mkfs.msdos on a 720k disk image and then dd the image over to the drive, df -h shows 713k free after mounting the drive.  Any idea why mkfs.msdos isn't filling out the full 720k on the disk itself, but is on the image?
    I would assume this happens because (as far as I can tell from skimming through the source over the course of a few minutes) mkfs.fat only applies floppy size heuristics to fd and lo devices -- your USB floppy drive is treated as a hard drive, and that leads to less-than-ideal default parameters. Run
    fsck.fat -v
    on the more spacious filesystem to figure out what options need to be set when you attempt formatting an actual disk. Or just make things easier by keeping newly-formatted blank disk images on hand so that dd and (maybe) ufiformat are all you need to worry about.

  • Working with large Artboards/Files in Illustrator

    Hello all!
    I'm currently designing a full size film poster for a client. The dimensions of the poster are 27" x 40" (industry standard film poster).
    I am a little uncertain in working with large files in Illustrator, so I have several problems that have come up in several design projects using similar large formats.
    The file size is MASSIVE. This poster uses several large, high-res images that I've embedded. I didn't want them to pixelate, so I made sure they were high quality. After embedding all these images, along with the vector graphics, the entire .ai file is 500MB. How can I reduce this file size? Can I do something with the images to make the .ai file smaller?
    I made my artboard 27" x 40" - the final size of the poster. Is this standard practice? Or when designing for a large print format, are you supposed to make a smaller, more manageable artboard size, and then scale up after to avoid these massive file sizes?
    I need to upload my support files for the project, including .ai and .eps - so it won't work if they're 500MB. This would be good info to understand for all projects I think.
    Any help with this would be appreciated. I can't seem to find any coherent information pertaining to this problem that seems to address my particular issues. Thank you very much!
    Asher

    Hi Asher,
    It's probably those high-res images you've embedded. Firstly, be sure your images are only as large as you need them Secondly, a solution would be to use linked images while you're working instead of embedding them into the file.
    Here is a link to a forum with a lot of great discussion about this issue, to get you started: http://www.cartotalk.com/lofiversion/index.php?t126.html
    And another: http://www.graphicdesignforum.com/forum/archive/index.php/t-1907.html
    Here is a great list of tips that someone in the above forum gave:
    -Properly scale files.  Do not take a 6x6' file then use the scaling tool to make it a 2x2'  Instead scale it in photoshop to 2x2 and reimport it.  Make a rule like over 20%, bring it back in to photoshop for rescaling.
    -Check resolutions. 600dpi is not going to be necessary for such and such printer.
    -Delete unused art.  Sloppy artists may leave old unused images under another image.  The old one is not being used but it still takes up space, therefore unecessarily inflating your file.
    -Choose to link instead of embedd.  This is your choice.  Either way you still have to send a large file but many times linking is less total MB then embedding.  Also embedding works well with duplicated images. That way multiple uses link to one original, whereas embedding would make copies.
    -When you are done, using compression software like ZIP or SIT (stuffit)
    http://www.maczipit.com/
    Compression can reduce file sizes alot, depending on the files.
    This business deals with alot of large files.  Generally people use FTP's to send large files, or plain old CD.  Another option is using segmented compression.  Something like winRAR/macRAR or dropsegment (peice of stuffit deluxe) compresses files, then breaks it up into smaller manageble pieces.   This way you can break up a 50mb file into say 10x 5mb pieces and send them 5mb at a time. 
    http://www.rarlab.com/download.htm
    *make sure your client knows how to uncompress those files.  You may want to link them to the site to download the software."
    Good luck!

  • What is better Mac Mini or Mac Pro and 4K 40" screen or iMac 5K for work with Exel

    I am working on Macbook pro retina for around 16 hours a day,  it is ok but the screen is to small. I am opening about 25 - 30 EXCEL windows, 20 - 25 safari windows at the same time. Excel windows are small, they only show 40 rows and S column   Now I am considering buying  iMac 5k ( but the screen is only 27", however a very high resolution but I think that fonts and digits displayed may be too small ). I am also considering Mac Mini or Pro with 40" 4K TV ( 4k computer monitors are 27 - 31" , so better idea would be purchasing iMac 5k ) . The problem is that 40 - 42" TV don't have Display Port ( I can't find  it anywhere )
    Please help me to choose what would suit my needs better. I need large screen to work comfortably with an Excel sheets.  What is better? Mac Mini or Mac Pro and 4K 40" screen or iMac 5K for work with Excel ?. The other problem I have is that when I spend  16 hours at work with retina macbook , my eyes are ok. Using the  30'' Dell screen 2560/1440 my eyes are tired after 2 hours.
    Thanks for your help on advance
    Best regards
    David

    Currently you can only get 4 K at 30 Hz via HDMI. Some users have reported jerky curser movement because of the 30 Hz.
    I would go for the 5K iMac. You can add a second screen at 4K.
    FaceTime HD camera
    Simultaneously supports full native resolution on the built-in display and up to 3840 by 2160 pixels on an external display
    Support for extended desktop and video mirroring modes
    The Mac Mini supports:
    Support for up to two displays at 2560 by 1600 pixels, both at millions of colors
    Thunderbolt digital video output
    Native Mini DisplayPort output
    DVI output using Mini DisplayPort to DVI Adapter (sold separately)
    VGA output using Mini DisplayPort to VGA Adapter (sold separately)
    Dual-link DVI output using Mini DisplayPort to Dual-Link DVI Adapter (sold separately)
    HDMI video output
    Support for 1080p resolution at up to 60Hz
    Support for 3840-by-2160 resolution at 30Hz
    Support for 4096-by-2160 resolution at 24Hz
    The 2013 Mac Pro can use three and more displays:
    Mac Pro (Late 2013): Using multiple displays - Apple Support
    I would go for multiple screens

  • Working with Large Numbers

    Hi there,
    I am currently doing a school assignment and not looking for answers but just a little guidance.
    I am working with large numbers and the modulo operator.
    I might have some numbers such as :
    int n = 221;
    int e = 5;
    int d = 77;
    int message = 84;
    int en = (int) (Math.pow(message, e) % n);
    int dn = (int) (Math.pow(en, d) % n);Would there be a better way to do this kind of calculation. The dn value should come out the same as message. But I always get something different and I think I might be losing something in the fact that an int can only hold smaller values.

    EJP wrote:
    It might make sense in some contexts to have a positive and negative infinity.
    Yes, perhaps that's a better name. Guess I was harking back to old COBOL days :-).(*)
    But the reason these things exist in FP is because the hardware can actually deliver them. That rationale doesn't apply to BIgInteger.Actually, it does. All I'm talking about is a value that compares higher or lower than any other. That could be done either by a special internal sign value (my slight preference) or by simply adding code to compareTo(), equals() and hashCode() methods that takes the two constants into account (as they already do with ZERO and ONE).
    Don't worry, I'm not holding my breath; but I have come across a few situations in which values like that would have been useful.
    Winston
    Edited by: YoungWinston on Mar 22, 2011 9:07 AM
    (*) Actually, '±infinity' tends to suggest a valid arithmetic value, and I wasn't thinking of changing existing BigInteger/BigDecimal maths (except perhaps to throw an exception if either value is involved).

  • Working with Large List in sharepoint 2010

    Hi All
    I have a list with almost 10k records in my sharepoint list and based on some business requirement i am binding (almost 6k records) the data to asp.net grid view and this will visible on the home page of the portal where most of the users will access. Can
    someone please guide the best method to reduce the performance inorder the program to hit the SP list everytime the page loads...
    Thanks & Regards
    Rakesh Kumar

    Hi,
    If you are Working with large data retrieval from the content database (SharePoint list), the points below for your reference:
    1. Limit the number of returned items.
    SPQuery query = new SPQuery();
    query.RowLimit =6000; // we want to retrieve2000 items
    query.ListItemCollectionPosition = prevItems.ListItemCollectionPosition; // starting at a previous position
    SPListItemCollection items = SPContext.Current.List.GetItems(query);
    2. Limit the number of returned columns.
    SPQuery query = new SPQuery();
    query.ViewFields = "";
    3. Query specific items using CAML (Collaborative Markup Language).
    SPQuery query = new SPQuery();
    query.Query = "15";
    4.Use ContentIterator class
    https://spcounselor-public.sharepoint.com/Blog/Post/2/Querying-a--big-list--with-ContentIterator-and-multiple-filters
    5. Create a Stored Procedure in Database to get the special data, create a web service to get the data, when create a web part to show the data in home page.
    Best Regards
    Dennis Guo
    TechNet Community Support

  • Is anyone working with large datasets ( 200M) in LabVIEW?

    I am working with external Bioinformatics databasesa and find the datasets to be quite large (2 files easily come out at 50M or more). Is anyone working with large datasets like these? What is your experience with performance?

    Colby, it all depends on how much memory you have in your system. You could be okay doing all that with 1GB of memory, but you still have to take care to not make copies of your data in your program. That said, I would not be surprised if your code could be written so that it would work on a machine with much less ram by using efficient algorithms. I am not a statistician, but I know that the averages & standard deviations can be calculated using a few bytes (even on arbitrary length data sets). Can't the ANOVA be performed using the standard deviations and means (and other information like the degrees of freedom, etc.)? Potentially, you could calculate all the various bits that are necessary and do the F-test with that information, and not need to ever have the entire data set in memory at one time. The tricky part for your application may be getting the desired data at the necessary times from all those different sources. I am usually working with files on disk where I grab x samples at a time, perform the statistics, dump the samples and get the next set, repeat as necessary. I can calculate the average of an arbitrary length data set easily by only loading one sample at a time from disk (it's still more efficient to work in small batches because the disk I/O overhead builds up).
    Let me use the calculation of the mean as an example (hopefully the notation makes sense): see the jpg. What this means in plain english is that the mean can be calculated solely as a function of the current data point, the previous mean, and the sample number. For instance, given the data set [1 2 3 4 5], sum it, and divide by 5, you get 3. Or take it a point at a time: the average of [1]=1, [2+1*1]/2=1.5, [3+1.5*2]/3=2, [4+2*3]/4=2.5, [5+2.5*4]/5=3. This second method required far more multiplications and divisions, but it only ever required remembering the previous mean and the sample number, in addition to the new data point. Using this technique, I can find the average of gigs of data without ever needing more than three doubles and an int32 in memory. A similar derivation can be done for the variance, but it's easier to look it up (I can provide it if you have trouble finding it). Also, I think this funtionality is built into the LabVIEW pt by pt statistics functions.
    I think you can probably get the data you need from those db's through some carefully crafted queries, but it's hard to say more without knowing a lot more about your application.
    Hope this helps!
    Chris
    Attachments:
    Mean Derivation.JPG ‏20 KB

  • How does the sync functionality work with large libraries on small devices?

    How does the sync functionality work with large libraries?
    Say I sync 100gb of photos with the new Photos app and turn on sync on a 16gb iphone. Will it fill the device up to 16gb? Can I tell it to limit to xgb so I leave room for music and apps? How does this work? Will it slow down my phone if its trying to sync 100gb across smaller devices?

    "Will the Apple TV now read directly from the Time Capsule?" ATV does not 'read' from the TC. It connects to the Mac and itunes library associated with the Mac. Of course the itunes program can 'point' to a library on the TC. You'll still need to have itunes open and the Mac powered on and not sleeping with the TC mounted to the Desktop to use ATV properly. So bottom line the use of TC just adds one more step to view files on the ATV.

  • JK Adobe TV - Top 5 Tips for Working with Vectors in CC

    Julieanne Kost has just blogged about her  Adobe TV video on working with shapes and paths in Photoshop CC.  It's actually not that new to Adobe TV, and has already had a lot of views, but we get a lot of questions here on the subject with CC, and there are some nice little tips in it.  I certainly learned a couple of things. :-)
    http://blogs.adobe.com/jkost/2014/01/top-5-tips-for-working-with-vectors-in-photoshop-cc.h tml
    http://tv.adobe.com/watch/the-complete-picture-with-julieanne-kost/top-5-tips-for-working- with-vectors-in-photoshop-cc/

    My apologies, but I really had no interest in a member's "answer", especially one that is so unhelpful.  Assuming that you were responding to me (we are the only two commenter's at this point), I would not be inclined to read the Creative Cloud offers since this is something that I am not interested in.  I bought the product the first day of offer, I did not rent it....just like I have in all the years past.

Maybe you are looking for

  • Problem creating LOB segment

    Hi all, I'm working with 11gR2 version in RAC environment with 2 nodes. I'm creating the schemes on my new database with empty tablespaces and 100Mb of initial free space. I'm having an error on the creation of this table: CREATE TABLE "IODBODB1"."DO

  • Content Conversion in Mail Adapter

    Hi folks, I would like to know whether it is possible to do content conversion when using a mail adapter. The scenario is : We will receive file as an attachment in a mail. We have to read the file and do the content conversion. Then we have to map t

  • Talent Management Reports

    We are implementing SSHR Talent Management for our client. Client using Global HRMS. I have read the doc of Talent management and found some workbook reports such as Appraisal Detail Workbook and Competence Updates after Appraisal Workbook. I think w

  • Installed ODAC 11g ver Release 3 (11.2.0.2.1) - no tnsping utility

    Hello Folks: Env: WIN XP PRO SP3, 32-bit Downloaded and installed ODAC 11.2.0.2.1 but the tnsping.exe utility is no wheres to be found - any ideas? Thanks!

  • Tolrence Limit at Service activity level similar to Article.

    Dear All, I am in edge of implementing the External service Process in our work place. Process will be as follow; Service Activity will be created. Pricing Conditions PO creation with Account Assignment "F" & Item category "D". And then normal servic