Newstand file sizes are huge!

The current issue of GQ is 580mb.  That's more than half a gig in size!  Several months of magazine subscriptions and my iPad is full, so how do I remove these issues onto my Mac?

The Martha Stewart issues of Everyday Food are about 380 MB each too.  Whoppers, yes, but they include embedded video and other enhancements. 
I think you can Archive the issues.  Martha has a button next to each issue that says Archive.

Similar Messages

  • Adobe Illustrator CS5 File sizes are huge!

    Try this:
    open a new Illustrator document
    create a simple box, black stroke, white fill
    save the file
    you're looking at something like 1 Mb for these 4 anchor points and the fill and stroke definition!
    now, back in Illustrator, downsave to Illustrator 3 format
    file size is in around 6Kb - much better!
    Okay, I know we can disable ICC profile embedding, PDF compatibility etc, but none of that even comes remotely close to the "correct" file size of < 10Kb.
    C'mon Adobe, what are you putting in that CS5 file that's so darned important? This isn't a complex illustration with gradient masks, 3D transforms and thousands of points.
    I get it, everyone has terrabyte drives and high-speed internet. Is that an excuse to make really inefficient file formats?

    hmmmm......
    As Monika pointed out... deleting the defualt library items will make files sizes drastically smaller.
    I'd guess that version 3, since it lacked the ability to work in preview mode, didn't save file previews with the file either... that could easily account for 50kb.

  • When I narrate my keynote presentation, why is the file size so huge?

    When I narrate my keynote presentation, why is the file size so huge?

    Here are the steps so you can see all the files that make up your Keynote…
    Go to the Finder and find your Keynote file and make sure that it has the .key extension on it. If it doesn't, click it and command i to view info and uncheck the box that says Hide Extension.
    Delete the .key and change to .zip and then click the use .zip in the dialog box that comes up
    Double click the new .zip file and it will change it to a folder
    Open the folder there are the files
    Close the folder and add .key to the end of the name and click Add in the dialog box (this will change it back to a Keynote file)
    You might be able to do this and convert the voice narration to a smaller format.

  • I have a few hundred duplicates in my iPhoto library, but the file sizes are different.  So one is 1.3mb and one is 567kb.  I want to delete the smaller ones, but short of comparing each duplicate, is there a way to do this?

    I have a few hundred duplicates in my iPhoto library, but the file sizes are different.  So one is 1.3mb and one is 567kb.  I want to delete the smaller ones, but short of comparing each duplicate, is there a way to do this?  I've been looking at Duplicate Annhilator but I don't think it can do it.
    Thanks!

    I just ran a test with iPhoto Library Manager, Duplicate Annihilator, iPhoto Duplicate Cleaner, Duplifinder and Photodedupo.  I imported a folder of 5 photos into a test library 3 times, allowing iPhoto to import duplicates.  I then ran the 5 photos thru resizer to reduce their jpeg compression but all other aspects of the file the same.
    None of the duplicate removal apps found set that was reduced in the file resizer. That's probably due to the fact that the file creation date was being used as a criteria and the resized photo would have a different file creation date even though the Image Capture date was the same.
    They all found the 3 regular duplicates and some of them would mark two and leave the 3rd unmarked.  iPhoto Duplicate Cleaner can sort the found duplicates by file size but if the file was edited to get the reduced file size it might not be found as it would have a different file creation/modification date. 
    iPhoto Library Manage was able to find all duplicates and mark them as such if the file names were the same the the filename option was selected.  Otherwise it also missed the modified, resized version.  It allowed one to select the one photo to save before going to work on the library.
    So if a photo has been reduced in image quality or pixel size it will not be considered a duplicate.
    OT

  • File SIZES are Different ?

    Hi Experts,
    I have implemented ESR By pass Scenario.
    The problem is we are getting the file sizes are differently in Source folder and Target folder
    How can we resolve this issue
    Please Guide Me
    Thanks and Regards,
    Ravi Teja.

    Refer to question no. 2 under the below wiki and see if that configuration helps.
    Sender File Adapter Frequently Asked Questions - Process Integration - SCN Wiki

  • FCP file sizes are twice the size of Quicktime's - any advantage to FCP?

    When I capture footage from my Canon HV40 MiniDV using FCP, the file sizes are almost twice the size as when I capture using Quicktime's Device Native setting. Is there any advantage to capturing with FCP, as the quality seems the same in QT? Or is there a way to reduce the file size being captured in FCP?
    I'd like to be able to capture in FCP because if there is a dropped frame in the recording, Quicktime stops recording, whereas FCP keeps going and starts a new file. But the larger file size in FCP puts me off.

    Not all codecs are editable...meaning, not all work in an editing application. I'm not sure what "device native" captures as in QT. Does it say in GET INFO? But FCP captures HDV and DV native, but in a native format that it can work with. HDV comes in as HDV...Dv as DV.
    I wouldn't trust what QT is capturing as DEVICE NATIVE. And 13GB/hour of footage is not that much. That's the lowest data rate format out there.
    Shane

  • FCE file sizes are twice the size of Quicktime's - any advantage to FCE?

    When I capture footage from my Canon HV40 MiniDV using FCE, the file sizes are almost twice the size as when I capture using Quicktime's Device Native setting. Is there any advantage to capturing with FCE, as the quality seems the same? Or is there a way to reduce the file size being captured in FCE?
    I'd like to be able to capture in FCE because if there is a dropped frame in the recording, Quicktime stops recording, whereas FCE keeps going and starts a new file. But the larger file size in FCE puts me off.

    There is no difference in the two files. They are both DV Stream (.dv).
    If you look in the Movie Properties window you'll notice that both tracks (audio and video) have the same file size.
    Obviously, this can't mean that the audio track is really as large as the video track in data size. It means that the only way QuickTime can "separate" the stream into two editable tracks is to duplicate them (but ignore the video playback of the audio track portion).
    I don't use Final Cut but I would suspect it might make two files (audio and video) instead of a single file made by QuickTime Player Pro.
    Confusing, isn't it?

  • Why the files my program create are created twice each file double ? And why sometimes the files size are too small ?

    My program is using Queue of type Uri to create Queue of urls and then i'm using webbrowser to navigate each Uri from the Queue in it's turn and get the url(html) source content and save it to the hard disk.
    The problem is sometimes the text files on the hard disk are small like 90KB or 60KB and sometimes they are as they are suppose to be 300KB or 200KB.
    This is a button click event where i'm calling two methods:
    private void toolStripButton3_Click(object sender, EventArgs e)
    GetHtmls();
    CheckQueue();
    This is the GetHtmls method code:
    private Queue<Uri> myUrls = new Queue<Uri>();
    private bool isBusy = false;
    private void GetHtmls()
    for (int i = 1; i < 49; i++)
    adrBarTextBox.Text = sourceUrl + i;
    targetHtmls = (combinedHtmlsDir + "\\Html" + i + ".txt");
    Uri targetUri = new Uri(sourceUrl + i);
    myUrls.Enqueue(targetUri);
    sourceUrl contain website address: http://www.tapuz.co.il/forums2008/forumpage.aspx?forumid=393&pagenumber=
    And i'm adding to it the numbers and create the pages.
    And add them to the Queue.
    THen the CheckQueue method:
    Uri uri;
    private void CheckQueue()
    if (isBusy)
    return; // We're downloading some page right now, don't disturb
    isBusy = true; // OK, let's get started
    if (myUrls.Count == 0) // No more pages to download, we're done
    isBusy = false;
    return;
    uri = myUrls.Dequeue(); // Get one URL from queue
    getCurrentBrowser().Navigate(uri);
    It suppose to Navigate to each Uri(html address) in the Queue.
    And the browser document completed event:
    private void Form1_DocumentCompleted(object sender, WebBrowserDocumentCompletedEventArgs e)
    // If page loaded completly then do something
    int urlnumber = uri.ToString().IndexOf("pagenumber=");
    string number = uri.ToString().Substring(urlnumber + 11);
    int num = Int32.Parse(number);
    targetHtmls = (combinedHtmlsDir + "\\Html" + num + ".txt");
    StreamWriter writer = File.CreateText(targetHtmls);
    writer.Write(getCurrentBrowser().DocumentText);
    writer.Close();
    isBusy = false; // We're done
    CheckQueue(); // Check next page in queue
    In the completed event i'm getting the page number and build the string for the text file then write the html source content to the text file.
    In the end i have on my hard disk 48 text files.
    The problems are:
    1. Sometimes it seems like it's not finishing navigating to the current uri or maybe some other reason maybe server side problem and creating small size files with source content inside but not all the source content. Sometimes the text files size are each
    file 99KB or 70KB and sometimes the size of them are about 300KB and 200KB and this is the right sizes 300KB 200KB.
    2. The text files on my hard disk suppose to be 48 different files each file should contain the source if the html page of the 48 pages. But on my hard disk the 48 text files are duplicated for some reason.
    This is the file on my hard disk:
    Some of the files are 205KB 350KB 175KB and some of the files sizes are 85KB 94KB 35KB 
    Why some of the files it didn't navigated to the end or maybe didn't got all the source ?
    And why it's making each second file the same like the one before ? It suppose to create 48 different files but i'm getting two identical files each navigation.

    I solved it now.
    This is what i did:
    It's a bit slow process since i'm waiting for each page to be loaded into the webbrowser but it does the work.
    using System;
    using System.Collections.Generic;
    using System.ComponentModel;
    using System.Data;
    using System.Drawing;
    using System.Linq;
    using System.Text;
    using System.Threading.Tasks;
    using System.Windows.Forms;
    using System.Collections;
    using System.IO;
    using System.Net;
    namespace WindowsFormsApplication1
    public partial class Form1 : Form
    private string sourceUrl = "http://test.test";
    private string htmlsTargetDirectory = "Test Htmls";
    private string appDir = Path.GetDirectoryName(@"C:\Users\chocolade1972\AppData\Local\Test_Images\Test Images\Test Htmls");
    private string combinedHtmlsDir;
    private String targetHtmls;
    private int counter = 1;
    private StreamWriter w;
    private string uri;
    private bool htmlloaded = false;
    public Form1()
    InitializeComponent();
    webBrowser1.ScriptErrorsSuppressed = true;
    combinedHtmlsDir = Path.Combine(appDir, htmlsTargetDirectory);
    if (!Directory.Exists(combinedHtmlsDir))
    Directory.CreateDirectory(combinedHtmlsDir);
    private void Form1_Load(object sender, EventArgs e)
    GetHtmls();
    timer1.Enabled = true;
    private void GetHtmls()
    uri = sourceUrl + counter;
    targetHtmls = (combinedHtmlsDir + "\\Html" + counter + ".txt");
    webBrowser1.Navigate(uri);
    private void webBrowser1_DocumentCompleted(object sender, WebBrowserDocumentCompletedEventArgs e)
    if (e.Url.ToString() == uri)
    targetHtmls = (combinedHtmlsDir + "\\Html" + counter + ".txt");
    htmlloaded = true;
    StreamWriter writer = File.CreateText(targetHtmls);
    writer.Write(webBrowser1.DocumentText);
    writer.Close();
    FileInfo fi = new FileInfo(targetHtmls);
    var size = fi.Length;
    w = new StreamWriter(combinedHtmlsDir + "\\test.txt", true);
    w.WriteLine("File Size " + size);
    w.Close();
    private void timer1_Tick(object sender, EventArgs e)
    if (htmlloaded == true)
    uri = sourceUrl + counter;
    myurls.Add(uri);
    webBrowser1.Navigate(uri);
    htmlloaded = false;
    counter += 1;

  • Trying to download update to CoPilot Live and CoPilot GPS with maps.  files sizes are large and taking hours to download on wireless connection.  How can I download App updates and new maps while connected to PC and Itunes through hard wire internet link?

    Trying to download update to CoPilot Live and CoPilot GPS with maps.  Files sizes are large and taking hours to download on wireless connection.  How can I download updates and new maps while connected to PC and Itunes through hard wire internet link?

    I'm on my iPad, so I don't know if this is the page with an actual download. I don't see a button, but assume that is because I  am on an iPad. It is in the DL section of Apple downloads.
    http://support.apple.com/kb/DL1708

  • After duplicate operation, file sizes(Checkpoint file size) are different

    HI
    I have a some questions.
    We are testing a 4-way Replication. After duplicate operation, file sizes(Checkpoint file size) are different in OS command(du -sh).
    Is the normal?
    TimesTen Version : TimesTen Release 7.0.5.0.0 (64 bit Solaris)
    OS Version : SunOS 5.10 Generic_141414-02 sun4u sparc SUNW,SPARC-Enterprise
    [TEST17A] side
    [TEST17A] /mmdb/DataStore # du -sh ./*
    6.3G ./SAMPLE
    410M ./SAMPLE_LOG
    [TEST17A] /mmdb/DataStore/SAMPLE # ls -lrt
    total 13259490
    -rw-rw-rw- 1 timesten other 501 Aug 14 2008 SAMPLE.inval
    -rw-rw-rw- 1 timesten other 4091428864 Jan 29 02:13 SAMPLE.ds1
    -rw-rw-rw- 1 timesten other 4113014784 Jan 29 02:23 SAMPLE.ds0
    [TEST17A] /mmdb/DataStore/SAMPLE # ttisql sample
    Command> dssize ;
    PERM_ALLOCATED_SIZE: 8388608
    PERM_IN_USE_SIZE: 36991
    PERM_IN_USE_HIGH_WATER: 36991
    TEMP_ALLOCATED_SIZE: 524288
    TEMP_IN_USE_SIZE: 5864
    TEMP_IN_USE_HIGH_WATER: 6757
    [TEST17B] side
    [TEST17B] /mmdb/DataStore # du -sh ./*
    911M ./SAMPLE
    453M ./SAMPLE_LOG
    [TEST17B] /mmdb/DataStore/SAMPLE # ls -lrt
    total 1865410
    -rw-rw-rw- 1 timesten other 334 Dec 11 2008 SAMPLE.inval
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:25 SAMPLE.ds1
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:25 SAMPLE.ds0
    [TEST17B] /mmdb/DataStore/SAMPLE # ttisql sample
    Command> dssize;
    PERM_ALLOCATED_SIZE: 8388608
    PERM_IN_USE_SIZE: 432128
    PERM_IN_USE_HIGH_WATER: 432128
    TEMP_ALLOCATED_SIZE: 524288
    TEMP_IN_USE_SIZE: 5422
    TEMP_IN_USE_HIGH_WATER: 6630
    [TEST18A] side
    [TEST18A] /mmdb/DataStore # du -sh ./*
    107M ./SAMPLE
    410M ./SAMPLE_LOG
    [TEST18A] /mmdb/DataStore/SAMPLE # ls -lrt
    total 218976
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:22 SAMPLE.ds0
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:32 SAMPLE.ds1
    [TEST18A] /mmdb/DataStore/SAMPLE # ttisql sample
    Command> dssize;
    PERM_ALLOCATED_SIZE: 8388608
    PERM_IN_USE_SIZE: 36825
    PERM_IN_USE_HIGH_WATER: 37230
    TEMP_ALLOCATED_SIZE: 524288
    TEMP_IN_USE_SIZE: 6117
    TEMP_IN_USE_HIGH_WATER: 7452
    [TEST18B] side
    [TEST18B] /mmdb/DataStore # du -sh ./*
    107M ./SAMPLE
    411M ./SAMPLE_LOG
    [TEST18B] /mmdb/DataStore/SAMPLE # ls -lrt
    total 218976
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:18 SAMPLE.ds1
    -rw-rw-rw- 1 timesten other 4091422064 Jan 29 02:28 SAMPLE.ds0
    [TEST18B] /mmdb/DataStore/SAMPLE # ttisql sample
    Command> dssize;
    PERM_ALLOCATED_SIZE: 8388608
    PERM_IN_USE_SIZE: 36785
    PERM_IN_USE_HIGH_WATER: 37140
    TEMP_ALLOCATED_SIZE: 524288
    TEMP_IN_USE_SIZE: 5927
    TEMP_IN_USE_HIGH_WATER: 7199
    Thank you very much.
    GooGyum

    You don't really give much detail on what operations were performed and in what sequence (e.g. duplicate from where to where...) nor if there was any workload running when you did the duplicate. In general checkpoint file sizes amongst replicas will not be the same / remain the same because:
    1. Replicas are logical replicas not physical replicas. Replication transfers logical operations and applies logical operations and even if you try and do exactly the same thing at both sides in exactly the same order there are internal operations etc. that are not necessarily synchronised which will cause the size of the files to vary somewhat.
    2. The size of the file as reported by 'ls -l' represents the maximum offset that has so far been written to in the file but the current 'usage' of the file may be less than this at present.
    3. Checkpoint files are 'sparse' files (unless created with PreAllocate=1) and so the space used as reported by 'du' will in general not correspond ot the size of the file as reported by 'ls -l'.
    Unless you are seeing some kind of problem I would not be concerned at an apparent difference in size.
    Chris

  • Why are Preview saved file sizes so HUGE?

    We have a neat copier at work that will scan in a document and email the pdf to you. So I have a 100 page document that I scanned in 4 chunks that yielded 4 pdf files of about 2.5MB each. I opened the first in Preview and then dragged the remaining 3 files in to create my final 100 page document. When I did the Save command I was shocked to see the resultant file size came in at 27MB! I searched around the menus and tried a Save As and selected the option for just black and white. The file that was created when that command finished ended up being 67MB.
    Anybody know how I can just get the originally expected 10MB file size? Many thanks in advance.

    I just tried the Combine PDFs program and it also yielded a 27MB file size when the sum of all the input files was only 10MB. So now I'm really confused. If I had to take a guess, I'd say the pdf format supports some kind of compression and both Preview and Combine PDFs can read in compressed pdfs but can't write them. I still need to get down to the 10MB size (because I can't email the file 27MB size I ended up with) so I'm still looking for help.

  • SSRS report Excel Export file size is huge

    Hello everyone,
    I am facing an issue with SSRS report exporting  to Excel.
    Issue : Exported file size is too big.
    I have a matrix report ( parametrized report ) it fetches data from multiple table.
    initially when
    1) I select all parameters all values
    2) load the report after that
    3) exporting to excel
    so exported file size was about 10 Mb to 12MB
    Now, suddenly from this month exported file size is too big around 62 MB for same data,
    I checked on different forums and tried the solutions but it didn't worked,
    can anyone please suggest me on this.
    My environment is
    Visual studio (BIDS) 2008
    SQL Server 2008 R2
    Excel 2007
    Thanks

    Hi UdayKGR,
    According to your description, when you export a report into an excel file, the exported file size is too huge. Now you want to reduce the file size. Right?
    In this scenario, do you export the report into excel 2003 file? When reports are first exported and saved to Excel 2003, they do not benefit from the file optimization that Excel automatically applies to its *.xls workbook files. The larger file size can
    cause problems for e-mail subscriptions and attachments. To reduce the size of the *.xls files for exported reports, open the *.xls files and then resave the workbooks. Resaving the workbooks typically reduces their file sizes by 40 to 50 percent. Also please
    try to install the Reporting Services 2008 R2 Service Pack 2 and reduce the merged columns in exported excel file. For more information, please see the links below:
    Exporting to Microsoft Excel (Report Builder and SSRS)
    problem
    regarding the size of exported report from reporting services to excel format
    SSRS Export to Excel - Performance
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Export to Word file size is huge

    Sometime after the end of March, the "Export to Word..." action in Pages has started producing very large file sizes.  The Pages file itself will be 119KB, but after exporting to .doc, it will become 1.5MB.  I have to export Pages files to Word all the time, to submit files to submission systems that have max file sizes, so I've been unable to upload my files, which is obviously a problem (and didn't used to be). 
    To fix this I have tried... restarting.  I have deleted the Pages plists, restarted.  I have tried exporting on a different computer with the newest version of Pages, which produced the same file size result.  I have copy/pasted the text into a new file, tried again.  Etc, etc.  In the end, I had to borrow a friend's computer to pull the 1.5MB Word export into Microsoft Word itself and resave, to get the expected smaller file size of 47KB.
    Please help.  Or, if there is no solution, I'll report as bug to Apple Feedback.  Thanks!

    To my knowledge all Pages 5 Exports have been particularly bloated.
    Since I have not considered Pages 5 worth using for serious work I can't compare between Pages 5.1 and Pages 5.2 as to whether the problem has actually got worse.
    All I know is that Pages 5 files & Exports are massively bigger than Page '09 files & Exports.
    Peter

  • LMS3.0.1 Log files' size are not purged.

    There is a LMS 3.0.1.
    Every day size of all log files only are increased.
    I have a 12G particion for log files but it's full every 2 weeks.
    In this case I have to stop LMS, delete all log files and start LMS manually every two weeks.
    Also, there is following message:
    CiscoWorks Folder Utilization is 8%Processes DFMCTMStartup, DataPurge are down.
    I tried to start these services several times but it failed all time.
    Hot to make LMS to prevent of increasing of log files over Recommended Size Limit?

    These two processes should be done in normal operation.  Do not try and start them manually.
    To rotate log files periodically, you will need to configure logrot.  Search the Common Services online help for "logrot", and you will get the instructions for configuring it.  Essentially, you will need to run NMSROOT/bin/perl NMSROOT/bin/logrot.pl -c, then walk through the menu prompts to configure the files you want to rotate.

  • Exported mp3 file size is huge!

    Does anyone know how to control the size of an exported song? For a 5 minute track I'm getting a 60Mb MP3!! I'm guessing this is done by setting the a different bit rate for the exported song but can't find any options in Garageband to do this.
    With such huge file sizes it makes it difficult to share the song through my website.

    No, you're not getting a 60 MB MP3, but a 60 MB AIFF file, and that's the regular size for uncompressed audio files. The bitrate in GB is fixed, and you wouldn't want to change it. Instead, convert the file into an MP3 or AAC in iTunes.
    (Come back if you need help with doing that.)

Maybe you are looking for