Most efficient way to use thumbnails of multiple sizes

When a user submits an image on my website, the upload script
currently creates thumbnails in three different sizes (120px, 90px,
and 20px). Different thumbnail sizes are used in different areas of
the site.
Is there a more storage-efficient way to display high-quality
thumbnails in different sizes, without requiring a separate
thumbnail file for each size used?
I cannot rely on browsers to resize images as the quality is
often very undesirable.

AngryCloud wrote:
> I may not have been clear in my last post...
>
> When an image is viewed normally on a page, it is saved
to the client's
> computer so that it will load instantly the next time
the image is called for.
>
> I do not want visitors to have to wait for the same
images they have already
> seen to re-download and resample. A file of each
resampled image should be
> saved to the client's computer to avoid this.
>
>
Is it possible to save resampled images on a page to the
client's
> computer?
It sounds like your page needs to check whether the cached
version
exists before creating the a new image, otherwise its always
going to
create a new version of the image and send it to the
browser... but I
don't know how this is possible.
Dooza
Posting Guidelines
http://www.adobe.com/support/forums/guidelines.html
How To Ask Smart Questions
http://www.catb.org/esr/faqs/smart-questions.html

Similar Messages

  • Fast/most efficient way to cut out of multiple tracks at the same time

    I try to select by dragging around but it lights up the entire length of each tract- I can use the scissors to manually cut the tracks but there has to be a better way...any ideas?
    Thanks, Don

    I don't know if this is most efficient but here's my "home grown" way:
    I place the section locator bars located at the top in the area I want to cut. You can fine tune this by zooming in close. I then right click on the track and use the split at locator function. I then can cut, drag or copy just that new, separate area. If you want to cut multiple tracks you can leave the locators where they are and repeat on all the tracks you want to split. I haven't tried selecting multiple tracks, and I'm not at my computer to try it right now.
    Maybe there's a quicker way but that has worked for me!

  • What is the most efficient way to use Remove Grain from After Effects on large Premiere Pro project?

    I've tried to use Dynamic Linking both ways but it doesn't the effects don't appear no matter how dramatic I make them. Why couldn't there just be Remove Grain in Premiere Pro so I wouldn't have to go through the headache of figuring this out. I've tried importing the Premiere Pro project into After effects where everything comes up and, since there's so many files, it doesn't seam feasibly or convenient at all. Trying Dynamic Link vice versa didn't lend results either. What can I do??? Projects range from about a half hour to an hour and a half. Thank you.

    The situations are well lit; it might be lack of knowledge on the camera settings itself.
    Sort that and your workflow issue dissapears instantly.
    There's nothing more efficient?
    Yep...read above... but ...any Plugin. eg Neat Video... will take long processing time.
    Fixing grain..actually "noise"...is always an image  degrading process as well.  It blurs!

  • Most efficient way to use streams & buffers ?

    Hya,
    I'm kinda new to Java & Network programming.
    I am trying to send a file over a network and store it at one end, However I am not sure how to use buffering to make both the data transfer over the connection and the write to storage efficient ?
    The code below is what I am using:
    InputStream inputStream = jxtaSocket.getInputStream();
    BufferedInputStream bufferedInputStream;
    bufferedInputStream = new BufferedInputStream( inputStream );
    File recievedFile = new File(searchResponse.getSharedFile().getFileName() );
    FileOutputStream fileOutputStream;
    fileOutputStream = new FileOutputStream(recievedFile, false);
    BufferedOutputStream bufferedOutputStream;
    bufferedOutputStream = new BufferedOutputStream( fileOutputStream );
    // bytes read from buffer in each go
    int bytes = 0;
    while( (bytes=bufferedInputStream.read()) != -1 )
    bufferedOutputStream.write( bytes );
    bytesCopied += bytes;
    bufferedInputStream.close();
    bufferedOutputStream.close();
    jxtaSocket.close();
    The type of file to be recieved is not known in advance, so would this code work regardless of the type of data that I recieve or should I use a byte[] array ?
    thanks in advance

    I'd use a byte[] instead of reading individual bytes. Haven't timed it, though, so it's more of a gut feeling.
    You could time it with a large file and educate all of us! Just stick a couple of System.currentTimeMillis() calls in there and print the difference. Do a couple of transfers before the actual measurement run, to give the JIT compiler time to compile everything.
    You'll probably need a file in the megabyte size range to have any measurable difference. And at least 100 Mbps networking; even your cell phone can swallow a 10 Mbps data stream without breaking a sweat, regardless of the reading method used.
        byte buf[] = new byte[8192];
        while (true) {
            int count = in.read(buf);
            if (count == 0)
                break;
            out.write(buf, 0, count);
        }The buffer size is a bit of a trade-off: a bigger buffer is marginally more expensive to create, while a smaller buffer has the cost of doing more read() and write() calls. As a guess, I'd say it is probably difficult to see any measurable difference between 1k...16k buffers. With "manual buffering" like that, the usefulness of BufferedXxxStream is likely to be nil, even negative.

  • Most efficiant way of using iPhoto with lightroom

    I am new to mac having moved from pc. I really like the interface and publish features of iphoto. I'm quite a serious photography so I want to use lightroom for the more professional features but continue with iPhoto for the family to browse my personal and better work through the simpler iphoto interface.
    2 questions really. I want lightroom to access all my photos so will I be importing photos to folders in pictures like I used to do with pc rather than importing them through iPhoto because of the package file.
    And what is the best way of putting better work into iPhoto without duplicating a lot of files.
    Hope you can help

    This post by Terence Devlin will give you some info:
    http://discussions.apple.com/message.jspa?messageID=7764845#7764845
    TIP: For insurance against the iPhoto database corruption that many users have experienced I recommend making a backup copy of the Library6.iPhoto (iPhoto.Library for iPhoto 5 and earlier) database file and keep it current. If problems crop up where iPhoto suddenly can't see any photos or thinks there are no photos in the library, replacing the working Library6.iPhoto file with the backup will often get the library back. By keeping it current I mean backup after each import and/or any serious editing or work on books, slideshows, calendars, cards, etc. That insures that if a problem pops up and you do need to replace the database file, you'll retain all those efforts. It doesn't take long to make the backup and it's good insurance.
    I've created an Automator workflow application (requires Tiger or later), iPhoto dB File Backup, that will copy the selected Library6.iPhoto file from your iPhoto Library folder to the Pictures folder, replacing any previous version of it. It's compatible with iPhoto 6 and 7 libraries and Tiger and Leopard. Just put the application in the Dock and click on it whenever you want to backup the dB file. iPhoto does not have to be closed to run the application, just idle. You can download it at Toad's Cellar. Be sure to read the Read Me pdf file.
    Note: There's now an Automator backup application for iPhoto 5 that will work with Tiger or Leopard.

  • What is the most efficient way to use my 2006ish iMac as a second screen for my MacBook pro?

    Hello.
    My line of work really requires the use of two screens, and I have a macbook pro from 2010 and a white iMac from 2006, I believe. My editing software lives on my laptop, so I'd like to use my iMac as the second display. I can't seem to find any male mini dvi to male mini displayport cables online, so I have a feeling they don't exist. Should I just get a mini dvi (male) to dvi (female) for the iMac and a (male) dvi to (male) mini displayport for the laptop? Is this going to cause latency if I'm running videos? Help please!

    Your iMac (from 2006) does not support Target Display Mode, that functionality only became available in Late 2009 with MiniDisplay Port. Even if you were to find a male-to-male cable the iMac isn't going to accept the video from the MacBook Pro in a method you're hoping to achieve (extended desktop mode). I've use ScreenRecycler and the lag is barely noticeable, I think even calling it "lag" is overkill.
    If you need a second monitor for work that is apparently an important accessory, why are they not supplying with a second monitor? Even if you were to find a valid connection solution, it's response time would still be slower than a direct link from the MacBook Pro.

  • Most efficient way to do multiple crops on many images?

    I have a large number of images shot in the default 4:3 aspect ratio. I need to print almost 200 as 4x6, and and undetermined but certainly large number as 8x10, so I have a lot of cropping to do. What would seasoned Aperture users suggest as the most efficient way to do this? I've thought of two possibilities:
    1. Duplicate every image I need to print in both sizes and crop one for each size print. This is the best option I've thought of, but it would certainly eat a lot of drive space.
    2. Do all the 8x10 crops, revert to original, and do the 4x6 crops. Saves disk space, but leave me with only the 4x6 crop in Aperture. (Sounds like I want to have my cake and eat it too, I suppose.)
    Anyway, there are a lot of you out there who have logged a lot more Aperture hours than I have. Is there a better workflow I have not considered?
    Thanks,
    Ben

    Hello Ben,
    the beauty of Aperture is, that you can have many versions of an image without needing much extra disk space.
    To have three versions of an image cropped to different aspect ratios, don't create duplicate master images but use (from the Aperture main menu)  "Photos -> Duplicate version" or "New Version from Master". Then crop this new version to a different aspect ratio. Aperture will not really render a new image file but just store the cropping rectangle to be able to create the cropped image when you export or print it.
    So you can have an original version, a 8x10 version, a 4x6 version in your library without needing much extra space - that is one of the rare occasions when you can have your cake and eat it too
    Regards
    Léonie

  • What is the most efficient way to post several stories to multiple html pages

    What is the most efficient way to post five stories to multiple html pages?
    Currently they are all html files saved in DW with some divs sharing content and other divs are unique.
    I've experimented with saving stories as library items and dropping into other html but wondered if there is a more efficient or dynamic process.

    Server-Side Includes.
    http://forums.adobe.com/message/2112460#2112460
    Nancy O.
    Alt-Web Design & Publishing
    Web | Graphics | Print | Media  Specialists
    http://alt-web.com/
    http://twitter.com/altweb

  • What is the most efficient way to convert a static site to a responsive site using Dreamweaver?

    I need to convert an old site made in Dreamweaver to be responsive to any monitor size. What is the most efficient way to do this?

    Depending on what you have to work with and how it was coded, it might be doable and then again not.  Suffice it to say, there are no magic buttons that will do this for you. Also consider that mobile & tablet users interact differently with their web devices. So your navigation & forms must be finger friendly.  Also images & content must make mobile users happy without killing their dataplans.  There's a lot of planning that goes into making a good Responsive Web site.
    Nancy O.

  • Most efficient way to delete "removed" photos from hard disk?

    Hello everyone! Glad to have this great community to come to for help. I searched for this question but came up with no hits. If it's already been discussed, I apologize and would love to be directed to the link.
    My wife and I have been using LR for a long time. We're currently on version 4. Unfortunately, she's not as tech-savvy or meticulous as I am, and she has been unknowingly "Removing" photos from the LR catalogues when she really meant to delete them from the hard disk. That means we have hundreds of unwanted raw photo files floating around in our computer and no way to pick them out from the ones we want! As a very organized and space-conscious person, I can't stand the thought. So my question is, what is the most efficient way to permanently delete these unwanted photos from the hard disk
    I did fine one suggestion that said to synchronize the parent folder with their respective catalogues, select all the photos in "Previous Import," and delete those, since they will be all of the photos that were previously removed from the catalogue.
    This is a great suggestion, but it probably wouldn't work for all of my catalogues since my file structure is organized by date (the default setting for LR). So, two catalogues will share the same "parent folder" in the sense that they both have photos from May 2013, but if I synchronize May 2013 with one, then it will get all the duds PLUS the photos that belong in the other catalogue.
    Does anyone have any suggestions? I know there's probably not an easy fix, and I'm willing to put in some time. I just want to know if there is a solution and make sure I'm working as efficiently as possible.
    Thank you!
    Kenneth

    I have to agree with the comment about multiple catalogs referring to images that are mixed in together... and the added difficulty that may have brought here.
    My suggestions (assuming you are prepared to combine the current catalogs into one)
    in each catalog, put a distinctive keyword onto all the images so that you can later discriminate these images as to which particular catalog they were formerly in (just in case this is useful information later)
    as John suggests, use File / "Import from Catalog" to bring all LR images together into one catalog.
    then in order to separate out the image files that ARE imported to LR, from those which either never were / have been removed, I would duplicate just the imported ones, to an entirely separate and dedicated disk location. This may require the temporary use of an external drive, with enough space for everything.
    to do this, highlight all the images in the whole catalog, then use File / "Export as Catalog" selecting the option "include negatives". Provide a filename and location for the catalog inside your chosen new saving location. All the image files that are imported to the catalog will be selectively copied into this same location alongside the new catalog. The same relative arrangement of subfolders will be created there, for them all to live inside, as is seen currently. But image files that do not feature in LR currently, will be left behind by this operation.
    your new catalog is now functional, referring to the copied image files. Making sure you have a full backup first, you can start deleting image files from the original location, that you believe to be unwanted. You can do this safe in the knowledge that anything LR is actively relying on, has already been duplicated elsewhere. So you can be quite aggressive at this, only watching out for image files that are required for other purposes (than as master data for Lightroom) - e.g., the exported JPG files you may have made.
    IMO it is a good idea to practice a full separation of image files used in your LR image library, from all other image files. This separation means you know where it is safe to manage images freely using the OS, vs where (what I think of as the LR-managed storage area) you need to bear LR's requirements constantly in mind. Better for discrete backup, too.
    In due course, as required, the copied image files plus catalog can be moved bodily to another drive (for example, if they have been temporarily put on an external drive, and you want to store them on your main internal one again). This then just requires a single re-browsing of their parent folder's location, in order to correct LR's records inside this catalog, as to the image files' changed addresses.
    If you don't want to combine the catalogs into one, a similar set of operations as above, can be carried out for each separate catalog you have now. This will create a separate folder structure in each case, containing just those duplicated image files. Once this has been done for all catalogs, you can start to clean up the present image files location. IMO this is very much the laborious and inflexible option, so far as future management of the total body of images is concerned... though there may still be some overriding reason for working that way.
    RP

  • The most efficient way to search a large String

    Hi All,
    2 Quick Questions
    QUESTION 1:
    I have about 50 String keywords -- I would like to use to search a big String object (between 300-3000 characters)
    Is the most efficient way to search it for my keywords like this ?
    if(myBigString.indexOf("string1")!=1 || myBigString.indexOf("string2")!=1 || myBigString.indexOf("string1")!=1 and so on for 50 strings.)
    System.out.println("it was found");
    QUESTION 2:
    Can someone help me out with a regular expression search of phone number in the format NNN-NNN-NNNN
    I would like it to return all instances of that pattern found on the page .
    I have done regular expressions, in javascript in vbscript but I have never done regular expressions in java.
    Thanks

    Answer 2:
    If you have the option of using Java 1.4, have a look at the new regular expressions library... whose package name I forget :-/ There have been articles published on it, both at JavaWorld and IBM's developerWorks.
    If you can't use Java 1.4, have a look at the jakarta regular expression projects, of which I think there are two (ORO and Perl-like, off the top of my head)
    http://jakarta.apache.org/
    Answer 1:
    If you have n search terms, and are searching through a string of length l (the haystack, as in looking for a needle in a haystack), then searching for each term in turn will take time O(n*l). In particular, it will take longer the more terms you add (in a linear fashion, assuming the haystack stays the same length)
    If this is sufficient, then do it! The simplest solution is (almost) always the easiest to maintain.
    An alternative is to create a finite state machine that defines the search terms (Or multiple parallel finite state machines would probably be easier). You can then loop over the haystack string a single time to find every search term at once. Such an algorithm will take O(n*k) time to construct the finite state information (given an average search term length of k), and then O(l) for the search. For a large number of search terms, or a very large search string, this method will be faster than the naive method.
    One example of a state-search for strings is the Boyer-Moore algorithm.
    http://www-igm.univ-mlv.fr/~lecroq/string/tunedbm.html
    Regards, and have fun,
    -Troy

  • What is the best, most efficient way to read a .xls File and create a pipe-delimited .csv File?

    What is the best and most efficient way to read a .xls File and create a pipe-delimited .csv File?
    Thanks in advance for your review and am hopeful for a reply.
    ITBobbyP85

    You should have no trouble doing this in SSIS. Simply add a data flow with connection managers to an existing .xls file (excel connection manager) and a new .csv file (flat file). Add a source to the xls and destination to the csv, and set the destination
    csv parameter "delay validation" to true. Use an expression to define the name of the new .csv file.
    In the flat file connection manager, set the column delimiter to the pipe character.

  • Most efficient way to open a new TextEdit doc at a specific place

    Suppose I've navigated in the Finder to some deep dark location in the folder hierarchy...   I want a new text file titled "notes" in this folder, i.e. to this path.
    Assuming TextEdit is already open, what is the most efficient way of creating and begin adding text to a new TextEdit file in that location?
    Tried this:  Maybe the usual Save dialog is aware of the current folder, so I can choose "New Document" in TextEdit's Dock icon pulldown, and then choose that path when I do File --> Save, choose the Save dialog's expanded view, and pull down the Where: selection.  Nope.  The current path is not there.
    Tried this:  Keeping an empty TextEdit document named untitled on my desktop.  Drag-copying that to the current folder.  Rename the file to "notes", open it, and start editing  That works, except it is clumsy.
    Is there a better way?
    Please forgive me if I'm missing something incredibly obvious.

    The problem with that is a new file has nothing in it.
    If you save a dummy file on the desktop someplace or use a real file in each of your locations, you could right click and duplicate it, then doubleclick to open it in the program of choice.
    Another option would to be to create a AppleScript that took the current open windows pathname and create and save a Textfile there.
    You save the app in the Dock and only have to click on it once, it automatically quits when mission accomplished.

  • Most efficient way to load XML file data into tables

    I have a complex XML file running into MBs. I want to load it's data into 7-8 tables.
    Which way will be better:
    1) Use SQL Loader to actually load directly into the 7-8 tables directly by modifying the control card.
    Is this really possible and feasible? I am not even sure about it
    2) Load data as XML Type in a table and register it. Then extract from there to load into various tables.
    Please help. I have to find the most efficient way of doing it.
    Regards,
    Sudhir

    Yes it is possible to use SQL*Loader to parse and load XML, but that is not what it was designed for and so is not recommended. You also don't need to register a schema, just to load/store/parse XML in the DB either.
    So where does that leave you?
    Some options
    {thread:id=410714} (see page 2)
    {thread:id=1090681}
    {thread:id=1070213}
    Those talk some about storage options and reading in XML from disk and parsing XML. They should also give you options to consider. Without knowing more about your requirements for the effort, it is difficult to give specific advice. Maybe your 7-8 tables don't exist and so using Object Relational Storage for the XML would be the best solution as you can query/update tables that Oracle creates based off the schema associated to the XML. Maybe an External Table definition works better for reading the XML into the system because this process will happen just once. Maybe using WebDAV makes more sense for loading XML to be parsed (I don't have much experience with this, just know it is possible from what I've read on the forums). Also, your version makes a difference as you have different options available depending upon the version of Oracle.
    Hope all that helps as a starter.
    Edited by: A_Non on Jul 8, 2010 4:31 PM
    A great example, see the answers by mdrake in {thread:id=1096784}

  • Most efficient way to do some string manipulation

    Greetings,
    I need to cleanse some data in a string by replacing unsafe characters with encoded equivalents. (FYI, this is for the purpose of transforming "unsafe" characters into encoded values as data inside an XML document).
    The following code accomplishes the task:
    Note that a string "currentValue" contains the data to be cleansed.
    A string, "encodedValue" contains the result.
      for (counter = 0; counter < currentValue.length(); counter++)
        addChar = (currentValue.substring(counter,counter+1));
        if (addChar.equals("<"))
          addChar = "#60;";
        if (addChar.equals(">"))
          addChar = "#62;";
        if (addChar.equals("="))
          addChar = "#61;";
        if (addChar.equals("\""))
          addChar = "#34;";
        if (addChar.equals("'"))
          addChar = "#39;";
        if (addChar.equals("'"))
          addChar = "#39;";
        if (addChar.equals("/"))
          addChar = "#47;";
        if (addChar.equals("\\"))
          addChar = "#92;";
        encodedValue += addChar;
      } // forI'm sure there is a way to make this more efficient. I'm not exactly "new" to java, but I am learning on my own with no formal training and often take a "brute force" approach with my initial effort.
    What would be the most efficient way to re-do the above?
    TIA,
    --Paul Galvin
    Integrated Systems & Services Group

    im a c++ programmer so im not totally up on these java classes either but...from a c++ stand point you might want to consider using the if else statment.
    by using if else, you only test the character until you find the actual "violating" character and skip the rest of the tests.
    also, you might trying using something to check for alphaNumeric cases first and use the continue keyword when you find one. since more of your characters are probably safe than unsafe you can skip all the ifs/if else statement and only do one test on the good characters. (i just looked for a way to test that and i didnt find one. c++ has a function that does that by checking the ascii number range. dont think that works in java. but maybe you can find one, it would reduce the number of tests probably.)
    happy hunting,
    txjump :)

Maybe you are looking for