Size of batch

Dear experts
Can we increase size of Batch,current size of Batch is 10chracters.I there any customisation setting for the same,
Regards
Sandeep

Hy Sandeep,
Up to My knowledge it is not possible.
Refer below link,
[Link|http://sap.ittoolbox.com/groups/technical-functional/sap-log-mm/batch-number-limitation-with-10-characters-1649702]
Regards,
Dhaval

Similar Messages

  • Reducing file size in batch processes (multiple files)

    I am now using a canon XTi 10.5 megapixel camera. I'm loving it, taking tons and pics, and see my memory on my computer shrinking rapidly. I'm not too worried about hard drive space as I can simply buy another hard drive. Where I'm running into problems is on my iPod. I have an 80gb video iPod. It is suddenly full, Almost 1/2 with audio, 1/2 with photos. I'm realizing that I need take the 5-7 megabyte files each (photos) in iPhoto and reduce them down a bit in dimensions and/or jpeg compression.
    I use photoshop elements and am familiar with the multiple files option. I've used that a bit, nice feature.
    Heres my question:
    I know iPhoto is sensitive to manipulation of photos/changes if the changes aren't done through its interface. How can I go about taking the original pics for a number of events (talking about perhaps 2000 pics right now) and batch process them down in size a bit without weirding out iPhoto?
    Does iPhoto have an option to batch change files in such a fashion? Is it as simple as finding the originals in the finder and simpy reducing them in size and making iPhoto update its thumbnails?
    and how do I make iPhoto update its thumbnails, I forget...
    Thanks

    If you just want displayable image files on the iPod and not the full sized files then a very good solution, IMO, is to export those photos you want on the iPod to a folder on the Desktop. Then use Resize! to batch change the pixel dimension and the jpg compression level of the files. For my iPod Nano I use 640 x 480 and have Resize! compress them to medium. That gives me files in the 100 KB size range. If you want a smaller file just increase the compressions level a bit. Resize! automatically creates a new folder for the resized images so you can try different levels on the same folder. You can get a lot of photos on your iPod at that size. I then put that folder of resized files in my Pictures folder and have iTunes use it instead of using iPhoto.
    Do you Twango?
    TIP: For insurance against the iPhoto database corruption that many users have experienced I recommend making a backup copy of the Library6.iPhoto database file and keep it current. If problems crop up where iPhoto suddenly can't see any photos or thinks there are no photos in the library, replacing the working Library6.iPhoto file with the backup will often get the library back. By keeping it current I mean backup after each import and/or any serious editing or work on books, slideshows, calendars, cards, etc. That insures that if a problem pops up and you do need to replace the database file, you'll retain all those efforts. It doesn't take long to make the backup and it's good insurance.
    I've created an Automator workflow application (requires Tiger), iPhoto dB File Backup, that will copy the selected Library6.iPhoto file from your iPhoto Library folder to the Pictures folder, replacing any previous version of it. It's compatible with iPhoto 08 libraries. iPhoto does not have to be closed to run the application, just idle. You can download it at Toad's Cellar. Be sure to read the Read Me pdf file.

  • Save for Web with Optimize File Size for Batch Process

    I have a client who needs to be able to run a batch process on jpg images ranging in size from 600KB to 2.5MB optimizing their file size as close to 200KB as possible. The feature is included in Photoshop's save for web optimize dialog but it doesn't work for batch processes. If you record an action with the optimize to file size option set it processes all the images in the batch based on the quality settings required to resize the image you used during recording, meaning if the image you recorded was 1.5MB the quality setting might be 70 percent to hit the 200KB mark but that wouldn't be the correct quality setting for a 600KB file nor a 2.5MB file those would either be too small or too large.
    I've done quite a bit of searching both online and in the photoshop scripting forums but I haven't been able to find anything that would point me in the right direction.
    My client's paying for this work so if there's someone who knows how to script this who'd rather be paid than just do some pointing we can arrange for that quite easily.
    Thanks in advance for your assistance and suggestions.
    In case it's of importance we're using Photoshop CS4.

    If you don't mind spending a few bucks on a shareware prog, Jpegsizer has the option to save to a filesize.
    You can find it here.
    http://www.tangotools.com/jpegsizer/register.htm
    There is a free trial version, but you can't save files.

  • Weblogic Bridge Batch Size setting ?

    What is the per-requisite to set JMS bridge batch size in weblogic 10.3.3
    Customer setting in production:
    QoS --> Exactly once
    Asynchronous Mode Enabled --> True
    Batch Size --> 1
    From weblogic console
    "A messaging bridge instance provides transaction semantics when the QOS is Exactly-once. This envelops a received message and sends it within a user transaction (XA/JTA)."
    From documentation
    Changing the Batch Size
    When the Asynchronous Mode Enabled attribute is set to false and the quality of service is Exactly-once, the Batch Size attribute can be used to reduce the number of transaction commits by increasing the number of messages per transaction (batch). The best batch size for a bridge instance depends on the combination of JMS providers used, the hardware, operating system, and other factors in the application environment. See “Configure transaction properties” in Administration Console Online Help.
    Can batch size be set when Asynchronous Mode Enabled --> True ? If yes, what can be the optimal batch size in a production environment ?
    Right now application works fines for non-batch orders from CRM. When customers submit batch of 500 orders during specific hours of the day, lot of messages get queued up affecting the order completion rate.
    Thanks,
    Murali
    Edited by: murali_ora123 on May 22, 2013 6:28 AM
    Edited by: murali_ora123 on May 22, 2013 6:34 AM

    BatchSize does not take effect in async mode.
    What is the provider of your source destination? If it is not WLS JMS destination, your bridge may actually work in sync mode although you have configured it to work in async mode. In order to work in async mode for exactly-once QOS, a WLS proprietary feature is needed. We should see a log message in your server log file when the switch to the sync mode happens.
    If your bridge is indeed work in sync mode, you need to tune your batch size and batch interval to find the best performance for your application load condition. BatchInterval helps you send a batch before the batch is filled with the number of messages defined by batch size.
    Hope this helps.

  • Insprction Lot and Batch Number

    Hi all,
    I need your help to resolve following situaltion:
    Process order will be created on the basis of one campaign say production of 2-3 days. But batch needs to be created per GR after say every one hour. Hence batch will be created at the time of GR and not order release. But order has two operation, each having inprocess inspection. And inspection for 2nd operation will be n times i.e. untill required quality is acheived, result for second operation should be recorded.
    Now my questions are,
    1) How to link result recording of both operation with the batch which will be created after GR is done? (OR batch should be created at the time of release?)
    2) System will create only one inspection lot per order, so how to record result of each operation after every hour? (If inspection point needs to be used then how?)
    3) How to record result of multiple inspection for second operation till desired quality is obtained?
    Scenario:
    Process Order will be created for 1000 kg which is say 3 days production. (Now as there is need to identify stock of each GR seperatly we are thinking of creating batch at the time of GR and not at the time of order release.) Order has two operation, operation 10 and operation 20, each with in-process inspection. Result recording for operation 10 will be only once but for operation 20 will be done untill desired quality is not achieved. Production will be done in the lot of 50 kg. Hence quality testing and GR will be done of every 50 kg.
    Points will be rewarded to every useful replies.
    Thanks and regards,
    Devang

    Dear,
    Changing the Batch Level
    In the standard SAP System, the plant level is set as the batch level. If you want to use a different level in your company, you have to change it in Customizing and carry out a conversion program.
    You choose the batch level in Customizing for Batch Management under Specify Batch Level and Activate Status Management.
    While the conversion program is running, you can only post batches at the old level. You can only post batches at the new level after the conversion has been completed.
    You cannot change back from a higher level to plant level in the standard system. However, you can change the batch level from client level back to material level .
    To change the batch level from plant level to material level, the material has to be subject to batch management in all plants.
    Number range
    Not possible,
    Please rfer this link also,
    [ Batch Number|size of batch]
    Also Check OSS note 533377 Conversion of batch level.
    Please refer this,
    http://help.sap.com/saphelp_46c/helpdata/en/25/2838714f7811d18a150000e816ae6e/frameset.htm
    Regards,
    R.Brahmankar

  • Invalid Fetch Size

    Hi,
    I am connecting to Progress database from ODI 11.1.1.1.7 using jdbcodbcdriver. However when I try to fetch data from the tables, i get below error.
    java.sql.SQLException: Invalid Fetch Size
    Do I need to set Fetch Size to a particular value. If so, where and what size should this value be set to?
    Thanks,
    Parag

    Hi Parag,
    In the physical topology, double-click on your dataserver. You should see Array Fetch Size and Batch Update Size. For Oracle RDBMS the default is 30 so you might try it.
    Regards,
    JeromeFr

  • Batch Processing in ODI

    Hi Experts,
    I am queries about the batch processing in ODI.
    We want transform bulk of data...right now I am using Merge IKM for integration.
    When merge step will integrate the data to target table, it is doing in one go..
    Suppose during integration there is some problem like tablespace full etc. At that time whole merge will rollback.
    Can we commit the data in Junks or we need to modify the KM's to introduce this facility?
    Just checked with Array Fetch Size and Batch Update properties of Data Server..but these are not for commit process of the transactions.
    Thanks in advance.
    Regards,
    Sandeep

    Yes you are right, i think that option will work when that insert/update statement executes > 1000 times.
    But in our scenario single insert statement transforming millions of rows.

  • Note: Due to heavy load, the latest workflow operation has been queued. It will attempt to resume at a later time

    Dear all,
    sorry for opening another thread on this.
    I think I have a performance issue with workflows attached to document sets in SharePoint. And I say “I think” because people keep telling me that this is the way it just is.
    The user creates a new document set, which triggers a workflow in which the user has to confirm/review/approve a series of tasks. The time it takes from clicking the OK button on those task form to the workflow status moving to the next step is about 4 seconds.
    And visiting that status page within those 4 seconds brings up the infamous “Note: Due to heavy load, the latest workflow operation has been queued. It will attempt to resume at a later time.” message.
    Hitting Refresh in the browser after those 4 seconds will make the new workflow status appear and the red text go away.
    Is that normal? Is that the performance that everyone else is seeing as well?
    I struggle to see why simply moving a workflow from one task to another should take that on a machine that isn’t doing anything else at the time.
    (1)   
    I have a standalone (non-clustered) SharePoint box, 4 CPUs, 8 GB of memory, more than half of that available, acting as application server and wfe; only the database is on different box.
    (2)   
    The CPU only goes up to 18 or 19%, so CPU does not seem to be the bottleneck. Half the RAM is also still free.
    (3)   
    The workflow is designed with Nintex, and has about 9 flexi and review tasks – the last 2 of them in a loop iterating over typically 3 or 4 items.
    (4)   
    Looking at the logs it looks like the processing in Nintex only takes about 1 second – I don’t know where the other 3 seconds are going.
    (5)   
    There is nothing obvious in the logs.
    (6)   
    We’ve looked at all the “theoretical” improvements around throttling and batch sizes etc. – none of them appeared make any difference. And the workflow is so small that it looks like my tasks gets executed straight away. The problem appears
    to be that the execution takes too long(?) and therefore has not finished by the time the page get redrawn.
    (7)   
    I am running perfmon and I can e.g. see one(!) workflow being loaded into memory – as expected as I am the only user.
    (8)   
    I am seeing a total of 3(?) SQL queries being executed(?). I get the Bytes Sent/sec spiking at 25K, and Bytes received at 18K. But is this good or bad or a bottleneck?
    Where do I take it from here?
    I have been told that “[…] most customers have no issue with this as they are used to the way SP operates and it can be slow at times.” Is it really that bad?
    If it is worth watching more performance counters then I’d need to know what to compare them to.
    Is there something else I am missing?
    Thanks
    Martin

    Hi,
    Before considering an additional hardware try to change following configurations for workflow:
    Increase Throttle Size
    Increase Batch Size
    Time Out
    Workflow Timer Interval
    AutoCleanUpDays
    Increase Throttle Size
    The Workflow throttle setting controls how many Workflows can be processing at any one time on the entire server farm. By increasing the throttle it will allow the number of Workflows execution or can be initiated at a time.
    Use below PowerShell command to get the current Throttle Size:
    Get-SPFarmConfig | Select WorkflowPostponeThreshold
    Use below PowerShell command to set new Throttle Size:
    Set-SPFarmConfig -WorkflowPostponeThreshold 100
    Increase Batch Size
    This is the size that determines number of events processed for a single Workflow instance. Default value is 100, but it can be range from 1 to any number.
    Use below PowerShell command to get the current Batch Size:
    Get-SPFarmConfig | Select WorkflowBatchSize
    Use below PowerShell command to set new Batch Size:
    Set-SPFarmConfig -WorkflowBatchSize 200
    Time Out
    This decides the time out of the Workflow event. The default value is 5 and can be any integer. The time is in minute.
    Use below STSADM command to get the current Time Out value:
    stsadm -o getproperty -pn workflow-eventdelivery-timeout
    Use below STSADM command to get the current Time Out value:
    stsadm -o setproperty -pn workflow-eventdelivery-timeout -pv “15″
    Workflow Timer Interval
    This setting is applicable at Web Application level and not the farm level. The workflow timer interval specifies how often the workflow SPTimer job fires to process pending workflow tasks. This interval also represents the granularity of delay timers within
    your workflow. If a timer is set to delay for one minute, but the interval timer fires only every five minutes, the workflow delays for five minutes, not one minute.
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o getproperty -pn job-workflow -url <Web Application Url>
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o setproperty -pn job-workflow -pv “Every 10 minutes between 0 and 30″ -url <Web Application Url>
    Here is the url for reference :
    http://praveenkasireddy.wordpress.com/2013/06/14/workflow-due-to-heavy-load-the-latest-workflow-operation-has-been-queued-it-will-attempt-to-resume-at-a-later-time/

  • Due to heavy load, the latest workflow operation has been queued. It will attempt to resume at a later time.

    I have SharePoint 2010 Enterprise running SP1. Configuration is one SharePoint server in the farm and a SQL 2008 R2 database for the backend. Our user environment is 80 users with very little load on the SharePoint server. I have the workflow timer
    set to 1 minute.
    I have a SPD workflow that starts manually on a form library. Whenever I publish a new version of the workflow, the next time I start the workflow it takes the full minute to finish. If I click on the workflow status before it finishes, I see the message
    "Due to heavy load, the latest workflow operation has been queued. It will
    attempt to resume at a later time.". After the minute completes the workflow finishes.
    Here's the weird thing, the next time I start the workflow, it runs in a couple of seconds - almost instantly. I've tried up to 15 times after the inital publishing and everything seems to work fine on initiation.
    Well, that would be fine for me, however, I intermintantly get this heavy load message during task processes that are running inside the workflow. It's probably less than 5% of the time. It's really frustrating though so I appreciate some help. I'm look
    online and haven't found anything that describes my situation.
    Thank you in advance!

    Hi,
    Before considering an additional hardware try to change following configurations for workflow:
    Increase Throttle Size
    Increase Batch Size
    Time Out
    Workflow Timer Interval
    AutoCleanUpDays
    Increase Throttle Size
    The Workflow throttle setting controls how many Workflows can be processing at any one time on the entire server farm. By increasing the throttle it will allow the number of Workflows execution or can be initiated at a time.
    Use below PowerShell command to get the current Throttle Size:
    Get-SPFarmConfig |
    Select WorkflowPostponeThreshold
    Use below PowerShell command to set new Throttle Size:
    Set-SPFarmConfig -WorkflowPostponeThreshold
    100
    Increase Batch Size
    This is the size that determines number of events processed for a single Workflow instance. Default value is 100, but it can be range from 1 to any number.
    Use below PowerShell command to get the current Batch Size:
    Get-SPFarmConfig |
    Select WorkflowBatchSize
    Use below PowerShell command to set new Batch Size:
    Set-SPFarmConfig -WorkflowBatchSize
    200
    Time Out
    This decides the time out of the Workflow event. The default value is 5 and can be any integer. The time is in minute.
    Use below STSADM command to get the current Time Out value:
    stsadm -o getproperty -pn workflow-eventdelivery-timeout
    Use below STSADM command to get the current Time Out value:
    stsadm -o setproperty -pn workflow-eventdelivery-timeout -pv “15″
    Workflow Timer Interval
    This setting is applicable at Web Application level and not the farm level. The workflow timer interval specifies how often the workflow SPTimer job fires to process pending workflow tasks. This interval also represents the granularity of delay timers within
    your workflow. If a timer is set to delay for one minute, but the interval timer fires only every five minutes, the workflow delays for five minutes, not one minute.
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o getproperty -pn job-workflow -url <Web Application Url>
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o setproperty -pn job-workflow -pv “Every 10 minutes between 0 and 30″ -url <Web Application Url>
    Here is the url for reference :
    http://praveenkasireddy.wordpress.com/2013/06/14/workflow-due-to-heavy-load-the-latest-workflow-operation-has-been-queued-it-will-attempt-to-resume-at-a-later-time/

  • Speed test: PL/SQL vs. Java Stored Procedures

    I performed tests on these two procedures:
    ===========================================
    // Create a Statement
    Statement stmt = conn.createStatement (ResultSet.TYPE_SCROLL_INSENSITIVE,
    ResultSet.CONCUR_UPDATABLE);
    // Query the table
    ResultSet rset = stmt.executeQuery ("select SIFOB, SIFRA_GU from sezko");
    while (rset.next ()) {
    salary = rset.getInt(1);
    rset.updateInt (1, salary + 1);
    salary = rset.getInt(2);
    rset.updateInt (2, salary + 1);
    rset.updateRow();
    } // while
    conn.commit();
    // Close the RseultSet
    rset.close();
    // Close the Statement
    stmt.close();
    ===========================================
    procedure updateTable is
    cursor c_updateTable is select rowid, SIFOB, SIFRA_GU from sezko;
    begin
    for r_updateTable in c_updateTable loop
    update sezko set SIFOB = SIFOB + 1, SIFRA_GU = SIFRA_GU + 1 where rowid = r_updateTable.rowid;
    end loop;
    commit;
    end;
    ===========================================
    First procedure is written in Java (as Java Stored procedure) and second is PL/SQL.
    Java is about 10x slower than PL/SQL code.
    Can you explain bad performance results?
    thank you
    Matic & Ales

    Hi,
    I suppose the problem is not with the connection object,but with make connection to the database for every executeupdate or executeupdaterow called .Similarly for fetching the data from the database you
    can use the fetch size technique.Please check the 8.1.6 Java Developers guide for using this.
    Update Batching(For Batch updates and commits)
    Fetch Size(For Batch fetching)
    Oracle Row Pre-Fetching
    Regards
    Anand
    null

  • Workflow. Latest workflow operation has been queued

    Good day,
    I have sequential workflow which Is started after item created in list. Sometimes, after start of workflow in status I can
    see note:
    Due to heavy load, the latest workflow
    operation has been queued. It will attempt to resume at a later time.
    After several page refreshes it disappears.
    Can I somehow make this note not to appear?

    Hi,
    Before considering an additional hardware try to change following configurations for workflow:
    Increase Throttle Size
    Increase Batch Size
    Time Out
    Workflow Timer Interval
    AutoCleanUpDays
    Increase Throttle Size
    The Workflow throttle setting controls how many Workflows can be processing at any one time on the entire server farm. By increasing the throttle it will allow the number of Workflows execution or can be initiated at a time.
    Use below PowerShell command to get the current Throttle Size:
    Get-SPFarmConfig | Select WorkflowPostponeThreshold
    Use below PowerShell command to set new Throttle Size:
    Set-SPFarmConfig -WorkflowPostponeThreshold 100
    Increase Batch Size
    This is the size that determines number of events processed for a single Workflow instance. Default value is 100, but it can be range from 1 to any number.
    Use below PowerShell command to get the current Batch Size:
    Get-SPFarmConfig | Select WorkflowBatchSize
    Use below PowerShell command to set new Batch Size:
    Set-SPFarmConfig -WorkflowBatchSize 200
    Time Out
    This decides the time out of the Workflow event. The default value is 5 and can be any integer. The time is in minute.
    Use below STSADM command to get the current Time Out value:
    stsadm -o getproperty -pn workflow-eventdelivery-timeout
    Use below STSADM command to get the current Time Out value:
    stsadm -o setproperty -pn workflow-eventdelivery-timeout -pv “15″
    Workflow Timer Interval
    This setting is applicable at Web Application level and not the farm level. The workflow timer interval specifies how often the workflow SPTimer job fires to process pending workflow tasks. This interval also represents the granularity of delay timers within
    your workflow. If a timer is set to delay for one minute, but the interval timer fires only every five minutes, the workflow delays for five minutes, not one minute.
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o getproperty -pn job-workflow -url <Web Application Url>
    Use below STSADM command to get the current Workflow Timer Interval value:
    stsadm -o setproperty -pn job-workflow -pv “Every 10 minutes between 0 and 30″ -url <Web Application Url>
    Here is the url for reference :
    http://praveenkasireddy.wordpress.com/2013/06/14/workflow-due-to-heavy-load-the-latest-workflow-operation-has-been-queued-it-will-attempt-to-resume-at-a-later-time/

  • [ADF-11.1.2] Proof of view performance tuning in oracle adf

    Hello,
    Take an example of : http://www.gebs.ro/blog/oracle/adf-view-object-performance-tuning-analysis/
    It tells me perfectly how to tune VO to achieve performance, but how to see it working ?
    For example: I Set Fetch size of 25, 'in Batch of' set to 1 or 26 I see following SQL Statement in Log
    [1028] SELECT Company.COMPANY_ID,         Company.CREATED_DATE,         Company.CREATED_BY,         Company.LAST_MODIFY_DATE,         Company.LAST_MODIFY_BY,         Company.NAME FROM COMPANY Companyas if it is fetching all the records from table at a time no matter what's the size of Batch. If I am seeing 50 records on UI at a time, then I would expect at least 2 SELECT statement fetching 26 records by each statement if I set Batch Size to 26... OR at least 50 SELECT statement for Batch size set to '1'.
    Please tell me how to see view performance tuning working ? How one can say that setting batch size = '1' is bad for performance?

    Anandsagar,
    why don't you just read up on http://download.oracle.com/docs/cd/E21764_01/core.1111/e10108/adf.htm#CIHHGADG
    there are more factors influencing performance than just query. Btw, indexing your queries also helps to tune performance
    Frank

  • Performance tuning in ODI

    Hi,
    Can any help how to do performance tuning in ODI. What are the procedure for performance tuning. What needs to checked for performace tuning. Is there any document for ODI performace tuning.
    Thanks,
    Gnanamanju

    Hi Gnanamanju,
    Let me try to contribute a little,
    ODI performance strategy tells,
    1. Its recommended to run the AGENT in the target host rather than source. In case if your target is a remote host.
    2.Selecting appropriate KMs to load data, for example, if its a simple file to oracle transfer, its recommended to use native DB utilities like SQLLDR, External Table and so on.
    3. In case, if your staging and target are on the same server then make sure in Physical architecture your selected the correct schema (Data Schema and Work Schema)
    4. Right area of execution of joins and filters. If your source has more records then use the filter condition to be executed on SOURCE rather staging.
    5. Increase the Array Fetch Size and Batch Fectch Size of your Data server ( so that agent can accordingly fetch/insert the data)
    And more....
    Thanks,
    Guru

  • How to find original iPhoto Album photos in Library (by Album content)

    I have run out of Macbook space (OSX 10.7.5) and want to put all my photos on an external hard drive and share them between MacBook and iMac (no cloud please).  I want to do this by "Album" name because that's how they are organized.  I don't get the "events" thing in iPhoto, nor am I interested in the "Faces" stuff.  Just simple file access is all I want. 
    QUESTION 1:  How can I find the original photos after exporting by album?  I want to delete these permanently from the Library -- from my MAC.
    QUESTION 2:  Is there a smarter way to get these photos onto my external drive than using "export?"  And, is there a way to re-size a batch of files as opposed to one by one?  I made all my camera photos the largest size possible, one reason I've run out of disk space.  But to individually re-size would be impossible.
    Note:  I'm in a mess with photos.  Have 20K on MacBook and about 25K on my iMac.  There are 8 different iPhoto Libraries on my iMac, created by me in an effort to organize but I didn't know what I was doing and they are a total mess.  My goal is to get them all into this hard drive from both computers then start all over again.  But they need to be by album.  I'm not an "Event" person -- doesn't make sense to me, so I've put all my photos in albums.  When I start over I may not use iPhoto.  Thinking of going back to Windows actually.

    Sorry I'm terrible at explaining and even using my macbook haha. Let me try again. I imported some photos that I wanted to group together into an album. I forgot to create the album before importing a new set of photos which then took the place of the previous ones in the "last import" album. Is there anywhere or any way I can easily find the selection of photos I forgot to put into an album together without having to select them one by one from the "all photos" section? These would be the only pictures in my Photos application that haven't been organized into an album yet if that helps; however, these photos were not taken on the same date.

  • Photos from Computer - Quality Reduced

    When I used Itunes to transfer photos from my compuer that where already fairly small in size (a batch of photos that I had already reduced in size for e-mail). I tunes further compressed / reduced them to the point they look horrible.
    From the computer is there a way to upload photos without itunes modifying them at all? Can you upload photos without I tunes? I'd upload some via e-mail, but I have a quality issue there also.
    Do any of the file manage type programs in the I store help with full resolution photo transfers?
    Thanks very much
    Jamie

    Hi,
    Whenever you upload photos to your Touch via iTunes they get "optimised" that is squashed (or indeed inflated) to 640x480 pixels and emerge as file size of around 656kb, this happens regardless of the original size. These files are then stored in the "iPod Photo Cache" but are unreadable directly. It is pointless to pre-optimise them before uploading as iTunes will just do it anyway, indeed it could just make them look worse, it is best to use originals.
    Nobody I know has much regard for the resulting picture quality but it does serve a purpose and is ideal for "snaps".
    Photos saved from the net can be of a higher resolution and will stand zooming to a higher level but still appear to show some degradation over images saved to a desktop.
    Emailing the images would probably work but such a fiddle.
    I'm not sure if any file management apps allow email access to files, something you would have to test or I'm sure someone with more experience will comment soon.
    Regards,
    Dud.

Maybe you are looking for

  • Populating the value of Sel Screen Parameter in the Free Characrtistic

    Hi  , The requirement is that we have 2 dates - Touched Date and Created on Date . The Touched date is present in the Selection Screen . Now we need to use th value entered  for the Touched Date as default value of the created on Date in  the Free Ch

  • Galleries/Photos unviewable once published...

    Hey Guys, This stuff is taking me for quite a ride! Finally we got our personal domain launched and linked to our .mac account (after a week and a half of trial and error....), now I have a Galleries Page set up with one completed album for Seniors a

  • Edit a Color Palette to Reduce Colors

    I am trying to reduce the number of colors in a file. If I save for web devices and reduce the number of colors, Photoshop appears to go by volume, rather rather than distinctiveness. Area of distinct colors get merged and multiple shades of similar

  • The setup button on my MX350 printer has stopped working.

    Mx350 setup button no longer working. Can't setup new computers or use wifi.

  • Migrating from Endeca v2.2.2 to OEID v2.3

    Hi, We are planning to upgrade an existing v2.2.2 Endeca project to v2.3 and the Integrator work-space is not available. All we have is a lar file containing the Studio components and v2.2.2 data store (app_indexes directory). I wonder if it is possi