Problems with larger documents.

Here's my problem: I am a composer, I need to print my scores (document size is 9x12) but it seems that adobe only supports 8.5x11. What I'm looking to do is to save my 9x12 document as a pdf, and have it actually show up in adobe reader as a 9x12 document so that it can be accurately printed at its intended size. Every time I try to do just that, the document either rescales to accomodate to adobe, or parts of the document get cut off. That's not what I want, I want a 9x12 document. If in the unfortunate case that what I'm trying to do is impossible, please notify me of an alternate universal document reader that will do what I want.

I set up the score for my music in my notation software (finale 2011.) I must then convert the score from finale into a PDF, as there are additional items that I need to add to the score that can't be added from finale. So I export the document from finale as a 9x12 file. I use one of those PDF creators that acts as a printer since Finale 2011 for windows doesn't have direct pdf conversion capabilities. Once the PDF is created, I open it only to see that my large document has been subject to the 8.5x11 parameters of Adobe Acrobat, and much of it has been cropped out. I don't want to rescale, because again, it is a 9x12 document and that is all it will ever be. Is there a better alternative to this process that might yeild the desired outcome?

Similar Messages

  • Pagecount: problem with larger documents

    Hi,
    To print the pagecount I print &SFSY-PAGE&/&SFSY-JOBPAGES& on every page. This works for smaller documents (2-3 pages) but if I have larger documents (12-15 pages) it will print 1/, 2/, 3/, 4/ etc untill close to the end where it will print 12/13, 13/13 correctly.
    Anyone had this issue before?
    Thanks!

    Hi,
    I think you should specific a length option for field, just like:
    &SFSY-PAGE(4ZC)& / &SFSY-FORMPAGES(4ZC)&
    "4: set length to 4
    "Z: Suppresses leading zeros in numbers
    "C: This effect corresponds to that of the ABAP statement CONDENSE
    Please try,
    Thanks

  • Out.println() problems with large amount of data in jsp page

    I have this kind of code in my jsp page:
    out.clearBuffer();
    out.println(myText); // size of myText is about 300 kbThe problem is that I manage to print the whole text only sometimes. Very often happens such that the receiving page gets only the first 40 kb and then the printing stops.
    I have made such tests that I split the myText to smaller parts and out.print() them one by one:
    Vector texts = splitTextToSmallerParts(myText);
    for(int i = 0; i < texts.size(); i++) {
      out.print(text.get(i));
      out.flush();
    }This produces the same kind of result. Sometimes all parts are printed but mostly only the first parts.
    I have tried to increase the buffer size but neither that makes the printing reliable. Also I have tried with autoFlush="false" so that I flush before the buffer size gets overflowed; again same result, sometimes works sometimes don't.
    Originally I use such a system where Visual Basic in Excel calls a jsp page. However, I don't think that this matters since the same problems occur if I use a browser.
    If anyone knows something about problems with large jsp pages, I would appreciate that.

    Well, there are many ways you could do this, but it depends on what you are looking for.
    For instance, generating an Excel Spreadsheet could be quite easy:
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.io.*;
    public class TableTest extends HttpServlet{
         public void doGet(HttpServletRequest request, HttpServletResponse response) throws IOException, ServletException {
              response.setContentType("application/xls");
              PrintWriter out = new PrintWriter(response.getOutputStream());
                    out.println("Col1\tCol2\tCol3\tCol4");
                    out.println("1\t2\t3\t4");
                    out.println("3\t1\t5\t7");
                    out.println("2\t9\t3\t3");
              out.flush();
              out.close();
    }Just try this simple code, it works just fine... I used the same approach to generate a report of 30000 rows and 40 cols (more or less 5MB), so it should do the job for you.
    Regards

  • Really having problems with organising documents in pages. For example, deleting pages deletes the whole section and I cannot insert a section break as its greyed out & sending an object to background deletes several pages - what the **** is going on???

    Really having problems with organising documents in pages. For example, deleting pages deletes the whole section and I cannot insert a section break as its greyed out & sending an object to background deletes several pages and trying to insert a text box autmatically creates new pages- what the **** is going on??? Is this programme really this badly thought out? I thought everything apple was supposed to be intuitive!?!?!? Help.

    You can not insert a section break into a textbox.
    You appear to have a lot of objects with wrap set, that once you take them out of the flow of text (by sending them to the background) causes the text to reflow and contract into fewer pages. Pages is almost unique in the way it forms up pages in Word Processing mode only so long as there is text on them.
    I suspect you probably have hammered away at returns, spaces and tabs as well to position things, which only works to add to the mess.
    Download the Pages09_UserGuide.pdf available under the Help menu and swot up a bit on how it works.
    You may find this a usueful resource as well:
    http://www.freeforum101.com/iworktipsntrick/
    Peter
    PS There is quite a lot of programming in OSX that is far from "intuitive". Pages is easy at one level, when using the templates, otherwise it can be quite frustrating.

  • Problems with attached documents in Mail and Docs To Go

    Hello everybody.
    I've found a problem with attached documents in Mail. In fact, if I click on attached documents with .doc (microsoft office word 97-2004) and files in .docx Office Open .xml, the iPad doesn't open them, also if I click in "Open with Docs To Go (which should support these files' format)".
    How can I do? There's an application or another way to open these files?
    Other Informations:
    Docs To Go version: 5.2.2

    Update the patch to the Sapgui.
    Regards,
    Ignacio.

  • Problems with large scanned images

    I have been giving Aperture another try since 1.1 came out, and I am still having problems with large tiff files derived from scanned 4x5 negatives. The files are 500mb or more, 16 bit RGB, with ProPhoto RGB or Ektaspace PS5 profiles, directly out of the scanner.
    Aperture imports the files correctly, and shows their thumbnails. When I select a thumbnail "Loading" is displayed briefly, and the the dreaded "Unsupported Image Format" is displayed. Sometimes "Loading" goes on for a while, and a geometric pattern (looking like a rendering of random memory) is displayed. Restarting Aperture doesn't help.
    Lower resolution (250mb, 16bit) files are handled properly. The scans are from an Epson 4870 scanner. I have tried pulling the scans into Photoshop and resaving with various tiff options, and as PSD with no improvement. I have the same problem with corrected/modified psd files coming out of Photoshop CS2.
    I am running on a Power Mac G5 dual 2ghz with 8gb of RAM and an NVIDIA GeForce 6800 GT DDL (250mb) video card, with all the latest OS and software updates.
    Has anyone else had similar problems? More importantly, is anyone else able to work with 500mb files of any kind? Is it my system, or is it the software? I sent feedback to Apple as well.
    dual g5 2ghz   Mac OS X (10.4.6)  

    I have a few (well actually about 100) scans on my system of >500Mb. I tried loading a few and am getting an inconsistent pattern of errors that correlates with what you are reporting.
    I imported 4 files and three were troubled, the fouth was OK. I imported another four files and the first one was OK and the three others had your reported error, also the previously good file from the first import was now showing the same 'unsupported' image' message.
    I would venture to say that if you shoot primarily 4x5 and work with scans of this size that Aperture is not the program for you--right now. I shoot 35mm and have a few images that I have scanned at 8000dpi on my Imacon 848 but most of my files are in the more reasonable 250Mb range (35mm @ 5000dpi).
    I will probably downsample my 8000dpi scans to 5000dpi and not worry to much about it. In a world where people believe that 16 megapixels is hi-res you are obviously on the extreme side.(Good for you!) You should definately file a bug report but I wouldn't expect much help anytime soon for your super-sized scans.

  • Error, problem with saving documents, Illustrator

    I have a problem with saving documents in Adobe Illustrator (CS6): when I import to my document some jpg, tiff etc (even it isn't huge) Illustrator is crashed ("program to stop working correctly because of a problem. Windows will close the program and notify you if there is a solution available."). Next: program close.
    But when I work only with objects from AI - everything is ok - any error, I can save it.
    I have reinstalled and updated AI. And nothing changed.
    What is a problem?

    <moved from Downloading, Installing, Setting Up to Illustrator>

  • RESOLVED - K8N Neo2 Platinum problem with large Seagate disk

    Hi All,
    I've been running this system for 4 years and I'm very happy with it. But I've run into a problem:
    I'm trying to exchange my 120 GB Seagate Barracuda systemdisk with a 500 GB Barracuda and I can't install XP with SP2.
    First I tried to clone the old HD with Norton Ghost and Acronis True Image. The cloning looked fine but after taking the old HD out and moving the new one to P-ATA 1 as Primary Master I get an "error - no operating system found". I also tried a fresh install from CD. After the CD reboots the system I get the same error.
    The HD is present in BIOS with the right parameters and all is looking well - but still no go.
    I've fiddled around in BIOS (vers. 1.3) with the different settings (LBA, Large, Auto, Csh.) The last 2 options gives the "no operating system" error and the first 2 gives a "disk error" although Seagate says that I should change settings to "LBA".
    I only have the new HD in the machine. I also tried fixboot and fixmbr. I can see all the installed files when I put it in as a slave.
    So my question is:
    Does the K8N Neo2 Platinum have a problem with large hd's as system disk?
    Or is this a Seagate or Windows problem ?
    Any help is much appreciated!
    Mads

    Actually that was the first thing I did just to see if the drive worked :-) - except that I had it as master on the second ide cable. It formats like it should and all partitions report "healthy". The only difference is that the new drives c-partition is set as active (can't find a way to change it back)  - while the old one is set as system. Could this make a difference ?
    Since neither cloning nor a fresh install works I think the problem is something else. Maybe I should try Seatools and see if that makes any difference.
    Thank's all for help so far! - Other suggestions ?

  • Problem with large data report

    I tried to run a template I got from release 12 using data from the release we are using (11i). The xml file is about 13,500 kb. when i run it from my desktop.
    I get the following error (mostly no output is generated sometimes its generated after a long time).
    Font Dir: C:\Program Files\Oracle\BI Publisher\BI Publisher Desktop\Template Builder for Word\fonts
    Run XDO Start
    RTFProcessor setLocale: en-us
    FOProcessor setData: C:\Documents and Settings\skiran\Desktop\working\2648119.xml
    FOProcessor setLocale: en-us
    I assumed there may be compatibility issues between 12i and 11i hence tried to write my own template and ran into same issue
    when i added the third nested loop.
    I also noticed javaws.exe runs in the background hogging a lot of memory. I am using Bi version 5.6.3
    I tried to run the template through template viewer. The process never completes.
    The log file is
    [010109_121009828][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.setData(InputStream) is called.
    [010109_121014796][][STATEMENT] Logger.init(): *** DEBUG MODE IS OFF. ***
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.setTemplate(InputStream)is called.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.setOutput(OutputStream)is called.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.setOutputFormat(byte)is called with ID=1.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.setLocale is called with 'en-US'.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.process() is called.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] FOProcessor.generate() called.
    [010109_121014796][oracle.apps.xdo.template.FOProcessor][STATEMENT] createFO(Object, Object) is called.
    [010109_121318828][oracle.apps.xdo.common.xml.XSLT10gR1][STATEMENT] oracle.xdo Developers Kit 10.1.0.5.0 - Production
    [010109_121318828][oracle.apps.xdo.common.xml.XSLT10gR1][STATEMENT] Scalable Feature Disabled
    End of Process.
    Time: 436.906 sec.
    FO Formatting failed.
    I cant seem to figure out whether this is a looping or large data or BI version issue. Please advice
    Thank you

    The report will probably fail in a production environment if you don't have enough heap. 13 megs is a big xml file for the parsers to handle, it will probably crush the opp. The whole document has to be loaded into memory and perserving the relationships in the documents is probably whats killing your performance. The opp or foprocessor is not using the sax parser like the bursting engine does. I would suggest setting a maximum range on the amount of documents that can be created and submit in a set of batches. That will reduce your xml file size and performance will increase.
    An alternative to the pervious approach would be to write a concurrent program that merges the pdfs using the document merger api. This would allow you to burst the document into a temp directory and then re-assimilate them it one document. One disadvantage of this approach is that the pdf is going to be freakin huge. Also, if you have to send that piggy to the printer your gonna have some problems too. When you convert it pdf to ps the files are going to be massive because of the loss of compression, it's gets even worse if the pdf has images......Then'll you have a more problems with disk on the server and or running out of memory on ps printers.
    All of things I have discussed I have done in some sort of fashion. Speaking from experience your idea of 13 meg xml file is just a really bad idea. I would go with option one.
    Ike Wiggins
    http://bipublisher.blogspot.com

  • Problem with large option class during runtime 11.5.10

    It seems like our Oracle Configurator has a problem with loading multiple large option classes with about 10,000 items under it. We are running 11.5.10. My guess is that Oracle has to load everything in the option class and at some point the server just gives up after loading so much data.
    Basically the model looks like this and was set up by Oracle themselves in 04. Overtime the option classes grew.
    --Base model
    ----Model A instance 1
    --------option class
    ------------10000 items
    ----Model A instance 2
    --------option class
    ------------10000 items
    ----Model A instance 50
    --------option class
    ------------10000 items
    Is there an optimal amount for an option class? Is the model structure possibly outdated and we should break up the option classes into models to alleviate some of the pain? Or are we at the limitation that Oracle Configurator can do and should peruse a custom solution?

    Funny thing really because the Case study for Chapter 4 is from our company that Oracle put together for us and its not working in the sense that it seems the servers can't handle the current model structure. I wouldn't blame the UI for these crashes and we already hide stuff from the user (ie, having to sort through all 10,000 items and hiding the BOM structure). Quick story of why i want to try and fix it is becuase the way its set up now in the model document is that 3 people could kill the server by basically saying "ok i selected this item for this location" (which would load the 10,000 items into memory on the server and not to the UI upon selection of the model) and then they would say... "ok now copy this selection to all these 50 places" (which would load those items again for each place that the item had to copied to) and BOOM server dead. I think sandeeps diagram of the changes is a really good idea and could help out alot and I tossed out that idea a few times before around my work place to help with the server issues. I feel that it would work but is it also possible that maybe configurator can't support our model the way they are wanting it to work?

  • Problem with scanning document in PDF format to computer

    has anyone had a problem with scanning to the computer a document in pdf format from hp Photosmart 7525  e-All-in-One series wireless printer? The document appears in My Documents but when I try and open the document I get an error message that says the document cannot be opened. I can scan and open the document in TIF format. I am running Windows 7 Home Premium and I just installed the Photosmart 7525 printer. I've uninstalled and reinstalled the printer using the CD download.
    This question was solved.
    View Solution.

    Hi,
    Can you open other pdf files ? Do you have Adobe Reader in your computer ? Please download and install it on your machine then try to open file(s) again:
       http://get.adobe.com/reader/
    Regards.
    BH
    **Click the KUDOS thumb up on the left to say 'Thanks'**
    Make it easier for other people to find solutions by marking a Reply 'Accept as Solution' if it solves your problem.

  • Problem with  large databases.

    Lightroom doesn't seem to like large databases.
    I am playing catch-up using Lightroom to enter keywords to all my past photos. I have about 150K photos spread over four drives.
    Even placing a separate database on each hard drive is causing problems.
    The program crashes when importing large numbers of photos from several folders. (I do not ask it to render previews.) If I relaunch the program, and try the import again, Lightroom adds about 500 more photos and then crashes, or freezes again.
    I may have to go back and import them one folder at a time, or use iView instead.
    This is a deal-breaker for me.
    I also note that it takes several minutes after opening a databese before the HD activity light stops flashing.
    I am using XP on a dual core machine with, 3Gigs of RAM
    Anyone else finding this?
    What is you work-around?

    Christopher,
    True, but given the number of posts where users have had similar problems ingesting images into LR--where LR runs without crashes and further trouble once the images are in--the probative evidence points to some LR problem ingesting large numbers.
    I may also be that users are attempting to use LR for editing during the ingestion of large numbers--I found that I simply could not do that without a crash occuring. When I limited it to 2k at a time--leaving my hands off the keyboard-- while the import occured, everything went without a hitch.
    However, as previously pointed out, it shouldn't require that--none of my other DAMs using SQLite do that, and I can multitask while they are ingesting.
    But, you are right--multiple single causes--and complexly interrated multiple causes--could account for it on a given configuration.

  • Problems with attach documents in solman

    Hi Experts,
    I need to attached to my organizational business structure (in Tx SOLAR01 and SOLAR02) many templates documents of functional and technical specifications and have a problem with that.
    When I make the solar_project_admin transaction and click onto project template buttom to create a personal document the system gives me a warning message:
    "Documentation types in SAP namespace cannot be copied
    Message no. SPROJECT371
    Diagnosis
    You tried to copy documentation types in the SAP namespace into the template. This is not possible.
    System Response
    The documentation types were not copied."
    Could you please tell me how can I create my own document (SOLAR01 Y SOLAR02) to attached onto the structure?.
    Tks!
    Ignacio.

    Update the patch to the Sapgui.
    Regards,
    Ignacio.

  • Archive Link problem with redundant document names

    Hi all,
    I have a tricky problem in archive link. I would like to attach some different documents to an HR/PD Organisation unit ( objekttyp pdotype_o )
    All customizing works fine but the problem is, all documents attached to one busines object ID have the same name ( in this case objectnumber.pdf)
    I am not lucky with this issue because  the users must be able to see all documents ( via document finder ) attached to the org. unit  and all what the users get, is a list of  one or more documents with exactly the same name.
    In some cases this are about 10 different documents and this issue is a very bad solution because users have to open all attached docs to recognize which one they need.
    During the import/check in of the documents via transaction OAAD and OAWD, SAP renames the original document name.pdf  into the related object ID.pdf and there is no way to rerename it or influence this process.
    I tried to use some scan software with SAP interface but unfort. with the same result because software of this kind use the archive link interface too and cannot influence the process of check in and rename.
    Has anybody an idea ( third party software, customer development or something like this ) to prevent this problem and give me the possibility to stay with the original document names or change it after check in?
    Thank you in advance
    Rayko

    How you fix this problem?
    I'm with the same problem. Could you help me please?
    Edited by: broullon on Jul 16, 2009 5:31 PM

  • Problem with large spool lists

    Hello everyone,
    we have the following problem with one of our printers. When we are trying to print larger lists (>30 pages) the spool shows an "unknown error" and only a few pages are printed out.
    We have the following setup. A Kyocera FS-9520DN printer with Device Type KYOAAB1C which is connected via saprouter connection and host spool access method U. The only error I get from the tracefile of the Spool WP is
    S Tue Feb 22 07:34:37 2011
    S  *** ERROR => cannot write to connection [rspoclpd.c   857]
    S  *** ERROR => Lost connection, control file not sent [rspoyunx.c   1754]
    T  new memory block 000000002B6D6CE0
    S  handle memory type is RSTSPROMMM
    and the SP01 error message
    Status:              Incorrect (Reason unknown)
    Job status:          Completed
    Area:                Unknown
    Error:               Reason unknown
    Last event:
    Message:             Processing completed by spool work process
    Any suggestions what might improve the situation? Thanks in advance
    Marco

    Hi,
    A direct connection to a remote printer is not recommended as per the SAP Note #64628.
    The best option is to define the printer locally on the SAP server with Access method L for Unix or Access Method C for windows. If you need to use remote Access Method U or S, you should use a print server in between the SAP system and the printers.
    Regards,
    Aidan

Maybe you are looking for

  • Multiclip - mutliple audio tracks

    This question has been asked before, but I'll reiterate it. If the support doesn't exist, hopefully apple dev will take it into consideration. I shoot a lot of weddings and events. FCP's 5.0.x multiclip feature has been a lifesaver. I know the short

  • How to register User custom field using PHP?

    Hi! I am looking for a way to share user specific data between connected users. For example, all participants should know each other IDs(our own system IDs, not LCCS). User's custom fields looks the best solution for this, but I can't find a way to r

  • LMS 4.2 Syslog Collector doesn't work

    Hi fellas, I need a few help for my LMS 4.2, syslog collector on LMS doesnt working even service syslog collector running normaly and also i saw in syslog_info is working to collect syslog from all router but not show up in dashboard monitoring. I ha

  • Time service load-balancing

    Hi, I have a question about load-balancing facilities for Time service. If I schedule a TimeScheduleDef object to run at certain intervals, does the run-time environment take care of load balancing it across multiple machines on a cluster? Thanks, De

  • How to get report symbol in cube

    Hi Experts, In one of my cube not showing report symbol (only for new request) so how to get the report symbol. Regards, Siva