Trail record size & check frecuency for end of uncommitted transactions

Hi, everyone,
Does anyone know if the trail record size can be changed from its default value of 4K?
What about how long the data pump process delays before searching for more data to proccess in its source trail while it waits for the end of a uncommitted transaction (which is 1 second)?

Thank you for your answer, MikeN
The delay I'm referring to cannot be set with EofDelay or EofDelayCSecs: these parameters establish how much time the Data Pump process sleeps when it has nothing new in its source trail. The delay that bothers me seems to happen when the Data Pump has nothing new in its source trail but it is in the middle of an open transaction processing.
I think is better explained with an example (which I think explains the goal of changing the record size too):
This is an excerpt for the Extract process trace:
*09:51:59.653622 write(20, "G\1\0\307H\0\0<E\4\0A\0F\5\377\2\361\361\317\21\232\265\300\0\0\0\0\0)h\20"..., 4096) = 4096* <----- Extract writes down the first 4K record of the transaction
09:51:59.653690 time(NULL) = 1349769119
09:51:59.653726 time(NULL) = 1349769119
09:51:59.653763 time(NULL) = 1349769119
09:51:59.653803 time(NULL) = 1349769119
09:51:59.653838 time(NULL) = 1349769119
09:51:59.653877 time(NULL) = 1349769119
09:51:59.653913 time(NULL) = 1349769119
09:51:59.653948 time(NULL) = 1349769119
09:51:59.653987 time(NULL) = 1349769119
09:51:59.654024 time(NULL) = 1349769119
09:51:59.654058 time(NULL) = 1349769119
09:51:59.654097 time(NULL) = 1349769119
09:51:59.654140 time(NULL) = 1349769119
09:51:59.654174 gettimeofday({1349769119, 654182}, NULL) = 0
09:51:59.654207 clock_gettime(CLOCK_REALTIME, {1349769119, 654216293}) = 0
09:51:59.654234 futex(0x9b62584, FUTEX_WAIT_PRIVATE, 957, {0, 999965707}) = 0
09:51:59.751502 futex(0x9b62568, FUTEX_WAKE_PRIVATE, 1) = 0
09:51:59.751554 llseek(19, 2722304, [2722304], SEEKSET) = 0
09:51:59.751608 futex(0x9b62534, FUTEX_WAKE_OP_PRIVATE, 1, 1, 0x9b62530, {FUTEX_OP_SET, 0, FUTEX_OP_CMP_GT, 1}) = 1
09:51:59.751682 nanosleep({0, 0}, NULL) = 0
*09:52:00.162689 write(20, "\0D\0\0O\0\0\0\30\0\0\0\0240000100050134977631"..., 2374) = 2374* <----- Extract writes down the remaining data for the transaction
And this is an excerpt of the corresponding Data Pump process trace:
09:51:59.653398 read(11, "F\0\4/0\0\1\3210\0\0\10GG\r\nTL\n\r1\0\0\2\0\0032\0\0\4 \0"..., 1048576) = 7604
09:51:59.653472 stat64("/stella_dat/ggate/tlstella/tl000195", 0xbfca2a0c) = -1 ENOENT (No such file or directory)
09:51:59.653543 nanosleep({0, 0}, NULL) = 0
09:51:59.653651 llseek(11, 0, [0], SEEKSET) = 0
*09:51:59.653543 nanosleep({0, 0}, NULL) = 0* <---- This is EOFDELAY: it's set to 0
09:51:59.653651 llseek(11, 0, [0], SEEKSET) = 0
*09:51:59.653709 read(11, "F\0\4/0\0\1\3210\0\0\10GG\r\nTL\n\r1\0\0\2\0\0032\0\0\4 \0"..., 1048576) = 11700* <----- Data Pump detects a new record in the source trail
09:51:59.653767 read(11, "", 1048576) = 0
09:51:59.653840 time(NULL) = 1349769119
09:51:59.653910 time(NULL) = 1349769119
09:51:59.653959 time(NULL) = 1349769119
09:51:59.654014 time(NULL) = 1349769119
09:51:59.654067 time(NULL) = 1349769119
09:51:59.654123 time(NULL) = 1349769119
09:51:59.654181 time(NULL) = 1349769119
09:51:59.654232 time(NULL) = 1349769119
09:51:59.654274 time(NULL) = 1349769119
09:51:59.654312 time(NULL) = 1349769119
09:51:59.654351 time(NULL) = 1349769119
09:51:59.654389 time(NULL) = 1349769119
09:51:59.654428 time(NULL) = 1349769119
09:51:59.654467 time(NULL) = 1349769119
09:51:59.654505 time(NULL) = 1349769119
09:51:59.654543 time(NULL) = 1349769119
09:51:59.654582 time(NULL) = 1349769119
09:51:59.654620 time(NULL) = 1349769119
09:51:59.654657 time(NULL) = 1349769119
09:51:59.654695 time(NULL) = 1349769119
09:51:59.654733 time(NULL) = 1349769119
09:51:59.654771 time(NULL) = 1349769119
09:51:59.654809 time(NULL) = 1349769119
09:51:59.654844 read(11, "", 1048576) = 0
*09:51:59.654881 nanosleep({1, 0}, NULL) = 0* <----- This is the 1 second delay that I want to get rid of
*09:52:00.655079 read(11, "\0D\0\0O\0\0\0\30\0\0\0\0240000100050134977631"..., 1048576) = 2374* <----- Data Pump reads the second record of the transaction

Similar Messages

  • Block size in tt for writing data to transaction log and checkpoint files

    Hello,
    what block size is TimesTen using when its writing data to transaction log and checkpoint files? Does it use some fixed block size during filesystem writes?

    Although in theory logging can write 2 KB blocks in almost all circumstances it will write 4 KB or larger so yes a filesystem with a 4 KB block size is fine for both checkpointing and logging.
    Chris

  • Customized LSMW (Batch Input Recording) to upload data for Vendor

    Hello Fiends,
       can u help me on this object and how to upload in xk01.
            Customized LSMW (Batch Input Recording) to upload data for Vendor Master using Transaction code XK01
    With best wishes,
    Chandu.
    Point will be rewarded....

    Hi,
    Go through the following link, you will find your answer
    http://www.sapbrain.com
    Regards,
    Bhaskar

  • Recording NWBC scripts for End User Experience

    Hi Experts,
    I am recording NWBC scripts using the Netweaver Business Client GUI for loading and running it through Robots for End User Experience Monitoring. I don't see an option to "LogOff" in NWBC client while recording  the script. We only have option to "Close all session and LogOff" in NWBC client .
    When I am using this option, it closes all the sessions including the EEM Recorder also without saving the recording .
    I need to include "LogOff" as the last step so that it doesn't leave any active sessions for each run by different robots.
    I have a Solution Manager 7.1 SP10 and NWBC client 3.5 Patch 13.
    Please help me achieve this.
    Regards,
    Manish K

    Hi,
    for a browser the recommendation is to open a new, empty tab to keep the browser alive while closing the tab with the  session. All that with the hope that the backend will use the triggered event of collapsing tab to close session.
    Maybe a similar strategy works for NWBC.. but I don't know the UI in detail.

  • HT6533 Where can I find the third party app? I'm really in trouble that I can't record and check my blood  glucose. Of cause I exported the data from app, but It's not readable for me. Please reply ASAP and give me other solution.

    Where can I find the third party app?
    I'm really in trouble that I can't record and check my blood  glucose.
    Of cause I exported the data from app, but It's not readable for me.
    Please reply ASAP and give me other solution.

    You find third-party apps in the App Store. What app are you looking for? Due to problems, Apple removed the ability to record that information in Health. You can always go back to whatever you used before three weeks Health was released. Why can you not check and record your blood glucose? I would assume you check it using blood samples and a glucose meter which has nothing to do with your phone. You should be able to record it in any number of apps, from Notes to purpose made third-party apps.

  • Serial records:Lot/Serial validation failed.  Please check log for details

    Hi,
    Received error while performing Quantity On Hand Conversion at MTL_TRANSACTIONS_INTERFACE table,
    "Serial records:Lot/Serial validation failed. Please check log for details."
    Perfomred following steps,
    1) Inserted data at MTL_TRANSACTIONS_INETRFACE table
    2) Executed "INV_TXN_MANAGER_PUB.process_Transactions"
    Working on Oracle EBS R12..
    Inventory Items having above mentioned error are not Lot Controlled and Serial Number generation is set to "At Receipt" . Can anyone tell, what is the cause of the error and where log file is created ?

    Hi;
    What is EBS version? Please see:
    ROI Fails with WMSINB-23840:Lot/Serial Validation Failed. Please Check Log for Details [ID 352570.1]
    Regard
    Helios

  • Error message:FRM-12001: Cannot Create the record group(check your query)

    Requirement: Need to get employee name and number in the LOV in search criteria.
    So I created LOV "full_name" and Record group Query under Employee Name property palette with
    select papf.title||' '||papf.last_name||', '||papf.first_name||' '||papf.middle_names emp_full_name
    ,papf.employee_number
    from apps.per_all_people_f papf, apps.per_person_types ppt
    where sysdate between papf.effective_start_date and papf.effective_end_date AND papf.person_type_id=ppt.person_type_id AND ppt.system_person_type IN ('EMP', 'OTHER', 'CWK','EMP_APL')
    AND PPT.default_flag='Y' and papf.BUSINESS_GROUP_ID=1
    order by papf.full_name
    I was unable to save and getting error message "FRM-12001: Cannot Create the record group(check your query)".
    I cant use PER_ALL_PEOPLE_F.FULL_NAME since full name here is last_name||title||middle_names||firstname.
    But my requiremnet is papf.title||' '||papf.last_name||', '||papf.first_name||' '||papf.middle_names emp_full_name .
    Can any one of you help me.

    First, Magoo wrote:
    <pre><font face = "Lucida Console, Courier New, Courier, Fixed" size = "1" color = "navy">create or replace function emp_full_name ( p_title in varchar2,
    p_last_name in varchar2,
    p_first_name in varchar2,
    p_mid_names in varchar2 ) return varchar2 is
    begin
    for l_rec in ( select decode ( p_title, null, null, p_title || ' ' ) ||
    p_last_name || ', ' || p_first_name ||
    decode ( p_mid_names, null, null, ' ' || p_mid_names ) full_name
    from dual ) loop
    return ( l_rec.full_name );
    end loop;
    end;</font></pre>
    Magoo, you don't ever need to use Select from Dual. And the loop is completely unnecessary, since Dual always returns only one record. This would be much simpler:
    <pre><font face = "Lucida Console, Courier New, Courier, Fixed" size = "1" color = "navy">create or replace function emp_full_name
    ( p_title in varchar2,
    p_last_name in varchar2,
    p_first_name in varchar2,
    p_mid_names in varchar2 ) return varchar2 is
    begin
    Return ( Ltrim( Rtrim ( p_title
    ||' ' ||p_last_name
    ||', '||p_first_name
    ||' ' ||p_middle_names )));
    end;</font></pre>
    And second:
    user606106, you did not mention how you got your record group working. However, you DO have an issue with spaces. If you change this:
    <pre><font face = "Lucida Console, Courier New, Courier, Fixed" size = "1" color = "navy">select papf.title||' '||papf.last_name||', '||papf.first_name||' '||papf.middle_names emp_full_name
    ,papf.employee_number </font></pre>
    to this:
    <pre><font face = "Lucida Console, Courier New, Courier, Fixed" size = "1" color = "navy">select Ltrim(Rtrim(papf.title||' '||papf.last_name||', '
    ||papf.first_name||' '||papf.middle_names)) AS emp_full_name,
    papf.employee_number</font></pre>
    it should work. The Ltrim(Rtrim()) removes leading and trailing spaces from the resulting full name.

  • R12 Positive Pay eText Template with File Header/Trailer Records

    A client for whom we are implementing R12.1.3 Positive Pay uses a bank which requires file header and trailer records in the file transmission. We used the standard IBYPOS_PAY_en.rtf template file and modified it to provide the additional, required detail. The new version of the template appears to be working exactly as needed when viewing the output from the template viewer. However, when positive pay is run through eBS there is no text output generated. The XML for report is being generated, as it can be pulled from both the concurrent request log file and the iby_trxn_documents table. Below is a snippet of the pertinent information from the log file:
    After formatting length of the output - baos::0
    After formatting, output is not null
    Output type is eText, so logging the output below:
    ==================================================
    ==================================================
    length of the output - baos::0
    After formatting, periodicSeq is null or empty.nullTimestamp: Fri Jan 11 12:02:53 EST 2013
    length of the output - baos::0
    Exit: iby.scheduler.FDFormatProgramUtils.doExtractAndFormat()Timestamp: Fri Jan 11 12:02:53 EST 2013
    Saving format output to file system.
    Enter: iby.scheduler.FDFormatProgramUtils.saveToFileSystemCustom()
    Concurrent request output file name: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/o455658.out
    File.separator: /
    Custom file name: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    Opened file in append mode: /u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    Updating FND request record for the custom file location/u02/app/appti/TCDXI/inst/apps/TCDXI_ncdx201/logs/appl/conc/out/CDX_455658.out
    request id: 455658
    Exit: iby.scheduler.FDFormatProgramUtils.saveToFileSystemCustom()
    Updating positive pay file created flag on the instruction record.
    Enter: iby.scheduler.FDFormatProgramUtils.postResults()
    Sql to be executed: BEGIN IBY_FD_POST_PICP_PROGS_PVT.set_pos_pay_created(:1); END;
    Payment instruction id: 10001
    Exit: iby.scheduler.FDFormatProgramUtils.postResults()
    Successfully updated instruction record.
    Exit: iby.scheduler.FDPositivePay.runPosPay
    Program completed successfully.
    java.util.logging.ErrorManager: 2:
    oracle.core.ojdl.LoggingException: oracle.core.ojdl.LoggingException: Attempt to flush a closed LogWriter
         at oracle.core.ojdl.ExceptionHandler.onException(Unknown Source)
         at oracle.core.ojdl.BufferedLogWriter.handleException(Unknown Source)
         at oracle.core.ojdl.BufferedLogWriter.flush(Unknown Source)
         at oracle.core.ojdl.logging.ODLHandler.flush(Unknown Source)
         at oracle.core.ojdl.logging.ODLHandler.publish(Unknown Source)
         at java.util.logging.Logger.log(Logger.java:478)
         at java.util.logging.Logger.doLog(Logger.java:500)
         at java.util.logging.Logger.log(Logger.java:523)
         at oracle.ias.cache.CacheInternal.logLifecycleEvent(Unknown Source)
         at oracle.ias.cache.CacheInternal.close(Unknown Source)
         at oracle.ias.cache.Cache.close(Unknown Source)
         at oracle.apps.jtf.cache.IASCacheProvider$CacheStopperThread.run(IASCacheProvider.java:1480)
    Caused by: oracle.core.ojdl.LoggingException: Attempt to flush a closed LogWriter
         ... 10 more
    Start of log messages from FND_FILE
    Enter:TIMESTAMP:: IBY_EXTRACTGEN_PVT.Create_Extract:: 11-JAN-13 12.02.52.760550000 PM -05:00
    Enter:TIMESTAMP::Pos Pay Extract(5) ::11-JAN-13 12.02.52.761687000 PM -05:00
    Enter:TIMESTAMP:: POS Payiby.plsql.IBY_FD_EXTRACT_GEN_PVT.Create_Pos_Pay_Extract_1_0:: 11-JAN-13 12.02.52.761942000 PM -05:00
    Exit:TIMESTAMP:: POS Payiby.plsql.IBY_FD_EXTRACT_GEN_PVT.Create_Pos_Pay_Extract_1_0:: 11-JAN-13 12.02.53.121536000 PM -05:00
    Exit:TIMESTAMP::Pos Pay Extract(5) ::11-JAN-13 12.02.53.121603000 PM -05:00
    End of log messages from FND_FILE
    Executing request completion options...
    Output file size:
    0
    ------------- 1) PRINT   -------------
    Disabling requested Output Post Processing. Nothing to process. The output of the request is zero byte.
    Any help that may be offered is greatly appreciated.
    Regards,
    Paul

    I faced this issue when I ran the program as a stand alone program. Please make sure that the Postive pay file is selected in the Payment Process profile which is used to make the payment. And submit the PPR and then run the positive pay from there(Payments Manager page) by selecting the check box.

  • External Table - possible bug related to record size and total bytes in fil

    I have an External Table defined with fixed record size, using Oracle 10.2.0.2.0 on HP/UX. At 279 byte records (1 or more fields, doesn't seem to matter), it can read almost 5M bytes in the file (17,421 records to be exact). At 280 byte records, it can not, but blows up with "partial record at end of file" - which is nonsense. It can read up to 3744 records, just below 1,048,320 bytes (1M bytes). 1 record over that, it blows up.
    Now, If I add READSIZE and set it to 1.5M, then it works. I found this extends further, for instance 280 recsize with READSIZE 1.5M will work for a while but blows up on 39M bytes in the file (I didn't bother figuring exactly where it stops working in this case). Increasing READSIZE to 5M works again, for 78M bytes in file. But change the definition to have 560 byte records and it blows up. Decrease the file size to 39M bytes and it still won't work with 560 byte records.
    Anyone have any explanation for this behavior? The docs say READSIZE is the read buffer, but only mentions that it is important to the largest record that can be processed - mine are only 280/560 bytes. My table definition is practically taken right out of the example in the docs for fixed length records (change the fields, sizes, names and it is identical - all clauses the same).
    We are going to be using these external tables a lot, and need them to be reliable, so increasing READSIZE to the largest value I can doesn't make me comfortable, since I can't be sure in production how large an input file may become.
    Should I report this as a bug to Oracle, or am I missing something?
    Thanks,
    Bob

    I have an External Table defined with fixed record size, using Oracle 10.2.0.2.0 on HP/UX. At 279 byte records (1 or more fields, doesn't seem to matter), it can read almost 5M bytes in the file (17,421 records to be exact). At 280 byte records, it can not, but blows up with "partial record at end of file" - which is nonsense. It can read up to 3744 records, just below 1,048,320 bytes (1M bytes). 1 record over that, it blows up.
    Now, If I add READSIZE and set it to 1.5M, then it works. I found this extends further, for instance 280 recsize with READSIZE 1.5M will work for a while but blows up on 39M bytes in the file (I didn't bother figuring exactly where it stops working in this case). Increasing READSIZE to 5M works again, for 78M bytes in file. But change the definition to have 560 byte records and it blows up. Decrease the file size to 39M bytes and it still won't work with 560 byte records.
    Anyone have any explanation for this behavior? The docs say READSIZE is the read buffer, but only mentions that it is important to the largest record that can be processed - mine are only 280/560 bytes. My table definition is practically taken right out of the example in the docs for fixed length records (change the fields, sizes, names and it is identical - all clauses the same).
    We are going to be using these external tables a lot, and need them to be reliable, so increasing READSIZE to the largest value I can doesn't make me comfortable, since I can't be sure in production how large an input file may become.
    Should I report this as a bug to Oracle, or am I missing something?
    Thanks,
    Bob

  • Open Hub Header and Trailer Record.

    Hi,
    For the Open Hub Destination
    Destination Type is File,
    Application server.
    Type of File Name: 2 Logical File name.
    How to to the get the Header and Trailer Record which will contain Creation Date, Creation Time and Total number of records.
    Header record Layput :
    Creation Date (YYYYMMDD)
    Creation Time (HHMMSS)
    Total record Count
    Trailer record Layout:
    Total number of Records in the file and
    XXXX( Key Figure ) Total.
    Thanks in advance.
    Regards,
    Adhvi.

    Hi Venkat,
    write a UDF in following way...
    pass the first parameter as the detail node (cache the whole queue) to the UDF pass the second parameter as the trailer countto the UDF
    now loop through the detail records get the count with a counter variable
    check the counter against the trailer count outside the loop
    if it doesnot match trigger the alert from the UDF itself
    Check the below link for triggering alert from an UDF
    /people/bhavesh.kantilal/blog/2006/07/25/triggering-xi-alerts-from-a-user-defined-function

  • Header and Trailer record validations

    Hi
    I have scenario file-xi-proxy. File contains the header record, detail record and trailer record
    In header record i am getting date field, in header i have to do validation like header record exits and it should be in date format.
    In Trailer record i have to do Total records and Total amount  equals to Total records processed and total amount.
    And amount value should be greater than zero..
    My source structure
    DT_ ACEAwardInformation
    <b>Header</b>
    BeginDate
    EndDate
    <b>DetailRecord</b>
    Field1
    Field2
    Field3
    Field4
    <b>TrailerRecord</b>
    TotalRecord
    TotalAmount
    Even content conversion parameters.
    venkat

    Hi Venkat,
    write a UDF in following way...
    pass the first parameter as the detail node (cache the whole queue) to the UDF pass the second parameter as the trailer countto the UDF
    now loop through the detail records get the count with a counter variable
    check the counter against the trailer count outside the loop
    if it doesnot match trigger the alert from the UDF itself
    Check the below link for triggering alert from an UDF
    /people/bhavesh.kantilal/blog/2006/07/25/triggering-xi-alerts-from-a-user-defined-function

  • Error maximum size of requests for one LUW

    Hi all,
    My problem is:
    In SAP ERP i call a function (ZQTC_NFE_CANCEL_XML_PI) that is implemented in SAP PI.
    Follow below my source code of a call:
    start----
    DATA: gv_rfcdest TYPE rfcdest,
          gs_cancel_xml TYPE zqtc_cancel_xml_layout.
    gv_rfcdest = 'SAPAVGXI'
    CALL FUNCTION 'ZQTC_NFE_CANCEL_XML_PI' IN BACKGROUND TASK
       DESTINATION gv_rfcdest
       EXPORTING
         i_cancel_xml          = gs_cancel_xml
       EXCEPTIONS
         communication_failure = 1
         system_failure        = 2
         OTHERS                = 3.
    IF sy-subrc <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    COMMIT WORK.
    end----
    When execute this function the error is "maximum size of requests for one LUW"
    Attention
    gs_cancel_xml is not big.
    Can anyone help me please!

    Hello,
    RFC destination progid created in R/3 should be same in XI system either D or Q system
    In your case check the RFC destination used and progid associated in that and make sure that the same progId exists in the system you want to receive data
    Note : progID is key for receiving data so make sure only one is active at a time.(in your case required D or Q s id in other system change the progid)
    Regards,
    Phani

  • Control/Trailer Record - use Pre-Reconciliation Workflow?

    preface: I am NEW to IdM.
    Would a best practice for verifying a control record in a flat file used for reconciliation be the use of the pre-reconciliation workflow? For instance, one implementation of the control record is a trailer record which indicates the file contents are complete. Other possibilities include verifying the control record outside of IdM with a script or implementing some other mechanism like a checksum on the file and do a verification prior to the reconciliation starting. Any assistance would be greatly appreciated.

    I vaguely remember encountering this last year and it was because the system did not recognise a blank value for the formula to work. What you ended up with was
    <> 0-3 AND ECR=0-3
    Can you put in a default value of "please select" that way when you change it to 0-3 the formula will go
    please select <> 0-3 AND ECR=0-3 and it will trigger
    cheers
    Alex

  • Check Box for each row in report -- all rows deleting when pressing DELETE

    Hello experts! I have set up a report with a check box for each row. When I click the DELETE button to delete the selected rows, every single one of the rows get deleted...even the ones that are not selected. I have my process point set to "On Submit - After computations and validations".
    This is my delete process (SHG is the table and SHG_ID is the primary key):
    FOR i in 1..HTMLDB_APPLICATION.G_F01.count
    LOOP
    DELETE FROM SHG
    WHERE SHG_ID = HTMLDB_APPLICATION.G_F01(i);
    END LOOP;
    Also, I've added to query in the region source this line:
    htmldb_item.checkbox(1, SHG_ID) del,
    Where does the "1" come into this? Not sure what I am doing wrong!
    Message was edited by:
    user477193
    Message was edited by:
    user477193

    The 1 (first argument to all the htmldb_item.* APIs) corresponds to the array number in htmldb_application.g_fNN. So 1 will populate array g_f01, 2 will populate g_f02 and so on.
    Your code seems fine, it should delete only the checked rows. Are you sure there is no other process on the page that might be deleting the rows? See if you can put up an example on htmldb.oracle.com

  • How can I print a full frame photo from iPhoto 11?  Even though I click on scale to fit paper size, it crops both ends of the photo in landscape set up.  I have an Epson R1800 with the latest drivers.

    How can I print a full frame photo from iPhoto 11?  Even though I click on scale to fit paper size, it crops both ends of the photo in landscape set up.  I have an Epson R1800 with the latest drivers. (I'm also 3 days new to iMac and iPhoto 11 from an eMac and iPhoto 6.)

    " Hello Jeff. Here at Oki Data, we do not support programming.  I would recommend contacting Adobe for further assistance. You can also browse their website at http://www.adobe.com/products/postscript/.

Maybe you are looking for

  • Apps removed from my iPhone b/c I don't have the most current OS version?

    Several apps including Yelp, London Tube, and other purchased apps were removed from my phone the last time I did a sync with iTunes with the message that a newer iPhone version was required. I currently have 3.0 and not 3.1.2 but why wouldn't it hav

  • F-22 Automatic Payment Transaction ( Payment Method )

    Hi we are not using Automatic payment transaction ( from customer ).  and not want to use it. Only for reporting purpose we are intended to use field (Lov)  Payment Method ( like cash, bank , DD etc ).  field. in transaction Incoming payment ( F-22)

  • Visa debit card do not work

    Hello! Can you translate my question from russian? Здравствуйте. У меня проблема с дебетной картой в iTunes. У меня карта visa classic и вчера я ввел ее данные в свой аккаунт, у меня списали со счета один доллар. После этого я купил три приложения, о

  • I wany to change to Firefox from IE8 to send and receive mails, how do I do this?

    I havw to use IE8 to access one program but I would like to send and receive my emails using Firefox. How do I make this change ?

  • Scope of MRP in REM

    hi gurus, i want to know the scope of MRP in a repetitive manufacturing scenario. regards sayeed