ESO application creates a large number of temporary file

Hello,
To summarize, when users run large queries, use attachments, of generate PDFs, etc, temporary files are created in /sourcing/tmp folder. Files older than 24 hours are cleared from this folder when logs roll over.
Our /sourcing filesystem is not sized to handle this u2018featureu2019 and I want to look for options. I am not keen on having the application filesystem /sourcing have a large quantity of temporary files going through it. The option that I can think of to get around it is to create a new filesystem or reuse an existing one and softlink /sourcing/tmp to it. Weu2019ll have to do this for each application server.
Does anybody have any suggestions as to how to avoid this problem?
Does anybody have any sizing inputs for sizing /sourcing/tmp.
Thanks,
Dnyandev

All,
The number of .blob and .bin files that get generated are large when u run volume testing. SAP has indentified it as a bug in the problem.
Does anybody have any solution.
We are currently using SAP ESourcing CLM 5.1
Thanks,
Dnyandev Kondekar

Similar Messages

  • Create a large number of purchase order in ME21N

    Hello,
    Is there a CATT or LSMW transaction or a program to create a large number of purchase orders with the positions?
    Thanks in advice
    Fany

    you can LSMW with direct input method
    Object - 0085, method - 0001
    venkat

  • Trouble loading a large number of csv files

    Hi All,
    I am having an issue loading a large number of csv files into my LabVIEW program. I have attached a png of the simplified code for the load sequence alone.
    What I want to do is load data from 5000 laser beam profiles, so 5000 csv files (68x68 elements), and then carry out some data analysis. However, the program will only ever load 2117 files, and I get no error messages. I have also tried, initially loading a single file, selecting a crop area - say 30x30 elements - and then loading the rest of the files cropped to these dimensions, but I still only get 2117 files.
    Any thoughts would be much appreciated,
    Kevin
    Kevin Conlisk
    Ph.D Student
    National Centre for Laser Applications
    National University of Ireland, Galway
    IRELAND
    Solved!
    Go to Solution.
    Attachments:
    Load csv files.PNG ‏14 KB

    How many elements are in the array of paths (your size(s) indicator) ?
    I suspect that the open file is somewhat limited to a certain number.
    You could also select a certain folder and use 'List Folder' to get a list of files and load those.
    Your data set is 170 MB, not really asthounising, however you should whatc your programming to prevent data-doublures.
    Ton
    Free Code Capture Tool! Version 2.1.3 with comments, web-upload, back-save and snippets!
    Nederlandse LabVIEW user groep www.lvug.nl
    My LabVIEW Ideas
    LabVIEW, programming like it should be!

  • I have an iPad 2 with iOS 5.1 and iBooks version 2.1.1. I have 64GB of storage, 80% is used. iBooks is using 250MB of storage. I have a large number of PDF files in my iBooks library. At this time I can not add another book or PDF file to my library.

    I have an iPad 2 with iOS 5.1 and iBooks version 2.1.1. I have 64GB of storage, 80% is used. iBooks is using 250MB of storage. I have a large number of PDF files in my iBooks library. At this time I can not add another book or PDF file to my library.  When I try to move a PDF file to iBooks the system works for a while...sometimes the file appears and than disappears....sometimes the file never appears. Is ther some limit to the number of books or total storage used in IBooks?  Thanks....

    Hi jybravo70, 
    Welcome to the Apple Support Communities!
    It sounds like you may be experiencing issues on your non iOS 8 devices because iOS 8 is required to set up or join a Family Sharing group.
    The following information is located in the print at the bottom of the article. 
    Apple - iCloud - Family Sharing
    Family Sharing requires a personal Apple ID signed in to iCloud and iTunes. Music, movies, TV shows, and books can be downloaded on up to 10 devices per account, five of which can be computers. iOS 8 and OS X Yosemite are required to set up or join a Family Sharing group and are recommended for full functionality. Not all content is eligible for Family Sharing.
    Have a great day, 
    Joe

  • Optimize a large number of pdf files using Acrobat Pro XI

    I have Adobe Acrobat Pro XI and need to optimize a large number of scanned files and other pdf's that are quite large. Is there any way to optimize multiple files at the same time, or do I have to do them one at a time (a major bummer if this is the case). I read on the forums that Acrobat Pro 9 had this capablity, so I can't imagine that a newer version wouldn't have it. Thanks for your help.

    You can use an Action to process multiple files in Acrobat Pro XI, via Tools - Action Wizard.
    Add to it a Save command and there you'll be able to specify that you want to optimize the file.

  • Approach to parse large number of XML files into the relational table.

    We are exploring the option of XML DB for processing a large number of files coming same day.
    The objective is to parse the XML file and store in multiple relational tables. Once in relational table we do not care about the XML file.
    The file can not be stored on the file server and need to be stored in a table before parsing due to security issues. A third party system will send the file and will store it in the XML DB.
    File size can be between 1MB to 50MB and high performance is very much expected other wise the solution will be tossed.
    Although we do not have XSD, the XML file is well structured. We are on 11g Release 2.
    Based on the reading this is what my approach.
    1. CREATE TABLE XML_DATA
    (xml_col XMLTYPE)
    XMLTYPE xml_col STORE AS SECUREFILE BINARY XML;
    2. Third party will store the data in XML_DATA table.
    3. Create XMLINDEX on the unique XML element
    4. Create views on XMLTYPE
    CREATE OR REPLACE FORCE VIEW V_XML_DATA(
       Stype,
       Mtype,
       MNAME,
       OIDT
    AS
       SELECT x."Stype",
              x."Mtype",
              x."Mname",
              x."OIDT"
       FROM   data_table t,
              XMLTABLE (
                 '/SectionMain'
                 PASSING t.data
                 COLUMNS Stype VARCHAR2 (30) PATH 'Stype',
                         Mtype VARCHAR2 (3) PATH 'Mtype',
                         MNAME VARCHAR2 (30) PATH 'MNAME',
                         OIDT VARCHAR2 (30) PATH 'OID') x;
    5. Bulk load the parse data in the staging table based on the index column.
    Please comment on the above approach any suggestion that can improve the performance.
    Thanks
    AnuragT

    Thanks for your response. It givies more confidence.
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    TNS for Linux: Version 11.2.0.3.0 - Production
    Example XML
    <SectionMain>
    <SectionState>Closed</SectionState>
    <FunctionalState>CP FINISHED</FunctionalState>
    <CreatedTime>2012-08</CreatedTime>
    <Number>106</Number>
    <SectionType>Reel</SectionType>
    <MachineType>CP</MachineType>
    <MachineName>CP_225</MachineName>
    <OID>99dd48cf-fd1b-46cf-9983-0026c04963d2</OID>
    </SectionMain>
    <SectionEvent>
    <SectionOID>99dd48cf-2</SectionOID>
    <EventName>CP.CP_225.Shredder</EventName>
    <OID>b3dd48cf-532d-4126-92d2</OID>
    </SectionEvent>
    <SectionAddData>
    <SectionOID>99dd48cf2</SectionOID>
    <AttributeName>ReelVersion</AttributeName>
    <AttributeValue>4</AttributeValue>
    <OID>b3dd48cf</OID>
    </SectionAddData>
    - <SectionAddData>
    <SectionOID>99dd48cf-fd1b-46cf-9983</SectionOID>
    <AttributeName>ReelNr</AttributeName>
    <AttributeValue>38</AttributeValue>
    <OID>b3dd48cf</OID>
    <BNCounter>
    <SectionID>99dd48cf-fd1b-46cf-9983-0026c04963d2</SectionID>
    <Run>CPFirstRun</Run>
    <SortingClass>84</SortingClass>
    <OutputStacker>D2</OutputStacker>
    <BNCounter>54605</BNCounter>
    </BNCounter>
    I was not aware of Virtual column but looks like we can use it and avoid creating views by just inserting directly into
    the staging table using virtual column.
    Suppose OID id is the unique identifier of each XML FILE and I created virtual column
    CREATE TABLE po_Virtual OF XMLTYPE
    XMLTYPE STORE AS BINARY XML
    VIRTUAL COLUMNS
    (OID_1 AS (XMLCAST(XMLQUERY('/SectionMain/OID'
    PASSING OBJECT_VALUE RETURNING CONTENT)
    AS VARCHAR2(30))));
    1. My question is how then I will write this query by NOT USING COLMUN XML_COL
    SELECT x."SECTIONTYPE",
    x."MACHINETYPE",
    x."MACHINENAME",
    x."OIDT"
    FROM po_Virtual t,
    XMLTABLE (
    '/SectionMain'
    PASSING t.xml_col                          <--WHAT WILL PASSING HERE SINCE NO XML_COL
    COLUMNS SectionType VARCHAR2 (30) PATH 'SectionType',
    MachineType VARCHAR2 (3) PATH 'MachineType',
    MachineName VARCHAR2 (30) PATH 'MachineName',
    OIDT VARCHAR2 (30) PATH 'OID') x;
    2. Insetead of creating the view then Can I do
    insert into STAGING_table_yyy ( col1 ,col2,col3,col4,
    SELECT x."SECTIONTYPE",
    x."MACHINETYPE",
    x."MACHINENAME",
    x."OIDT"
    FROM xml_data t,
    XMLTABLE (
    '/SectionMain'
    PASSING t.xml_col                         <--WHAT WILL PASSING HERE SINCE NO XML_COL
    COLUMNS SectionType VARCHAR2 (30) PATH 'SectionType',
    MachineType VARCHAR2 (3) PATH 'MachineType',
    MachineName VARCHAR2 (30) PATH 'MachineName',
    OIDT VARCHAR2 (30) PATH 'OID') x
    where oid_1 = '99dd48cf-fd1b-46cf-9983';<--VIRTUAL COLUMN
    insert into STAGING_table_yyy ( col1 ,col2,col3
    SELECT x."SectionOID",
    x."EventName",
    x."OIDT"
    FROM xml_data t,
    XMLTABLE (
    '/SectionMain'
    PASSING t.xml_col                         <--WHAT WILL PASSING HERE SINCE NO XML_COL
    COLUMNS SectionOID PATH 'SectionOID',
    EventName VARCHAR2 (30) PATH 'EventName',
    OID VARCHAR2 (30) PATH 'OID',
    ) x
    where oid_1 = '99dd48cf-fd1b-46cf-9983';<--VIRTUAL COLUMN
    Same insert for other tables usind the OID_1 virtual coulmn
    3. Finaly Once done how can I delete the XML document from XML.
    If I am using virtual column then I beleive it will be easy
    DELETE table po_Virtual where oid_1 = '99dd48cf-fd1b-46cf-9983';
    But in case we can not use the Virtual column how we can delete the data
    Thanks in advance
    AnuragT

  • Can i create a large number of identical tabs?

    I have searched around and found nothing on the topic.. what i am looking at here is this.
    I have currently set up two tabs that are identical, each representing a component on a device and the component is selected when the user clicks on the tab.
    For example, if i switch to tab two, i select component two.
    Now these tabs are identical, same controls same indicators and so on.. what i want to know is if it is possible to make it so that say.. if i tell it i have 5 components, it would show 5 identical tabs. if i told it it had 20 components it would bring up 20 and so on. At the moment i am looking at having to create each tab one by one and then hiding them..
    Thanks for your time and any help at all is greatly appreciated

    bracker wrote:
    altenbach--this problem could be phrased a little more generally though:  some, but not all types of controls could appear on each tab, while other controls are unique to particular tabs.  I've encountered exactly this situation, and I don't have a good way to duplicate the exact location of the common controls on each tab.  Can "copy and paste" be done "in place" somehow?
    Using altenbach's approach would be best. If you have a handful of specific controls but most are common you can control this by showing/hiding control. When the user selects the type of device you could show/hide the controls which are not common to all. Using this approach you could place controls on top of each other as long as only one would be visible at a time. This saves on diagram space and gives you a clean interface. Internally you use an array of clusters. If you may have different clusters you can decide if the differences are significant enough to warrant using different cluster definitions or simply having a few parameters that do not apply to all devices included in a single cluster definition. The controls on the front panel itself do not need to be in a cluster since they are only there for the UI. This gives you a bit more freedom in laying out the front panel. Anything that uses typedefs (ENUMs for example) should share the same typedef though.
    If you feel very ambitious and think you may reuse this control in several applications you could implement this as an X-Control.
    Mark Yedinak
    "Does anyone know where the love of God goes when the waves turn the minutes to hours?"
    Wreck of the Edmund Fitzgerald - Gordon Lightfoot

  • Large number of trace files generated

    Many of the following trace files are being generted throughout the day, sometimes 4/5 per minute
    There is nothing in the alert log
    Any ideas?
    Many Thanks in advance
    Dump file e:\oracle\admin\nauti1\udump\nauti1_ora_5552.trc
    Tue Nov 18 17:36:11 2008
    ORACLE V10.2.0.4.0 - Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - Production
    Windows Server 2003 Version V5.2 Service Pack 2
    CPU : 4 - type 586, 4 Physical Cores
    Process Affinity : 0x00000000
    Memory (Avail/Total): Ph:2045M/3839M, Ph+PgF:3718M/5724M, VA:649M/2047M
    Instance name: nauti1
    Redo thread mounted by this instance: 1
    Oracle process number: 32
    Windows thread id: 5552, image: ORACLE.EXE (SHAD)
    *** ACTION NAME:() 2008-11-18 17:36:11.432
    *** MODULE NAME:(Nautilus.Exe) 2008-11-18 17:36:11.432
    *** SERVICE NAME:(nauti1) 2008-11-18 17:36:11.432
    *** SESSION ID:(130.42066) 2008-11-18 17:36:11.432
    KGX cleanup...
    KGX Atomic Operation Log 342CD2A4
    Mutex 452CC5F8(130, 0) idn 0 oper EXAM
    Cursor Parent uid 130 efd 17 whr 26 slp 0
    oper=DEFAULT pt1=00000000 pt2=00000000 pt3=00000000
    pt4=00000000 u41=0 stt=0
    KGX cleanup...
    KGX Atomic Operation Log 342CD2A4
    Mutex 452CC5F8(130, 0) idn 0 oper EXAM
    Cursor Parent uid 130 efd 17 whr 26 slp 0
    oper=DEFAULT pt1=48265D6C pt2=48265E68 pt3=48265D3C
    pt4=00000000 u41=0 stt=0
    Dump file e:\oracle\admin\nauti1\udump\nauti1_ora_5552.trc
    Sat Nov 22 12:52:32 2008
    ORACLE V10.2.0.4.0 - Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - Production
    Windows Server 2003 Version V5.2 Service Pack 2
    CPU : 4 - type 586, 4 Physical Cores
    Process Affinity : 0x00000000
    Memory (Avail/Total): Ph:2070M/3839M, Ph+PgF:3896M/5724M, VA:673M/2047M
    Instance name: nauti1
    Redo thread mounted by this instance: 1
    Oracle process number: 29
    Windows thread id: 5552, image: ORACLE.EXE (SHAD)

    Check out metalink bug description for Bug 6638558

  • Large number of swf files causes Flash plugin to fail

    I'm displaying up to 700 swf files on a page as links to launch an AIR app.  These appear within a worklist with one swf for each worklist item.  This causes the Flash plugin for IE to fail.  Any ideas what causes this and how it can be addressed?  We are already looking at reducing the number of swf's as a short-term fix.

    There is no limit that I am aware of, but you should keep mindful of what is reasonable. Chances are your viewer will lose focus after about 15-30 min, so anything longer than that may not have the effectiveness that you are looking for.

  • Automated encoding of large number of video files

    I am trying to find some machine-friendly tools for video encoding. They must not require direct user interaction, e.g. through a GUI. Ideally, it will be something that I can control through an XML or command line interface. For example, I would like to be able to say something like "encode this file or network stream, subject to such-and-such parameters, and write the resulting bits to that file or network stream".
    Is there something in the QuickTime constellation -- QtPro/QTSS/DSS/Compressor/??? -- that will help me with this? I'm obviously not an expert in any of these products or SDKs and would appreciate any pointers, suggestions, recommendations, etc.
    Thanks,
    Scott Herzinger

    The QuickTime Pro Player can be controlled via AppleScript but I'm not certain what the range of capabilities are. Compressor allows batch processing and version 2 also has a command-line interface (can be run from the Terminal application or by using scripts). There are also third-party tools that allow batch processing using QuickTime codecs. Some of the latter are shareware or freeware, check on VersionTracker.com (for example, the freeware utility QTAmateur does batch exporting but works only with QT7 -- Pro or non-Pro).
    As for the command-line interface to Compressor 2, here is a thread that talks a little about what can be done (actually, a problem resolution to get the command line interface to work):
    w kucz, "Compressor & Command line" #2, 10:12am Aug 4, 2005 CDT

  • Batch process to create large number of metadata fields in UCM?

    Is there any kind of scripting available to create a large number of metadata fields in UCM?.
    thanks

    I can give you a code in java, which adds two metadata fields. It's quite hardcoded, but you could use it as a basis for a script:
    String FIELD1_NAME = "xsecretKey";
    String FIELD2_NAME = "xattachmentID";
    if (binder == null)
    binder = new DataBinder(SharedObjects.getSafeEnvironment());
    String maxOrderQuery = "select MAX(dOrder) as dOrder from DocMetaDefinition";
    try
    String val = loader.getDBConfigValue(COMPONENT_NAME, COMPONENT_SUBNAME, null);
    //binder.putLocal("dName", FIELD1_NAME);
    binder.putLocal("dName", FIELD2_NAME);
    ResultSet tempRs = ws.createResultSet("Qmetadef", binder);
    if (val == null || tempRs.isEmpty())
    ResultSet rset = ws.createResultSetSQL(maxOrderQuery);
    String orderStr = ResultSetUtils.getValue(rset, "dOrder");
    int order = NumberUtils.parseInteger(orderStr, 0);
    setNewFieldParams(binder, "BigText", "secretKey", "<$dateCurrent()*1$>");
    addNewMetadataField(FIELD1_NAME, binder, ++order, null, null, ws, cxt);
    setNewFieldParams(binder, "Text", "attachmentID", null);
    addNewMetadataField(FIELD2_NAME, binder, ++order, null, null, ws, cxt);
    rset = ws.createResultSetSQL("select * from DocMeta where dID=0");
    if (!testResultSetFieldExists(rset, FIELD1_NAME))
    FieldInfo fi[] = loader.createFieldInfo(new String[] {
    FIELD1_NAME
    }, 200);
    //fi[0].m_type = 5;
    ws.alterTable("DocMeta", fi, null, new String[] {
    "dName"
    if (!testResultSetFieldExists(rset, FIELD2_NAME))
    FieldInfo fi[] = loader.createFieldInfo(new String[] {
    FIELD2_NAME
    }, 30);
    //fi[0].m_type = 3;
    ws.alterTable("DocMeta", fi, null, new String[] {
    "dName"
    loader.setDBConfigValue(COMPONENT_NAME, COMPONENT_SUBNAME, "7.0", "1");
    catch (Exception e)
    SystemUtils.dumpException(null, e);
    }

  • Submit submit a large number of task to a thread pool (more than 10,000)

    i want to submit a large number of task to a thread pool (more than 10,000).
    Since a thread pool take runnable as input i have to create as many objects of Runnable as the number of task, but since the number of task is very large it causes the memory overflow and my application crashes.
    Can you suggest me some way to overcome this problem?

    Ravi_Gupta wrote:
    I have to serve them infinitely depending upon the choice of the user.
    Take a look at my code (code of MyCustomRunnable is already posted)
    public void start(Vector<String> addresses)
    searching = true;What is this for? Is it a kind of comment?
    >
    Vector<MyCustomRunnable> runnables = new Vector<MyCustomRunnable>(1,1);
    for (String address : addresses)
    try
    runnables.addElement(new MyCustomRunnable(address));
    catch (IOException ex)
    ex.printStackTrace();
    }Why does MyCustomRunnable throw an IOException? Why is using up resources when it hasn't started. Why build this vector at all?
    >
    //ThreadPoolExecutor pool = new ThreadPoolExecutor(100,100,50000L,TimeUnit.MILLISECONDS,new LinkedBlockingQueue());
    ExecutorService pool = Executors.newFixedThreadPool(100);You have 100 CPUs wow! I can only assume your operations are blocking on a Socket connection most of the time.
    >
    boolean interrupted = false;
    Vector<Future<String>> futures = new Vector<Future<String>>(1,1);You don't save much by reusing your vector here.
    for(int i=1; !interrupted; i++)You are looping here until the thread is interrupted, why are you doing this? Are you trying to generate loading on a remote server?
    System.out.println("Cycle: " + i);
    for(MyCustomRunnable runnable : runnables)Change the name of you Runnable as it clearly does much more than that. Typically a Runnable is executed once and does not create resources in its constructor nor have a cleanup method.
    futures.addElement((Future<String>) pool.submit(runnable));Again, it unclear why you would use a vector rather than a list here.
    >
    for(Future<String> future : futures)
    try
    future.get();
    catch (InterruptedException ex)
    interrupted = true;If you want this to break the loop put the try/catch outside the loop.
    ex.printStackTrace();
    catch (ExecutionException ex)
    ex.printStackTrace();If you are generating a load test you may want to record this kind of failure. e.g. count them.
    futures.clear();
    try
    Thread.sleep(60000);Why do you sleep even if you have been interrupted? For better timing, you should sleep, before check if you futures have finished.
    catch(InterruptedException e)
    searching = false;again does nothing.
    System.out.println("Thread pool terminated..................");
    //return;remove this comment. its dangerous.
    break;why do you have two way of breaking the loop. why not interrupted = true here.
    searching = false;
    System.out.println("Shut downing pool");
    pool.shutdownNow();
    try
    for(MyCustomRunnable runnable : runnables)
    runnable.close(); //release resources associated with it.
    catch(IOException e)put the try/catch inside the loop. You may want to ignore the exception but if one fails, the rest of the resources won't get cleaned up.
    The above code serve the task infinitely untill it is terminated by user.
    i had created a large number of runnables and future objects and they remain in memory until
    user terminates the operation might be the cause of the memory overflow.It could be the size of the resources each runnable holds. Have you tried increasing your maximum memory? e.g. -Xmx512m

  • Experience with creation of a large number of key figures

    Hi:
    I have a need to create a large number of Key Figures (at least about 50 of them) for DP use. Do you have any experience with automating this task if so, can you share your experience?  Thanks.
    Satish

    Hi,
    So far we have not come across a need to automate KF creation. And 50 KF is not a huge number, it would take may be a couple of hours to create them. Also, since it is a one time activity, developing an automation script for the same does not sound useful. You might want to consider creating them manually so that you can ensure all settings are proper.
    Hope this helps.
    Thanks & Regards
    Mani Suresh

  • NetUI file upload creates temporary files without cleanup

    When using the NetUI File Upload in my web application, I notice that it
    stores temporary files in the server config directory. The file names
    are netxxxxx.tmp and they are never deleted on server startup/shutdown,
    etc. I am uploading large files (about 500k) so these temporary files
    start to take up space very quickly and I have to manually delete them.
    Is there a way to disable this or what? I suppose I could edit a
    startup/shutdown script to delete them, but I'd rather not have to
    change the provided software.
    Thanks

    Bumping this up... any BEA reps read this NG?
    Steven Ostrowski wrote:
    When using the NetUI File Upload in my web application, I notice that it
    stores temporary files in the server config directory. The file names
    are netxxxxx.tmp and they are never deleted on server startup/shutdown,
    etc. I am uploading large files (about 500k) so these temporary files
    start to take up space very quickly and I have to manually delete them.
    Is there a way to disable this or what? I suppose I could edit a
    startup/shutdown script to delete them, but I'd rather not have to
    change the provided software.
    Thanks

  • Huge memory usage while parsing a large number of files

    hi,
    I use ORACLE8i XML parser class in my app.The problem is that
    when I parse a large number of xml files ,memory usage is so
    huge as more than 100M byte.Even worse,sometime I saw virtual
    memory low dialog.Could anybody give me some idea about this.
    any reply are welcome.Thanks a lot.

    Hi Mike. Yes I do have this enabled but this should only be requesting login to view the normal (open) applications. What I end up getting, after I login, is all the applications restarting. In the case of the synchronising files the database has to rebuild itself every time (160,000 files), and nothing has been happening while logged out. Using the Caffeine app I have still enabled the screensaver using a hot corner and had to log back in after 10 minutes but all the applications were still running, including the synchronisation. I am pretty sure my laptop has not gone into sleep mode, or it wouldn't have to reopen the apps, but it also has not shutdown as I get a login box as soon as I press any key. I can't understand the problem, and what makes it even stranger is that it is happening on two separate MacBooks at the same time. Thanks for the suggestion – really appreciated.
    OldGnome - thanks for the suggestion for the other app. I chose Caffeine as this has really good user reviews but the other one doesn't seem to have any, but I might still try it. Again thanks for your help.

Maybe you are looking for

  • HP 8620 printer two sided printing reverses pages

    Is there a way to change the order that this printer prints double sided prints? It prints last page first, so if it is 3 pages pages 2 & 3 are one page and page 1 is just one-sided.  Need this to be reversed.  Can you do this in the Mac driver? Lisa

  • Is it possible to connect Sqlfire server with oracle using DBSynchronizer using default license

    is it possible to connect Sqlfire server with oracle using DBSynchronizer using default license Sql fire is my machine Oracle is another machine When i connect sqlfire with oracle using DBsynchronizer It shows the error message as Wan is not supporte

  • Additional field in the sales order header Purchase Order Data tab

    Hi, I want to add an additional field in the below screen: VAO1 > Go to > Header > Purchase Order Data > Under Sold to party section i want to add an additional field. Is this possible through configuration? Please let me know. Thanks in advance Edit

  • SuSE 8.0 9iAS R2 and 9iDB R2 - Reporting Success

    Hi All, I'm just writing in to report Success on SuSE ver. 8.0 (The Internet D/L'ed version) for... Oracle 9iAS Release 2 (Yes Portal is working) and Oracle 9iDB Release 2 I currently have install documents that I put together for these particular in

  • My iPod 5 charger broke can I get a new one from Apple.

    I have a iPod 5th generation charger that is bent at a certain point and is critical for charging it. The spot where the wires are connected to the motherboard of the USB. The cable is bent on the USB that goes in the iPod. I've heard that the 5th ge