What's the best practice to manage the page file?

       
We have one Hyper-v Server running windows 2012 R2 with 128 GB RAM and 2 drives (C and D). It setup Automatically manage page file size for all drives. What's the best practice to manage the page file?
Bob Lin, MCSE & CNE Networking, Internet, Routing, VPN Networking, Internet, Routing, VPN Troubleshooting on http://www.ChicagoTech.net How to Install and Configure Windows, VMware, Virtualization and Cisco on http://www.HowToNetworking.com

For Hyper-V systems, my general recommendation is to set the page file to 1-4 GB. This allows for a mini-dump should something happen. 99.99% of the time, Microsoft will be able to figure out the cause of the problem from the mini-dump. It does not make
sense on a Hyper-V system to set aside enough space to capture all the memory on the system because only a very small portion of that memory is used by the parent partition. Most of the memory is under control of the individual VMs.
Yes, I had one of the Hyper-V product group tell me that I should let Windows manage it.  A couple of times I saw space on my system disk disappear because the algorithm decided it wanted all the space for the page file.  Made it so I couldn't
patch my systems.  Went back in and set the page file to 1-4 GB and have not had any issues since.
. : | : . : | : . tim

Similar Messages

  • What is the best practice in securing deployed source files

    hi guys,
    Just yesterday, I developed a simple image cropper using ajax
    and flash. After compiling the package, I notice the
    package/installer delivers the same exact source files as in
    developed to the installed folder.
    This doesnt concern me much at first, but coming to think of
    it. This question keeps coming out of my head.
    "What is the best practice in securing deployed source
    files?"
    How do we secure application installed source files from
    being tampered. Especially, when it comes to tampering of the
    source files after it's been installed. E.g. modifying spraydata.js
    files for example can be done easily with an editor.

    Hi,
    You could compute a SHA or MD5 hash of your source files on
    first run and save these hashes to EncryptedLocalStore.
    On startup, recompute and verify. (This, of course, fails to
    address when the main app's swf / swc / html itself is
    decompiled)

  • What is the best practice for using the Calendar control with the Dispatcher?

    It seems as if the Dispatcher is restricting access to the Query Builder (/bin/querybuilder.json) as a best practice regarding security.  However, the Calendar relies on this endpoint to build the events for the calendar.  On Author / Publish this works fine but once we place the Dispatcher in front, the Calendar no longer works.  We've noticed the same behavior on the Geometrixx site.
    What is the best practice for using the Calendar control with Dispatcher?
    Thanks in advance.
    Scott

    Not sure what exactly you are asking but Muse handles the different orientations nicely without having to do anything.
    Example: http://www.cariboowoodshop.com/wood-shop.html

  • What are the best practices to extend the overall lifespan of my MacBook Pro and its battery?

    In general what are the recomended practices to extend the lifespan of my batter and other general practice to extend the lifespan and characteristics(such as performance and speed) like new on my MacBook Pro which this past fall (2011)?

    About Batteries in Modern Apple Laptops
    Apple - Batteries - Notebooks
    Extending the Life of Your Laptop Battery
    Apple - Batteries
    Determining Battery Cycle Count
    Calibrating your computer's battery for best performance
    MacBook and MacBook Pro- Mac reduces processor speed when battery is removed while operating from an A-C adaptor
    Battery University
    Kappy's Personal Suggestions for OS X Maintenance
    For disk repairs use Disk Utility.  For situations DU cannot handle the best third-party utilities are: Disk Warrior;  DW only fixes problems with the disk directory, but most disk problems are caused by directory corruption; Disk Warrior 4.x is now Intel Mac compatible. Drive Genius provides additional tools not found in Disk Warrior.  Versions 1.5.1 and later are Intel Mac compatible.
    OS X performs certain maintenance functions that are scheduled to occur on a daily, weekly, or monthly period. The maintenance scripts run in the early AM only if the computer is turned on 24/7 (no sleep.) If this isn't the case, then an excellent solution is to download and install a shareware utility such as Macaroni, JAW PseudoAnacron, or Anacron that will automate the maintenance activity regardless of whether the computer is turned off or asleep.  Dependence upon third-party utilities to run the periodic maintenance scripts was significantly reduced since Tiger.  These utilities have limited or no functionality with Snow Leopard or Lion and should not be installed.
    OS X automatically defragments files less than 20 MBs in size, so unless you have a disk full of very large files there's little need for defragmenting the hard drive. As for virus protection there are few if any such animals affecting OS X. You can protect the computer easily using the freeware Open Source virus protection software ClamXAV. Personally I would avoid most commercial anti-virus software because of their potential for causing problems. For more about malware see Macintosh Virus Guide.
    I would also recommend downloading a utility such as TinkerTool System, OnyX 2.4.3, or Cocktail 5.1.1 that you can use for periodic maintenance such as removing old log files and archives, clearing caches, etc.
    For emergency repairs install the freeware utility Applejack.  If you cannot start up in OS X, you may be able to start in single-user mode from which you can run Applejack to do a whole set of repair and maintenance routines from the command line.  Note that AppleJack 1.5 is required for Leopard. AppleJack 1.6 is compatible with Snow Leopard. There is no confirmation that this version also works with Lion.
    When you install any new system software or updates be sure to repair the hard drive and permissions beforehand. I also recommend booting into safe mode before doing system software updates.
    Get an external Firewire drive at least equal in size to the internal hard drive and make (and maintain) a bootable clone/backup. You can make a bootable clone using the Restore option of Disk Utility. You can also make and maintain clones with good backup software. My personal recommendations are (order is not significant):
    Carbon Copy Cloner
    Data Backup
    Deja Vu
    SuperDuper!
    SyncTwoFolders
    Synk Pro
    Synk Standard
    Tri-Backup
    Visit The XLab FAQs and read the FAQs on maintenance, optimization, virus protection, and backup and restore.
    Additional suggestions will be found in Mac Maintenance Quick Assist.
    Referenced software can be found at CNet Downloads or MacUpdate.
    Be sure you have an adequate amount of RAM installed for the number of applications you run concurrently. Be sure you leave a minimum of 10% of the hard drive's capacity as free space.

  • Advice re best practice for managing the scan listener logs and list logs

    Hi friends,
    I've just started a job as a RAC dba administrator for some big 24*7 systems, I've never worked with clusterware and RAC.
    2 Space problems
    1) Very large listener_scan2.log in /u01/11.2.0/grid/log/diag/tnslsnr/<server name>/listener_scan2/trace folder
    2) Heaps of log_nnn.xml files in /u01/11.2.0/grid/log/diag/tnslsnr/<server name>/listener_scan2/alert folder (4Gb used up)
    Welcome advice on the best way to manage these in the short term (i.e. delete manually) and the recommended practice and safest way (adri maybe not sure how it works with scan listeners)
    Welcome advice and commands that could be used to safely clean these up and put a robust mechanism in place for logfile management in RAC and CLusterware systems.
    Finally should I be checking the log files in /u01/11.2.0/grid/log/diag/tnslsnr/<server name>/listener_scan2/alert regulalrly ?
    My experience with listener logs is that they are only looked at when there are major connectivity issues and on the whole are ignored.
    Thanks for your help,
    Cheers, Rob

    Have you had any issues that require them for investigative purposes? If not, just remove them. Are the logs required for some sort of audit process? If yes, gzip them to a location where you can use your OS tape backup policies to retain them for n-days. Once you remove an active file, it should recreate the file and continue without interruption.

  • What is the best practice for creating master pages and styles with translated text?

    I format translated text all the time for my company. I want to create a set of master pages and styles for each language and then import those styles into future translated documents. That way, the formatting can be done quickly and easily.
    What are the best practices for doing this? As a company this has been tried in the past, but without success. I'd like to know what other people are doing in this regard.
    Thank you!

    I create a master template that is usually void of content, with the exception I define as many of the paragraph styles I believe can/will be used with examples of their use in the body of the document--a style guide for that client. When beginning a new document for that client, I import those styles from the paragraph styles panel.
    Exception to this is when in a rush I begin documentation first, then begin new work. Then in the new work, I still pull in those defined paragraph and or object styles via their panels into the new work.
    There are times I need new styles. If they have broader applicability than a one-off instance or publication, then I open the style template for that client and import that style(s) from the publication containing the new style(s) and create example paragraphs and usage instructions.
    Take care, Mike

  • What is the best way to manage the file space given how much GB the video clips use up?

    I just downloaded a years worth of video off my HD camera.  I took up 125 GB on my hard drive.  What is the best way to manage this content going forward as there will be a lot more video to come in future years with small children!

    rule of thumb:
    SDef ~13GBs/h
    HDef ~30-40GBs/h ...
    latest investment of mine:
    ext.HDD, usb2 (=I have no trouble using usb2 for editing my short projects), 1.5TB=70€  ...
    ... which adds to ~4.2TB for my hobby, incl. back-up drives ...

  • What is the best way to manage the large FinalCutProX files

    We are continously running out of space on our desktop with the FinalCutProX files being so large. We have moved most of the files (events & projects) to an external server and working off the single video projects from our desktop. How are you working and what is you set up to manage these large files.?
    Computers we are using  new Mac Pro:
    Memory  16 GB 1600 MHz DDR3
    Processor  2.7 GHz Intel Core i7
    OS X 10.8.5
    What type of external servers?
    Are both the events and projects stored  together
    Do you keep the video you are currently working on your desktop?
    Thanks,
    Ang

    I have a MacBook Pro 17", 2.2 Ghz i7, 16 GB of RAM, 960 GB internal SSD (2 platters RAID 0). I have about 250 GB free on the internal drive, where I keep my current project. I also have a 3 TB external Thunderbolt desktop drive where I store the completed projects. I backup both drives to multiple drives that are stored away between daily backups.
    If my current project takes up a lot of space, I work it on the 3 TB external. I also have a 750 GB Seagate Momentus XT hybrid HD/SSD Thunderbolt bus powered portable just in case I need it in the field.

  • What is a best practice for managing a large amount of ever-changing hyperlinks?

    I am moving an 80+ page printed catalog online. We need to add hyperlinks to our Learning Management System courses to each reference of a class - there are 100s of them. I'm having difficulty understanding what the best practice is for consistent results when I need to go back and edit (which we will have to do regularly).
    These seem like my options:
    Link the actual text - sometimes when I go back to edit the link I can't find it in InDesign but can see it's there when I open up the PDF in Acrobat
    Draw an invisible box over the text and link it - this seems to work better but seems like an extra step
    Do all of the linking in Acrobat
    Am I missing anything?
    Here is the document in case anyone wants to see it so far. For the links that are in there, I used a combination of adding the links in InDesign then perfecting them using Acrobat (removing additional links or correcting others that I couldn't see in InDesign). This part of the process gives me anxiety each month we have to make edits. Nothing seems consistent. Maybe I'm missing something obvious?

    what exatly needs to be edited, the hyperlink or content or?

  • How is the best way to manage the stats table?

    Hello!
    I have the Integration 2.1 working with an Oracle 8.1.7 db. I noticed that the table
    STATS is growing pretty fast.
    How is the best way to manage this table?... I haven't found something related with
    this issue in the documentation, but at least I want to know how to safely delete
    records from this table.
    For example, if I know the minimal time I have to keep in the table, is quite simple
    to create a shell script and/or Oracle pl/sql job to trim the table.
    I hope somebody can help me!!!!
    Thank you!
    Ulises Sandoval

    Write an app people want to buy and rate highly.

  • What is the best practice to manage versions in XI?

    Hi!
    Is there any <b>good</b> “best practice” ways to manage versions in XI.
    Have a challenging scenario with many legacy systems and many interfaces per legacy system.
    Should I put all the different interfaces for one legacy system under one namespace in the Integration Repository or should I make one own namespace for each interface.
    Or is there other approaches, that I should consider to get a environment that is “easily” maintained.
    br.samuli

    Hi,
    In our project we have defined our own naming conventions/namespaces.
    For instance, we have agreed that we will group all interfaces (from all offices worldwide and different projects) into one single product version.
    This global custom product will in turn be divided into different SWC (Software Components) and SWCV (Software Component versions).
    So for each business scenario we will create separate SWC's and when necessary create new versions of these SWC's.
    This means that each SWC should contain all the required objects to support a complete integration scenario i.e. inbound/outbound-interfaces, data types, msg types, business scenario's etc...
    Regards,
    Rob.

  • What is the best practice to deploy the SharePoint site from test to production environment?

    We are beginning to start a new SharePoint 2010 and 2013 development projects, soon developing new features, lists, workflows, customizations to the SharePoint site, customization to list forms and would like to put good practice (that will help in deployment)
    in place before going ahead with development.
    What is the best way to go about deploying my site from Development to Production?
    I am using Visual Studio 2012 and also have Designer 2013...
    I have already read that this can be done through powershell, also through visual studio and also via designer. But at this point I am confused as to which are best practices specifically for lists, configurations; workflows; site customizations; Visual studio
    development features; customization to list forms etc. You can also provide me reference to links/ebook covering this topic.
    Thanks in advance for any help.

    Hi Nachiket,
    You can follow below approach where the environments has been built in similar fashion
    http://thesharepointfarm.com/sharepoint-test-environments/
    if you have less data then you can use  http://spdeploymentwizard.codeplex.com/
    http://social.technet.microsoft.com/Forums/sharepoint/en-US/b0bdb2ec-4005-441a-a233-7194e4fef7f7/best-way-to-replicate-production-sitecolletion-to-test-environment?forum=sharepointadminprevious
    For custom solutions like workflows etc you can always build the WSP packages and deploy across the environments using powershell scripts.
    Hope this helps.
    My Blog- http://www.sharepoint-journey.com|
    If a post answers your question, please click Mark As Answer on that post and Vote as Helpful
    Hi, can you answer me specifically with regards to the foll:-
    lists
    configurations
    workflows
    site customizations like changes to css/masterpage
    Visual studio webparts
    customization to list forms
    Thanks.

  • What are the best practices for using the enhancement framework?

    Hello enhancement framework experts,
    Recently, my company upgraded to SAP NW 7.1 EhP6.  This presents us with the capability to use the enhancement framework.
    A couple of senior programmers were asked to deliver a guideline for use of the framework.  They published the following statement:
    "SAP does not guarantee the validity of the enhancement points in future releases/versions. As a result, any implemented enhancement points may require significant work during upgrades. So, enhancement points should essentially be used as an alternative to core modifications, which is a rare scenario.".
    I am looking for confirmation or contradiction to the statement  "SAP does not guarantee the validity of enhancement points in future releases/versions..." .  Is this a true statement for both implicit and explicit enhancement points?
    Is the impact of activated explicit and implicit enhancements much greater to an SAP upgrade than BAdi's and user exits?
    Is there any SAP published guidelines/best practices for use of the enhancement framework?
    Thank you,
    Kimberly
    Edited by: Kimberly Carmack on Aug 11, 2011 5:31 PM

    Found an article that answers this question quite well:
    [How to Get the Most From the Enhancement and Switch Framework as a Customer or Partner - Tips from the Experts|http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/c0f0373e-a915-2e10-6e88-d4de0c725ab3]
    Thank you Thomas Weiss!

  • What is the best way to manage the two mailbox situation?

    My iPhone is working great. When I used a Treo, it seems like the use of Gmail was a little smoother. I got all the mail on the Treo, and all the mail on the home computer. On the Treo, there is a 'select all' feature that made it easy to clean out mail and move it to trash, but it looks like there's nothing like that on the iPhone. Did I miss it?
    It seems like I'm not getting some of the mail I read first on my iPhone on the home computer. Is there a 'leave mail on the server' setting? I can't find it.
    How are people managing this?
    What do you do with the mail on the iPhone? Delete it? Leave it there and hope your iPhone doesn't fill up with mail and eventually explode?

    As with any other email client, there is an account preference setting for removing messages from the server when downloaded for a POP account.
    Are you accessing your Gmail account as an IMAP or POP account with the iPhone's Mail client, and the same question with the email client on your computer used for accessing the account?
    Using the Gmail account preset on the iPhone creates the account as an IMAP account. There is no remove messages from the server when downloaded setting with an IMAP account.

  • What is the best practice to replicate the change of an object name in HANA AT/AN view where there are dependent views in it?

    Dear All HANA Studio Experts,
    I'm using SAP HANA Studio Version: 1.0.68 in our current landscape and looking for a expert and intelligent solution for a problem that we have encountered recently.
    We have created multiple AT views, AN views and created CA views on top of these AN views to cater different LOB related to our organization. We have also designed BO universe on top of the CA views from HANA and finally created webi reports out of these universes.
    In a particular scenario, we have one AT view which is been used by 10 different AN views.
    Now we had 3 Requirements like below:
    1. To add a new field in this AT view,
    2. To update an existing field in this AT view and
    3. To remove a field from this AT view which is not required anymore.
    Problems Encountered:
    Req 1. When we added a new field (e.g. XX_ADD) in the AT view and activate it, its throwing warning that there are dependent objects on this AT view. However no error thrown and it leads to re-work to check if this new field is available in all those dependent views and re-deployed those all.
    Question: How to ensure if this addition of object will not impact the rest of the dependent views?
    Req 2. When we updated an existing field (e.g. XX_UPDATE_NEW) in the AT view and activate it, its throwing warning that there are dependent objects on this AT view. However the BO reports built on top of these different dependent AN views became inoperative.
    Question: How to ensure if we need to update the name of an existing object it will not impact the rest of the dependent AN views?
    Req 3. Finally we have removed one object from the existing fields as it is not required. Again we faced warnings that there are dependent AN views on this AT view.
    Question: How to ensure if we need to delete an existing object, it will automatically reflect in all dependent AN views?
    Please suggest, if it is a good practice to create separate AT,AN and CA views for each of the LOB's and avoiding any dependent views (hence gradually the total no. of logical objects will be increase) or to reuse the AT/AN views whenever possible and in case of any updation/deletion we need to do a huge rework to check and redeploy the dependent objects again and again?
    Thanks for reading this detailed explanation and please let me know in case you need any other inputs....don't hesitate to share your expert thoughts on this.

    Hi Avishek,
    Are you using Development or Modeller Prespective?
    Avishek Chakraborty wrote:
    Now we had 3 Requirements like below:
    1. To add a new field in this AT view,
    2. To update an existing field in this AT view and
    3. To remove a field from this AT view which is not required anymore.
    1) Adding a new Field in the AT View and after activating the view. Then you have to Refresh/Check out the dependent views to see if it is reflecting.
    2)  We cannot update the existing field in the AT view i.e like name change if it is being already used in joins in any of the dependent views. It will not allow you to activate
    3) To remove the field which is not being used any where, we can make the changes in AT View and Refresh/Check out to see if the dependent views are reflecting the change
    Regards,
    Krishna Tangudu

Maybe you are looking for