ISE 1.2 VM file system recommendation

Hi,
According to http://www.cisco.com/en/US/docs/security/ise/1.2/installation_guide/ise_vmware.html#wp1056074, tabl 4-1 discusses storage requirement for ISE 1.2 in VM environment.  It recommends VMFS.
What are the implications when using NFS instead?  Is this just a recommendation or an actual requirement?
At the moment, we use Netapp array which uses NFS for all vApps.  It will be difficult to justify a creation an additional FC HBA just for this one vApp.  Please explain.
TIA,
Byung

If you refer to
http://www.cisco.com/en/US/docs/security/ise/1.2/installation_guide/ise_vmware.html
It says :
Storage
•File System—VMFS
We recommend that you use VMFS for storage. Other storage protocols are not tested and might result in some file system errors.
•Internal Storage—SCSI/SAS
•External Storage—iSCSI/SAN
We do not recommend the use of NFS storage.

Similar Messages

  • File system recommendation in SAP -AIX

    Dear All ,
    Pls check my current file system of PROD .
    Filesystem    GB blocks      Used      Free %Used Mounted on
    /dev/prddata1lv    120.00     66.61     53.39   56% /oracle/IRP/sapdata1
    /dev/prddata2lv    120.00     94.24     25.76   79% /oracle/IRP/sapdata2
    /dev/prddata3lv    120.00     74.55     45.45   63% /oracle/IRP/sapdata3
    /dev/prddata4lv    120.00     89.14     30.86   75% /oracle/IRP/sapdata4
    1, How much Space is recommended to increase the total space of the sapdata1..4 file system in SAP.
    Currently we have assigned 120GB to each datafile. if its any recommendation of sapdata1,2,3,4 total space , how can proceed for further ?
    ie ---> is there any limitation to increase the sapdata file system in SAP .
    example sapdata1,...4 ---> max limit to increase is 200 GB like that .....
    2,secondly only sapdata4 is keep on growing and rest of the file system are not much .
    Kindly suggets
    Edited by: satheesh0812 on Dec 20, 2010 7:13 AM

    Hi,
    Please identify what are the activities that were taking place which was consuming your space so rapidly.
    Analyze if the same activities are going to be performed in future, if so for how many months.
    Based on the comparision of present growth trend you can predict future trend and decide. Also analyze which table/tablespace is growing fast and what were its related data files and where are they located, you need to increase that file system to high level rather increasing all file systems sapdata1...4.
    Based on gathered inputs you can easily predict future growth trends and accordingly increase your size with additional space of 10%.
    You can increase your file system there is no limitation such, but you will face performance related problems. For this you need to create a separate tablespace and move the heavily growing table data files into that table space with proper index build.
    Regards.
    Edited by: Sita Rr Uppalapati on Dec 20, 2010 3:20 PM

  • Recommended File System - 3.5 TB

    I'm about to build a large file server (3.5 TB):
      - 8 * 500GB Sata drives
      - Raid 5 (hardware)
    The server will be primarly storing cd/dvd images and large graphics files.
    I'm just starting my research on different file system.  Any recommendations?
    Thanks,
    Chris....

    I was about to test a couple of different file systems and ran into a problem partitioning the drive.
    To recap I am using 8 * 500GB WD Sata drives on a 3ware 9500 Card.  I have created one large Raid 5 array (in the 9500 BIOS program).
    The 3ware instructions tell me to use parted to partition the drive (since it is large than 2TB).  However, parted does not support XFS...so I'm looking for another way to partition the drive.
    I used fdisk.  When I start fdisk /dev/sda, it informs me that I have to set the cylinders before I can perform any actions.  I set the cylinders to the maximum fdisk will allow (some very large number).  I then created a partition using 1 as the start cylinder and 3500GB as the stop (which is the size of my RAID array).
    I'm a little concered about the cylinder size I put in.  I know it is wrong, since I don't have any idea what the cylinder size should be.  Not sure cylinder size really makes sense across multiple disks though.
    Also, every time I start fdisk (even to just verify the partition table), it asks me to input the cylinder size.
    Any help/insights into how I can correctly partition the raid array would be appreciated.
    Thanks,
    Chris....

  • SAP GoLive : File System Response Times and Online Redologs design

    Hello,
    A SAP Going Live Verification session has just been performed on our SAP Production environnement.
    SAP ECC6
    Oracle 10.2.0.2
    Solaris 10
    As usual, we received database configuration instructions, but I'm a little bit skeptical about two of them :
    1/
    We have been told that our file system read response times "do not meet the standard requirements"
    The following datafile has ben considered having a too high average read time per block.
    File name -Blocks read  -  Avg. read time (ms)  -Total read time per datafile (ms)
    /oracle/PMA/sapdata5/sr3700_10/sr3700.data10          67534                         23                               1553282
    I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    2/
    We have been asked  to increase the size of the online redo logs which are already quite large (54Mb).
    Actually we have BW loading that generates "Chekpoint not comlete" message every night.
    I've read in sap note 79341 that :
    "The disadvantage of big redo log files is the lower checkpoint frequency and the longer time Oracle needs for an instance recovery."
    Frankly, I have problems undertanding this sentence.
    Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right ?
    But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    Thank you.
    Any useful help would be appreciated.

    Hello
    >> I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    The recommended ("standard") values are published at the end of sapnote #322896.
    23 ms seems really a little bit high to me - for example we have round about 4 to 6 ms on our productive system (with SAN storage).
    >> Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right?
    Correct.
    >> But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    A checkpoint is occured on every logswitch (of the online redologfiles). On a checkpoint event the following 3 things are happening in an oracle database:
    Every dirty block in the buffer cache is written down to the datafiles
    The latest SCN is written (updated) into the datafile header
    The latest SCN is also written to the controlfiles
    If your redologfiles are larger ... checkpoints are not happening so often and in this case the dirty buffers are not written down to the datafiles (in the case of no free space in the buffer cache is needed). So if your instance crashes you need to apply more redologs to the datafiles to be in a consistent state (roll forward). If you have smaller redologfiles more log switches are occured and so the SCNs in the data file headers (and the corresponding data) are closer to the newest SCN -> ergo the recovery is faster.
    But this concept does not really fit the reality because of oracle implements some algorithm to reduce the workload for the DBWR in the case of a checkpoint.
    There are also several parameters (depends on the oracle version) which control that a required recovery time is kept. (for example FAST_START_MTTR_TARGET)
    Regards
    Stefan

  • 10G Database install on Linux 64bit File system Vs. Automatic Storage

    Hi,
    I'm installing Oracle 10G on a Dell 2950 server with an Intel 64 bit processor. The operating system is Red Hat Linux 4. It has a hardware RAID.
    I've downloaded the Install Guide but I have a question about installing.
    I'm confused about the File system vs. Automatic Storage management option. I will be installing on the local system, I will not be using a NAS or SAN.
    Under Database Storage Options, the guide says that "Oracle recommends that the file system you choose be separate from the file systems used by the operating system or the Oracle software.".
    1. Do I need to do that since I'm already using hardware RAID??
    2. Which way is recommended / what do most people do: File system or Automatic Storage managment??
    3. For Automatic Storage Management I read that I'd have to create an ASM Disk group that can consist of "a multiple disk device such as a RAID storage array or logical volume. However, in most cases disk groups consist of one or more individual physical disks". Do I need to reconfigure my partitions ??
    I just need some input on what I should do since this is my first time installing Oracle on Linux.
    Thank you.

    Besides documentation there's a step-by-step guide :
    http://www.oracle.com/technology/pub/articles/smiley_10gdb_install.html#asm
    Many questions should be answered here.
    Werner

  • Difference between ASM Disk Group, ADVM Volume and ACFS File system

    Q1. What is the difference between an ASM Disk Group and an ADVM Volume ?
    To my mind, an ASM Disk Group is effectively a logical volume for Database files ( including FRA files ).
    11gR2 seems to have introduced the concepts of ADVM volumes and ACFS File Systems.
    An 11gR2 ASM Disk Group can contain :
    ASM Disks
    ADVM volumes
    ACFS file systems
    Q2. ADVM volumes appear to be dynamic volumes.
    However is this therefore not effectively layering a logical volume ( the ADVM volume ) beneath an ASM Disk Group ( conceptually a logical volume as well ) ?
    Worse still if you have left ASM Disk Group Redundancy to the hardware RAID / SAN level ( as Oracle recommend ), you could effectively have 3 layers of logical disk ? ( ASM on top of ADVM on top of RAID/SAN ) ?
    Q3. if it is 2 layers of logical disk ( i.e. ASM on top of ADVM ), what makes this better than 2 layers using a 3rd party volume manager ( eg ASM on top of 3rd party LVM ) - something Oracle encourages against ?
    Q4. ACFS File systems, seem to be clustered file systems for non database files including ORACLE_HOMEs, application exe's etc ( but NOT GRID_HOME, OS root, OCR's or Voting disks )
    Can you create / modify ACFS file systems using ASM.
    The oracle toplogy diagram for ASM in the 11gR2 ASM Admin guide, shows ACFS as part of ASM. I am not sure from this if ACFS is part of ASM or ASM sits on top of ACFS ?
    Q5. Connected to Q4. there seems to be a number of different ways, ACFS file systems can be created ? Which of the below are valid methods ?
    through ASM ?
    through native OS file system creation ?
    through OEM ?
    through acfsutil ?
    my head is exploding
    Any help and clarification greatly appreciated
    Jim

    Q1 - ADVM volume is a type of special file created in the ASM DG.  Once created, it creates a block device on the OS itself that can be used just like any other block device.  http://docs.oracle.com/cd/E16655_01/server.121/e17612/asmfilesystem.htm#OSTMG30000
    Q2 - the asm disk group is a disk group, not really a logical volume.  It combines attributes of both when used for database purposes, as the database and certain other applications know how to talk "ASM" protocol.  However, you won't find any general purpose applications that can do so.  In addition, some customers prefer to deal directly with file systems and volume devices, which ADVM is made to do.  In your way of thinking, you could have 3 layers of logical disk, but each of them provides different attributes and characteristics.  This is not a bad thing though, as each has a slightly different focus - os file system\device, database specific, and storage centric.
    Q3 - ADVM is specifically developed to extend the characteristics of ASM for use by general OS applications.  It understands the database performance characteristics and is tuned to work well in that situation.  Because it is developed in house, it takes advantage of the ASM design model.  Additionally, rather than having to contact multiple vendors for support, your support is limited to calling Oracle, a one-stop shop.
    Q4 - You can create and modify ACFS file systems using command line tools and ASMCA.  Creating and modifying logical volumes happens through SQL(ASM), asmcmd, and ASMCA.  EM can also be used for both items.  ACFS sits on top of ADVM, which is a file in an ASM disk group.  ACFS is aware of the characteristics of ASM\ADVM volumes, and tunes it's IO to make best use of those characteristics. 
    Q5 - several ways:
    1) Connect to ASM with SQL, use 'alter diskgroup add volume' as Mihael points out.  This creates an ADVM volume.  Then, format the volume using 'mkfs' (*nix) or acfsformat (windows).
    2) Use ASMCA - A gui to create a volume and format a file system.  Probably the easiest if your head is exploding.
    3) Use 'asmcmd' to create a volume, and 'mkfs' to format the ACFS file system.
    Here is information on ASMCA, with examples:
    http://docs.oracle.com/cd/E16655_01/server.121/e17612/asmca_acfs.htm#OSTMG94348
    Information on command line tools, with examples:
    Basic Steps to Manage Oracle ACFS Systems

  • What company can retrieve photos from a micro sd card? Another casualty of SD Card blank or has unsupported file system

         I have a Droid X with a sandisk 16gb micro sdhc class 4 installed.  This card has been in use for about a year and wasnt even half full.  I was playing music on thanksgiving and all of a sudden the music stopped.  I thought hmm, It must be doing one of its many reboots.  It wasnt shut off but was on my home screen and the music player had closed.  I went to my music and the songs were still listed.  I tried to play one and got the error message that i  cannot play this file.  I tried again and the songs disappeared and up came "SD Card blank or has unsupported file system" on the top bar.  I tried to view files on the card on my phone and on my computer by using an sd adapter with no luck.  I popped another card into my phone and it worked with no error message and i could mount and unmount the card and view the picture I took when viewing the new card on my computer.  I have a 9 year old daughter who suffers from multiple chronic illnesses and have probably 200 photos on my card of her including an extensive surgery and recovery earlier this year.  If a professional can retrieve my photos, I would like to know more.  Im coming up on my contracts end and this has me ready to zip up a life vest and get on another boat altogether. I have seen hundreds post about this on boards but I hope this can be resolved.  Thanks in advance.

        Hi Matt73,
    I use my phone to store all my precious moments too and I can just imagine how frustrating it must be to have memory card issues like the ones outlined in your post. I see that you have tried different things to retrieve your files from your memory card to no available. It appears that your SD memory card somehow became corrupted and most likely the data in it is now lost (sorry). While not a common issue, there is always a possibility for memory cards and any other memory device to fail. SanDisk Memory cards carry 5 year warranty, to get a replacement please contact SanDisk at 866-726-3475. Their hours of operation are 8:00 a.m. - 7:00 p.m. PT Monday to Friday or by email at [email protected] at your convenience.
    I recommend backing up your media (specially pictures) on a regular basis to your computer or to an online service to protect those special memories, I personally have my phone setup to automatically save the pictures I take to my Google  account and even if lost my phone I have a copy of all my pictures online. There are several website that allow you to sync your phone to upload pictures automatically. Please consider a similar option going forward.
    On a personal note, I hope that your little girl's health improves that all those chronic illnesses are kept under control by your physician; I have 2 kids and I Know what is like when they don't feel well. Have a great day!
    AntonioC_VZWSupport
    Follow us on Twitter at www.twitter.com/VZWSupport

  • How do I determine if my file system is case-sensitive before I buy Photoshop CS5?

    How do I determine if my file system is case-sensitive before I buy Photoshop CS5? I currently have Mac OSX 10.5.8 with an Intel Core 2 Duo Processor and 4GB of Memory.. Adobe says the following requirements are needed..
    2GB of available hard-disk space for installation; additional free space required during installation (cannot install on a volume that uses a case-sensitive file system or on removable flash storage devices)
    1024x768 display (1280x800 recommended) with qualified hardware-accelerated OpenGL graphics card, 16-bit color, and 256MB of VRAM
    Some GPU-accelerated features require graphics support for Shader Model 3.0 and OpenGL 2.0
    I don't know if I have a "hardware-accelerated OpenGL graphics card" or "256MB or VRAM".. I've never heard of "VRAM" before and the last bullet points is just as confusing. The last thing I want to do is buy this program and discover my Mac Book is just a tad bit to old for it, I've already done that once with a video game.

    Disk Utility highlighting the drive or partition & clicking the blue Info icon may report it, but since I don't have Case Sensitive I can't tell for sure.
    A simple test would be to save a small txt doc as...
    test
    Then save to the place same as...
    TEST
    If both show on the Disk, then it is, if it asks to replace test with TEST then it isn't.

  • Target file system layout in homogeneous DB copy

    Gurus:
    We have a target system whose file system layout is totally different from the source system in that:
    1) the number of the file system is less than that in the source;
    2) the size of each targte file system is also different from the source system.
    BTW way the target has a DB running.
    Can we go ahead to do the import without doing anything on  above differences? i.e just go ahead to overwrite the target
    DB?
    Oracle on Unix.
    Thanks a lot!

    Hi,
    its better and recommemded by SAP if you have the same DB instance SID and SAPSID.
    If the target system has different SID than the source that will be more work to do for you. its makes things easyer if you have the same name as the source system.
    You have two different file system and different SID. So its better to minimize those things if you can.
    I think in this step where you are, I would prefer to deinstall oracle on the target system and install it again with the SAP recommendations for SAP Software which you will find on the installation guide. Because you wrote that oracle is already installed on the target system but without SAP. This means the installation was not for SAP and its different to SAP recommended..
    Chooce also Oracle DB instance name and SAPSID the same name.
    Good luck
    HaneAtik

  • CS4 and Case Sensitive File Systems

    Does anyone know if CS4 will support case sensitive file systems?

    While I do wish all software companies including Apple completely
    supported the case-sensitive file system. Many don't. I cannot pull up
    an article, but I remember reading that Apple only recommended the use
    of the case-sensitive file system for OS X server and if memory serves
    even for that it only recommended it for the attached hard drives.
    We can lament the lack of support but if you really want to change the
    status quo for Adobe products then the only useful way to do so is:
    ...Mike

  • Determining file system case-sensitivity

    Is there any way that isn't a hack to determine whether the specific platform that the JVM is running on uses case-sensitive file names or not?
    Thanks,
    Jason

    > Beware that the use of case-sensitive file names is not platform dependant but rather file system dependent.
    I didn't know that. That's good to know. So, would you just recommend creating a temporary file in the path in question, then checking to see if it exists using the same name with different case, then deleting the temporary file?
    That's what I was trying to not do... I was looking for a more solid way. But since it isn't platform dependent, is it the only way? Or is there still a better method?
    Thanks,
    Jason

  • [SOLVED] Ring file-system for a surveillance server?

    In my home, I have a (almost) dedicated (Arch-32) computer to use as a surveillance server. It continuously fetches video from a IP camera using motion and saving various GB of images and videos. Obviously, I have to manually delete older videos when the disk is getting full, but it would be better to have some kind of file-system that overrides the older files with the newer ones when it gets full.
    Is there anything like that? Any recommendations?
    Thanks!
    PS: If your suggestion/solution could only delete videos (not images), better.
    Last edited by sironitomas (2011-03-12 04:06:12)

    limaunion wrote:
    One possiblity is to schedule a daily cron job with something like this:
    find /home/videos/ -mindepth 2 -cmin +2880 -exec rm -rf {} \;
    In this case cmin=2880 (in minutes) equals 48 hours; 'man find' is your friend
    Regards (from baires)
    LU
    Gracias! I didn't know about find's cmin, it is a great option. Exec gave me some problems so I ended up using more than one command. This is the script actually:
    find /home/user/videos -cmin +4320 > temp
    rm `cat temp | grep avi`
    rm temp
    The only thing bad about it, is that when find doesn't produce any output, rm shows a error message because there are no arguments. I know that's not the best way, but it works.
    Any recommendation is accepted.
    Thanks!
    Last edited by sironitomas (2011-03-12 04:05:38)

  • Question about a file system storage option for RAC on 10g

    Hello everyone,
    I am in the beginning of connecting our storage and switches, and building RAC on them but there is a little argument between our specialists.
    We have two database servers(10g with OEL 5) to be clustered and two visible disk groups to each of those nodes. So question is can we choose only one disk group as a shared storage leaving the rest one as the redundant copy during database a creation window while installing the database.  Because some of us argue that oracle database has a built-in capability to decide on what level of RAID we store our data.
    Thank you for your help.

    "some of us argue that oracle database has a built-in capability to decide on what level of RAID we store our data". 
    That statement is not true.  Oracle has optional multiplexing for control files, redo logs, and archive logs but this is not enabled by default and Oracle will not automatically enable it.  If you want redundancy of tables, indexes, temp, and undo you must provide this because Oracle does not offer it standard or as an option.  You can achieve redundancy with RAID at the array level, or host based mirroring (like ASM redundancy groups or Linux mdadm).  This can also depend on your file system because, I think, OCFS2 does not support host based mirroring (so you cannot use mdadm or lvm to mirror the storage if you are using OCFS2).
    Redundancy is not required, but it is recommended if you are using hard disks because they are prone to failures.  You can configure RAID 10 across all disks in the array and present this as one big LUN to the database server.  If you have two storage arrays and you want to mirror the data across the two arrays, then present all of the devices as JBOD and use Linux mdadm to create your RAID group.
    RAC requires shared storage.  Maybe you have a NAS or SAN device, and you will present LUNs to the Oracle database servers.  That is no problem.  The problem is making those LUNs usable by Oracle RAC.  When I used Oracle 10g RAC, I used the Linux raw device facility to manage these LUNs and make them ready for Oracle RAC.  However, raw has been desupported.  Today I would use either ASM or OCFS2.  This has nothing to do with redundancy, this is just because you are using RAC.

  • Warming up File System Cache for BDB Performance

    Hi,
    We are using BDB DPL - JE package for our application.
    With our current machine configuration, we have
    1) 64 GB RAM
    2) 40-50 GB -- Berkley DB Data Size
    To warm up File System Cache, we cat the .jdb files to /dev/null (To minimize the disk access)
    e.g
         // Read all jdb files in the directory
         p = Runtime.getRuntime().exec("cat " + dirPath + "*.jdb >/dev/null 2>&1");
    Our application checks if new data is available every 15 minutes, If new Data is available then it clears all old reference and loads new data along with Cat *.jdb > /dev/null
    I would like to know that if something like this can be done to improve the BDB Read performance, if not is there any better method to Warm Up File System Cache ?
    Thanks,

    We've done a lot of performance testing with how to best utilize memory to maximize BDB performance.
    You'll get the best and most predictable performance by having everything in the DB cache. If the on-disk size of 40-50GB that you mention includes the default 50% utilization, then it should be able to fit. I probably wouldn't use a JVM larger than 56GB and a database cache percentage larger than 80%. But this depends a lot on the size of the keys and values in the database. The larger the keys and values, the closer the DB cache size will be to the on disk size. The preload option that Charles points out can pull everything into the cache to get to peak performance as soon as possible, but depending on your disk subsystem this still might take 30+ minutes.
    If everything does not fit in the DB cache, then your best bet is to devote as much memory as possible to the file system cache. You'll still need a large enough database cache to store the internal nodes of the btree databases. For our application and a dataset of this size, this would mean a JVM of about 5GB and a database cache percentage around 50%.
    I would also experiment with using CacheMode.EVICT_LN or even CacheMode.EVICT_BIN to reduce the presure on the garbage collector. If you have something in the file system cache, you'll get reasonably fast access to it (maybe 25-50% as fast as if it's in the database cache whereas pulling it from disk is 1-5% as fast), so unless you have very high locality between requests you might not want to put it into the database cache. What we found was that data was pulled in from disk, put into the DB cache, stayed there long enough to be promoted during GC to the old generation, and then it was evicted from the DB cache. This long-lived garbage put a lot of strain on the garbage collector, and led to very high stop-the-world GC times. If your application doesn't have latency requirements, then this might not matter as much to you. By setting the cache mode for a database to CacheMode.EVICT_LN, you effectively tell BDB to not to put the value or (leaf node = LN) into the cache.
    Relying on the file system cache is more unpredictable unless you control everything else that happens on the system since it's easy for parts of the BDB database to get evicted. To keep this from happening, I would recommend reading the files more frequently than every 15 minutes. If the files are in the file system cache, then cat'ing them should be fast. (During one test we ran, "cat *.jdb > /dev/null" took 1 minute when the files were on disk, but only 8 seconds when they were in the file system cache.) And if the files are not all in the file system cache, then you want to get them there sooner rather than later. By the way, if you're using Linux, then you can use "echo 1 > /proc/sys/vm/drop_caches" to clear out the file system cache. This might come in handy during testing. Something else to watch out for with ZFS on Solaris is that sequentially reading a large file might not pull it into the file system cache. To prevent the cache from being polluted, it assumes that sequentially reading through a large file doesn't imply that you're going to do a lot of random reads in that file later, so "cat *.jdb > /dev/null" might not pull the files into the ZFS cache.
    That sums up our experience with using the file system cache for BDB data, but I don't know how much of it will translate to your application.

  • Every boot requires fsck. Makes file system read-only

    Suddenly when I booted my Arch I got a message saying that there was a corruption in my /home and /var directories. The same message suggested to pass fsck as solution. The booting process hanged there without entering to the emergency login. I took a live cd an executed a fsck in all my ext4 file system and “repaired” the problem.
    When I could log into the system if, for example, I want to write a file o do any operation (sudo for example) I get the message:
    sudo: unable to open /var/db/sudo/lola/0: Read-only file system
    If I reboot the PC again, the cycle begins again, this is: errors, asking for fsck, then, read-only FS.
    Some info that you may find useful:
    fstab
    /dev/sda3 / ext4 defaults 0 1
    /dev/sda5 /var ext4 defaults 0 1
    /dev/sda6 swap swap defaults 0 0
    /dev/sda7 /home ext4 defaults,user_xattr 0 1
    df -h
    Filesystem Size Used Avail Use% Mounted on
    rootfs 15G 7.2G 6.6G 52% /
    dev 1.9G 0 1.9G 0% /dev
    run 1.9G 740K 1.9G 1% /run
    /dev/sda3 15G 7.2G 6.6G 52% /
    tmpfs 1.9G 536K 1.9G 1% /dev/shm
    tmpfs 1.9G 0 1.9G 0% /sys/fs/cgroup
    tmpfs 1.9G 80K 1.9G 1% /tmp
    /dev/sda7 80G 39G 37G 52% /home
    /dev/sda5 6.8G 3.7G 2.8G 58% /var
    /dev/sda2 358G 334G 25G 94% /media/windows
    systemctl
    UNIT LOAD ACTIVE SUB DESCRIPTION
    proc-sys-fs-binfmt_misc.automount loaded active running Arbitrary Executable File Formats File System Au
    sys-devices-pci...000:00:1b.0-sound-card0.device loaded active plugged /sys/devices/pci0000:00/0000:00:1b.0/sound/card0
    sys-devices-pci...0-0000:02:00.0-net-eth0.device loaded active plugged /sys/devices/pci0000:00/0000:00:1c.0/0000:02:00.
    sys-devices-pci...-0000:03:00.0-net-wlan0.device loaded active plugged /sys/devices/pci0000:00/0000:00:1c.1/0000:03:00.
    sys-devices-pci...-1:0:0:0-block-sda-sda1.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda2.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda3.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda4.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda5.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda6.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...-1:0:0:0-block-sda-sda7.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...1:0:0-1:0:0:0-block-sda.device loaded active plugged TOSHIBA_MK5055GSX
    sys-devices-pci...5:0:0-5:0:0:0-block-sr0.device loaded active plugged TSSTcorp_CDDVDW_TS-L633C
    sys-devices-platform-serial8250-tty-ttyS0.device loaded active plugged /sys/devices/platform/serial8250/tty/ttyS0
    sys-devices-platform-serial8250-tty-ttyS1.device loaded active plugged /sys/devices/platform/serial8250/tty/ttyS1
    tail -100 /var/log/errors.log
    Jan 23 14:43:52 localhost kernel: [ 283.449542] ata2.00: error: { UNC }
    Jan 23 14:43:52 localhost kernel: [ 283.451224] end_request: I/O error, dev sda, sector 513442820
    Jan 23 14:43:52 localhost kernel: [ 283.451228] Buffer I/O error on device sda2, logical block 63796096
    Jan 23 14:43:56 localhost kernel: [ 287.394421] ata2.00: exception Emask 0x0 SAct 0xf SErr 0x0 action 0x0
    Jan 23 14:43:56 localhost kernel: [ 287.394426] ata2.00: irq_stat 0x40000001
    Jan 23 14:43:56 localhost kernel: [ 287.394430] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 14:43:56 localhost kernel: [ 287.394435] ata2.00: cmd 60/08:00:00:84:9a/00:00:1e:00:00/40 tag 0 ncq 4096 in
    Jan 23 14:43:56 localhost kernel: [ 287.394435] res 41/40:08:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 14:43:56 localhost kernel: [ 287.394438] ata2.00: status: { DRDY ERR }
    Jan 23 14:43:56 localhost kernel: [ 287.394440] ata2.00: error: { UNC }
    Jan 23 14:43:56 localhost kernel: [ 287.394442] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 14:43:56 localhost kernel: [ 287.394447] ata2.00: cmd 60/00:08:1f:53:4d/01:00:2d:00:00/40 tag 1 ncq 131072 in
    Jan 23 14:43:56 localhost kernel: [ 287.394447] res 41/40:02:04:84:9a/00:00:1e:00:00/4e Emask 0x9 (media error)
    Jan 23 14:43:56 localhost kernel: [ 287.394449] ata2.00: status: { DRDY ERR }
    Jan 23 14:43:56 localhost kernel: [ 287.394451] ata2.00: error: { UNC }
    Jan 23 14:43:56 localhost kernel: [ 287.394453] ata2.00: failed command: WRITE FPDMA QUEUED
    Jan 23 14:43:56 localhost kernel: [ 287.394458] ata2.00: cmd 61/08:10:7f:6e:b7/00:00:30:00:00/40 tag 2 ncq 4096 out
    Jan 23 14:43:56 localhost kernel: [ 287.394458] res 41/40:02:04:84:9a/00:00:1e:00:00/4e Emask 0x9 (media error)
    Jan 23 14:43:56 localhost kernel: [ 287.394460] ata2.00: status: { DRDY ERR }
    Jan 23 14:43:56 localhost kernel: [ 287.394462] ata2.00: error: { UNC }
    Jan 23 14:43:56 localhost kernel: [ 287.394464] ata2.00: failed command: WRITE FPDMA QUEUED
    Jan 23 14:43:56 localhost kernel: [ 287.394468] ata2.00: cmd 61/90:18:37:42:31/00:00:35:00:00/40 tag 3 ncq 73728 out
    Jan 23 14:43:56 localhost kernel: [ 287.394468] res 41/40:02:04:84:9a/00:00:1e:00:00/4e Emask 0x9 (media error)
    Jan 23 14:43:56 localhost kernel: [ 287.394471] ata2.00: status: { DRDY ERR }
    Jan 23 14:43:56 localhost kernel: [ 287.394472] ata2.00: error: { UNC }
    Jan 23 14:43:56 localhost kernel: [ 287.396122] end_request: I/O error, dev sda, sector 513442820
    Jan 23 14:43:56 localhost kernel: [ 287.396125] Buffer I/O error on device sda2, logical block 63796096
    Jan 23 14:43:56 localhost kernel: [ 287.396179] end_request: I/O error, dev sda, sector 760042271
    Jan 23 14:43:56 localhost kernel: [ 287.396232] end_request: I/O error, dev sda, sector 817327743
    Jan 23 14:43:56 localhost kernel: [ 287.396235] Buffer I/O error on device sda7, logical block 1133175
    Jan 23 14:43:56 localhost kernel: [ 287.396282] end_request: I/O error, dev sda, sector 892420663
    Jan 23 14:44:01 localhost kernel: [ 291.317005] ata2.00: exception Emask 0x0 SAct 0x2 SErr 0x0 action 0x0
    Jan 23 14:44:01 localhost kernel: [ 291.317010] ata2.00: irq_stat 0x40000008
    Jan 23 14:44:01 localhost kernel: [ 291.317014] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 14:44:01 localhost kernel: [ 291.317019] ata2.00: cmd 60/08:08:00:84:9a/00:00:1e:00:00/40 tag 1 ncq 4096 in
    Jan 23 14:44:01 localhost kernel: [ 291.317019] res 41/40:08:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 14:44:01 localhost kernel: [ 291.317022] ata2.00: status: { DRDY ERR }
    Jan 23 14:44:01 localhost kernel: [ 291.317024] ata2.00: error: { UNC }
    Jan 23 14:44:01 localhost kernel: [ 291.318643] end_request: I/O error, dev sda, sector 513442820
    Jan 23 14:44:01 localhost kernel: [ 291.318647] Buffer I/O error on device sda2, logical block 63796096
    Jan 23 14:44:01 localhost kernel: [ 291.546162] Aborting journal on device sda7-8.
    Jan 23 14:54:20 localhost [ 7.837917] systemd[1]: Socket service syslog.service not loaded, refusing.
    Jan 23 14:54:20 localhost [ 7.838515] systemd[1]: Failed to listen on Syslog Socket.
    Jan 23 14:54:21 localhost kernel: [ 10.930157] [drm:intel_panel_setup_backlight] *ERROR* Failed to get maximum backlight value
    Jan 23 14:54:34 localhost dhcpcd[564]: dhcpcd not running
    Jan 23 14:54:35 localhost dhcpcd[649]: dhcpcd not running
    Jan 23 14:54:35 localhost dhcpcd[662]: dhcpcd not running
    Jan 23 14:58:04 localhost su: pam_unix(su:auth): conversation failed
    Jan 23 17:00:38 localhost systemd[1]: Failed to start Netcfg multi-profile daemon.
    Jan 23 17:00:39 localhost [ 7.790297] systemd[1]: Socket service syslog.service not loaded, refusing.
    Jan 23 17:00:39 localhost [ 7.790891] systemd[1]: Failed to listen on Syslog Socket.
    Jan 23 17:00:39 localhost kernel: [ 11.363432] [drm:intel_panel_setup_backlight] *ERROR* Failed to get maximum backlight value
    Jan 23 17:00:55 localhost dhcpcd[612]: dhcpcd not running
    Jan 23 17:00:55 localhost dhcpcd[618]: dhcpcd not running
    Jan 23 17:00:55 localhost dhcpcd[626]: dhcpcd not running
    Jan 23 17:01:23 localhost su: pam_unix(su:auth): conversation failed
    Jan 23 17:04:55 localhost kernel: [ 274.346016] ata2.00: exception Emask 0x0 SAct 0xf SErr 0x0 action 0x0
    Jan 23 17:04:55 localhost kernel: [ 274.346022] ata2.00: irq_stat 0x40000008
    Jan 23 17:04:55 localhost kernel: [ 274.346025] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 17:04:55 localhost kernel: [ 274.346031] ata2.00: cmd 60/70:08:e8:83:9a/00:00:1e:00:00/40 tag 1 ncq 57344 in
    Jan 23 17:04:55 localhost kernel: [ 274.346031] res 41/40:70:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 17:04:55 localhost kernel: [ 274.346034] ata2.00: status: { DRDY ERR }
    Jan 23 17:04:55 localhost kernel: [ 274.346035] ata2.00: error: { UNC }
    Jan 23 17:04:55 localhost kernel: [ 274.348151] end_request: I/O error, dev sda, sector 513442820
    Jan 23 17:04:55 localhost kernel: [ 274.348154] Buffer I/O error on device sda2, logical block 63796096
    Jan 23 17:04:55 localhost kernel: [ 274.348159] Buffer I/O error on device sda2, logical block 63796097
    Jan 23 17:04:55 localhost kernel: [ 274.348162] Buffer I/O error on device sda2, logical block 63796098
    Jan 23 17:04:55 localhost kernel: [ 274.348164] Buffer I/O error on device sda2, logical block 63796099
    Jan 23 17:04:55 localhost kernel: [ 274.348166] Buffer I/O error on device sda2, logical block 63796100
    Jan 23 17:04:55 localhost kernel: [ 274.348168] Buffer I/O error on device sda2, logical block 63796101
    Jan 23 17:04:55 localhost kernel: [ 274.348171] Buffer I/O error on device sda2, logical block 63796102
    Jan 23 17:04:55 localhost kernel: [ 274.348173] Buffer I/O error on device sda2, logical block 63796103
    Jan 23 17:04:55 localhost kernel: [ 274.348176] Buffer I/O error on device sda2, logical block 63796104
    Jan 23 17:04:55 localhost kernel: [ 274.348178] Buffer I/O error on device sda2, logical block 63796105
    Jan 23 17:04:59 localhost kernel: [ 278.257533] ata2.00: exception Emask 0x0 SAct 0x84 SErr 0x0 action 0x0
    Jan 23 17:04:59 localhost kernel: [ 278.257539] ata2.00: irq_stat 0x40000008
    Jan 23 17:04:59 localhost kernel: [ 278.257543] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 17:04:59 localhost kernel: [ 278.257549] ata2.00: cmd 60/08:38:00:84:9a/00:00:1e:00:00/40 tag 7 ncq 4096 in
    Jan 23 17:04:59 localhost kernel: [ 278.257549] res 41/40:08:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 17:04:59 localhost kernel: [ 278.257551] ata2.00: status: { DRDY ERR }
    Jan 23 17:04:59 localhost kernel: [ 278.257553] ata2.00: error: { UNC }
    Jan 23 17:04:59 localhost kernel: [ 278.259233] end_request: I/O error, dev sda, sector 513442820
    Jan 23 17:05:03 localhost kernel: [ 282.157929] ata2.00: exception Emask 0x0 SAct 0x4 SErr 0x0 action 0x0
    Jan 23 17:05:03 localhost kernel: [ 282.157934] ata2.00: irq_stat 0x40000008
    Jan 23 17:05:03 localhost kernel: [ 282.157938] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 17:05:03 localhost kernel: [ 282.157943] ata2.00: cmd 60/08:10:00:84:9a/00:00:1e:00:00/40 tag 2 ncq 4096 in
    Jan 23 17:05:03 localhost kernel: [ 282.157943] res 41/40:08:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 17:05:03 localhost kernel: [ 282.157946] ata2.00: status: { DRDY ERR }
    Jan 23 17:05:03 localhost kernel: [ 282.157947] ata2.00: error: { UNC }
    Jan 23 17:05:03 localhost kernel: [ 282.159582] end_request: I/O error, dev sda, sector 513442820
    Jan 23 17:05:03 localhost kernel: [ 282.159588] Buffer I/O error on device sda2, logical block 63796096
    Jan 23 17:05:07 localhost kernel: [ 286.069482] ata2.00: exception Emask 0x0 SAct 0x40 SErr 0x0 action 0x0
    Jan 23 17:05:07 localhost kernel: [ 286.069487] ata2.00: irq_stat 0x40000008
    Jan 23 17:05:07 localhost kernel: [ 286.069491] ata2.00: failed command: READ FPDMA QUEUED
    Jan 23 17:05:07 localhost kernel: [ 286.069496] ata2.00: cmd 60/08:30:00:84:9a/00:00:1e:00:00/40 tag 6 ncq 4096 in
    Jan 23 17:05:07 localhost kernel: [ 286.069496] res 41/40:08:04:84:9a/00:00:1e:00:00/6e Emask 0x409 (media error) <F>
    Jan 23 17:05:07 localhost kernel: [ 286.069499] ata2.00: status: { DRDY ERR }
    Jan 23 17:05:07 localhost kernel: [ 286.069501] ata2.00: error: { UNC }
    Jan 23 17:05:07 localhost kernel: [ 286.071955] end_request: I/O error, dev sda, sector 513442820
    Jan 23 17:05:07 localhost kernel: [ 286.071959] Buffer I/O error on device sda2, logical block 63796096
    Waiting for your help.
    Thank you.

    oh god, if my HD dies I will kill myself.
    I did the SMART, short and long (following the ArchWiki) and got (will post the long as soon it ends):
    === START OF READ SMART DATA SECTION ===
    SMART Self-test log structure revision number 1
    Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
    # 1 Short offline Completed without error 00% 7967 -
    With more details:
    smartctl -a /dev/sda
    smartctl 6.0 2012-10-10 r3643 [x86_64-linux-3.6.9-1-ARCH] (local build)
    Copyright (C) 2002-12, Bruce Allen, Christian Franke, www.smartmontools.org
    === START OF INFORMATION SECTION ===
    Model Family: Toshiba 2.5" HDD MK..55GSX
    Device Model: TOSHIBA MK5055GSX
    Serial Number: 897JC07RT
    LU WWN Device Id: 5 000039 1f3c01f05
    Firmware Version: FG001M
    User Capacity: 500,107,862,016 bytes [500 GB]
    Sector Size: 512 bytes logical/physical
    Device is: In smartctl database [for details use: -P show]
    ATA Version is: ATA8-ACS (minor revision not indicated)
    SATA Version is: SATA 2.6, 3.0 Gb/s
    Local Time is: Wed Jan 23 13:46:24 2013 PYST
    SMART support is: Available - device has SMART capability.
    SMART support is: Enabled
    === START OF READ SMART DATA SECTION ===
    SMART overall-health self-assessment test result: PASSED
    General SMART Values:
    Offline data collection status: (0x00) Offline data collection activity
    was never started.
    Auto Offline Data Collection: Disabled.
    Self-test execution status: ( 249) Self-test routine in progress...
    90% of test remaining.
    Total time to complete Offline
    data collection: ( 120) seconds.
    Offline data collection
    capabilities: (0x5b) SMART execute Offline immediate.
    Auto Offline data collection on/off support.
    Suspend Offline collection upon new
    command.
    Offline surface scan supported.
    Self-test supported.
    No Conveyance Self-test supported.
    Selective Self-test supported.
    SMART capabilities: (0x0003) Saves SMART data before entering
    power-saving mode.
    Supports SMART auto save timer.
    Error logging capability: (0x01) Error logging supported.
    General Purpose Logging supported.
    Short self-test routine
    recommended polling time: ( 2) minutes.
    Extended self-test routine
    recommended polling time: ( 183) minutes.
    SCT capabilities: (0x0039) SCT Status supported.
    SCT Error Recovery Control supported.
    SCT Feature Control supported.
    SCT Data Table supported.
    SMART Attributes Data Structure revision number: 16
    Vendor Specific SMART Attributes with Thresholds:
    ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
    1 Raw_Read_Error_Rate 0x000b 100 100 050 Pre-fail Always - 0
    2 Throughput_Performance 0x0005 100 100 050 Pre-fail Offline - 0
    3 Spin_Up_Time 0x0027 100 100 001 Pre-fail Always - 1320
    4 Start_Stop_Count 0x0032 100 100 000 Old_age Always - 1807
    5 Reallocated_Sector_Ct 0x0033 100 100 050 Pre-fail Always - 7
    7 Seek_Error_Rate 0x000b 100 100 050 Pre-fail Always - 0
    8 Seek_Time_Performance 0x0005 100 100 050 Pre-fail Offline - 0
    9 Power_On_Hours 0x0032 081 081 000 Old_age Always - 7967
    10 Spin_Retry_Count 0x0033 136 100 030 Pre-fail Always - 0
    12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 1789
    191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age Always - 97
    192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 33
    193 Load_Cycle_Count 0x0032 082 082 000 Old_age Always - 189405
    194 Temperature_Celsius 0x0022 100 100 000 Old_age Always - 39 (Min/Max 9/50)
    196 Reallocated_Event_Count 0x0032 100 100 000 Old_age Always - 1
    197 Current_Pending_Sector 0x0032 100 100 000 Old_age Always - 9
    198 Offline_Uncorrectable 0x0030 100 100 000 Old_age Offline - 0
    199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age Always - 0
    220 Disk_Shift 0x0002 100 100 000 Old_age Always - 87
    222 Loaded_Hours 0x0032 085 085 000 Old_age Always - 6001
    223 Load_Retry_Count 0x0032 100 100 000 Old_age Always - 0
    224 Load_Friction 0x0022 100 100 000 Old_age Always - 0
    226 Load-in_Time 0x0026 100 100 000 Old_age Always - 320
    240 Head_Flying_Hours 0x0001 100 100 001 Pre-fail Offline - 0
    SMART Error Log Version: 1
    ATA Error Count: 634 (device log contains only the most recent five errors)
    CR = Command Register [HEX]
    FR = Features Register [HEX]
    SC = Sector Count Register [HEX]
    SN = Sector Number Register [HEX]
    CL = Cylinder Low Register [HEX]
    CH = Cylinder High Register [HEX]
    DH = Device/Head Register [HEX]
    DC = Device Command Register [HEX]
    ER = Error register [HEX]
    ST = Status register [HEX]
    Powered_Up_Time is measured from power on, and printed as
    DDd+hh:mm:SS.sss where DD=days, hh=hours, mm=minutes,
    SS=sec, and sss=millisec. It "wraps" after 49.710 days.
    Error 634 occurred at disk power-on lifetime: 7967 hours (331 days + 23 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    40 41 02 04 84 9a 6e Error: UNC at LBA = 0x0e9a8404 = 245007364
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    60 08 00 00 84 9a 40 08 03:31:25.735 READ FPDMA QUEUED
    ea 00 00 00 00 00 a0 00 03:31:25.714 FLUSH CACHE EXT
    61 18 08 07 82 b1 40 08 03:31:25.714 WRITE FPDMA QUEUED
    61 08 00 0f 7d 31 40 08 03:31:25.714 WRITE FPDMA QUEUED
    ea 00 00 00 00 00 a0 00 03:31:25.682 FLUSH CACHE EXT
    Error 633 occurred at disk power-on lifetime: 7967 hours (331 days + 23 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    40 41 02 04 84 9a 6e Error: UNC at LBA = 0x0e9a8404 = 245007364
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    60 08 00 00 84 9a 40 08 03:31:21.832 READ FPDMA QUEUED
    61 18 00 27 81 b1 40 08 03:31:21.451 WRITE FPDMA QUEUED
    61 18 00 27 77 b1 40 08 03:31:20.425 WRITE FPDMA QUEUED
    61 18 00 2f 8c b1 40 08 03:31:19.451 WRITE FPDMA QUEUED
    61 18 00 ef 81 b1 40 08 03:31:18.445 WRITE FPDMA QUEUED
    Error 632 occurred at disk power-on lifetime: 7967 hours (331 days + 23 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    40 41 0a 04 84 9a 6e Error: UNC at LBA = 0x0e9a8404 = 245007364
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    60 00 00 07 7f 09 40 08 03:16:43.581 READ FPDMA QUEUED
    60 08 08 00 84 9a 40 08 03:16:43.559 READ FPDMA QUEUED
    60 00 00 a7 5b 09 40 08 03:16:43.552 READ FPDMA QUEUED
    ef 10 02 00 00 00 a0 00 03:16:43.551 SET FEATURES [Reserved for Serial ATA]
    27 00 00 00 00 00 e0 00 03:16:43.551 READ NATIVE MAX ADDRESS EXT
    Error 631 occurred at disk power-on lifetime: 7967 hours (331 days + 23 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    40 41 0a 04 84 9a 6e Error: UNC at LBA = 0x0e9a8404 = 245007364
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    60 00 00 a7 5b 09 40 08 03:16:39.713 READ FPDMA QUEUED
    60 08 00 87 b5 25 40 08 03:16:39.705 READ FPDMA QUEUED
    60 00 00 8f 60 05 40 08 03:16:39.693 READ FPDMA QUEUED
    60 08 00 e7 d4 54 40 08 03:16:39.692 READ FPDMA QUEUED
    60 08 08 00 84 9a 40 08 03:16:39.678 READ FPDMA QUEUED
    Error 630 occurred at disk power-on lifetime: 7967 hours (331 days + 23 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    40 41 12 04 84 9a 6e Error: WP at LBA = 0x0e9a8404 = 245007364
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    61 08 08 0e 4c b6 40 08 03:16:35.710 WRITE FPDMA QUEUED
    61 08 08 0e 4c b6 40 08 03:16:35.710 WRITE FPDMA QUEUED
    61 08 08 0e 4c b6 40 08 03:16:35.709 WRITE FPDMA QUEUED
    61 98 18 8f 41 31 40 08 03:16:35.709 WRITE FPDMA QUEUED
    61 08 08 0e 4c b6 40 08 03:16:35.708 WRITE FPDMA QUEUED
    SMART Self-test log structure revision number 1
    Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
    # 1 Short offline Completed without error 00% 7967 -
    SMART Selective self-test log data structure revision number 1
    SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
    1 0 0 Not_testing
    2 0 0 Not_testing
    3 0 0 Not_testing
    4 0 0 Not_testing
    5 0 0 Not_testing
    Selective self-test flags (0x0):
    After scanning selected spans, do NOT read-scan remainder of disk.
    If Selective self-test is pending on power-up, resume after 0 minute delay.
    Is there an equivalent to Solaris iostat -En in Linux? Havent found any. Remember that I can't install anything: read-only FS
    Last edited by gromlok (2013-01-23 16:47:35)

Maybe you are looking for

  • Dynamically want to display a row of a table in BI Publisher

    Hi Gurus, Could any body help me regarding the following thing . I want to display a complete row of a table as a first record(on thenext page) if the particular Group having some data goes to next page . This Grouph is the inner group. I have tried

  • Hard drive reformatted. Need to delete iweb site. No backups available

    Hard drive was reformatted and the iweb site material obviously lost. I would like to delete all the iweb pages i have posted using 'mobile me' as they do not show anything. Is this possible. I though it would be possible to delete the pages using 'm

  • TS3276 Lost messages in the sent messages folder?

    Has anyone lost mobileme (icloud) messages in the sent messages folder?

  • How to send a push notification in smp

    Hi All,     I am doing a sample with Push notification in SMP2.3, when i insert a record in database then it has to send a notification. form the work space when i send a notification it is coming. Now i want the notification when i insert a record t

  • Problem with revoking accounts on deprovisioned Date

    Hello, I am setting the De-provisioning date for a user manually from UI to the current date(tried with yesterday's date too). After I submit, I ran the "Set User Deprovisioned Date" Job. It has updated the user deprovisioned date field, but the user