[SOLVED] Ring file-system for a surveillance server?

In my home, I have a (almost) dedicated (Arch-32) computer to use as a surveillance server. It continuously fetches video from a IP camera using motion and saving various GB of images and videos. Obviously, I have to manually delete older videos when the disk is getting full, but it would be better to have some kind of file-system that overrides the older files with the newer ones when it gets full.
Is there anything like that? Any recommendations?
Thanks!
PS: If your suggestion/solution could only delete videos (not images), better.
Last edited by sironitomas (2011-03-12 04:06:12)

limaunion wrote:
One possiblity is to schedule a daily cron job with something like this:
find /home/videos/ -mindepth 2 -cmin +2880 -exec rm -rf {} \;
In this case cmin=2880 (in minutes) equals 48 hours; 'man find' is your friend
Regards (from baires)
LU
Gracias! I didn't know about find's cmin, it is a great option. Exec gave me some problems so I ended up using more than one command. This is the script actually:
find /home/user/videos -cmin +4320 > temp
rm `cat temp | grep avi`
rm temp
The only thing bad about it, is that when find doesn't produce any output, rm shows a error message because there are no arguments. I know that's not the best way, but it works.
Any recommendation is accepted.
Thanks!
Last edited by sironitomas (2011-03-12 04:05:38)

Similar Messages

  • Confusion with OCFS2 File system for OCR and Voting disk RHEL 5, Oracle11g,

    Dear all,
    I am in the process of installing Oracle 11g 3 Node RAC database
    The environment on which i have to do this implementation is as follows:
    Oracle 11g.
    Red Hat Linux 5 x86
    Oracle Clusterware
    ASM
    EMC Storage
    250 Gb of Storage drive.
    SAN
    As of now i am in the process of installing Oracle Clusterware on the 3 nodes.
    I have performed these tasks for the cluster installs.
    1. Configure Kernel Parameters
    2. Configure User Limits
    3. Modify the /etc/pam.d/login file
    4. Configure Operating System Users and Groups for Oracle Clusterware
    5. Configure Oracle Clusterware Owner Environment
    6. Install CVUQDISK rpm package
    7. Configure the Hosts file
    8. Verify the Network Setup
    9. Configure the SSH on all Cluster Nodes (User Equivalence)
    9. Enable the SSH on all Cluster Nodes (User Equivalence)
    10. Install Oracle Cluster File System (OCFS2)
    11.Verify the Installation of Oracle Cluster File System (OCFS2)
    12. Configure the OCFS2 (/etc/ocfs2/cluster.conf)
    13. Configure the O2CB Cluster Stack for OCFS2
    BUT, here after i am a little bit confused on how to proceed further. The next step is to Format the disk and mount the OCFS2, Create Software Directories... so and so forth.
    I asked my system admin to provide me two partitions so that i could format them with OCFS2 file system.
    He wrote back to me saying.
    *"Is what you want before I do it??*
    */dev/emcpowera1 is 3GB and formatted OCFS2.*
    */dev/emcpowera2 is 3GB and formatted OCFS2.*
    *Are those big enough for you? If not, I can re-size and re-format them*
    *before I mount them on the servers.*
    *the SAN is shared storage. /dev/emcpowera is one of three LUNs on*
    *the shared storage, and it's 214GB. Right now there are only two*
    *partitions on it- the ones I listed below. I can repartition the LUN any*
    *way you want it.*
    *Where do you want these mounted at:*
    */dev/emcpowera1*
    */dev/emcpowera2*
    *I was thinking if this mounting techique would work like so:*
    *emcpowera1: /u01/shared_config/OCR_config*
    *emcpowera2: /u01/shared_config/voting_disk*
    *Let me know how you'd like them mounted."*
    Please recommend me what i should convey to him so that i can ask him to the exact same thing.
    My second question is, as we are using ASM, for which i am gonna configure ASM after clusterware installation, should i install Openfiler??
    Pls refer the enviroment information i provided above and make recommendations.
    As of now i am using Jeffery Hunters guide to install the entire setup. You think the entire install guide goes well with my enviroment??
    http://www.oracle.com/technology/pub/articles/hunter_rac11gr1_iscsi.html?rssid=rss_otn_articles
    Kind regards
    MK

    Thanks for ur reply Mufalani,
    You have managed to solve half part of my query. But still i am stuck with what kind of mount point i should ask the system admin to create for OCR and Voting disk. Should i go with the mount point he is mentioning??
    Let me put forth few more questions here.
    1. Is 280 MB ok for OCR and voting disks respectively??
    2. Should i ask the system admin to create 4 voting disk mount points and two for ocr??
    3. As mentioned by the system admin.
    */u01/shared_config/OCR_config*
    */u01/shared_config/voting_disk*
    Is this ok for creating the ocr and voting disks?
    4. Can i use OCFS2 file system for formating the disk instead of using them as RAW device!!?
    5. As u mentioned that Openfiler is not needed for Configuring ASM... Could you provide me the links which will guide me to create partition disks, voting disks and ocr disks!! I could not locate them on the doc or else were. I did find a couple of them, but was unable to identify a suitable one for my envoirement.
    Regards
    MK

  • File system for SAP ECC, EP , BI and CRM installation on Solaris/DB2

    Hello,
    We are going to implement SAP ECC 6.0 with EP 7, BI and CRM on Solaris operating system with IBM DB2 database.
    All these applications are going to be installed on single server, as being a basis person, I know this is not all recommended.
    But due to client's requirement and keeping cost factor in mind I need to install all these application on single box.
    Now here I need your help. as I basis person, I know the required Solaris file system for SAP ECC 6.0 but not having any Idea about other application like EP/ CRM and BI .
    If anyone able to help me with required Solaris file system. it will be great help.
    Please let me know if there is any query.
    Thanks.

    > All these applications are going to be installed on single server, as being a basis person, I know this is not all recommended.
    > But due to client's requirement and keeping cost factor in mind I need to install all these application on single box.
    Why not using Solaris zones/container? This decreases the administrative amount tremendeously since you will deal as with "single machines" but they all run together on one box.
    > Now here I need your help. as I basis person, I know the required Solaris file system for SAP ECC 6.0 but not having any Idea about other application like EP/ CRM and BI .
    So you'd need to read the installation guides
    Markus

  • How to create common file systems for both development & quality systems

    Hello Team
    Need some calrification & advise
    we are planning to install ECC6 with EHP5 on AIX7.1 with DB2 9.7 version on single physical server ( NO Virutal IP'S) for both development(SID : D10 & System number 20) & quality(SID : Q10 & System number 20) systems.And it will be running in single LPAR for development & quality systems.
    please advise me how can we create the below common file systems for both development & quality to avoid file permission & installation issues.
    /sapmnt/<SID>
    /usr/sap/
    /usr/sap/trans/
    /db2/IBM
    Thanks in advance & Regards
    BM

    Thanks you Juan for your response
    We are planning for seperate DB instance only.
    what could be the owner(UID) & group(GID) for the below file systems
    /sapmnt/
    usr/sap/
    usr/sap/trans
    Thanks in advance
    BM

  • Regarding shared file system requirement in endeca server cluster

    Hi,
    Our solution involves running a single data domain in an endeca server cluster.
    As per the documentation, endeca server cluster requires shared file system for keeping index file for follower node to read.
    My questions are,
    Can I run the endeca cluster with out shared file system by having the index file on each node of the endeca server cluster?
    Can dependency on shared file system be a single point of failure, if yes how can it be avoided?
    I really appreciate your feedback on these questions.
    thanks,
    rp

    Hi rp,
    The requirement for a shared file system in the Endeca Server cluster is a must. As this diagram shows, the shared file system maintans the index, and also maintains the state of the Cluster Coordinator, which ensures cluster services (automatic leader election, propagation of the latest index version to all nodes in the data domain). A dependency on a shared file system can be a single point of failure and requires to run a backup, -- this is a standard IT approach, that is, it is not specific to the Endeca Server cluster in particular.
    See this section on Cluster Behavior, for info on how the shared file system is used (the topic "how updates are processed"), and on how increased availability is achieved.
    HTH,
    Julia

  • Choosing a highly-available, networked file system for VM images

    What's the consensus here on highly-available, networked file systems for VM disk images?
    HA-NFS seems difficult and overly-complicated.
    GlusterFS doesn't seem to be well-suited for this task.
    Ceph/RADOS seems popular but there isn't much information available.
    OCFS2 seem to be suitable but there also isn't much information available.
    Eric P.
    Truckee, CA

    Actually, I still would suggest taking a closer look at what I had mentioned in your other thread:
    [Use OCFS2 to Share Local Storage (DAS) Across OVM Compute Nodes|https://forums.oracle.com/forums/thread.jspa?threadID=2438632&tstart=0]
    I still think that this would also solve thus thread. In the end, there is no how-to on how to build that up and you probably will need some help from some experts. I for myself, also didnÄt configure all of that on my own - i left ASM/Cluster part for some contractor to setup and focused on the iscsi-target/initiator setup along with the multipath setup on m VM servers.
    Maybe I will do a little write-up, once we get it all sorted out, since we do experience some strage behaviour of our storages boxes, while ASM hammers ob them quite hardly, but I suspect the raid controllers to blame and not the concept… ;)

  • RAID/ shared file system for clustering

    could somebody give me the h/w config,vendor etc.. used generally to
    provide failover to the shared file system for cluster of weblogic
    servers?

    Many of our customers use either a hardware or software raid to protect
    their filesystem for the cluster. Most of this is a vendor specific --
    you should check out the documentation for your specific platform.
    Another option is to not do a single filesystem for the cluster. They
    should be free self-explanatory -- you simply need to replicate all of
    the files, application code, etc. across the cluster in each individual
    server machine.
    shivu wrote:
    >
    could somebody give me the h/w config,vendor etc.. used generally to
    provide failover to the shared file system for cluster of weblogic
    servers?--
    Thanks,
    Michael
    -- BEA WebLogic is hiring!
    Check our website: http://www.bea.com/

  • Lucreate - „Cannot make file systems for boot environment“

    Hello!
    I'm trying to use LiveUpgrade to upgrade one "my" Sparc servers from Solaris 10 U5 to Solaris 10 U6. To do that, I first installed the patches listed on [Infodoc 72099|http://sunsolve.sun.com/search/document.do?assetkey=1-9-72099-1] and then installed SUNWlucfg, SUNWlur and SUNWluufrom the S10U6 sparc DVD iso. I then did:
    --($ ~)-- time sudo env LC_ALL=C LANG=C PATH=/usr/bin:/bin:/sbin:/usr/sbin:$PATH lucreate -n S10U6_20081207  -m /:/dev/md/dsk/d200:ufs
    Discovering physical storage devices
    Discovering logical storage devices
    Cross referencing storage devices with boot environment configurations
    Determining types of file systems supported
    Validating file system requests
    Preparing logical storage devices
    Preparing physical storage devices
    Configuring physical storage devices
    Configuring logical storage devices
    Analyzing system configuration.
    Comparing source boot environment <d100> file systems with the file
    system(s) you specified for the new boot environment. Determining which
    file systems should be in the new boot environment.
    Updating boot environment description database on all BEs.
    Searching /dev for possible boot environment filesystem devices
    Updating system configuration files.
    The device </dev/dsk/c1t1d0s0> is not a root device for any boot environment; cannot get BE ID.
    Creating configuration for boot environment <S10U6_20081207>.
    Source boot environment is <d100>.
    Creating boot environment <S10U6_20081207>.
    Creating file systems on boot environment <S10U6_20081207>.
    Creating <ufs> file system for </> in zone <global> on </dev/md/dsk/d200>.
    Mounting file systems for boot environment <S10U6_20081207>.
    Calculating required sizes of file systems              for boot environment <S10U6_20081207>.
    ERROR: Cannot make file systems for boot environment <S10U6_20081207>.So the problem is:
    ERROR: Cannot make file systems for boot environment <S10U6_20081207>.
    Well - why's that?
    I can do a "newfs /dev/md/dsk/d200" just fine.
    When I try to remove the incomplete S10U6_20081207 BE, I get yet another error :(
    /bin/nawk: can't open file /etc/lu/ICF.2
    Quellcodezeilennummer 1
    Boot environment <S10U6_20081207> deleted.I get this error consistently (I ran the lucreate many times now).
    lucreate used to work fine, "once upon a time", when I brought the system from S10U4 to S10U5.
    Would anyone maybe have an idea about what's broken there?
    --($ ~)-- LC_ALL=C metastat
    d200: Mirror
        Submirror 0: d20
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 31458321 blocks (15 GB)
    d20: Submirror of d200
        State: Okay        
        Size: 31458321 blocks (15 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t1d0s0          0     No            Okay   Yes
    d100: Mirror
        Submirror 0: d10
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 31458321 blocks (15 GB)
    d10: Submirror of d100
        State: Okay        
        Size: 31458321 blocks (15 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t0d0s0          0     No            Okay   Yes
    d201: Mirror
        Submirror 0: d21
          State: Okay        
        Submirror 1: d11
          State: Okay        
        Pass: 1
        Read option: roundrobin (default)
        Write option: parallel (default)
        Size: 2097414 blocks (1.0 GB)
    d21: Submirror of d201
        State: Okay        
        Size: 2097414 blocks (1.0 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t1d0s1          0     No            Okay   Yes
    d11: Submirror of d201
        State: Okay        
        Size: 2097414 blocks (1.0 GB)
        Stripe 0:
            Device     Start Block  Dbase        State Reloc Hot Spare
            c1t0d0s1          0     No            Okay   Yes
    hsp001: is empty
    Device Relocation Information:
    Device   Reloc  Device ID
    c1t1d0   Yes    id1,sd@THITACHI_DK32EJ-36NC_____434N5641
    c1t0d0   Yes    id1,sd@SSEAGATE_ST336607LSUN36G_3JA659W600007412LQFN
    --($ ~)-- /bin/df -k | grep md
    /dev/md/dsk/d100     15490539 10772770 4562864    71%    /Thanks,
    Michael

    Hello.
    (sys01)root# devfsadm -Cv
    (sys01)root# To be on the safe side, I even rebooted after having run devfsadm.
    --($ ~)-- sudo env LC_ALL=C LANG=C lustatus
    Boot Environment           Is       Active Active    Can    Copy     
    Name                       Complete Now    On Reboot Delete Status   
    d100                       yes      yes    yes       no     -        
    --($ ~)-- sudo env LC_ALL=C LANG=C lufslist d100
                   boot environment name: d100
                   This boot environment is currently active.
                   This boot environment will be active on next system boot.
    Filesystem              fstype    device size Mounted on          Mount Options
    /dev/md/dsk/d100        ufs       16106660352 /                   logging
    /dev/md/dsk/d201        swap       1073875968 -                   -In the rebooted system, I re-did the original lucreate:
    <code>--($ ~)-- time sudo env LC_ALL=C LANG=C PATH=/usr/bin:/bin:/sbin:/usr/sbin:$PATH lucreate -n S10U6_20081207 -m /:/dev/md/dsk/d200:ufs</code>
    Copying.
    *{color:#ff0000}Excellent! It now works!{color}*
    Thanks a lot,
    Michael

  • Problem by Maintaining a Business System for the Integration Server

    Hi to all,
    I have the following problem. I tried to change some Business systems configuration in SLD but accidentally deleted defined Business System for our Integration Server. After that I tried to create that business system again but I have the following error:
    Integration Server already defined. There is already an integration server defined for the selected technical system. Please select role Application System.
    Since than I’m receiving the following error message when I try to refresh my adapter engine cache:
    com.sap.aii.ib.server.abapcache.CacheRefreshException: Unable to find an associated SLD element
    Do you know how can I define the Business System again currently I can't see it in SLD Business systems?
    I’ve also checked the IS URL (TR: SXMB_MONI) and it seems that it’s ok. 
    Regards
    Naycho

    Hi Naycho,
    i think you have to clean up some sld rubbish
    Go to SLD / Administration / Content Maintenance. Choose left side "all classes" or "landscape description" and right side  "business system". Delete all entries you find for your business system, after that try to recreate it.
    Regards,
    Udo

  • Btrfs with different file systems for different mount points?

    Hey,
    I finally bought a SSD, and I want to format it to f2fs (and fat to boot on UEFI) and install Arch on it, and in my old HDD I intend to have /home and /var and  try btrfs on them, but I saw in Arch wiki that btrfs "Cannot use different file systems for different mount points.", it means I cannot have a / in f2fs and a /home in btrfs? What can I do? Better use XFS, ZFS or ext4 (I want the faster one)?
    Thanks in advance and sorry my english.

    pedrofleck wrote:Gosh, what was I thinking, thank you! (I still have a doubt: is btrfs the best option?)
    Just a few weeks ago many of us were worrying about massive data loss due to a bug introduced in kernel 3.17 that caused corruption when using snapshots. Because btrfs is under heavy developement, this sort of thing can be expected. That said, I have my entire system running with btrfs. I have 4 volumes, two raid1, a raid0 and a jbod. I also run rsync to an ext4 partition and ntfs. Furthermore I make offline backups as well.
    If you use btrfs make sure you have backups and make sure you are ready to use them. Also, make sure you check sum your backups. rsync has the option to use checksums in place of access times to determine what to sync.

  • Is there native file system for solaris cluster?

    As GFS for RedHat Linux and GPFS for AIX, is there any native file system for solaris cluster? Thanks for any infomation!

    What do you mean by 'native'?
    Oracle Solaris Cluster has a feature called the Global File System (or sometimes the Proxy File System). It allows UFS or Symantec VxFS to be mounted globally (on all cluster nodes simultaneously) in read/write mode. There is also a shared QFS option. All these are described in the documentation and in the book "Oracle Solaris Cluster Essentials".
    Tim
    ---

  • File systems available on Windows Server 2012 R2?

    What are the supported file systems in Windows Server 2012 R2? I mean the complete list. I know you can create, read and write on Fat32, NTFS and ReFS. What about non-Microsoft file systems, like EXT4 or HFS+? If I create a VM with a Linux OS, will
    I be able to acces the virtual hard disk natively from WS 2012 R2, or will I need a third party tool, like the one from Paragon? If I have a drive formated in EXT4 or HFS+, will I be able to acces it from Windows, without any third party tool? Acces it,
    I mean both read and write on them. I know that on the client OS, Windows 8.1, this is not possible natively, this is why I am asking here, I guess it is very possible for the server OS to have build-in support for accesing thoose file systems. If Hyper-V
    has been optimised to run not just Windows VMs, but also Linux VMs, it would make sense to me that file systems like thoose from Linux or OS X to be available using a build-in feature. I have tried to mount the vhd from a Linux VM I have created in HyperV,
    Windows Explorer could not read the hard drive.

    Installed Paragon ExtFS free. With it loaded, tried to mount on Windows Explorer a ext4 formated vhd, created on a Linux Hyper-V vm, it failed, and Paragon ExtFS crashed. Uninstalled Paragon ExtFS. The free version was not supported on WS 2012 R2
    by Paragon, if Windows has no build-in support for ext4, this means this free software has not messed around anything in the OS, I guess.
    Don't mess with third-party kernel-mode file systems as it's basically begging for troubles: crash inside them will make whole system BSOD and third-party FS are typically buggy... Because a) FS development for Windows is VERY complex and b) there are very
    few external adopters so not that many people actually theist them. What you can do however:
    1) Spawn an OS with a supported FS inside VM and configure loopback connectivity (even over SMB) with your host. So you'll read and write your volume inside a VM and copy content to / from host.
    (I personally use this approach in a reversed direction, my primary OS is MacOS X but I read/write NTFS-formatted disks from inside a Windows 7 VM I run on VMware Fusion)
    2) Use user-mode file system explorer (see sample links below, I'm NOT affiliated with that companie). So you'll copy content from the volume as it would be some sort of a shell extension.
    Crashes in 1) and 2) would not touch your whole OS stability. 
    HFS Explorer for Windows
    http://www.heise.de/download/hfsexplorer.html
    Ext2Read
    http://sourceforge.net/projects/ext2read/
    (both are user-land applications for HFS(+) and EXT2/3/4 accordingly)
    Hope this helped :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • File system getting full and Server node getting down.

    Hi Team,
    Currently we are using IBM Power 6 AIX operating system.
    And in our environment, for development system, file system is getting full and development system is getting slow while accessing.
    Can you please let me know , what exactly the problem & which command is used to see the file system size and how to resolve the issue by deleting the core files or some ting. Please help me .
    Thanks
    Manoj K

    Hi      Orkun Gedik,
    When i executed the command df -lg and find . -name core noting is displayed, but if i execute the command df is showed me the below information. below is an original file which i have modified the sid.
    Filesystem    512-blocks      Free %Used    Iused %Iused Mounted on
    /dev/fslv10     52428800  16279744   69%   389631    15% /usr/sap/SID
    Server 0 node is giving the problem. its getting down all the times.
    And if i check it in the /usr/sap/SID/<Instance>/work for the server node "std_server0.out file , the below information is written in the file.
    framework started for 73278 ms.
    SAP J2EE Engine Version 7.00   PatchLevel 81863.450 is running! PatchLevel 81863.450 March 10, 2010 11:48 GMT
    94.539: [GC 94.539: [ParNew: 239760K->74856K(261888K), 0.2705150 secs] 239760K->74856K(2009856K), 0.2708720 secs] [Times: user=0.00 sys=0.36, real=0.27 secs]
    105.163: [GC 105.164: [ParNew: 249448K->80797K(261888K), 0.2317650 secs] 249448K->80797K(2009856K), 0.2320960 secs] [Times: user=0.00 sys=0.44, real=0.23 secs]
    113.248: [GC 113.248: [ParNew: 255389K->87296K(261888K), 0.3284190 secs] 255389K->91531K(2009856K), 0.3287400 secs] [Times: user=0.00 sys=0.58, real=0.33 secs]
    Please advise.
    thanks in advance
    Manoj K

  • Image: display from local drive, upload to server file system, and display from server

    Hello,
    I am using jdev 11.1.2.4.0...
    The requirement is that users would like to upload images from their local drive to server drive and they would like display images from server drive later.
    We don't want to keep images in database. I am not sure what the solution should be; however, I plan to ....
    1. create a table to keep an information of images -- image_id, image_location (server drive), image_filename
    2. create a page where users can enter image information and specific filename from a local drive (I think I will use inputFile component) and provide the preview button to display an image on screen before save it. To save data is to save information to the database and copy an image file to the server drive (the destination location for image files is predefined.)
    3. create another page where users can browse information from the database and display an image for a selected record by getting from the server.
    I need suggestions on...
    1. how to display an image on page from the local drive?
    2. how to copy a file from a local drive to a server?
    3. how to display an image on page from the server drive?
    Thank you.
    nat

    See:
    http://tompeez.wordpress.com/2011/11/26/jdev11-1-2-1-0-handling-imagesfiles-in-adf-part-1/
    http://tompeez.wordpress.com/2011/11/26/jdev11-1-2-1-0-handling-imagesfiles-in-adf-part-2/
    http://tompeez.wordpress.com/2011/12/16/jdev11-1-2-1-0-handling-imagesfiles-in-adf-part-3/
    Where Timo saves images to the database, you save it to the file system (examples on how to do this from Java are available plenty if you just Google for it). Similar to Timo you then use the image tag to display images. The difference is that you can directly add the URL from the database table.
    The benefit of using a database to host images is that you are not dependent on file server structures though
    Frank

  • SAP file system restoration on other server

    Dear Experts,
    To check that our offline file system backup is successful, we are planning to restore the offline file system backup from the tape on to a new test server.
    Our current SAP system (ABAP only) is in cluster with CI running on one node (using virtual host name cicep) & DB running on another node (using virtual host name dbcep).
    Now, is it possible to restore the offline file system backup of the above said cluster server on to a single server with a different host name?
    Please help on this.
    Regards,
    Ashish Khanduri

    Dear Ashish
    We want to include file system backup process as part of our backup strategy.  To test the waters, we are planning to take a backup of the at filesystems level.  Following are the filesystems in our production systems.
    We have a test server (hostname is different), without any filesystems created beforehand. 
    I want to know:
    1. Which filesystems will be required from the below:
    /dev/hd4         4194304   3772184   11%     5621     2% /
    /dev/hd2        10485760   6151688   42%    43526     6% /usr
    /dev/hd9var      4194304   4048944    4%     4510     1% /var
    /dev/hd3         4194304   2571760   39%     1543     1% /tmp
    /dev/hd1          131072    129248    2%       85     1% /home
    /proc                  -         -    -         -     -  /proc
    /dev/hd10opt      655360    211232   68%     5356    18% /opt
    /dev/oraclelv   83886080  73188656   13%    11091     1% /oracle
    /dev/optoralv   20971520  20967664    1%        4     1% /opt/oracle
    /dev/oracleGSPlv   83886080  74783824   11%    18989     1% /oracle/GSP
    /dev/sapdata1lv  833617920 137990760   84%     3189     1% /oracle/GSP/sapdata1
    /dev/sapdata2lv  623902720 215847400   66%       82     1% /oracle/GSP/sapdata2
    /dev/sapdata3lv  207093760 108510632   48%       24     1% /oracle/GSP/sapdata3
    /dev/sapdata4lv  207093760 127516424   39%       28     1% /oracle/GSP/sapdata4
    /dev/origlogAlv   20971520  20730080    2%        8     1% /oracle/GSP/origlogA
    /dev/origlogBlv   20971520  20730080    2%        8     1% /oracle/GSP/origlogB
    /dev/mirrlogAlv   20971520  20762848    1%        6     1% /oracle/GSP/mirrlogA
    /dev/mirrlogBlv   20971520  20762848    1%        6     1% /oracle/GSP/mirrlogB
    /dev/oraarchlv  311951360 265915600   15%      526     1% /oracle/GSP/oraarch
    /dev/usrsaplv   41943040  41449440    2%      165     1% /usr/sap
    /dev/sapmntlv   41943040  20149168   52%   565823    21% /sapmnt
    /dev/usrsapGSPlv   41943040  25406768   40%   120250     5% /usr/sap/GSP
    /dev/saptranslv   41943040   5244424   88%   136618    18% /usr/sap/trans
    IDES:/sapcd     83886080   4791136   95%    18878     4% /sapcd
    GILSAPED:/usr/sap/trans   41943040   5244424   88%   136618    18% /usr/sap/trans
    2. Is it possible to directly backup the filesystems (like /dev/oracleGSPlv)?  This requirement is because, when I backup (using tar) /oracle, all the folders in /oracle, like /oracle/GSP, /oracle/GSP/sapdata1 etc, are also backed up.  I do not want it.  I would like to backup the filesystems directly.
    3. Which unix backup tools are used to backup the individual filesystems?
    4. How do we restore the filesystems to the test server?
    Thanks for your advise.
    Abdul
    Edited by: Abdul Rahim Shaik on Feb 8, 2010 12:10 PM

Maybe you are looking for

  • How to setup the timeout time about container mangement transaction?

    hi i want to modify the time-out time about cantainer management transaction. but i don't know where and how to setup it.... please tell me how to do it, thanks a lot. Joe

  • New bookmarks not being saved

    Beginning today my new bookmarks are not being saved. I read several KB articles and have tried: - making sure the bookmarks file is not read-only - reset browser.places in about:config - renamed the places.sqlite and places.sqlite-journal, and resta

  • Files on the base URL

    How can I explore / search for files on a Applets base URL. Is there a simple way to do this without using servlets or some kind of java security programing ?

  • How to configure output device option

    DEAR ALL ,                   I have been facing problem in taking printout in line printer(lp04)  with page dimension 305 mm * 215mm for transaction mb90--> output type -->we03 which is not giving the option to choose output device type before print

  • Confusion with Membership and Billing

    I currently have Creative Cloud (Student and Teacher Edition) installed on my laptop for about a year now. I downloaded it through free subscription offered on a website, with a redemption code (thanks to an agreement between my college and Adobe.) F