Question on File System Types

Hi,
I would like to have a dual boot Solaris 10, Windows system. I want to create a partition such that both systems can read and write to it.
Any ideas?

Best choice today is a FAT filesystem. It has limitations on file size and total size, but most operating systems can use it easily.
Darren

Similar Messages

  • [SOLVED] ERROR: Unable to determine the file system type of /dev/root:

    :: Running Hook [udev]
    :: Triggering uevents...done
    Root device '804' doesn't exist.
    Creating root device /dev/root with major 8 and minor 4.
    error: /dev/root: No such device or address
    ERROR: Unable to determine the file system type of /dev/root:
    Either it contains no filesystem, an unknown filesystem,
    or more than one valid file system signature was found.
    Try adding
    rootfstype=your_filesystem_type
    to the kernelcommand line.
    You are now being dropped into an emergency shell.
    /bin/sh: can't access tty; job control turned off
    [ramfs /]# [ 1.376738] Refined TSC clocksource calibration: 3013.000 MHz.
    [ 1.376775] Switching to clocksource tsc
    That's what I get when I boot my Arch system. It worked fine for quite a while, but suddenly it ran into an error where the SCSI driver module was corrupt. I fixed it by reinstalling util-linux-ng and kernel26, however, I run into this issue now. http://www.pastie.org/2163181 < Link to /var/log/pacman.log for the month of July, just in case. Yes, I bought a new ATI/AMD Radeon HD 5450 this Saturday, but it seemed to work fine till this broke and it works fine on Ubuntu too, so I suppose we can rule it out.
    Last edited by SgrA (2011-07-05 20:45:36)

    Autodetection failed on your first image, in both your previous kernel installs:
    [2011-07-04 16:14] find: `/sys/devices': No such file or directory
    Which means that sysfs was not mounted. You should be able to boot from the fallback image, which does not use autodetect. Figure out why /sys isn't mounted, as well, and fix that.
    This is also a somewhat crappy bug in mkinitcpio that lets you create an autodetect image that's useless. It'll be fixed in the next version of mkinitcpio that makes it to core.
    Last edited by falconindy (2011-07-04 17:41:19)

  • Oracle VM Server 2.2.1 FIle System Type

    I have installed VM server 2.2.1, but i recently ran into disk size problems and want to resize. I have no space on the / partition and all my space on the /var/ovs/mount/"randomnumber". When i used the parted command "print" to see my partitions, the /var partition has no file system type such as ext3 or ext2. The / partition has ext3. Does anyone know why the /var/ovs/mount/"randomnumber" has no file system type because without this I can not resize the partition. Let me also say that when i installed VM server i pretty much took all the defaults so the partitions were created this way by default.
    Thanks

    Paul_RealityTech wrote:
    I have installed VM server 2.2.1, but i recently ran into disk size problems and want to resize. I have no space on the / partition and all my space on the /var/ovs/mount/"randomnumber". When i used the parted command "print" to see my partitions, the /var partition has no file system type such as ext3 or ext2. The / partition has ext3. Does anyone know why the /var/ovs/mount/"randomnumber" has no file system type because without this I can not resize the partition. Let me also say that when i installed VM server i pretty much took all the defaults so the partitions were created this way by default./var/ovs/mount/UUID is ocfs2 and cannot be shrunk in size.

  • Command To ID File System Type

    I was found a server running Arch. Powered on fine and seems to be running great. I was able to recover the root login however I would like to know what command I can run in a terminal to identify the file system type on the disk partitions.
    [root@tiger /]# fdisk -l
    Disk /dev/sda: 160.0 GB, 160041885696 bytes
    255 heads, 63 sectors/track, 19457 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Disk identifier: 0x000b7d3e
    Device Boot Start End Blocks Id System
    /dev/sda1 1 62 497983+ 82 Linux swap / Solaris
    /dev/sda2 * 63 311 2000092+ 83 Linux
    /dev/sda3 312 19457 153790245 83 Linux

    Does fdisk -l not supply you with what you were asking for? You could also use the command blkid, but it does not show the filesystem ID...

  • OVM Manager 2.2.2: server pool error with file system type

    I'm just getting started with OVM. I've installed OVM on one machine and the manager on another.
    I created a server pool, which seemed to work OK but shows "Error" in the table under the "Server Pools" tab in the manager interface
    When I edit it I see
    Error: OVM-1011 OVM Manager communication with NNN.NNN.NN.NNN for operation Pre-check cluster root for Server Pool failed: <Exception: SR '/dev/sda3' not supported: type 'ocfs2.local' not in ['nfs', 'ocfs2.cluster']>
    Can anyone explain this? Does this mean I can't use a local file system in OVM 2.2.2? I understood this was the case with OVM 3, which is why I went with 2.2.2.
    Thanks.

    Roger Ford wrote:
    Error: OVM-1011 OVM Manager communication with NNN.NNN.NN.NNN for operation Pre-check cluster root for Server Pool failed: <Exception: SR '/dev/sda3' not supported: type 'ocfs2.local' not in ['nfs', 'ocfs2.cluster']>
    Can anyone explain this? Does this mean I can't use a local file system in OVM 2.2.2? I understood this was the case with OVM 3, which is why I went with 2.2.2.You can't created a clustered pool with a local filesystem. You need to format the filesystem with ocfs2 in clustered mode.

  • What are the characteristics of the procfs file system type ?

    Hi Solaris guys,
    I walked through the Student Guide SA-239 to find information about procfs file system (/proc), but not yet. Could anyone here explain it for me ?
    1. File ownership is determined by the credentials of the process.
    2. It contains reference by file names to the opened files of the process.
    3. Each process ID named directory has files that contain more detailed information about the process.
    4. It contains a decimal number directory entry corresponding to a process ID.
    I wonder whether these characteristics are correct or not. Please help me.
    Thanks a lot !

    man -s4 proc
    or
    http://docs.sun.com/app/docs/doc/817-0683/6mgff29c4?q=procfs&a=view
    Its a virual directory structure with a directory per process. These per process directories contain more directories and files that supply detailed information on that process eg per open file information.
    Access to the per process directories is controlled by checking the credentials of the accessing process against the credentials of the
    process whose procfs directory is being accessed.
    tim

  • Get the HDD file system type

    Hi All,
    I want to know about the file system of my hard disk (Fat32 or NTFS) using java. Can we do it using java
    regards,
    Maheshwaran Devaraj

    When you say path you mean you want the path to print out in your HTML? Is that accurate? If so you generally need to constructe that path based on the path to the current component. So if you component is located at /content/mysite/en/about/mypage/jcr:content/parsys/image then the path to the image would generally be something like /content/mysite/en/about/mypage/jcr:content/parsys/image.img.jpg/1283829292873.jpg. The .img. selector triggers the servlet associated with the foundation parbase - /libs/foundation/components/parbase/img.GET.java. The reason you reference it this way is that there is no filesystem path to the image - it is stored in the repository not on the file system, and it requires a servlet or script to ge the binary from the repository and steam it.
    Normally the way you'd construct this is to use the out of the box Image class - so look at /libs/foundation/components/image/image.jsp. Now this example assumes that your component where you loaded the image extends /libs/foundation/components/parbase. If it doesn't then you either have to change your sling:superResourceType to /libs/foundation/components/parbase or some other component that does exten /libs/foundation/components/parbase.

  • General question about file system of MAC OS

    hello,
    I like to know the difference  between  HFS+ and HFSX  , I never used MAC in my life but now I have some interest.
    I looked in wiki but there is no clear explanation.
    thanks
    johan

    I've never heard of "HFSX" unless you mean HFS Extended (or more correctly Mac OS Extended), which is the more current term for HFS+.
    Regards.

  • Mounting the Root File System into RAM

    Hi,
    I had been wondering, recently, how can one copy the entire root hierarchy, or wanted parts of it, into RAM, mount it at startup, and use it as the root itself.  At shutdown, the modified files and directories would be synchronized back to the non-volatile storage. This synchronization could also be performed manually, before shutting down.
    I have now succeeded, at least it seems, in performing such a task. There are still some issues.
    For anyone interested, I will be describing how I have done it, and I will provide the files that I have worked with.
    A custom kernel hook is used to (overall):
    Mount the non-volatile root in a mountpoint in the initramfs. I used /root_source
    Mount the volatile ramdisk in a mountpoint in the initramfs. I used /root_ram
    Copy the non-volatile content into the ramdisk.
    Remount by binding each of these two mountpoints in the new root, so that we can have access to both volumes in the new ramdisk root itself once the root is changed, to synchronize back any modified RAM content to the non-volatile storage medium: /rootfs/rootfs_{source,ram}
    A mount handler is set (mount_handler) to a custom function, which mounts, by binding, the new ramdisk root into a root that will be switched to by the kernel.
    To integrate this hook into a initramfs, a preset is needed.
    I added this hook (named "ram") as the last one in mkinitcpio.conf. -- Adding it before some other hooks did not seem to work; and even now, it sometimes does not detect the physical disk.
    The kernel needs to be passed some custom arguments; at a minimum, these are required: ram=1
    When shutting down, the ramdisk contents is synchronized back with the source root, by the means of a bash script. This script can be run manually to save one's work before/without shutting down. For this (shutdown) event, I made a custom systemd service file.
    I chose to use unison to synchronize between the volatile and the non-volatile mediums. When synchronizing, nothing in the directory structure should be modified, because unison will not synchronize those changes in the end; it will complain, and exit with an error, although it will still synchronize the rest. Thus, I recommend that if you synch manually (by running /root/Documents/rootfs/unmount-root-fs.sh, for example), do not execute any other command before synchronization has completed, because ~/.bash_history, for example, would be updated, and unison would not update this file.
    Some prerequisites exist (by default):
        Packages: unison(, cp), find, cpio, rsync and, of course, any any other packages which you can mount your root file system (type) with. I have included these: mount.{,cifs,fuse,ntfs,ntfs-3g,lowntfs-3g,nfs,nfs4}, so you may need to install ntfs-3g the nfs-related packages (nfs-utils?), or remove the unwanted "mount.+" entires from /etc/initcpio/install/ram.
        Referencing paths:
            The variables:
                source=
                temporary=
            ...should have the same value in all of these files:
                "/etc/initcpio/hooks/ram"
                "/root/Documents/rootfs/unmount-root-fs.sh"
                "/root/.rsync/exclude.txt"    -- Should correspond.
            This is needed to sync the RAM disk back to the hard disk.
        I think that it is required to have the old root and the new root mountpoints directly residing at the root / of the initramfs, from what I have noticed. For example, "/new_root" and "/old_root".
    Here are all the accepted and used parameters:
        Parameter                       Allowed Values                                          Default Value        Considered Values                         Description
        root                                 Default (UUID=+,/dev/disk/by-*/*)            None                     Any string                                      The source root
        rootfstype                       Default of "-t <types>" of "mount"           "auto"                    Any string                                      The FS type of the source root.
        rootflags                         Default of "-o <options>" of "mount"        None                     Any string                                      Options when mounting the source root.
        ram                                 Any string                                                  None                     "1"                                                  If this hook sould be run.
        ramfstype                       Default of "-t <types>" of "mount"           "auto"                     Any string                                      The FS type of the RAM disk.
        ramflags                         Default of "-o <options>" of "mount"        "size=50%"           Any string                                       Options when mounting the RAM disk.
        ramcleanup                    Any string                                                   None                     "0"                                                  If any left-overs should be cleaned.
        ramcleanup_source       Any string                                                   None                     "1"                                                  If the source root should be unmounted.
        ram_transfer_tool          cp,find,cpio,rsync,unison                            unison                   cp,find,cpio,rsync                           What tool to use to transfer the root into RAM.
        ram_unison_fastcheck   true,false,default,yes,no,auto                    "default"                true,false,default,yes,no,auto        Argument to unison's "fastcheck" parameter. Relevant if ram_transfer_tool=unison.
        ramdisk_cache_use        0,1                                                              None                    0                                                      If unison should use any available cache. Relevant if ram_transfer_tool=unison.
        ramdisk_cache_update   0,1                                                              None                    0                                                     If unison should copy the cache to the RAM disk. Relevant if ram_transfer_tool=unison.
    This is the basic setup.
    Optionally:
        I disabled /tmp as a tmpfs mountpoint: "systemctl mask tmp.mount" which executes "ln -s '/dev/null' '/etc/systemd/system/tmp.mount' ". I have included "/etc/systemd/system/tmp.mount" amongst the files.
        I unmount /dev/shm at each startup, using ExecStart from "/etc/systemd/system/ram.service".
    Here are the updated (version 3) files, archived: Root_RAM_FS.tar (I did not find a way to attach files -- does Arch forums allow attachments?)
    I decided to separate the functionalities "mounting from various sources", and "mounting the root into RAM". Currently, I am working only on mounting the root into RAM. This is why the names of some files changed.
    Of course, use what you need from the provided files.
    Here are the values for the time spend copying during startup for each transfer tool. The size of the entire root FS was 1.2 GB:
        find+cpio:  2:10s (2:12s on slower hardware)
        unison:      3:10s - 4:00s
        cp:             4 minutes (31 minutes on slower hardware)
        rsync:        4:40s (55 minutes on slower hardware)
        Beware that the find/cpio option is currently broken; it is available to be selected, but it will not work when being used.
    These are the remaining issues:
        find+cpio option does not create any destination files.
        (On some older hardware) When booting up, the source disk is not always detected.
        When booting up, the custom initramfs is not detected, after it has been updated from the RAM disk. I think this represents an issue with synchronizing back to the source root.
    Inconveniences:
        Unison needs to perform an update detection at each startup.
        initramfs' ash does not parse wild characters to use "cp".
    That's about what I can think of for now.
    I will gladly try to answer any questions.
    I don't consider myself a UNIX expert, so I would like to know your suggestions for improvement, especially from who consider themselves so.
    Last edited by AGT (2014-05-20 23:21:45)

    How did you use/test unison? In my case, unison, of course, is used in the cpio image, where there are no cache files, because unison has not been run yet in the initcpio image, before it had a chance to be used during boot time, to generate them; and during start up is when it is used; when it creates the archives. ...a circular dependency. Yet, files changed by the user would still need to be traversed to detect changes. So, I think that even providing pre-made cache files would not guarantee that they would be valid at start up, for all configurations of installation. -- I think, though, that these cache files could be copied/saved from the initcpio image to the root (disk and RAM), after they have been created, and used next time by copying them in the initcpio image during each start up. I think $HOME would need to be set.
    Unison was not using any cache previously anyway. I was aware of that, but I wanted to prove it by deleting any cache files remaining.
    Unison, actually, was slower (4 minutes) the first time it ran in the VM, compared to the physical hardware (3:10s). I have not measured the time for its subsequent runs, but It seemed that it was faster after the first run. The VM was hosted on a newer machine than what I have used so far: the VM host has an i3-3227U at 1.9 GHz CPU with 2 cores/4 threads and 8 GB of RAM (4 GB ware dedicated to the VM); my hardware has a Pentium B940 at 2 GHz CPU with 2 cores/2 threads and 4 GB of RAM.
    I could see that, in the VM, rsync and cp were copying faster than on my hardware; they were scrolling quicker.
    Grub, initially complains that there is no image, and shows a "Press any key to continue" message; if you continue, the kernel panics.
    I'll try using "poll_device()". What arguments does it need? More than just the device; also the number of seconds to wait?
    Last edited by AGT (2014-05-20 16:49:35)

  • DB_UNIQUE_NAME vs DB_NAME in standby databases of ASM file systems

    Question : Do we need to have the db_unique_name paramter set differently in standby database compared to db_name in standby .
    Problem we are facing :
    Here is the little background
    Primary Server     : SERVER1
    db_name          : VENKAT
    db_unique_name     : VENKAT
    Standby server     : SERVER2
    db_name          : VENKAT
    db_unique_name     : VENKAT_stb
    Sever          : Linux
    Database Version: 11.20.3
    File system type: ASM ( 11.2.0.3)
    stanby type     : Physical
    Disk group names: Identical on both primary and standby servers
    Data          : +DATA_OP01027_128
    FRA          : +FRA_VENKAT_128
    How datafiles are layed out on primary server:
    sample datafile name location : +DATA_VENKAT_128/VENKAT/datafile/venkat.277.789579565
    How standby was build : using Active duplicte command
    Once we have the standby database build we have the datafiles created under this location
    Sample datafile name location on standby server : +DATA_VENKAT_128/VENKAT_stb/datafile/venkat.280.789579597
    with this we have learnt that the directory VENKAT_stb is getting created in standby ASM off the db_unique_name that
    was given in the database , we have not seen this issue in the normal file system even we are using db_unique_name different that
    the db_name in standby database .
    Can you please help us how we can prevent this situation of having datafiles getting created under differnt direction in standby compared to prod.
    Can you also let us know what impacts we might be having if we don't specify the db_unique_name different than db_name in standby database.
    Hope this explains the problem what we are facing currently .
    What steps i follwed to fix this issues :
    I have db_unique_name set to the same name as db_name and when i did the restore all datafiles are in the identical location to prod standby server .
    Note : We do fully understand the need for having the db_unique_name set different to db_name in standby db in standby and primary db's are residing on the same physical server .
    Thanks
    Venkat

    first of all, this is not an issue or problem
    works as intended
    Question : Do we need to have the db_unique_name paramter set differently in standby database compared to db_name in standby .yes
    Sample datafile name location on standby server : +DATA_VENKAT_128/VENKAT_stb/datafile/venkat.280.789579597
    with this we have learnt that the directory VENKAT_stb is getting created in standby ASM off the db_unique_name that
    was given in the database , we have not seen this issue in the normal file system even we are using db_unique_name different that
    the db_name in standby database .
    Can you please help us how we can prevent this situation of having datafiles getting created under differnt direction in standby compared to prod.well, dont use OMF then
    OMF format for datafiles in ASM is: +DISKGROUP/DB_UNIQUE_NAME/DATAFILE/TABLESPACE_NAME.FILE.INCARNATION
    datafiles will be created this way no matter what you do
    the difference is, that if you dont use OMF, there will be an alias created referencing the file, with the path you gave
    for example:
    OMF:
    create tablespace test size 10M;
    a datafile is created: +DATA_VENKAT_128/VENKAT/DATAFILE/test.280.789581212 (i wrote some random numbers here)
    non-OMF:
    create tablespace test datafile '+DATA_VENKAT_128/dummy/test01.dbf' size 10M;
    what actually happens:
    a datafile is created: +DATA_VENKAT_128/VENKAT/DATAFILE/test.280.789581212 (i wrote some random numbers here)
    and an ASM alias is created: +DATA_VENKAT_128/dummy/test01.dbf
    and this alias is used by the database
    while OMF files have their specified path format, and their path (db_unique_name) and even name (numbers at the end) will change when duplicated, aliases dont necessarily do this
    however this is just extra work and administration, OMF is your friend

  • How to determine the file system on Solaris

    Friends,
    How to determine which file system I have installed UFS or ZFS on Solaris
    Thanks

    Other methods would include looking at the /etc/vfstab if it's in there or fstyp(1M):
    System Administration Commands fstyp(1M)
    NAME
    fstyp - determine file system type
    SYNOPSIS
    fstyp [-a | -v] special [:logical-drive]

  • Dfc: Display file system space usage using graph and colors

    Hi all,
    I wrote a little tool, somewhat similar to df(1) which I named dfc.
    To present it, nothing better than a screenshot (because of colors):
    And there is a few options available (as of version 3.0.0):
    Usage: dfc [OPTIONS(S)] [-c WHEN] [-e FORMAT] [-p FSNAME] [-q SORTBY] [-t FSTYPE]
    [-u UNIT]
    Available options:
    -a print all mounted filesystem
    -b do not show the graph bar
    -c choose color mode. Read the manpage
    for details
    -d show used size
    -e export to specified format. Read the manpage
    for details
    -f disable auto-adjust mode (force display)
    -h print this message
    -i info about inodes
    -l only show information about locally mounted
    file systems
    -m use metric (SI unit)
    -n do not print header
    -o show mount flags
    -p filter by file system name. Read the manpage
    for details
    -q sort the output. Read the manpage
    for details
    -s sum the total usage
    -t filter by file system type. Read the manpage
    for details
    -T show filesystem type
    -u choose the unit in which
    to show the values. Read the manpage
    for details
    -v print program version
    -w use a wider bar
    -W wide filename (un truncate)
    If you find it interesting, you may install it from the AUR: http://aur.archlinux.org/packages.php?ID=57770
    (it is also available on the archlinuxfr repository for those who have it enabled).
    For further explanations, there is a manpage or the wiki on the official website.
    Here is the official website: http://projects.gw-computing.net/projects/dfc
    If you encounter a bug (or several!), it would be nice to inform me. If you wish a new feature to be implemented, you can always ask me by sending me an email (you can find my email address in the manpage or on the official website).
    Cheers,
    Rolinh
    Last edited by Rolinh (2012-05-31 00:36:48)

    bencahill wrote:There were the decently major changes (e.g. -t changing from 'don't show type' to 'filter by type'), but I suppose this is to be expected from such young software.
    I know I changed the options a lot with 2.1.0 release. I thought it would be better to have -t for filtering and -T for printing the file system type so someone using the original df would not be surprised.
    I'm sorry for the inconvenience. There should not be any changes like this one in the future though but I thought it was needed (especially because of the unit options).
    bencahill wrote:
    Anyway, I now cannot find any way of having colored output showing only some mounts (that aren't all the same type), without modifying the code.
    Two suggestions:
    1. Introduce a --color option like ls and grep (--color=WHEN, where WHEN is always,never,auto)
    Ok, I'll implement this one for 2.2.0 release It'll be more like "-c always", "-c never" and "-c auto" (default) because I do not use long options but I think this would be OK, right?
    bencahill wrote:2. Change -t to be able to filter multiple types (-t ext4,ext3,etc), and support negative matching (! -t tmpfs,devtmpfs,etc)
    This was already planned for 2.2.0 release
    bencahill wrote:Both of these would be awesome, if you have time. I've simply reverted for now.
    This is what I would have suggested.
    bencahill wrote:By the way, awesome software.
    Thanks I'm glad you like it!
    bencahill wrote:P.S. I'd already written this up before I noticed the part in your post about sending feature requests to your email. I decided to post it anyway, as I figured others could benefit from your answer as well. Please forgive me if this is not acceptable.
    This is perfectly fine Moreover, I seem to have some troubles with my e-mail addressee... So it's actually better that you posted your requests here!

  • OEL file system questions - XFS and encrypted file system

    Does anyone know if Oracle Enterprise Linux 64 bit can support / install the XFS file type ?
    Different question.
    Does anyone know if Oracle Enterprise Linux 64 bit can encrypt the whole file system?
    Steve

    Oracle Enterprise Linux (OEL), which is based on RHEL source, does not contain everything you need to create an XFS filesystem.
    Sergio

  • Unix admin question - file system is filling up

    Oracle 9.2.0.1.0
    UNIX AIX 5.2
    The oracli3 filesystem was hitting the upper 90%s and I informaed the DBA. Learning from him is just leaning over his shoulder scribbling down as much as I can see while he thrashes through it.
    I gleaned some things today that I wanted to check with tthe UNIX pros out there.
    As I explained, that particular filesystem was filling up, and so he made some searches, first in the /patches directory. There were not too many there (in size), so he next went to /oracli1/oracle/iAS and did a search for logs like this:
    sudo find . -name '*.log' -ls
    He found one huge log file, and so had me do the cat /dev/null > command to send it to the black hole.
    My question is this - when I next see that we are filling up as it were, should I automatically look for logs? Are they the culprits some of the time? Can offending logs be dealt with in this way (cat /dev/null >) ?
    Is there something else I should be looking for ? (The oracli3 is now at 93% so I think we still have things to look at)
    Many thanks.
    DA
    Sorry - one other thing - when I use the
    find . -name '*.log' -ls command, how can I change it so that the output is more like that of the "more" suffix? The problem is that I would like to scroll through the output of .log files that this command throws back at me.
    Message was edited by:
    Dan A

    It all depends.
    When you assume there's nothing useful in those log files, you hire an Unix admin with zero knowledge of Oracle and have hem wipe those files away, 'because the file system fills up'. In the past we had similar administrators throwing away archived redo log files for exactly the same reason.
    YOU on the other hand should set up proper administration procedures to move them away, tar them, and compress the tar file, so they take less space. After another 30 days you throw away the compressed tars.
    What you have now is a very unprofessional method to 'administrate' a server.
    Actually you are not administrating it all, because when disaster strikes, or disaster stroke in the past, and you need to explain and find out what happened and especially WHY, you will be yelling at the Unix 'admin' because he just has thrown away all log files, 'because the disk was getting full'
    As to the second question:
    There are a few fundamental concepts in Unix
    1 ALL output can be redirected using > < and |, The latter symbol is called the pipe symbol and feeds the output into a filter. The filter is a command which reads the output from a previous command and processes. More, pg, and sort are filters, even sqlplus can be a filter
    2 NO special switches should be implemented in any command if you can accomplish it using redirection or filters. So you won't find
    TYPE/PAGE (VMS, Dos), but you will find cat <filename> | more
    So the answer is you need to add | more to the command line
    find . -name '*.log' -ls | more
    You would also be advised to try to follow an Unix course, as your question is very fundamental. Alternatively buy the book Classic Shell scripting
    Sybrand Bakker
    Senior Oracle DBA

  • Shared file system and other questions

    Hi,
              Is shared file system common for all WL in the cluster or each WL has its
              own one.
              Next questin is connected with proxy server.
              I'm not sure if there should be one proxy server for two servers working in
              one cluster or each server should has own proxy server (I mean WL proxy
              server).???
              In my opinion there should be one proxy server which redirect requests to
              server in cluster using some balancing method. But when I read documentation
              about cluster's architectures I saw the picture where each WL cluster server
              had proxy own server. I'm not sure if this picture showed only outline of
              architecture or it was real solution???
              Is it better to install another proxy server (for example Apache) than WL
              proxy server?
              In previous questions I asked if it is possible to replic static classes,
              fields etc. Prasad Peddada told that it is not replicated. But I wonder if
              there exists some method to force replicating it.? Maybe we can implement
              some method and run it?
              Thanks,
              Best regards,
              Robert
              

    Mike.
              In WLS51 cluster, it is better to specify different workingDir for each
              server. You can create a weblogic.properties in your per-server directory
              and add jspc properties in that file.
              Cheers - Wei
              "Mike Reiche" <[email protected]> wrote in message
              news:[email protected]...
              >
              > If you share directories between instances -
              > what happens when two WL instances decide to compile
              > the same JSP at the same time?
              >
              > Mike
              >
              >
              > Viresh Garg <[email protected]> wrote:
              > >
              > >
              > >
              > >
              > >"Robert Zaczyñski" wrote:
              > >
              > >> Hi,
              > >> Is shared file system common for all WL in
              > >the cluster or each WL has its
              > >> own one.
              > >
              > >I am not sure I understand your question. We
              > >recommend using shared file system
              > >for ease of maintanence and to get rid of pain
              > >for versioning problems for
              > >application code. It is not a requirement of
              > >WLS cluster. I really don't
              > >understand what you mean by
              > >"Is shared file system common for all WL in
              > >the cluster or each WL has its
              > >own one
              > >"
              > >
              > >>
              > >>
              > >> Next questin is connected with proxy server.
              > >> I'm not sure if there should be one proxy
              > >server for two servers working in
              > >> one cluster or each server should has own
              > >proxy server (I mean WL proxy
              > >> server).???
              > >> In my opinion there should be one proxy server
              > >which redirect requests to
              > >> server in cluster using some balancing method.
              > >But when I read documentation
              > >> about cluster's architectures I saw the picture
              > >where each WL cluster server
              > >> had proxy own server. I'm not sure if this
              > >picture showed only outline of
              > >> architecture or it was real solution???
              > >
              > >One or more proxy servers can be used. One is
              > >just fine, however you **might**
              > >want to use multiple proxies for heavily loaded
              > >application to avoid contention.
              > >All proxy servers will load balance between
              > >all servers in clsuter. a proxy
              > >servers is not attached or bound to an instance
              > >in cluster.
              > >
              > >>
              > >>
              > >> Is it better to install another proxy server
              > >(for example Apache) than WL
              > >> proxy server?
              > >>
              > >> In previous questions I asked if it is possible
              > >to replic static classes,
              > >> fields etc. Prasad Peddada told that it is
              > >not replicated. But I wonder if
              > >> there exists some method to force replicating
              > >it.? Maybe we can implement
              > >> some method and run it?
              > >
              > >I would recommend that for static classes, you
              > >use context parameters of web
              > >application or something similar, why do you
              > >need to replicate them. WHy not
              > >each cluster instance have a local copy and
              > >use their own copy for such
              > >classes/fields?
              > >
              > >>
              > >>
              > >> Thanks,
              > >> Best regards,
              > >> Robert
              > >
              > >Viresh Garg
              > >Principal Developer Relations Engineer
              > >BEA Systems
              > >
              > >
              > >
              > ><!doctype html public "-//w3c//dtd html 4.0
              > >transitional//en">
              > ><html>
              > >
              > ><p>"Robert Zaczyñski" wrote:
              > ><blockquote TYPE=CITE>Hi,
              > ><br>Is shared file system common for all WL
              > >in the cluster or each WL has
              > >its
              > ><br>own one.</blockquote>
              > >
              > ><p><br>I am not sure I understand your question.
              > >We recommend using shared
              > >file system for ease of maintanence and to get
              > >rid of pain for versioning
              > >problems for application code. It is not a requirement
              > >of WLS cluster.
              > >I really don't understand what you mean by
              > ><br>"<i>Is shared file system common for all
              > >WL in the cluster or each
              > >WL has its</i>
              > ><br><i>own one</i>
              > ><br>"
              > ><blockquote TYPE=CITE>
              > ><p>Next questin is connected with proxy server.
              > ><br>I'm not sure if there should be one proxy
              > >server for two servers working
              > >in
              > ><br>one cluster or each server should has own
              > >proxy server (I mean WL proxy
              > ><br>server).???
              > ><br>In my opinion there should be one proxy
              > >server which redirect requests
              > >to
              > ><br>server in cluster using some balancing method.
              > >But when I read documentation
              > ><br>about cluster's architectures I saw the
              > >picture where each WL cluster
              > >server
              > ><br>had proxy own server. I'm not sure if this
              > >picture showed only outline
              > >of
              > ><br>architecture or it was real solution???</blockquote>
              > >
              > ><p><br>One or more proxy servers can be used.
              > >One is just fine, however
              > >you **might** want to use multiple proxies for
              > >heavily loaded application
              > >to avoid contention. All proxy servers will
              > >load balance between all servers
              > >in clsuter. a proxy servers is not attached
              > >or bound to an instance in
              > >cluster.
              > ><blockquote TYPE=CITE>
              > ><p>Is it better to install another proxy server
              > >(for example Apache) than
              > >WL
              > ><br>proxy server?
              > ><p>In previous questions I asked if it is possible
              > >to replic static classes,
              > ><br>fields etc. Prasad Peddada told that it
              > >is not replicated. But I wonder
              > >if
              > ><br>there exists some method to force replicating
              > >it.? Maybe we can implement
              > ><br>some method and run it?</blockquote>
              > >
              > ><p><br>I would recommend that for static classes,
              > >you use context parameters
              > >of web application or something similar, why
              > >do you need to replicate them.
              > >WHy not each cluster instance have a local copy
              > >and use their own copy
              > >for such classes/fields?
              > ><blockquote TYPE=CITE>
              > ><p>Thanks,
              > ><br>Best regards,
              > ><br>Robert</blockquote>
              > >
              > ><p>Viresh Garg
              > ><br>Principal Developer Relations Engineer
              > ><br>BEA Systems
              > ><br> </html>
              > >
              > >
              >
              

Maybe you are looking for

  • SAPPHIRE RADEON HD2600XT 256MB AGP RETAIL on PM8PM-V

    Hello to all happy new year    Can i install SAPPHIRE RADEON HD2600XT 256MB AGP RETAIL to my system without having problems. Can ayonyone tested this? My System Specs are: CASE: INNOVATOR MIDI TOWER 450W CPU: Intel Pentium 4 Prescott m2 630-3,0GHZ-LG

  • Payslip distribution (in pdf format) via email

    Dear All, Please help with the configuration steps to distribute payslip in PDF Format to their official email.Kindly suggest the steps to configure the same. Best Regards,

  • [Trouble] The Report has no tables

    Hi there, i have trouble in using Crystal Reports for .NET 2.0. I load a CSV file via JetDB Provider as Text database and try to use it as DataSource for my report. But it crashes all time, with a error like this "report has no tables". Actually i do

  • Server keeps refusing HTTP request while uploading file

    Hello, I have written an applet that uses sockets to communicate with HTTP server. I use following string to send a HTTP request to a server: req = "POST /!pages/upload.php HTTP/1.1\r\n";     req += "Host: www.austyn.sk\r\nUser-Agent: Mozilla/5.0 (Wi

  • Cannot print to pdf -There is not enough memory or disk space to repaginate this document

    I don't know if this is an Adobe pdf issue or just a Word 2013 thing. But I'm now using Word 2013 on Win 8.1, 64 bit, 8Gs of Ram and when I go to print a few select pages of my document a message pop out at me saying 'There is not enough memory or di