Aperture backup strategi - file system compatibility

Hi all,
I am beginning to run out of space for a managed library so I'm starting to look for alternatives. Currently I have a managed library, with a vault on another disk on the same computer. This vault is then mirrored to another computer (Linux) via rsync. This seems to work fine.
The solution I'm thinking about it to skip the vault and mirror the library + referenced files directly, while Aperture is not running of course.
The question that arises is whether Aperture does/needs anything "nonstandard" in it's files - resource forks for example. The file names inside the library are a bit weird but within specs for a Unix filesystem so that isn't as problem, but I would hate to have to recover a library and then discover that all the files are "disconnected" because of some minor change that occurred during backup/restore...
Anyone with experience of this (mirroring libraries to a non-Apple filesystem and restoring it)?
PowerMac G5 Dual 2.0 GHz, 2GB RAM Mac OS X (10.4.8) ATI Radeon 9650

OK,
I actually tried this myself instead. I created a new Library, imported a project into it to get some photos with metadata and versions, etc. The masters were relocated to outside the library. I then copied the library to and from a Linux server with rsync, and moved the masters to the file server as well.
After opening Aperture again and reconnecting the masters, all seemed well.

Similar Messages

  • Backup to file system and sbt_tape at the same time?

    hello!
    is it possible to do a rman backup to disk and sbt_tape at the same time (just one backup, not two)? the reason is, that i want to copy the rman backup files from the local file system via robocopy to another server (in a different building), so that i can use them for fast restores in the case of a crash.
    if not, what is the recommended strategy in this case? backups should be available on tape AND the file system of a different machine.
    environment: oracle 10g, windows server 2008, commvault for backup on tape
    thanks for your advice.
    best regards,
    christian

    If you manually copy backupsets out of the FRA to another location, but still accessible as "local disks", you can use the CATALOG command in RMAN to "catalog" the copies.
    Thus, if you opy or move files from the FRA to /mybackups/MYDB, you would use
    "CATALOG START WITH /mybackups/MYDB" or individually "CATALOG BACKUPPIECE /mybackups/MYDB/backuppiece_1" etc
    Once you have moved or removed the backupsets out of the FRA, you must use
    "CROSSCHECK BACKUP"
    and
    "DELETE EXPIRED BACKUP"
    to update the RMAN repository. Else, Oracle will continue to "account" for disks pace consumed by the BackupPieces in V$FLASH_RECOVERY_AREA_USAGE and will soon "run out of space".
    You would do the same for ArchiveLogs (CROSSCHECK ARCHIVELOG ALL; DELETE EXPIRED ARCHIVELOG ALL) if your ArchiveLogs go to the FRA as USE_DB_RECOVERY_FILE_DEST

  • After Restoring/Backup of File System XI Java Instances are not up!

    Hello all,
    We are facing problem in restoring the SAP XI System, after taking backup of the system the <b>java instances</b> in SAP XI System are not starting again. ABAP connections are fine.
    Can anyone provide suggestions/solutions in order to restore the XI System back.
    The system information is as follows.
    System Component:     SAP NetWeaver 2004s, <b>PI 7.0</b>
    Operating System:     SunOS 5.9, SunOS 5.10
    Database:          ORACLE 9.2.0.
    Regards,
    Ketan Patel

    If it´s REALLY a PI 7.0 (SAP_BASIS 700 and WebAS Java 7.00) then it´s not compatible. WebAS 7.00 needs Oracle 10g (http://service.sap.com/pam)
    Also see
    http://service.sap.com/nw2004s
    --> Availibility
    --> SAP NetWeaver 7.0 (2004s) PAM
    If you open the Powerpoint, you will see that Oracle 9 is not listed. I wonder, how you got that installed.
    Neverless, if you recover a Java instance, both filesystem and database content (of the Java schema) must be in sync, means, you need to restore both, database (schema) and filesystem, that have been backed up at the same time.
    Check Java Backup and Restore :
    Restoring the System
           1.      Shut down the system.
           2.      Install a new AS Java system using SAPInst, or restore the file system from the offline backups that you created.
           3.      Import the database backup using the relevant tools provided by the database vendor.
           4.      Overwrite the SAP system directory /usr/sap/.
           5.      Start the system (see Starting and Stopping SAP NetWeaver ABAP and Java.)
    The J2EE Engine is restored with the last backup.
    Markus

  • Backup into file system

    Hi
    Setting backup-storage with the follwoing configuration is not generating backup files under said location - we are pumping huge volume of data and data(few GB) is not getting backuped up into file system - can you let me know if what is that I missing here?
    Thanks
    sunder
    <distributed-scheme>
         <scheme-name>distributed-Customer</scheme-name>
         <service-name>DistributedCache</service-name>
         <!-- <thread-count>5</thread-count> -->
         <backup-count>1</backup-count>
         <backup-storage>
         <type>file-mapped</type>
         <directory>/data/xx/backupstorage</directory>
         <initial-size>1KB</initial-size>
         <maximum-size>1KB</maximum-size>
         </backup-storage>
         <backing-map-scheme>
              <read-write-backing-map-scheme>
                   <scheme-name>DBCacheLoaderScheme</scheme-name>
                   <internal-cache-scheme>
                   <local-scheme>
                        <scheme-ref>blaze-binary-backing-map</scheme-ref>
                   </local-scheme>
                   </internal-cache-scheme>
                   <cachestore-scheme>
                        <class-scheme>
                             <class-name>com.xxloader.DataBeanInitialLoadImpl
                             </class-name>
                             <init-params>
                                  <init-param>
                                       <param-type>java.lang.String</param-type>
                                       <param-value>{cache-name}</param-value>
                                  </init-param>
                                  <init-param>
                                       <param-type>java.lang.String</param-type>
                                       <param-value>com.xx.CustomerProduct
                                       </param-value>
                                  </init-param>
                                  <init-param>
                                       <param-type>java.lang.String</param-type>
                                       <param-value>CUSTOMER</param-value>
                                  </init-param>
                             </init-params>
                        </class-scheme>
                   </cachestore-scheme>
                   <read-only>true</read-only>
              </read-write-backing-map-scheme>
         </backing-map-scheme>
         <autostart>true</autostart>
    </distributed-scheme>
    <local-scheme>
    <scheme-name>blaze-binary-backing-map</scheme-name>
    <high-units>{back-size-limit 1}</high-units>
    <unit-calculator>BINARY</unit-calculator>
    <expiry-delay>{back-expiry 0}</expiry-delay>
    <cachestore-scheme></cachestore-scheme>
    </local-scheme>

    Hi
    We did try out with the following configuration
    <near-scheme>
         <scheme-name>blaze-near-HeaderData</scheme-name>
    <front-scheme>
    <local-scheme>
    <eviction-policy>HYBRID</eviction-policy>
    <high-units>{front-size-limit 0}</high-units>
    <unit-calculator>FIXED</unit-calculator>
    <expiry-delay>{back-expiry 1h}</expiry-delay>
    <flush-delay>1m</flush-delay>
    </local-scheme>
    </front-scheme>
    <back-scheme>
    <distributed-scheme>
    <scheme-ref>blaze-distributed-HeaderData</scheme-ref>
    </distributed-scheme>
    </back-scheme>
    <invalidation-strategy>present</invalidation-strategy>
    <autostart>true</autostart>
    </near-scheme>
    <distributed-scheme>
    <scheme-name>blaze-distributed-HeaderData</scheme-name>
    <service-name>DistributedCache</service-name>
    <partition-count>200</partition-count>
    <backing-map-scheme>
    <partitioned>true</partitioned>
    <read-write-backing-map-scheme>
    <internal-cache-scheme>
    <external-scheme>
    <high-units>20</high-units>
    <unit-calculator>BINARY</unit-calculator>
    <unit-factor>1073741824</unit-factor>
    <nio-memory-manager>
    <initial-size>1MB</initial-size>
    <maximum-size>50MB</maximum-size>
    </nio-memory-manager>
    </external-scheme>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <class-name>
    com.xx.loader.DataBeanInitialLoadImpl
    </class-name>
    <init-params>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>com.xx.bean.HeaderData</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>SDR.TABLE_NAME_XYZ</param-value>
    </init-param>
    </init-params>
    </class-scheme>
    </cachestore-scheme>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
    <backup-count>1</backup-count>
    <backup-storage>
    <type>off-heap</type>
    <initial-size>1MB</initial-size>
    <maximum-size>50MB</maximum-size>
    </backup-storage>
    <autostart>true</autostart>
    </distributed-scheme>
    With configuration the amount of residual main memory consumption is like 15 G.
    When we changed this configuration to
    <near-scheme>
         <scheme-name>blaze-near-HeaderData</scheme-name>
    <front-scheme>
    <local-scheme>
    <eviction-policy>HYBRID</eviction-policy>
    <high-units>{front-size-limit 0}</high-units>
    <unit-calculator>FIXED</unit-calculator>
    <expiry-delay>{back-expiry 1h}</expiry-delay>
    <flush-delay>1m</flush-delay>
    </local-scheme>
    </front-scheme>
    <back-scheme>
    <distributed-scheme>
    <scheme-ref>blaze-distributed-HeaderData</scheme-ref>
    </distributed-scheme>
    </back-scheme>
    <invalidation-strategy>present</invalidation-strategy>
    <autostart>true</autostart>
    </near-scheme>
    <distributed-scheme>
    <scheme-name>blaze-distributed-HeaderData</scheme-name>
    <service-name>DistributedCache</service-name>
    <partition-count>200</partition-count>
    <backing-map-scheme>
    <partitioned>true</partitioned>
    <read-write-backing-map-scheme>
    <internal-cache-scheme>
    <external-scheme>
    <high-units>20</high-units>
    <unit-calculator>BINARY</unit-calculator>
    <unit-factor>1073741824</unit-factor>
    <nio-memory-manager>
    <initial-size>1MB</initial-size>
    <maximum-size>50MB</maximum-size>
    </nio-memory-manager>
    </external-scheme>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <class-name>
    com.xx.loader.DataBeanInitialLoadImpl
    </class-name>
    <init-params>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>com.xx.bean.HeaderData</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>SDR.TABLE_NAME_XYZ</param-value>
    </init-param>
    </init-params>
    </class-scheme>
    </cachestore-scheme>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
    <backup-count>1</backup-count>
    <backup-storage>
    <type>file-mapped</type>
    <initial-size>1MB</initial-size>
    <maximum-size>100MB</maximum-size>
    <directory>/data/xxcache/blazeload/backupstorage</directory>
    <file-name>{cache-name}.store</file-name>
    </backup-storage>
    <autostart>true</autostart>
    </distributed-scheme>
    Note backup storage is file-mapped
    <backup-storage>
    <type>file-mapped</type>
    <initial-size>1MB</initial-size>
    <maximum-size>100MB</maximum-size>
    <directory>/data/xxcache/blazeload/backupstorage</directory>
    <file-name>{cache-name}.store</file-name>
    </backup-storage>
    We still see that process residual main memory consumption is 15 G and we also see that /data/xxcache/blazeload/backupstorage folder is empty.
    Wanted to check where does backup storage maintains the information - we would like offload this to flat file.
    Appreciate any pointers in this regard.
    Thanks
    sunder

  • ASM RMAN backup to File System

    Hi all,
    I have a rman backup (datafile and controlfile) which was took in an ASM instance (not a RAC) ORACLE 11.2.0.2 in a Linux server, now I want restore the backup in a new database in windows/Linux OS using general File System storage (single instance rdbms) instead of ASM.
    Is this possible?
    Can I restrore an ASM rman backup in a file system storage mechanisim in a new server?
    Kindly clarify my question.
    Thanks in Advance..
    Nonuday

    Nonuday wrote:
    Hi Levi,
    Thanks for your invaluable script and blog.
    can you clarify me on this query:
    I have a RMAN backup taken from ASM and the backup is database and controlf file backup which contains datafiles and controlfiles.
    Now I need to restore this on my system and here I dont use ASM or archive log, I use single instance in no archive log mode database.
    I have restored the control file from the RMAN controfile backup.
    Before restoring the control file I have checked the orginal pfile of the backup database which had parameters like
    'db_create_file_dest',
    'db_create_online_log_dest',
    'db_recovery_file_dest_size',
    'db_recovery_dest',
    'log_archive_dest'.
    Since I am not gng to create a DB in no archive log mode, I didnt use any of the above parameters and created a database.
    Now my question is:
    If i restore the database and the datafile will get restored and after renaming all the logfiles, database will be opened.
    I want to know whether this method is correct or wrong and will the database work as it was working previously. Or do i need create the db_file_recovery and other parameters also for this database.About Parameter:
    All these parameters should reflect your current environment any reference to the old environment must be modified.
    About Filesystem used:
    Does not matter what Filesystem you are using the File (datafile/redolog/controlfile/archivelog/backuppiece) are created on Binary Format which depend on Platform only. So, The same binary file ( e.g datafile) have same format and content on raw device, ASM, ext3, ext2, and so on. So, to database it's only a location where file are stored, but the file are the same. ASM has a different architecture from Regular Filesystem and need be managed in a different manner (i.e using RMAN).
    About Database:
    Since your database files are the same even using different filesystem what you need is rename your datafiles/redofiles on controlfile during restore, the redo files will be recreated.
    So, does not matter if you database are noarchivelog or archivelog, the same way which you will do a restore on ASM is the same way to restore on Regular Filesystem. (it's only about renaming database file on controlfile during restore)
    On blog the post "How Migrate All Files on ASM to Non-ASM (Unix/Linux)" is about move the file from filesystem to another. But you can modify the script used to restore purposes;
    ## set newname tell to RMAN where file will be restored and keep this files location on memory buffer
    RMAN> set newname for datafile 1 to <location>;
    ### swich get list of files from memory buffer (rman) and rename on controlfile the files already restored.
    RMAN>switch datafile/tempfile all ;With database mounted use this script below:
    I just commented three lines that are unnecessary in your case.
    SET serveroutput ON;
    DECLARE
      vcount  NUMBER:=0;
      vfname VARCHAR2(1024);
      CURSOR df
      IS
        SELECT file#,
          rtrim(REPLACE(name,'+DG_DATA/drop/datafile/','/u01/app/oracle/oradata/drop/'),'.0123456789') AS name
        FROM v$datafile;
      CURSOR tp
      IS
        SELECT file#,
          rtrim(REPLACE(name,'+DG_DATA/drop/tempfile/','/u01/app/oracle/oradata/drop/'),'.0123456789') AS name
        FROM v$tempfile;
    BEGIN
    --  dbms_output.put_line('CONFIGURE CONTROLFILE AUTOBACKUP ON;'); ### commented
      FOR dfrec IN df
      LOOP
        IF dfrec.name  != vfname THEN
          vcount      :=1;
          vfname     := dfrec.name;
        ELSE
          vcount := vcount+1;
          vfname:= dfrec.name;
        END IF;
      --  dbms_output.put_line('backup as copy datafile ' || dfrec.file# ||' format  "'||dfrec.name ||vcount||'.dbf";');  ### commented
      END LOOP;
      dbms_output.put_line('run');
      dbms_output.put_line('{');
      FOR dfrec IN df
      LOOP
        IF dfrec.name  != vfname THEN
          vcount      :=1;
          vfname     := dfrec.name;
        ELSE
          vcount := vcount+1;
          vfname:= dfrec.name;
        END IF;
        dbms_output.put_line('set newname for datafile ' || dfrec.file# ||'  to  '''||dfrec.name ||vcount||'.dbf'' ;');
      END LOOP;
      FOR tprec IN tp
      LOOP
        IF tprec.name  !=  vfname THEN
          vcount      :=1;
          vfname     := tprec.name;
        ELSE
          vcount := vcount+1;
          vfname:= tprec.name;
        END IF;
        dbms_output.put_line('set newname for tempfile ' || tprec.file# ||'  to  '''||tprec.name ||vcount||'.dbf'' ;');
        END LOOP;
          dbms_output.put_line('restore database;');
        dbms_output.put_line('switch tempfile all;');
        dbms_output.put_line('switch datafile all;');
        dbms_output.put_line('recover database;');
        dbms_output.put_line('}');
    ---   dbms_output.put_line('alter database open;');  ### comented because you need rename your redologs on controlfile before open database
        dbms_output.put_line('exit');
    END;
    /After restore you must rename your redologs on controlfile from old location to new location:
    e.g
    ##  use this query to get current location of redolog
    SQL>  select group#,member from v$logfile order by 1;
    ## and change from <old_location> to <new_location>
    SQL > ALTER DATABASE
      RENAME FILE '+DG_TSM_DATA/tsm/onlinelog/group_3.263.720532229' 
               TO  '/u01/app/oracle/oradata/logs/log3a.rdo'  When you change all redolog on controlfile issue command below:
    SQL> alter database open resetlogs;PS: Always track database in real time using alert log file of database.
    HTH,
    Levi Pereira

  • Backup root file system on DVD

    I have a root file system about 4Gb large.
    Can I back it up on a DVD?
    I read the ufsdump works with tapes.
    Can it work with a DVD?
    If yes, what is the command format?

    I have a root file system about 4Gb large.
    Can I back it up on a DVD?
    I read the ufsdump works with tapes.
    Can it work with a DVD?
    If yes, what is the command format?

  • Raw Device Backup to file system(OPS 8i)

    Hi
    Our currently setup is
    Oracle database 8.1.6 (Oracle Parallel Server) Two Node
    Noarchive Mode
    Solaris 2.6
    all database file ,redo logfiles,controlfiles under raw device.
    database size 16 G.B
    oracle block size 8192
    currently we are using only export backup of oracle.
    But now i want to take cold backup of oracle database to disk.
    cold backup Raw --> Disk
    How we can take cold backup with dd command and skip parameter ?
    Is anybody have practical idea of dd command with skip parameter.
    Thanks and regards
    Kuljeet pal singh

    you can use ufsdump instead of dd

  • Do we need to backup OS (file systems) on Exadata storage cells?

    We got confusing messages about if we need to or not. I'd like to hear your opinons.
    Thanks!

    Hi,
    The answer is no.
    There is no need to backup the OS of the storage cell.
    Worst case a complete storage cell needs to be replaced. A field engineer will open your broken storage cell and take out the onboard USB and put it inside the new storage cell.
    The system will boot from this USB and you can choose to implement the 'old' configuration on the new cell.
    Regards,
    Tycho

  • Using old file system backup for Cloning

    I have taken an off-line backup of Oracle 11i (11.5.10.2) 15 days ago. Before taking backup of file system , I verified that all the latest Rapid Clone Patches are applied. No changes or patch work in APPL_TOP or DB has been done since that backup. Now I need to do cloning of this instance, how I can use this backup for Cloning.
    Rapid Clone scripts create and generate some files/directories so I am not sure whether my Old backup of file system will work or not. What is the best way to use old backup for cloning , also what are the file and directories in addition to the old backup of file system which I need to copy to Target System.
    Thanks for reviewing and suggestions.
    Samar

    Samar,
    If you have run preclone before backing it up, your backup should be valid for cloning.
    2.1 in the cloning doc has to be in the backup.
    These doc's should clear out yours doubts on cloning.
    Cloning Oracle Applications Release 11i with Rapid Clone
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=230672.1
    FAQ: Cloning Oracle Applications Release 11i
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=216664.1

  • RMAN to Disk (shared mounted file system) then secure backup to Tape

    Hi
    This is a little away from the norm and I am new to Oracle secure Backup.
    We have several databases on physically separate servers, all backup to a central disk (a ZFS shared file-system).
    We have a media server also with the same mounted file-system so it can see all of the RMAN backups there.
    Secure backup is installed on the media server and is configured as such.
    The question I have is I need to backup the file system to tape where all the RMAN backups live, I have configured the data set but I get file permission errors for each of the RMAN backup files in the directory.
    I have tried to change these permissions but to no avail (assuming it is just a read write access change that is needed - but this may be a problem in the long run). What is the general process for backup of already created RMAN backups sat in a shared area? I know its not the norm to do to disk then to tape backups but can this be done? I would have installed Secure backup client on each server and managed the whole backup through secure backup but this is not possible I must do the to tape from the file system, any advise and guidance on this would be much appreciated.
    Kind regards
    Vicky
    Edited by: user10090654 on Oct 4, 2011 4:50 AM

    You can easily accomplish a very streamlined D2D2T strategy! RMAN backup to disk...then backup that disk to tape via RMAN and OSB. Upon restore, RMAN will restore from best media disk or tape based on where files are located.
    Donna

  • Using ufsrestore on multiple file systems

    I have a ufsdump script that allows me to backup the file systems on c0t0d0 all at once. Now I am looking for a way to restore all of these file systems at one time using a script. I would greatly appreciate any help.
    Cletus

    Is this the boot disk ?

  • ASM Vs File system

    1. With file system, we were able to set some threshold alerts at the OS mount point level. Is this possible with ASM since it is a raw device at the OS level?
    2. The ASM directories are logical and is visible ONLY at the oracle ASM level. True?
    3. Is cold backup an option with ASM?
    4. Is RMAN the only solution with ASM disks?
    5. Are there any suggested Hardware snap/clone available for ASM storage?
    Thanks for your insight on this.

    KR wrote:
    1. With file system, we were able to set some threshold alerts at the OS mount point level. Is this possible with ASM since it is a raw device at the OS level?
    2. The ASM directories are logical and is visible ONLY at the oracle ASM level. True?
    3. Is cold backup an option with ASM?
    4. Is RMAN the only solution with ASM disks?
    5. Are there any suggested Hardware snap/clone available for ASM storage?1. Not possible from an OS point-of-view. See earlier reply for workaround;
    2. Yep, or use asmcmd (from the ASM ORACLE_HOME) or ftp
    3. If you use RMAN to cold-backup: yes. You cannot use a regular backup to back up your devices, unless you have designated files as 'disks' to ASM (But who would want that?). In this case you should shut down ASM and backup the file system where these files reside.
    4. Yes, it is.
    5. That depends on your hardware configuration. It is possible to run ASM on concurrent/shared, host based mirrored volume groups/volumes. If you use SAN technology, you can always verify whether these options are available from the storage manufacturer. However, I would be really careful with these, always taking things down before taking snapshots/flashcopies/clones/whatever you call them.
    HTH.
    Arnoud Roth

  • Back up - file system

    This question is on backup.
    Every time i take backup of file system for app folder(\\Hyperion\AnalyticServices) from my essbase server.
    Is this the right way backing up of file system.if any of database corroupts, Can i just rplace the app folder with the backup app folder , does this work.

    This is one way to do the backup. If an application corrupted, you would just have to restore the application directory under the app folder. That is unless you have stored files on another drive. If you have gone into the storage tab of properties and changed where the page and index files reside, you would also have to back up that drive (same directory starting point) at the same time as your other drive backup.

  • Why would anyone want to use ASM Clustered File system?

    DB Version: 11gR2
    OS : Solaris, AIX, HP-UX
    I've read about the new feature ACFS.
    http://www.oracle-base.com/articles/11g/ACFS_11gR2.php
    But why would anyone want to store database binaries in a separate Filesystem created by Oracle?

    Hi Vitamind,
    how do these binaries interact with the CPU when they want something to be done?
    ACFS should work with Local OS (Solaris) to communicate with the CPU . Isn't this kind of double work?ACFS dont work with .... but provide filesystem to Local S.O
    There may be extra work, but that's because there are more resources that a common filesystem.
    Oracle ACFS executes on operating system platforms as a native file system technology supporting native operating system file system application programming interfaces (APIs).
    ACFS is a general purpose POSIX compliant cluster file system. Being POSIX compliant, all operating system utilities we use with ext3 and other file systems can also be used with Oracle ACFS given it belongs to the same family of related standards.
    ACFS Driver Model
    An Oracle ACFS file system is installed as a dynamically loadable vendor operating system (OS) file system driver and tool set that is developed for each supported operating system platform. The driver is implemented as a Virtual File System (VFS) and processes all file and directory operations directed to a specific file system.
    It makes sense you use the ACFS if you use some of the features below:
    • Oracle RAC / RAC ONE NODE
    • Oracle ACFS Snapshots
    • Oracle ASM Dynamic Volume Manager
    • Cluster Filesystem for regular files
    ACFS Use Cases
    • Shared Oracle DB home
    • Other “file system” data
    • External tables, data loads, data extracts
    • BFILES and other data customer chooses not to store in db
    • Log files (consolidates access)
    • Test environments
    • Copy back a previous snapshot after testing
    • Backups
    • Snapshot file system for point-intime backups
    • General purpose local or cluster file system
    • Leverage ASM manageability
    Note : Oracle ACFS file systems cannot be used for an Oracle base directory or an Oracle grid infrastructure home that contains the software for Oracle Clusterware, Oracle ASM, Oracle ACFS, and Oracle ADVM components.
    Regards,
    Levi Pereira

  • Hi, I am using my icloud on my iphone but am unable to find it on my mac book pro. It's not under my system preferences like it shows under the setup. Can I download it again? Or something. I just need to backup some files on my computer and am unable.

    Hi, I am using my icloud on my iphone but am unable to find it on my mac book pro. It's not under my system preferences like it shows under the setup. Can I download it again? Or something. I just need to backup some files on my computer and am unable.

    The minimum requirement for iCloud is Lion 10.7.5 (Mavericks preferred): the iCloud Preference Pane does not appear on earlier systems - the MobileMe pane appears on Lion and earlier but is now non-functional - you cannot now open or access a MobileMe account.
    To make use of iCloud you will have to upgrade your Mac to Lion or Mavericks, provided it meets the requirements.
    The requirements for Lion are:
    Mac computer with an Intel Core 2 Duo, Core i3, Core i5, Core i7, or Xeon processor
    2GB of memory
    OS X v10.6.6 or later (v10.6.8 recommended)
    7GB of available space
    Lion is available in the Online Apple Store ($19.99). Mountain Lion (10.8.x) is also available there at the same price but there seems little point as the system requirements are the same for Mavericks (10.9.x) - which is free - unless you need to run specific software which will run on Mountain Lion only.
    The requirements for Mountain Lion and Mavericks are:
    OS X v10.6.8 or later
    2GB of memory
    8GB of available space
      and the supported models are:
    iMac (Mid 2007 or newer)
    MacBook (Late 2008 Aluminum, or Early 2009 or newer)
    MacBook Pro (Mid/Late 2007 or newer)
    Xserve (Early 2009)
    MacBook Air (Late 2008 or newer)
    Mac mini (Early 2009 or newer)
    Mac Pro (Early 2008 or newer)
    It is available from the Mac App Store (in Applications).
      You should be aware that PPC programs (such as AppleWorks) will not run on Lion or above; and some other applications may not be compatible - there is a useful compatibility checklist at http://roaringapps.com/apps:table

Maybe you are looking for

  • Warning - using HomePlug Mains network adapters wi...

    Thanks to working with the kind mods from BT I've now figured out the cause of my Infinity "go slow". Been using Homeplug Turbo 85Mbps network adaptors to connect from the Hub to the PC in my Study (upstairs, when the hub is downstairs). Despite init

  • ORA-279 signalled during: ALTER DATABASE RECOVER  automatic standby databas

    Hi, oracle server:8.1.7.0.0 os:solaris 5.9 we created standby server by taking coldbackup of primary server and we are using rsync to syschronize the archive log files between production and standby servers but every over we recover the standby serve

  • Interactive pdf/excel

    Hello everybody. How to create an interactive pdf file, that has a button (calculate), when pressed puts input data from pdf to excel (few numbers), runs the excel calculation, and writes the excel output back to pdf? When created, the interaction sh

  • Trying to open rosetta but not an option in menu

    I'm trying to open black and white the game using rosetta as it is a power pc game and not intel. I cant seem to find the option to use rosetta in the 'get info' menu. However i can find it for almost every other application. What am i doing wrong? I

  • Lightroom 5.6, issue with previews displayed blank!

    Try to regenerate previews, not working and by the end of the process, LR says they already exist! This is the same issue for all my catalogs! I also tried to restore a backup, doing the same! First issue appears under LR5.5, I migrated to 5.6 to see