Question about archive destination.

Hi All,
When we set the archived destination to this parameter log_archive_dest like:
SQL>alter system set LOG_ARCHIVE_DEST_.. = 'location=/archivelogpath reopen';
Could you help me to know about 'reopen' keyword here. Please help me know, what does it mean and whats the use of it. I have failed to get this in oracle documentation.
Best Regards,
Michel.

From the documentation:
Rearchiving to a Failed Destination
Use the REOPEN attribute of the LOG_ARCHIVE_DEST_n parameter to specify whether and when ARCn should attempt to rearchive to a failed destination following an error. REOPEN applies to all errors, not just OPEN errors.
REOPEN=n sets the minimum number of seconds before ARCn should try to reopen a failed destination. The default value for n is 300 seconds. A value of 0 is the same as turning off the REOPEN attribute; ARCn will not attempt to archive after a failure. If you do not specify the REOPEN keyword, ARCn will never reopen a destination following an error.
You cannot use REOPEN to specify the number of attempts ARCn should make to reconnect and transfer archived logs. The REOPEN attempt either succeeds or fails.
When you specify REOPEN for an OPTIONAL destination, the database can overwrite online logs if there is an error. If you specify REOPEN for a MANDATORY destination, the database stalls the production database when it cannot successfully archive. In this situation, consider the following options:
Archive manually to the failed destination.
Change the destination by deferring the destination, specifying the destination as optional, or changing the service.
Drop the destination.
When using the REOPEN keyword, note the following:
ARCn reopens a destination only when starting an archive operation from the beginning of the log file, never during a current operation. ARCn always retries the log copy from the beginning.
If you specified REOPEN, either with a specified time the default, ARCn checks to see whether the time of the recorded error plus the REOPEN interval is less than the current time. If it is, ARCn retries the log copy.
The REOPEN clause successfully affects the ACTIVE=TRUE destination state. The VALID and ENABLED states are not changed.

Similar Messages

  • Subtitling (question about archived material)

    OK, I've looked through the archived questions about how to subtitle...and I just want to make sure I do this right. If I do subtitles in DVDSP, do I write them up in a text file first and then import them, or just type them in manually?
    If I do subtitles in FCP I just use something like Title3D and type them in manually...creating a whole buncha overlay (title) clips, right?
    Forgive the redundency, I know how subtitles are supposed to look, just trying to find the best (and least time consuming) way to produce them.
    FYI...the subtitles will be of a person from Spain speaking English (it's kinda hard to understand him since he talks fast and has a heavy Spanish accent).
    Thanks, all, in advance.
    Jonathan

    subtitling (question about archived material)< </div>
    So you are trying put subtitles on your archived material?
    If you use the subtitle track in DVDSP, the text will be switchable by the user. (This is made more complicated by several factors like your desire to make the subtitles permanent or uncontrollable by the user. These are options you can control by programming the DVDSP project. Not difficult but very confusing.)
    If you use a text tool in FCP, the titles will be permanently burned into the video image.
    If you use some fo the more sophisticated (read: Expensive) subtitling tools for FCP, you can export a caption-compliant file that can be printed to videotape or mastered to DVD. If you own one of these products, you are at the complete mercy of their manual's writer.
    bogiesan

  • Question about Archive Log Deletion policy

    I've a problem to understand the Archive Log Deletion policy, and I I'd like to this problem explain with the following example.
    Messages of the database are in German, but I guess you'll understand them.
    SQL> startup
    ORACLE-Instance hochgefahren.
    Total System Global Area 5344731136 bytes
    Fixed Size                  2129240 bytes
    Variable Size            2684355240 bytes
    Database Buffers         2617245696 bytes
    Redo Buffers               41000960 bytes
    Datenbank mounted.
    Datenbank geöffnet.
    SQL> archive log list
    Datenbank-Log-Modus              Archive-Modus
    Automatische Archivierung             Aktiviert
    Archivierungsziel            E:\oracle\thetis_iv\arch
    Älteste Online-Log-Sequenz     17917
    Nächste zu archivierende Log-Sequenz   17919
    Aktuelle Log-Sequenz           17919
    SQL> alter system switch logfile;
    System wurde geändert.I created a brand new archive log.
    SQL> exit
    Verbindung zu Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options beendet
    D:\OracleDB\product\11.1.0\db_1\BIN>dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:04    <DIR>          .
    06.04.2011  15:04    <DIR>          ..
    06.04.2011  15:04        17.137.152 ARC17919_0721667907.001
                   1 Datei(en),     17.137.152 Bytes
                   2 Verzeichnis(se), 41.073.258.496 Bytes freiand this is the only archive log in the directory. Now I start rman:
    D:\OracleDB\product\11.1.0\db_1\BIN>rman target / catalog rmanrepo@rmanrepo
    Recovery Manager: Release 11.1.0.7.0 - Production on Mi Apr 6 15:05:35 2011
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Mit Ziel-Datenbank verbunden: ENTWIV (DBID=21045568)
    Kennwort für Recovery-Katalog-Datenbank:
    Verbindung mit Datenbank des Recovery-Katalogs
    RMAN> show all;
    RMAN-Konfigurationsparameter für Datenbank mit db_unique_name ENTWIV sind:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO 'SBT_TAPE';
    CONFIGURE CONTROLFILE AUTOBACKUP OFF;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO 'E:\oracle\thetis_iv\backup\CF_%F_ENTWIV.ORA';
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE SBT_TAPE TO '%F'; # default
    CONFIGURE DEVICE TYPE SBT_TAPE PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE 'SBT_TAPE' PARMS  'ENV=(TPDO_OPTFILE=D:\Services\Tivoli\TSM\AgentOBA64\tpdo.opt)';
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BZIP2'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO 'SBT_TAPE';
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO 'D:\ORACLEDB\PRODUCT\11.1.0\DB_1\DATABASE\SNCFENTWIV.ORA'; # defaultThe archive log deletion policy says the the logfiles have to be backed up for two times before they get deleted.
    Now I backup all archive logs, that havn't been backed up for at least two times.
    RMAN> run { backup archivelog all not backed up 2 times
    2>       format '%d_AR_%Y%M%D_%s_%t'
    3>       tag 'ARCHIVE LOGS'
    4>       DELETE ALL INPUT;
    5>     }
    Starten backup um 06.04.2011 15:08:01
    Aktuelles Log archiviert
    Zugewiesener Kanal: ORA_SBT_TAPE_1
    Kanal ORA_SBT_TAPE_1: SID=253 Device-Typ=SBT_TAPE
    Kanal ORA_SBT_TAPE_1: Data Protection for Oracle: version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set für Archive Log wird begonnen
    Kanal ORA_SBT_TAPE_1: Archive Logs in Backup Set werden angegeben
    Eingabe-Archive-Log-Thread=1 Sequence=17919 RECID=147 STAMP=747759899
    Eingabe-Archive-Log-Thread=1 Sequence=17920 RECID=148 STAMP=747760081
    Kanal ORA_SBT_TAPE_1: Piece 1 wird auf 06.04.2011 15:08:02 begonnen
    Kanal ORA_SBT_TAPE_1: Piece 1 auf 06.04.2011 15:08:09 beendet
    Piece Handle=ENTWIV_AR_20110406_23_747760082 Tag=ARCHIVE LOGS Kommentar=API Version 2.0,MMS Version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set vollstõndig, abgelaufene Zeit: 00:00:08
    Kanal ORA_SBT_TAPE_1: Archive Logs werden gel÷scht
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17919_0721667907.001 RECID=147 STAMP=747759899
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17920_0721667907.001 RECID=148 STAMP=747760081
    Beendet backup um 06.04.2011 15:08:10
    RMAN> exit
    Recovery Manager abgeschlossen.
    D:\OracleDB\product\11.1.0\db_1\BIN> dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:08    <DIR>          .
    06.04.2011  15:08    <DIR>          ..
                   0 Datei(en),              0 Bytes
                   2 Verzeichnis(se), 41.090.396.160 Bytes freirman deleted all archive logs, even I they are on tape only once by now.
    Thats not what I expected. Where is my mistake?

    Hi,
    I do new tests it's very strange.
    BACKUP ARCHIVELOG command is not obeying the policy of archivelog.
    You can open a SR on MOS. (to check bugs)
    I reproduce the same test and the result was the same, it seems that this is a bug.
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO DISK;
    RMAN> backup archivelog all not backed up 2 times delete all input;
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=1 sequence=15 RECID=16 STAMP=747753711
    input archived log thread=2 sequence=20 RECID=17 STAMP=747753714
    input archived log thread=1 sequence=16 RECID=19 STAMP=747753729
    input archived log thread=2 sequence=21 RECID=18 STAMP=747753729
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t132210_0.304.747753731 tag=TAG20110406T132210 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    channel ORA_DISK_1: deleting archived log(s)
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_15.293.747753711 RECID=16 STAMP=747753711
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_20.295.747753715 RECID=17 STAMP=747753714
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_16.294.747753729 RECID=19 STAMP=747753729
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_21.298.747753729 RECID=18 STAMP=747753729
    Finished backup at 06-APR-11
    RMAN> list archivelog all;
    specification does not match any archived log in the repositoryOracle Docs Says:
    The BACKUP ARCHIVELOG ... DELETE INPUT command deletes archived log files after they are backed up.
    This command eliminates the separate step of manually deleting archived redo logs.
    With DELETE INPUT, RMAN deletes only the specific copy of the archived log chosen for the backup set.
    With DELETE ALL INPUT, RMAN deletes each backed-up archived redo log file from all log archiving destinations.
    As explained in "Configuring an Archived Redo Log Deletion Policy",
    the BACKUP ... DELETE INPUT and DELETE ARCHIVELOG commands obey the archived redo log deletion policy
    for logs in all archiving locations. For example, if you specify that logs should only be deleted when backed
    up at least twice to tape, then BACKUP ... DELETE honors this policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10642/rcmbckba.htm#BRADV89524
    But in ours case it's not honors this policy.
    Only with the FORCE command should this happen. But it is not our case.
    Oracle Docs:
    If FORCE is not specified on the deletion commands,
    then these deletion commands obey the archived log deletion policy.
    If FORCE is specified, then the deletion commands ignore the archived log deletion policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10643/rcmsynta010.htm#RCMRF113
    Alternatively you can do the following:
    Set the commands separately.
    Check this:
    RMAN>  run {
    2> backup archivelog all not backed up 2 times ;
    3> delete archivelog all backed up 2 times to disk;
    4> }
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=2 sequence=22 RECID=21 STAMP=747755128
    input archived log thread=1 sequence=17 RECID=20 STAMP=747755127
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t134528_0.295.747755129 tag=TAG20110406T134528 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    Finished backup at 06-APR-11
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=78 instance=orcl1 device type=DISK
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_17.298.747755127 thread=1 sequence=17
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_22.294.747755129 thread=2 sequence=22
    RMAN>Edited by: Levi Pereira on Apr 6, 2011 1:35 PM

  • Question about archived logs

    Hi!
    I wonder if there is possibility to configure archiving policy in a way:
    set one destination (/arch1/oracle) as primary, and another (/arch2/oracle) as optionally.
    redo Logs will be copied to /arch2/oracle only if /arch1/oracle fails, but after it will be available again, files will be moved (or copied) from /arch2/oracle to /arch1/oracle. Anybody tried to store backup files in that way? is it possible? I tried various combinations of log_archive_dest_n and log_archive_dest_state_n parameters, but it won't work in a way I expected.
    Message was edited by:
    wojaczek

    it's about limitation of disk place on server where DB operates. Archived logs are stored on NFS filesystem, but time to time there are some problems with remote filesystem. In such cases DB will hang and manual change of arch log destination to local FS is needed. But this destination could not be set permanently because of disk space limitations.
    When i set destination 1 as mandatory and 2 as optionally db hangs when arc can't use mandatory destination.
    Setting both destinations as optionally gives finally inconsistency and storing files on local fs, what is not desirable

  • A question about archive and install

    Recently my system crashed due to a faulty OS instal.
    So, I hauled my G4 to the mac store and they resoved the problem for me and told me to do a archive and instal, and even gave me instructions.
    Everything was going fine untill it came the time to choose which directory to install to... I clicked on my hard drive and a dialouge box popped up saying that it couldn't install older versions of the files over newer versions....
    So now I've hit a dead end.
    How can I continue the install?

    WanderingAimlessly:
    Do you currently have a functioniing OS on your HDD?
    Do you have your computer backed up?
    If your answers are "no" to the first question and "yes" to the second, I suggest you go a completer reformat, zero all data and install.
    Post back with your responses, and we will decide where to go from there.
    Good luck.
    cornelius

  • Question about Archiving Mailboxes

    I keep almost all of my mail On My Mac > Personal Mail, then divided into subfolders beneath that.  I back up to Time Machine.
    This morning I noticed one of those sub folders was missing, and had to go back more than 24 hours to recover it.  Not sure how it went missing, but it did.
    I am wondering what is the diffference between using Archive Mailbox under the Mailbox menu and keeping the messages in folders On My Mac?
    I think my mailboxes are just getting too darn big, try as I might to contain them.  I'm a mail hoarder.
    Thanks for any insight

    How you backup and what you backup to is all about taking a calculated risk. In other words, what's it worth to you?
    Ultimately, whatever method you choose, I think redundancy is the most important factor. It doesn't matter if you have half-a-dozen backups, if they're all in the same place and the building burns down, they're all gone.
    Yes, hard drives are flimsy and fail, but you can easily duplicate a backup and store them in separate locations; then, make regular maintenance checks to see if they're viable. The probability of two backups failing simultaneously is astronomically low; the probability of one backup failing is astronomically high. For the average person, most dedicated backup systems are too far out of reach, pricewise and from an infrastructure standpoint. Duplicating hard drives is pretty crude, but it's cheap so long as you exercise due diligence.

  • Questions about archive

    Hope anyone can help me understand this…
    I found there are various kinds of archive feature in Exchange or Outlook. And I need help to choose the appropriate method to do archive in my Exchange.
    Thanks

    Hi David,
    Exchange Server 2010 has a new built-in archiving solution, which lets you create a personal archive mailbox within the Exchange organization.
    You can use the Enable-Mailbox "xxx" -Archive cmdlet to enable the personal archive for a user's mailbox.
    What's more, here is an article for your reference.
    Understanding Personal Archives
    http://technet.microsoft.com/en-us/library/dd979795(v=exchg.141).aspx
    Hope it helps.
    If you need further assistance, please feel free to let me know.
    Best regards,
    Amy
    Amy Wang
    TechNet Community Support

  • Question about ALTER SYSTEM ARCHIVE LOG START

    Good morning,
    I'm trying (unsuccessfully) to get my database to be in archive log mode.
    These are the steps I followed:
    SQL> shutdown immediate;
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> startup mount;
    ORACLE instance started.
    Total System Global Area  535662592 bytes
    Fixed Size                  1375792 bytes
    Variable Size             377487824 bytes
    Database Buffers          150994944 bytes
    Redo Buffers                5804032 bytes
    Database mounted.
    SQL> select log_mode from v$database;
    LOG_MODE
    ARCHIVELOG
    SQL> show parameter log_archive_start;
    NAME                                 TYPE        VALUE
    log_archive_start                    boolean     FALSE
    SQL> alter system archive log start;
    System altered.
    SQL> show parameter log_archive_start;
    NAME                                 TYPE        VALUE
    log_archive_start                    boolean     FALSE
    SQL>I've gone thru that process twice but, I don't seem to be able to get the ARCH process to start. (newbie mistake I'm sure...)
    Thank you for your help (again!),
    John.

    The parameter log_archive_start is no more needed John (as suggested already) and the best way to check the archive options is through the archive log list command.
    [oracle@edhdr2p0-orcl oui]$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.1.0 Production on Tue Aug 10 10:51:57 2010
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, Automatic Storage Management, OLAP, Data Mining
    and Real Application Testing options
    SQL> archive log list
    Database log mode              No Archive Mode
    Automatic archival             Disabled
    Archive destination            USE_DB_RECOVERY_FILE_DEST
    Oldest online log sequence     11
    Current log sequence           13
    SQL> shut immediate
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> startup mount
    ORACLE instance started.
    Total System Global Area  418484224 bytes
    Fixed Size                  1336932 bytes
    Variable Size             318769564 bytes
    Database Buffers           92274688 bytes
    Redo Buffers                6103040 bytes
    Database mounted.
    SQL> select process,status from V$archive_processes;
       PROCESS STATUS
             0 STOPPED
             1 STOPPED
             2 STOPPED
             3 STOPPED
             4 STOPPED
             5 STOPPED
             6 STOPPED
             7 STOPPED
             8 STOPPED
             9 STOPPED
            10 STOPPED
       PROCESS STATUS
            11 STOPPED
            12 STOPPED
            13 STOPPED
            14 STOPPED
            15 STOPPED
            16 STOPPED
            17 STOPPED
            18 STOPPED
            19 STOPPED
            20 STOPPED
            21 STOPPED
       PROCESS STATUS
            22 STOPPED
            23 STOPPED
            24 STOPPED
            25 STOPPED
            26 STOPPED
            27 STOPPED
            28 STOPPED
            29 STOPPED
    30 rows selected.
    SQL>
    SQL> alter database archivelog;
    Database altered.
    SQL> alter database open;
    archive log list;
    Database altered.
    SQL> Database log mode         Archive Mode
    Automatic archival             Enabled
    Archive destination            USE_DB_RECOVERY_FILE_DEST
    Oldest online log sequence     11
    Next log sequence to archive   13
    Current log sequence           13
    SQL>
    SQL> select * from V$archive_processes where status <> 'STOPPED';
       PROCESS STATUS     LOG_SEQUENCE STAT
             0 ACTIVE                0 IDLE
             1 ACTIVE                0 IDLE
             2 ACTIVE                0 IDLE
             3 ACTIVE                0 IDLEHTH
    Aman....

  • Questions about using Bitlocker without TPM

    We currently use Bitlocker to encrypt our Windows 7 computers with TPM. Now we are looking at encrypting some Windows 7 computers without a TPM. I see how to change the group policy setting to allow Bitlocker without a TPM. I have looked at a lot of other
    threads and I have a few questions about how the Bitlocker without TPM works.
    1) I see a USB drive containing a key is required for Bitlocker configurations without a TPM, say the end user loses this USB drive, what are the recovery options for their computer? 
    This article seems to indicate that without the USB drive connected, you are unable to even access recovery options http://blogs.technet.com/b/hugofe/archive/2010/10/29/bitlocker-without-tpm.aspx
    We have recovery backed up to AD when Bitlocker is enabled, but how could we do this recovery on a computer on computer where it's USB is lost? Would we have to remove the HD itself and attach it to another computer to access?
    2) After enabling Bitlocker on a computer without a TPM and using the USB Drive for the key, is there a way to also add a PIN or password protection at bootup?

    Hi,
    Sorry for my dilatory reply, 
    Configuring a startup key is another method to enable a higher level of security with the TPM. The startup key is a key stored on a USB flash drive, and the USB flash drive must be inserted every time the computer starts. The startup key is used to provide
    another factor of authentication in conjunction with TPM authentication. To use a USB flash drive as a startup key, the USB flash drive must be formatted by using the NTFS, FAT, or FAT32 file system.
    You must have a startup key to use BitLocker on a non-TPM computer.
    From: http://technet.microsoft.com/de-de/library/ee449438(v=ws.10).aspx#BKMK_Key
    For more Q&A about BitLocker, you can refer to the link above.
    hope this is helpful.
    Roger Lu
    TechNet Community Support

  • Recovery Catalog showing wrong Archive Destination

    Hi All,
    I am using Oracle 10g on Solaris 10.
    The problem I am facing is this, while taking backup using recover catalog..............
    Starting backup at 27-JUL-11
    current log archived
    released channel: ch00
    released channel: ch01
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of backup command at 07/27/2011 11:16:47
    RMAN-06059: expected archived log not found, lost of archived log compromises recoverability
    ORA-19625: error identifying file /u01/oracle/product/10.2.0/db_1/dbs/arch1_56_757089402.dbf
    ORA-27037: unable to obtain file status
    SVR4 Error: 2: No such file or directory
    Additional information: 3
    This is because of the following reason
    RMAN> list archivelog all;
    List of Archived Log Copies
    Key Thrd Seq S Low Time Name
    39401 1 1567 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1567_756523261.dbf
    39402 1 1568 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1568_756523261.dbf
    39403 1 1569 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1569_756523261.dbf
    39404 1 1570 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1570_756523261.dbf
    39405 1 1571 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1571_756523261.dbf
    39406 1 1572 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1572_756523261.dbf
    39407 1 1573 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1573_756523261.dbf
    39408 1 1574 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1574_756523261.dbf
    39409 1 1575 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1575_756523261.dbf
    39410 1 1576 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1576_756523261.dbf
    39411 1 1577 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1577_756523261.dbf
    39412 1 1578 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1578_756523261.dbf
    39413 1 1579 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1579_756523261.dbf
    39414 1 1580 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1580_756523261.dbf
    39415 1 1581 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1581_756523261.dbf
    39416 1 1582 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1582_756523261.dbf
    39417 1 1583 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1583_756523261.dbf
    39418 1 1584 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1584_756523261.dbf
    39419 1 1585 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1585_756523261.dbf
    39420 1 1586 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1586_756523261.dbf
    39421 1 1587 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1587_756523261.dbf
    39422 1 1588 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1588_756523261.dbf
    39423 1 1589 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1589_756523261.dbf
    39424 1 1590 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1590_756523261.dbf
    39425 1 1591 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1591_756523261.dbf
    39426 1 1592 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1592_756523261.dbf
    39427 1 1593 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1593_756523261.dbf
    39428 1 1594 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1594_756523261.dbf
    39429 1 1595 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1595_756523261.dbf
    39430 1 1596 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1596_756523261.dbf
    39431 1 1597 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1597_756523261.dbf
    39432 1 1598 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1598_756523261.dbf
    39433 1 1599 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1599_756523261.dbf
    39434 1 1600 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1600_756523261.dbf
    39435 1 1601 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1601_756523261.dbf
    39436 1 1602 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1602_756523261.dbf
    39437 1 1603 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1603_756523261.dbf
    39438 1 1604 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1604_756523261.dbf
    39439 1 1605 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1605_756523261.dbf
    39440 1 1606 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1606_756523261.dbf
    39441 1 1607 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1607_756523261.dbf
    39442 1 1608 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1608_756523261.dbf
    39443 1 1609 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1609_756523261.dbf
    39444 1 1610 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1610_756523261.dbf
    39445 1 1611 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1611_756523261.dbf
    39446 1 1612 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1612_756523261.dbf
    39447 1 1613 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1613_756523261.dbf
    39448 1 1614 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1614_756523261.dbf
    39449 1 1615 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1615_756523261.dbf
    39450 1 1616 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1616_756523261.dbf
    39451 1 1617 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1617_756523261.dbf
    39452 1 1618 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1618_756523261.dbf
    39453 1 1619 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1619_756523261.dbf
    39454 1 1620 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1620_756523261.dbf
    39455 1 1621 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1621_756523261.dbf
    39456 1 1622 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1622_756523261.dbf
    39457 1 1623 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1623_756523261.dbf
    39458 1 1624 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1624_756523261.dbf
    39459 1 1625 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1625_756523261.dbf
    39460 1 1626 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1626_756523261.dbf
    39461 1 1627 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1627_756523261.dbf
    39462 1 1628 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1628_756523261.dbf
    39463 1 1629 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1629_756523261.dbf
    39464 1 1630 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1630_756523261.dbf
    39465 1 1631 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1631_756523261.dbf
    39466 1 1632 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1632_756523261.dbf
    39467 1 1633 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1633_756523261.dbf
    39468 1 1634 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1634_756523261.dbf
    39469 1 1635 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1635_756523261.dbf
    39470 1 1636 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1636_756523261.dbf
    39471 1 1637 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1637_756523261.dbf
    39682 1 1638 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1638_756523261.dbf
    39689 1 1639 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1639_756523261.dbf
    40642 1 1640 A 26-JUL-11 /oradb2/Archive/HCSPRD91/1_1640_756523261.dbf
    40643 1 1641 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1641_756523261.dbf
    40644 1 1642 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1642_756523261.dbf
    40645 1 1643 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1643_756523261.dbf
    40646 1 1644 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1644_756523261.dbf
    40647 1 1645 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1645_756523261.dbf
    40648 1 1646 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1646_756523261.dbf
    41310 1 1647 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1647_756523261.dbf
    41317 1 1648 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1648_756523261.dbf
    41373 1 1649 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1649_756523261.dbf
    41374 1 1650 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1650_756523261.dbf
    41375 1 1651 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1651_756523261.dbf
    41383 1 1652 A 27-JUL-11 /oradb2/Archive/HCSPRD91/1_1652_756523261.dbf
    *38725 1 56 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_56_757089402.dbf*
    *38946 1 57 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_57_757089402.dbf*
    *39000 1 58 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_58_757089402.dbf*
    *39022 1 59 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_59_757089402.dbf*
    *39243 1 60 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_60_757089402.dbf*
    *39297 1 61 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_61_757089402.dbf*
    *39703 1 62 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_62_757089402.dbf*
    *39704 1 63 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_63_757089402.dbf*
    *39705 1 64 A 26-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_64_757089402.dbf*
    *39933 1 65 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_65_757089402.dbf*
    *39991 1 66 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_66_757089402.dbf*
    *40023 1 67 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_67_757089402.dbf*
    *40024 1 68 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_68_757089402.dbf*
    *40025 1 69 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_69_757089402.dbf*
    *40254 1 70 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_70_757089402.dbf*
    *40308 1 71 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_71_757089402.dbf*
    *40345 1 72 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_72_757089402.dbf*
    *40570 1 73 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_73_757089402.dbf*
    *41329 1 74 A 27-JUL-11 /u01/oracle/product/10.2.0/db_1/dbs/arch1_74_757089402.dbf*
    I dont know from where recovery catalog is getting the information in bold.
    I have crosschecked and deleted archives and resynced the catalog several times. Immediatly after doing that the backup works fine, but the other day its the same thing all over again.
    Can some one tell me from where the catalog is getting this wrong infomation about archivelogs???
    By the way here is some more info about my target Database:
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    SQL> archive log list
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination /oradb2/Archive/HCSPRD91
    Oldest online log sequence 1651
    Next log sequence to archive 1653
    Current log sequence 1653
    NAME TYPE VALUE
    archive_lag_target integer 0
    log_archive_config string
    log_archive_dest string
    log_archive_dest_1 string LOCATION=/oradb2/Archive/HCSPR
    D91
    log_archive_dest_10 string
    log_archive_dest_2 string
    log_archive_dest_3 string
    log_archive_dest_4 string
    log_archive_dest_5 string
    log_archive_dest_6 string
    NAME TYPE VALUE
    log_archive_dest_7 string
    log_archive_dest_8 string
    log_archive_dest_9 string
    log_archive_dest_state_1 string enable
    log_archive_dest_state_10 string enable
    log_archive_dest_state_2 string enable
    log_archive_dest_state_3 string enable
    log_archive_dest_state_4 string enable
    log_archive_dest_state_5 string enable
    log_archive_dest_state_6 string enable
    log_archive_dest_state_7 string enable
    NAME TYPE VALUE
    log_archive_dest_state_8 string enable
    log_archive_dest_state_9 string enable
    log_archive_duplex_dest string
    log_archive_format string %t_%s_%r.dbf
    log_archive_local_first boolean TRUE
    log_archive_max_processes integer 2
    log_archive_min_succeed_dest integer 1
    log_archive_start boolean FALSE
    log_archive_trace integer 0
    remote_archive_enable string true
    standby_archive_dest string ?/dbs/arch
    Regards .....

    Do you have any other DBs with the same DBID as the one you are backing up? Are they connect to the same RMAN catalog?
    Iordan Iotzov
    http://iiotzov.wordpress.com/

  • Follow-up on question about copying 10.4

    I only got one reply to my question about how to copy OS 10.4 from one internal drive to another and unfortunately, the response was cut short. Something about if I had another computer with Firewire, using it and that's as far as the response went. Yes, I have a  laptop and have Firewire, but still don't know how to use it to help.

    If you just wish to make a duplicate of one hard drive to another hard drive in the same G5, then do this:
    Clone using Restore Option of Disk Utility
    Open Disk Utility from the Utilities folder.
    Select the destination volume from the left side list.
    Click on the Restore tab in the DU main window.
    Check the box labeled Erase destination.
    Select the destination volume from the left side list and drag it to the Destination entry field.
    Select the source volume from the left side list and drag it to the Source entry field.
    Double-check you got it right, then click on the Restore button.
    Destination means the second internal drive. Source means the internal startup drive.

  • Questions about Indexing and Using an Indexing POA

    Although I have only about 50 users, at least 15 of them have in excess of 100,000 messages in their accounts and the POA (version 7.0.2) is regularly slowing to a crawl. (I just know that plans for revolution are fomenting!) I have embarked on a campaign to reduce these accounts by archiving everything off to get mail accounts down to 3000 or fewer pieces. I have achieved user buy-in, but have worked on only a few users so far.
    In another closely related thread, it was suggested to me that the PO speed issues relate to broken indexes. And I suspect that given so many messages, the indexes were never getting fully rebuilt with the default QF POA settings. I am trying to fix that situation in addition to reducing mail account sizes. So, I have set up a second POA on another server and dedicated it to the indexing task. The /qfinterval is set for 1 hour, other /qf switches at default. The POA-QF does no mail delivery, but it does do nightly user upkeep.
    The POA-QF seems to be steadily working away and making progress at reducing the number of unindexed messages. However, I have questions about what I am seeing and what more I can do:
    1. Is the progress I am seeing real progress? For example I have a user with over 100,000 messages to be indexed and every time I check the logs, the count drops by about 500 messages per hourly QF cycle. I assume that if I just let it keep running, it will eventually get caught up and fixed. Not only with this user, but with all the others as well. Will my patience (and theirs) be rewarded? Are there any gotchas I need to prepare for?
    2. One user has recently had virtually all of her messages successfully moved to archive. I can see them in the Archive, and do not see them in the online account. However, now over a week later, QF still shows >130,000 items still left to index for that user. The POA-QF is making slow, steady progress reducing that number, but why is this user's QF count still so high? Does it just need more time, or is there something amiss for this user?
    3. I may want to rebuild indexes for single users from scratch. I have seen the TID 3105742 which tells how to do this: Essentially you turn off mail delivery functions, and make some other switch changes to dedicate the POA to indexing for just a single user, and then you let the POA rebuild the indexes. The implication of that scenario is that the POA is now enjoying exclusive access to the user's databases.
    If I want to use my secondary POA-QF to rebuild a user's index from scratch, does the main POA have to be offline and the user out of GWise? That is, Does the QF process require exclusive access in order to rebuild indexes from scratch?
    Thanks for any thoughts or suggestions.
    Peter Smick

    pgsmick wrote:
    > 1. Is the progress I am seeing real progress? For example I have a user with
    > over 100,000 messages to be indexed and every time I check the logs, the count
    > drops by about 500 messages per hourly QF cycle. I assume that if I just let
    > it keep running, it will eventually get caught up and fixed. Not only with
    > this user, but with all the others as well. Will my patience (and theirs) be
    > rewarded? Are there any gotchas I need to prepare for?
    Set this switch for this indexing POA - /qflevel=999 - this will index
    everything in one run. It will take a long time, but with no qflevel switch you
    are indeed only indexing 500 messages at a time, and if the user has that much
    mail, it might never really catch up.
    >
    > 2. One user has recently had virtually all of her messages successfully moved
    > to archive. I can see them in the Archive, and do not see them in the online
    > account. However, now over a week later, QF still shows >130,000 items still
    > left to index for that user. The POA-QF is making slow, steady progress
    > reducing that number, but why is this user's QF count still so high? Does it
    > just need more time, or is there something amiss for this user?
    >
    This is odd, because really the index count should drop to nothing, but with the
    above switch this might get resolved as well.
    > 3. I may want to rebuild indexes for single users from scratch. I have seen
    > the TID 3105742 which tells how to do this: Essentially you turn off mail
    > delivery functions, and make some other switch changes to dedicate the POA to
    > indexing for just a single user, and then you let the POA rebuild the indexes.
    > The implication of that scenario is that the POA is now enjoying exclusive
    > access to the user's databases.
    Not really - the POA is not enjoying exclusive access to the user's database,
    the indexer is just avoiding an attempt to index anything else.
    > If I want to use my secondary POA-QF to rebuild a user's index from scratch,
    > does the main POA have to be offline and the user out of GWise? That is, Does
    > the QF process require exclusive access in order to rebuild indexes from
    > scratch?
    No - QF never requires exclusive access. That said, you may find that an
    extremely vigorous QF can cause slowdowns for the user.
    Danita
    Novell Knowledge Partner
    Moving GroupWise to Linux?
    http://www.caledonia.net/gwmove.html

  • Re: Questions about Plan structure

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

  • Question about Kurts comments discussing the seperation of AIA & CDP - Test Lab Guide: Deploying an AD CS Two-Tier PKI Hierarchy - Kurt L Hudson MSFT

    Question about the sentence in bold. What is the meaning behind this comment?
    How would you separate the role of the AIA and CDP from a CA subordinate server? I can see where I add a CES and CEP server which has those as well, but I don't completely understand his comment. Because in this second step, (http://technet.microsoft.com/en-us/library/tlg-key-based-renewal.aspx)
    he shows how to implement CES and CEP.
    This is from the guide located at: http://technet.microsoft.com/library/hh831348.aspx
    Step 3: Configure APP1 to distribute certificates and CRLs
    In the extensions of the root CA, it was stated that the CRL from the root CA would be available via http://www.contoso.com/pki. Currently, there is not a PKI virtual directory on APP1, so one must be created.
    In a production environment, you would typically separate the issuing CA role from the role of hosting the AIA and CDP.
    However, this lab combines both in order to reduce the number of resources needed to complete the lab.
    Thanks,
    James

    My concern is, they have a 2-3k base of xp systems, over this year they are migrating them to Windows 7. During this time they will also be upgrading hardware for the existing windows 7 machines. The turnover of certificates are going to be high, which
    from what I've read here, it worries me.
    http://blogs.technet.com/b/askds/archive/2009/06/24/implementing-an-ocsp-responder-part-i-introducing-ocsp.aspx
    The application then can go to those locations to download the CRL. There are, however, some potential issues with this scenario. CRLs over time can get rather large
    depending on the number of certificates issued and revoked. If CRLs grow to a large size, and many clients have to download CRLs, this can have a negative impact on network performance. More importantly, by
    default Windows clients will timeout after 15 seconds while trying to download a CRL. Additionally,
    CRLs have information about every currently valid certificate that has been revoked, which is an excessive amount of data given the fact that an application may only need the revocation status for a few certificates. So,
    aside from downloading the CRL, the application or the OS has to parse the CRL and find a match for the serial number of the certificate that has been revoked.
    With the above limitations, which mostly revolve around scalability, it is clear that there are some drawbacks to using CRLs. Hence, the introduction of Online Certificate
    Status Protocol (OCSP). OCSP reduces the overhead associated with CRLs. There are server/client components to OCSP: The OCSP responder, which is the server component, and the OCSP Client. The OCSP Responder accepts status
    requests from OCSP Clients. When the OCSP Responder receives the request from the client it then needs to determine the status of the certificate using the serial number presented by the client. First the OCSP Responder determines if it has any cached responses
    for the same request. If it does, it can then send that response to the client. If there is no cached response, the OCSP Responder then checks to see if it has the CRL issued by the CA cached locally on the OCSP. If it does, it can check the revocation status
    locally, and send a response to the client stating whether the certificate is valid or revoked. The response is signed by the OCSP Signing Certificate that is selected during installation. If the OCSP does not have the CRL cached locally, the OCSP Responder
    can retrieve the CRL from the CDP locations listed in the certificate. The OCSP Responder then can parse the CRL to determine the revocation status, and send the appropriate response to the client.

  • A quick question about WebDynpro SLD and R/3 with concurrent users

    Hello ,
    I have a very quick question about Webdynpros and SLD connecting to an R/3 system, when you configure a webdynpro to connect to an R/3 system using SLD, you configure a user name and password from the R/3  for the SLD to use. What I would like to know is when I have concurrent users of my webdynpro, how can I know what one user did in R/3 and what another user did? Is there a way for the users of the web dynpro to use their R/3 credentials so SLD can access the R/3? Like dynamically configuring the SLD for each user?
    - I would like to avoid leaving their their passwords open in the code ( configuring two variable to get the users username and password and use these variables as JCO username and password )
    Thanks Ubergeeks,
    Guy

    Hi Guy
    You will have to use Single Sign On to achieve this. In the destination you have defined to connect to R/3 , there is an option to 'useSSO' instead of userid and password. This will ensure that calls to R/3 will be with the userid that has logged into WAS. You wont need to pass any passwords because  a login ticket is generated from WAS and passed on to R/3. The userid is derived from this ticket.
    For this to happen you will have to maintain a trust relation ship between R/3 and your WAS ,there is detailed documentation of this in help files. Configuration is very straight forward and is easy to perform
    Regards
    Pran

Maybe you are looking for

  • Ipod nano recognized by My PC but not by I-tunes 7

    I had upgrade to I-Tunes 7 and My Ipod Nano is recognized by My Computer but isn´t recognized by I-Tunes, somebody help me please. I already reinstalled USB drivers nothing happened. I already selected the manual Terminal Services and nothing happene

  • Difference impact on Net Value

    Hi All, Similar questions have been answered but this has not solved my queries so I have to post. the user wants to post the difference in amount received and invoice amount in some loss account. this is happening but in doing this, I have to give t

  • Problems with embedded font with medium or heavy weight

    Hey guys, I'm switching a current Project from Flashbuilder Beta2 to the Flashbuilder final release. I have a big problem concerning embedded fonts. The fonts I have to use are Futura Com Medium & Futura Com Heavy. Here's some of my code: @font-face

  • File DSN

    Hi, We  connected Crystal to Progress DB  via FileDSN(ODBC RDO) . We got the below error. logon failed. Details :HY000:[DataDirect][ODBC Progress OpenEdge Wire Protocol driver] Error parsing connect string at offset 198   attributes specified twise u

  • Can it be fixed??

    i was recently listening to my ipod when its battery died. i let it sit a lil while and then put it on the charger. after it charged a little bit the scroll wheel wont respond and neither will the buttons. What can i do