Question about Archiving Mailboxes

I keep almost all of my mail On My Mac > Personal Mail, then divided into subfolders beneath that.  I back up to Time Machine.
This morning I noticed one of those sub folders was missing, and had to go back more than 24 hours to recover it.  Not sure how it went missing, but it did.
I am wondering what is the diffference between using Archive Mailbox under the Mailbox menu and keeping the messages in folders On My Mac?
I think my mailboxes are just getting too darn big, try as I might to contain them.  I'm a mail hoarder.
Thanks for any insight

How you backup and what you backup to is all about taking a calculated risk. In other words, what's it worth to you?
Ultimately, whatever method you choose, I think redundancy is the most important factor. It doesn't matter if you have half-a-dozen backups, if they're all in the same place and the building burns down, they're all gone.
Yes, hard drives are flimsy and fail, but you can easily duplicate a backup and store them in separate locations; then, make regular maintenance checks to see if they're viable. The probability of two backups failing simultaneously is astronomically low; the probability of one backup failing is astronomically high. For the average person, most dedicated backup systems are too far out of reach, pricewise and from an infrastructure standpoint. Duplicating hard drives is pretty crude, but it's cheap so long as you exercise due diligence.

Similar Messages

  • Subtitling (question about archived material)

    OK, I've looked through the archived questions about how to subtitle...and I just want to make sure I do this right. If I do subtitles in DVDSP, do I write them up in a text file first and then import them, or just type them in manually?
    If I do subtitles in FCP I just use something like Title3D and type them in manually...creating a whole buncha overlay (title) clips, right?
    Forgive the redundency, I know how subtitles are supposed to look, just trying to find the best (and least time consuming) way to produce them.
    FYI...the subtitles will be of a person from Spain speaking English (it's kinda hard to understand him since he talks fast and has a heavy Spanish accent).
    Thanks, all, in advance.
    Jonathan

    subtitling (question about archived material)< </div>
    So you are trying put subtitles on your archived material?
    If you use the subtitle track in DVDSP, the text will be switchable by the user. (This is made more complicated by several factors like your desire to make the subtitles permanent or uncontrollable by the user. These are options you can control by programming the DVDSP project. Not difficult but very confusing.)
    If you use a text tool in FCP, the titles will be permanently burned into the video image.
    If you use some fo the more sophisticated (read: Expensive) subtitling tools for FCP, you can export a caption-compliant file that can be printed to videotape or mastered to DVD. If you own one of these products, you are at the complete mercy of their manual's writer.
    bogiesan

  • Questions about archive

    Hope anyone can help me understand this…
    I found there are various kinds of archive feature in Exchange or Outlook. And I need help to choose the appropriate method to do archive in my Exchange.
    Thanks

    Hi David,
    Exchange Server 2010 has a new built-in archiving solution, which lets you create a personal archive mailbox within the Exchange organization.
    You can use the Enable-Mailbox "xxx" -Archive cmdlet to enable the personal archive for a user's mailbox.
    What's more, here is an article for your reference.
    Understanding Personal Archives
    http://technet.microsoft.com/en-us/library/dd979795(v=exchg.141).aspx
    Hope it helps.
    If you need further assistance, please feel free to let me know.
    Best regards,
    Amy
    Amy Wang
    TechNet Community Support

  • A question about archive and install

    Recently my system crashed due to a faulty OS instal.
    So, I hauled my G4 to the mac store and they resoved the problem for me and told me to do a archive and instal, and even gave me instructions.
    Everything was going fine untill it came the time to choose which directory to install to... I clicked on my hard drive and a dialouge box popped up saying that it couldn't install older versions of the files over newer versions....
    So now I've hit a dead end.
    How can I continue the install?

    WanderingAimlessly:
    Do you currently have a functioniing OS on your HDD?
    Do you have your computer backed up?
    If your answers are "no" to the first question and "yes" to the second, I suggest you go a completer reformat, zero all data and install.
    Post back with your responses, and we will decide where to go from there.
    Good luck.
    cornelius

  • Question about archived logs

    Hi!
    I wonder if there is possibility to configure archiving policy in a way:
    set one destination (/arch1/oracle) as primary, and another (/arch2/oracle) as optionally.
    redo Logs will be copied to /arch2/oracle only if /arch1/oracle fails, but after it will be available again, files will be moved (or copied) from /arch2/oracle to /arch1/oracle. Anybody tried to store backup files in that way? is it possible? I tried various combinations of log_archive_dest_n and log_archive_dest_state_n parameters, but it won't work in a way I expected.
    Message was edited by:
    wojaczek

    it's about limitation of disk place on server where DB operates. Archived logs are stored on NFS filesystem, but time to time there are some problems with remote filesystem. In such cases DB will hang and manual change of arch log destination to local FS is needed. But this destination could not be set permanently because of disk space limitations.
    When i set destination 1 as mandatory and 2 as optionally db hangs when arc can't use mandatory destination.
    Setting both destinations as optionally gives finally inconsistency and storing files on local fs, what is not desirable

  • Question about archive destination.

    Hi All,
    When we set the archived destination to this parameter log_archive_dest like:
    SQL>alter system set LOG_ARCHIVE_DEST_.. = 'location=/archivelogpath reopen';
    Could you help me to know about 'reopen' keyword here. Please help me know, what does it mean and whats the use of it. I have failed to get this in oracle documentation.
    Best Regards,
    Michel.

    From the documentation:
    Rearchiving to a Failed Destination
    Use the REOPEN attribute of the LOG_ARCHIVE_DEST_n parameter to specify whether and when ARCn should attempt to rearchive to a failed destination following an error. REOPEN applies to all errors, not just OPEN errors.
    REOPEN=n sets the minimum number of seconds before ARCn should try to reopen a failed destination. The default value for n is 300 seconds. A value of 0 is the same as turning off the REOPEN attribute; ARCn will not attempt to archive after a failure. If you do not specify the REOPEN keyword, ARCn will never reopen a destination following an error.
    You cannot use REOPEN to specify the number of attempts ARCn should make to reconnect and transfer archived logs. The REOPEN attempt either succeeds or fails.
    When you specify REOPEN for an OPTIONAL destination, the database can overwrite online logs if there is an error. If you specify REOPEN for a MANDATORY destination, the database stalls the production database when it cannot successfully archive. In this situation, consider the following options:
    Archive manually to the failed destination.
    Change the destination by deferring the destination, specifying the destination as optional, or changing the service.
    Drop the destination.
    When using the REOPEN keyword, note the following:
    ARCn reopens a destination only when starting an archive operation from the beginning of the log file, never during a current operation. ARCn always retries the log copy from the beginning.
    If you specified REOPEN, either with a specified time the default, ARCn checks to see whether the time of the recorded error plus the REOPEN interval is less than the current time. If it is, ARCn retries the log copy.
    The REOPEN clause successfully affects the ACTIVE=TRUE destination state. The VALID and ENABLED states are not changed.

  • Question about Archive Log Deletion policy

    I've a problem to understand the Archive Log Deletion policy, and I I'd like to this problem explain with the following example.
    Messages of the database are in German, but I guess you'll understand them.
    SQL> startup
    ORACLE-Instance hochgefahren.
    Total System Global Area 5344731136 bytes
    Fixed Size                  2129240 bytes
    Variable Size            2684355240 bytes
    Database Buffers         2617245696 bytes
    Redo Buffers               41000960 bytes
    Datenbank mounted.
    Datenbank geöffnet.
    SQL> archive log list
    Datenbank-Log-Modus              Archive-Modus
    Automatische Archivierung             Aktiviert
    Archivierungsziel            E:\oracle\thetis_iv\arch
    Älteste Online-Log-Sequenz     17917
    Nächste zu archivierende Log-Sequenz   17919
    Aktuelle Log-Sequenz           17919
    SQL> alter system switch logfile;
    System wurde geändert.I created a brand new archive log.
    SQL> exit
    Verbindung zu Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options beendet
    D:\OracleDB\product\11.1.0\db_1\BIN>dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:04    <DIR>          .
    06.04.2011  15:04    <DIR>          ..
    06.04.2011  15:04        17.137.152 ARC17919_0721667907.001
                   1 Datei(en),     17.137.152 Bytes
                   2 Verzeichnis(se), 41.073.258.496 Bytes freiand this is the only archive log in the directory. Now I start rman:
    D:\OracleDB\product\11.1.0\db_1\BIN>rman target / catalog rmanrepo@rmanrepo
    Recovery Manager: Release 11.1.0.7.0 - Production on Mi Apr 6 15:05:35 2011
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Mit Ziel-Datenbank verbunden: ENTWIV (DBID=21045568)
    Kennwort für Recovery-Katalog-Datenbank:
    Verbindung mit Datenbank des Recovery-Katalogs
    RMAN> show all;
    RMAN-Konfigurationsparameter für Datenbank mit db_unique_name ENTWIV sind:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO 'SBT_TAPE';
    CONFIGURE CONTROLFILE AUTOBACKUP OFF;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO 'E:\oracle\thetis_iv\backup\CF_%F_ENTWIV.ORA';
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE SBT_TAPE TO '%F'; # default
    CONFIGURE DEVICE TYPE SBT_TAPE PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE 'SBT_TAPE' PARMS  'ENV=(TPDO_OPTFILE=D:\Services\Tivoli\TSM\AgentOBA64\tpdo.opt)';
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BZIP2'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO 'SBT_TAPE';
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO 'D:\ORACLEDB\PRODUCT\11.1.0\DB_1\DATABASE\SNCFENTWIV.ORA'; # defaultThe archive log deletion policy says the the logfiles have to be backed up for two times before they get deleted.
    Now I backup all archive logs, that havn't been backed up for at least two times.
    RMAN> run { backup archivelog all not backed up 2 times
    2>       format '%d_AR_%Y%M%D_%s_%t'
    3>       tag 'ARCHIVE LOGS'
    4>       DELETE ALL INPUT;
    5>     }
    Starten backup um 06.04.2011 15:08:01
    Aktuelles Log archiviert
    Zugewiesener Kanal: ORA_SBT_TAPE_1
    Kanal ORA_SBT_TAPE_1: SID=253 Device-Typ=SBT_TAPE
    Kanal ORA_SBT_TAPE_1: Data Protection for Oracle: version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set für Archive Log wird begonnen
    Kanal ORA_SBT_TAPE_1: Archive Logs in Backup Set werden angegeben
    Eingabe-Archive-Log-Thread=1 Sequence=17919 RECID=147 STAMP=747759899
    Eingabe-Archive-Log-Thread=1 Sequence=17920 RECID=148 STAMP=747760081
    Kanal ORA_SBT_TAPE_1: Piece 1 wird auf 06.04.2011 15:08:02 begonnen
    Kanal ORA_SBT_TAPE_1: Piece 1 auf 06.04.2011 15:08:09 beendet
    Piece Handle=ENTWIV_AR_20110406_23_747760082 Tag=ARCHIVE LOGS Kommentar=API Version 2.0,MMS Version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set vollstõndig, abgelaufene Zeit: 00:00:08
    Kanal ORA_SBT_TAPE_1: Archive Logs werden gel÷scht
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17919_0721667907.001 RECID=147 STAMP=747759899
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17920_0721667907.001 RECID=148 STAMP=747760081
    Beendet backup um 06.04.2011 15:08:10
    RMAN> exit
    Recovery Manager abgeschlossen.
    D:\OracleDB\product\11.1.0\db_1\BIN> dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:08    <DIR>          .
    06.04.2011  15:08    <DIR>          ..
                   0 Datei(en),              0 Bytes
                   2 Verzeichnis(se), 41.090.396.160 Bytes freirman deleted all archive logs, even I they are on tape only once by now.
    Thats not what I expected. Where is my mistake?

    Hi,
    I do new tests it's very strange.
    BACKUP ARCHIVELOG command is not obeying the policy of archivelog.
    You can open a SR on MOS. (to check bugs)
    I reproduce the same test and the result was the same, it seems that this is a bug.
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO DISK;
    RMAN> backup archivelog all not backed up 2 times delete all input;
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=1 sequence=15 RECID=16 STAMP=747753711
    input archived log thread=2 sequence=20 RECID=17 STAMP=747753714
    input archived log thread=1 sequence=16 RECID=19 STAMP=747753729
    input archived log thread=2 sequence=21 RECID=18 STAMP=747753729
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t132210_0.304.747753731 tag=TAG20110406T132210 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    channel ORA_DISK_1: deleting archived log(s)
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_15.293.747753711 RECID=16 STAMP=747753711
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_20.295.747753715 RECID=17 STAMP=747753714
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_16.294.747753729 RECID=19 STAMP=747753729
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_21.298.747753729 RECID=18 STAMP=747753729
    Finished backup at 06-APR-11
    RMAN> list archivelog all;
    specification does not match any archived log in the repositoryOracle Docs Says:
    The BACKUP ARCHIVELOG ... DELETE INPUT command deletes archived log files after they are backed up.
    This command eliminates the separate step of manually deleting archived redo logs.
    With DELETE INPUT, RMAN deletes only the specific copy of the archived log chosen for the backup set.
    With DELETE ALL INPUT, RMAN deletes each backed-up archived redo log file from all log archiving destinations.
    As explained in "Configuring an Archived Redo Log Deletion Policy",
    the BACKUP ... DELETE INPUT and DELETE ARCHIVELOG commands obey the archived redo log deletion policy
    for logs in all archiving locations. For example, if you specify that logs should only be deleted when backed
    up at least twice to tape, then BACKUP ... DELETE honors this policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10642/rcmbckba.htm#BRADV89524
    But in ours case it's not honors this policy.
    Only with the FORCE command should this happen. But it is not our case.
    Oracle Docs:
    If FORCE is not specified on the deletion commands,
    then these deletion commands obey the archived log deletion policy.
    If FORCE is specified, then the deletion commands ignore the archived log deletion policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10643/rcmsynta010.htm#RCMRF113
    Alternatively you can do the following:
    Set the commands separately.
    Check this:
    RMAN>  run {
    2> backup archivelog all not backed up 2 times ;
    3> delete archivelog all backed up 2 times to disk;
    4> }
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=2 sequence=22 RECID=21 STAMP=747755128
    input archived log thread=1 sequence=17 RECID=20 STAMP=747755127
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t134528_0.295.747755129 tag=TAG20110406T134528 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    Finished backup at 06-APR-11
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=78 instance=orcl1 device type=DISK
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_17.298.747755127 thread=1 sequence=17
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_22.294.747755129 thread=2 sequence=22
    RMAN>Edited by: Levi Pereira on Apr 6, 2011 1:35 PM

  • Outlook 2013 cached mode "mail to keep offline" - does this setting affect archive mailbox or Office 365 hosted archive mailbox?

    I have an Exchange 2013 Hybrid environment with Exchange 2007 in coexistance mode as well as Office 365. We are hosting the majority of our Exchange accounts on-premise (migrating them from 2007), and most (soon to be ALL) Archive mailboxes in Office 365.
    I have through Group policy set the policy "Select Cached Exchange Mode sync settings for new profiles" to 6 months. This works fine for the contents of the regular mailbox, but the setting is NOT being set at all for the Online Archive mailbox,
    whether or not it is hosted on-premise or in Office 365. Is this the expected behavior?
    Another question: Is it possible to have the regular mailbox in cached mode and the archive in Online mode or vice versa?

    Hi,
    The key to your questions is: Archive mailboxes are online only, they cannot be cached for offline viewing, so you can keep your regular mailboxes offline(In Cached Exchange Mode) or Online, but cannot switch Archive mailboxes to Offline.
    The whole idea of the archive mailbox is to keep your "online" mailbox with recent / important mails small and accessible on and offline. The archive mailbox is intended to store older / less important emails which are not neccessary to access immediately.
    One important reason of archive mailbox is to prevent the growth of offline files.
    Regards,
    Melon Chen
    TechNet Community Support

  • Migration from 2007 and Archive Mailbox PST import question

    Our enterprise will be migrating from Exchange 2007 SP3 to Exchange 2013 CU3 very shortly. Exchange 2013 is set up and I am in coexistance mode currently. Everything is working fine with the legacy redirection, I have tested some mailbox migrations and they
    work fine. We currently use Mimosa Nearpoint as a third party email archive product.
    As part of this migration from 2007 to 2013, we plan on moving away from Mimosa Nearpoint and utilizing the Exchange 2013 Archive Mailbox feature. We will give each user an Archive Mailbox. We have the ability to export the users existing archive out of
    NearPoint to .PST files and then migrating them directly into the users Archive Mailbox using the "New-MailboxImportRequest -IsArchive" powershell commandlet. This has been tested and works fine with one very important exception.
    When I export from Nearpoint, I have to export the entire contents of the archive, I cannot set date ranges. So when I do an export, it also contains the contents of the user's mailbox. I have managed folder policies in place in the Exchange 2007 environment
    which delete all contents of the mailbox older than 60 days, but when I import the .PST to the archive, it imports ALL items into the Archive Mailbox. So I have 60 days worth duplication of items.
    It is my understanding that if I imported the Archive .PST directly into the users production mailbox in the 2013 side, that it would not import duplicate items. I am trying to find out if there is any process which will do the same thing with me importing
    directly into the users Archive Mailbox.
    Thanks you,

    Hi,
    We used a third patry email archive product to export to .pst files, so it is too hard to say that whether it cause this issue.
    However, I found a method to delete the duplicate items from our mailbox via MFCMAPI Tool.
    Please make sure that the Exchange Server mailbox has been backed up first.
    Details in the following KB:
    How to remove duplicate folders in Outlook when connected to an Exchange Server mailbox
    http://support.microsoft.com/kb/2509983
    I also found a script to remove duplicate items.
    This script will scan each folder of a given mailbox and removes duplicate items from those folders. You can specify how the items should be deleted and what items to process, e.g. mail items or appointments. Sample usage is after a misbehaving synchronization
    tool creates duplicate items or (accidental) import of PST file with duplicate items.
    *****Please also make a full back up first*****
    Removing Duplicate Items from a Mailbox
    http://gallery.technet.microsoft.com/office/Removing-Duplicate-Items-f706e1cc#content
    Disclaimer       
    The sample scripts are not supported under any Microsoft standard support program or service.
    The sample scripts are provided AS IS without warranty of any kind. Microsoft further disclaims all implied warranties including, without limitation, any implied warranties of merchantability or of fitness for a particular purpose.
    The entire risk arising out of the use or performance of the sample scripts and documentation remains with you.
    In no event shall Microsoft, its authors, or anyone else involved in the creation, production, or delivery of the scripts be liable for any damages whatsoever (including, without limitation, damages for loss of business profits, business interruption, loss
    of business information, or other pecuniary loss) arising out of the use of or inability to use the sample scripts or documentation, even if Microsoft has been advised of the possibility of such damages.
    Hope it is helpful
    Thanks
    Mavis
    Mavis Huang
    TechNet Community Support

  • Retention Policy between Mailbox and Archive Mailbox level

    Dear All
    I have some questions about the Retention Policy between Mailbox and Archive Mailbox.
    I want to apply different retention tag on different Mailbox level. 
    In normal case , I would like apply the Retention Policy for Mailbox after 28 days moving to archive Mailbox and over 365 days to be deleted.
    And the Archive Mailbox should be liked deleted the mail over 90 or 180 days.
    But I found that the archive Mailbox is applied the same Retention Policy as the Mailbox one.
    Is it Normal case ? How can I set another retention Policy for My archive Mailbox ?
    Best Regards
    Vincent Chi

    You cannot have a different retention policy for primary and archive mailbox. The policy applied to the primary mailbox also applies
    to the archive mailbox.
    http://blogs.technet.com/b/vikass_blog/archive/2013/07/04/retention-policy-on-archive-mailbox.aspx
    Regards
    John

  • How do I create an Archive Mailbox for a pop3 account on an iPhone4?

    How do  create an Archive Mailbox for a pop3 email account on an iPhone4 running iOs6?

    You can use classifications but there is no auto feature to archive like that on web apps.
    In terms of the blog, Like I have said to everyone that has posted about blog preview images:
    http://www.prettypollution.com.au/business-catalyst-blog
    Just one example of an image at the start of the blog post rendering out, not hard at all.

  • Questions about using Bitlocker without TPM

    We currently use Bitlocker to encrypt our Windows 7 computers with TPM. Now we are looking at encrypting some Windows 7 computers without a TPM. I see how to change the group policy setting to allow Bitlocker without a TPM. I have looked at a lot of other
    threads and I have a few questions about how the Bitlocker without TPM works.
    1) I see a USB drive containing a key is required for Bitlocker configurations without a TPM, say the end user loses this USB drive, what are the recovery options for their computer? 
    This article seems to indicate that without the USB drive connected, you are unable to even access recovery options http://blogs.technet.com/b/hugofe/archive/2010/10/29/bitlocker-without-tpm.aspx
    We have recovery backed up to AD when Bitlocker is enabled, but how could we do this recovery on a computer on computer where it's USB is lost? Would we have to remove the HD itself and attach it to another computer to access?
    2) After enabling Bitlocker on a computer without a TPM and using the USB Drive for the key, is there a way to also add a PIN or password protection at bootup?

    Hi,
    Sorry for my dilatory reply, 
    Configuring a startup key is another method to enable a higher level of security with the TPM. The startup key is a key stored on a USB flash drive, and the USB flash drive must be inserted every time the computer starts. The startup key is used to provide
    another factor of authentication in conjunction with TPM authentication. To use a USB flash drive as a startup key, the USB flash drive must be formatted by using the NTFS, FAT, or FAT32 file system.
    You must have a startup key to use BitLocker on a non-TPM computer.
    From: http://technet.microsoft.com/de-de/library/ee449438(v=ws.10).aspx#BKMK_Key
    For more Q&A about BitLocker, you can refer to the link above.
    hope this is helpful.
    Roger Lu
    TechNet Community Support

  • Questions about Indexing and Using an Indexing POA

    Although I have only about 50 users, at least 15 of them have in excess of 100,000 messages in their accounts and the POA (version 7.0.2) is regularly slowing to a crawl. (I just know that plans for revolution are fomenting!) I have embarked on a campaign to reduce these accounts by archiving everything off to get mail accounts down to 3000 or fewer pieces. I have achieved user buy-in, but have worked on only a few users so far.
    In another closely related thread, it was suggested to me that the PO speed issues relate to broken indexes. And I suspect that given so many messages, the indexes were never getting fully rebuilt with the default QF POA settings. I am trying to fix that situation in addition to reducing mail account sizes. So, I have set up a second POA on another server and dedicated it to the indexing task. The /qfinterval is set for 1 hour, other /qf switches at default. The POA-QF does no mail delivery, but it does do nightly user upkeep.
    The POA-QF seems to be steadily working away and making progress at reducing the number of unindexed messages. However, I have questions about what I am seeing and what more I can do:
    1. Is the progress I am seeing real progress? For example I have a user with over 100,000 messages to be indexed and every time I check the logs, the count drops by about 500 messages per hourly QF cycle. I assume that if I just let it keep running, it will eventually get caught up and fixed. Not only with this user, but with all the others as well. Will my patience (and theirs) be rewarded? Are there any gotchas I need to prepare for?
    2. One user has recently had virtually all of her messages successfully moved to archive. I can see them in the Archive, and do not see them in the online account. However, now over a week later, QF still shows >130,000 items still left to index for that user. The POA-QF is making slow, steady progress reducing that number, but why is this user's QF count still so high? Does it just need more time, or is there something amiss for this user?
    3. I may want to rebuild indexes for single users from scratch. I have seen the TID 3105742 which tells how to do this: Essentially you turn off mail delivery functions, and make some other switch changes to dedicate the POA to indexing for just a single user, and then you let the POA rebuild the indexes. The implication of that scenario is that the POA is now enjoying exclusive access to the user's databases.
    If I want to use my secondary POA-QF to rebuild a user's index from scratch, does the main POA have to be offline and the user out of GWise? That is, Does the QF process require exclusive access in order to rebuild indexes from scratch?
    Thanks for any thoughts or suggestions.
    Peter Smick

    pgsmick wrote:
    > 1. Is the progress I am seeing real progress? For example I have a user with
    > over 100,000 messages to be indexed and every time I check the logs, the count
    > drops by about 500 messages per hourly QF cycle. I assume that if I just let
    > it keep running, it will eventually get caught up and fixed. Not only with
    > this user, but with all the others as well. Will my patience (and theirs) be
    > rewarded? Are there any gotchas I need to prepare for?
    Set this switch for this indexing POA - /qflevel=999 - this will index
    everything in one run. It will take a long time, but with no qflevel switch you
    are indeed only indexing 500 messages at a time, and if the user has that much
    mail, it might never really catch up.
    >
    > 2. One user has recently had virtually all of her messages successfully moved
    > to archive. I can see them in the Archive, and do not see them in the online
    > account. However, now over a week later, QF still shows >130,000 items still
    > left to index for that user. The POA-QF is making slow, steady progress
    > reducing that number, but why is this user's QF count still so high? Does it
    > just need more time, or is there something amiss for this user?
    >
    This is odd, because really the index count should drop to nothing, but with the
    above switch this might get resolved as well.
    > 3. I may want to rebuild indexes for single users from scratch. I have seen
    > the TID 3105742 which tells how to do this: Essentially you turn off mail
    > delivery functions, and make some other switch changes to dedicate the POA to
    > indexing for just a single user, and then you let the POA rebuild the indexes.
    > The implication of that scenario is that the POA is now enjoying exclusive
    > access to the user's databases.
    Not really - the POA is not enjoying exclusive access to the user's database,
    the indexer is just avoiding an attempt to index anything else.
    > If I want to use my secondary POA-QF to rebuild a user's index from scratch,
    > does the main POA have to be offline and the user out of GWise? That is, Does
    > the QF process require exclusive access in order to rebuild indexes from
    > scratch?
    No - QF never requires exclusive access. That said, you may find that an
    extremely vigorous QF can cause slowdowns for the user.
    Danita
    Novell Knowledge Partner
    Moving GroupWise to Linux?
    http://www.caledonia.net/gwmove.html

  • Re: Questions about Plan structure

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

  • Question about Kurts comments discussing the seperation of AIA & CDP - Test Lab Guide: Deploying an AD CS Two-Tier PKI Hierarchy - Kurt L Hudson MSFT

    Question about the sentence in bold. What is the meaning behind this comment?
    How would you separate the role of the AIA and CDP from a CA subordinate server? I can see where I add a CES and CEP server which has those as well, but I don't completely understand his comment. Because in this second step, (http://technet.microsoft.com/en-us/library/tlg-key-based-renewal.aspx)
    he shows how to implement CES and CEP.
    This is from the guide located at: http://technet.microsoft.com/library/hh831348.aspx
    Step 3: Configure APP1 to distribute certificates and CRLs
    In the extensions of the root CA, it was stated that the CRL from the root CA would be available via http://www.contoso.com/pki. Currently, there is not a PKI virtual directory on APP1, so one must be created.
    In a production environment, you would typically separate the issuing CA role from the role of hosting the AIA and CDP.
    However, this lab combines both in order to reduce the number of resources needed to complete the lab.
    Thanks,
    James

    My concern is, they have a 2-3k base of xp systems, over this year they are migrating them to Windows 7. During this time they will also be upgrading hardware for the existing windows 7 machines. The turnover of certificates are going to be high, which
    from what I've read here, it worries me.
    http://blogs.technet.com/b/askds/archive/2009/06/24/implementing-an-ocsp-responder-part-i-introducing-ocsp.aspx
    The application then can go to those locations to download the CRL. There are, however, some potential issues with this scenario. CRLs over time can get rather large
    depending on the number of certificates issued and revoked. If CRLs grow to a large size, and many clients have to download CRLs, this can have a negative impact on network performance. More importantly, by
    default Windows clients will timeout after 15 seconds while trying to download a CRL. Additionally,
    CRLs have information about every currently valid certificate that has been revoked, which is an excessive amount of data given the fact that an application may only need the revocation status for a few certificates. So,
    aside from downloading the CRL, the application or the OS has to parse the CRL and find a match for the serial number of the certificate that has been revoked.
    With the above limitations, which mostly revolve around scalability, it is clear that there are some drawbacks to using CRLs. Hence, the introduction of Online Certificate
    Status Protocol (OCSP). OCSP reduces the overhead associated with CRLs. There are server/client components to OCSP: The OCSP responder, which is the server component, and the OCSP Client. The OCSP Responder accepts status
    requests from OCSP Clients. When the OCSP Responder receives the request from the client it then needs to determine the status of the certificate using the serial number presented by the client. First the OCSP Responder determines if it has any cached responses
    for the same request. If it does, it can then send that response to the client. If there is no cached response, the OCSP Responder then checks to see if it has the CRL issued by the CA cached locally on the OCSP. If it does, it can check the revocation status
    locally, and send a response to the client stating whether the certificate is valid or revoked. The response is signed by the OCSP Signing Certificate that is selected during installation. If the OCSP does not have the CRL cached locally, the OCSP Responder
    can retrieve the CRL from the CDP locations listed in the certificate. The OCSP Responder then can parse the CRL to determine the revocation status, and send the appropriate response to the client.

Maybe you are looking for

  • Need advice on importing for quality  since Lame won't work on itunes 7

    I had been using Lame to import CDs. It would record at a variable rate based on the music and made really good sounding recordings to iTunes. Since I updated iTunes to v.7.3, Lame no longer works and is now available only as source code (let me admi

  • How can i get the code html and css of my form please help me

    how can i get the code html and css of my form

  • Upgradation of Lenovo S868t smartphone

    Hello, I'd like to upgrade my Lenovo S868t smartphone which is in Ice cream sandwitch (4.0.3). I'd like to upgrade into Jelly bean (4.1). Would you please tell what I should do? Thanks Solved! Go to Solution.

  • Mail will not take password

    I changed my gmail password on the website and now it is asking me for it on my MacBook Air Mail and my Iphone and when I enter the new password it is telling me it is incorrect. I works at gmail.com - I reset it and it still works at gmail.com and n

  • Getting a BT Yahoo email address into a sub-accoun...

    I've got BT Premium Email with sub-accounts - still have some spare. How do I 'transfer' an old, stand alone BT Yahoo email address into one of my sub accounts. Can I do this or do I have to contact BT - if so do you have a number ? Thanks Solved! Go