Subtitling (question about archived material)

OK, I've looked through the archived questions about how to subtitle...and I just want to make sure I do this right. If I do subtitles in DVDSP, do I write them up in a text file first and then import them, or just type them in manually?
If I do subtitles in FCP I just use something like Title3D and type them in manually...creating a whole buncha overlay (title) clips, right?
Forgive the redundency, I know how subtitles are supposed to look, just trying to find the best (and least time consuming) way to produce them.
FYI...the subtitles will be of a person from Spain speaking English (it's kinda hard to understand him since he talks fast and has a heavy Spanish accent).
Thanks, all, in advance.
Jonathan

subtitling (question about archived material)< </div>
So you are trying put subtitles on your archived material?
If you use the subtitle track in DVDSP, the text will be switchable by the user. (This is made more complicated by several factors like your desire to make the subtitles permanent or uncontrollable by the user. These are options you can control by programming the DVDSP project. Not difficult but very confusing.)
If you use a text tool in FCP, the titles will be permanently burned into the video image.
If you use some fo the more sophisticated (read: Expensive) subtitling tools for FCP, you can export a caption-compliant file that can be printed to videotape or mastered to DVD. If you own one of these products, you are at the complete mercy of their manual's writer.
bogiesan

Similar Messages

  • A question about archive and install

    Recently my system crashed due to a faulty OS instal.
    So, I hauled my G4 to the mac store and they resoved the problem for me and told me to do a archive and instal, and even gave me instructions.
    Everything was going fine untill it came the time to choose which directory to install to... I clicked on my hard drive and a dialouge box popped up saying that it couldn't install older versions of the files over newer versions....
    So now I've hit a dead end.
    How can I continue the install?

    WanderingAimlessly:
    Do you currently have a functioniing OS on your HDD?
    Do you have your computer backed up?
    If your answers are "no" to the first question and "yes" to the second, I suggest you go a completer reformat, zero all data and install.
    Post back with your responses, and we will decide where to go from there.
    Good luck.
    cornelius

  • Question about archived logs

    Hi!
    I wonder if there is possibility to configure archiving policy in a way:
    set one destination (/arch1/oracle) as primary, and another (/arch2/oracle) as optionally.
    redo Logs will be copied to /arch2/oracle only if /arch1/oracle fails, but after it will be available again, files will be moved (or copied) from /arch2/oracle to /arch1/oracle. Anybody tried to store backup files in that way? is it possible? I tried various combinations of log_archive_dest_n and log_archive_dest_state_n parameters, but it won't work in a way I expected.
    Message was edited by:
    wojaczek

    it's about limitation of disk place on server where DB operates. Archived logs are stored on NFS filesystem, but time to time there are some problems with remote filesystem. In such cases DB will hang and manual change of arch log destination to local FS is needed. But this destination could not be set permanently because of disk space limitations.
    When i set destination 1 as mandatory and 2 as optionally db hangs when arc can't use mandatory destination.
    Setting both destinations as optionally gives finally inconsistency and storing files on local fs, what is not desirable

  • Question about archive destination.

    Hi All,
    When we set the archived destination to this parameter log_archive_dest like:
    SQL>alter system set LOG_ARCHIVE_DEST_.. = 'location=/archivelogpath reopen';
    Could you help me to know about 'reopen' keyword here. Please help me know, what does it mean and whats the use of it. I have failed to get this in oracle documentation.
    Best Regards,
    Michel.

    From the documentation:
    Rearchiving to a Failed Destination
    Use the REOPEN attribute of the LOG_ARCHIVE_DEST_n parameter to specify whether and when ARCn should attempt to rearchive to a failed destination following an error. REOPEN applies to all errors, not just OPEN errors.
    REOPEN=n sets the minimum number of seconds before ARCn should try to reopen a failed destination. The default value for n is 300 seconds. A value of 0 is the same as turning off the REOPEN attribute; ARCn will not attempt to archive after a failure. If you do not specify the REOPEN keyword, ARCn will never reopen a destination following an error.
    You cannot use REOPEN to specify the number of attempts ARCn should make to reconnect and transfer archived logs. The REOPEN attempt either succeeds or fails.
    When you specify REOPEN for an OPTIONAL destination, the database can overwrite online logs if there is an error. If you specify REOPEN for a MANDATORY destination, the database stalls the production database when it cannot successfully archive. In this situation, consider the following options:
    Archive manually to the failed destination.
    Change the destination by deferring the destination, specifying the destination as optional, or changing the service.
    Drop the destination.
    When using the REOPEN keyword, note the following:
    ARCn reopens a destination only when starting an archive operation from the beginning of the log file, never during a current operation. ARCn always retries the log copy from the beginning.
    If you specified REOPEN, either with a specified time the default, ARCn checks to see whether the time of the recorded error plus the REOPEN interval is less than the current time. If it is, ARCn retries the log copy.
    The REOPEN clause successfully affects the ACTIVE=TRUE destination state. The VALID and ENABLED states are not changed.

  • Question about Archiving Mailboxes

    I keep almost all of my mail On My Mac > Personal Mail, then divided into subfolders beneath that.  I back up to Time Machine.
    This morning I noticed one of those sub folders was missing, and had to go back more than 24 hours to recover it.  Not sure how it went missing, but it did.
    I am wondering what is the diffference between using Archive Mailbox under the Mailbox menu and keeping the messages in folders On My Mac?
    I think my mailboxes are just getting too darn big, try as I might to contain them.  I'm a mail hoarder.
    Thanks for any insight

    How you backup and what you backup to is all about taking a calculated risk. In other words, what's it worth to you?
    Ultimately, whatever method you choose, I think redundancy is the most important factor. It doesn't matter if you have half-a-dozen backups, if they're all in the same place and the building burns down, they're all gone.
    Yes, hard drives are flimsy and fail, but you can easily duplicate a backup and store them in separate locations; then, make regular maintenance checks to see if they're viable. The probability of two backups failing simultaneously is astronomically low; the probability of one backup failing is astronomically high. For the average person, most dedicated backup systems are too far out of reach, pricewise and from an infrastructure standpoint. Duplicating hard drives is pretty crude, but it's cheap so long as you exercise due diligence.

  • Question about Archive Log Deletion policy

    I've a problem to understand the Archive Log Deletion policy, and I I'd like to this problem explain with the following example.
    Messages of the database are in German, but I guess you'll understand them.
    SQL> startup
    ORACLE-Instance hochgefahren.
    Total System Global Area 5344731136 bytes
    Fixed Size                  2129240 bytes
    Variable Size            2684355240 bytes
    Database Buffers         2617245696 bytes
    Redo Buffers               41000960 bytes
    Datenbank mounted.
    Datenbank geöffnet.
    SQL> archive log list
    Datenbank-Log-Modus              Archive-Modus
    Automatische Archivierung             Aktiviert
    Archivierungsziel            E:\oracle\thetis_iv\arch
    Älteste Online-Log-Sequenz     17917
    Nächste zu archivierende Log-Sequenz   17919
    Aktuelle Log-Sequenz           17919
    SQL> alter system switch logfile;
    System wurde geändert.I created a brand new archive log.
    SQL> exit
    Verbindung zu Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options beendet
    D:\OracleDB\product\11.1.0\db_1\BIN>dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:04    <DIR>          .
    06.04.2011  15:04    <DIR>          ..
    06.04.2011  15:04        17.137.152 ARC17919_0721667907.001
                   1 Datei(en),     17.137.152 Bytes
                   2 Verzeichnis(se), 41.073.258.496 Bytes freiand this is the only archive log in the directory. Now I start rman:
    D:\OracleDB\product\11.1.0\db_1\BIN>rman target / catalog rmanrepo@rmanrepo
    Recovery Manager: Release 11.1.0.7.0 - Production on Mi Apr 6 15:05:35 2011
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Mit Ziel-Datenbank verbunden: ENTWIV (DBID=21045568)
    Kennwort für Recovery-Katalog-Datenbank:
    Verbindung mit Datenbank des Recovery-Katalogs
    RMAN> show all;
    RMAN-Konfigurationsparameter für Datenbank mit db_unique_name ENTWIV sind:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO 'SBT_TAPE';
    CONFIGURE CONTROLFILE AUTOBACKUP OFF;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO 'E:\oracle\thetis_iv\backup\CF_%F_ENTWIV.ORA';
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE SBT_TAPE TO '%F'; # default
    CONFIGURE DEVICE TYPE SBT_TAPE PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 1 BACKUP TYPE TO BACKUPSET; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE SBT_TAPE TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE 'SBT_TAPE' PARMS  'ENV=(TPDO_OPTFILE=D:\Services\Tivoli\TSM\AgentOBA64\tpdo.opt)';
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BZIP2'; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO 'SBT_TAPE';
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO 'D:\ORACLEDB\PRODUCT\11.1.0\DB_1\DATABASE\SNCFENTWIV.ORA'; # defaultThe archive log deletion policy says the the logfiles have to be backed up for two times before they get deleted.
    Now I backup all archive logs, that havn't been backed up for at least two times.
    RMAN> run { backup archivelog all not backed up 2 times
    2>       format '%d_AR_%Y%M%D_%s_%t'
    3>       tag 'ARCHIVE LOGS'
    4>       DELETE ALL INPUT;
    5>     }
    Starten backup um 06.04.2011 15:08:01
    Aktuelles Log archiviert
    Zugewiesener Kanal: ORA_SBT_TAPE_1
    Kanal ORA_SBT_TAPE_1: SID=253 Device-Typ=SBT_TAPE
    Kanal ORA_SBT_TAPE_1: Data Protection for Oracle: version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set für Archive Log wird begonnen
    Kanal ORA_SBT_TAPE_1: Archive Logs in Backup Set werden angegeben
    Eingabe-Archive-Log-Thread=1 Sequence=17919 RECID=147 STAMP=747759899
    Eingabe-Archive-Log-Thread=1 Sequence=17920 RECID=148 STAMP=747760081
    Kanal ORA_SBT_TAPE_1: Piece 1 wird auf 06.04.2011 15:08:02 begonnen
    Kanal ORA_SBT_TAPE_1: Piece 1 auf 06.04.2011 15:08:09 beendet
    Piece Handle=ENTWIV_AR_20110406_23_747760082 Tag=ARCHIVE LOGS Kommentar=API Version 2.0,MMS Version 5.5.1.0
    Kanal ORA_SBT_TAPE_1: Backup Set vollstõndig, abgelaufene Zeit: 00:00:08
    Kanal ORA_SBT_TAPE_1: Archive Logs werden gel÷scht
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17919_0721667907.001 RECID=147 STAMP=747759899
    Archive Log-Dateiname=E:\ORACLE\THETIS_IV\ARCH\ARC17920_0721667907.001 RECID=148 STAMP=747760081
    Beendet backup um 06.04.2011 15:08:10
    RMAN> exit
    Recovery Manager abgeschlossen.
    D:\OracleDB\product\11.1.0\db_1\BIN> dir E:\oracle\thetis_iv\arch
    Datenträger in Laufwerk E: ist Volume
    Volumeseriennummer: 3EBD-77E5
    Verzeichnis von E:\oracle\thetis_iv\arch
    06.04.2011  15:08    <DIR>          .
    06.04.2011  15:08    <DIR>          ..
                   0 Datei(en),              0 Bytes
                   2 Verzeichnis(se), 41.090.396.160 Bytes freirman deleted all archive logs, even I they are on tape only once by now.
    Thats not what I expected. Where is my mistake?

    Hi,
    I do new tests it's very strange.
    BACKUP ARCHIVELOG command is not obeying the policy of archivelog.
    You can open a SR on MOS. (to check bugs)
    I reproduce the same test and the result was the same, it seems that this is a bug.
    CONFIGURE ARCHIVELOG DELETION POLICY TO BACKED UP 2 TIMES TO DISK;
    RMAN> backup archivelog all not backed up 2 times delete all input;
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=1 sequence=15 RECID=16 STAMP=747753711
    input archived log thread=2 sequence=20 RECID=17 STAMP=747753714
    input archived log thread=1 sequence=16 RECID=19 STAMP=747753729
    input archived log thread=2 sequence=21 RECID=18 STAMP=747753729
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t132210_0.304.747753731 tag=TAG20110406T132210 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    channel ORA_DISK_1: deleting archived log(s)
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_15.293.747753711 RECID=16 STAMP=747753711
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_20.295.747753715 RECID=17 STAMP=747753714
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_16.294.747753729 RECID=19 STAMP=747753729
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_21.298.747753729 RECID=18 STAMP=747753729
    Finished backup at 06-APR-11
    RMAN> list archivelog all;
    specification does not match any archived log in the repositoryOracle Docs Says:
    The BACKUP ARCHIVELOG ... DELETE INPUT command deletes archived log files after they are backed up.
    This command eliminates the separate step of manually deleting archived redo logs.
    With DELETE INPUT, RMAN deletes only the specific copy of the archived log chosen for the backup set.
    With DELETE ALL INPUT, RMAN deletes each backed-up archived redo log file from all log archiving destinations.
    As explained in "Configuring an Archived Redo Log Deletion Policy",
    the BACKUP ... DELETE INPUT and DELETE ARCHIVELOG commands obey the archived redo log deletion policy
    for logs in all archiving locations. For example, if you specify that logs should only be deleted when backed
    up at least twice to tape, then BACKUP ... DELETE honors this policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10642/rcmbckba.htm#BRADV89524
    But in ours case it's not honors this policy.
    Only with the FORCE command should this happen. But it is not our case.
    Oracle Docs:
    If FORCE is not specified on the deletion commands,
    then these deletion commands obey the archived log deletion policy.
    If FORCE is specified, then the deletion commands ignore the archived log deletion policy.http://download.oracle.com/docs/cd/E11882_01/backup.112/e10643/rcmsynta010.htm#RCMRF113
    Alternatively you can do the following:
    Set the commands separately.
    Check this:
    RMAN>  run {
    2> backup archivelog all not backed up 2 times ;
    3> delete archivelog all backed up 2 times to disk;
    4> }
    Starting backup at 06-APR-11
    current log archived
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting archived log backup set
    channel ORA_DISK_1: specifying archived log(s) in backup set
    input archived log thread=2 sequence=22 RECID=21 STAMP=747755128
    input archived log thread=1 sequence=17 RECID=20 STAMP=747755127
    channel ORA_DISK_1: starting piece 1 at 06-APR-11
    channel ORA_DISK_1: finished piece 1 at 06-APR-11
    piece handle=+DATA/orcl/backupset/2011_04_06/annnf0_tag20110406t134528_0.295.747755129 tag=TAG20110406T134528 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:01
    Finished backup at 06-APR-11
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=78 instance=orcl1 device type=DISK
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_1_seq_17.298.747755127 thread=1 sequence=17
    RMAN-08138: WARNING: archived log not deleted - must create more backups
    archived log file name=+DATA/orcl/archivelog/2011_04_06/thread_2_seq_22.294.747755129 thread=2 sequence=22
    RMAN>Edited by: Levi Pereira on Apr 6, 2011 1:35 PM

  • Question about packaging material

    Hello all,
    i have encountered an interesting problem.
    I create an order with material r-1150 qty = 10.
    and in delivery i used 2 packaging material, each for 5 pc.
    1. PK-100
    2. PK-095,
    thus 2 HUs are created
    however only PK-095 can be displayed in delivery item list ( together with Material r-1150) .the PK-100 is missing.
    and in the following TO, by picking , i can also only confirm the material and PK-095. and the material qty = 10 .
    then I tried to package with only pk-100, and it is still missing from the delivery item list.
    one thing more weird is , if i change manually in Picking the pk-100 to pk-095, even Pk-095 wont be displayed. Only when I empty the box, and delete the HU and create a new HU with Pk-095, then it could be displayed.
    So do you know how to control this ? and where i can configure to hidden and display the packaging material ?
    thanks!
    regards
    Recke

    Hello, Sridhar,
    first, thanx for ur answer,
    but I think u misunderstood it.
    In the first 1. example
    In the delivery, I packed the qty 10 into 2 different packaging material
    1. PK-100 ---- 5 pc
    2. PK-095 ---- 5 pc.
    but in the item list of delivery, i can only see PK-095, and the following TO, I can pick the total 10 pc and PK-095
    but the PK-100 is missing ( despite that it contains 5 pcs)
    In the 2nd example.
    I only use PK-100 to pack in delivery, but still missing in the item list. That means only my 10 pc product is there.
    see attached file
    in my example 3
    I only use PK-095 as packaging material, it displays together product.with out any problem.
    so do u know what is the matter here ??

  • Question about configurable material

    Hi :
    Can somebody tell me what's required step to set up a configurable material? What are dependencies,profiles?And how do they work?

    Create configurable material mmk1/in mm01 use material type KMAT, characteristics CT04, classes Cl02, configuration profile Cu41, define the class in material master in classification view, assign objects to class Cl24N
    Create BOM Work center, Routing and proceed.
    If you have any doubt refer
    http://help.sap.com/erp2005_ehp_03/helpdata/EN/92/58c09b417011d189ec0000e81ddfac/frameset.htm

  • Question about planning material

    Is it possible in SAP to have MRP use the planning material field in the material master (MPGD-PRGRP) do the following function or if not is there an SAP solution to my issue:
    Customer wants to use part ABC in BOM's, and also stock this material in inventory. However, the customer gets this material from another plant in their SAP system via a stock transport order but wants MRP to place the requirement on the supplying plant as part PABC which is the part in the supplying plant. The vision would be for the supplying plant to ship PABC and once it is received a 309 movement type would be done to convert the part back to ABC.
    Any thoughts on this?

    Hi,
    It will be better if you use Special procurement key as 40 in MRP 2 view of the Material PABC in the Receiving Plant. Also Material PABC will be maintained for both the Plants (Supplying & Receiving both)
    Then do MRP Run in Receiving Plant for Material PABC, as a result of this it will generate a Stock Transport PR in the Receiving Plant and then proceed with the STO Process and receive material in Receiving Plant. Then do transfer posting using movement type 309.

  • White vs black macbooks (question about the material)

    Hi, I'm trying to buy my first mac and I am really excited. But I'm not sure if I should get the black or the white macbook (same specs). Any advice is appreciated.
    I've heard a lot about how the new Black Macbooks have a different finish than the white ones - matte compared to the glossy finish. In layman's terms, basically all I have learnt is that the black ones have a lot of stains on them. But I would think that the white ones get dirty pretty fast. Also I've read some reviews where some folks have said that the black has a different keyboard ... I tried both of them out at an Apple store and didn't notice much of a different.
    Is there really that much of a difference between the two colors? If there isn't, I'd like to get the Black one just cuz it won't get dirty that quickly.
    Any thoughts? Thanks for your time

    I just went through the same dilemna. I had to decide whether paying $200 more for 40GB more of hard drive space and the black color. (The 40 GB is only worth $75 so you are paying $125 for the black color.) I ended up choosing the black. Sure, it smudges easily, however, that is really easy to clean. As for the scratches, once they are in the white ones, they are pretty much there for good. The black ones, you have a great chance of buffing them out if it is just a surface scratch.
    I didn't need the 40GB of space, however, the black looks really cool, classy and professional and is easier to keep clean. Plus it accentuates the glowing apple far better than the white one. It really sounds like you want the black, so GO FOR IT!
    2.16 Intel Core Duo MacBook (Black)   Mac OS X (10.4.9)  

  • Questions about archive

    Hope anyone can help me understand this…
    I found there are various kinds of archive feature in Exchange or Outlook. And I need help to choose the appropriate method to do archive in my Exchange.
    Thanks

    Hi David,
    Exchange Server 2010 has a new built-in archiving solution, which lets you create a personal archive mailbox within the Exchange organization.
    You can use the Enable-Mailbox "xxx" -Archive cmdlet to enable the personal archive for a user's mailbox.
    What's more, here is an article for your reference.
    Understanding Personal Archives
    http://technet.microsoft.com/en-us/library/dd979795(v=exchg.141).aspx
    Hope it helps.
    If you need further assistance, please feel free to let me know.
    Best regards,
    Amy
    Amy Wang
    TechNet Community Support

  • Questions about using Bitlocker without TPM

    We currently use Bitlocker to encrypt our Windows 7 computers with TPM. Now we are looking at encrypting some Windows 7 computers without a TPM. I see how to change the group policy setting to allow Bitlocker without a TPM. I have looked at a lot of other
    threads and I have a few questions about how the Bitlocker without TPM works.
    1) I see a USB drive containing a key is required for Bitlocker configurations without a TPM, say the end user loses this USB drive, what are the recovery options for their computer? 
    This article seems to indicate that without the USB drive connected, you are unable to even access recovery options http://blogs.technet.com/b/hugofe/archive/2010/10/29/bitlocker-without-tpm.aspx
    We have recovery backed up to AD when Bitlocker is enabled, but how could we do this recovery on a computer on computer where it's USB is lost? Would we have to remove the HD itself and attach it to another computer to access?
    2) After enabling Bitlocker on a computer without a TPM and using the USB Drive for the key, is there a way to also add a PIN or password protection at bootup?

    Hi,
    Sorry for my dilatory reply, 
    Configuring a startup key is another method to enable a higher level of security with the TPM. The startup key is a key stored on a USB flash drive, and the USB flash drive must be inserted every time the computer starts. The startup key is used to provide
    another factor of authentication in conjunction with TPM authentication. To use a USB flash drive as a startup key, the USB flash drive must be formatted by using the NTFS, FAT, or FAT32 file system.
    You must have a startup key to use BitLocker on a non-TPM computer.
    From: http://technet.microsoft.com/de-de/library/ee449438(v=ws.10).aspx#BKMK_Key
    For more Q&A about BitLocker, you can refer to the link above.
    hope this is helpful.
    Roger Lu
    TechNet Community Support

  • Questions about Indexing and Using an Indexing POA

    Although I have only about 50 users, at least 15 of them have in excess of 100,000 messages in their accounts and the POA (version 7.0.2) is regularly slowing to a crawl. (I just know that plans for revolution are fomenting!) I have embarked on a campaign to reduce these accounts by archiving everything off to get mail accounts down to 3000 or fewer pieces. I have achieved user buy-in, but have worked on only a few users so far.
    In another closely related thread, it was suggested to me that the PO speed issues relate to broken indexes. And I suspect that given so many messages, the indexes were never getting fully rebuilt with the default QF POA settings. I am trying to fix that situation in addition to reducing mail account sizes. So, I have set up a second POA on another server and dedicated it to the indexing task. The /qfinterval is set for 1 hour, other /qf switches at default. The POA-QF does no mail delivery, but it does do nightly user upkeep.
    The POA-QF seems to be steadily working away and making progress at reducing the number of unindexed messages. However, I have questions about what I am seeing and what more I can do:
    1. Is the progress I am seeing real progress? For example I have a user with over 100,000 messages to be indexed and every time I check the logs, the count drops by about 500 messages per hourly QF cycle. I assume that if I just let it keep running, it will eventually get caught up and fixed. Not only with this user, but with all the others as well. Will my patience (and theirs) be rewarded? Are there any gotchas I need to prepare for?
    2. One user has recently had virtually all of her messages successfully moved to archive. I can see them in the Archive, and do not see them in the online account. However, now over a week later, QF still shows >130,000 items still left to index for that user. The POA-QF is making slow, steady progress reducing that number, but why is this user's QF count still so high? Does it just need more time, or is there something amiss for this user?
    3. I may want to rebuild indexes for single users from scratch. I have seen the TID 3105742 which tells how to do this: Essentially you turn off mail delivery functions, and make some other switch changes to dedicate the POA to indexing for just a single user, and then you let the POA rebuild the indexes. The implication of that scenario is that the POA is now enjoying exclusive access to the user's databases.
    If I want to use my secondary POA-QF to rebuild a user's index from scratch, does the main POA have to be offline and the user out of GWise? That is, Does the QF process require exclusive access in order to rebuild indexes from scratch?
    Thanks for any thoughts or suggestions.
    Peter Smick

    pgsmick wrote:
    > 1. Is the progress I am seeing real progress? For example I have a user with
    > over 100,000 messages to be indexed and every time I check the logs, the count
    > drops by about 500 messages per hourly QF cycle. I assume that if I just let
    > it keep running, it will eventually get caught up and fixed. Not only with
    > this user, but with all the others as well. Will my patience (and theirs) be
    > rewarded? Are there any gotchas I need to prepare for?
    Set this switch for this indexing POA - /qflevel=999 - this will index
    everything in one run. It will take a long time, but with no qflevel switch you
    are indeed only indexing 500 messages at a time, and if the user has that much
    mail, it might never really catch up.
    >
    > 2. One user has recently had virtually all of her messages successfully moved
    > to archive. I can see them in the Archive, and do not see them in the online
    > account. However, now over a week later, QF still shows >130,000 items still
    > left to index for that user. The POA-QF is making slow, steady progress
    > reducing that number, but why is this user's QF count still so high? Does it
    > just need more time, or is there something amiss for this user?
    >
    This is odd, because really the index count should drop to nothing, but with the
    above switch this might get resolved as well.
    > 3. I may want to rebuild indexes for single users from scratch. I have seen
    > the TID 3105742 which tells how to do this: Essentially you turn off mail
    > delivery functions, and make some other switch changes to dedicate the POA to
    > indexing for just a single user, and then you let the POA rebuild the indexes.
    > The implication of that scenario is that the POA is now enjoying exclusive
    > access to the user's databases.
    Not really - the POA is not enjoying exclusive access to the user's database,
    the indexer is just avoiding an attempt to index anything else.
    > If I want to use my secondary POA-QF to rebuild a user's index from scratch,
    > does the main POA have to be offline and the user out of GWise? That is, Does
    > the QF process require exclusive access in order to rebuild indexes from
    > scratch?
    No - QF never requires exclusive access. That said, you may find that an
    extremely vigorous QF can cause slowdowns for the user.
    Danita
    Novell Knowledge Partner
    Moving GroupWise to Linux?
    http://www.caledonia.net/gwmove.html

  • Re: Questions about Plan structure

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

    Reply-To: "Duncan Kinnear" <[email protected]>
    Q. if you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    A It is better to have them in the same plan.It depends on the design
    you have
    Q Why should the Managers be separated from their business classes?
    A Managers are usually service objects which might require different
    resources.
    These managers when deployed might be required by several other
    applications.
    Seperating them as a different plan will help in just using one
    installed partition to be
    Used by different applications (refer about Reference Partition)
    Q If you have the Database managers separate, what scope does each DB
    manager cover?
    A It is better to have the DB managers in user scope.
    It depends on the numbers of users, u have for the system.
    Since u are talking about 100 tables. It's a huge system.
    It also depends on the user licenses u have for the backend.
    Take care that u use proper load balancing of DB Managers for the
    system.
    Krishna CVSR
    GoldStone Softech Inc
    >
    Hi there,
    We are in the middle of designing the structure of a new system.
    I have read/heard that it is best to break down the plans into the
    following categories:
    Business Classes
    Managers/Services
    Clients
    I have a few questions about this structure:
    Should related business classes be grouped together in Plans? E.g. if
    you have an Employee class with related EmployeeSubordinate and
    EmployeeSalaryHistory classes, should these all be in the same plan?
    Why should the Managers be separated from their business classes?
    E.g. if there is an EmployeeMgr service which deals with anything to do
    with the Employee business class, why separate them in different
    plans? If you need the manager to access the class, you will always
    need both.
    Some of the Forte documentation talks about "Policy" managers and
    "Database" managers. How do the functions of these managers differ
    for a simple CRUD (Create Read Update Delete) class?
    If you have the Database managers separate, what scope does each
    DB manager cover? I.e. do you have one for the entire database (over
    100 tables in our case), or do you break it down by sub-system?
    Thanks in advance for any answers.
    Cheers,
    Duncan Kinnear,
    McCarthy and Associates, Email: [email protected]
    PO Box 764, McLean Towers, Phone: +64 6 834 3360
    Shakespeare Road, Napier, New Zealand. Fax: +64 6 834 3369
    Providing Integrated Software to the Meat Processing Industry for over 10 years
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>
    Get Your Private, Free Email at http://www.hotmail.com
    To unsubscribe, email '[email protected]' with
    'unsubscribe forte-users' as the body of the message.
    Searchable thread archive <URL:http://pinehurst.sageit.com/listarchive/>

  • Question about Kurts comments discussing the seperation of AIA & CDP - Test Lab Guide: Deploying an AD CS Two-Tier PKI Hierarchy - Kurt L Hudson MSFT

    Question about the sentence in bold. What is the meaning behind this comment?
    How would you separate the role of the AIA and CDP from a CA subordinate server? I can see where I add a CES and CEP server which has those as well, but I don't completely understand his comment. Because in this second step, (http://technet.microsoft.com/en-us/library/tlg-key-based-renewal.aspx)
    he shows how to implement CES and CEP.
    This is from the guide located at: http://technet.microsoft.com/library/hh831348.aspx
    Step 3: Configure APP1 to distribute certificates and CRLs
    In the extensions of the root CA, it was stated that the CRL from the root CA would be available via http://www.contoso.com/pki. Currently, there is not a PKI virtual directory on APP1, so one must be created.
    In a production environment, you would typically separate the issuing CA role from the role of hosting the AIA and CDP.
    However, this lab combines both in order to reduce the number of resources needed to complete the lab.
    Thanks,
    James

    My concern is, they have a 2-3k base of xp systems, over this year they are migrating them to Windows 7. During this time they will also be upgrading hardware for the existing windows 7 machines. The turnover of certificates are going to be high, which
    from what I've read here, it worries me.
    http://blogs.technet.com/b/askds/archive/2009/06/24/implementing-an-ocsp-responder-part-i-introducing-ocsp.aspx
    The application then can go to those locations to download the CRL. There are, however, some potential issues with this scenario. CRLs over time can get rather large
    depending on the number of certificates issued and revoked. If CRLs grow to a large size, and many clients have to download CRLs, this can have a negative impact on network performance. More importantly, by
    default Windows clients will timeout after 15 seconds while trying to download a CRL. Additionally,
    CRLs have information about every currently valid certificate that has been revoked, which is an excessive amount of data given the fact that an application may only need the revocation status for a few certificates. So,
    aside from downloading the CRL, the application or the OS has to parse the CRL and find a match for the serial number of the certificate that has been revoked.
    With the above limitations, which mostly revolve around scalability, it is clear that there are some drawbacks to using CRLs. Hence, the introduction of Online Certificate
    Status Protocol (OCSP). OCSP reduces the overhead associated with CRLs. There are server/client components to OCSP: The OCSP responder, which is the server component, and the OCSP Client. The OCSP Responder accepts status
    requests from OCSP Clients. When the OCSP Responder receives the request from the client it then needs to determine the status of the certificate using the serial number presented by the client. First the OCSP Responder determines if it has any cached responses
    for the same request. If it does, it can then send that response to the client. If there is no cached response, the OCSP Responder then checks to see if it has the CRL issued by the CA cached locally on the OCSP. If it does, it can check the revocation status
    locally, and send a response to the client stating whether the certificate is valid or revoked. The response is signed by the OCSP Signing Certificate that is selected during installation. If the OCSP does not have the CRL cached locally, the OCSP Responder
    can retrieve the CRL from the CDP locations listed in the certificate. The OCSP Responder then can parse the CRL to determine the revocation status, and send the appropriate response to the client.

Maybe you are looking for

  • Different formats of the flat file for the same target

    In our deployment, we use plugin code to extract the csv files in the required format. The customers are on same version of datamart, but they are on different versions of source database - from 3.x to 4.5 depending on which version of application th

  • Fsck ok; Time Machine backup not

    Just that. Time Machine gave me the old "Time Machine completed a verification of your backups on "WDMyCloud". To improve reliability, Time Machine must create a new backup for you." I ran fsck on the drive, and it returned: /dev/rdisk3s2: ** The vol

  • Saving the result set in a Search page

    Hi , We have a manual search page where the results are displayed in the form of 10 columns. The first 5 are read only fields whereas rest 5 are editable checkboxes. There is no Query region created in our page. Suppose, fields a, b, c are search par

  • ITunes doesn't load - saying "Directory iTunes doesn't exist" ????

    Dear all, I have reinstalled iTunes 7.xx and get the same error when trying to start iTunes. The error message tells me: Directory "iTunes" doesn't exist in Dir: "Eigene Musik", but it exists. What am I doing wrong ? Could I please get some help ? Th

  • How can I save photos to iPhoto direct from Photoshop Elements

    Does anyone know how I can save a photo direct to iPhoto after editing in PSE?  Currently I save to the desktop and then "import".  Its a bit of a fiddly way espectially if you  are editing a number of photos direct from iPhoto.  It would be nice to