OIM Database Connector - ID PK from DB Sequence Best Practice?

Figured I'd ask around before I hacked something together for something that is a very common scenario. When provisioning my PK (ID) is not set and is set via a Oracle Sequence. I'm currently using the Database Application Tables GTC connector. A couple approaches, based on my limited OIM knowledge:
1. OnInsert DB trigger to handle. Requires nothing from OIM for this, but I really feel like this is way too invasive and doesn't support the "passive" nature of Identity Management.
2. Deviate from the GTC and get the SEQUENCE from a custom adapter task.
3. Go have a beer.
If there is some documentation or posts regarding this please let me know. I checked the best practices & connector docs and found nothing about this.

I think you have a good grasp of the problem. Option 3 is of course superior to the others :)
If you start using pre-pops on GTC generated forms you can't regenerate the GTC or you risk losing the pre pop connectors.
There has been some discussions around what Oracle supports when it comes to GTC but it seems like you now can add prepop and entity adapters and still stay in support.
Best regards
/M

Similar Messages

  • OIM Database Connector recon not working

    I am using OIM 9.1.0.0. with connector pack 9.0.4.1 on JBOSS 4.0.3SP1 Windows 2k3sp2
    I imported the database connector and tried to recon but i get the following error message
    14:36:31,444 INFO [STDOUT] 14:36:31,444 INFO [DBADAPTERLOGGER] DBReconciliation::execute : Check for create/update reconciliation
    14:36:31,444 INFO [STDOUT] 14:36:31,444 INFO [DBADAPTERLOGGER] DBReconciliation::doTasks : query is SELECT idmuser.USER_ID FROM idmuser
    14:36:31,444 INFO [STDOUT] 14:36:31,444 INFO [DBADAPTERLOGGER] DBReconciliation::createReconEvents : ##### Users present in database ##### 3
    14:36:31,459 INFO [STDOUT] 14:36:31,459 INFO [DBADAPTERLOGGER] DBReconciliation::createReconEvents : Reconciliation in Trusted Mode
    14:36:31,897 INFO [STDOUT] 14:36:31,897 INFO [DBADAPTERLOGGER] DBReconciliation::createReconEvents : Reconciliation in Trusted Mode
    14:36:32,225 INFO [STDOUT] 14:36:32,225 INFO [DBADAPTERLOGGER] DBReconciliation::createReconEvents : Reconciliation in Trusted Mode
    14:36:32,881 INFO [STDOUT] 14:36:32,881 INFO [DBADAPTERLOGGER] DBReconciliation::createReconEvents : Create/Update Reconciliation Successfully Completed
    14:36:32,897 INFO [STDOUT] 14:36:32,897 ERROR [TASK] Class/Method: SchedulerBaseTask/run encounter some problems: {1}
    java.lang.NullPointerException
    at com.thortech.xl.integration.dbadapter.reconciliation.DBReconciliation.execute(DBReconciliation.java:267)
    at com.thortech.xl.scheduler.tasks.SchedulerBaseTask.run(Unknown Source)
    at com.thortech.xl.scheduler.core.quartz.QuartzWrapper$TaskExecutionAction.run(Unknown Source)
    at Thor.API.Security.LoginHandler.jbossLoginSession.runAs(Unknown Source)
    at com.thortech.xl.scheduler.core.quartz.QuartzWrapper.execute(Unknown Source)
    at org.quartz.core.JobRunShell.run(JobRunShell.java:178)
    at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:477)
    I am using Jboss 4.03SP1 with log4j.xml as
    <category name="DBADAPTERLOGGER">
    <priority value="ALL"/>
    </category>
    <category name="XL_INTG.OID">
    <priority value="WARN"/>
    </category>
    <category name="WEBAPP"> <priority value="WARN" /></category>
    <category name="PERFORMANCE"> <priority value="WARN" /></category>
    <category name="ADAPTERS"> <priority value="DEBUG" /></category>
    <category name="XELLERATE.DDM">
    <priority value="WARN"/>
    </category>
    <category name="OIMCP.DATC">
    <priority value="WARN"/>
    </category>
    <category name="XELLERATE.DATABASE">
    <priority value="WARN"/>
    </category>
    <category name="Adapter.Oracle-eBizUM">
    <priority value="WARN"/>
    </category>
    <category name="XELLERATE">
    <priority value="WARN"/>
    </category>
    <category name="com.nexaweb.server">
    <priority value="WARN"/>
    </category>
    <category name="com.opensymphony.oscache">
    <priority value="WARN"/>
    </category>
    Here is the xml file for the database adapter
    <?xml version="1.0"?>
    <xdb_app_map     xmlns:xsi = "http://www.w3.org/2001/XMLSchema-instance"     name = "mapping" xsi:noNamespaceSchemaLocation = "file:///c:/reconapp/xdb_app_map.xsd" >
         <target_application>
              <target_app_name>reconapp</target_app_name>
              <target_app_ver>1.0</target_app_ver>
              <target_app_provider>xxxx</target_app_provider>
         </target_application>
         <target_database><database name="Oracle"><properties /></database></target_database>
         <database_adapter>
              <operation name="create">
                   <task table_name="idmuser" xeltask_type="insert">
                        <column col_name="USER_ID"          data_type="VARCHAR2"     data_typ_size="1024"     col_info="primary"     col_type="xellerate" xel_data_source="xel_usr_id"               required="true" />
                        <column col_name = "FNAME"     data_type = "VARCHAR2"     data_typ_size = "100"                              col_type = "xellerate" xel_data_source = "xel_usr_first_name"     required = "true"/>
                        <column col_name = "LNAME"     data_type = "VARCHAR2"     data_typ_size = "100"                              col_type = "xellerate" xel_data_source = "xel_usr_last_name"     required = "true"/>
                        <column col_name = "PASSWORD"     data_type = "VARCHAR2"     data_typ_size = "100"                              col_type = "xellerate" xel_data_source = "xel_usr_password"     required = "true"/>
                        <column col_name = "STATUS"     data_type = "VARCHAR2"     data_typ_size = "100"                              col_type = "xellerate" xel_data_source = "xel_usr_status"     required = "true"/>
                        <column col_name = "ORG"     data_type = "VARCHAR2"     data_typ_size = "100"                              col_type = "xellerate" xel_data_source = "xel_usr_org"     required = "true"/>
                   </task>
              </operation>
              <operation name="reconcileCreateUpdate" enabled="true">
                   <task table_name="idmuser" xeltask_type="select">
                        <column table_name="idmuser" col_name="USER_ID" data_type="VARCHAR2" data_typ_size="1024" col_info="primary" required="true" col_type="xellerate" xel_data_source="xel_usr_id" />
                   </task>
                   <task table_name="idmuser" xeltask_type="select">
                        <column table_name="idmuser" col_name="USER_ID" data_type="VARCHAR2" data_typ_size="1024" col_info="primary" required="true" col_type="xellerate" xel_data_source="xel_usr_id" />
                        <column table_name="idmuser" col_name="FNAME" data_type="VARCHAR2" data_typ_size="100" required="true" col_type="xellerate" xel_data_source="xel_usr_first_name" />
                        <column table_name="idmuser" col_name="LNAME" data_type="VARCHAR2" data_typ_size="100" required="true" col_type="xellerate" xel_data_source="xel_usr_last_name" />
                        <column table_name="idmuser" col_name="PASSWORD" data_type="VARCHAR2" data_typ_size="100" required="true" col_type="xellerate" xel_data_source="xel_usr_password" />
                        <column table_name="idmuser" col_name="STATUS" data_type="VARCHAR2" data_typ_size="100" required="true" col_type="xellerate" xel_data_source="xel_usr_status" />
                   </task>
              </operation>
         </database_adapter>
         <mapping_data>
              <mapping_date>30/7/2008 00:00:00</mapping_date>
              <mapping_version>1.0</mapping_version>
         </mapping_data>
    </xdb_app_map>
    What can be the problem ?? is there something with recon timestamp to do with this ?

    fixed it ...
    missing operation reconDelete
    added the following snippet
    <operation name="reconcileDelete" enabled="false" />

  • OIM Database Connectors version upgrade

    Hi,
    I need to upgrade the OIM DB User Management Connector from release 9.0.4.1 to 9.0.4.4. and the OIM Database Application Table Connector from release 9.0.4.1 to 9.1.0.1. Any one knows which patches should be applied for these upgrades? I can't find any documentation regarding this.
    Thank you.
    Regards.

    First of all have a look at the read me which comes along with the connector patch and see if the upgrade you are trying to achieve is supported or not. For example:
    The *4 Deploying the Patch Set* section in the latest release *9.0.4.5* of the DB User Management Connector clearly says the following:
    Note:
    Direct upgrade to release 9.0.4.5 from release 9.0.4.3 or earlier releases is not supported.
    Same applies to DB App Tables Connector as well. See and confirm it by yourself because I don't have those versions of the connector.
    Thanks
    Sunny

  • When to use unattend.xml in task sequence - best practice?

    Hi, I've tried researching this but not found an answer to my specific query.
    We have ConfigMgr 2012 R2 with MDT 2013 although I don't think this is an MDT specific question.
    I'm trying to create a Build and Capture task sequence for our Windows Server 2008 R2 and Server 2012 /2012R2 server builds utilising an UNATTEND.XML file to make some customisations that can be deployed for every build afterwards in a Deployment Task Sequence.
    Specifically the addition of some Windows Features like SNMP and it's configuration and the addition of the Telnet Client. There are other bits like language settings and configuration items but I'm specifically interested in the Features part for my question.
    In CM 2012R2 you now have the option under the "Apply Operating System" to use a captured image or an original installation source. However they work differently if you specify the use of the same unattended answer file.
    The "image" deployment ignores all of the "add features" sections of the XML file and the "installation source" loses the  configuration options from SNMP from the XML file. When you then deploy the captured image using
    the same unattend.xml again the one from the "installer" now has all the SNMP features required and the one from the "image" is still missing everything.
    So my question is as follows.
    What is best practice for specifying an unattend.xml file in a task sequence. Is it in the build and capture TS or in the Deployment TS ?
    or
    Do I need multiple XML files, one for build and capture with some bits in and another for deployment with the rest in?
    or
    Should I be doing something else?
    Although this is specifically asking about Server O/S we will be using the same methodology for Windows 7 deployment.

    In this case DISM is only used to add the actual features... for configuration you could use a simple script that runs afterwards. Sample registry file:
    SAMPLE REG FILE - HKLM-SNMP.reg
    Windows Registry Editor Version 5.00
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters]
    "NameResolutionRetries"=dword:00000010
    "EnableAuthenticationTraps"=dword:00000001
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\PermittedManagers]
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\RFC1156Agent]
    "sysServices"=dword:0000004f
    "sysLocation"=""
    "sysContact"=""
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\TrapConfiguration]
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\TrapConfiguration\public]
    "1"="127.0.0.1"
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\ValidCommunities]
    "public"=dword:00000004
    Sample batch file:
    SAMPLE SCRIPT FILE - ConfigureSNMPService.bat
    @ECHO OFF
    net stop "SNMP Service"
    regedit /s HKLM-SNMP.reg
    net start "SNMP Service"
    Also some settings for SNMP can be controlled through group policy:
    http://serverfault.com/questions/285762/group-policy-for-multiple-snmp-permitted-managers

  • Import data from excel file - best practice in the CQ?

    Hi,
    I have question related to importing data from excel file and creates from those data a table in the CQ page. Is inside CQ some OOTB component which provides this kind of functionalities? Maybe somebody implement this kind of functionality or there is best practice to do this kind of functionalities?
    Thanks in advance for any answer,
    Regards
    kasq

    You can check a working example package [1] (use your Adobe ID to log in)
    After installing it, go to [2] for immediate example.
    Unfortunately it only supports the old OLE-2 Excel format (.xls and not .xlsx)
    [1] - http://dev.day.com/content/packageshare/packages/public/day/cq540/demo/xlstable.html
    [2] - http://localhost:4502/cf#/content/geometrixx/en/company/news/pressreleases/my_personal_bes ts.html

  • (Swing) Reading fields from other forms - Best practice

    Hi All,
    I have been learning Java and Swing through NetBeans (quite a cool way to do it). I would like some 'best-practice' advice please. Google and search on this forum are normally my friends, can't find anything relevant for this though. I have created an application that seems to be functional, all my navigation is through tabs.
    To make my app portable I have created a ?Settings? part. To save having another tab, I have basically copied the way NetBeans deals with opening the ?About Box? in its example, and have set-up fields in here. This essentially opens a JDialog with fields through an option in the menu.
    I figure the best way to store the data is in a simple text file, and have the application read the values on start-up? Then when somebody changes the fields and saves it just overwrites the text file with the new values. I am not sure how to best read these in though, I can?t find a way of reading direct from the JDialog (i.e. var.getText()) which makes sense as the JDialog will be closed.
    Will my best practice simply be writing a class which reads the file, then whenever I need a value just calling the class?
    Thanks a lot

    You should not be using preference files unless you really know what you are doing. This is one area of the very few areas of Java programming that can have lasting nasty negative affects on a computer that runs the code if done wrong.
    If you want the user to save preferences it is better to just use xml. The advantage of preference files is they can be used by multiple programs and are permanent even after the program is done running/deleted/writen-over/forgotten. The disadvantages of preference files is they can be used by multiple programs, are permanent even after the program is done running/deleted/writen-over/forgotten, can cause damage to your OS, may be seen as harmfull by add blockers and/or anti-virus software, may lead to conflicts with other software that uses similar naming schemes, and are pretty much unportable.
    You are going to be much better off in the long run learning XML and just saving them there, an added benefit of XML is if you screw something up and it causes your program to no longer run you can easily edit the xml by hand.
    JSG

  • Working with many sequences- best practice

    Hi.
    I´ve just started using Adobe Premiere CS6. My goal is to create a 2 hour long movie, based on 30 hours of raw gopro footage recorded on a recent vacation.
    Now my question is, what is the best practice for working with so many sequences/movie clips?
    Do you have one heavy project file, with all the clips?
    Or do you make small chapters that contains x number of sequences/is x minutes long, and in the end combine all these?
    Or how would you do it the best way, so its easiest to work with?
    Thanks alot for your help.
    Kind regards,
    Lars

    I'll answer your second question first, as it's more relevant to the topic.
    You should export in the very highest quality you can based on what you started with.
    The exception to this is if you have some end medium in mind. For example, it would be best to export 30 FPS if you are going to upload it to YouTube.
    On the other hand, if you just want it as a video file on your computer, you should export it as 50 FPS because that retains the smooth, higher framerate.
    Also, if you are making slow-motion scenes with that higher framerate, then export at the lowest framerate (for example, if you slow down a scene to 50% speed, your export should be at 25 FPS).
    About my computer:
    It was for both, but I built it more with gaming in mind as I wasn't as heavily into editing then as I am now.
    Now, I am upgrading components based on the editing performance gains I could get rather than gaming performance gains.

  • Import files into bridge from a CMS - best practice

    Hi All,
    i am trying to come up with the simplest solution for the following:
    we have a CMS in which our content is managed and edited.
    the content editor might like to edit video files that reside in our asset management system.
    our editing tool is Adobe Premiere.
    My thought was to transfer the video file along with its metadata and edit instructions to an FTP server from our asset repository and then transfer these files into the video editing local environment using bridge. after editing the file - it should be transferred back into the asset repository via FTP server or such.
    it seems the way to this would be to extend the bridge functionality using the java script SDK, but i would love to know if there is a best practice solution before we start detailed design and development.
    your help is highly appreciated and thanks in advance,
    Deena

    Not sure what advice you have been given and how you have interpreted it.
    You are of course referring to im6 here since you can't drag projects from im08 to iDVD.
    I don't make many DVD's anymore and may well be wrong here but I'm not aware that dragging an im6 project to iDVD is any different than using share/iDVD and I'm wondering if the advice relates to exporting from im6 and then importing the exported movie into iDVD which is indeed a different workflow.

  • OIM Database Connector User not created

    I am using OIM 9.1.0.0 with connector pack 9.0.4.1. When I run trusted recon i see the new user appear in the recon manager, but the user is not created. When I open the event and click on Create User the user is created.
    But before this the recon event shows data missing.
    How can automate it ??

    The GTC framework will generate the rule for you if you configure the matching only property. Please take another look at the docs.
    Or you can follow PInk's suggestion and crate the rule manually.
    Best regards
    /M

  • Primary key, column, sequence : best practice.

    Hi I was wondering when designing table should you limit the length of primary key.
    Because if you have limited column length, and your adding and removing items than you have sequence gap, than you will need to resize column length and if their foreign key (resize them in foreign key relationship too) or you need to reset sequence to go back to the maximum (id).
    WHat are the benefits and disbenefits of having limited column length. Is their way to restrict the maximum number of items without putting column length in their (inside the database) not the application.
    Thanks.

    Unless you know the exact number of rows your table will have, it is not convenient to limit the size. Now it all depend on what you mean by limiting the size and which column type you are talking about.
    Here I have some notes about --> Why is Varchar2 oversizing bad?, where I state this:
    Apparently the same space consumption would be the same if a Varchar2(4) is used versus a Varchar2(4000), if you require one byte, then one byte will be allocated, and performance should be the same, but when taking a look at the performance statistics there are some surprises that arise.
    When fetching information from Oracle, buffers and fetching structures are being prepared beforehand to be able to manage the maximum column lenght, applications such as toad may run out of memory trying to allocate resources to retrieve the information from an unnecessarily over sized varchar2 column.
    The answer to your case is: you should better plan according to your actual requirements, if you oversize columns you will degrade performance and you will have memory issues on the long run. If you under size then you will have problems with maximum capacity in your tables and you will be forced to -most probably- perform online maintenance to your objects to resize them.
    ~ Madrid
    http://hrivera99.blogspot.com/

  • Access AM method from backing bean - best practice

    Hello,
    I need to call an appModule method from a backing bean. What is better?
    1. Resolve bindings.AMDataControl.dataProvider, get AM and invoke the method directly
    or
    2. Create method action in page definition, resolve OperationBinding from it's binding expression and execute the operationBinding
    I don't need to have a button on the page for this method.
    Rado

    Good question Rado.
    With a "flexible" framework like Oracle ADF, I also ask myself sometimes what would be the preferred approach to do things.
    See also these forum threads:
    - about "ADF Faces: Passing values from a managed bean to ADF BC module"
    http://forums.oracle.com/forums/thread.jspa?threadID=486742
    - about : "ADF Faces: getting selected row data in multi select table ..."
    http://forums.oracle.com/forums/thread.jspa?threadID=503248
    regards
    Jan Vervecken

  • HCM Master data upload sequence & best practices

    Experts,
    What would be the best method and recommended sequence to upload HCM master data into the below infotypes?
    0,1,2,3   6,7,8,9   207,208,209,210  (payroll)
    21,167,168,169,170,171 and 3.series (Benefits)
    PA0795
    PA2006
    PA2012
    PBO795
    T529T
    T530T
    Please advice.
    Thanks in advance.
    NW

    Hi,
    The best method to mass upload is LSMW
    the sequence will be
    First you need to create the master data so the Action tables need to be configured first
    T529T
    T530T
    also other related PA config need to be completed
    Then when you will start uploading data the sequence will be
    0, 1, 2 , 6, 7, 8, 9 , 207, 208, 209, 210, 21, 171, 167, 168 169, 170, 2006, 2012, 795
    The benefits features (BAREA, BENGR, BSTAT) also need to be configured with all other benefits related config prior to uploading benefits information
    Some other imp. features like LGMST, TARIF, ABKRS, SCHKZ etc.. also need to be configured prior to uploading the employee master data.
    Hope this will be of help
    Regards,
    Guds

  • Using Accessor Method from Within Considered Best Practice?

    I was wondering what is considered best practice in the AS3 world regarding this topic.
    In C++ and .NET world, as well as PHP I would consider using accessor methods, instead of directly accessing the class property, bad practice due to unnecessary overhead.
    What are your thoughts in this?

    Getters and setters and bindable vars can be public, protected or private.  Which one you choose follows standard object-oriented practices.
    is just like a macro that defines a particular get/set pair.  I would only use it on vars.  Using it on a getter simply wraps the getter in another get/set pair which is wasteful.  The Flex Framework rarely uses .  Instead, we typically define get/set pairs with metadata so we can control what event fires and when.  It also saves SWF size because the compiler doesn't have to generate a long and ugly name to avoid name collisions.
    Alex Harui
    Flex SDK Developer
    Adobe Systems Inc.
    Blog: http://blogs.adobe.com/aharui

  • Looking For Guidance: Best Practices for Source Control of Database Assets

    Database Version: 11.2.0.3
    OS: RHEL 6.2
    Source Control: subversion
    This is a general question aimed at database professionals, however, it is not specific to any oracle version, etc.  Its a leadership question for other Oracle shops regarding source control.
    The current trunk, in my client's source control, is the implementation of a previous employee who used ER Studio.  After walking the batch scripts and subordinate files , it was determined that there would be no formal or elegant way to recreate the current version of the database from our source control - the engineers who have contributed to these assets are no longer employed or available for consulting.  The batch scripts are stale, if you will.
    To clean this up and to leverage best practices, I need some guidance on whether or not to baseline the current repository and how to move forward with additions of assets; tables, procs, pkgs, etc.  I'm really interested in how larger oracle shops organize their repository - what directories do you use, how are they labeled...are they labeled with respect to version?
    Assumptions:
    1. repository (database assets only) needs to be baselined (?)
    2. I have approval to change this database directory under the trunk to support best practices and get the client steered straight in terms of recovery and
    Knowns:
    1. the current application version in the database is 5.11.0 (that's my client's application version)
    2. this is for one schema/user of a database (other schemas under the database belong to different trunks)
    This is the layout that we currently have and for the privacy of the
    client I've made this rather generic.  I'd love to have a fresh
    start...how do I go about doing that...initially, I like using
    SqlDeveloper's ability to create sql scripts from a connected target. 
    product_name
      |_trunk
         |_database
           |_config
           |_data
           |_database
           |_integration
           |_patch
           |   |_5.2A.2
           |   |_5.2A.4
           |   |_5.3.0
           |   |_5.3.1
           |
           |_scripts
           |   |_config
           |   |_logs
           |
           |_server
    Thank you in advance.

    HiWe are using Data ONTAP 8.2.3p3 on our FAS8020 in 7-mode and we have 2 aggregates, a SATA and SAS aggregate. I want to decommission the SATA aggregate as I want to move that tray to another site. If I have a flexvol containing 3 qtrees CIFS shares can I use data motion (vol copy) to move the flex vol on the same controller but to a different aggregate without major downtime? I know this article is old and it says here that CIFS are not supported however I am reading mix message that on the version of data ONTAP we are now on does support CIFS and data motion however there will be a small downtime with the CIFS share terminating. Is this correct? Thanks

  • Database Administration - Best Practices

    Hello Gurus,
    I would like to know various best practices for managing and administering Oracle databases. To give you all an example what I am thinking about - for example, if you join a new company and would like to see if all the database conform to some kind of standard/best practices, what would you look for - for instance - are the control files multiplexed, are there more than one member for each redo log group, is the temp tablespace using TEMPFILE or otherwise...something of that nature.
    Do you guys have some thing in place which you use on a regular basis. If yes, I would like to get your thoughts and insights on this.
    Appreciate your time and help with this.
    Thanks
    SS

    I have a template that I use to gather preliminary information so that I can at least get a glimar of what is going on. I have posted the text below...it looks better as a spreedsheet.
    System Name               
    System Description               
         Name      Phone     Pager
    System Administrator               
    Security Administrator               
    Backup Administrator               
    Below This Line Filled Out for Each Server in The System               
    Server Name               
    Description (Application, Database, Infrastructure,..)               
    ORACLE version/patch level          CSI     
              Next Pwd Exp     
    Server Login               
    Application Schema Owner               
    SYS               
    SYSTEM               
         Location          
    ORACLE_HOME               
    ORACLE_BASE               
    Oracle User Home               
    Oracle SQL scripts               
    Oracle RMAN/backup scripts               
    Oracle BIN scripts               
    Oracle backup logs               
    Oracle audit logs               
    Oracle backup storage               
    Control File 1               
    Control File 2               
    Control File 3                    
    Archive Log Destination 1                    
    Archive Log Destination 2                    
    Datafiles Base Directory                    
    Backup Type     Day     Time     Est. Time to Comp.     Approx. Size
    archive log                    
    full backup                    
    incremental backup                    
    As for "Best" practices, well I think that you know the basics from your posting but a lot of it will also depend on the individual system and how it is integrated overall.
    Some thoughts I have for best practices:
    Backups ---
    1) Nightly if possible
    2) Tapes stored off site
    3) Archives backed up through out day
    4) To Disk then to Tape and leave backup on disk until next backup
    Datafiles ---
    1) Depending on hardware used.
    a) separate datafiles from indexes
    b) separate high I/O datafiles/indexes on dedicated disks/lungs/trays
    2) file names representative of usage (similar to its tablespace name)
    3) Keep them of reasonable size < 2 GB (again system architecture dependent)
    Security ---
    At least meet DOD - DISA standards where/when possible
    http://iase.disa.mil/stigs/stig/database-stig-v7r2.pdf
    Hope that gives you a start
    Regards
    tim

Maybe you are looking for

  • I have re-installed PS CS4 and Bridge onto my new Macbook Pro, but I cannot open or edit RAW files?

    I just bought a new Macbook, and I tried transferring my Photoshop CS4 and Bridge to my new one via my external Hard Drive. It's worked in the past (I did it when transferring my desktop to my laptop), but this time, I am unable to open or edit RAW f

  • Ipod does not end update

    Hi - Ongoing problem since loading ITunes 7. When I update the iPod after music transfers, iTunes says "Update complete", but the iPod continues to say "Updating - Do not remove" (I once let it go for two hours and it continued to say this). If I try

  • Auto populate a date field in pdf

    in the form I am working on I have a date/time field. I need it to populate the current date when the form is open. I get it to work but it has the time on it. I only need the date. when I try to change it to display just the date. It will not auto p

  • Delete emails????? HELP 3892 inbox!!

    Hi iPad 3 on ios8. I currently have over 3800 emails in my inbox. HOW do you complete a bulk delete all on the iPad? I have tried the check one-hold move-uncheck, as described elsewhere. I have also change settings so deleted emails are also deleted

  • Im locked out of my phone I've forgot my password what do i do?

    im locked out of my iPhone forgot password what do i do????