Log Groups/Log Files Oracle8i

Sorry for the basic question but I need clarification help.
The Oracle8i DBA Certification Exam Guide indicates:
1. You must have at least two Log Groups each with at least one Log File in each. For Recoverability you
must have more than one Log File in each Log Group (and obviously the same number in each Log Group).
2. LGWR writes to only one Log Group at a time. If more than one Log File is in the Log Group, LGWR will write to each at the same time. For recoverability it is advised that each Log File within a Log Group should
reside on a separate disk.
3. As the Log File fills, a Checkpoint occurs and LGWR will then begin writing to the next Log Group as
previously described.
PROBLEM:
My current DBA Instructor maintains the understanding above is incorrect. He maintains:
1. LGWR will write to the first Log File in each Log Group simultaneously (for recoverability).
2. As the first log file is filled, a Checkpoint occurs, and LGWR will write to the second (etc) Log File in
each Log Group.
3. All Log Files within one Log Group must reside on the same disk (Syntax problem if you don't).
4. Each Log Group should be placed on a different disk for recoverability.
Who is correct? Did I misunderstand the DBA Certification Exam Guide?
Thanks for any help.

Hi,
When you take an online back including logs (which is the default option for online backups in 700 systems)
this means that a backup is taken and all the logs that are written to during the time of the backup are also inlcuded in the backup. this means that you have the backup image and the logs to enable a restore and rollforward through the logs to a consistent point in time. If you took an online backup without including the logs, you would need to ensure you have those logs from the time of that backup to be able to restore that image and rollforward to ensure consisntecny.
One mjor point, never ever delete or touch active log files from /db2/NMS/log_dir as this will result in the system becoming inconsistent due to the logging mechanism being interrupted. If such a case does occur, please contact support.
Regards,
Paul

Similar Messages

  • How to delete log files in wheel group

    Let me start by saying that I've never used the terminal mode on a mac before, and I know nothing about UNIX. But here's my problem:
    My MacBook Pro, running OS 10.6, began filling up its boot disk for no apparent reason. I ran Disk Inventory X and found that a group of files had just been created in a folder called "Private" that is in the wheel group, to which I do not have access (I understand that this is a root admin group used by OSX). These files include, among others, a log file that is 20GB and was just created today. Since I can't access this folder as a normal user, how do I delete these files, and how can stop this from continuing to fill up my disk?

    Open the Console in the /Applications/Utilities/ folder and look at the various log files for an entry or group of entries which is repeated many times. When done, choose Go to Folder from the Finder's Go menu, provide /private/var/log/ as the path, and delete the files as you would any other.
    (60027)

  • Do I need to create new group for standby redo log files?

    I have 10 group of redo log files with 2 members for each group for my primary database , Do I need to create new group for standby redo log files for the standby database
    Group#     Members
    ==============
    1              2
    2              2
    3             2
    4             2
    5             2
    6             2
    7             2
    8             2
    9             2
    10           2
    If So, The following statment is correct? or nto
    ALTER DATABASE ADD STANDBY LOGFILE GROUP 1 ('D:\Databases\epprod\StandbyRedoLog\REDO01.LOG',D:\Databases\epprod\StandbyRedoLog\REDO01_1.LOG');
    please correct me if am doin mistake
    becuase when I issue the statment I getting error message sayin the group is already created.

    Thanks John
    I just find the answer
    Yes, it's recomeded to add new group , for instnace If I have 10 group from 1 to 10 then the standby shoudl be from 11 to 20
    Thanks I found the answer.

  • Dataguard Solution for standby redo log file groups

    Respected Experts,
    My database version is 10.2.0.1.0 and Red Hat 5 os.I want to create a standby database using RMAN.
    Can any one help me with the full steps.And i'm also confuse about number of standby redo log file members
    need to be created.
    Thanks and Regards
    Monoj Das

    My database version is 10.2.0.1.0 and Red Hat 5 os.I want to create a standby database using RMAN.To configure standby either you can use duplicate target database for standby
    or
    1) restore standby controlfile
    2) mount standby database
    3) restore database
    and configure standby paraemter then start MRP, will do.
    http://docs.oracle.com/cd/B19306_01/server.102/b14239/create_ps.htm
    Can any one help me with the full steps.And i'm also confuse about number of standby redo log file members
    need to be created.It depends which parameter you want to use, if you mention log_archive_dest_2='service ARCH ' then no need to create any standby redo log file groups,
    If you use log_archive_dest_2='service LGWR ' here transport will be in terms of redo and you need standby redo log files on standby database. Which is realtime.
    When you use LGWR, data lost will be less if in case of any online redo log file lost. which is recommended.
    HTH.

  • Larger redo log file members or more log groups

    Oracle 11gR1 RHEL5 64 bit
    Hi,
    I was wondering what is better from a perfomance tuning perspective. I have log swiches occuring every 2 minutes in our production database. I know definitely that our log file members are too small (100MB). The redo log sizing tool in OEM told me to make it 40G according to the fast_start_mttr_target setting which is set to 600. Now, my question is what is better to do?
    1. Increase the size of my current redo log members? Right now there are 4 groups with 2 members in each.
    OR
    2. Should I create additonal redo log groups (4 more) and then re-rerun the sizing tool or query the v$instance_recovery view?
    Which is better? tradeoffs?
    Thanks all.

    If you want to reduce the number (frequency) of Log Switches, you should increase the size of the Online Redo Logs -- ie create new Log File Groups of a larger size and drop the older ones.
    If the issue is "checkpoint not complete" waits, then either
    a. Increasing the size of the Log Files
    or
    b. Increasing the number of Log Files
    is doable
    Note that if you increase the number but not the size, you still have a checkpoint every N Mbytes -- ie, possibly too frequently !
    On the other hand if you increase your size to be very large, at every switch, the Archiver is going to kick in with a large Read + large Write operation -- reading that Redo Log of N GBytes and writing it out to the archive log target location, imposing that additional I/O spike on your system. (Writing to filesystem will go through the FileSystem Buffers so if your database SGA isn't very large and your database performance relies on hitting the FileSystem Buffer Cache to avoid Disk Reads, that performance will be impacted as a large portion of the FileSystem Buffer Cache will be taken over by the Archiver for some time).
    Hemant K Chitale
    http://hemantoracledba.blogspot.com

  • ADS MA : the XML exported file (export to log file) doesn't export Distinguished Name in the attribute member for a group

    Hello,
    I am facing a weird issue during the export of a group to a log file (xml).
    I have configured my ADLDS management agent such as the export run profile exports data into a XML Ffile:
    Everything is fine in the XML, I see my new accounts, the attributes updated for accounts but for an unknown reason the group which should contains accounts does
    not contain the DN values.
    It contains the tags <dn-value> and <dn> but <dn> is empty
    e.g:
    <delta operation="update" dn="CN=GroupX,OU=Users,DC=ZZZZ">
     <anchor encoding="base64">XDSQDQDQ</anchor>
     <dn-attr name="member" operation="add" multivalued="true">
      <dn-value>
       <dn/>
      </dn-value>
      <dn-value>
       <dn/>
      </dn-value>
     </dn-attr>
    During the export, FIM updates the attribute "member" of the group:
    Member attribute seems to be caught by FIM during synchro profile and export profil bt not translates correctly in the final xml file.
    Any ideas?
    Thanks for your reply.

    Thinking the same thing as David - sounds like a bug - but that's curious because I've never had a problem with the AD MA doing exactly the same thing, albeit with FIM R1 most recently.  What version of FIM are you using, and have you checked the
    release notes of any subsequent versions to see if any such issue is mentioned?
    Bob Bradley (FIMBob @
    TheFIMTeam.com) ... now using FIM Event Broker for just-in-time delivery of FIM 2010 policy via the sync engine, and continuous compliance for FIM

  • AlwaysOn Availability groups and unable to backup LOG files

    Hi,
    I have an issue where we are using AlwaysOn availability groups, however my transaction log backup jobs are not working for any databases that are part of the groups.  Databases which are standalone backup fine.  I've tried using the builtin maintenance
    plans as well as the 'maintenance script' from Ola Hallengren which is my preference.
    All the jobs are reporting success even though nothing is bring written to disk, the jobs complete within a few seconds.
    If I backup the log files individually myself they work fine.  After
    going a bit of Googling this appears to have been a known issue that was fixed in CU7, unfortunately I'm running CU9 and still have the issue.
    I've tried changing the backup preference to any, as well as primary or secondary only to no avail.
    I'm using SQL Server SP1 CU9, 2 nodes in the group.  Anyone have any suggestions?
    Marcus

    Hi,
    I understand that you have tried every options in Backup Preferences. I still suggest you enable the “For availability database. Ignore Replica Priority for Backup and Backup on Primary Settings” when you define the maintenance plan and check how it works.
    This settings is off by default. And the maintenance plan will detect the availability group's AUTOMATED_BACKUP_PREFERENCE setting when deciding to backup a database or log if that database is defined in the group.
    'Given that an availability group's default AUTOMATED_BACKUP_PREFERENCE setting is 'SECONDARY, a maintenance plan, defined and running on a SQL Server instance hosting the primary replica, will NOT backup databases defined in the availability group.
    Reference:
    http://blogs.msdn.com/b/alwaysonpro/archive/2014/01/02/maintenance-plan-does-not-backup-database-or-log-of-database-that-belongs-to-availability-group.aspx
    Hope it helps.
    Tracy Cai
    TechNet Community Support

  • How do we track client deployment via group policy by referring log files globally

    How do we track client deployment via group policy by referring log file centrally?

    need answer from  both CM07/CM012 by using GPO
    There is NO Centralized tracking for GPOs.
    Garth Jones | My blogs: Enhansoft and
    Old Blog site | Twitter:
    @GarthMJ

  • How to Properly Protect a Virtualized Exchange Server - Log File Discontinuity When Performing Child Partition Snapshot

    I'm having problems backing up a Hyper-V virtualized Exchange 2007 server with DPM 2012. The guest has one VHD for the OS, and two pass-through volumes, one for logs and one for the databases. I have three protection groups:
    System State - protects only the system state of the mail server, runs at 4AM every morning
    Exchange Databases - protects the Exchange stores, 15 minute syncs with an express full at 6:30PM every day
    VM - Protecting the server hosting the Exchange VM. Does an child partition snapshot backup of the Exchange server guest with an express full at 9:30PM every day
    The problem I'm experiencing is that every time the VM express full completes I start receiving errors on the Exchange Database synchronizations stating that a log file discontinuity was detected. I did some poking around in the logs on the Exchange server
    and sure enough, it looks like the child partition snapshot backup is causing Exchange to truncate the log files even though the logs and databases are on pass-through disks and aren't covered by the child partition snapshot.
    What is the correct way to back up an entire virtualized Exchange server, system state, databases, OS drive and all?

    I just created a new protection group. I added "Backup Using Child Partition Snapshot\MailServer", short-term protection using disk, and automatically create the replica over the network immediately. This new protection group contains only the child partition
    snapshot backup. No Exchange backups of any kind.
    The replica creation begins. Soon after, the following events show up in the Application log:
    =================================
    Log Name:      Application
    Source:        MSExchangeIS
    Date:          10/23/2012 10:41:53 AM
    Event ID:      9818
    Task Category: Exchange VSS Writer
    Level:         Information
    Keywords:      Classic
    User:          N/A
    Computer:      PLYMAIL.mcquay.com
    Description:
    Exchange VSS Writer (instance 7d26282d-5dec-4a73-bf1c-f55d5c1d1ac7) has been called for "CVssIExchWriter::OnPrepareSnapshot".
    =================================
    Log Name:      Application
    Source:        ESE
    Date:          10/23/2012 10:41:53 AM
    Event ID:      2005
    Task Category: ShadowCopy
    Level:         Information
    Keywords:      Classic
    User:          N/A
    Computer:      PLYMAIL.mcquay.com
    Description:
    Information Store (3572) Shadow copy instance 2051 starting. This will be a Full shadow copy.
    =================================
    The events continue on, basically snapshotting all of Exchange. From the DPM side, the total amount of data transferred tells me that even though Exhange is trunctating its logs, nothing is actually being sent to the DPM server. So this snapshot operation
    seems to be superfluous. ~30 minutes later, when my regularly scheduled Exchange job runs, it fails because of a log file discontinuity.
    So, in this case at least, a Hyper-V snapshot backup is definitely causing Exchange to truncate the log files. What can I look at to figure out why this is happening?

  • Error in log file (WWC-41439)

    Hi,
    When i installed Oracle Portal, i had the error WWc-41439 when i tried log in Oracle Portal. I revised the log file and i saw the following error.(I have isntalled Oracle Portal 3.0., Oracle Database 8.1.7 on windows NT).
    STEP 24 : Installing SSO packages to public
    INSTALL_ACTION : installSSOLayer()..\..\bin\sqlplus portal30_sso/portal30_sso@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoinsg.sql
    STEP 25 : Associating Login Server & Oracle Portal
    INSTALL_ACTION :assocNewLoginServer: Portal Url Prefix: http://pablo/pls/portal30/
    INSTALL_ACTION :assocNewLoginServer: SSO URL Prefix: http://pablo/pls/portal30_sso/
    INSTALL_ACTION : assocNewLoginServer: ..\..\bin\sqlplus portal30_sso/portal30_sso@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoseedl.sql portal30 http://pablo/pls/portal30/ portal30_sso http://pablo/pls/portal30_sso/ NO
    INSTALL_ACTION : assocNewLoginServer: ..\..\bin\sqlplus portal30_sso/portal30_sso@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoinsgp.sql portal30
    INSTALL_ACTION : assocNewLoginServer:..\..\bin\sqlplus portal30/portal30@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoseedw.sql portal30 portal30_sso http://pablo/pls/portal30_sso/ http://pablo/pls/portal30/
    INSTALL_ACTION : STEP 19 : assocNewLoginServer:..\..\bin\sqlplus portal30/portal30@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\wwc\secpsr.sql
    INSTALL_ACTION : STEP 19 : assocNewLoginServer:..\..\bin\sqlplus portal30/portal30@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\wwc\secps.sql portal30_sso_PS N N N N
    INSTALL_ACTION : assocNewLoginServer:..\..\bin\sqlplus portal30_sso/portal30_sso@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoinsrp.sql portal30
    INSTALL_ACTION : assocNewLoginServer:..\..\bin\sqlplus portal30_sso/portal30_sso@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\sso\ssoseedu.sql portal30
    Installing and running diagnostics
    INSTALL_ACTION:installDiagnostics() : ..\..\bin\loadjava -resolve -verbose -thin -user portal30/portal30@localhost:1521:or8i ..\..\portal30\admin\plsql\wwc\Diagnose.class
    C:\ORACLE\iSuites\assistants\opca>REM
    C:\ORACLE\iSuites\assistants\opca>REM $Header: runljava.bat@@/main/3 \
    C:\ORACLE\iSuites\assistants\opca>REM Checked in on Fri Nov 17 15:32:36 PST 2000 by meoropez \
    C:\ORACLE\iSuites\assistants\opca>REM Copyright (c) 2000 by Oracle Corporation. All Rights Reserved. \
    C:\ORACLE\iSuites\assistants\opca>REM $
    C:\ORACLE\iSuites\assistants\opca>REM
    C:\ORACLE\iSuites\assistants\opca>Rem For running the Loadjava from the Configuration Assistant.
    C:\ORACLE\iSuites\assistants\opca>..\..\bin\loadjava -resolve -verbose -thin -user portal30/portal30@localhost:1521:or8i ..\..\portal30\admin\plsql\wwc\Diagnose.class
    INSTALL_ACTION : Running Diagnostics ..\..\bin\sqlplus portal30/portal30@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=localhost)(PORT=1521)))(CONNECT_DATA=(SID=or8i))) @..\..\portal30\admin\plsql\wwc\secdiag.sql
    SQL*Plus: Release 8.1.7.0.0 - Production on Dom Abr 20 20:23:37 2003
    (c) Copyright 2000 Oracle Corporation. All rights reserved.
    Conectado a:
    Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    Creating Table 'wwsec_diagnostic$'
    Creating Sequence 'wwsec_diagnostic_seq'
    Diagnostics Report v 1.01: Oracle Portal v 3.0.9.8.0
    As of 20-Abr-2003 20:23:41 Schema Name: PORTAL30 SSO Schema Name: portal30_sso
    Proxy Server Settings:
    HTTP Server:
    HTTP Server Port:
    No Proxy Servers for Domains beginning with:
    URL Connection Time-Out (seconds):
    PORTAL30.wwsec_enabler_config_info$
    Login Server URL : http://pablo/pls/portal30_sso/portal30_sso.wwsso_app_admin.ls_login
    DAD          : portal30_sso
    Host connection : *** FAILED ***
    Unable to find the Schema Name for the Login Server
    Recommendations:
    Please check the DAD settings for the Login Server
    Desconectado de Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    INSTALL_ACTION : End of Installation.
    How i resolve the problem?
    Is there somebody in the same situacion?
    Thank you

    Did you promoted/demoted a server? ( like from consumer to master , master to consumer , hub to consumer/master OR disabled replication and re-enabled with another unique ID and recreated replication agreements again?
    IF YES, find out the current Unique ID in all the related server and delete the old serverIDs in the nsslapd-rererral.
    Also the error will tell you for which server it is trying to update the duplicate entry. Mostly , that entry will be the old value as I explained before.
    I have seen this before and mostly by doing above it gets corrected. I recommend that you should think before doing it.
    In any case dont forget to update us since this is a tricky situation to me.
    -Kunal

  • SQLException Error in Log file

    Does anyone know of the following error, and how to fix this? I keep getting it in my engine log file like every 30 mn.
    Thank you,
    -Lilach
    --------------------------------------------- Error -----------------------------------------------------------------
    Exception [java.sql.SQLException: [BEA][SQLServer JDBC Driver][SQLServer]String or binary data would be truncated.].
    Caused by: [BEA][SQLServer JDBC Driver][SQLServer]String or binary data would be truncated.
    Caused by: [BEA][SQLServer JDBC Driver][SQLServer]The statement has been terminated.
    fuego.directory.DirectoryRuntimeException: Exception [java.sql.SQLException: [BEA][SQLServer JDBC Driver][SQLServer]String or binary data would be truncated.].
         at fuego.directory.DirectoryRuntimeException.wrapException(DirectoryRuntimeException.java:85)
         at fuego.directory.provider.jdbc.mssqlserver.MSSQLServerPersistenceManager.mapSQLException(MSSQLServerPersistenceManager.java:116)
         at fuego.directory.provider.jdbc.JDBCServiceAccessor.mapSQLException(JDBCServiceAccessor.java:78)
         at fuego.directory.provider.jdbc.JDBCParticipantsAccessor.createHumanParticipant(JDBCParticipantsAccessor.java:221)
         at fuego.directory.provider.jdbc.JDBCParticipantsAccessor.createHumanParticipant(JDBCParticipantsAccessor.java:180)
         at fuego.directory.hybrid.jdbc.DelegatorFuegoParticipantAccessor.updateOrPersistHumanParticipant(DelegatorFuegoParticipantAccessor.java:224)
         at fuego.directory.hybrid.HybridParticipantAccessor.updateHumanParticipant(HybridParticipantAccessor.java:377)
         at sun.reflect.GeneratedMethodAccessor38.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at fuego.directory.provider.DirectorySessionImpl$AccessorProxy.invoke(DirectorySessionImpl.java:756)
         at $Proxy6.updateHumanParticipant(Unknown Source)
         at fuego.directory.hybrid.HybridParticipantAccessor.fetchHumanParticipant(HybridParticipantAccessor.java:189)
         at sun.reflect.GeneratedMethodAccessor29.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at fuego.directory.provider.DirectorySessionImpl$AccessorProxy.invoke(DirectorySessionImpl.java:756)
         at $Proxy6.fetchHumanParticipant(Unknown Source)
         at fuego.directory.provider.notifiers.DirObjectType$3.fetchDirObjectFor(DirObjectType.java:74)
         at fuego.directory.provider.notifiers.DirectoryObjectProxy.getObject(DirectoryObjectProxy.java:160)
         at fuego.directory.provider.notifiers.DirectoryObjectProxy.createAddEvent(DirectoryObjectProxy.java:91)
         at fuego.directory.provider.notifiers.BasePollingEventGenerator.buildEvents(BasePollingEventGenerator.java:148)
         at fuego.directory.hybrid.ldap.LDAPPollingEventGenerator.buildEvents(LDAPPollingEventGenerator.java:109)
         at fuego.directory.provider.notifiers.BasePollingEventGenerator.generateEvents(BasePollingEventGenerator.java:51)
         at fuego.directory.hybrid.HybridMultipleEventGenerator.generateEvents(HybridMultipleEventGenerator.java:43)
         at fuego.directory.provider.notifiers.DirectoryNotifier.notifyChanges(DirectoryNotifier.java:403)
         at fuego.server.service.DirectoryListener.updateEngineFromDirectoryImpl(DirectoryListener.java:328)
         at fuego.server.service.DirectoryListener$DirectoryPollingItem.execute(DirectoryListener.java:381)
         at fuego.server.execution.DefaultEngineExecution$AtomicExecutionTA.runTransaction(DefaultEngineExecution.java:304)
         at fuego.transaction.TransactionAction.startNestedTransaction(TransactionAction.java:527)
         at fuego.transaction.TransactionAction.startTransaction(TransactionAction.java:548)
         at fuego.transaction.TransactionAction.start(TransactionAction.java:212)
         at fuego.server.execution.DefaultEngineExecution.executeImmediate(DefaultEngineExecution.java:123)
         at fuego.fengine.service.FEngineExecution.executeImmediate(FEngineExecution.java:65)
         at fuego.server.execution.DefaultEngineExecution.executeWithoutComponentImmediate(DefaultEngineExecution.java:185)
         at fuego.server.execution.EngineExecution.executeWithoutComponentImmediate(EngineExecution.java:86)
         at fuego.server.execution.PollingDirectory.execute(PollingDirectory.java:46)
         at fuego.server.execution.DefaultEngineExecution$AtomicExecutionTA.runTransaction(DefaultEngineExecution.java:304)
         at fuego.transaction.TransactionAction.startBaseTransaction(TransactionAction.java:470)
         at fuego.transaction.TransactionAction.startTransaction(TransactionAction.java:551)
         at fuego.transaction.TransactionAction.start(TransactionAction.java:212)
         at fuego.server.execution.DefaultEngineExecution.executeImmediate(DefaultEngineExecution.java:123)
         at fuego.server.execution.DefaultEngineExecution.executeAutomaticWork(DefaultEngineExecution.java:62)
         at fuego.server.execution.EngineExecution.executeAutomaticWork(EngineExecution.java:42)
         at fuego.server.execution.ToDoItem.executeAutomaticWork(ToDoItem.java:251)
         at fuego.server.execution.ToDoItem.run(ToDoItem.java:536)
         at fuego.component.ExecutionThread.processMessage(ExecutionThread.java:775)
         at fuego.component.ExecutionThread.processBatch(ExecutionThread.java:755)
         at fuego.component.ExecutionThread.doProcessBatch(ExecutionThread.java:142)
         at fuego.component.ExecutionThread.doProcessBatch(ExecutionThread.java:134)
         at fuego.fengine.ToDoQueueThread$PrincipalWrapper.processBatch(ToDoQueueThread.java:446)
         at fuego.component.ExecutionThread.work(ExecutionThread.java:839)
         at fuego.component.ExecutionThread.run(ExecutionThread.java:408)
    Caused by: java.sql.SQLException: [BEA][SQLServer JDBC Driver][SQLServer]String or binary data would be truncated.
         at albpm.jdbc.base.BaseExceptions.createException(Unknown Source)
         at albpm.jdbc.base.BaseExceptions.getException(Unknown Source)
         at albpm.jdbc.sqlserver.tds.TDSRequest.processErrorToken(Unknown Source)
         at albpm.jdbc.sqlserver.tds.TDSRequest.processReplyToken(Unknown Source)
         at albpm.jdbc.sqlserver.tds.TDSRPCRequest.processReplyToken(Unknown Source)
         at albpm.jdbc.sqlserver.tds.TDSRequest.processReply(Unknown Source)
         at albpm.jdbc.sqlserver.SQLServerImplStatement.getNextResultType(Unknown Source)
         at albpm.jdbc.base.BaseStatement.commonTransitionToState(Unknown Source)
         at albpm.jdbc.base.BaseStatement.postImplExecute(Unknown Source)
         at albpm.jdbc.base.BasePreparedStatement.postImplExecute(Unknown Source)
         at albpm.jdbc.base.BaseStatement.commonExecute(Unknown Source)
         at albpm.jdbc.base.BaseStatement.executeUpdateInternal(Unknown Source)
         at albpm.jdbc.base.BasePreparedStatement.executeUpdate(Unknown Source)
         at fuego.jdbc.FaultTolerantPreparedStatement.executeUpdate(FaultTolerantPreparedStatement.java:623)
         at fuego.directory.provider.jdbc.JDBCParticipantsAccessor.createParticipantInfo(JDBCParticipantsAccessor.java:1280)
         at fuego.directory.provider.jdbc.JDBCParticipantsAccessor.createHumanParticipantInternal(JDBCParticipantsAccessor.java:1195)
         at fuego.directory.provider.jdbc.JDBCParticipantsAccessor.createHumanParticipant(JDBCParticipantsAccessor.java:217)
         ... 49 more

    Hi All ,
    I having this error . Are there any fixes for this bug ?
    my configuration file is like the following ;
    +<?xml version="1.0" encodxxx="UTF-8"?>+
    +<?fuego version="6.1 ALPHA" application="albpmenterprise"?>+
    +<!-- This file contains the propper attribute mappxxx for the FDI Generic Ldap Provider usxxx Oracle Internet Directory.+
    +     * Preference for group object+
    +          <preference id="assignedParticipants.containsId" value="true"/>+
    +          This preference is useful to speed up the provider and it can only be used if the assignedParticipant value is the dn of the user and the dn contains the participant id+
    +          <preference id="assignedParticipants.containsId" value="true"/>+
    +          This preference is useful to speed up the provider and it can only be used if the assignedGroup value is the dn of the group and the dn contains the group id+
    +          <preference id="modifyTimeStamp.suffix" value="Z"/>+
    +          This preference is useful when the suffix mofidyTimeStamp format of your ldap is not .OZ.+
    -->
    +<config>+
    +     <object id="person">+
    +          <object-filter>+
    +               <![CDATA[+
    +                    (objectclass=person)+
    +               ]]>+
    +          </object-filter>+
    +          <relative-dn>ou=xxxBANK,cn=Users</relative-dn>+
    +          <attribute id="id" value="uid"/>+
    +          <attribute id="lastName" value="sn"/>+
    +          <attribute id="firstName" value="givenname"/>+
    +          <attribute id="displayName" value="displayname"/>+
    +          <attribute id="mail" value="mail"/>+
    +          <attribute id="telephoneNumber" value="telephonenumber"/>+
    +          <attribute id="employeeId" value="employeeNumber"/>+
    +          <attribute id="thumbnailPhoto" value="thumbnailPhoto"/>+
    +          <attribute id="manager" value="manager"/>+
    +          <attribute id="modifyTimeStamp" value="modifyTimestamp"/>+
    +     </object>+
    +     <object id="group">+
    +          <object-filter>+
    +               <![CDATA[+
    +                    (objectclass=xxxGroupObject)+
    +               ]]>+
    +          </object-filter>+
    +          <relative-dn>+
    +               <!-- the relative dn for group --></relative-dn>+
    +          <attribute id="id" value="cn"/>+
    +          <attribute id="modifyTimeStamp" value="modifyTimestamp"/>+
    +          <attribute id="displayName" value="cn"/>+
    +          <attribute id="name" value="cn"/>+
    +          <attribute id="description" value="description"/>+
    +          <attribute id="assignedParticipants" value="uniquemember"/>+
    +          <attribute id="assignedGroups" value="uniquemember"/>+
    +          <attribute id="ou" value="distxxxuishedName"/>+
    +     </object>+
    +     <object id="ou">+
    +          <object-filter>+
    +               <![CDATA[+
    +                    (objectclass=xxxOUObject)+
    +               ]]>+
    +          </object-filter>+
    +          <relative-dn>+
    +               <!-- the relative dn for ous -->+
    +          </relative-dn>+
    +          <attribute id="name" value="ou"/>+
    +          <attribute id="description" value="description"/>+
    +     </object>+
    +</config>+

  • How to disable write to redo log file in oracle7.3.4

    in oracle 8, alter table no logged in redo log file like: alter table tablename nologging;
    how to do this in oracle 7.3.4?
    thanks.

    user652965 wrote:
    Thanks very much for your help guys. I appreciate it. unfortunately none of these commands worked for me. I kept getting error on clearing logs that redo log is needed to perform recovery so it can't be cleared. So I ended up restoring from earlier snapshot of my db volume. Database is now open.
    Thanks again for your input.And now, as a follow-up, at a minimum you should make sure that all redo log groups have at least 3 members. Then, if you lose a single redo log file, all you have to do is shutdown the db and copy one of the good members (of the same group as the lost member) over the lost member.
    And as an additional follow-up, if you value your data you will run in archivelog mode and take regular backups of the database and archivelogs. If you fail to do this you are saying that your data is not worth saving.

  • Collecting Log Files

    When providing feedback around reported issues we may request you access and share certain log files from your computer in order to better understand the operation of the Creative Cloud app on your system.
    Below is information about various log files, their location, and when they are created.
    Note: (Windows) If unable to locate a specific folder it may be necessary to ensure the option to show hidden folders is turned on
    http://helpx.adobe.com/x-productkb/global/show-hidden-files-folders-extensions.html
    Note: (Mac) The Users Library user folder is hidden on starting with 10.7 or later
    see http://helpx.adobe.com/x-productkb/global/access-hidden-user-library-files.html
    ~ on Mac denotes the logged in user's directory.  The Library folder under it is hidden. The easiest way to access it is to use the "Go to folder" option from the "Go" menu of Finder and type the path exactly as I typed starting from ~ symbol.)
    Please always include these logs - PDApp.log and AMT3.log from this location:
        Win7/8 : %TEMP%
        Mac : ~/Library/Logs/
    For issues with installation - include these log files:
        Win: <drive>:\Program Files\Common Files\Adobe\Installers
        Mac: /Library/Logs/Adobe/Installers
    For issues with updates
        Win7/8: <%localappdata%>\Adobe\AAMUpdater\1.0
        Mac: ~/Library/Application Support/Adobe/AAMUpdater/1.0
    Please zip this complete 1.0 folder to share
    For issues with activation or licensing
    Please zip the entire SLCache folder and all files in it, into a single zip, and attach to bug)
        Windows 32 bit: <drive>:\Program Files\Common Files\Adobe
        Windows 64 bit: <drive>:\Program Files (x86)\Common Files\Adobe
        Mac: ~/Library/Application Support/Adobe
    For issues with Adobe Application Manager and downloading
        Windows: %temp%[AdobeDownloads]\
        Mac: ~/Library/Logs/Adobe/AdobeDownloads/
    For issues related to file sync and Typekit
    Mac:
    The log file can be found here:
    <Mac Hard Drive>/Users/<username>/Library/Application Support/Adobe/CloudSync/CoreSync-YYYY-MM-DD.log
    (where YYYY-MM-DD indicate the date of the last log)
    Please see note at the top about for accessing the user library folder
    Windows:
    The log file can be found here:
    C:\Users\<username>\AppData\Roaming\CloudSync\CoreSync-YYYY-MM-DD.log
    (where YYYY-MM-DD indicate the date of the last log)

    I am able to run the commands individually from oracle user, but when I run "raccheck" script its not working.
    oracle & grid user id:_
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ id oracle
    uid=501(oracle) gid=501(oinstall) groups=501(oinstall),502(dba),504(asmdba),506(oper)
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ id grid
    uid=502(grid) gid=501(oinstall) groups=501(oinstall),503(asmadmin),504(asmdba),505(asmoper)
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$I am able to run crsctl command or any other GI command from oracle user. But "raccheck" script is not working.
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ echo $CRS_HOME
    /app/grid/product/11.2.0.3
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ $CRS_HOME/bin/crsctl query crs activeversion
    Oracle Clusterware active version on the cluster is [11.2.0.3.0]
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ $CRS_HOME/bin/crsctl query crs softwareversion
    Oracle Clusterware version on node [hublpr1] is [11.2.0.3.0]
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ $CRS_HOME/bin/crsctl check crs
    CRS-4638: Oracle High Availability Services is online
    CRS-4537: Cluster Ready Services is online
    CRS-4529: Cluster Synchronization Services is online
    CRS-4533: Event Manager is online
    oracle@hublpr1:/admintmp/ORACLE/RAC-Check$ cat /proc/cpuinfo
    processor       : 0
    vendor_id       : GenuineIntel
    cpu family      : 6
    model           : 44
    ...................................................

  • Problem in creating database -Missing Redo log file

    I am try to create a new database using DBCA .While creating a database it shows the error oracle instance terminated.Force Disconnected.
    My alert log file is
    Errors in file /u01/app/oracle/diag/rdbms/oracl/oracl/trace/oracl_ora_5424.trc:
    ORA-00313: open failed for members of log group 1 of thread 1
    ORA-00312: online log 1 thread 1: '/u01/app/oracle/oradata/oracl/redo01.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Wed Nov 06 10:07:27 2013
    Errors in file /u01/app/oracle/diag/rdbms/oracl/oracl/trace/oracl_ora_5424.trc:
    ORA-00313: open failed for members of log group 2 of thread 1
    ORA-00312: online log 2 thread 1: '/u01/app/oracle/oradata/oracl/redo02.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Errors in file /u01/app/oracle/diag/rdbms/oracl/oracl/trace/oracl_ora_5424.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/u01/app/oracle/oradata/oracl/redo03.log'
    ORA-27037: unable to obtain file status
    Linux-x86_64 Error: 2: No such file or directory
    Additional information: 3
    Wed Nov 06 10:07:38 2013
    Setting recovery target incarnation to 2
    Wed Nov 06 10:07:38 2013
    Assigning activation ID 1876274518 (0x6fd5ad56)
    Thread 1 opened at log sequence 1
      Current log# 1 seq# 1 mem# 0: /u01/app/oracle/oradata/oracl/redo01.log
    Successful open of redo thread 1
    Wed Nov 06 10:07:38 2013
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Nov 06 10:07:38 2013
    SMON: enabling cache recovery
    Errors in file /u01/app/oracle/diag/rdbms/oracl/oracl/trace/oracl_ora_5424.trc  (incident=1345):
    ORA-00600: internal error code, arguments: [kpotcgah-7], [12534], [ORA-12534: TNS:operation not supported
    Incident details in: /u01/app/oracle/diag/rdbms/oracl/oracl/incident/incdir_1345/oracl_ora_5424_i1345.trc
    Wed Nov 06 10:07:46 2013
    Trace dumping is performing id=[cdmp_20131106100746]
    Errors in file /u01/app/oracle/diag/rdbms/oracl/oracl/trace/oracl_ora_5424.trc:
    ORA-00600: internal error code, arguments: [kpotcgah-7], [12534], [ORA-12534: TNS:operation not supported
    Error 600 happened during db open, shutting down database
    USER (ospid: 5424): terminating the instance due to error 600
    Instance terminated by USER, pid = 5424
    ORA-1092 signalled during: alter database "oracl" open resetlogs...
    ORA-1092 : opiodr aborting process unknown ospid (5424_47935551851664)
    Wed Nov 06 10:07:47 2013
    ORA-1092 : opitsk aborting process
                                                                                                                                   251,1         95%

    >I am try to create a new database using DBCA
    >Please help me to resolve this issue.My redo log file was missing
    DROP and recreate the database.  It is a *new* database without any data.
    Check what datafile locations and redo log file locations you specify when creating the new database. Check if you have permissions and enough disk space.
    Hemant K Chitale

  • Not able to add new log file to the 11g database.

    Hi DBA's
    I am not able to add the log file i am getting error while adding the database.
    SQL> alter database add logfile group 3 ('/oracle/DEV/db/apps_st/data/log03a.dbf','/oracle/DEV/db/apps_st/data/log03a.dbf') size 50m reuse;
    alter database add logfile group 3 ('/oracle/DEV/db/apps_st/data/log03a.dbf','/oracle/DEV/db/apps_st/data/log03a.dbf') size 50m reuse
    ERROR at line 1:
    ORA-01505: error in adding log files
    ORA-01577: cannot add log file '/oracle/DEV/db/apps_st/data/log03a.dbf' - file
    already part of database
    SQL> select a.group#, member, a.status from v$log a, v$logfile b where a.group# = b.group# order by 1;
    GROUP# MEMBER STATUS
    1 /oracle/DEV/db/apps_st/data/log01a.dbf ACTIVE
    1 /oracle/DEV/db/apps_st/data/log01b.dbf ACTIVE
    2 /oracle/DEV/db/apps_st/data/log02a.dbf CURRENT
    2 /oracle/DEV/db/apps_st/data/log02b.dbf CURRENT
    Kindly help me to add the new log file to my database.
    Thanks,
    SG

    Hi Sawwan,
    V$LOGMEMBER was written in the document,
    I query the log members as bellow
    1)select a.group#, member, a.status from v$log a, v$logfile b where a.group# = b.group# order by 1;
    GROUP# MEMBER STATUS
    1 /oracle/DEV/db/apps_st/data/log01a.dbf INACTIVE
    1 /oracle/DEV/db/apps_st/data/log01b.dbf INACTIVE
    2 /oracle/DEV/db/apps_st/data/log02a.dbf CURRENT
    2 /oracle/DEV/db/apps_st/data/log02b.dbf CURRENT
    2)SQL> select group#,member,status from v$logfile;
    GROUP# MEMBER STATUS
    2 /oracle/DEV/db/apps_st/data/log02a.dbf
    2 /oracle/DEV/db/apps_st/data/log02b.dbf
    1 /oracle/DEV/db/apps_st/data/log01a.dbf
    1 /oracle/DEV/db/apps_st/data/log01b.dbf
    But i am littile bit confused that there is no group or datafile called " Group 3 and log03a.dbf" as per the above query, how can i drop tease group and datafile.
    and i crossverified in the data top the files are exist or not but those are not existing. but still i am getting the same error that i can't create that already exist.
    can issue the bellow queris to drop those group which i dont think so it will exist?
    SQL>alter database drop logfile group 3;
    Thanks in advance.
    Regards,
    SG

Maybe you are looking for

  • Java ME 8 Permission check failed when opening a serial port

    I have a larger Jave ME8.1 application that was going well until I tried to add one last piece, reading and writing data from a serial port. This was left to last because it is trivial, at least in most programming languages. The is IDE NetBeans 8.0.

  • PDF printing problems with Canon Color C3220

    I'm having an issue with our Canon Color Imagerunner C3220. When I send a PDF from Acrobat Professional 7.0 from a Mac to the Canon I get a rule that overprints an image. The image should be knocking out the rule. When I print from a Windows machine

  • NEED HELP. Created Website Template in Photoshop CS5, but cant center it into Dreamweaver CS6

    I created a website layout template in Photoshop cs5 once i opened it into Dreamweaver i was trying to Align it to Center but it would not go center still stays to the left. I know that something is not right i am missing something and i am new to th

  • About voided payment accounting line

    I voided a payment before submitting Payables Accounting Process or creating accounting lines. Then I submitted request of Payables Accounting Process. And I found there's no accounting lines created. The version is 11.5.6 . Could anybody tell me thi

  • HT4859 icloud, do i need to back up my computers if i have icloud?

    do i have to back up itunes if i have icloud? will also be using icloud for photos and video, right now i have norton on-line storage and would like to take itunes off of it because running out of online storage and don't want to buy more from norton