Active Data Guard-How to configure Auxiliary Data Base's Listener.

HI,
I am stuck in creating listener/tnsnames for auxiliary database as the problem happens because of the auxiliary instance has just started. It is now in NOMOUNT stage. The instance registration with the listener is performed by PMON process and to start the PMON process database need to be in mount stage. So, before the instance registration by PMON with the listener there is actually noting to register and hence the instance is BLOCKED.
http://www.oracle.com/technology/deploy/availability/pdf/oracle-openworld-2009/adg_hol_2009.pdf
I am following the above document.
LSNRCTL> status
Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
STATUS of the LISTENER
Alias LISTENER
Version TNSLSNR for Linux: Version 11.2.0.1.0 - Production
Start Date 05-JUL-2010 11:06:23
Uptime 0 days 0 hr. 11 min. 44 sec
Trace Level off
Security ON: Local OS Authentication
SNMP OFF
Listener Parameter File /u01/app/11.2.0/grid/network/admin/listener.ora
Listener Log File /u02/app/oracle/diag/tnslsnr/oracle/listener/alert/log.xml
Listening Endpoints Summary...
(DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(KEY=LISTENER)))
(DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=192.168.127.133)(PORT=1521)))
(DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=192.168.127.140)(PORT=1521)))
Services Summary...
Service "+ASM" has 1 instance(s).
Instance "+ASM1", status READY, has 1 handler(s) for this service...
Service "eagledb" has 1 instance(s).
Instance "eagledb1", status READY, has 1 handler(s) for this service...
Service "eagledbXDB" has 1 instance(s).
Instance "eagledb1", status READY, has 1 handler(s) for this service...
Service "eagledb_DGB" has 1 instance(s).
Instance "eagledb1", status READY, has 1 handler(s) for this service...
Service "nyc" has 1 instance(s).
Instance "nyc", status BLOCKED, has 1 handler(s) for this service...
The command completed successfully
I found some simillar solution on below link but it is aplicable for 10g or below.
http://arjudba.blogspot.com/2008/05/connection-to-auxilary-instance-failed.html
I am using enterprise linux with clusterware 11g rel 2 with data base 11f rel 2 on single machine.
Thanks
Anis
99826412
Muscat
Edited by: user12979506 on Jul 5, 2010 10:13 PM

The "auxiliary" is, I presume, actually a physical standby.
If so why is it in NOMOUNT? Why don't you start it?
alter database recover managed standby database cancel;
alter database open read only;

Similar Messages

  • [Forum FAQ] How to configure a Data Driven Subscription which get multi-value parameters from one column of a database table?

    Introduction
    In SQL Server Reporting Services, we can define a mapping between the fields that are returned in the query to specific delivery options and to report parameters in a data-driven subscription.
    For a report with a parameter (such as YEAR) that allow multiple values, when creating a data-driven subscription, how can we pass a record like below to show correct data (data for year 2012, 2013 and 2014).
    EmailAddress                             Parameter                      
    Comment
    [email protected]              2012,2013,2014               NULL
    In this article, I will demonstrate how to configure a Data Driven Subscription which get multi-value parameters from one column of a database table
    Workaround
    Generally, if we pass the “Parameter” column to report directly in the step 5 when creating data-driven subscription.
    The value “2012,2013,2014” will be regarded as a single value, Reporting Services will use “2012,2013,2014” to filter data. However, there are no any records that YEAR filed equal to “2012,2013,2014”, and we will get an error when the subscription executed
    on the log. (C:\Program Files\Microsoft SQL Server\MSRS10_50.MSSQLSERVER\Reporting Services\LogFiles)
    Microsoft.ReportingServices.Diagnostics.Utilities.InvalidReportParameterException: Default value or value provided for the report parameter 'Name' is not a valid value.
    This means that there is no such a value on parameter’s available value list, this is an invalid parameter value. If we change the parameter records like below.
    EmailAddress                        Parameter             Comment
    [email protected]         2012                     NULL
    [email protected]         2013                     NULL
    [email protected]         2014                     NULL
    In this case, Reporting Services will generate 3 reports for one data-driven subscription. Each report for only one year which cannot fit the requirement obviously.
    Currently, there is no a solution to solve this issue. The workaround for it is that create two report, one is used for view report for end users, another one is used for create data-driven subscription.
    On the report that used create data-driven subscription, uncheck “Allow multiple values” option for the parameter, do not specify and available values and default values for this parameter. Then change the Filter
    From
    Expression:[ParameterName]
    Operator   :In
    Value         :[@ParameterName]
    To
    Expression:[ParameterName]
    Operator   :In
    Value         :Split(Parameters!ParameterName.Value,",")
    In this case, we can specify a value like "2012,2013,2014" from database to the data-driven subscription.
    Applies to
    Microsoft SQL Server 2005
    Microsoft SQL Server 2008
    Microsoft SQL Server 2008 R2
    Microsoft SQL Server 2012
    Please click to vote if the post helps you. This can be beneficial to other community members reading the thread.

    For every Auftrag, there are multiple Position entries.
    Rest of the blocks don't seems to have any relation.
    So you can check this code to see how internal table lt_str is built whose first 3 fields have data contained in Auftrag, and next 3 fields have Position data. The structure is flat, assuming that every Position record is related to preceding Auftrag.
    Try out this snippet.
    DATA lt_data TYPE TABLE OF string.
    DATA lv_data TYPE string.
    CALL METHOD cl_gui_frontend_services=>gui_upload
      EXPORTING
        filename = 'C:\temp\test.txt'
      CHANGING
        data_tab = lt_data
      EXCEPTIONS
        OTHERS   = 19.
    CHECK sy-subrc EQ 0.
    TYPES:
    BEGIN OF ty_str,
      a1 TYPE string,
      a2 TYPE string,
      a3 TYPE string,
      p1 TYPE string,
      p2 TYPE string,
      p3 TYPE string,
    END OF ty_str.
    DATA: lt_str TYPE TABLE OF ty_str,
          ls_str TYPE ty_str,
          lv_block TYPE string,
          lv_flag TYPE boolean.
    LOOP AT lt_data INTO lv_data.
      CASE lv_data.
        WHEN '[Version]' OR '[StdSatz]' OR '[Arbeitstag]' OR '[Pecunia]'
             OR '[Mita]' OR '[Kunde]' OR '[Auftrag]' OR '[Position]'.
          lv_block = lv_data.
          lv_flag = abap_false.
        WHEN OTHERS.
          lv_flag = abap_true.
      ENDCASE.
      CHECK lv_flag EQ abap_true.
      CASE lv_block.
        WHEN '[Auftrag]'.
          SPLIT lv_data AT ';' INTO ls_str-a1 ls_str-a2 ls_str-a3.
        WHEN '[Position]'.
          SPLIT lv_data AT ';' INTO ls_str-p1 ls_str-p2 ls_str-p3.
          APPEND ls_str TO lt_str.
      ENDCASE.
    ENDLOOP.

  • How to configure Essbase data source in OBIEE11g on Unix system?

    Hi,
    I am looking for documentation/link for how to configure Essbase data source in OBIEE 11g on UNIX system.
    Thanks in advance

    Hi Fayaz
    First You need "BI Administrator Client Tool" cause you need to make changes with in RPD (repository) BUT "BI Administrator Client Tool" for Unix is not Available.
    So you have download OBIEE 11.1.1.5 ""BI Administrator Client Tool" and Install it on Windows Platform.
    Now regarding your Original Issue I am also looking for this Info.
    Regards
    Sher
    Edited by: Sher Ullah Baig on Apr 17, 2012 4:00 PM

  • How to configure multiple databases in single listener

    Hi,
    I want to configure upto 8 standby databases in a server. But I do not know how to configure multiple databases within the listener.
    Also can I use the same port number for all the databases.
    please give your suggestions.
    thankyou
    satyanag

    Please check (http://download.oracle.com/docs/cd/B10501_01/network.920/a96580/toc.htm)
    Jonathan Ferreira
    http://oracle4dbas.blogspot.com

  • How to configure Dynamic Data Connection for Business View

    Hi,
    How can we configure Dynamic Data connection that we can save the profile of the connection to somewhere that we do not need to enter it everytime when we refresh the report?
    thanks and regards
    nora

    Hi James,
    Thanks for the reply. Its solved now. For anybody if interested you can set the dynamic email address either i) having it as part of payload - In this case use the xpath to query the payload varaible ii) use identity service - follow the following steps
    1.Create a user in Application enterprise manager/also you can use a existing account, if u are creating a new one assign the correct role
    2. In either case edit the user-properties.xml(bpel/system/services/config) file for the corresponding user and add an attribute called email
    3. Bounce the server for this changes to take effect
    4. In the notification properties in the to address use ids:getUserProperty and pass the attribute name

  • How to configure once data load then trigerd or run ibot?

    Hi Experts,
    I have a one requirement,
    1) Every day run one workflow( means data load into data warehouse)
    2) After, ibot should be run and delivery to users.
    3) We scheduled the workflows in DAC for every day morning.
    Requirement:
    Once data loaded, then IBot should be run and send to users dynamically (without scheduling).
    If workflow failed, IBot won’t be delivered.
    How to find out or configure once data load then trigerd or run ibot?
    I am using obi 10g and informatica 8 and os xp.
    Advance thanks..
    Thanks,
    Raja

    Hi,
    Below are the details for automating the OBIEE Scheduler.
    Create a batch file or Sh file with following command
    D:\OracleBI\server\Bin\saschinvoke -u Administrator/udrbiee007 -j 8
    -u is username/Password for Scheduler (username/password that u given while configuration)
    -j is a job id number when u create a ibot it will assign the new job number that can be identified from"Jobmanager"
    Refer the below thread for more information.
    iBot scheduling after ETL load
    Or ,
    What you the above also it will work but problem is we need specify the time like every day 6.30 am .
    Note: The condition report is true then the report will be delivered at 6.30 pm only but the condition is false the report will not triggered.
    I also implemented this but that is little bit different .
    Hope this help's
    Thanks
    Satya
    Edited by: Satya Ranki Reddy on Jul 13, 2012 12:05 PM

  • How to configure Spring Data JPA to handle sdo_geometry and Timestamp(0)?

    Hi all,
    I'm new to this community, not sure if I'm posting in the right place.
    I'm an Oracle old-hand, but new to Spring Data, having done a bit on postGIS, but now trying to apply it to Oracle.
    My problem:  I have a new maven project, with Entity and Repository classes, and a JUnit test program that uses DbUnit to preload test data.
    The current config uses a beans.xml file to set up:
      <bean id="entityManagerFactory"
        class="org.springframework.orm.jpa.LocalContainerEntityManagerFactoryBean">...
        <property name="jpaProperties">....
      <bean id="vendorAdapter"
        class="org.springframework.orm.jpa.vendor.HibernateJpaVendorAdapter">
        <property name="databasePlatform" value="org.hibernate.dialect.Oracle10gDialect" />
        <property name="generateDdl" value="false" />
      </bean>
      <bean id="dataSource" class="com.mchange.v2.c3p0.ComboPooledDataSource"
        destroy-method="close">
        <property name="driverClass" value="oracle.jdbc.OracleDriver" />
        <property name="jdbcUrl"
         value="jdbc:oracle:thin:@localhost:2632:BLADE" />...
    <bean id="transactionManager" class="org.springframework.orm.jpa.JpaTransactionManager">
      <property name="entityManagerFactory" ref="entityManagerFactory" />
    </bean>
    Test of repos.count() works ok.
    Test to call repos.findByName(String) fails, (stack below).
    Now, the current setup doesn't handle SDO_GEOMETRY, or TIMESTAMP(0), but will manage TIMESTAMP(6).
    The warnings below show that I'm  using "DefaultDataTypeFactory" and should probably be using OracleDataTypeFactory.
    I've also found the OracleDataSource class, and tried that, but still get the same data type issues.
    Ok, the big questions:
    1. Can anyone point me at an example to configure the dataSource, or vendorAdapter, to use the OracleDataTypeFactory, and retrieve these data types?
    2. What data type should I be using in Java for the geometry column?
    Thanks,
    Phil
    Test output warnings:
    16:24:05.846 [main] INFO  org.springframework.test.context.transaction.TransactionalTestExecutionListener - Began transaction (2) for test context [DefaultTestContext@990034 testClass = NotamRepositoryTest, testInstance = com.thalesgroup.uk.airscape.common.data.domain.dao.notam.NotamRepositoryTest@3ff1d6, testMethod = testNotamFetchByDatasetName@NotamRepositoryTest, testException = [null], mergedContextConfiguration = [MergedContextConfiguration@1346193 testClass = NotamRepositoryTest, locations = '{classpath:testBeans.xml}', classes = '{}', contextInitializerClasses = '[]', activeProfiles = '{}', contextLoader = 'org.springframework.test.context.support.DelegatingSmartContextLoader', parent = [null]]]; transaction manager [org.springframework.orm.jpa.JpaTransactionManager@1f1fe9d]; rollback [true]
    16:24:08.674 [main] WARN  org.dbunit.dataset.AbstractTableMetaData - Potential problem found: The configured data type factory 'class org.dbunit.dataset.datatype.DefaultDataTypeFactory' might cause problems with the current database 'Oracle' (e.g. some datatypes may not be supported properly). In rare cases you might see this message because the list of supported database products is incomplete (list=[derby]). If so please request a java-class update via the forums.If you are using your own IDataTypeFactory extending DefaultDataTypeFactory, ensure that you override getValidDbProducts() to specify the supported database products.
    16:24:08.674 [main] WARN  org.dbunit.util.SQLHelper - DATA_SET.SEC_TAG_MODIFIED data type (1111, 'TIMESTAMP(0)') not recognized and will be ignored. See FAQ for more information.
    16:24:08.674 [main] WARN  org.dbunit.util.SQLHelper - DATA_SET.CREATED data type (1111, 'TIMESTAMP(0)') not recognized and will be ignored. See FAQ for more information.
    16:24:08.674 [main] WARN  org.dbunit.util.SQLHelper - DATA_SET.MODIFIED data type (1111, 'TIMESTAMP(0)') not recognized and will be ignored. See FAQ for more information.
    16:24:08.690 [main] WARN  org.dbunit.util.SQLHelper - DATA_SET.VALID_FROM data type (1111, 'TIMESTAMP(0)') not recognized and will be ignored. See FAQ for more information.
    16:24:08.690 [main] WARN  org.dbunit.util.SQLHelper - DATA_SET.VALID_TO data type (1111, 'TIMESTAMP(0)') not recognized and will be ignored. See FAQ for more information.
    16:24:08.690 [main] WARN  org.dbunit.dataset.AbstractTableMetaData - Potential problem found: The configured data type factory 'class org.dbunit.dataset.datatype.DefaultDataTypeFactory' might cause problems with the current database 'Oracle' (e.g. some datatypes may not be supported properly). In rare cases you might see this message because the list of supported database products is incomplete (list=[derby]). If so please request a java-class update via the forums.If you are using your own IDataTypeFactory extending DefaultDataTypeFactory, ensure that you override getValidDbProducts() to specify the supported database products.
    16:24:08.690 [main] WARN  org.dbunit.util.SQLHelper - NOTAM.Q_POINT data type (1111, 'SDO_GEOMETRY') not recognized and will be ignored. See FAQ for more information.
    16:24:08.690 [main] WARN  org.dbunit.util.SQLHelper - NOTAM.Q_RANGE_RING data type (1111, 'SDO_GEOMETRY') not recognized and will be ignored. See FAQ for more information.
    java.lang.UnsupportedOperationException
        at org.hibernate.spatial.GeometrySqlTypeDescriptor.getExtractor(GeometrySqlTypeDescriptor.java:57)
        at org.hibernate.type.AbstractStandardBasicType.nullSafeGet(AbstractStandardBasicType.java:267)
        at org.hibernate.type.AbstractStandardBasicType.nullSafeGet(AbstractStandardBasicType.java:263)
        at org.hibernate.type.AbstractStandardBasicType.nullSafeGet(AbstractStandardBasicType.java:253)
        at org.hibernate.type.AbstractStandardBasicType.hydrate(AbstractStandardBasicType.java:338)
        at org.hibernate.persister.entity.AbstractEntityPersister.hydrate(AbstractEntityPersister.java:2969)
        at org.hibernate.loader.Loader.loadFromResultSet(Loader.java:1695)
        at org.hibernate.loader.Loader.instanceNotYetLoaded(Loader.java:1627)
        at org.hibernate.loader.Loader.getRow(Loader.java:1514)
        at org.hibernate.loader.Loader.getRowFromResultSet(Loader.java:725)
        at org.hibernate.loader.Loader.processResultSet(Loader.java:952)
        at org.hibernate.loader.Loader.doQuery(Loader.java:920)
        at org.hibernate.loader.Loader.doQueryAndInitializeNonLazyCollections(Loader.java:354)
        at org.hibernate.loader.Loader.doList(Loader.java:2553)
        at org.hibernate.loader.Loader.doList(Loader.java:2539)
        at org.hibernate.loader.Loader.listIgnoreQueryCache(Loader.java:2369)
        at org.hibernate.loader.Loader.list(Loader.java:2364)
        at org.hibernate.loader.hql.QueryLoader.list(QueryLoader.java:496)
        at org.hibernate.hql.internal.ast.QueryTranslatorImpl.list(QueryTranslatorImpl.java:387)
        at org.hibernate.engine.query.spi.HQLQueryPlan.performList(HQLQueryPlan.java:231)
        at org.hibernate.internal.SessionImpl.list(SessionImpl.java:1264)
        at org.hibernate.internal.QueryImpl.list(QueryImpl.java:103)
        at org.hibernate.jpa.internal.QueryImpl.list(QueryImpl.java:573)
        at org.hibernate.jpa.internal.QueryImpl.getResultList(QueryImpl.java:449)
        at org.hibernate.jpa.criteria.compile.CriteriaQueryTypeQueryAdapter.getResultList(CriteriaQueryTypeQueryAdapter.java:67)
        at org.springframework.data.jpa.repository.query.JpaQueryExecution$CollectionExecution.doExecute(JpaQueryExecution.java:81)
        at org.springframework.data.jpa.repository.query.JpaQueryExecution.execute(JpaQueryExecution.java:59)
        at org.springframework.data.jpa.repository.query.AbstractJpaQuery.doExecute(AbstractJpaQuery.java:97)
        at org.springframework.data.jpa.repository.query.AbstractJpaQuery.execute(AbstractJpaQuery.java:88)
        at org.springframework.data.repository.core.support.RepositoryFactorySupport$QueryExecutorMethodInterceptor.doInvoke(RepositoryFactorySupport.java:384)
        at org.springframework.data.repository.core.support.RepositoryFactorySupport$QueryExecutorMethodInterceptor.invoke(RepositoryFactorySupport.java:344)
        at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179)
        at org.springframework.transaction.interceptor.TransactionInterceptor$1.proceedWithInvocation(TransactionInterceptor.java:98)
        at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:262)
        at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:95)
        at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179)
        at org.springframework.dao.support.PersistenceExceptionTranslationInterceptor.invoke(PersistenceExceptionTranslationInterceptor.java:136)
        at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179)
        at org.springframework.data.jpa.repository.support.CrudMethodMetadataPostProcessor$CrudMethodMetadataPopulatingMethodIntercceptor.invoke(CrudMethodMetadataPostProcessor.java:111)
        at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179)
        at org.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:92)
        at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179)
        at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:207)
        at com.sun.proxy.$Proxy67.findByDataSetNameAndDeletedIsNull(Unknown Source)
        at com.thalesgroup.uk.airscape.common.data.domain.dao.notam.NotamRepositoryTest.testNotamFetchByDatasetName(NotamRepositoryTest.java:103)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
        at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
        at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
        at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
        at org.springframework.test.context.junit4.statements.RunBeforeTestMethodCallbacks.evaluate(RunBeforeTestMethodCallbacks.java:74)
        at org.springframework.test.context.junit4.statements.RunAfterTestMethodCallbacks.evaluate(RunAfterTestMethodCallbacks.java:83)
        at org.springframework.test.context.junit4.statements.SpringRepeat.evaluate(SpringRepeat.java:72)
        at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.runChild(SpringJUnit4ClassRunner.java:233)
        at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.runChild(SpringJUnit4ClassRunner.java:87)
        at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
        at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
        at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
        at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
        at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
        at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61)
        at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:71)
        at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
        at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:176)
        at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
        at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
        at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
        at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
        at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
        at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)

    Hi
    You can use T-code NACE and then select your application based on your output type you can assign your  Print program and Form.
    Reward points, if it is helpful.
    Regards
    Raja.

  • Data Guard - How?

    Hi!
    What I need to install Data Guard? I read, I need OEM Grid Control too. On download page there are 3 zips! This is more than database software! Am I right? Generally I'm looking for nice "step by step" tutorial for Data Guard.
    And License question: Now, I'm using downloable software from Oracle, but later I will use CDs from my customer: 10gR2 EE. Is this version enough?
    Thanx,
    Jacek

    And License question: Now, I'm using downloable software from Oracle, but later I will
    use CDs from my customer: 10gR2 EE. Is this version enough?There is no (none, nada, zero, zilch) diffference between the downloaded version and the CD version if the version number is the same.
    Long ago, Oracle separated the software from the license. Some consideration:
    a) Software
    - [Most] Oracle software is free;
    - All Oracle software is used under specific license;
    b) License
    - The license defines the 'right to use' conditions of using the software;
    - Some forms of license require associated payment;
    Patches
    - In this discussion (ie. in my definition), patches are not considered software;
    - Patches require paid support;
    - Paid support can only be attached to specific 'right to use' licenses (usually for-fee licenses);
    And finally, Oracle reserves the right to audit you (and your customers) to make sure you have the correct license for your use.
    By the way - Data Guard is part of the database and does not require anything from Grid Control. Go to this page to get help: http://www.oracle.com/technology/deploy/availability/htdocs/DataGuardOverview.html

  • Date sync: how can I sync dates from 2 cameras

    hi
    how can I sync dates from 2 cameras. From a wedding, my camera had the correct date, the borrowed one was about 11hrs out but not exactly. I want to change all the pics from the 2nd camera to sync with the first. I've already imported.
    Many thanks

    Use "Metadata→Adjust date and time".
    Instructions in the User Manual here.  Not that the change is an offset to the time-stamp listed in the EXIF.
    Message was edited by: Kirby Krieger

  • Data Grid, how to avoid unintentional data manipulation

    Hi,
    Is there a option to avoid unintentional writing in the data grid ?
    And if not, please implement such a toggle button for Security reasons.
    Andre

    Hello K.
    thank you for interesting in and care about my mental health.
    I'm alive and kicking.
    But if I follow your logic the developers of TOAD must have a real big (mental) problem.
    They are obviously that kind of paranoid that the say:
    “Our data grid is protected always against data modification as long as the user does not explicitly put the rowid to the column list to undoubtedly make clear that they want to manipulate some data.”
    I asked them if this is really needed, because in SQL Navigator there is just a simple button to toggle between RW and R only (What I definitely prefer) .
    And the said YES – to avoid unexpected / accidentally changes.
    Do you find that still strange?
    If yes these guys must be more ill then the two of as can imagine.
    But I’m very sure that they are alive and kicking too.
    By the way - one of my questions remained unanswered:
    Did you ever analysed other tools to pick the best feature of them in order to implement they into your tool?
    If not, believe me you should. It is worth the effort.
    Otherwise it is quite possible that you will take an unfounded amount of time till you will have reached a equivalent level.
    Or do you want to be behind the others for eternity?
    If you had the task to develop an new car would it ever come into your mind to build it with without a backward gear and simply to wait until someone requests something like that.
    Some (useful) standard has been set before you came up.
    What about a little less self satisfaction and a little bit more respect.
    That’s my advice for you (also) beyond the sqldev tool .
    Regards
    Andre

  • How to configure old AirPort Base Stations with current Mac

    Hi everyone,
    I noticed that you can't use the current AirPort Utility to configure old AirPort Base Stations. I'm talking about the pre-Extreme models that only had 802.11b (Graphite and Snow). It won't find them and even entering their IP manually won't let you enter the configuration.
    Copying over the older AirPort Admin Utility from a 10.3 system won't let you configure them either, the app itself runs fine though. You need to use an older PPC Mac with OS 9 or max OS X 10.3 if I recall correctly.
    Is there a way to do it directly from a current Intel Mac running Leopard?
    Thanks
    Björn

    *Airport Admin Utility* for Graphite and Snow
    <http://www.apple.com/downloads/macosx/apple/application_updates/airportadminuti lityforgraphiteandsnow425.html>

  • Configuring data guard: how many standby redolog do I need?

    Hi all, I want to add a standby databse to my RAC test environment but I don't understanad how many redolog I need.
    The main db is a 2 node RAC (10.2.0.4.0); every node has 3 log groups; every log group has 2 logfile (every log group is multiplexed on DG_DATA1/ASR/ and FLASH_RECOVERY_AREA/ASR/ )
    The standby db will have all its datafile on the same filesystem, so I already configured the trasnslations parameters:
    DB_FILE_NAME_CONVERT='/opt/oracle/app/oracle/oradata1/ASRSB','+DG_DATA1/ASR','/opt/oracle/app/oracle/oradata2/ASRSB','+FLASH_RECOVERY_AREA/ASR'
    LOG_FILE_NAME_CONVERT='/opt/oracle/app/oracle/oradata1/ASRSB','+DG_DATA1/ASR','/opt/oracle/app/oracle/oradata2/ASRSB','+FLASH_RECOVERY_AREA/ASR'
    Replication via rman works well, but I have to manually add the standby redolog on the standby db and which formula to use to calculate the number of the standby redolog needed. Which is the correct formula between :
    1)
    # of standby log = ( # of redolog files + 1) * # of thread = (6 +1)*2 = 14 standby redolog files
    2)
    # of standby log = ( # of redolog groups + 1) * # of thread = (3 +1)*2 = 8 standby redolog files
    thanks,
    andrea

    Andrea,
    I think 8 should be the answer on your question. Because you have 3 log groups and 2 threads therefore (3+1)*2 = 8.
    According to the documentation :
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ps.htm#SBYDB00426
    Minimally, the configuration should have one more standby redo log file group than the number of online redo log file groups on the primary database. However, the recommended number of standby redo log file groups is dependent on the number of threads on the primary database. Use the following equation to determine an appropriate number of standby redo log file groups:
    (maximum number of logfiles for each thread + 1) * maximum number of threads
    I think you may get confused with maximum number of logfiles. But sometimes logfile is used as a synonym to log group. The actual log group files are usually called log members. (for instance MAXLOGFILES parameter specifies the maximum number of log groups and MAXLOGMEMBERS specifies the maximum number of members for each group).
    Each additional member in a redo log group is just a multiplexed copy. You can also consider multiplexing the standby redo log files, the same why you do for redo log members.
    In order to learn more about the usage, benefits and limitations of Standby Redo Logs (SRL) you can refer to Metalink Note 2193444.1.
    I hope this helps.
    --Mihajlo                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • How to configure the date format in toolbar of ADF Calendar

    Hi,
    In the weekly view of the ADF Calendar, the date on the toolbar of the calendar component is something like this "May 1,2011 - May 7,2011".
    Is there a way to configure this. What changes need to be done to show this as "1 May 2011 - 7 May 2011".
    Thanks,
    Deepak

    See Bug 13825615 - backport to 11.1.1.7.0 bug 13045234 - date format not supported in toolbar by ca

  • How to configure optimal data connections to allow excel to retrieve multiple data sets form a single data source

    Hi all,
    I would like to have a data connection file used by excel to extract data form a SQL Server DB (so that change in location of the DB needs only a change to the  single data connection file). The excel file will retrieve data from only a single database,
    but there are multiple queries (30+) stored on SQL server that are used - each SQL server views' data is returned to a pivot table in an excel worksheet, which has an associated chart with various slicers on the main dashboard worksheet.
    Do I need a seperate data conenction file for each SQL server query being retrieved, or can a single connection file be created as all data comes from a single database?
    (all my attempts and research on the net have led me to believe that a seperate file is needed for each view, but this seems unnecessary)

    Hi, im learning my way with this, so apologies if I am providing too much or too little info.
    There are 8 source files which are very loosely related in that they capture infromation regarding what has happened within a metro railway over a day. However there are only a few relationships between the contents of tehse files.
    These tables are imported into SQL Server using SSIS where I have developed a number of views that query these 8 source tables to generate a number of metrics to provide insight into the service provided to the passengers. Some examples of metrics are: Number
    of passengers in transit at any given time, time taken to travel between adjacent stations, how much power was used during the day, what distance did the trains travel during the day, etc. Some views provide only a handful of rows, some provide 1M plus. There
    are now approx 40 seperate views created.
    I have then used a spreadsheet with a worksheet associated with each SQL server view. Each worksheet is set up as a pivot table, which allows the related chart on the main dashboard worksheet to use standard excel capability to slice and present the data
    in different ways.
    At the moment if the server on which teh database is stored moves, I have had to recreate the spreadsheet from scratch as I dont know how to change the connection information in any other way. Ideally I would like to have connection info in a single place
    to reduce ongoing maintenance, particularly as I would like to place the spreadsheet on a SharePoint server to distribute it to other users.
    Thanks

  • Data Guard DB Broker configuration failed after adding Standby DB

    I am trying to set up DG Broker configuration between Primary and Physical Standby DB. I have added Standby db and enabled it. But when i try to show configuration it failed. Given below are the errors.
    Any help regarding this error is appreciated.
    GMGRL> show configuration
    Configuration
    Name: DBTEST
    Enabled: YES
    Protection Mode: MaxPerformance
    Databases:
    DBSOURCE - Primary database
    DBTARGET - Physical standby database
    Fast-Start Failover: DISABLED
    Current status for "DBTEST":
    Warning: ORA-16607: one or more databases have failed
    DGMGRL> show database verbose DBTARGET
    Object "dbtarget" was not found
    Given below is the alert.log error
    Fatal NI connect error 12514, connecting to:
    (DESCRIPTION=(ADDRESS=(PROTOCOL=TCP)(HOST=pdg2.localdomain)(PORT=1521))(CONNECT_DATA=(SERVER=DEDICATED)(SERVICE_NAME=DBTARGET_DGB.US.JT.COM)(CID=(PROGRAM=oracle)(HOST=pdg1.localdomain)(USER=oracle))))
    VERSION INFORMATION:
    TNS for Linux: Version 11.1.0.6.0 - Production
    TCP/IP NT Protocol Adapter for Linux: Version 11.1.0.6.0 - Production
    Time: 29-AUG-2011 14:58:08
    Tracing not turned on.
    Tns error struct:
    ns main err code: 12564
    TNS-12564: TNS:connection refused
    ns secondary err code: 0
    nt main err code: 0
    nt secondary err code: 0
    nt OS err code: 0
    # tnsnames.ora Network Configuration File: /u01/app/oracle/product/11.1.0/db_1/network/admin/tnsnames.ora
    # Generated by Oracle configuration tools.
    DBSOURCE =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = pdg1.localdomain)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = DBSOURCE.US.JT.COM)
    DBTARGET =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = pdg2.localdomain)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = DBTARGET.US.JT.COM)
    I can connect from Primary to Standby db (DBTARGET). Given below is the status of that
    [oracle@pdg1 ~]$ sqlplus sys@dbtarget as sysdba
    SQL*Plus: Release 11.1.0.6.0 - Production on Mon Aug 29 15:01:43 2011
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    Enter password:
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    SQL>

    MSEBERG,
    Hi yes i am using spfile in both Primary and Standby. Given below is the list for inconsistent properties. It did not report anything.
    While running the status report on DBTARGET i get the following error. It waited for about 5 mins before giving the err
    Current status for "DBTARGET":
    Error: ORA-16662: network timeout when contacting a database
    Given below is the complete status.
    DGMGRL> SHOW DATABASE 'DBSOURCE' 'StatusReport'
    STATUS REPORT
    INSTANCE_NAME SEVERITY ERROR_TEXT
    DGMGRL> show database 'DBSOURCE' 'LogXptStatus';
    LOG TRANSPORT STATUS
    PRIMARY_INSTANCE_NAME STANDBY_DATABASE_NAME STATUS
    DBSOURCE DBTARGET
    DGMGRL> SHOW DATABASE 'DBSOURCE' 'InconsistentProperties';
    INCONSISTENT PROPERTIES
    INSTANCE_NAME PROPERTY_NAME MEMORY_VALUE SPFILE_VALUE BROKER_VALUE
    DGMGRL> SHOW DATABASE 'DBSOURCE' 'InconsistentLogXptProps';
    INCONSISTENT LOG TRANSPORT PROPERTIES
    INSTANCE_NAME STANDBY_NAME PROPERTY_NAME MEMORY_VALUE BROKER_VALUE
    DGMGRL> SHOW DATABASE VERBOSE 'DBSOURCE';
    Database
    Name: DBSOURCE
    Role: PRIMARY
    Enabled: YES
    Intended State: TRANSPORT-ON
    Instance(s):
    DBSOURCE
    Properties:
    DGConnectIdentifier = 'dbsource'
    ObserverConnectIdentifier = ''
    LogXptMode = 'ASYNC'
    DelayMins = '0'
    Binding = 'OPTIONAL'
    MaxFailure = '0'
    MaxConnections = '1'
    ReopenSecs = '300'
    NetTimeout = '30'
    RedoCompression = 'DISABLE'
    LogShipping = 'ON'
    PreferredApplyInstance = ''
    ApplyInstanceTimeout = '0'
    ApplyParallel = 'AUTO'
    StandbyFileManagement = 'MANUAL'
    ArchiveLagTarget = '0'
    LogArchiveMaxProcesses = '3'
    LogArchiveMinSucceedDest = '1'
    DbFileNameConvert = '/u01/app/oracle/oradata/DBSOURCE,
    /u01/app/oracle/oradata/DBSOURCE'
    LogFileNameConvert = '/u01/app/oracle/oradata/DBSOURCE,
    /u01/app/oracle/oradata/DBSOURCE, /u01/app/oracle/flash_recovery_area/DBSOURCE/onlinelog,
    /u01/app/oracle/flash_recovery_area/DBSOURCE/onlinelog'
    FastStartFailoverTarget = ''
    StatusReport = '(monitor)'
    InconsistentProperties = '(monitor)'
    InconsistentLogXptProps = '(monitor)'
    SendQEntries = '(monitor)'
    LogXptStatus = '(monitor)'
    RecvQEntries = '(monitor)'
    HostName = 'pdg1.localdomain'
    SidName = 'DBSOURCE'
    StandbyArchiveLocation = '/u01/app/oracle/oradata/DBSOURCE'
    AlternateLocation = ''
    LogArchiveTrace = '0'
    LogArchiveFormat = '%t_%s_%r.arc'
    LatestLog = '(monitor)'
    TopWaitEvents = '(monitor)'
    Current status for "DBSOURCE":
    SUCCESS
    DGMGRL> SHOW DATABASE VERBOSE 'DBTARGET'
    Database
    Name: DBTARGET
    Role: PHYSICAL STANDBY
    Enabled: YES
    Intended State: APPLY-ON
    Instance(s):
    DBTARGET
    Properties:
    DGConnectIdentifier = 'dbtarget'
    ObserverConnectIdentifier = ''
    LogXptMode = 'ASYNC'
    DelayMins = '0'
    Binding = 'OPTIONAL'
    MaxFailure = '0'
    MaxConnections = '1'
    ReopenSecs = '300'
    NetTimeout = '30'
    RedoCompression = 'DISABLE'
    LogShipping = 'ON'
    PreferredApplyInstance = ''
    ApplyInstanceTimeout = '0'
    ApplyParallel = 'AUTO'
    StandbyFileManagement = 'MANUAL'
    ArchiveLagTarget = '0'
    LogArchiveMaxProcesses = '3'
    LogArchiveMinSucceedDest = '1'
    DbFileNameConvert = '/u01/app/oracle/oradata/DBSOURCE,
    /u01/app/oracle/oradata/DBSOURCE'
    LogFileNameConvert = '/u01/app/oracle/oradata/DBSOURCE/archive,
    /u01/app/oracle/oradata/DBSOURCE/archive, /u01/app/oracle/oradata/flash_recovery_area/DBSOURCE,
    /u01/app/oracle/flash_recovery_area/DBSOURCE/onlinelog'
    FastStartFailoverTarget = ''
    StatusReport = '(monitor)'
    InconsistentProperties = '(monitor)'
    InconsistentLogXptProps = '(monitor)'
    SendQEntries = '(monitor)'
    LogXptStatus = '(monitor)'
    RecvQEntries = '(monitor)'
    HostName = 'pdg2.localdomain'
    SidName = 'DBTARGET'
    StandbyArchiveLocation = '/u01/app/oracle/oradata/DBSOURCE/archive'
    AlternateLocation = ''
    LogArchiveTrace = '0'
    LogArchiveFormat = '%t_%s_%r.arc'
    LatestLog = '(monitor)'
    TopWaitEvents = '(monitor)'
    Current status for "DBTARGET":
    Error: ORA-16662: network timeout when contacting a database
    DGMGRL> show configuration
    Configuration
    Name: DBTEST
    Enabled: YES
    Protection Mode: MaxPerformance
    Databases:
    DBSOURCE - Primary database
    DBTARGET - Physical standby database
    Fast-Start Failover: DISABLED
    Current status for "DBTEST":
    Warning: ORA-16607: one or more databases have failed
    Edited by: 880215 on Aug 29, 2011 2:47 PM

Maybe you are looking for

  • Need help in setting up a Multi Client chat server.

    I've been stuck at this problem longer than I should be, and it's starting to get to me. In my quest to learn more about networking in Java, I've decided to start with a simple multi client chat program. The problem is that I don't know exactly how t

  • I can't: mount disks / go to Safe Mode

    So, I wake up the other day, turn on my PowerBook G4 and my external disk G-Drive Q 250 GB only to find that it didn't mount. I turn my G-Drive off, disconnect it from FireWire 800, turn it back on, plug the cable in, and nothing happens. Then, I ope

  • How do I get rid of the pre-appended dbo_  in schema name ?

    A customer opened an SR which asked the following question I have a SQL Server database called SQLDB1. When Migrating the database (objects and data) to Oracle, the created schemas come over as dbo_OWNERNAME. How can I change the target schemas in Or

  • Contacts crash

    Hello, I have a problem with default contacts problem. Every time I try to use it it is crashing into the homescreen, but still working on background, same problem is happening when I am trying to configure it through the settings. I can't understand

  • Why is Lightroom 5 hanging when trying to export images?

    Why is Lightroom 5 hanging when trying to export images?