Issue at create physical schema

Hello,
I created a data server using the technology Hyperion Essbase and also put the port because I'm not use the default, then I'am creating a Physical Schema but I can not see any other value than Undefined.
So I can not continue.
What is happening?

You have to manually type in the Application/Database names for example http://2.bp.blogspot.com/__2AaArK5lW8/SNZib4cqZBI/AAAAAAAAAZg/v6SgHh4ux5A/s1600-h/6.png
Cheers
John
http://john-goodwin.blogspot.com/

Similar Messages

  • Issue in creating OPSS Schema with rcu.

    Hi,
      There is an issue in creating the OPSS schema in oracle DB 11.2.0.3.0 even though the rcu (Oracle Fusion Middleware Repository Creation Utility 11g (11.1.2.1.0)
       completed it without any error with status of success opss schema at the end of it's creation operation.
          But the validation of opss schema is getting Failed as shown by this query ! I 've refferenced this Doc's Configuring the OPSS Security Store
    SQL> desc jps_dn;
    ERROR:
    ORA-04043: object jps_dn does not exist
      There is no error in rcu log And in opss.log as well.  Also There is NO issue in createing the Other schema's like OIM,OAM,OAAM
        Wonder what am missing here that causing the shema not reflecting or Validation of it getting Failed  in the database. 
         Also i 've tried by changing this sec_case_sensitive_logon value to False  from true in the DB but to no effect.
         in both cases the OPSS schema validation is getting failed. 
         Greately appreciate any suggestion.  
    Please see the following summary of rcu operation
    Repository Creation Utility: Create - Completion Summary
    Database details:
    Host Name                       : ebs.oracle.com
    Port                            : 1521
    Service Name                    : IAM.ORACLE.COM
    Connected As                    : sys
    Prefix for (prefixable) Schema Owners : DEV
    RCU Logfile                     : /data/Rootdownloads/rcuHome/rcu/log/logdir.2013-09-25_04-57/rcu.log
    Component schemas created:
    Component                       Status  Logfile /data/Rootdownloads/rcuHome/rcu/log/logdir.2013-09-25_04-57/opss.log
    Oracle Platform Security Services               Success        
      Please suggest.
    Thanks
    Priya

    Hi,
    Thanks for the response Hussein and Helios,
       I am able to solve the issue as it was due to incorrect rcu version.
      But I am really struggling to Configure Policy store for the last 10 days with the following  Error.  I would be highly obliged if you could guide me on this issue.
    I am getting stuck at this  Error while running configureSecurityStore.py.  for the error I 've refferenced these Metalink Note ID's
      But to No Avail.
    configureSecurityStore.py Fails With ORA-00001: unique constraint (DEV_OPSS.IDX_JPS_RDN_PDN) violated. (Doc ID 1547423.1) (1549203.1)
       Summarining  
       The steps that I 've followed
    1. Created the Schema's Using rcu 11g (11.1.2.0.0) on 11.2.0.3.0 Oracle DB. (OS :RHEL 64 bit)
    2. Insatlled Weblogic 10.3.6
    3. Installed   Oracle Identity and Access Management 11g (11.1.2.0.0) And  SOA 11.1.1.7.0
    4. Run config.sh to create OIM,OAM and OAAM domains
    5. Run setDomainEnv.sh from user_projects/domains//bin
      After that I tried to Configure the DB Policy store before starting the Admin Server. But getting the same error. every time it get  failed
    I 've dropped schemas and Reinstall the WLS,IAM,SOA software again But got the same Error.
      at least 20 times in the last 10 days I 've Dropped schema even created New DB and Reinstalled All the IAM Software But to No Avail
    Please see the  following Exact Error. And I would be highly Obliged if you could drop few lines as what I am missing here in the entire process.
    [oracle@ebs Middleware]$ /oracle/Middleware/oracle_common/common/bin/wlst.sh /oracle/Middleware/Oracle_IAM/common/tools/configureSecurityStore.py -d /oracle/Middleware/user_projects/domains/IAMDomain/  -c IAM -p welcome1 -m create
    CLASSPATH=/oracle/Middleware/patch_wls1036/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/Middleware/patch_ocp371/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/jdk1.6.0_30/lib/tools.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic_sp.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic.jar:/oracle/Middleware/modules/features/weblogic.server.modules_10.3.6.0.jar:/oracle/Middleware/wlserver_10.3/server/lib/webservices.jar:/oracle/Middleware/modules/org.apache.ant_1.7.1/lib/ant-all.jar:/oracle/Middleware/modules/net.sf.antcontrib_1.1.0.0_1-0b2/lib/ant-contrib.jar:/oracle/Middleware/patch_wls1036/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/Middleware/patch_ocp371/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/jdk1.6.0_30/lib/tools.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic_sp.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic.jar:/oracle/Middleware/modules/features/weblogic.server.modules_10.3.6.0.jar:/oracle/Middleware/wlserver_10.3/server/lib/webservices.jar:/oracle/Middleware/modules/org.apache.ant_1.7.1/lib/ant-all.jar:/oracle/Middleware/modules/net.sf.antcontrib_1.1.0.0_1-0b2/lib/ant-contrib.jar:/oracle/Middleware/patch_wls1036/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/Middleware/patch_ocp371/profiles/default/sys_manifest_classpath/weblogic_patch.jar:/oracle/jdk1.6.0_30/lib/tools.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic_sp.jar:/oracle/Middleware/wlserver_10.3/server/lib/weblogic.jar:/oracle/Middleware/modules/features/weblogic.server.modules_10.3.6.0.jar:/oracle/Middleware/wlserver_10.3/server/lib/webservices.jar:/oracle/Middleware/modules/org.apache.ant_1.7.1/lib/ant-all.jar:/oracle/Middleware/modules/net.sf.antcontrib_1.1.0.0_1-0b2/lib/ant-contrib.jar::/oracle/Middleware/oracle_common/modules/oracle.jrf_11.1.1/jrf-wlstman.jar:/oracle/Middleware/oracle_common/common/wlst/lib/adfscripting.jar:/oracle/Middleware/oracle_common/common/wlst/lib/adf-share-mbeans-wlst.jar:/oracle/Middleware/oracle_common/common/wlst/lib/mdswlst.jar:/oracle/Middleware/oracle_common/common/wlst/resources/auditwlst.jar:/oracle/Middleware/oracle_common/common/wlst/resources/igfwlsthelp.jar:/oracle/Middleware/oracle_common/common/wlst/resources/jps-wlst.jar:/oracle/Middleware/oracle_common/common/wlst/resources/jps-wls-trustprovider.jar:/oracle/Middleware/oracle_common/common/wlst/resources/jrf-wlst.jar:/oracle/Middleware/oracle_common/common/wlst/resources/oamap_help.jar:/oracle/Middleware/oracle_common/common/wlst/resources/oamAuthnProvider.jar:/oracle/Middleware/oracle_common/common/wlst/resources/ossoiap_help.jar:/oracle/Middleware/oracle_common/common/wlst/resources/ossoiap.jar:/oracle/Middleware/oracle_common/common/wlst/resources/ovdwlsthelp.jar:/oracle/Middleware/oracle_common/common/wlst/resources/sslconfigwlst.jar:/oracle/Middleware/oracle_common/common/wlst/resources/wsm-wlst.jar:/oracle/Middleware/utils/config/10.3/config-launch.jar::/oracle/Middleware/wlserver_10.3/common/derby/lib/derbynet.jar:/oracle/Middleware/wlserver_10.3/common/derby/lib/derbyclient.jar:/oracle/Middleware/wlserver_10.3/common/derby/lib/derbytools.jar::
    Initializing WebLogic Scripting Tool (WLST) ...
    Welcome to WebLogic Server Administration Scripting Shell
    Type help() for help on available commands
    Info: Data source is: opss-DBDS
    Info: DB JDBC driver: oracle.jdbc.OracleDriver
    Info: DB JDBC URL: jdbc:oracle:thin:@ebs.oracle.com:1521/iam.oracle.com
    INFO: Found persistence provider "org.eclipse.persistence.jpa.PersistenceProvider". OpenJPA will not be used.
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  checkServiceSetup - done
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  checkServiceSchema - Store schema has been seeded completely
    Sep 26, 2013 1:35:36 AM oracle.security.jps.internal.config.ldap.LdapCredStoreServiceConfigurator schemaCompatibleHandler
    INFO: Credential store schema upgrade not required. Store Schema version 11.1.1.6.0 is compatible to the seed schema version 11.1.1.4.0
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  updateServiceConfiguration - done
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  seedSchemaAndCreateDIT - done
    Sep 26, 2013 1:35:40 AM oracle.security.jps.internal.tools.utility.JpsUtilMigrationCredImpl migrateCredentialData
    INFO: Migration of Credential Store data in progress.....
    Sep 26, 2013 1:35:40 AM oracle.security.jps.internal.tools.utility.JpsUtilMigrationCredImpl migrateCredentialData
    INFO: Migration of Credential Store data completed, Time taken for migration is 00:00:00
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  migrateData - done
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  testJpsService - done
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  checkServiceSetup - done
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  checkServiceSchema - Store schema has been seeded completely
    Sep 26, 2013 1:35:41 AM oracle.security.jps.internal.config.ldap.LdapKeyStoreServiceConfigurator schemaCompatibleHandler
    INFO: Keystore schema upgrade not required. Store Schema version 11.1.1.6.0 is compatible to the seed schema version 11.1.1.4.0
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  updateServiceConfiguration - done
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  seedSchemaAndCreateDIT - done
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  migrateData - done
    [oracle.security.jps.internal.config.db.DbKeyStoreServiceConfigurator]  testJpsService - done
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  checkServiceSetup - done
    Sep 26, 2013 1:35:45 AM oracle.security.jps.internal.config.ldap.LdapPolicyStoreServiceConfigurator schemaCompatibleHandler
    INFO: Policy schema upgrade not required. Store Schema version 11.1.1.6.0 is compatible to the seed schema version 11.1.1.4.0
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  checkServiceSchema - Store schema has been seeded completely
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  updateServiceConfiguration - done
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  seedSchemaAndCreateDIT - done
    WLS ManagedService is not up running. Fall back to use system properties for configuration.
    Sep 26, 2013 1:36:00 AM oracle.security.jps.internal.tools.utility.destination.apibased.JpsDstPolicy migrateDataInternal
    INFO: Migration of Admin Role Members started
    Sep 26, 2013 1:36:00 AM oracle.security.jps.internal.tools.utility.destination.apibased.JpsDstPolicy migrateDataInternal
    INFO: Migration of Admin Role Members completed in 00:00:00
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  migrateData - done
    [oracle.security.jps.internal.config.db.DbPolicyStoreServiceConfigurator]  testJpsService - done
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  checkServiceSetup - done
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  checkServiceSchema - Store schema has been seeded completely
    Sep 26, 2013 1:36:00 AM oracle.security.jps.internal.config.ldap.LdapAuditServiceConfigurator schemaCompatibleHandler
    INFO: Audit store schema upgrade not required. Store Schema version 11.1.1.6.0 is compatible to the seed schema version 11.1.1.4.0
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  updateServiceConfiguration - done
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  seedSchemaAndCreateDIT - done
    Sep 26, 2013 1:36:00 AM oracle.security.jps.internal.audit.AuditServiceImpl registerInternal
    WARNING: Cannot register to audit service for component "JPS".
    Sep 26, 2013 1:36:00 AM oracle.security.jps.internal.tools.utility.JpsUtilMigrationAuditStoreImpl migrateAuditStoreData
    INFO: Migration of Audit Store data in progress.....
    Sep 26, 2013 1:36:51 AM oracle.security.jps.internal.tools.utility.JpsUtilMigrationAuditStoreImpl migrateAuditStoreData
    INFO: Migration of Audit Store data completed, Time taken for migration is 00:00:50
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  migrateData - done
    [oracle.security.jps.internal.config.db.DbAuditStoreServiceConfigurator]  testJpsService - done
    persist to output: /oracle/Middleware/user_projects/domains/IAMDomain/config/fmwconfig - done
    INFO: Found persistence provider "org.eclipse.persistence.jpa.PersistenceProvider". OpenJPA will not be used.
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  checkServiceSetup - done
    Sep 26, 2013 1:36:55 AM oracle.security.jps.internal.config.ldap.LdapCredStoreServiceConfigurator schemaCompatibleHandler
    INFO: Credential store schema upgrade not required. Store Schema version 11.1.1.6.0 is compatible to the seed schema version 11.1.1.4.0
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  checkServiceSchema - Store schema has been seeded completely
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  updateServiceConfiguration - done
    [oracle.security.jps.internal.config.db.DbCredStoreServiceConfigurator]  seedSchemaAndCreateDIT - failed JPS-10000: There was an internal error in the policy store.
    Exception in thread "main" java.lang.RuntimeException: JPS-10000: There was an internal error in the policy store.
    oracle.security.jps.internal.api.common.JpsCredentialStoreLdapNodeCreationException: JPS-10000: There was an internal error in the policy store.
         at oracle.security.jps.internal.common.rdbms.util.JpsDbBootstrapImpl.createJpsCredentailStoreInLdap(JpsDbBootstrapImpl.java:303)
        at oracle.security.jps.internal.config.ldap.LdapCredStoreServiceConfigurator.addServiceStoreBase(LdapCredStoreServiceConfigurator.java:113)
         at oracle.security.jps.internal.config.ldap.LdapCredStoreServiceConfigurator.seedSchemaAndCreateDIT(LdapCredStoreServiceConfigurator.java:142)
        at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.runConfiguration(LdapServiceEnabler.java:484)
         at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.configureCredentialStoreService(LdapServiceEnabler.java:232)
        at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.configureSecurityServices(LdapServiceEnabler.java:170)
         at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.main(LdapServiceEnabler.java:129)
    Caused by: oracle.security.jps.service.policystore.PolicyStoreConnectivityException: JPS-10000: There was an internal error in the policy store.
         at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager.handleRollbackException(JpsDBDataManager.java:1345)
        at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager.internalCommitTxn(JpsDBDataManager.java:1508)
         at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager.commitTransactionInDoAs(JpsDBDataManager.java:1475)
        at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager.commitTransaction(JpsDBDataManager.java:1466)
         at oracle.security.jps.internal.common.rdbms.util.JpsDbBootstrapImpl.createJpsCredentailStoreInLdap(JpsDbBootstrapImpl.java:296)
        ... 6 more
    Caused by: javax.persistence.RollbackException: Exception [EclipseLink-4002] (Eclipse Persistence Services - 2.3.1.v20111018-r10243): org.eclipse.persistence.exceptions.DatabaseException
    Internal Exception: java.sql.SQLIntegrityConstraintViolationException: ORA-00001: unique constraint (DEV1_OPSS.IDX_JPS_RDN_PDN) violated
    Error Code: 1
    Call: INSERT INTO JPS_DN (ENTRYID, PARENTDN, RDN) VALUES (?, ?, ?)
         bind => [3 parameters bound]
    Query: InsertObjectQuery(EntryId=11437:rdn=cn=credentialstore:pdn=cn=jpsroot,cn=jpscontext,cn=iam,: JpsStore Entry={[EntryId = 11437:Attribute RowId = 45348
    dn = cn=CredentialStore,cn=IAM,cn=JPSContext,cn=jpsroot, EntryId = 11437:Attribute RowId = 45349
    objectclass = top, EntryId = 11437:Attribute RowId = 45350
    objectclass = orclContainer, EntryId = 11437:Attribute RowId = 45351
    cn = CredentialStore]})
        at org.eclipse.persistence.internal.jpa.transaction.EntityTransactionImpl.commitInternal(EntityTransactionImpl.java:102)
         at org.eclipse.persistence.internal.jpa.transaction.EntityTransactionImpl.commit(EntityTransactionImpl.java:63)
        at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager$8.run(JpsDBDataManager.java:1487)
         at oracle.security.jps.internal.policystore.rdbms.JpsDBDataManager.internalCommitTxn(JpsDBDataManager.java:1492)
        ... 9 more
    Caused by: Exception [EclipseLink-4002] (Eclipse Persistence Services - 2.3.1.v20111018-r10243): org.eclipse.persistence.exceptions.DatabaseException
    Internal Exception: java.sql.SQLIntegrityConstraintViolationException: ORA-00001: unique constraint (DEV1_OPSS.IDX_JPS_RDN_PDN) violated
    Error Code: 1
    Call: INSERT INTO JPS_DN (ENTRYID, PARENTDN, RDN) VALUES (?, ?, ?)
         bind => [3 parameters bound]
    Query: InsertObjectQuery(EntryId=11437:rdn=cn=credentialstore:pdn=cn=jpsroot,cn=jpscontext,cn=iam,: JpsStore Entry={[EntryId = 11437:Attribute RowId = 45348
    dn = cn=CredentialStore,cn=IAM,cn=JPSContext,cn=jpsroot, EntryId = 11437:Attribute RowId = 45349
    objectclass = top, EntryId = 11437:Attribute RowId = 45350
    objectclass = orclContainer, EntryId = 11437:Attribute RowId = 45351
    cn = CredentialStore]})
        at org.eclipse.persistence.exceptions.DatabaseException.sqlException(DatabaseException.java:324)
         at org.eclipse.persistence.internal.databaseaccess.DatabaseAccessor.executeDirectNoSelect(DatabaseAccessor.java:840)
        at org.eclipse.persistence.internal.databaseaccess.DatabaseAccessor.executeNoSelect(DatabaseAccessor.java:906)
         at org.eclipse.persistence.internal.databaseaccess.DatabaseAccessor.basicExecuteCall(DatabaseAccessor.java:592)
        at org.eclipse.persistence.internal.databaseaccess.DatabaseAccessor.executeCall(DatabaseAccessor.java:535)
         at org.eclipse.persistence.internal.sessions.AbstractSession.basicExecuteCall(AbstractSession.java:1717)
        at org.eclipse.persistence.sessions.server.ClientSession.executeCall(ClientSession.java:253)
        at org.eclipse.persistence.internal.queries.DatasourceCallQueryMechanism.executeCall(DatasourceCallQueryMechanism.java:207)
         at org.eclipse.persistence.internal.queries.DatasourceCallQueryMechanism.executeCall(DatasourceCallQueryMechanism.java:193)
        at org.eclipse.persistence.internal.queries.DatasourceCallQueryMechanism.insertObject(DatasourceCallQueryMechanism.java:342)
         at org.eclipse.persistence.internal.queries.StatementQueryMechanism.insertObject(StatementQueryMechanism.java:162)
        at org.eclipse.persistence.internal.queries.StatementQueryMechanism.insertObject(StatementQueryMechanism.java:177)
         at org.eclipse.persistence.internal.queries.DatabaseQueryMechanism.insertObjectForWrite(DatabaseQueryMechanism.java:472)
        at org.eclipse.persistence.queries.InsertObjectQuery.executeCommit(InsertObjectQuery.java:80)
         at org.eclipse.persistence.queries.InsertObjectQuery.executeCommitWithChangeSet(InsertObjectQuery.java:90)
        at org.eclipse.persistence.internal.queries.DatabaseQueryMechanism.executeWriteWithChangeSet(DatabaseQueryMechanism.java:287)
         at org.eclipse.persistence.queries.WriteObjectQuery.executeDatabaseQuery(WriteObjectQuery.java:58)
        at org.eclipse.persistence.queries.DatabaseQuery.execute(DatabaseQuery.java:844)
        at org.eclipse.persistence.queries.DatabaseQuery.executeInUnitOfWork(DatabaseQuery.java:743)
         at org.eclipse.persistence.queries.ObjectLevelModifyQuery.executeInUnitOfWorkObjectLevelModifyQuery(ObjectLevelModifyQuery.java:108)
        at org.eclipse.persistence.queries.ObjectLevelModifyQuery.executeInUnitOfWork(ObjectLevelModifyQuery.java:85)
         at org.eclipse.persistence.internal.sessions.UnitOfWorkImpl.internalExecuteQuery(UnitOfWorkImpl.java:2871)
        at org.eclipse.persistence.internal.sessions.AbstractSession.executeQuery(AbstractSession.java:1516)
         at org.eclipse.persistence.internal.sessions.AbstractSession.executeQuery(AbstractSession.java:1498)
        at org.eclipse.persistence.internal.sessions.AbstractSession.executeQuery(AbstractSession.java:1449)
        at org.eclipse.persistence.internal.sessions.CommitManager.commitNewObjectsForClassWithChangeSet(CommitManager.java:224)
         at org.eclipse.persistence.internal.sessions.CommitManager.commitAllObjectsForClassWithChangeSet(CommitManager.java:191)
        at org.eclipse.persistence.internal.sessions.CommitManager.commitAllObjectsWithChangeSet(CommitManager.java:136)
         at org.eclipse.persistence.internal.sessions.AbstractSession.writeAllObjectsWithChangeSet(AbstractSession.java:3799)
        at org.eclipse.persistence.internal.sessions.UnitOfWorkImpl.commitToDatabase(UnitOfWorkImpl.java:1415)
         at org.eclipse.persistence.internal.sessions.RepeatableWriteUnitOfWork.commitToDatabase(RepeatableWriteUnitOfWork.java:636)
        at org.eclipse.persistence.internal.sessions.UnitOfWorkImpl.commitToDatabaseWithChangeSet(UnitOfWorkImpl.java:1505)
         at org.eclipse.persistence.internal.sessions.RepeatableWriteUnitOfWork.commitRootUnitOfWork(RepeatableWriteUnitOfWork.java:267)
        at org.eclipse.persistence.internal.sessions.UnitOfWorkImpl.commitAndResume(UnitOfWorkImpl.java:1143)
         at org.eclipse.persistence.internal.jpa.transaction.EntityTransactionImpl.commitInternal(EntityTransactionImpl.java:84)
        ... 12 more
    Caused by: java.sql.SQLIntegrityConstraintViolationException: ORA-00001: unique constraint (DEV1_OPSS.IDX_JPS_RDN_PDN) violated
        at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:445)
        at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:396)
        at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:879)
        at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:450)
         at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:192)
        at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:531)
        at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:207)
        at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1044)
         at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1329)
        at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3593)
        at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3674)
         at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeUpdate(OraclePreparedStatementWrapper.java:1354)
        at org.eclipse.persistence.internal.databaseaccess.DatabaseAccessor.executeDirectNoSelect(DatabaseAccessor.java:831)
         ... 45 more
        at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.throwExceptionWithStackTrace(LdapServiceEnabler.java:145)
        at oracle.security.jps.internal.tools.configuration.ldap.LdapServiceEnabler.main(LdapServiceEnabler.java:137)
    Error: Failed to initialize security store.
    Error: Create operation has failed.
    [oracle@ebs Middleware]$
      Also OPSS schema is FIne as it's get validated by this query.
    SQL>  select * from JPS_DN where rdn like '%cn=credentialstore%';
       ENTRYID
    RDN
    PARENTDN
          3004
    cn=credentialstore
    cn=oracleschemaversion,cn=opss,
          5004
    cn=credentialstore
    cn=jpsroot,cn=jpscontext,cn=iam,
       ENTRYID
    RDN
    PARENTDN
    Please suggest.
    Thanks
    Priya

  • Problem when creating a new physical schema

    Hello,
    I'm a new user of ODI. I installed on windows 7 -64 bits. I used java of 32 bits.
    The error is that when I right-click on physical architectures to create a new physical schema, nothing happens.
    I have already reinstalled ODI twice. I have seen previous posts in this forum about this topic, but they haven't found the solution.
    Thank you very much for your help,
    david

    Hi David,
    If you do following one to resolve your issue.
    First of all, please un-install Java 32 bit and install Java (JDK 1.6.0_20 updated or higher version) 64-bit later install ODi 11g 64 bit in your windows 7
    Regards,
    Phani

  • Can we create logical/ physical schema by running any script (unix/window)?

    I wish to create logical/ physical schema in oracle data integrator
    by running any windows or preferably unix script.
    is there any method or scripts provided in ODI?

    I think you could insert info into Matser Respository tables:
    Connect info: snp_connect
    Logical Schemas : snp_lschema
    Physical Schemas: snp_pschema
    Or you may write some sql to do this.

  • A separate work schema for a physical schema ?

    Hi,
    The ODI best practices white paper tells us (p.47/48) to create a separate schema and designate that as the work schema for a physical schema. This, however means that the user with which we connect to the database needs to have powerful system privileges like CREATE ANY TABLE, DROP ANY TABLE depending on what you want to do. Many DBA's will object to this. A simpler solution seems to be to set the work schema to be identical to the physical schema.
    Has anybody done this and experienced drawbacks/issues with it ?
    Can anybody give any good arguments in favor of the best practice advising a separate work schema ?
    thanks, Arjan

    Arjan ,
    I would like to suggest my opinion. The main reason for allocating the work Schema separate to the main schema is to have all the $ tables created separated so that while debugging we can choose to have them and drop them later and also mostly when the CKM is enabled the E$ tables are created and they are not dropped. there are also situation where dynamic $ tables are created so each time you run they $ tables are created and in case of failure or de bugging they remain . On the second attempt we have a new dynamic $ table. Think of a scenario where we have both of them in one place and the amount of unnecessary $ tables we have with our DW tables . Also if such a scenario have to happen in production box. Nobody would like to have such a scenario and then you need to either drop them manually or through another pl/sql codes . Thats the main reason its good to have a separate Work Schema.
    Also we use the Work schema for creating stagging table at run time while ODI processing.
    Secondly regarding permission and using the schema i have seen both the kind of environment where user used in topology is either the Work Schema or the Data Schema. Yes you would need to give certain permission or priviledges like Drop , create as $ tables have to have them but some dont give permission like Delete , truncate target tables .
    Now if you make your DBA understand the architecture , he would surely provide the required permission for the Work schema tables to be created and dropped and i dont see any thing to worry, since $ tables gets created and dropped but yes the permission have to be in such a way the user used in ODI data server dont have the permission to drop the Data schema table , if you want have more restricited permission for truncate and delete too, but in development we would surely need truncate and delete too since we tend to load wrong data and logic . Think about it in a different scenarios and requirement.
    Look at this link Cezar has spoken on this topic. -http://odiexperts.com/?p=672
    hope this helps.

  • T-code WV31 create physical inventory for vendor consignment stock

    Hi all,
    I have some issue with T-code WV31i n Sap Retail. Even though it supposed to create physical inventory doc for unrestricted use stock, it also create physical inventory Docs for special stocks (vendor consignment stock). What could be the reason?
    Best regards
    Sameera
    Edited by: Sameera on Dec 23, 2011 7:07 AM

    Hi,
    If i do physical inventory for 2 quantity.
    1. Consignment Payables account debited with 2 qty ( Now stock becomes 4 nos)
    If i issue 4 quantity
    2. Consignment Payables account credited with 4 qty.
    If i do consignment settelment, it is setteling 4-2 (for 2 quantity).
    But I want to settele for 4 quantity.
    Thanks,
    Maruthi

  • Issues while creating a search project in BCC

    Hi There,
    Migrating the apps from 2007.3 to 10.0.3.Experiencing issues while creating a search project.
    Step #1.In Search Administration,navigating through Project > Global Settings & creating a Content Label named as "Catalog"
    Step #2.Navigating through Project > Create Production Project.Creating a project with Content Set as "CatalogSet"
    Step #3.Adding a Content with this info:
              Content Name: Catalog
              Type of Content : ATG Repository
              IndexingOutputConfig Path : /atg/commerce/search/ProductCatalogOutputConfig
              Host name : localhost
              RMI : 10160 ( RMI port of commerce instance)
         Creating a content & getting the message as New Content "CatalogTest" has been created.
         Issue #1: Content is not displaying in the Content eventhough content has been created successfully.
         Issue #2: Getting error for "Default Text Processing Options"
         You can refer these issues in the attached screenshot "Content-Page-with-error.jpg"
    Step #4.Creating a new TPO as "English" & trying to "Associate Search Project" & getting an error.Refer the attached error log "Associate-Search-Project.log"
    Due to these issues, "Build Index" button is disabled.Did anybody experienced these issues? Any help/comments are appreciated.
    Also attached the full server logs as commerce-server.log & publishing-server.log.
    Since no options for attachment in this forum, I posted the attachment in this link.Please refer it.
    https://groups.google.com/forum/?fromgroups#!topic/atg_tech/csEHi-x3Nyc
    Thanks in advance!
    Mani
    Edited by: user11263665 on Apr 25, 2012 10:38 PM

    We faced the same problem when migrating from 2007 to ATG 10.0.3. The reason this error comes up is because the Search depends on a local customization adapter /atg/searchadmin/adapter/customization/LocalCustomizationDataAdapter which adds English as a language in the text processing options.
    After the migration you need to import the initial ATG search related data into your Search schema. These instructions have been mentioned in the ATG install guide. Here is the extract from the docs.
    Note: The import scripts import data using startSQLRepository. Make sure your JTDataSource and
    FakeXADatasource are configured to permit this in home/localconfig. JTDataSource should automatically
    point to FakeXADataSource. FakeXADataSource should point to the administration database, and
    FaxeXADataSource_production should point to the production database. Note that some product
    installation processes, including Oracle ATG Web Knowledge Manager Self Service, need FakeXADataSource
    pointed at the production database, so you may need to change the FakeXADataSource between installations.
    once you are done with the configurations above, import only the searchadmindata xml into search schema as this imports the LocalCustomizationDataAdapter config information into your search schema. Once this is done, you should be good to index.
    <ATG10dir>\home\bin\startSQLRepository.bat|sh -m SearchAdmin.AdminUI
    -repository /atg/searchadmin/SearchAdminRepository -import
    <ATG10dir>\Search10.1\SearchAdmin\common\data\initial\searchadmindata.
    xml

  • Issue in Production Core Schema of ATG 10.0.3

    Hi All,
    I tried to create a schema in Production Core from CIM. The schema created successfully and when I tried to import data , it finished up saying that "No Data to Import".
    In the table listing, none of the DCS tables are found.
    I am using ATG 10.0.3 version with weblogic server.
    Kindly help to resolve this issue.
    Thanks,
    Dinesh

    What option did you select under CIM product selection? If you are not able to see DCS tables then you might have selected "ATG Platform" which would not include commerce. You would have to select B2C/B2B commerce or CRS depending on what you are setting up. It will include commerce module along with ATG platform and CA.

  • Significance of DEFAULT check in Physical Schema

    Hi guys! Can you tell me the significance of Default check in physical schema(in topology manager)?
    When I remove this check from my current physical schema and run any Interface based on the same. I get errors like Delete previous checksome error
    And If i create two or more than two physical schema and everytime when I want to Excecute interfaces based on these different schemas all I need to go to that particular schema and make it my default Schema; Otherwise excecuting Interface gives checksome error if that interface doesnot belong to our default physical schema.
    Can you please comment on this.
    Thankyou.

    Hi Diwakar,
    Adding my comment on this is,
    The below is coded in document,
    "If this box is checked, then the physical schema will be the data server default schema, when no schema has been specified. Only one physical schema can be marked by default."
    The bottom line is for one data server even though it has one/more physical schemas ,atleast one should be default schema.
    So every time if you are using different schema you have to make it as default to took that connection.
    P.S: Experts comments are welcome !!!
    Thanks,
    Guru

  • Creating Physical Directory

    hii...
    I am using CREATE DIRECTORY method.
    CREATE DIRECTORY MYDIR AS 'C:\TEMP1';
    it is working. But, i want to create physical directory if TEMP1 folder isn't in the C: drive.
    How can i do this? do you have any idea? thanx

    Oracle implemented the DIRECTORY model primarily for security reasons, so normal users should not be issuing CREATE DIRECTORY statements. A privileged user should create the directory and then give normal users appropriate permissions on the directory. The privileged user should know whether or not the physical directory exists.
    Just in case you did not know, the O/S directory pointed to by an Oracle DIRECTORY object needs to be on the same server as the Oracle instance itself.
    John

  • Creating physical standby using rman fails with ORA-19558: error de-allocat

    Dear All,
    We are creating physical stadnby database from 2 node RAC ( 2 node RAC to standalone physical standby).
    While in the rman duplicate process we are getting below error, we were not able to sorted out..
    No third party storage has been used ....
    DB version : 11gR2 and the OS is RHEL5
    Appreciate if any one hepl us to resolve this issue ...
    Thanks in advance ...
    RMAN-03009: failure of backup command on prmy1 channel at 10/05/2011 17:59:26
    ORA-19558: error de-allocating device
    ORA-19557: device error, device type: DISK, device name:
    ORA-17627: ORA-01041: internal error. hostdef extension doesn't exist
    ORA-17627: ORA-01041: internal error. hostdef extension doesn't exist
    ORA-03113: end-of-file on communication channel
    Edited by: 889828 on 2011/10/06 2:17 AM

    The problem is well decribed in your alert log.....you are using Oracle Managed File names which means you will not be able to duplicate the database as RMAN won't be able to automatically generate new names for the files using the DB_FILE_NAME_CONVERT and LOG_FILE_NAME_CONVERT initialization parameters.
    Youll have to use the SET NEWNAME command to specify the names of the database files. Done this a few times , and I don't recommend using OMF specifically beacuse of this hassle.
    WARNING: OMF is enabled on this database. Creating a physical standby controlfile, when OMF is enabled on the primary database, requires manual RMAN intervention to resolve OMF
    datafile pathnames.
    NOTE: Please refer to the RMAN documentation for procedures describing how to manually resolve OMF datafile pathnames.
    Problem handling is described in
    Oracle® Database Backup and Recovery Reference
    10g Release 2 (10.2)
    Part Number B14194-03

  • Problem while creating Physical standby database on remote Server

    Hi
    I try to create a physical standby datbaase using OEM 10gr2, it is working when I create on same server. No issue but when I try to create same on remote server it is giving below warning and process failing. Please let me know what needs to be down in this senario...
    It is showing redo log information in v$log and v$log_file but not creating physically on PHYSICAL Standby instance..
    I appriciate your help and ur time too..
    These messages are from PRIMARY database alert*.log file
    There is space for up to 12 standby redo logfiles
    Use the following SQL commands on the standby database to create
    standby redo logfiles that match the primary database:
    ALTER DATABASE ADD STANDBY LOGFILE 'srl1.f' SIZE 52428800;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl2.f' SIZE 52428800;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl3.f' SIZE 52428800;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl4.f' SIZE 52428800;
    ALTER DATABASE ADD STANDBY LOGFILE 'srl5.f' SIZE 52428800;
    WARNING: OMF is enabled on this database. Creating a physical
    standby controlfile, when OMF is enabled on the primary
    database, requires manual RMAN intervention to resolve OMF
    datafile pathnames.
    NOTE: Please refer to the RMAN documentation for procedures
    describing how to manually resolve OMF datafile pathnames.
    Thanks and Regards,
    RK

    Well, it is simply warning that "Creating a physical
    standby controlfile, when OMF is enabled on the primary
    database, requires manual RMAN intervention to resolve OMF
    datafile pathnames.
    NOTE: Please refer to the RMAN documentation for procedures
    describing how to manually resolve OMF datafile pathnames"
    For instance, the path for the logfile name in "ALTER DATABASE ADD STANDBY LOGFILE 'srl5.f' SIZE 52428800" may need to be changed manually. It is advisable to check the mentioned documentation.

  • Problem While Creating Physical Standby Using RMAN

    Hi Guru's
    May be this incidence you all face while creating physical standby DB. I try to create Physical standby database Using RMAN Duplicate Command from one server
    (pri machine) to Standby Machine.
    The steps i followed to create the above are as follows:
    Step 1:- Enable Forced Logging
    SQL> ALTER DATABASE FORCE LOGGING;
    Step 2:- Configure a Standby Redo Log
    SQL> ALTER DATABASE ADD STANDBY LOGFILE GROUP 10
      2> ('/oracle/dbs/log1c.rdo','/oracle/dbs/log2c.rdo') SIZE 500M;
    Step 3:- Set Primary Database Initialization Parameters
    SQL> create pfile='?/dbs/pfileorcl.ora' from spfile;
    Edit the pfile to add the standby parameters, here shown:
    db_unique_name='orcl'
    LOG_ARCHIVE_CONFIG='DG_CONFIG=(orcl,orcldr)'
    LOG_ARCHIVE_DEST_2='SERVICE=orcldr LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=orcldr'
    *.fal_server=orcldr
    *.fal_client=orcl
    *.standby_file_management=auto
    Step 4:- Enable Archiving
    SQL> SHUTDOWN IMMEDIATE;
    SQL> STARTUP MOUNT;
    SQL> ALTER DATABASE ARCHIVELOG;
    SQL> ALTER DATABASE OPEN;
    Step 5:- Setup tnsnames for standby database
    This should be done on primary database by altering tnsnames.ora or using NetCA command, and create it by the name orcldr
    orcldr =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = standby-svr)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = orcldr)
    Step 6:- Backup database and copy it to standby
    This backup script should be run on primary and copied to standby on the same mount point, running this scripts using RMAN, all files under /rman_backup should be copied to /rman_backup on standby server.
    $rman target /
    RMAN> run
    allocate channel c1 type disk;
    allocate channel c2 type disk;
    backup database format '/rman_backup/%U';
    backup archivelog all format '/rman_backup /%U';
    backup current controlfile for standby format '/rman_backup/%U';
    Step 7:- Standby Database Steps
    Installing Oracle Software
    This should be same release and patchset with exactly same ORACLE_HOME mount point.
    Setting Up listener
    Create and start a listener on standby database using NetCA, or creating listener.ora in $ORACLE_HOME/network/admin
    LISTENER =
      (DESCRIPTION =
        (ADDRESS = (PROTOCOL = TCP)(HOST = standby-svr)(PORT = 1521))
    Step 8:- Set Standby Database Initialization Parameters
    Copying the pfile created in primary database and renames it to initorcldr.ora, and changes these parameters:
    db_unique_name='orcldr'
    LOG_ARCHIVE_CONFIG='DG_CONFIG=(orcldr,orcl)'
    LOG_ARCHIVE_DEST_2='SERVICE=orcl LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=orcl'
    *.fal_server=orcl
    *.fal_client=orcldr
    *.LOG_ARCHIVE_DEST_STATE_2='ENABLE'
    Step 9:- Setup tnsnames for primary database
    This should be done on standby database by altering tnsnames.ora or using NetCA command, and create it by the name orcl
    orcl =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = primary-svr)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = orcl)
    Step 10:- Copy a Password File
    A password file must be created on the Primary and copied over to the Standby site. The sys password must be identical on both sites. Copy orapworcl to $ORACLE_HOME/dbs and rename file to orapworcldr
    Step 11:- Create mount points for Oracle
    Mount point should be created on standby server with the same convention used in primary, this includes the location for controlfiles, redo logs, datafiles, archive log destination and alert logs.
    $ mkdir -p /u01/app/oracle/oradata/orcl/
    $ mkdir -p /u01/app/oracle/flash_recovery_area/
    $ mkdir -p /u01/app/oracle/admin/orcl/adump
    $ mkdir -p /u01/app/oracle/admin/orcl/bdump
    $ mkdir -p /u01/app/oracle/admin/orcl/cdump
    $ mkdir -p /u01/app/oracle/admin/orcl/udump
    Step 12:- Use RMAN to restore backup and setup standby
    Connect to RMAN and execute the following command to create standby database, this should be performed on standby server after copying backup and setting tnsnames.
    $ export ORACLE_SID=orcldr
    $ rman target sys/tiger@orcl auxiliary sys/tiger@orcldr
    RMAN> duplicate target database for standby dorecover;
    Here i am getting problem as :
    While trying to connect to traget database and auxiliary database i notice that when RMAN is connect to both databases it shows
    orcl ( DBID xyz)
    orcl (not mounted)
    Here in my view the second database must be standby database name or auxiiary db name is standby db name which in my case is orcldr
    after this issue i am facing the second issues as :
    in command
    duplicate target database for standby dorecover;
    After modifying the command to
    RMAN> duplicate target database to "standby";
    rman-05520 database name mismatch
    Can u please let me know where i am mistaking

    ok for pri:
    db_name=orcl
    db_unique_name=orcl
    for standby:
    db_name=orcl 
    db_unique_name=sbyorcl ( i change orcldr to sbyorcl later)for new testing
    but now what i am getting on
      Verify connectivity
    On Primary Server:
    C:\> lsnrctl stop LISTENERI (working fine)
    C:\> lsnrctl start LISTENER (working fine)
    C:\> tnsping orcl (working fine)
    C:\> tnsping sbyorcl (working fine)
    C:\> sqlplus sys/xxxxx@orcl (working fine)
    C:\> sqlplus sys/xxxxx@sbyorcl (not working fine)
    ERROR:
    ORA-12514: TNS:listener does not currently know of service requested in connect
    descriptor
    On Standby Server:
    C:\>lsnrctl stop LISTENER
    C:\> lsnrctl start LISTENER
    C:\> tnsping orcl
    C:\> tnsping sbyorcl
    C:\> sqlplus sys/xxxxx@orcl
    C:\> sqlplus sys/xxxxx@sbyorcl
    My listner file for pri :-
    SID_LIST_LISTENER =
      (SID_LIST =
        (SID_DESC =
          (GLOBAL_DBNAME = orcl)
          (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
          (SID_NAME = ORCL)
    LISTENER =
      (DESCRIPTION =
        (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.129)(PORT = 1521))
    tnsnames.ora file on pri is:
    sbyorcl =
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.128)(PORT = 1521))
        (CONNECT_DATA =
          (SERVICE_NAME = sbyorcl)
    ORCL =
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.129)(PORT = 1521))
        (CONNECT_DATA =
          (SERVICE_NAME = orcl)
    and listener file on standby :
    SID_LIST_LISTENER =
      (SID_LIST =
        (SID_DESC =
          (SID_NAME = PLSExtProc)
          (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
          (PROGRAM = extproc)
        (SID_DESC =
          (GLOBAL_DBNAME = orcl)
          (ORACLE_HOME = C:\oracle\product\10.2.0\db_1)
          (SID_NAME = orcl)
    LISTENER =
      (DESCRIPTION_LIST =
        (DESCRIPTION =
          (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.128)(PORT = 1521))
        (DESCRIPTION =
          (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC0))
    and tnsnames.ora file on standby:
    ORCL =
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.128)(PORT = 1521))
        (CONNECT_DATA =
          (SERVICE_NAME = orcl)
    sbyorcl =
      (DESCRIPTION =
        (ADDRESS = (PROTOCOL = TCP)(HOST = 192.168.202.129)(PORT = 1521))
    (CONNECT_DATA =
           SERVICE_NAME = sbyorcl)
    EXTPROC_CONNECTION_DATA =
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC0))
        (CONNECT_DATA =
          (SID = PLSExtProc)
          (PRESENTATION = RO)
    Kinldy guide me where i am mistaking it's urgent

  • How to 1 logical schema to many physical schemas?

    I have a database schema which is instantiated on many different servers. I set up a physical schema pointing to one, and a logical schema pointing to that physical schema. I imported the schema to a model, created interfaces for the tables, and created a package to execute it; and that is all working for that one physical instance.
    1) How can I implement that same model, interfaces, and package for each of the physical instances?
    1a) Can I change the JDBC parameters at package run time to point to a different database? How?
    1b) Can I select a different physical schema for the logical schema at package run time so that I only have to set up a different physical schema for each database? How?
    Thank you.

    "But if you have a lot of context (for example 1000 stores), you can define a generic physical schema, a logical one. The physical is based on variables (host, port,..). "
    Using contexts is working for me, but at least one of my schemas has more than 50 server instances, so this approach would be beneficial. Before I posted this question, I had tried to use variables for the host, port, and SID without success. I used a global variable and gave it default values, but it failed. Then I tried setting the value in a package and creating a scenario, but that too failed. What am I missing?

  • Physical Schema and logical schema

    Hi,
    When creating the data server in the topology corresponding to the appropriate technology we are creating a physical schema. But then why do we need to create logical schema. Is it created for execution of the interface? And can multiple physical schemas be mapped to same logical schema?

    Hi
    Physical schema represents the actual connection to the data source or data target. Logical schema represents the logical name associated to that source or target.
    One logical schema can be associated with multiple physical schema along with context, i.e. one logical schema is associated with different physical schema using different context.
    It can be understood with following example:
    You have 3 environments: Dev, QA, Prod, each having different database servers as DB1, DB2, DB3, respectively. Similarly we have 3 context corresponding to Dev, QA and Prod. You create logical schema with name DB_source
    Now you associate physical DB servers to logical schema (DB_source) for each context:
    DEV: DB1
    QA: DB2
    PROD: DB3
    Now when u develop ODI interfaces, you use one context DEV which associates DB_source to DB1. While mentioning context for execution, keep it as "Execution". This means, whatever context you choose during execution, corresponding physical DBs will be used.
    Thus if you change the execution context, corresponding physical schema will be used during execution.
    Let me know if you have further questions !!

Maybe you are looking for