LDAP as a target

HI! I'm trying to load data from oracle to ldap server. I mapped the both tables but when i execute it i get this error:
1747 : 42000 : java.sql.SQLException: ORA-01747: invalid user.table.column, table.column, or column specification
If someone can help me i will appreciate it. Thnaks

HI VivaLaVida,
Please take a look at the following EM12c documentation:
http://docs.oracle.com/cd/E24628_01/doc.121/e24473/security.htm#BABGAGIJ
You can connect EM12c with the following authentication systems:
•Oracle Access Manager (OAM) SSO
•Repository-Based Authentication
•SSO-Based Authentication
•Enterprise User Security Based Authentication
•LDAP Authentication Options: Oracle Internet Directory and Microsoft Active Directory
for OID there are non-GUI configuration methods:
http://docs.oracle.com/cd/E24628_01/doc.121/e24473/security.htm#autoId12

Similar Messages

  • Howto: pull Authn Questions and Answers during recon from ldap

    hi Sun team,
    ENV:
    IDM6.0 , LDAP is a target resource and Psft is source resource.
    Business case:
    I have a CU Requirement, where I need to recon from LDAP and to do the initial seeding of idm.
    This LDAP has questions and answers stored for each user as it was being used for their current self service app. When I recon against LDAP, I am able to pull the users, but I somehow cannot populate the waveset.questions[#0].question and answer fields as well as waveset.suppliedQuestions[#0].question and answer fields. How would I go about pulling that info from LDAP into IDM so that IDM becomes the self-service app.

    Steve,
    I have the same range of user population. I have tested the functionality of it, but not the performance of it. Here is how my file looked, it had to have the username.
    **********FILE format*******************
    command,user,waveset.questions[#0].answer,waveset.questions[#1].answer,
    waveset.suppliedQuestions[#0].suppliedQuestion,waveset.suppliedQuestions[#0].answer
    update,ac5234,helloA1,helloA2,what is my dogs name?,puppy
    update,ac7234,helloAC1,helloAC2,what is my moms name?,angie
    ***********************************************************************************

  • Oracle.security.jps.service.policystore.PolicyObjectNotFoundException

    Hi Experts,
    I am getting the following erron when statting the managed server, Could you please let me know what could be the cause.
    oracle.security.jps.service.policystore.PolicyObjectNotFoundException: JPS-04028: Application with name "CallistaSMS" does not exist.
    at oracle.security.jps.internal.policystore.ldap.LdapPolicyStore.unsync_getApplicationPolicy(LdapPolicyStore.java:636)
    at oracle.security.jps.internal.policystore.ldap.LdapPolicyStore.getApplicationPolicy(LdapPolicyStore.java:733)
    at au.com.callista.security.policy.mgmt.PolicyManager.appStripeExistsInPolicy(PolicyManager.java:141)
    at au.com.callista.security.policy.mgmt.PolicyManager.createAppStripeEntryInPolicyIfNeeded(PolicyManager.java:152)
    at au.com.callista.security.policy.mgmt.PolicyManager.getApplicationPolicy(PolicyManager.java:161)
    at au.com.callista.security.policy.mgmt.PolicyManager.getAllAppRoleEntiesByName(PolicyManager.java:119)
    at au.com.callista.security.policy.mgmt.PolicyManager.appRoleExists(PolicyManager.java:131)
    at au.com.callista.security.policy.mgmt.PolicyManager.createAppRole(PolicyManager.java:203)
    at au.com.callista.security.policy.mgmt.PolicyManagerFacade.createAppRoleWithPrincipalEntry(PolicyManagerFacade.java:58)
    at au.com.callista.security.policy.init.PolicyInitiator.postStart(PolicyInitiator.java:39)
    at weblogic.application.internal.flow.BaseLifecycleFlow$PostStartAction.run(BaseLifecycleFlow.java:297)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.application.internal.flow.BaseLifecycleFlow$LifecycleListenerAction.invoke(BaseLifecycleFlow.java:199)
    at weblogic.application.internal.flow.BaseLifecycleFlow.postStart(BaseLifecycleFlow.java:71)
    at weblogic.application.internal.flow.TailLifecycleFlow.activate(TailLifecycleFlow.java:34)
    at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:637)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:205)
    at weblogic.application.internal.EarDeployment.activate(EarDeployment.java:58)
    at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:161)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:79)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activate(BasicDeployment.java:184)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activateFromServerLifecycle(BasicDeployment.java:361)
    at weblogic.management.deploy.internal.DeploymentAdapter$1.doActivate(DeploymentAdapter.java:52)
    at weblogic.management.deploy.internal.DeploymentAdapter.activate(DeploymentAdapter.java:200)
    at weblogic.management.deploy.internal.AppTransition$2.transitionApp(AppTransition.java:31)
    at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:261)
    at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:246)
    at weblogic.management.deploy.internal.ConfiguredDeployments.activate(ConfiguredDeployments.java:170)
    at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(ConfiguredDeployments.java:124)
    at weblogic.management.deploy.internal.DeploymentServerService.resume(DeploymentServerService.java:181)
    at weblogic.management.deploy.internal.DeploymentServerService.start(DeploymentServerService.java:97)
    at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    Regards,
    Salai

    hi Kal,
    I did the following,
    1. Created new repository with new prefix.
    2. REdirected the datasource to point new repository
    3. executed the the the following command
    wls:/jsf_domain/serverConfig> reassociateSecurityStore (domain="jsf_domain", servertype="DB_ORACLE", datasourcename="jdbc/DBPolicyDS", jpsroot="cn=CallistaSecurityStore")
    Starting policy store reassociation.
    The store and ServiceConfigurator setup done.
    Schema is seeded into the store
    Command FAILED, Reason: JPS-01055: Could not create ldap credential store instance. Reason oracle.security.jps.JpsException:
    JPS-01013: The credential store DN cn=CredentialStore,cn=jsf_domain,cn=JPSContext,cn=CallistaSecurityStore is missing in the LDAP store;
    the target DN must be pre-configured.
    Traceback (innermost last):
    File "<console>", line 1, in ?
    File "/uc01/app/oracle/product/middleware_jsf/oracle_common/common/wlst/jpsWlstCmd.py", line 1116, in reassociateSecurityStore
    File "/uc01/app/oracle/product/middleware_jsf/oracle_common/common/wlst/jpsWlstCmd.py", line 1097, in reassociateSecurityStoreImpl
    at weblogic.rjvm.ResponseImpl.unmarshalReturn(ResponseImpl.java:237)
    at weblogic.rmi.internal.BasicRemoteRef.invoke(BasicRemoteRef.java:223)
    at javax.management.remote.rmi.RMIConnectionImpl_1035_WLStub.invoke(Unknown Source)
    at weblogic.management.remote.common.RMIConnectionWrapper$16.run(ClientProviderBase.java:910)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:146)
    at weblogic.security.Security.runAs(Security.java:61)
    at weblogic.management.remote.common.RMIConnectionWrapper.invoke(ClientProviderBase.java:908)
    at javax.management.remote.rmi.RMIConnector$RemoteMBeanServerConnection.invoke(RMIConnector.java:993)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    javax.management.MBeanException: javax.management.MBeanException: JPS-01055: Could not create ldap credential store instance.
    Reason oracle.security.jps.JpsException: JPS-01013: The credential store DN cn=CredentialStore,cn=jsf_domain,cn=JPSContext,cn=CallistaSecurityStore
    is missing in the LDAP store; the target DN must be pre-configured.
    Could you please help me out.
    Regards,
    Salai

  • Mounting a local home directory

    Here is a scenario:
    A distributed environment with offices in cities worldwide.
    Requirement: When a user logs into a local workstation, say in Chicago, the user mounts a homedirectory on a Chicago server. When a user logs into a local workstation, say in London, the user mounts a homedirectory on a London server. etc etc
    Can this be done with DS 5.2 on Solaris9 ?
    I have not tried anything but these are my ideas:
    My first solution is to implement local ldap servers, then configure the client to use the local server as the default server. The local server then has an auto_home entry to refer to the local file server. The problem is that then one cannot set up a replicated environment because the master LDAP server at HQ would overwrite the local LDAP server with a different auto_home entry (unless one excluded the automountinformation attribute from replication ?)
    Any other ideas?

    [This is drifting too much from DS, but see my note at the end]
    For automount see the man page automount(1M) and look for "Replicated File Systems". It has hard-coded selection rules which favor local networks, and you can include client-side variables (predefined + automountd(1M) options) in your values.juser    $MYSRVR:/export/home
    *    chi-srvr,lon-srvr:/export/home/&amd (am-utils) is much more flexible, see http://www.am-utils.org/docs/am-utils/am-utils_8.html#SEC58.
    home/*  -type:=nfs;sublink=${key};rfs:=/export/home netgrp(eur-hosts);rhost:=lon-srvr \
        network(192.168.1.0);rhost=lon-srvr network(17.16.0.0);rhost=chi-srvr rhost=chi-srvrThe relevance to Directory Server and LDAP is that older techniques (often used with NIS) may still be valid and might just need a couple of simple, global LDAP entries. Targeting the tool which uses your data might avoid twisting your DIT into knots trying to control what data gets seen by those tools.
    That being said, I am personally very curious to see the DS and LDAP based solutions.

  • Edir service not showing up in migration list

    When I run miggui on my oes2 server, I can authenticate to both servers and can add some services to the migration list like iprint and file services, but edirectory isn't one of the options. I installed the target server with the pre-migration pattern and joined the existing tree, but the target is not a replica. Running dsrepair on my source server, my target server shows up in the "servers known to this database" list.
    The only thing I can think of is that I didn't configure ldap during the target server installation. Is that required? If not, any other ideas? Thanks.
    Tim

    Hmmm, just discovered that dns on target server has some issues. I can ping the source server by IP, but not by name. The source server is properly registered, so nslookup by IP and name bring back the correct results.

  • Novell eDirectory Target Recon: Unable to search LDAP LDAP: error code 53

    Hi All,
    I am getting following error while running the Novell eDirectory Target Reconciliation in batch mode:
    *ERROR,27 Oct 2009 22:59:54,263,[XL_INTG.EDIRECTORY],Paged Search failed.javax.naming.OperationNotSupportedException: [LDAP: error code 53 - Unwilling To Perform]; remaining name 'ou=centrica,ou=Regular,o=Infosys'*
    *DEBUG,27 Oct 2009 22:59:54,263,[XL_INTG.EDIRECTORY],tcUtilLDAPOperations -> ~~~~~~~~~~Entering disconnectFromLDAP()~~~~~~~~~~*
    *DEBUG,27 Oct 2009 22:59:54,263,[XL_INTG.EDIRECTORY],tcUtilLDAPOperations -> Closing initial directory context*
    *DEBUG,27 Oct 2009 22:59:54,263,[XL_INTG.EDIRECTORY],tcUtilLDAPOperations -> ~~~~~~~~~~Leaving disconnectFromLDAP()~~~~~~~~~~*
    *ERROR,27 Oct 2009 22:59:54,263,[XL_INTG.EDIRECTORY],tcTskLDAPUserReconciliation] --> execute(): failed. Exception in execute() method.*
    *javax.naming.NamingException: tcUtilLDAPOperations  -> : NamingException : Unable to search LDAP [[LDAP: error code 53 - Unwilling To Perform]]*
    *     at com.thortech.xl.integration.ldap.util.tcUtilLDAPOperations.search(Unknown Source)*
    *     at com.thortech.xl.util.schedule.tasks.tcTskLDAPUserReconciliation.processChange(tcTskLDAPUserReconciliation.java:2752)*
    *     at com.thortech.xl.util.schedule.tasks.tcTskLDAPUserReconciliation.execute(tcTskLDAPUserReconciliation.java:344)*
    *     at com.thortech.xl.scheduler.tasks.SchedulerBaseTask.run(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper$TaskExecutionAction.run(Unknown Source)*
    *     at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)*
    *     at weblogic.security.service.SecurityManager.runAs(Unknown Source)*
    *     at weblogic.security.Security.runAs(Security.java:41)*
    *     at Thor.API.Security.LoginHandler.weblogicLoginSession.runAs(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper.execute(Unknown Source)*
    *     at org.quartz.core.JobRunShell.run(JobRunShell.java:178)*
    *     at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:477)*
    *DEBUG,27 Oct 2009 22:59:54,264,[XL_INTG.EDIRECTORY],tcTskLDAPUserReconciliation] --> execute(): exit*
    This error is coming only for OU's which have more than 200 entries in them. The
    if anyone has any clue on this problem, then kindly help.
    Cheers,
    Sunny

    See if your e-directory supports paged searches.
    If not then you have to configure your e-directory to support paged searches. I think the connector by default only uses paged searches.
    You need to look for OID 1.2.840.113556.1.4.319 (page control) in supported control list of the rootDSE.
    Hope this helps,
    Sagar

  • Target Doesn't Exist in LDAP

    Hi,
    I installed Oracle COllaboration Suite on Widows 2000.After installation of Mailstorage , I tried to configure email, but under Unified messaging all services are down(ie. UM IMAP,UM POP,UM LIST SERVER,UM HOUSE KEEPING,UM INBOUND SMTP,UM OUTBOUNBD SMTP all are showing down). If I try to start ,that time it is showing "No process instance found".In mailstorage installation I didn't face any error, it was successfully insttaled.
    I checked using follwing commands
    D:\>oesctl show targets
    TARGET: neena:um_system:smtp_in
    TARGET: neena:um_system:smtp_out
    TARGET: neena:um_system:imap
    TARGET: neena:um_system:gc
    TARGET: neena:um_system:list
    TARGET: neena:um_system:pop
    D:\>oesctl show status nee:um_system:smtp_in
    Could not get a list of process for the target nee:um_system:smtp_in
    oracle.mail.ldap.ESDSException: [ESDS error code - 10019] Target Doesn't Exist
    In LDAP
    No processes defined for nee:um_system:smtp_in
    All Process showing same error message. How i can solve this, anyone face this type of problems
    Best regards
    satheesh kumar

    HI,
    After that I checked these things also I got the follwing types of errors for all process
    C:\>oesctl startup neena:um_system:smtp_in
    Exception in thread "main" java.lang.NullPointerException
    at oracle.mail.admin.util.AQHandler.dequeue(AQHandler.java:325)
    at oracle.mail.admin.targets.ProcessCommand.waitForResponse(ProcessComma
    nd.java:105)
    at oracle.mail.admin.targets.ProcessCommand.doCommands(ProcessCommand.ja
    va:153)
    at oracle.mail.admin.targets.ESProcess.startup(ESProcess.java:125)
    at oracle.mail.admin.targets.UMService.startup(UMService.java:157)
    at oracle.mail.admin.util.OesAdmin.startProcesses(OesAdmin.java:507)
    at oracle.mail.admin.util.OesAdmin.doStartupCommand(OesAdmin.java:483)
    at oracle.mail.admin.util.OesAdmin.processCommand(OesAdmin.java:217)
    at oracle.mail.admin.util.OesAdmin.main(OesAdmin.java:198)
    C:\>oesctl show status neena:um_system:smtp_in
    Operation cancelled: oracle.mail.admin.AdminException: AQ Session creation faile
    d. See log for details.
    C:\>oesctl refresh neena:um_system:smtp_in
    Unfortunate exception: oracle.mail.admin.AdminException: AQ Session creation fai
    led. See log for details.
    neena:um_system:smtp_in:107825826962564557 null
    Best Regards
    satheesh

  • Group Policy Item level targeting LDAP Query for specific AD Sites

    Hi Everyone,
    I'm looking to try and take advantage of Group Policy Preference Item Level Targeting to publish user Proxy settings based on what AD Site a user is located in.
    The company I work for has multiple proxies (for multiple regions). We have hundreds of AD Sites listed within our AD S and S setup. I know that I could potentially list every AD Site that requires a particular proxy, but this would generate additional admin
    overhead each time a new site is commissioned, as well as each time a site is decommissioned (this is a construction company, so may "sites" can fluctuate rapidly).
    Due to this, I would like to know if it is possible to filter by LDAP query, and filter for each proxy, based on a portion of the text/name of the AD Sites (as we use a specific naming convention for our sites, this could be dynamic enough for us to not
    have to add or remove additional AD sites).
    Is this possible, and, if so, how would I write the LDAP Filter/Query??
    Right now I would assume I would do it in the following manner:
    (&(objectCategory=site)(objectClass=site)(cn=AU-*)
    Any assistance would be greatly appreciated.
    Cheers,
    Simon

    > Right now I would assume I would do it in the following manner:
    >
    > (&(objectCategory=site)(objectClass=site)(cn=AU-*)
    Sites do NOT reside in the domain partition, but in the configuration
    partition... You can verify your LDAP filter with
    dsquery * -filter "(your filter here)"
    This query will return all matching objects' distinguished names (DN).
    But why don't you use the "Sites" ILT instead? This ILT supports ? and *
    as wildcards, so it might be sufficient.
    Martin
    Mal ein
    GUTES Buch über GPOs lesen?
    NO THEY ARE NOT EVIL, if you know what you are doing:
    Good or bad GPOs?
    And if IT bothers me - coke bottle design refreshment :))

  • Logging configuration is missing or invalid for target

    Hi,
    I am getting this error while starting the SOA Server
    HTTP to server: '[2001:0:cf2e:3096:30be:1076:3f57:fefc]', port: '7001'
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:314)
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:404)
    at weblogic.net.http.HttpClient.New(HttpClient.java:240)
    at weblogic.net.http.HttpURLConnection.connect(HttpURLConnection.java:17
    2)
    at weblogic.deploy.service.datatransferhandlers.HttpDataTransferHandler.
    getDataAsStream(HttpDataTransferHandler.java:75)
    Truncated. see log file for complete stacktrace
    >
    java.net.ConnectException: Tried all: '1' addresses, but could not connect over
    HTTP to server: '[2001:0:cf2e:3096:30be:1076:3f57:fefc]', port: '7001'
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:314)
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:404)
    at weblogic.net.http.HttpClient.New(HttpClient.java:240)
    at weblogic.net.http.HttpURLConnection.connect(HttpURLConnection.java:17
    2)
    at weblogic.deploy.service.datatransferhandlers.HttpDataTransferHandler.
    getDataAsStream(HttpDataTransferHandler.java:75)
    at weblogic.deploy.service.datatransferhandlers.DataHandlerManager$Remot
    eDataTransferHandler.getDataAsStream(DataHandlerManager.java:153)
    at weblogic.deploy.internal.targetserver.datamanagement.AppDataUpdate.do
    Download(AppDataUpdate.java:39)
    at weblogic.deploy.internal.targetserver.datamanagement.DataUpdate.downl
    oad(DataUpdate.java:56)
    at weblogic.deploy.internal.targetserver.datamanagement.Data.prepareData
    Update(Data.java:97)
    at weblogic.deploy.internal.targetserver.BasicDeployment.prepareDataUpda
    te(BasicDeployment.java:682)
    at weblogic.deploy.internal.targetserver.BasicDeployment.stageFilesForSt
    atic(BasicDeployment.java:725)
    at weblogic.deploy.internal.targetserver.AppDeployment.prepare(AppDeploy
    ment.java:104)
    at weblogic.management.deploy.internal.DeploymentAdapter$1.doPrepare(Dep
    loymentAdapter.java:39)
    at weblogic.management.deploy.internal.DeploymentAdapter.prepare(Deploym
    entAdapter.java:187)
    at weblogic.management.deploy.internal.AppTransition$1.transitionApp(App
    Transition.java:21)
    at weblogic.management.deploy.internal.ConfiguredDeployments.transitionA
    pps(ConfiguredDeployments.java:233)
    at weblogic.management.deploy.internal.ConfiguredDeployments.prepare(Con
    figuredDeployments.java:165)
    at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(Conf
    iguredDeployments.java:122)
    at weblogic.management.deploy.internal.DeploymentServerService.resume(De
    ploymentServerService.java:173)
    at weblogic.management.deploy.internal.DeploymentServerService.start(Dep
    loymentServerService.java:89)
    at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    <Aug 27, 2009 12:15:16 PM IST> <Error> <Deployer> <BEA-149205> <Failed to initia
    lize the application 'usermessagingdriver-email' due to error weblogic.managemen
    t.DeploymentException: Exception occured while downloading files.
    weblogic.management.DeploymentException: Exception occured while downloading fil
    es
    at weblogic.deploy.internal.targetserver.datamanagement.AppDataUpdate.do
    Download(AppDataUpdate.java:43)
    at weblogic.deploy.internal.targetserver.datamanagement.DataUpdate.downl
    oad(DataUpdate.java:56)
    at weblogic.deploy.internal.targetserver.datamanagement.Data.prepareData
    Update(Data.java:97)
    at weblogic.deploy.internal.targetserver.BasicDeployment.prepareDataUpda
    te(BasicDeployment.java:682)
    at weblogic.deploy.internal.targetserver.BasicDeployment.stageFilesForSt
    atic(BasicDeployment.java:725)
    Truncated. see log file for complete stacktrace
    java.net.ConnectException: Tried all: '1' addresses, but could not connect over
    HTTP to server: '[2001:0:cf2e:3096:30be:1076:3f57:fefc]', port: '7001'
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:314)
    at weblogic.net.http.HttpClient.openServer(HttpClient.java:404)
    at weblogic.net.http.HttpClient.New(HttpClient.java:240)
    at weblogic.net.http.HttpURLConnection.connect(HttpURLConnection.java:17
    2)
    at weblogic.deploy.service.datatransferhandlers.HttpDataTransferHandler.
    getDataAsStream(HttpDataTransferHandler.java:75)
    Truncated. see log file for complete stacktrace
    >
    <Aug 27, 2009 12:15:17 PM IST> <Error> <Deployer> <BEA-149205> <Failed to initia
    lize the application 'soa-infra' due to error weblogic.management.DeploymentExce
    ption: [J2EE:160149]Error while processing library references. Unresolved applic
    ation library references, defined in weblogic-application.xml: [Extension-Name:
    adf.oracle.domain, exact-match: false], [Extension-Name: oracle.sdp.messaging, e
    xact-match: false]..
    weblogic.management.DeploymentException: [J2EE:160149]Error while processing lib
    rary references. Unresolved application library references, defined in weblogic-
    application.xml: [Extension-Name: adf.oracle.domain, exact-match: false], [Exten
    sion-Name: oracle.sdp.messaging, exact-match: false].
    at weblogic.application.internal.flow.CheckLibraryReferenceFlow.prepare(
    CheckLibraryReferenceFlow.java:26)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.ja
    va:609)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineD
    river.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.j
    ava:185)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.jav
    a:53)
    Truncated. see log file for complete stacktrace
    >
    <Aug 27, 2009 12:15:18 PM IST> <Error> <Deployer> <BEA-149205> <Failed to initia
    lize the application 'worklistapp' due to error weblogic.management.DeploymentEx
    ception: [J2EE:160149]Error while processing library references. Unresolved appl
    ication library references, defined in weblogic-application.xml: [Extension-Name
    : adf.oracle.domain, exact-match: false], [Extension-Name: oracle.jsp.next, exac
    t-match: false]..
    weblogic.management.DeploymentException: [J2EE:160149]Error while processing lib
    rary references. Unresolved application library references, defined in weblogic-
    application.xml: [Extension-Name: adf.oracle.domain, exact-match: false], [Exten
    sion-Name: oracle.jsp.next, exact-match: false].
    at weblogic.application.internal.flow.CheckLibraryReferenceFlow.prepare(
    CheckLibraryReferenceFlow.java:26)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.ja
    va:609)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineD
    river.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.j
    ava:185)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.jav
    a:53)
    Truncated. see log file for complete stacktrace
    >
    <Aug 27, 2009 12:15:20 PM IST> <Error> <Deployer> <BEA-149205> <Failed to initia
    lize the application 'b2bui' due to error weblogic.management.DeploymentExceptio
    n: [J2EE:160149]Error while processing library references. Unresolved applicatio
    n library references, defined in weblogic-application.xml: [Extension-Name: adf.
    oracle.domain, exact-match: false]..
    weblogic.management.DeploymentException: [J2EE:160149]Error while processing lib
    rary references. Unresolved application library references, defined in weblogic-
    application.xml: [Extension-Name: adf.oracle.domain, exact-match: false].
    at weblogic.application.internal.flow.CheckLibraryReferenceFlow.prepare(
    CheckLibraryReferenceFlow.java:26)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.ja
    va:609)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineD
    river.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.j
    ava:185)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.jav
    a:53)
    Truncated. see log file for complete stacktrace
    >
    <Aug 27, 2009 12:15:20 PM IST> <Error> <Deployer> <BEA-149205> <Failed to initia
    lize the application 'DefaultToDoTaskFlow' due to error weblogic.management.Depl
    oymentException: [J2EE:160149]Error while processing library references. Unresol
    ved application library references, defined in weblogic-application.xml: [Extens
    ion-Name: adf.oracle.domain, Implementation-Version: 11.1.1.1.0, exact-match: fa
    lse], [Extension-Name: oracle.jsp.next, exact-match: false]..
    weblogic.management.DeploymentException: [J2EE:160149]Error while processing lib
    rary references. Unresolved application library references, defined in weblogic-
    application.xml: [Extension-Name: adf.oracle.domain, Implementation-Version: 11.
    1.1.1.0, exact-match: false], [Extension-Name: oracle.jsp.next, exact-match: fal
    se].
    at weblogic.application.internal.flow.CheckLibraryReferenceFlow.prepare(
    CheckLibraryReferenceFlow.java:26)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.ja
    va:609)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineD
    river.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.j
    ava:185)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.jav
    a:53)
    Truncated. see log file for complete stacktrace
    >
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    JmsConnectionFactory: Unable to set connection parameters for OracleConnectionMa
    nager
    Aug 27, 2009 12:15:47 PM oracle.wsm.common.logging.WsmMessageLogger logSevere
    SEVERE: Failure in looking up EJB component QueryService#oracle.wsm.policymanage
    r.ejb.IStringQueryServiceRemote.
    <Aug 27, 2009 12:16:29 PM IST> <Notice> <WebLogicServer> <BEA-000365> <Server st
    ate changed to ADMIN>
    <Aug 27, 2009 12:16:34 PM IST> <Warning> <Log Management> <BEA-170011> <The LogB
    roadcaster on this server failed to broadcast log messages to the admin server.
    The Admin server may not be running. Message broadcasts to the admin server will
    be disabled.>
    <Aug 27, 2009 12:17:11 PM IST> <Notice> <WebLogicServer> <BEA-000365> <Server st
    ate changed to RESUMING>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    3]" is now listening on fe80:0:0:0:39b9:eee1:8736:8262:8001 for protocols iiop,
    t3, ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    4]" is now listening on fe80:0:0:0:30be:1076:3f57:fefc:8001 for protocols iiop,
    t3, ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    7]" is now listening on 0:0:0:0:0:0:0:1:8001 for protocols iiop, t3, ldap, snmp,
    http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    1]" is now listening on 192.168.1.3:8001 for protocols iiop, t3, ldap, snmp, htt
    p.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default"
    is now listening on 2001:0:cf2e:3096:30be:1076:3f57:fefc:8001 for protocols iio
    p, t3, ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    5]" is now listening on fe80:0:0:0:c931:e3ad:23b7:a24c:8001 for protocols iiop,
    t3, ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    2]" is now listening on fe80:0:0:0:39e8:36d5:9f8b:f988:8001 for protocols iiop,
    t3, ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    6]" is now listening on fe80:0:0:0:0:5efe:c0a8:103:8001 for protocols iiop, t3,
    ldap, snmp, http.>
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <Server> <BEA-002613> <Channel "Default[
    8]" is now listening on 127.0.0.1:8001 for protocols iiop, t3, ldap, snmp, http.
    >
    <Aug 27, 2009 12:17:16 PM IST> <Notice> <WebLogicServer> <BEA-000357> <Started W
    ebLogic Independent Managed Server "soa_server1" for domain "domain1" running in
    Development Mode>
    <Aug 27, 2009 12:17:16 PM IST> <Warning> <Server> <BEA-002611> <Hostname "Metamo
    rph26-PC", maps to multiple IP addresses: 192.168.1.3, fe80:0:0:0:30be:1076:3f57
    :fefc%11, fe80:0:0:0:c931:e3ad:23b7:a24c%13, fe80:0:0:0:0:5efe:c0a8:103%14, 2001
    :0:cf2e:3096:30be:1076:3f57:fefc>
    <Aug 27, 2009 12:19:24 PM IST> <Warning> <JMX> <BEA-149509> <Unable to establish
    JMX Connectivity with the Adminstration Server AdminServer at service:jmx:t3://
    [2001:0:cf2e:3096:30be:1076:3f57:fefc]:7001/jndi/weblogic.management.mbeanserver
    s.domainruntime.
    java.io.IOException
    at weblogic.management.remote.common.ClientProviderBase.makeConnection(C
    lientProviderBase.java:187)
    at weblogic.management.remote.common.ClientProviderBase.newJMXConnector(
    ClientProviderBase.java:81)
    at javax.management.remote.JMXConnectorFactory.newJMXConnector(JMXConnec
    torFactory.java:338)
    at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFacto
    ry.java:247)
    at weblogic.management.mbeanservers.runtime.internal.RegisterWithDomainR
    untimeService.getDomainMBeanServerConnection(RegisterWithDomainRuntimeService.ja
    va:206)
    Truncated. see log file for complete stacktrace
    javax.naming.CommunicationException [Root exception is java.net.ConnectException
    : t3://[2001:0:cf2e:3096:30be:1076:3f57:fefc]:7001: Destination unreachable; nes
    ted exception is:
    java.net.ConnectException: Connection timed out: connect; No available r
    outer to destination]
    at weblogic.jndi.internal.ExceptionTranslator.toNamingException(Exceptio
    nTranslator.java:40)
    at weblogic.jndi.WLInitialContextFactoryDelegate.toNamingException(WLIni
    tialContextFactoryDelegate.java:783)
    at weblogic.jndi.WLInitialContextFactoryDelegate.getInitialContext(WLIni
    tialContextFactoryDelegate.java:367)
    at weblogic.jndi.Environment.getContext(Environment.java:315)
    at weblogic.jndi.Environment.getContext(Environment.java:285)
    Truncated. see log file for complete stacktrace
    java.net.ConnectException: t3://[2001:0:cf2e:3096:30be:1076:3f57:fefc]:7001: Des
    tination unreachable; nested exception is:
    java.net.ConnectException: Connection timed out: connect; No available r
    outer to destination
    at weblogic.rjvm.RJVMFinder.findOrCreateInternal(RJVMFinder.java:216)
    at weblogic.rjvm.RJVMFinder.findOrCreate(RJVMFinder.java:170)
    at weblogic.rjvm.ServerURL.findOrCreateRJVM(ServerURL.java:153)
    at weblogic.jndi.WLInitialContextFactoryDelegate.getInitialReference(WLI
    nitialContextFactoryDelegate.java:400)
    at weblogic.jndi.Environment.getInitialReference(Environment.java:245)
    Truncated. see log file for complete stacktrace
    java.rmi.ConnectException: Destination unreachable; nested exception is:
    java.net.ConnectException: Connection timed out: connect; No available r
    outer to destination
    at weblogic.rjvm.ConnectionManager.bootstrap(ConnectionManager.java:464)
    at weblogic.rjvm.ConnectionManager.bootstrap(ConnectionManager.java:315)
    at weblogic.rjvm.RJVMManager.findOrCreateRemoteInternal(RJVMManager.java
    :251)
    at weblogic.rjvm.RJVMManager.findOrCreate(RJVMManager.java:194)t
    at weblogic.rjvm.RJVMFinder.findOrCreateRemoteServer(RJVMFinder.java:238
    Truncated. see log file for complete stacktrace
    >
    <Aug 27, 2009 12:20:06 PM IST> <Notice> <WebLogicServer> <BEA-000365> <Server st
    ate changed to RUNNING>
    <Aug 27, 2009 12:20:06 PM IST> <Notice> <WebLogicServer> <BEA-000360> <Server st
    arted in RUNNING mode>
    I have reinstall it again and again but the same exception is coming.Please help me out.
    Thanks
    Ap

    It seem that you are missing libraries to start the application:
    Error while processing library references. Unresolved applic
    +ation library references, defined in weblogic-application.xml: [Extension-Name:+
    adf.oracle.domain, exact-match: false], [Extension-Name: *oracle.sdp.messaging*, e+
    +xact-match: false]..+
    weblogic.management.DeploymentException: J2EE:160149Error while processing lib
    rary references. Unresolved application library references, defined in weblogic-
    +application.xml: Extension-Name: adf.oracle.domain, exact-match: false, [Exten+
    sion-Name: oracle.sdp.messaging, exact-match: false].
    Either you haven't installed the libraries or you have installed them to the Admin server and need to activate them on the managed server. You should be able to see if they are installed if you look at the deployments overview in the weblogic console. I found this blog entry usefull when I needed to install the libraries for adf:
    http://one-size-doesnt-fit-all.blogspot.com/2009/01/configuring-weblogic-server.html
    If they are installed you might want to check each of them if they are assigned to you managed server target. You can select them in the deployment overview and choose the target option. There should be a check mark at the server you deployed your application to.
    Best regards,
    Sturla

  • How to get All Users from OID LDAP

    Hi all,
    I have Oracle Internet Directory(OID) and have created the users in it manually.
    Now I want to extract all the users from OID. How can I get Users from OID??
    Any response will be appritiated. If some one could show me demo code for that I shall be greatful to you.
    Thanks and reagards
    Pravy

    hi,
    the notes from metalink:
    bgards
    elvis
    Doc ID: Note:276688.1
    Subject: How to copy (export/import) the Portal database schemas of IAS 9.0.4 to another database
    Type: BULLETIN
    Status: PUBLISHED
    Content Type: TEXT/X-HTML
    Creation Date: 18-JUN-2004
    Last Revision Date: 05-AUG-2005
    How to copy (export/import) Portal database schemas of IAS 9.0.4 to another database
    Note 276688.1
    Download scripts Unix: Attachment 276688.1:1
    Download Perl scripts (Unix/NT) :Attachment 276688.1:2
    This article is being delivered in Draft form and may contain errors. Please use the MetaLink "Feedback" button to advise Oracle of any issues related to this article.
    HISTORY
    Version 1.0 : 24-JUN-2004: creation
    Version 1.1 : 25-JUN-2004: added a link to download the scripts from Metalink
    Version 1.2 : 29-JUN-2004: Import script: Intermedia indexes are recreated. Imported jobs are reassigned to Portal. ptlconfig replaces ptlasst.
    Version 1.3 : 09-JUL-2004: Additional updates. Usage of iasconfig.xml. Need only 3 environment variables to import.
    Version 1.4 : 18-AUG-2004: Remark about 9.2.0.5 and 10.1.0.2 database
    Version 1.5 : 26-AUG-2004: Duplicate job id
    Version 1.6 : 29-NOV-2004: Remark about WWC-44131 and WWSBR_DOC_CTX_54
    Version 1.7 : 07-JAN-2005: Attached perl scripts (for NT/Unix) at the end of the note
    Version 1.8 : 12-MAY-2005: added a work-around for the WWSTO_SESS_FK1 issue
    Version 1.9 : 07-JUL-2005: logoff trigger and 9.0.1 database export, import in 10g database
    Version 1.10: 05-AUG-2005: reference to the 10.1.2 note
    PURPOSE
    This document explains how to copy a Portal database schema from a database to another database.
    It allows restoring the Portal repository and the OID security associated with Portal.
    It can be used to go in production by copying physically a database from a development portal to a production environment and avoid to use the export/import utilities of Portal.
    This note:
    uses the export/import on the database level
    allows the export/import to be done between different platforms
    The script are Unix based and for the BASH shell. They can be adapted for other platforms.
    For the persons familiar with this technics in Portal 9.0.2, there is a list of the main differences with Portal 9.0.2 at the end of the note.
    These scripts are based on the experience of a lot of persons in Portal 902.
    The scripts are attached to the note. Download them here: Attachment 276688.1:1 : exp_schema_904.zip
    A new version of the script was written in Perl. You can also download them, here: Attachment 276688.1:2 : exp_schema_904_v2.zip. They do exactly the same than the bash ones. But they have the advantage of working on all platforms.
    SCOPE & APPLICATION
    This document is intented for Portal administrators. For using this note, you need basic DBA skills.
    This notes is for Portal 9.0.4.x only. The notes for Portal 9.0.2 are :
    Note 228516.1 : How to copy (export/import) Portal database schemas of IAS 9.0.2 to another database
    Note 217187.1 : How to restore a cold backup of a Portal IAS 9.0.2 on another machine
    The note for Portal 10.1.2 is:
    Note 330391.1 : How to copy (export/import) Portal database schemas of IAS 10.1.2 to another databaseMethod
    The method that we will follow in the document is the following one:
    Export:
    - export of the 4 portal schemas of a database (DEV / development)
    - export the LDAP OID users and groups (optional)
    Install a new machine with fresh IAS installation (PROD / production)
    Import:
    - delete the new and empty portal schema on PROD
    - import the schemas in the production database in place of the deleted schemas
    - import the LDAP OID users and groups (optional)
    - modify the configuration such that the infrastructure uses the portal repository of the backup
    - modify the configuration such that the portal repository uses the OID, webcache and SSO of the new infrastructure
    The export and the import are divided in several steps. All of these steps are included in 2 sample scripts:
    export : exp_portal_schema.sh
    import : imp_portal_schema.sh
    In the 2 scripts, all the steps are runned in one shot. It is just an example. Depending of the configuration and circonstance, all the steps can be runned independently.
    Convention
    Development (DEV) is the name of the machine where resides the copied database
    Production (PROD) is the name of the machine where the database is copied
    Prerequisite
    Some prerequisite first.
    A. Environment variables
    To run the import/export, you will need 3 environment variables. In the given scripts, they are defined in 'portal_env.sh'
    SYS_PASSWORD - the password of user sys in the Portal database
    IAS_PASSWORD - the password of IAS
    ORACLE_HOME - the ORACLE_HOME of the midtier
    The rest of the settings are found automatically by reading the iasconfig.xml file and querying the OID. It is done in 'portal_automatic_env.sh'. I wish to write a note on iasconfig.xml and the way to transform it in usefull environment variables. But it is not done yet. In the meanwhile, you can read the old 902 doc, that explains the meaning of most variables :
    < Note 223438.1 : Shell script to find your portal passwords, settings and place them in environment variables on Unix >
    B. Definition: Cutter database
    A 'Cutter Database' is the term used to designate a Database created by RepCA or OUI and that contains all the schemas used by a IAS 9.0.4 infrastructure. Even if in most cases, several schemas are not used.
    In Portal 9.0.4, the option to install only the portal repository in an empty database has been removed. It has been replaced by RepCA, a tool that creates an infrastructure database. Inside all the infrastucture database schemas, there are the portal schemas.
    This does not stop people to use 2 databases for running portal. One for OID and one for Portal. But in comparison with Portal 9.0.2, all schemas exist in both databases even if some are not used.
    The main idea of Cutter database is to have only 1 database type. And in the future, simplify the upgrades of customer installation
    For an installation where Portal and OID/SSO are in 2 separate databases, it looks like this
    Portal 9.0.2 Portal 9.0.4
    Infrastructure database
    (INFRA_SID)
    The infrastructure contains:
    - OID (used)
    - OEM (used)
    - Single Sign-on / orasso (used)
    - Portal (not used)
    The infrastructure contains:
    - OID (used)
    - OEM (used)
    - Single Sign-on / orasso (used)
    - Portal (not used)
    Portal database
    (PORTAL_SID)
    The custom Portal database contains:
    - Portal (used)
    The custom Portal database (is also an infrastructure):
    - OID (not used)
    - OEM (not used)
    - Single Sign-on / orasso (not used)
    - Portal (used)
    Whatever, the note will suppose there is only one single database. But it works also for 2 databases installation like the one explained above.
    C. Directory structure.
    The sample scripts given inside this note will be explained in the next paragraphs. But first, the scripts are done to use a directory structure that helps to classify the files.
    Here is a list of important files used during the process of export/import:
    File Name
    Description
    exp_portal_schema.sh
    Sample script that exports all the data needed from a development machine
    imp_portal_schema.sh
    Sample script that import all the data into a production machine
    portal_env.sh
    Script that defines the env variable specific to your system (to configure)
    portal_automatic_env.sh
    Helper script to get all the rest of the Portal settings automatically
    xsl
    Directory containing all the XSL files (helper scripts)
    del_authpassword.xsl
    Helper script to remove the authpassword tags in the DSML files
    portal_env_unix.sql
    Helper script to get Portal settings from the iasconfig.xml file
    exp_data
    Directory containing all the exported data
    portal_exp.dmp
    export on the database level of the portal, portal_app, ... database schemas
    iasconfig.xml
    copy the name of iasconfig.xml of the midtier of DEV. Used to get the hostname and port of Webcache
    portal_users.xml
    export from LDAP of the OID users used by Portal (optional)
    portal_groups.xml export from LDAP of the OID groups used by Portal (optional)
    imp_log
    Directory containing several spool and logs files generated during the import
    import.log Log file generated when running the imp command
    ptlconfig.log
    Log generated by ptlconfig when rewiring portal to the infrastructure.
    Some other spool files.
    D. Known limitations
    The scripts given in this note have the following known limitations:
    It does not copy the data stored in the SSO schema: external applications definitions and the passwords stored for them.
    See in the post steps: SSO migration to know how to do.
    The ssomig command resides in the Infrastructure Oracle home. And all commands of Portal in the Midtier home. And practically, these 2 Oracle homes are most of the time not on the same machine. This is the reason.
    The export of the users in OID exports from the default user location:
    ldapsearch .... -b "cn=users,dc=domain,dc=com"
    This is not 100% correct. The users are by default stored in something like "cn=users,dc=domain,dc=com". So, if the users are stored in the default location, it works. But if this location (user install base) is customized, it does not work.
    The reason is that such settings means that the LDAP most of the time highly customized. And I prefer that the administrator to copy the real LDAP himself. The right command will probably depend of the customer case. So, I prefered not to take the risk..
    orclCommonNicknameAttribute must match in the Target and Source OID .
    The orclCommonNicknameAttribute must match on both the source and target OID. By default this attribute is set to "uid", so if this has been changed, it must be changed in both systems.
    Reference Note 282698.1
    Migration of custom Java portlets.
    The script migrates all the data of Portal stored in the database. If you have custom java portlet deployed in your development machine, you will need to copy them in the production system.
    Step 1 - Export in Development (DEV)
    To export a full Portal installation to another machine, you need to follow 3 steps:
    Export at the database level the portal schemas + related schemas
    Get the midtier hostname and port of DEV
    Export of the users and groups with LDAPSEARCH in 2 XML files
    A script combining all the steps is available here.
    A. Export the 4 portals schemas (DEV)
    You need to export 3 types of database schemas:
    The 4 portal schemas created by default by the portal installation :
    portal,
    portal_app,
    portal_demo,
    portal_public
    The schemas where your custom database portlets / providers resides (if any)
    - The custom schemas you have created for storing your portlet / provider code
    The schemas where your custom tables resides. (if any)
    - Your custom schemas accessed by portal and containing only data (tables, views ...)
    You can get an approximate list of the schemas: default portal schemas (1) and database portlets schemas (2) with this query.
    SELECT USERNAME, DEFAULT_TABLESPACE, TEMPORARY_TABLESPACE
    FROM DBA_USERS
    WHERE USERNAME IN (user, user||'_PUBLIC', user||'_DEMO', user||'_APP')
    OR USERNAME IN (SELECT DISTINCT OWNER FROM WWAPP_APPLICATION$ WHERE NAME != 'WWV_SYSTEM');
    It still misses your custom schemas containing data only (3).
    We will export the 4 schemas and your custom ones in an export file with the user sys.
    Please, use a command like this one
    exp userid="'sys/change_on_install@dev as sysdba'" file=portal_exp.dmp grants=y log=portal_exp.log owner=(portal,portal_app,portal_demo,portal_public)The result is a dump file: 'portal_exp.dmp'. If you are using a database 9.2.0.5 or 10.1.0.2, the database of the exp/imp dump file has changed. Please read this.
    B. Hostname and port
    For the URL to access the portal, you need the 2 following infos to run the script 'imp_portal_schema.sh below :
    Webcache hostname
    Webcache listen port
    These values are contained in the iasconfig.xml file of the midtier.
    iasconfig.xml
    <IASConfig XSDVersion="1.0">
    <IASInstance Name="ias904.dev.dev_domain.com" Host="dev.dev_domain.com" Version="9.0.4">
    <OIDComponent AdminPassword="@BfgIaXrX1jYsifcgEhwxciglM+pXod0dNw==" AdminDN="cn=orcladmin" SSLEnabled="false" LDAPPort="3060"/>
    <WebCacheComponent AdminPort="4037" ListenPort="7782" InvalidationPort="4038" InvalidationUsername="invalidator" InvalidationPassword="@BR9LXXoXbvW1iH/IEFb2rqBrxSu11LuSdg==" SSLEnabled="false"/>
    <EMComponent ConsoleHTTPPort="1813" SSLEnabled="false"/>
    </IASInstance>
    <PortalInstance DADLocation="/pls/portal" SchemaUsername="portal" SchemaPassword="@BR9LXXoXbvW1c5ZkK8t3KJJivRb0Uus9og==" ConnectString="cn=asdb,cn=oraclecontext">
    <WebCacheDependency ContainerType="IASInstance" Name="ias904.dev.dev_domain.com"/>
    <OIDDependency ContainerType="IASInstance" Name="ias904.dev.dev_domain.com"/>
    <EMDependency ContainerType="IASInstance" Name="ias904.dev.dev_domain.com"/>
    </PortalInstance>
    </IASConfig>
    It corresponds to a portal URL like this:
    http://dev.dev_domain.com:7782/pls/portalThe script exp_portal_schema.sh copy the iasconfig.xml file in the exp_data directory.
    C. Export the security: users and groups (optional)
    If you use other Single Sing-On uses than the portal user, you probably need to restore the full security, the users and groups stored in OID on the production machine. 5 steps need to be executed for this operation:
    Export the OID entries with LDAPSEARCH
    Before to import, change the domain in the generated file (optional)
    Before to import, remove the 'authpassword' attributes from the generated files
    Import them with LDAPADD
    Update the GUID/DN of the groups in portal tables
    Part 1 - LDAPSEARCH
    The typical commands to do this operation look like this:
    ldapsearch -h $OID_HOSTNAME -p $OID_PORT -X -b "cn=portal.040127.1384,cn=groups,dc=dev_domain,dc=com" -s sub "objectclass=*" > portal_group.xml
    ldapsearch -h $OID_HOSTNAME -p $OID_PORT -X -D "cn=orcladmin" -w $IAS_PASSWORD -b "cn=users,dc=dev_domain,dc=com" -s sub "objectclass=inetorgperson" > portal_users.xmlTake care about the following points
    The groups are stored in a LDAP directory containing the date of installation
    ( in this example: portal.040127.1384,cn=groups,dc=dev_domain,dc=com )
    If the domain of dev and prod is different, the exported files contains the name of the development domain in the form of 'dc=dev_domain,dc=com' in a lot of place. The domain name needs to be replaced by the production domain name everywhere in the files.
    Ldapsearch uses the option '- X '. It it to export to DSML files (XML). It avoids a problem related with common LDAP files, LDIF files. LDIF files are wrapped at 78 characters. The wrapping to 78 characters make difficult to change the domain name contained in the LDIF files. XML files are not wrapped and do not have this problem.
    A sample script to export the 2 XML files is given here in : step 3 - export the users and groups (optional) of the export script.
    Part 2 : change the domain in the DSML files
    If the domain of dev and prod is different, the exported files contains the name of the development domain in the form of 'dc=dev_domain,dc=com' in a lot of place. The domain name need to be replaced by the production domain name everywhere in the files.
    To do this, we can use these commands:
    cat exp_data/portal_groups.xml | sed -e "s/$DEV_DN/$PROD_DN/" > imp_log/portal_groups.xml
    cat exp_data/portal_users.xml | sed -e "s/$DEV_DN/$PROD_DN/" > imp_log/temp_users.xml
    Part 3 : Remove the authpassword attribute
    The export of all attributes from the all users has also exported an automatically generated attribute in OID called 'authpassword'.
    'authpassword' is a list automatically generated passwords for several types of application. But mostly, it can not be imported. Also, there is no option in ldapsearch (that I know) that allows removing an attribute. In place of giving to the ldapsearch command the list of all the attributes that is very long, without 'authpassword', we will remove the attribute after the export.
    For that we will use the fact that the DSML files are XML files. There is a XSLT in the Oracle IAS, in the executable '$ORACLE_HOME/bin/xml'. XSLT is a standard specification of the internet consortium W3C to transform a XML file with the help of a XSL file.
    Here is the XSL file to remove the authpassword tag.
    del_autpassword.xsl
    <!--
    File : del_authpassword.xsl
    Version : 1.0
    Author : mgueury
    Description:
    Remove the authpassword from the DSML files
    -->
    <xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">
    <xml:output method="xml"/>
    <xsl:template match="*|@*|node()">
    <xsl:copy>
    <xsl:apply-templates select="*|@*|node()"/>
    </xsl:copy>
    </xsl:template>
    <xsl:template match="attr">
    <xsl:choose>
    <xsl:when test="@name='authpassword;oid'">
    </xsl:when>
    <xsl:when test="@name='authpassword;orclcommonpwd'">
    </xsl:when>
    <xsl:otherwise>
    <xsl:copy>
    <xsl:apply-templates select="*|@*|node()"/>
    </xsl:copy>
    </xsl:otherwise>
    </xsl:choose>
    </xsl:template>
    </xsl:stylesheet>
    And the command to make the transfomation:
    xml -f -s del_authpassword.xsl -o imp_log/portal_users.xml imp_log/temp_users.xmlWhere :
    imp_log/portal_users.xml is the final file without authpassword tags
    imp_log/temp_users.xml is the input file with the authpassword tags that can not be imported.
    Part 4 : LDAPADD
    The typical commands to do this operation look like this:
    ldapadd -h $OID_HOSTNAME -p $OID_PORT -D "cn=orcladmin" -w $IAS_PASSWORD -c -X portal_group.xml
    ldapadd -h $OID_HOSTNAME -p $OID_PORT -D "cn=orcladmin" -w $IAS_PASSWORD -c -X portal_users.xmlTake care about the following points
    Ldapadd uses the option ' -c '. Existing users/groups are generating an error. The option -c allows continuing and ignoring these errors. Whatever, the errors should be checked to see if it is just existing entries.
    A sample script to import the 2 XML files given in the step 5 - import the users and groups (optional) of the import script.
    Part 5 : Update the GUID/DN
    In Portal 9.0.4, the update of the GUID is taken care by PTLCONFIG during the import. (Import step 7)
    D. Example script for export
    Here is a example script that combines the 3 steps.
    Depending of you need, you will :
    or execute all the steps
    or just execute the 1rst one (export of the database users). It will be enough you just want to login with the portal user on the production instance.
    if your portal repository resides in a database 9.2.0.5 or 10.1.0.2, please read this
    you can download all the scripts here, Attachment 276688.1:1
    Do not forget to modify the script to your need and mostly add the list of users like explained in point A above.
    exp_portal_schema.sh
    # BASH Script : exp_portal_schema.sh
    # Version : 1.3
    # Portal : 9.0.4.0
    # History :
    # mgueury - creation
    # Description:
    # This script export a portal dump file from a dev instance
    # -------------------------- Environment variables --------------------------
    . portal_env.sh
    # In case you do not use portal_env.sh you have to define all the variables
    # For exporting the dump file only.
    # export SYS_PASSWORD=change_on_install
    # export PORTAL_TNS=asdb
    # For the security (optional)
    # export IAS_PASSWORD=welcome1
    # export PORTAL_USER=portal
    # export PORTAL_PASSWORD=A1b2c3de
    # export OID_HOSTNAME=development.domain.com
    # export OID_PORT=3060
    # export OID_DOMAIN_DN=dc=`echo $OID_HOSTNAME | cut -d '.' -f2,3,4,5,6 --output-delimiter=',dc='`
    # ------------------------------ Help function -----------------------------------
    function press_any_key() {
    if [ $PRESS_ANY_KEY_AFTER_EACH_STEP = "Y" ]; then
    echo
    echo Press enter to continue
    read $ANY_KEY
    else
    echo
    fi
    echo "------------------------------- Export ------------------------------------"
    # create a directory for the export
    mkdir exp_data
    # copy the env variables in the log just in case
    export > exp_data/exp_env_variable.txt
    echo "--------------------- step 1 - export"
    # export the portal users, but take care to add:
    # - your users containing DB providers
    # - your users containing data (tables)
    exp userid="'sys/$SYS_PASSWORD@$PORTAL_TNS as sysdba'" file=exp_data/portal_exp.dmp grants=y log=exp_data/portal_exp.log owner=(portal,portal_app,portal_demo,portal_public)
    press_any_key
    echo "--------------------- step 2 - store iasconfig.xml file of the MIDTIER"
    cp $MIDTIER_ORACLE_HOME/portal/conf/iasconfig.xml exp_data
    press_any_key
    echo "--------------------- step 3 - export the users and groups (optional)"
    # Export the groups and users from OID in 2 XML files (not LDIF)
    # The OID groups of portal are stored in GROUP_INSTALL_BASE that depends
    # of the installation date.
    # For the user, I use the default place. If it does not work,
    # you can find the user place with:
    # > exec dbms_output.put_line(wwsec_oid.get_user_search_base);
    # Get the GROUP_INSTALL_BASE used in security export
    sqlplus $PORTAL_USER/$PORTAL_PASSWORD@$PORTAL_TNS <<IASDB
    set serveroutput on
    spool exp_data/group_base.log
    begin
    dbms_output.put_line(wwsec_oid.get_group_install_base);
    end;
    IASDB
    export GROUP_INSTALL_BASE=`grep cn= exp_data/group_base.log`
    echo '--- Exporting Groups'
    echo 'creating portal_groups.xml'
    ldapsearch -h $OID_HOSTNAME -p $OID_PORT -X -s sub -b "$GROUP_INSTALL_BASE" -s sub "objectclass=*" > exp_data/portal_groups.xml
    echo '--- Exporting Users'
    echo 'creating portal_users.xml'
    ldapsearch -h $OID_HOSTNAME -p $OID_PORT -D "cn=orcladmin" -w $IAS_PASSWORD -X -s sub -b "cn=users,$OID_DOMAIN_DN" -s sub "objectclass=inetorgperson" > exp_data/portal_users.xml
    The script is done to run from the midtier.
    Step 2 - Install IAS in a new machine (PROD)
    A. Installation
    This note does not distinguish if Portal is sharing the same database than Single-Sign On and OID. For simplicity, I will speak only about 1 database. But I could also create a second infrastructure database just for the portal repository. This way is better for production system, because the Portal repository is only product used in the 2nd database. Having 2 separate databases allows taking easily backup of the portal repository.
    On the production machine, you need to install a fresh install of IAS 9.0.4. Take care to use :
    the same IAS patchset 9.0.4.1, 9.0.4.2, ...on the middle-tier and infrastruture than in development
    and same characterset than in development (or UTF8)
    The result will be 2 ORACLE_HOMES and 1 infrastructure database:
    the ORACLE_HOME of the infrastructure (SID:infra904)
    the ORACLE_HOME of the midtier (SID:ias904)
    an infrastructure database (SID:asdb)
    The empty new Portal install should work fine before to go to the next step.
    B. About tablespaces (optional)
    The size of the tablespace of the production should match the one of the Developement machine. If not, the tablespace will autoextend. It is not really a concern, but it is slow. You should modify the tablespaces for to have as much space on prod and dev.
    Also, it is safer to check that there is enough free space on the hard disk to import in the database.
    To modify the tablespace size, you can use Oracle Entreprise Manager console,
    On Unix, . oraenv
    infra904oemapp dbastudio
    On NT Start/ Programs/ Oracle Application server - infra904 / Enterprise Manager Console
    Launch standalone
    Choose the portal database (typically asdb.domain.com)
    Connect with a DBA user, sys or system
    Click Storage/Tablespaces
    Change the size of the PORTAL, PORTAL_DOC, PORTAL_LOGS, PORTAL_IDX tablespaces
    C. Backup
    It could be a good idea to take a backup of the MIDTIER and INFRASTRUCTURE Oracle Homes at that point to allow retesting the import process if it fails for any reason as much as you want without needing to reinstall everything.
    Step 3 - Import in production (on PROD)
    The following script is a sample of an Unix script that combines all the steps to import a portal repository to the production machine.
    To import a portal reporistory and his users and group in OID, you need to do 8 things:
    Stop the midtier to avoid errors while dropping the portal schema
    SQL*Plus with Portal
    Drop the 4 default portal schemas
    Create the portal users with the same passwords than the just deleted users and give them grants (you need to create your own custom shemas too if you have some).
    Import the dump file
    Import the users and groups into OID (optional)
    SQL*Plus with SYS : Post import changes
    Recompile everything in the database
    Reassign the imported jobs to portal
    SQL*Plus with Portal : Post import changes
    Recreate the Portal intermedia indexes
    Correct an import errror on wwsrc_preference$
    Make additional post import changes, by updating some portal tables, and replacing the development hostname, port or domain by the production ones.
    Rewire the portal repository with ptlconfig -dad portal
    Restart the midtier
    Here is a sample script to do this on Unix. You will need to adapt the script to your needs.
    imp_portal_schema.sh
    # BASH Script : imp_portal_schema.sh
    # Version : 1.3
    # Portal : 9.0.4.0
    # History :
    # mgueury - creation
    # Description:
    # This script import a portal dump file and relink it with an
    # infrastructure.
    # Script to be started from the MIDTIER
    # -------------------------- Environment variables --------------------------
    . portal_env.sh
    # Development and Production machine hostname and port
    # Example
    # .._HOSTNAME machine.domain.com (name of the MIDTIER)
    # .._PORT 7782 (http port of the MIDTIER)
    # .._DN dc=domain,dc=com (domain name in a LDAP way)
    # These values can be determined automatically with the iasconfig.xml file of dev
    # and prod. But if you do not know or remember the dev hostname and port, this
    # query should find it.
    # > select name, http_url from wwpro_providers$ where http_url like 'http%'
    # These variables are used in the
    # > step 4 - security / import OID users and groups
    # > step 6 - post import changes (PORTAL)
    # Set the env variables of the DEV instance
    rm /tmp/iasconfig_env.sh
    xml -f -s xsl/portal_env_unix.xsl -o /tmp/iasconfig_env.sh exp_data/iasconfig.xml
    . /tmp/iasconfig_env.sh
    export DEV_HOSTNAME=$WEBCACHE_HOSTNAME
    export DEV_PORT=$WEBCACHE_LISTEN_PORT
    export DEV_DN=dc=`echo $OID_HOSTNAME | cut -d '.' -f2,3,4,5,6 --output-delimiter=',dc='`
    # Set the env variables of the PROD instance
    . portal_env.sh
    export PROD_HOSTNAME=$WEBCACHE_HOSTNAME
    export PROD_PORT=$WEBCACHE_LISTEN_PORT
    export PROD_DN=dc=`echo $OID_HOSTNAME | cut -d '.' -f2,3,4,5,6 --output-delimiter=',dc='`
    # ------------------------------ Help function -----------------------------------
    function press_any_key() {
    if [ $PRESS_ANY_KEY_AFTER_EACH_STEP = "Y" ]; then
    echo
    echo Press enter to continue
    read $ANY_KEY
    else
    echo
    fi
    echo "------------------------------- Import ------------------------------------"
    # create a directory for the logs
    mkdir imp_log
    # copy the env variables in the log just in case
    export > imp_log/imp_env_variable.txt
    echo "--------------------- step 1 - stop the midtier"
    # This step is needed to avoid most case of ORA-01940: user connected
    # when dropping the portal user
    $MIDTIER_ORACLE_HOME/opmn/bin/opmnctl stopall
    press_any_key
    echo "--------------------- step 2 - drop and create empty users"
    sqlplus "sys/$SYS_PASSWORD@$PORTAL_TNS as sysdba" <<IASDB
    spool imp_log/drop_create_user.log
    ---- Drop users
    -- Warning: You need to stop all SQL*Plus connection to the
    -- portal schema before that else the drop will give an
    -- ORA-01940: cannot drop a user that is currently connected
    drop user portal_public cascade;
    drop user portal_app cascade;
    drop user portal_demo cascade;
    drop user portal cascade;
    ---- Recreate the users and give them grants"
    -- The new users will have the same passwords as the users we just dropped
    -- above. Do not forget to add your exported custom users
    create user portal identified by $PORTAL_PASSWORD default tablespace portal;
    grant connect,resource,dba to portal;
    create user portal_app identified by $PORTAL_APP_PASSWORD default tablespace portal;
    grant connect,resource to portal_app;
    create user portal_demo identified by $PORTAL_DEMO_PASSWORD default tablespace portal;
    grant connect,resource to portal_demo;
    create user portal_public identified by $PORTAL_PUBLIC_PASSWORD default tablespace portal;
    grant connect,resource to portal_public;
    alter user portal_public grant connect through portal;
    start $MIDTIER_ORACLE_HOME/portal/admin/plsql/wwv/wdbigra.sql portal
    exit
    IASDB
    press_any_key
    echo "--------------------- step 3 - import"
    imp userid="'sys/$SYS_PASSWORD@$PORTAL_TNS as sysdba'" file=exp_data/portal_exp.dmp grants=y log=imp_log/import.log full=y
    press_any_key
    echo "--------------------- step 4 - import the OID users and groups (optional)"
    # Some errors will be raised when running the ldapadd because at least the
    # default entries will not be able to be inserted. Remove them from the
    # ldif file if you want to avoid them. Due to the flag '-c', ldapadd ignores
    # duplicate entries. Another more radical solution is to erase all the entries
    # of the users and groups in OID before to run the import.
    # Replace the domain name in the XML files.
    cat exp_data/portal_groups.xml | sed -e "s/$DEV_DN/$PROD_DN/" > imp_log/portal_groups.xml
    cat exp_data/portal_users.xml | sed -e "s/$DEV_DN/$PROD_DN/" > imp_log/temp_users.xml
    # Remove the authpassword attributes with a XSL stylesheet
    xml -f -s xsl/del_authpassword.xsl -o imp_log/portal_users.xml imp_log/temp_users.xml
    echo '--- Importing Groups'
    ldapadd -h $OID_HOSTNAME -p $OID_PORT -D "cn=orcladmin" -w $IAS_PASSWORD -c -X imp_log/portal_groups.xml -v
    echo '--- Importing Users'
    ldapadd -h $OID_HOSTNAME -p $OID_PORT -D "cn=orcladmin" -w $IAS_PASSWORD -c -X imp_log/portal_users.xml -v
    press_any_key
    echo "--------------------- step 5 - post import changes (SYS)"
    sqlplus "sys/$SYS_PASSWORD@$PORTAL_TNS as sysdba" <<IASDB
    spool imp_log/sys_post_changes.log
    ---- Recompile the invalid packages"
    -- On the midtier, the script utlrp is not present. This step
    -- uses a copy of it stored in patch/utlrp.sql
    select count(*) INVALID_OBJECT_BEFORE from all_objects where status='INVALID';
    start patch/utlrp.sql
    set lines 999
    select count(*) INVALID_OBJECT_AFTER from all_objects where status='INVALID';
    ---- Jobs
    -- Reassign the JOBS imported to PORTAL. After the import, they belong
    -- incorrectly to the user SYS.
    update dba_jobs set LOG_USER='PORTAL', PRIV_USER='PORTAL' where schema_user='PORTAL';
    commit;
    exit
    IASDB
    press_any_key
    echo "--------------------- step 6 - post import changes (PORTAL)"
    sqlplus $PORTAL_USER/$PORTAL_PASSWORD@$PORTAL_TNS <<IASDB
    set serveroutput on
    spool imp_log/portal_post_changes.log
    ---- Intermedia
    -- Recreate the portal indexes.
    -- inctxgrn.sql is missing from the 9040 CD-ROMS. This is the bug 3536937.
    -- Fixed in 9041. The missing script is contained in the downloadable zip file.
    start patch/inctxgrn.sql
    start $MIDTIER_ORACLE_HOME/portal/admin/plsql/wws/ctxcrind.sql
    ---- Import error
    alter table "WWSRC_PREFERENCE$" add constraint wwsrc_preference_pk
    primary key (subscriber_id, id)
    using index wwsrc_preference_idx1
    begin
    DBMS_RLS.ADD_POLICY ('', 'WWSRC_PREFERENCE$', 'WEBDB_VPD_POLICY',
    '', 'webdb_vpd_sec', 'select, insert, update, delete', TRUE,
    static_policy=>true);
    end ;
    ---- Modify tables with full URLs
    -- If the domain name of prod and dev are different, this step is really important.
    -- It modifies the portal tables that contains reference to the hostname or port
    -- of the development machine. (For more explanation: see Addional steps in the note)
    -- groups (dn)
    update wwsec_group$
    set dn=replace( dn, '$DEV_DN', '$PROD_DN' )
    update wwsec_group$
    set dn_hash = wwsec_api_private.get_dn_hash( dn )
    -- users (dn)
    update wwsec_person$
    set dn=replace( dn, '$DEV_DN', '$PROD_DN' )
    update wwsec_person$
    set dn_hash = wwsec_api_private.get_dn_hash( dn)
    -- subscriber
    update wwsub_model$
    set dn=replace( dn, '$DEV_DN', '$PROD_DN' ), GUID=':1'
    where dn like '%$DEV_DN%'
    -- preferences
    update wwpre_value$
    set varchar2_value=replace( varchar2_value, '$DEV_DN', '$PROD_DN' )
    where varchar2_value like '%$DEV_DN%'
    update wwpre_value$
    set varchar2_value=replace( varchar2_value, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where varchar2_value like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- page url items
    update wwv_things
    set title_link=replace( title_link, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where title_link like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- web providers
    update wwpro_providers$
    set http_url=replace( http_url, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where http_url like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- html links created by the RTF editor inside text items
    update wwv_text
    set text=replace( text, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where text like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- Portlet metadata nls: help URL
    update wwpro_portlet_metadata_nls$
    set help_url=replace( help_url, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where help_url like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- URL items (There is a trigger on this table building absolute_url automatically)
    update wwsbr_url$
    set absolute_url=replace( absolute_url, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where absolute_url like '%$DEV_HOSTNAME:$DEV_PORT%'
    -- Things attributes
    update wwv_thingattributes
    set value=replace( value, '$DEV_HOSTNAME:$DEV_PORT', '$PROD_HOSTNAME:$PROD_PORT' )
    where value like '%$DEV_HOSTNAME:$DEV_PORT%'
    commit;
    exit
    IASDB
    press_any_key
    echo "--------------------- step 7 - ptlconfig"
    # Configure portal such that portal uses the infrastructure database
    cd $MIDTIER_ORACLE_HOME/portal/conf/
    ./ptlconfig -dad portal
    cd -
    mv $MIDTIER_ORACLE_HOME/portal/logs/ptlconfig.log imp_log
    press_any_key
    echo "--------------------- step 8 - restart the midtier"
    $MIDTIER_ORACLE_HOME/opmn/bin/opmnctl startall
    date
    Each step can generate his own errors due to a lot of factors. It is better to run the import step by step the first time.
    Do not forget to check the output of log files created during the various steps of the import:
    imp_log/drop_create_user.log
    Spool when dropping and recreating the portal users
    imp_log/import.log Import log file when importing the portal_exp.dmp file
    imp_log/sys_post_changes.log
    Spool when making post changes with SYS
    imp_log/portal_post_changes.log
    Spool when making post changes with PORTAL
    imp_log/ptlconfig.log
    Log file of ptconfig when rewiring the midtier
    Step 4 - Test
    A. Check the log files
    B. Test the website and see if it works fine.
    Step 5 - take a backup
    Take a backup of all ORACLE_HOME and DATABASES to prevent all hardware problems. You need to copy:
    All the files of the 2 ORACLE_HOME
    And all the database files.
    Step 6 - Additional steps
    Here are some additional steps.
    SSO external application ( that are part of the orasso schema and not imported yet )
    Page URL items ( they seems to store the full URL ) - included in imp_portal_schema.sh
    Web Providers ( the URL needs to be changed ) - included in imp_portal_schema.sh
    Text items edited with the RTF editor in IE and containing links - included in imp_portal_schema.sh
    Most of them are taken care by the "step 8 - post import changes". Except the first one.
    1. SSO import
    This script imports only Portal and the users/groups of OID. Not the list of the external application contained in the orasso user.
    In Portal 9.0.4, there is a script called SSOMIG that resides in $INFRA_ORACLE_HOME/sso/bin and allows to move :
    Definitions and user data for external applications
    Registration URLs and tokens for partner applications
    Connection information used by OracleAS Discoverer to access various data sources
    See:
    Oracle® Application Server Single Sign-On Administrator's Guide 10g (9.0.4) Part Number B10851-01
    14. Exporting and Importing Data
    2. Page items: the page URL items store the full URL.
    This is Bug 2661805 fixed in Portal 9.0.2.6.
    This following work-around is implemented in post import step of imp_portal_schema.sh
    -- page url items
    update wwv_things
    set title_link=replace( title_link, 'dev.dev_domain.com:7778', 'prod.prod_domain.com:7778' )
    where title_link like '%$DEV_HOSTNAME:$DEV_PORT%'
    2. Web Providers
    The URL to the Web providers needs also change. Like for the Page items, they contain the full path of the webserver.
    Or you can get the list of the URLs to change with this query
    select name, http_url from PORTAL.WWPRO_PROVIDERS$ where http_url like '%';
    This following work-around is implemented in post import step of imp_portal_schema.sh
    -- web providers
    update wwpro_providers$
    set http_url=replace( http_url, 'dev.dev_domain.com:7778', 'prod.prod_domain.com:7778' )
    where http_url like '%$DEV_HOSTNAME:$DEV_PORT%'
    4. The production and development machine do not share the same domain
    If the domain of the production and the development are not the same, the DN (name in LDAP) of all users needs to change.
    Let's say from
    dc=dev_domain,dc=com -> dc=prod_domain,dc=com
    1. before to upload the ldif files. All the strings in the 2 ldifs files that contain 'dc=dev_domain,dc=com', have to be replaced by 'dc=prod_domain,dc=com'
    2. in the wwsec_group$ and wwsec_person$ tables in portal, the DN need to change too.
    This following work-around is implemented in post import step of imp_portal_schema.sh
    -- groups (dn)
    update wwsec_group$
    set dn=replace( dn, 'dc=dev_domain,dc=com', 'dc=prod_domain,dc=com' )
    update wwsec_group$
    set dn_hash = wwsec_api_private.get_dn_hash( dn )
    -- users (dn)
    update wwsec_person$
    set dn=replace( dn, 'dc=dev_domain,dc=com', 'dc=prod_domain,dc=com' )
    update wwsec_person$
    set dn_hash = wwsec_api_private.get_dn_hash( dn)
    5. Text items with HTML links
    Sometimes people stores full URL inside their text items, it happens mostly when they use link with the RichText Editor in IE .
    This following work-around is implemented in post import step in imp_portal_schema.sh
    -- html links created by the RTF editor inside text items
    update wwv_text
    set text=replace( text, 'dev.dev_domain.com:7778', 'prod.prod_domain.com:7778' )
    where text like '%$DEV_HOSTNAME:$DEV_PORT%'
    6. OID Custom password policy
    It happens quite often that the people change the password policy of the OID server. The reason is that with the default policy, the password expires after 60 days. If so, do not forget to make the same changes in the new installation.
    PROBLEMS
    1. Import log has some errors
    A. EXP-00091 -Exporting questionable statistics
    You can ignore this error.
    B. IMP-00017 - WWSRC_PREFERENCE$
    When importing, there is one import error:
    IMP-00017: following statement failed with ORACLE error 921:
    "ALTER TABLE "WWSRC_PREFERENCE$" ADD "
    IMP-00003: ORACLE error 921 encountered
    ORA-00921: unexpected end of SQL commandThe primary key is not created. You can create it with this commmand
    in SQL*Plus with the user portal.. Then readd the missing VPD policy.
    alter table "WWSRC_PREFERENCE$" add constraint wwsrc_preference_pk
    primary key (subscriber_id, id)
    using index wwsrc_preference_idx1
    begin
    DBMS_RLS.ADD_POLICY ('', 'WWSRC_PREFERENCE$', 'WEBDB_VPD_POLICY',
    '', 'webdb_vpd_sec', 'select, insert, update, delete', TRUE,
    static_policy=>true);
    end ;
    Step 8 in the script "imp_portal_schema.sh" take care of this. This can also possibly be solved by the
    C. IMP-00017 - WWDAV$ASL
    . importing table "WWDAV$ASL"
    Note: table contains ROWID column, values may be obsolete 113 rows importedThis error is normal, the table really contains a ROWID column.
    D. IMP-00041 - Warning: object created with compilation warnings
    This error is normal too. The packages giving these error have
    dependencies on package not yet imported. A recompilation is done
    after the import.
    E. ldapadd error 'cannot add add entries containing authpasswords'
    # ldap_add: DSA is unwilling to perform
    # ldap_add: additional info: You cannot add entries containing authpasswords.
    "authpasswords" are automatically generated values from the real password of the user stored in userpassword. These values do not have to be exported from ldap.
    In the import script, I remove the additional tag with a XSL stylesheet 'del_authpassword.xsl'. See above.
    F. IMP-00017: WWSTO_SESSION$
    IMP-00017: following statement failed with ORACLE error 2298:
    "ALTER TABLE "WWSTO_SESSION$" ENABLE CONSTRAINT "WWSTO_SESS_FK1""
    IMP-00003: ORACLE error 2298 encountered
    ORA-02298: cannot validate (PORTAL.WWSTO_SESS_FK1) - parent keys not found
    Here is a work-around for the problem. I will normally integrate it in a next version of the scripts.
    SQL> delete from WWSTO_SESSION_DATA$;
    7690 rows deleted.
    SQL> delete from WWSTO_SESSION$;
    1073 rows deleted.
    SQL> commit;
    Commit complete.
    SQL> ALTER TABLE "WWSTO_SESSION$" ENABLE CONSTRAINT "WWSTO_SESS_FK1";
    Table altered.
    G. IMP-00017 - ORACLE error 1 - DBMS_JOB.ISUBMIT
    This error can appear during the import when the import database is not empty and is already customized for some reasons. For example, you export from an infrastructure and you import in a database with a lot of other programs that uses jobs. And unhappily the same job id.
    Due to the way the export/import of jobs is done, the jobs keeps their id after the import. And they may conflict.
    IMP-00017: following statement failed with ORACLE error 1: "BEGIN DBMS_JOB.ISUBMIT(JOB=>42,WHAT=>'begin execute immediate " "''begin wwutl_cache_sys.process_background_inval; end;'' ; exc" "eption when others then wwlog_api.log(p_domain=> ''utl'', " " p_subdomain=>''cache'', p_name=>''background'', " " p_action=>''process_background_inval'', p_information => ''E" "rror in process_background_inval ''|| sqlerrm);end;', NEXT_DATE=" ">TO_DATE('2004-08-19:17:32:16','YYYY-MM-DD:HH24:MI:SS'),INTERVAL=>'SYSDATE " "+ 60/(24*60)',NO_PARSE=>TRUE); END;"
    IMP-00003: ORACLE error 1 encountered ORA-00001: unique constraint (SYS.I_JOB_JOB) violated
    ORA-06512: at "SYS.DBMS_JOB", line 97 ORA-06512: at line 1
    Solutions:
    1. use a fresh installed database,
    2. Due that the jobs conflicting are different because it happens only in custom installation, there is no clear rule. But you can
    recreate the jobs lost after the import with other_ids
    and/or change the job id of the other program before to import. This type of commands can help you (you need to do it with SYS)
    select * from dba_jobs;
    update dba_jobs set job=99 where job=52;
    commit
    2. Import in a RAC environment
    Be aware of the Bug 2479882 when the portal database is in a RAC database.
    Bug 2479882 : NEEDED TO BOUNCE DB NODES AFTER INSTALLING PORTAL 9.0.2 IN RAC NODE3. Intermedia
    After importing a environment, the intermedia indexes are invalid. To correct the error you need to run in SQL*Plus with Portal
    start $MIDTIER_ORACLE_HOME/portal/admin/plsql/wws/inctxgrn.sql
    start $MIDTIER_ORACLE_HOME/portal/admin/plsql/wws/ctxcrind.sql
    But $MIDTIER_ORACLE_HOME/portal/admin/plsql/wws/inctxgrn.sql is missing in IAS 9.0.4.0. This is Bug 3536937. Fixed in 9041. The missing scripts are contained in the downloadable zip file (exp_schema904.zip : Attachment 276688.1:1 ), directory sql. This means that practically in 9040, you have to run
    start sql/inctxgrn.sql
    start $MIDTIER_ORACLE_HOME/portal/admin/plsql/wws/ctxcrind.sql
    In the import script, it is done in the step 6 - recreate Portal Intermedia indexes.
    You can not WA the problem without the scripts. Running ctxcrind.sql alone does not work. You will have this error:
    ORA-06510: PL/SQL: unhandled user-defined exception
    ORA-06512: at "PORTAL.WWERR_API_EXCEPTION", line 164
    ORA-06512: at "PORTAL.WWV_CONTEXT", line 1035
    ORA-06510: PL/SQL: unhandled user-defined exception
    ORA-06512: at "PORTAL.WWERR_API_EXCEPTION", line 164
    ORA-06512: at "PORTAL.WWV_CONTEXT", line 476
    ORA-06510: PL/SQL: unhandled user-defined exception
    ORA-20000: Oracle Text error:
    DRG-12603: CTXSYS does not own user datastore procedure: WWSBR_THING_CTX_69
    ORA-06512: at line 13
    4. ptlconfig
    If you try to run ptlconfig simply after an import you will get an error:
    Problem processing Portal instance: Configuring HTTP server settings : Installing cache data : SQL exception: ERROR: ORA-23421: job number 32 is not a job in the job queue
    This is because the import done by user SYS has imported the PORTAL jobs to the SYS schema in place of portal. The solution is to run
    update dba_jobs set LOG_USER='PORTAL', PRIV_USER='PORTAL' where schema_user='PORTAL';
    In the import script, it is done in the step 8 - post import changes.
    5. WWC-41417 - invalid credentials.
    When you try to login you get:
    Unexpected error encountered in wwsec_app_priv.process_signon (User-Defined Exception) (WWC-41417)
    An exception was raised when accessing the Oracle Internet Directory: 49: Invalid credentials
    Details
    Error:Operation: dbms_ldap.simple_bind_s
    OID host: machine.domain.com
    OID port number: 4032
    Entry DN: orclApplicationCommonName=PORTAL,cn=Portal,cn=Products,cn=OracleContext. (WWC-41743)Solution:
    - run secupoid.sql
    - rerun ptlconfig
    This problem has been seen after using ptlasst in place of ptlconfig.
    6. EXP-003 with a database 9.2.0.5 or 10.1.0.2
    In fact, the DB format of imp/exp has changed in 9.2.0.5 or 10.1.0.2. The EXP-3 error only occurs when the export from the 9.2.0.5.0 or 10.1.0.2.0 database is done with a lower release export utility, e.g. 9.2.0.4.0.
    Due to the way this note is written, the imp/exp utility used is the one of the midtier (9014), if your portal resides in a 9.2.0.5 database, it will not work. To work-around the problem, there are 2 solutions:
    Change the script so that it uses the exp and imp command of database.
    Make a change to the 9.2.0.5 or 10.1.0.2 database to make them compatible with previous version. The change is to modify a database internal view before to export/import the data.
    A work-around is given in Bug 3784697
    1. Make a note of the export definition of exu9tne from
    $OH/rdbms/admin/catexp.sql
    2. Copy this to a new file and add "UNION ALL select * from sys.exu9tneb" to the end of the definition
    3. Run this as sys against the DB to be exported.
    4. Export as required
    5. Put back the original definition of exu9tne
    eg: For 9204 the workaround view would be:
    CREATE OR REPLACE VIEW exu9tne (
    tsno, fileno, blockno, length) AS
    SELECT ts#, segfile#, segblock#, length
    FROM sys.uet$
    WHERE ext# = 1
    UNION ALL
    select * from sys.exu9tneb
    7. EXP-00006: INTERNAL INCONSISTENCY ERROR
    This is Bug 2906613.
    The work-around given in this bug is the following:
    - create the following view, connected as sys, before running export:
    CREATE OR REPLACE VIEW exu8con (
    objid, owner, ownerid, tname, type, cname,
    cno, condition, condlength, enabled, defer,
    sqlver, iname) AS
    SELECT o.obj#, u.name, c.owner#, o.name,
    decode(cd.type#, 11, 7, cd.type#),
    c.name, c.con#, cd.condition, cd.condlength,
    NVL(cd.enabled, 0), NVL(cd.defer, 0),
    sv.sql_version, NVL(oi.name, '')
    FROM sys.obj$ o, sys.user$ u, sys.con$ c,
    sys.cdef$ cd, sys.exu816sqv sv, sys.obj$ oi
    WHERE u.user# = c.owner# AND
    o.obj# = cd.obj# AND
    cd.con# = c.con# AND
    cd.spare1 = sv.version# (+) AND
    cd.enabled = oi.obj# (+) AND
    NOT EXISTS (
    SELECT owner, name
    FROM sys.noexp$ ne
    WHERE ne.owner = u.name AND
    ne.name = o.name AND
    ne.obj_type = 2)
    The modification of exu8con simply adds support for a constraint type that had not previously been supported by this view. There is no negative impact.
    8. WWSBR_DOC_CTX_54 is invalid
    After the recompilation of the package, one package remains invalid (in sys_post_changes.log):
    INVALID_OBJECT_AFTER
    1
    select owner, object_name from all_objects where status='INVALID'
    CTXSYS WWSBR_DOC_CTX_54
    CREATE OR REPLACE procedure WWSBR_DOC_CTX_54
    (rid in rowid, bilob in out NOCOPY blob)
    is begin PORTAL.WWSBR_CTX_PROCS.DOC_CTX(rid,bilob);end;
    This object is not used anymore by portal. The error can be ignored. The procedure can be removed too. This is Bug 3559731.
    9. You do not have permission to perform this operation. (WWC-44131)
    It seems that there are problems if
    - groups on the production machine are not residing in the default place in OID,
    - and that the group creation base and group search base where changed.
    After this, the cloning of the repository work without problem. But it seems that the command 'ptlconfig -dad portal' does not reset the GUID and DN of the groups correctly. I have not checked this yet.
    The solution seems to use the script given in the 9.0.2 Note 228516.1. And run group_sec.sql to reset all the DN and GUID in the copied instance.
    10. Invalid Java objects when exporting from a 9.x database and importing in a 10g database
    If you export from a 9.x database and import in a 10g database, after running utlrp.sql, 18 Java objects will be invalid.
    select object_name, object_type from user_objects where status='INVALID'
    SQL> /
    OBJECT_NAME OBJECT_TYPE
    /556ab159_Handler JAVA CLASS
    /41bf3951_HttpsURLConnection JAVA CLASS
    /ce2fa28e_ProviderManagerClien JAVA CLASS
    /c5b98d35_ServiceManagerClient JAVA CLASS
    /d77cf2ab_SOAPServlet JAVA CLASS
    /649bf254_JavaProvider JAVA CLASS
    /a9164b8b_SpProvider JAVA CLASS
    /2ee43ac9_StatefulEJBProvider JAVA CLASS
    /ad45acec_StatelessEJBProvider JAVA CLASS
    /da1c4a59_EntityEJBProvider JAVA CLASS
    /66fdac3e_OracleSOAPHTTPConnec JAVA CLASS
    /939c36f5_OracleSOAPHTTPConnec JAVA CLASS
    org/apache/soap/rpc/Call JAVA CLASS
    org/apache/soap/rpc/RPCMessage JAVA CLASS
    org/apache/soap/rpc/Response JAVA CLASS
    /198a7089_Message JAVA CLASS
    /2cffd799_ProviderGroupUtils JAVA CLASS
    /32ebb779_ProviderGroupMgrProx JAVA CLASS
    18 rows selected.
    This is a known issue. This can be solved by applying patch one of the following patch depending of your IAS version.
    Bug 3405173 - PORTAL 9.0.4.0.0 PATCH FOR 10G DB UPGRADE (FROM 9.0.X AND 9.2.X)
    Bug 4100409 - PORTAL 9.0.4.1.0 PATCH FOR 10G DB UPGRADE (FROM 9.0.X AND 9.2.X)
    Bug 4100417 - PORTAL 9.0.4.2.0 PATCH FOR 10G DB UPGRADE (FROM 9.0.X AND 9.2.X)
    11. Import : IMP-00003: ORACLE error 30510 encountered
    When importing Portal 9.0.4.x, it could be that the import of the database side produces an error ORA-30510.The new perl script work-around the issue in the portal_post_import.sql script. But not the BASH scripts. If you use the BASH scripts, after the import, please run this command manually in SQL*Plus logged as portal.
    ---- Import error 2 - ORA-30510 when importing
    CREATE OR REPLACE TRIGGER logoff_trigger
    before logoff on schema
    begin
    -- Call wwsec_oid.unbind to close open OID connections if any.
    wwsec_oid.unbind;
    exception
    when others then
    -- Ignore all the errors encountered while unbinding.
    null;
    end logoff_trigger;
    This is logged as <Bug;4458413>.
    12. Exporting from a 9.0.1 database and import in a 9.2.0.5+ or 10g DB
    It could be that when exporting from a 9.0.1 database to a 10g database that the java classes do not get compiled correctly. The following errors are seen
    ORA-29534: referenced object PORTAL.oracle/net/www/proto/https/HttpsURLConnection could not be resolved
    errors:: class oracle/net/www/proto/https/HttpsURLConnection
    ORA-29521: referenced name oracle/security/ssl/OracleSSLSocketFactoryImpl could not be found
    ORA-29521: referenced name oracle/security/ssl/OracleSSLSocketFactory could not be found
    In such a case, please apply the following patches after the import in the 10g database.
    Bug 3405173 PORTAL REPOS DB UPGRADE TO 10G: for Portal 9.0.4.0
    Bug 4100409 PORTAL REPOS DB UPGRADE TO 10G: for Portal 9.0.4.1
    Main Differences with Portal 9.0.2
    For the persons used to this technics in Portal 9.0.2, you could be interested to read the main differences with the same note for Portal 9.0.2
    Portal 9.0.2
    Portal 9.0.4
    Cutter database
    Portal 9.0.2 can be part of an infrastructure database or in a custom external database.
    In Portal 9.0.2, the portal schema is imported in an empty database.
    Portal 9.0.4 can only be installed in a 'Cutter database', a database created with RepCA or OUI containing always OID, DCM and so on...
    In Portal 9.0.4, the portal schema is imported in an 'Cutter database' (new)
    group_sec.sql
    group_sec.sql is used to correct the GUIDs of OID stored in Portal
    ptlconfig -dad portal -oid is used to correct the GUIDs of OID stored in Portal (new)
    1 script
    The import / export are divided by several steps with several scripts
    The import script is done in one step
    Additional steps are included in the script
    This requires to know the hostname and port of the original development machine. (new)
    Import
    The steps are:
    creation of an empty database
    creation of the users with password=username
    import
    The steps are:
    creation of an IAS 10g infrastructure DB (repca or OUI)
    deletion of new portal schemas (new)
    creation of the users with the same password than the schemas just dropped.
    import
    DAD
    The dad needed to be changed
    The passwords are not changed, the dad does not need to be changed.
    Bugs
    In portal 9.0.2, 2 bugs were workarounded by change_host.sh
    In Portal 9.0.4, some tables additional tables needs to be updated manually before to run ptlasst. This is #Bug:3762961#.
    export of LDAP
    The export is done in LDIF files. If the prod and the dev have different domain, it is quite difficult to change the domain name in these file due to the line wrapping at 78 characters.
    The export is done in XML files, in the DSML format (new). It is a lot easier to change the XML files if the domain name is different from PROD to DEV.
    Download
    You have to cut and paste the scripts
    The scripts are attached to the note. Just donwload them.
    Rewiring
    9.0.2 uses ptlasst.
    ptlasst.csh -mode MIDTIER -i custom -s $PORTAL_USER -sp $PORTAL_PASSWORD -c $PORTAL_HOSTNAME:$PORTAL_DB_PORT:$PORTAL_SERVICE_NAME -sdad $PORTAL_DAD -o orasso -op $ORASSO_PASSWORD -odad orasso -host $MIDTIER_HOSTNAME -port $MIDTIER_HTTP_PORT -ldap_h $INFRA_HOSTNAME -ldap_p $OID_PORT -ldap_w $IAS_PASSWORD -pwd $IAS_PASSWORD -sso_c $INFRA_HOSTNAME:$INFRA_DB_PORT:$INFRA_SERVICE_NAME -sso_h $INFRA_HOSTNAME -sso_p $INFRA_HTTP_PORT -ultrasearch -oh $MIDTIER_ORACLE_HOME -mc false -mi true -chost $MIDTIER_HOSTNAME -cport_i $WEBCACHE_INV_PORT -cport_a $WEBCACHE_ADM_PORT -wc_i_pwd $IAS_PASSWORD -emhost $INFRA_HOSTNAME -emport $EM_PORT -pa orasso_pa -pap $ORASSO_PA_PASSWORD -ps orasso_ps -pp $ORASSO_PS_PASSWORD -iasname $IAS_NAME -verbose -portal_only
    9.0.4 uses ptlconfig (new)
    ptlconfig -dad portal
    Environment variables
    A lot of environment variables are needed
    Just 3 environment variables are needed:
    - password of SYS
    - password of IAS,
    - ORACLE_HOME of the Midtier
    All the rest is found in iasconfig.xml and LDAP (new)
    TO DO
    - Check if the orclcommonapplication name fits SID.hostname
    - Check what gives the import of a portal30 upgraded schema inside a schema named portal
    - Explain how to copy the portal*.dbf files in place of export/import and the limitation of tra

  • Error in invoking target 'relink' of makefile precomp//lib/ins_precomp.mk

    Hi Experts,
    I am getting following error while installing oracle at the linking phase
    ORACLE 10.2 /AIX 5.3
    Error in invoking target 'relink' of makefile '/oracle/DNB/102_64/precomp/lib/ins_precomp.mk'.
    See '/oracle/oraInventory/logs/installActions2007-05-15_03-35-04PM.log' for details.
    And the make.log has following entry
    SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid values are between 4 and 915.
    ld: 0711-251 SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid values are between 4 and 915.
    ld: 0711-866 INTERNAL ERROR: Output symbol table size miscalculated.
    ld: 0711-759 INTERNAL ERROR: Source file xoutxx.c, line 470.
    Depending on where this product was acquired, contact your service
    representative or the approved supplier.
    make: The error code from the last command is 16.
    Regards
    Arif AHmed
    919820008732
    make.log
    ld: 0711-870 WARNING: Symbol table entry 17 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 18. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 115 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 251 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-593 SEVERE ERROR: Symbol C_BSTAT (entry 252) in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    The symbol refers to a csect with symbol number 0, which was not
    found. The new symbol cannot be associated with a csect and
    is being ignored.
    ld: 0711-870 WARNING: Symbol table entry 7 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 1. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 8 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 17 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-251 SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid values are between 4 and 915.
    ld: 0711-251 SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid values are between 4 and 915.
    ld: 0711-866 INTERNAL ERROR: Output symbol table size miscalculated.
    ld: 0711-759 INTERNAL ERROR: Source file xoutxx.c, line 470.
    Depending on where this product was acquired, contact your service
    representative or the approved supplier.
    make: The error code from the last command is 16.
    Stop.
    Install Action
    ld: 0711-870 WARNING: Symbol table entry 7 in object /oracle/DNB/102_64/lib//libnnz10.a[pk
    INFO: cs11db.o]:
    Unknown symbol type 1. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 8 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 17 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz1
    INFO: 0.a[pkcs11db.o]:
    Unknown symbol type 18. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 115 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 251 in object /oracle/DNB/102_64/lib//libnnz10.a[pkcs11db.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-593 SEVERE ERROR: Symbol C_BSTAT (entry 252) in object /oracle/DNB
    INFO: /102_64/lib//libnnz10.a[pkcs11db.o]:
    The symbol refers to a csect with symbol number 0, which was not
    found. The new symbol cannot be associated with a csect and
    is being ignored.
    ld: 0711-870 WARNING: Symbol table entry 7 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 1. The symbol is being copied to the output file.
    ld: 0711-870 WARNING: Symbol table entry 8 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 0. The symbol is being copied to the
    INFO: output file.
    ld: 0711-870 WARNING: Symbol table entry 17 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Unknown symbol type 0. The symbol is being copied to the output file.
    ld: 0711-251 SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid values are between 4 and 915.
    ld: 0711-251 SEVERE ERROR: Symbol table entry 18 in object /oracle/DNB/102_64/lib//libnnz10.a[truntim.o]:
    Field n_offset contains 56885504. Valid val
    INFO: ues are between 4 and 915.
    ld: 0711-866 INTERNAL ERROR: Output symbol table size miscalculated.
    INFO: ld: 0711-759 INTERNAL ERROR: Source file xoutxx.c, line 470.
    Depending on where this product was acquired, contact your service
    representative or the approved supplier.
    INFO: make: The error code from the last command is 16.
    Stop.
    INFO: End output from spawned process.
    INFO: -
    INFO: Exception thrown from action: make
    Exception Name: MakefileException
    Exception String: Error in invoking target 'relink' of makefile '/oracle/DNB/102_64/precomp/lib/ins_precomp.mk'. See '/oracle/oraInventory/logs/installActions2007-05-17_11-25-50AM.log' for details.
    Exception Severity: 1
    INFO: Exception handling set to prompt user with options to Retry Ignore
    User Choice: Retry
    INFO: The output of this make operation is also available at: '/oracle/DNB/102_64/install/make.log'
    INFO:
    INFO: Start output from spawned process:
    INFO: -
    INFO:
    INFO: Linking /oracle/DNB/102_64/precomp/lib/proc
    System Information
    <Host>:/oracle/DNB/install/ORA>lsps -a
    Page Space Physical Volume Volume Group Size %Used Active Auto Type
    paging01 hdiskpower4 swapvg 10240MB 1 yes yes lv
    paging00 hdiskpower4 swapvg 10240MB 1 yes yes lv
    hd6 hdisk0 rootvg 2048MB 1 yes yes lv
    <Host>:/oracle/DNB/install/ORA>lsattr -El sys0 -a realmem
    realmem 4194304 Amount of usable physical memory in Kbytes False
    <Host>:/oracle/DNB/install/ORA>lsdev | grep aio
    aio0 Available Asynchronous I/O (Legacy)
    posix_aio0 Available Posix Asynchronous I/O
    <Host>:/oracle/DNB/install/ORA>lsattr -EHl sys0 | grep maxuproc
    maxuproc 2048 Maximum number of PROCESSES allowed per user True
    <Host>:/oracle/DNB/install/ORA>oslevel -s
    5300-05-03
    <Host>:/oracle/DNB/install/ORA>lsuser oradnb
    oradnb id=3861 pgrp=dba groups=dba,oper,sapinst home=/oracle/DNB shell=/bin/csh gecos=SAP Database Administrator login=true su=true rlogin=true daemon=true admin=false sugroups=ALL admgroups= tpath=nosak ttys=ALL expires=0 auth1=SYSTEM auth2=NONE umask=22 registry=files SYSTEM=compat or LDAP logintimes= loginretries=0 pwdwarntime=0 account_locked=false minage=0 maxage=0 maxexpired=-1 minalpha=0 minother=0 mindiff=0 maxrepeats=8 minlen=0 histexpire=0 histsize=0 pwdchecks= dictionlist= fsize=-1 cpu=-1 data=-1 stack=-1 core=8388604 rss=65536 nofiles=2000 fsize_hard=-1 cpu_hard=-1 data_hard=-1 stack_hard=-1 core_hard=8388604 roles=
    <Host>:/oracle/DNB/install/ORA>java -fullversion
    java full version "J2RE 1.4.2 IBM AIX 5L for PowerPC (64 bit JVM) build caix64142-20061124 (SR7)"
    Checking the Software Requirements on AIX
    http://download-east.oracle.com/docs/html/B10811_05/ch2.htm#i1009960
    <Host>:/oracle/<SID>/install/ORA>oslevel -r
    5300-05
    <Host>:/oracle/<SID>/install/ORA>lslpp -l bos.adt.base bos.adt.lib bos.adt.libm bos.perf.perfstat \
    > bos.perf.libperfstat bos.perf.proctools
    Fileset Level State Description
    Path: /usr/lib/objrepos
    bos.adt.base 5.3.0.51 COMMITTED Base Application Development
    Toolkit
    bos.adt.lib 5.3.0.50 COMMITTED Base Application Development
    Libraries
    bos.adt.libm 5.3.0.40 COMMITTED Base Application Development
    Math Library
    bos.perf.libperfstat 5.3.0.50 COMMITTED Performance Statistics Library
    Interface
    bos.perf.perfstat 5.3.0.50 COMMITTED Performance Statistics
    Interface
    bos.perf.proctools 5.3.0.50 COMMITTED Proc Filesystem Tools
    Path: /etc/objrepos
    bos.perf.libperfstat 5.3.0.0 COMMITTED Performance Statistics Library
    Interface
    bos.perf.perfstat 5.3.0.50 COMMITTED Performance Statistics
    Interface
    <Host>:/oracle/<SID>/install/ORA>/usr/sbin/instfix -i -k "IY43980 IY44810 IY45462 IY45707 IY46214 IY46605 \
    > IY48525 IY51801 IY56024"
    There was no data for IY43980 in the fix database.
    There was no data for IY44810 in the fix database.
    There was no data for IY45462 in the fix database.
    There was no data for IY45707 in the fix database.
    There was no data for IY46214 in the fix database.
    There was no data for IY46605 in the fix database.
    All filesets for IY48525 were found.
    There was no data for 5 in the fix database.
    There was no data for IY51801 in the fix database.
    There was no data for IY56024 in the fix database.
    <Host>:/oracle/<SID>/install/ORA>lslpp -l X11.motif.lib mqm.Client.Bnd mqm.Server.Bnd
    Fileset Level State Description
    Path: /usr/lib/objrepos
    X11.motif.lib 5.3.0.50 COMMITTED AIXwindows Motif Libraries
    lslpp: 0504-132 Fileset mqm.Client.Bnd not installed.
    lslpp: 0504-132 Fileset mqm.Server.Bnd not installed.
    <Host>:/oracle/<SID>/install/ORA>lslpp -l bos.rte.aio
    Fileset Level State Description
    Path: /usr/lib/objrepos
    bos.rte.aio 5.3.0.51 COMMITTED Asynchronous I/O Extension
    Path: /etc/objrepos
    bos.rte.aio 5.3.0.0 COMMITTED Asynchronous I/O Extension
    <Host>:/oracle/<SID>/install/ORA>/usr/sbin/instfix -i -k "IY48526 IY47055 IY40034"
    There was no data for IY48526 in the fix database.
    There was no data for IY47055 in the fix database.
    There was no data for IY40034 in the fix database.

    Dear @Prince Jose
    how to enamble CIO?
    maybe your can give a step by step..
    regards,
    RhA

  • Error while compiling BerkeleyDB: Command failed for target `util_sig.lo'

    Hi,
    While trying to compile BerkeleyDB on Solaris 10 x86 ( is there an easier way to get BerkeleyDB running than compiling from source? ) I was confronted with the error message appended.
    I used Berkeley 4.4.20 and configured by executing
    PATH=/usr/ccs/bin:${PATH} # for ar
    cd build_unix
    CC=gcc ../dist/configure
    Here the error message:
    UNIX /app/berkeleydb/db/4.4.20/build_unix
    $ make
    /usr/bin/sh ./libtool --mode=compile gcc -c -I. -I../dist/..  -D_REENTRANT -O2 ../common/util_sig.c
    gcc -c -I. -I../dist/.. -D_REENTRANT -O2 ../common/util_sig.c -fPIC -DPIC -o .libs/util_sig.o
    In file included from /usr/include/sys/signal.h:34,
    from /usr/include/signal.h:26,
    from ../common/util_sig.c:15:
    /usr/include/sys/siginfo.h:259: error: parse error before "ctid_t"
    /usr/include/sys/siginfo.h:292: error: parse error before '}' token
    /usr/include/sys/siginfo.h:294: error: parse error before '}' token
    /usr/include/sys/siginfo.h:390: error: parse error before "ctid_t"
    /usr/include/sys/siginfo.h:392: error: conflicting types for `__proc'
    /usr/include/sys/siginfo.h:261: error: previous declaration of `__proc'
    /usr/include/sys/siginfo.h:398: error: conflicting types for `__fault'
    /usr/include/sys/siginfo.h:267: error: previous declaration of `__fault'
    /usr/include/sys/siginfo.h:404: error: conflicting types for `__file'
    /usr/include/sys/siginfo.h:273: error: previous declaration of `__file'
    /usr/include/sys/siginfo.h:420: error: conflicting types for `__prof'
    /usr/include/sys/siginfo.h:287: error: previous declaration of `__prof'
    /usr/include/sys/siginfo.h:424: error: conflicting types for `__rctl'
    /usr/include/sys/siginfo.h:291: error: previous declaration of `__rctl'
    /usr/include/sys/siginfo.h:426: error: parse error before '}' token
    /usr/include/sys/siginfo.h:428: error: parse error before '}' token
    /usr/include/sys/siginfo.h:432: error: parse error before "k_siginfo_t"
    /usr/include/sys/siginfo.h:437: error: parse error before '}' token
    In file included from /usr/include/signal.h:26,
    from ../common/util_sig.c:15:
    /usr/include/sys/signal.h:85: error: parse error before "siginfo_t"
    In file included from ../common/util_sig.c:15:
    /usr/include/signal.h:111: error: parse error before "siginfo_t"
    /usr/include/signal.h:113: error: parse error before "siginfo_t"
    *** Error code 1
    make: Fatal error: Command failed for target `util_sig.lo'
    Thanks, Dietrich
    PS:
    I need the LDAP server for playing with a JAAS LDAP login module and something minimal would be enough. Is there something easier to install than OpenLDAP?

    I fail to understand why you're taking all this trouble. Both the Berkeley DB as OpenLDAP are both shipped with Solaris 10/x86 natively. Look into the so called companion cd / dvd. That makes setting all of this up a piece of cake.
    Another option could be looking into Blastwave or Sun Freeware.

  • 'error in invoking target 'client_sharedlib' of makefile'   for help

    I am installing oracle 10g(10202_database_soxl86)on solaris 10 (sol-10-u10-ga2-x86-dvd.iso) throgh VMware workstation 7.1.4
    I fellowing the steps below:
    1、 # pkgadd -R /cdrom/Solaris_10/Product SUNWarc SUNWbtool SUNWhea SUNWlibm SUNWlibms SUNWsprot SUNWtoo SUNWi1of SUNWi1cs SUNWi15cs SUNWxwfnt SUNWgcc
    2、 # mkdir -p /export/home/app/oracle
    # chown -R oracle:oinstall /export/home/app/oracle
    3、 # groupadd oinstall
    # groupadd dba
    # useradd -g oinstall -G dba -s /usr/bin/bash -d /export/home/app/oracle oracle
    # passwd oracle
    4、$vi .profile
    export ORACLE_BASE=/export/home/app
    export ORACLE_HOME=$ORACLE_BASE/oracle
    export ORACLE_SID=sun
    export PATH=$ORACLE_HOME/bin:$PATH
    export DISPLAY=192.168.0.1:0.0
    stty erase ^h
    $ env | grep ORA
    results: ORACLE_BASE=/export/home/app
    ORACLE_HOME=/export/home/app/oracle
    ORACLE_SID=sun
    $ env | grep DIS
    results:DISPLAY=192.168.0.1:0.0
    5、# uname -a
    result:SunOS solaris10 5.10 Generic_147441-01 i86pc i386 i86pc
    # isainfo -v
    result:32-bit i386 applications
    pclmulqdq aes sse4.2 sse4.1 ssse3 popcnt tscp sse3 sse2 sse fxsr mmx
    cmov sep cx8 tsc fpu
    6、# df -k
    Filesystem kbytes used avail capacity Mounted on
    swap 2437932 112 2437820 1% /tmp
    7 、# cd /tmp
    # unzip / 10202_database_soxl86.zip
    # chown -R oracle:oinstall database/
    8、 # su - oracle
    $ cd /tmp/database
    $ ./runInstaller
    Error information:
    1、error in invoking target 'client_sharedlib' of makefile
    2、Error in invoking target 'relink' of makefile
    3、Error in invoking target 'ioracle' of makefile '/export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk'.
    4、Error in invoking target 'nnfgt.o mkldflags client_sharedlib' of makefile '/export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk'
    5、Error in invoking target 'install' of makefile '/export/home/oracle/product/10.2.2/sqlplus/lib/ins_sqlplus.mk'.
    6、Error in invoking target 'irman' of makefile '/export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk'.
    7、Error in invoking target 'install' of makefile '/export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk'.
    Error in invoking target 'install' of makefile '/export/home/oracle/product/10.2.2/plsql/lib/ins_plsql.mk'.
    8、Error in invoking target 'ioklist iokinit iokdstry' of makefile '/export/home/oracle/product/10.2.2/network/lib/ins_nau.mk'.
    9、Error in invoking target 'utilities ctx_on' of makefile '/export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk'.
    10、Error in invoking target 'agent nmo nmb' of makefile '/export/home/oracle/product/10.2.2/sysman/lib/ins_sysman.mk'.
    11、Error in invoking target 'clientonlyinstall' of makefile '/export/home/oracle/product/10.2.2/ldap/lib/ins_ldap.mk'.
    12、Error in invoking target 'install' of makefile '/export/home/oracle/product/10.2.2/srvm/lib/ins_srvm.mk'.
    13、Error in invoking target 'racg_install' of makefile '/export/home/oracle/product/10.2.2/racg/lib/ins_has.mk'.
    14、Error in invoking target 'install' of makefile '/export/home/oracle/product/10.2.2/network/lib/ins_net_server.mk'.
    15、Error in invoking target 'all_no_orcl ipc_g' of makefile '/export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk'.
    16、Error in invoking target 'collector' of makefile '/export/home/oracle/product/10.2.2/sysman/lib/ins_emdb.mk'.
    details:
    Checking installer requirements...
    INFO:
    INFO: Checking operating
    system version: must be 5.10. Actual 5.10
    INFO:
    Passed
    INFO:
    INFO: Checking Temp space: must be
    greater than 250 MB. Actual 2403 MB Passed
    INFO: Checking swap
    space: must be greater than 500 MB. Actual 2805 MB Passed
    INFO:
    Checking monitor: must be configured to display at least 256 colors.
    Actual 16777216 Passed
    INFO:
    INFO: All installer requirements
    met.
    Exception String: Error
    in invoking target 'client_sharedlib' of makefile
    '/export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk'.
    See '/export/home/oracle/oraInventory/logs/installActions2011-11-
    04_08-21-57PM.log' for details.
    Exception Severity: 1
    INFO: Exception
    handling set to prompt user with options to Retry Ignore
    User
    Choice: Ignore
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link
    Phase of component Enterprise Manager Agent DB
    INFO: 11/4/11 8:37:54
    PM CST: Starting install Link Phase of component Enterprise Manager
    Common Files
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link
    Phase of component Oracle Display Fonts
    INFO: 11/4/11 8:37:54 PM CST:
    Starting install Link Phase of component Oracle Extended Windowing
    Toolkit
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase of
    component Oracle UIX
    INFO: 11/4/11 8:37:54 PM CST: Starting install
    Link Phase of component SQL*Plus Required Support Files
    INFO: 11/4/11
    8:37:54 PM CST: Starting install Link Phase of component DBJAVA
    Required Support Files
    INFO: 11/4/11 8:37:54 PM CST: Starting install
    Link Phase of component Oracle OLAP RDBMS Files
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Unlinking oracle
    INFO: Calling Action unixActions10.2.0.1.0 make
    installMakePath = /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = olap_on
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = olap_off
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    null
    INFO: The output of this make operation is also available at:
    '/export/home/oracle/product/10.2.2/install/make.log'
    INFO:
    INFO:
    Start output from spawned process:
    INFO:
    INFO:
    INFO: /usr/ccs/bin/ar d
    /export/home/oracle/product/10.2.2/rdbms/lib/libknlopt.a xsnoolap.o
    INFO: /usr/ccs/bin/ar cr
    /export/home/oracle/product/10.2.2/rdbms/lib/libknlopt.a
    /export/home/oracle/product/10.2.2/rdbms/lib/xsyeolap.o
    INFO: End
    output from spawned process.
    INFO:
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking Oracle
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link
    Phase of component Oracle OLAP API
    INFO: 11/4/11 8:37:54 PM CST:
    Starting install Link Phase of component XDK Required Support Files
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase of
    component RDBMS Required Support Files for Instant Client
    INFO:
    11/4/11 8:37:54 PM CST: Starting install Link Phase of component
    RDBMS Required Support Files
    INFO: 11/4/11 8:37:54 PM CST: Starting
    install Link Phase of component Oracle JFC Extended Windowing
    Toolkit
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase of
    component Agent Required Support Files
    INFO: 11/4/11 8:37:54 PM CST:
    Starting install Link Phase of component Parser Generator Required
    Support Files
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link
    Phase of component Sample Schema Data
    INFO: 11/4/11 8:37:54 PM CST:
    Starting install Link Phase of component Oracle Starter Database
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase of
    component Oracle Message Gateway Common Files
    INFO: 11/4/11 8:37:54
    PM CST: Starting install Link Phase of component XML Parser for
    Oracle JVM
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase
    of component Precompiler Required Support Files
    INFO: 11/4/11 8:37:54
    PM CST: Starting install Link Phase of component XML Parser for
    Java
    INFO: 11/4/11 8:37:54 PM CST: Starting install Link Phase of
    component Oracle Help for the Web
    INFO: 11/4/11 8:37:54 PM CST:
    Starting install Link Phase of component Oracle Ice Browser
    INFO:
    11/4/11 8:37:54 PM CST: Starting install Link Phase of component
    Enterprise Manager Baseline
    INFO: 11/4/11 8:37:54 PM CST: Starting
    install Link Phase of component Oracle Data Mining RDBMS Files
    INFO:
    Calling Action unixActions10.2.0.1.0 registerForMake
    installMakePath = /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Unlinking oracle
    INFO: Calling Action unixActions10.2.0.1.0 make
    installMakePath = /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = dm_on
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = dm_off
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking Label Security Option
    INFO: The output of this make operation
    is also available at:
    '/export/home/oracle/product/10.2.2/install/make.log'
    INFO:
    INFO:
    Start output from spawned process:
    INFO:
    INFO:
    INFO: /usr/ccs/bin/ar cr
    /export/home/oracle/product/10.2.2/rdbms/lib/libknlopt.a
    /export/home/oracle/product/10.2.2/rdbms/lib/dmndmse.o      
    INFO:
    /usr/ccs/bin/ar cr
    /export/home/oracle/product/10.2.2/rdbms/lib/libknlopt.a
    /export/home/oracle/product/10.2.2/rdbms/lib/dmwdm.o
    INFO: End
    output from spawned process.
    INFO:
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Unlinking oracle
    INFO: 11/4/11 8:37:55 PM CST: Starting install Link
    Phase of component Oracle Database 10g interMedia Files
    INFO: 11/4/11
    8:37:55 PM CST: Starting install Link Phase of component Oracle
    interMedia Java Advanced Imaging
    INFO: 11/4/11 8:37:55 PM CST:
    Starting install Link Phase of component SQLJ Runtime
    INFO: 11/4/11
    8:37:55 PM CST: Starting install Link Phase of component Oracle
    interMedia Annotator
    INFO: 11/4/11 8:37:55 PM CST: Starting install
    Link Phase of component Oracle JDBC/OCI Instant Client
    INFO: 11/4/11
    8:37:55 PM CST: Starting install Link Phase of component Oracle
    interMedia Locator RDBMS Files
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking oracle
    INFO: Calling Action unixActions10.2.0.1.0 make
    installMakePath = /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = sdo_on
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = sdo_off
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking sdo options
    INFO: The output of this make operation is also
    available at: '/export/home/oracle/product/10.2.2/install/make.log'
    INFO:
    INFO: Start output from spawned process:
    INFO:
    INFO:
    INFO: /usr/ccs/bin/ar cr
    /export/home/oracle/product/10.2.2/rdbms/lib/libknlopt.a
    /export/home/oracle/product/10.2.2/rdbms/lib/kxmwsd.o
    INFO: End
    output from spawned process.
    INFO:
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = ioracle
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    undoTarget = ioracle
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking Oracle
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link
    Phase of component Platform Required Support Files
    INFO: 11/4/11
    8:37:56 PM CST: Starting install Link Phase of component Oracle Core
    Required Support Files
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component Database Workspace Manager
    INFO: 11/4/11
    8:37:56 PM CST: Starting install Link Phase of component Oracle
    Containers for Java
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component Oracle Help For Java
    INFO: 11/4/11 8:37:56 PM
    CST: Starting install Link Phase of component JDBC Common Files
    INFO:
    11/4/11 8:37:56 PM CST: Starting install Link Phase of component
    Perl Interpreter
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link
    Phase of component Oracle Code Editor
    INFO: 11/4/11 8:37:56 PM CST:
    Starting install Link Phase of component Oracle Notification
    Service
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link Phase of
    component Oracle interMedia Client Option
    INFO: 11/4/11 8:37:56 PM
    CST: Starting install Link Phase of component Oracle JDBC Thin
    Driver for JDK 1.2
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component Oracle JDBC Thin Driver for JDK 1.4
    INFO:
    11/4/11 8:37:56 PM CST: Starting install Link Phase of component
    Enterprise Manager plugin Common Files
    INFO: 11/4/11 8:37:56 PM CST:
    Starting install Link Phase of component LDAP Required Support
    Files
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link Phase of
    component Character Set Migration Utility
    INFO: 11/4/11 8:37:56 PM
    CST: Starting install Link Phase of component Oracle Locale Builder
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link Phase of
    component PL/SQL Embedded Gateway
    INFO: 11/4/11 8:37:56 PM CST:
    Starting install Link Phase of component OLAP SQL Scripts
    INFO:
    11/4/11 8:37:56 PM CST: Starting install Link Phase of component
    Database SQL Scripts
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component Oracle Globalization Support
    INFO: 11/4/11
    8:37:56 PM CST: Starting install Link Phase of component Required
    Support Files
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link
    Phase of component Secure Socket Layer
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk
    installTarget = nnfgt.o
         installArguments = ORACLE_HOME
    =/export/home/oracle/product/10.2.2
         undoMakeFileName = null
    undoTarget = null
         undoArguments = null
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking nnfgt
    INFO: Calling Action unixActions10.2.0.1.0
    registerForMake
         installMakePath = /usr/ccs/bin/make
    installMakeFileName =
    /export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk
    installTarget = mkldflags
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    null
         undoTarget = null
         undoArguments = null
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking ldflags
    INFO: Calling Action unixActions10.2.0.1.0
    registerForMake
         installMakePath = /usr/ccs/bin/make
    installMakeFileName =
    /export/home/oracle/product/10.2.2/network/lib/ins_net_client.mk
    installTarget = client_sharedlib
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    null
         undoTarget = null
         undoArguments = null
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Building client shared library
    INFO: 11/4/11 8:37:56 PM CST: Starting
    install Link Phase of component Oracle Database User Interface
    INFO:
    11/4/11 8:37:56 PM CST: Starting install Link Phase of component
    Enterprise Manager Minimal Integration
    INFO: 11/4/11 8:37:56 PM CST:
    Starting install Link Phase of component Oracle Wallet Manager
    INFO:
    11/4/11 8:37:56 PM CST: Starting install Link Phase of component
    Oracle Clusterware RDBMS Files
    INFO: 11/4/11 8:37:56 PM CST: Starting
    install Link Phase of component Oracle RAC Required Support Files-
    HAS
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link Phase of
    component SQL*Plus
    INFO: Calling Action unixActions10.2.0.1.0
    registerForMake
         installMakePath = /usr/ccs/bin/make
    installMakeFileName =
    /export/home/oracle/product/10.2.2/sqlplus/lib/ins_sqlplus.mk
    installTarget = install
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    null
         undoTarget = null
         undoArguments = null
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking SQLPLUS executables
    INFO: 11/4/11 8:37:56 PM CST: Starting
    install Link Phase of component Enterprise Manager plugin Common
    Files
    INFO: 11/4/11 8:37:56 PM CST: Starting install Link Phase of
    component iSQL*Plus
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component Oracle Recovery Manager
    INFO: Calling Action
    unixActions10.2.0.1.0 registerForMake
         installMakePath =
    /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/rdbms/lib/ins_rdbms.mk
    installTarget = irman
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    null
         undoTarget = null
         undoArguments = null
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking RMAN utility
    INFO: 11/4/11 8:37:56 PM CST: Starting install
    Link Phase of component HAS Common Files
    INFO: 11/4/11 8:37:56 PM
    CST: Starting install Link Phase of component Precompiler Common
    Files
    INFO: Calling Action unixActions10.2.0.1.0 make
    installMakePath = /usr/ccs/bin/make
         installMakeFileName =
    /export/home/oracle/product/10.2.2/precomp/lib/ins_precomp.mk
    installTarget = links
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         undoMakeFileName =
    /export/home/oracle/product/10.2.2/precomp/lib/ins_precomp.mk
    undoTarget = rmlinks
         undoArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2
         logFile =
    /export/home/oracle/product/10.2.2/install/make.log
         progMsg =
    Linking links
    INFO: The output of this make operation is also
    available at: '/export/home/oracle/product/10.2.2/install/make.log'
    INFO:
    INFO: Start output from spawned process:
    INFO:
    INFO:
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; rm -f SQLCA.H)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public ; rm -f
    ORACA.H)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public
    ; rm -f SQLDA.H)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; rm -f SQLCA.COB)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public ; rm -f
    ORACA.COB)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; rm -f SQLCA.FOR)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public ; rm -f
    ORACA.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; rm -f
    BNDDSC.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; rm -f
    SELDSC.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s sqlca.h
    SQLCA.H)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public
    ; ln -s oraca.h ORACA.H)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s sqlda.h
    SQLDA.H)
    INFO: (cd /export/home/oracle/product/10.2.2/precomp/public
    ; ln -s sqlca.cob SQLCA.COB)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s oraca.cob
    ORACA.COB)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s sqlca.for
    SQLCA.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s oraca.for
    ORACA.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s bnddsc.for
    BNDDSC.FOR)
    INFO: (cd
    /export/home/oracle/product/10.2.2/precomp/public ; ln -s seldsc.for
    SELDSC.FOR)
    INFO: End output from spawned process.
    INFO:
    INFO: Calling Action
    unixActions10.2.0.1.0 make
         installMakePath = /usr/ccs/bin/make
    installMakeFileName =
    /export/home/oracle/product/10.2.2/precomp/lib/ins_precomp.mk
    installTarget = relink
         installArguments =
    ORACLE_HOME=/export/home/oracle/product/10.2.2,EXENAME=proc
    undoMakeFileName = null
         undoTarget = null
         undoArguments = null
    logFile = /export/home/oracle/product/10.2.2/install/make.log
    progMsg = Linking proc
    INFO: The output of this make operation is
    also available at:
    '/export/home/oracle/product/10.2.2/install/make.log'
    INFO:
    INFO:
    Start output from spawned process:
    INFO:
    INFO:
    INFO: Linking
    /export/home/oracle/product/10.2.2/precomp/lib/proc
    INFO: ld:
    INFO:
    fatal:
    INFO: library -l
    INFO: clntsh
    INFO: : not found
    INFO:
    INFO: ld:
    INFO: fatal:
    INFO: file processing errors. No output written to
    INFO: /export/home/oracle/product/10.2.2/precomp/lib/proc
    INFO:
    INFO:
    *** Error code 1
    The following command caused the error:
    echo "Linking
    /export/home/oracle/product/10.2.2/precomp/lib/proc"; \
    rm -f
    /export/home/oracle/product/10.2.2/precomp/lib/proc; \
    /usr/ccs/bin/ld -o
    /export/home/oracle/product/10.2.2/precomp/lib/proc -
    L/export/home/oracle/product/10.2.2/precomp/lib/ -
    L/export/home/oracle/product/10.2.2/lib/ -dy
    /export/home/oracle/product/10.2.2/lib/prod/lib/crti.o
    /export/home/oracle/product/10.2.2/lib/prod/lib/crt1.o
    /export/home/oracle/product/10.2.2/lib/s0m
    INFO: ain.o
    /export/home/oracle/product/10.2.2/precomp/lib/pdc.o
    /export/home/oracle/product/10.2.2/precomp/lib/pds.o
    /export/home/oracle/product/10.2.2/precomp/lib/libproc2.a
    /export/home/oracle/product/10.2.2/lib/libslax10.a
    /export/home/oracle/product/10.2.2/precomp/lib/libpgp.a -lpls10 -
    lplp10 /export/home/oracle/product/10.2.2/lib/libplc10.a -lsnls10 -
    lunls10 -lnls10
    /export/home/oracle/product/10.2.2/rdbms/lib/kpudfo.o -lclntsh `cat
    /export/home/oracle/product/10.2.2/lib/ldflags` -lnsslb10 -
    lncrypt10
    INFO: -lnsgr10 -lnzjs10 -ln10 -lnnz10 -lnl10 -lnro10 `cat
    /export/home/oracle/product/10.2.2/lib/ldflags` -lnsslb10 -
    lncrypt10 -lnsgr10 -lnzjs10 -ln10 -lnnz10 -lnl10 -lclient10 -
    lnnetd10 -lvsn10 -lcommon10 -lgeneric10 -lmm -lsnls10 -lnls10 -
    lcore10 -lsnls10 -lnls10 -lcore10 -lsnls10 -lnls10 -lxml10 -
    lcore10 -lunls10 -lsnls10 -lnls10 -lcore10 -lnls10 `cat
    /export/home/oracle/product/10.2.2/lib/ldflags` -lnsslb10 -
    lncrypt10 -lnsgr10 -lnzjs10 -ln10 -lnnz10 -lnl10 -lnro10 `cat
    /export/home/oracle/product/1
    INFO: 0.2.2/lib/ldflags` -lnsslb10
    -lncrypt10 -lnsgr10 -lnzjs10 -ln10 -lnnz10 -lnl10 -lclient10 -
    lnnetd10 -lvsn10 -lcommon10 -lgeneric10 -lpls10 -lsnls10 -lnls10
    -lcore10 -lsnls10 -lnls10 -lcore10 -lsnls10 -lnls10 -lxml10 -
    lcore10 -lunls10 -lsnls10 -lnls10 -lcore10 -lnls10 -lclient10 -
    lnnetd10 -lvsn10 -lcommon10 -lgeneric10 -lsnls10 -lnls10 -lcore10
    -lsnls10 -lnls10 -lcore10 -lsnls10 -lnls10 -lxml10 -lcore10 -
    lunls10 -lsnls10 -lnls10 -lcore10 -lnls10 `cat
    /export/home/oracle/product/10.2.2/lib/sy
    INFO: sliblist` -R
    /opt/SUNWcluster/lib:/export/home/oracle/product/10.2.2/lib:/opt/ORC
    Lcluster/lib/ -Y
    P,:/opt/SUNWcluster/lib:/opt/ORCLcluster/lib/:/usr/ccs/lib:/usr/lib
    -Qy -lc -lrt -laio -lposix4 -lm
    /export/home/oracle/product/10.2.2/lib/prod/lib/crtn.o; \
    /bin/chmod
    755 /export/home/oracle/product/10.2.2/precomp/lib/proc
    INFO: make:
    Fatal error: Command failed for target
    `/export/home/oracle/product/10.2.2/precomp/lib/proc'
    INFO: End
    output from spawned process.
    INFO:
    INFO: Exception thrown from
    action: make
    Exception Name: MakefileException

    895824 wrote:
    I am installing oracle 10g(10202_database_soxl86)on solaris 10 (sol-10-u10-ga2-x86-dvd.iso) throgh VMware workstation 7.1.4
    I fellowing the steps below:
    1、 # pkgadd -R /cdrom/Solaris_10/Product SUNWarc SUNWbtool SUNWhea SUNWlibm SUNWlibms SUNWsprot SUNWtoo SUNWi1of SUNWi1cs SUNWi15cs SUNWxwfnt SUNWgcc
    2、 # mkdir -p /export/home/app/oracle
    # chown -R oracle:oinstall /export/home/app/oracle
    3、 # groupadd oinstall
    # groupadd dba
    # useradd -g oinstall -G dba -s /usr/bin/bash -d /export/home/app/oracle oracle
    # passwd oracle
    4、$vi .profile
    export ORACLE_BASE=/export/home/app
    export ORACLE_HOME=$ORACLE_BASE/oracle
    export ORACLE_SID=sun
    export PATH=$ORACLE_HOME/bin:$PATH
    export DISPLAY=192.168.0.1:0.0
    stty erase ^h
    $ env | grep ORA
    results: ORACLE_BASE=/export/home/app
    ORACLE_HOME=/export/home/app/oracle
    ORACLE_SID=sun
    $ env | grep DIS
    results:DISPLAY=192.168.0.1:0.0
    5、# uname -a
    result:SunOS solaris10 5.10 Generic_147441-01 i86pc i386 i86pc
    # isainfo -v
    result:32-bit i386 applications
    pclmulqdq aes sse4.2 sse4.1 ssse3 popcnt tscp sse3 sse2 sse fxsr mmx
    cmov sep cx8 tsc fpu
    6、# df -k
    Filesystem kbytes used avail capacity Mounted on
    swap 2437932 112 2437820 1% /tmp
    7 、# cd /tmp
    # unzip / 10202_database_soxl86.zip
    # chown -R oracle:oinstall database/
    8、 # su - oracleLook no further. You modified your .profile to set the environment, but you did that as root, then did an su to oracle. That oracle session never saw the environment changes (ORACLE_HOME, etc) you did to your (root) profile. All Oracle specific tasks should be done logged on as the oracle software owner - oracle. Once root made the necessary prereqs at the OS level (pkgadd, groupadd, useradd) you should have logged off as root and logged on as oracle, THEN proceeded to fix your .profile, etc., and run the installer. NEVER act as a dba while logged on as root. NEVER log on as root to do ANYTHING that does not REQUIRE root access.
    $ cd /tmp/database
    $ ./runInstaller
    <snip>

  • How to only synchronize one specific LDAP user group with SAP?

    Hi,
    Hopefully this is the correct forum to post this in. I want to have continuous one-way synchronization of users from my LDAP server to my SAP central system. I've started configure in SAP using transaction SM59 and LDAP. Can I somewhere set that only one specific LDAP user group shall be transferred to SAP (they do not need to be assigned to any specific group, profile, role in SAP) - or should this be done on the LDAP server side (or is it at all possible)?
    Correct me if I'm wrong, but the User Group field in the report RSLDAPSYNC_USER only concerns SAP user groups right? This would therefore not be sufficient since I want to select the users to synchronize based on user groups in the directory.
    Thanks, Oscar

    We've used a repository constant to specify the LDAP filter for reading users / groups from the LDAP target.
    E.g. LDAP_FILTER_USERS (&(objectCategory=person)(objectClass=user))
    Then we also have a constant for the LDAP_STARTING_POINT
    For our AD Group Initial Load we filter according to these settings:
    LDAP_FILTER_GROUPS = (objectclass=group)
    LDAP_STARTING_POINT_GROUPS = ou=IDMManagedGroups,ou=Groups,dc=cfstest,dc=le,dc=ac,dc=uk
    The above example only reads AD groups starting at the specified OU
    Then in a Job From LDAP Pass the LDAP URL looks like this:
    LDAP://%$rep.LDAP_HOST%:%$rep.LDAP_PORT%/%$rep.LDAP_STARTING_POINT_GROUPS%?*?SUB?%$rep.LDAP_FILTER_GROUPS%
    I hope this helps
    Paul

  • Error while doing the Ldap sync for UDFs

    Hi All,
    I am doing LDAP sync for UDFs,
    Created users in OID.
    assigned to orclIDXPerson object modified the ldapconfig.props and created the input file.
    Now I am running the ldapsyncudf.sh then I getting the below error.
    Exception in thread "main" java.lang.NullPointerException
    at oracle.ods.virtualization.schema.AttributeTypeDefinition.getOID(AttributeTypeDefinition.java:117)
    at oracle.ods.virtualization.jndi.OVDSchemaContext.convertAttrDefnToJNDIAttrs(OVDSchemaContext.java:655)
    at oracle.ods.virtualization.jndi.OVDSchemaContext.getAttributes(OVDSchemaContext.java:137)
    at oracle.ods.virtualization.jndi.OVDSchemaContext.getAttributes(OVDSchemaContext.java:109)
    at oracle.iam.configservice.impl.LDAPUDFSyncImpl.isAttrExistsInLDAP(LDAPUDFSyncImpl.java:555)
    at oracle.iam.configservice.impl.LDAPUDFSyncImpl.validateOVDSchema(LDAPUDFSyncImpl.java:519)
    at oracle.iam.configservice.impl.LDAPUDFSyncImpl.addUDFwithLDAP(LDAPUDFSyncImpl.java:1082)
    at oracle.iam.configservice.api.LDAPUDFSyncEJB.addUDFwithLDAPx(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at com.bea.core.repackaged.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:310)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.jee.spi.MethodInvocationVisitorImpl.visit(MethodInvocationVisitorImpl.java:37)
    at weblogic.ejb.container.injection.EnvironmentInterceptorCallbackImpl.callback(EnvironmentInterceptorCallbackImpl.java:54)
    at com.bea.core.repackaged.springframework.jee.spi.EnvironmentInterceptor.invoke(EnvironmentInterceptor.java:50)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:89)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
    at $Proxy631.addUDFwithLDAPx(Unknown Source)
    can anyone please unblock me.
    Thanks,
    Valli

    Hi,
    Please see if these help (for 11gR2)
    Export the LDAPUser.xml file from MDS using weblogicExportMetatdata.bat. This xml contains the attributes mapping between OIM and OID for LDAP synchronization.
    Include the entry for OIM attribute (if entry does not exist for the attribute in the XML) under entity-attributes node. For e.g. use the following xml snippet to add the entry for ISD Code for Phone attribute
    <entity-attributes><attribute name=”ISD Code for Phone”> <type>string</type> <required>false</required> <attribute-group>Extended </attribute-group> <searchable>true</searchable> </attribute> </entity-attributes>
    Include the entry for OID attribute under target-fields node. For e.g. use the following xml snippet to add the entry for CountryCode
    <target-fields><field name=”CountryCode”><type>String</type> <required>false</required> </target-fields>
    Now map the OIM attribute with the OID attribute using the following xml snippet under attribute-maps node
    <attribute-maps><attribute-map> <entity-attribute> ISD Code for Phone </entity-attribute> <target-field>CountryCode</target-field> </attribute-map></attribute-maps>
    Save the changes and import the file back into MDS using WebLogic import utilities.

Maybe you are looking for