Execution Plan Failed for Full Load

Hi Team,
When i run the full Load the execution plan failed and verified the log and found below information from SEBL_VERT_8_1_1_FLATFILE.DATAWAREHOUSE.SIL_Vert.SIL_InsertRowInRunTable
DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
DIRECTOR> VAR_27028 Use override value [SEBL_VERT_8_1_1_FLATFILE.DATAWAREHOUSE.SIL_Vert.SIL_InsertRowInRunTable.log] for session parameter:[$PMSessionLogFile].
DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[mplt_SIL_InsertRowInRunTable.$$DATASOURCE_NUM_ID].
DIRECTOR> VAR_27028 Use override value [21950495] for mapping parameter:[MPLT_GET_ETL_PROC_WID.$$ETL_PROC_WID].
DIRECTOR> TM_6014 Initializing session [SIL_InsertRowInRunTable] at [Mon Sep 26 15:53:45 2011].
DIRECTOR> TM_6683 Repository Name: [infa_rep]
DIRECTOR> TM_6684 Server Name: [infa_service]
DIRECTOR> TM_6686 Folder: [SIL_Vert]
DIRECTOR> TM_6685 Workflow: [SIL_InsertRowInRunTable] Run Instance Name: [] Run Id: [8]
DIRECTOR> TM_6101 Mapping name: SIL_InsertRowInRunTable [version 1].
DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
DIRECTOR> TM_6827 [H:\Informatica901\server\infa_shared\Storage] will be used as storage directory for session [SIL_InsertRowInRunTable].
DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
DIRECTOR> TM_6703 Session [SIL_InsertRowInRunTable] is run by 32-bit Integration Service [node01_eblnhif-czc80685], version [9.0.1 HotFix2], build [1111].
MANAGER> PETL_24058 Running Partition Group [1].
MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
MANAGER> PETL_24001 Parallel Pipeline Engine running.
MANAGER> PETL_24003 Initializing session run.
MAPPING> CMN_1569 Server Mode: [ASCII]
MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
MAPPING> TM_6151 The session sort order is [Binary].
MAPPING> TM_6156 Using low precision processing.
MAPPING> TM_6180 Deadlock retry logic will not be implemented.
MAPPING> TM_6187 Session target-based commit interval is [10000].
MAPPING> TM_6307 DTM error log disabled.
MAPPING> TE_7022 TShmWriter: Initialized
MAPPING> DBG_21075 Connecting to database [Connect_to_OLAP], user [OBAW]
MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
MAPPING> CMN_1022 Database driver error...
CMN_1022 [
Database driver error...
Function Name : Logon
ORA-12154: TNS:could not resolve service name
Database driver error...
Function Name : Connect
Database Error: Failed to connect to database using user [OBAW] and connection string [Connect_to_OLAP].]
MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
MAPPING> CMN_1076 ERROR creating database connection.
MAPPING> DBG_21520 Transform : LKP_W_PARAM_G_Get_ETL_PROC_WID, connect string : Relational:DataWarehouse
MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
MAPPING> TE_7017 Internal error. Failed to initialize transformation [MPLT_GET_ETL_PROC_WID.LKP_ETL_PROC_WID]. Contact Informatica Global Customer Support.
MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
MAPPING> TM_6006 Error initializing DTM for session [SIL_InsertRowInRunTable].
MANAGER> PETL_24005 Starting post-session tasks. : (Mon Sep 26 15:53:45 2011)
MANAGER> PETL_24029 Post-session task completed successfully. : (Mon Sep 26 15:53:45 2011)
MAPPING> TM_6018 The session completed with [0] row transformation errors.
MANAGER> PETL_24002 Parallel Pipeline Engine finished.
DIRECTOR> PETL_24013 Session run completed with failure.
DIRECTOR> TM_6022
SESSION LOAD SUMMARY
================================================
DIRECTOR> TM_6252 Source Load Summary.
DIRECTOR> CMN_1740 Table: [SQ_FILE_DUAL] (Instance Name: [SQ_FILE_DUAL])
     Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
DIRECTOR> TM_6253 Target Load Summary.
DIRECTOR> TM_6023
===================================================
DIRECTOR> TM_6020 Session [SIL_InsertRowInRunTable] completed at [Mon Sep 26 15:53:46 2011].
I checked physical datasource connection in DAC and workflow manager and tested and verified but still facing same issue. I connected through oracle merant ODBC driver.
Pls. let me know the solution for this error
Regards,
VSR

Hi,
Did you try using Oracle 10g/11g drivers as datasource? If not I think you should try that but before that ensure that from your DAC server box you are able to tnsping the OLAP database. Hope this helps
If this is helpful mark is helpful
Regards,
BI Learner

Similar Messages

  • Profiler execution plan ONLY for long running queries

    The duration only applies to specific profiler events however I'd like to capture the execution plan ONLY for queries over 10 minutes.
    Is there a way to do this using Xevents?
    Anyone knows?
    Thanks!
    Paula

    I've wanted that too but could not find a way to get it from profiler.
    But it may be possible with xevents (or without xevents!) to watch for long-running queries and then get the plan from the cache,where it will probably stick for some time, using DMVs.
    Josh

  • Process chain for full load.

    Hi,
    I have developed a Process chain for delta loads. Now my question is, can we use process chain to load full load data source for master data, if yeas then what are the steps.
    If we can use process chain for full load to master data then can we use delta and full load infopackages in the process chain.
    Please update.

    Hi Sata,
      You can Include both Full and Delta in the Process Chain. But make sure that you execute the Init InfoPackage manually before executing the delta InfoPackage in the Process Chain.
    For Full Load Process chain, add the infopackage
    >DTP for Full Load---->delete PSA data variant(after 2days) -
    >attribute change run
    Normally load the hierarchy first then the attribute and text.
    Assign points if it helped you.
    Regards,
    Senoy

  • Record count is different for Full load and Init with data transfer

    Hi all,
    We have a data source for which delta is enabled on Calendar Day.
    If I run the full load record count is 2670.But If I run the delta record count is 2665.5 records are missing.
    Would you please help on this.Why those records are missing.

    Hi,
    Actually for Full Load say you have 50 records. When you run it will upload 50 records.
    Now if you run delta if there is no records changes it wouldnt update because there is no update in the existing records.
    only delta is done when there is changes in the records.

  • HR Execution plan failing because target table needs to be extended?

    I'm going a fresh install of OBIEE 10.1.3.4.1 with Dac Build AN 10.1.3.4.1.20090415.0146. This is first one so is it common to have this issue or is something else going on that I just can't see? Should I just enable auto extend on this one table or could there be corruption causing this error?
    Thanks
    In DAC I can validate Informatic and physical servers all connect and are OK and iunder execute, I've ran analyze repository tables and it reporting OK, ran create repository report and no errors or mission objects reports. When I run Execution plan for Human Resources - Oracle R1211 - Flexfield, it failed 5 of the 6 phases wherne 3 are stopped (never ran) and two failed (SDE_ORA_Flx_EBSSegDataTmpLoad, SDE_ORA_Flx_EBSValidationTableDataTmpLoad).
    When I review the logs on the unix server under the /u02/BIAPPS_SUITE/INFR/server/infa_shared/SessLogs/ORA_R1211.DATAWAREHOUSE.SDE_ORAR1211_Adaptor.SDE_ORA_Flx_EBSSegDataTmpLoad.log, I see that it is failing because it has a problem with the target table W_ORA_FLX_EBS_SEG_DATA_TMP
    READER_1_1_1> RR_4049 SQL Query issued to database : (Fri May 28 23:04:22 2010)
    READER_1_1_1> RR_4050 First row returned from database to reader : (Fri May 28 23:04:22 2010)
    WRITER_1_*_1> WRT_8167 Start loading table [W_ORA_FLX_EBS_SEG_DATA_TMP] at: Fri May 28 23:04:21 2010
    WRITER_1_*_1> CMN_1761 Timestamp Event: [Fri May 28 23:04:22 2010]
    WRITER_1_*_1> WRT_8229 Database errors occurred:
    **************ORA-01652: unable to extend temp segment by 128 in tablespace BIA_DW_TBS*
    When I review the ORA_R1211.DATAWAREHOUSE.SDE_ORAR1211_Adaptor.SDE_ORA_Flx_EBSValidationTableDataTmpLoad.log i found the following errors:
    INSERT INTO
    W_ORA_FLX_EBS_VALID_TAB_TMP(FLEX_VALUE_SET_ID,APPLICATION_TABLE_NAME,VALUE_COLUMN_NAME,ID_COLUMN_NAME,VALUE_SET_WHERE_CLAUSE, DATASOURCE_NUM_ID,VALUE_SET_SQL) VALUES ( ?, ?, ?, ?, ?, ?, ?) WRITER_1_*_1> WRT_8020
    ************ No column marked as primary key for table[W_ORA_FLX_EBS_VALID_TAB_TMP]. UPDATEs Not Supported.
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_ORA_FLX_EBS_VALID_TAB_TMP]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158 And then again the can't extend error WRT_8036 Target: W_ORA_FLX_EBS_VALID_TAB_TMP (Instance Name: [W_ORA_FLX_EBS_VALID_TAB_TMP])
    WRT_8038 Inserted rows - Requested: 10080 Applied: 10080 Rejected: 0 Affected: 10080
    READER_1_1_1> BLKR_16019 Read [13892] rows, read [0] error rows for source table [FND_FLEX_VALIDATION_TABLES]
    instance name [FND_FLEX_VALIDATION_TABLES]
    READER_1_1_1> BLKR_16008 Reader run completed.
    WRITER_1_*_1> CMN_1761 Timestamp Event: [Fri May 28 23:04:21 2010]
    WRITER_1_*_1> WRT_8229 Database errors occurred:
    ***** ORA-01652: unable to extend temp segment by 128 in tablespace BIA_DW_TBS

    Worked with DBA and though source tables were not huge, we extended manually and proccess completed once we re-ran.
    Thanks all.

  • 0BBP_BPART_BUYID_ATTR Datasource taking too much time for full load

    Hello Gurus,
    This Data source is fetching 95000 records for five hours in full load. There is no Delta capability. When I run trace and try to see code, It has Oracle and MS SQL inxes but not DB2, and we are on DB2 database. Also all the fields are hidden except Business partner and Buyer ID. I didnot find any notes on this datasource.
    Did any one faced this issue, Help me how can I make this loading Faster in Daily Full loads.

    Hi,
    Divide the load using selections in different infopackages
    Thanks
    HP

  • Infospoke error for full load

    Experts
    when I run full load info spoke through process chain ,It failed ,saying sql error1652. Can you please tell me why this is happening and it is happening only since 2 days. Is it b cos of large volume of data ?
    Any advice greatly appreciated.
    Many thanks

    This error occurs due to temp space problem..Unable to extend TEMP segment tablespace...
    You need to alter the tablespace TEMP default storage (pctincrease 1) and immediately after alter the tablespace TEMP default storage(pctincrease 0)
    Check with your DBA to do this..
    Regards
    Manga(Assign points if it helps!!)

  • Is an execution plan produced for a user lacking grants?

    Will Oracle generate an execution plan for syntactically correct PL/SQL submitted by a user lacking authority to read a table or will it 1st check syntax, then the dictionary, then the rights BEFORE generating the execution plan?

    xerosaburu wrote:
    Will Oracle generate an execution plan for syntactically correct PL/SQL submitted by a user lacking authority to read a table or will it 1st check syntax, then the dictionary, then the rights BEFORE generating the execution plan?incapable, incompetent, or just too lazy to run your own simple test?
    SQL> connect / as sysdba
    Connected.
    SQL> create user user2 identified by  user2;
    User created.
    SQL> grant create session to user2;
    Grant succeeded.
    SQL> connect user2/user2
    Connected.
    SQL> set autotrace on explain
    SQL> select count(*) from hr.employees;
    select count(*) from hr.employees
    ERROR at line 1:
    ORA-00942: table or view does not exist
    SQL> Handle:     xerosaburu
    Status Level:     Newbie (5)
    Registered:     Oct 18, 2007
    Total Posts:     104
    Total Questions:     21 (10 unresolved)
    why so many unanswered questions?

  • Deployment plan fails for weblogic-ejb-jar.xml

    FILE FOR PLAN TO ACT ON:
    application.ear\ReferenceTableSession.jar\META-INF\weblogic-ejb-jar.xml
    <?xml version="1.0" encoding="utf-8"?>
    <weblogic-ejb-jar xmlns="http://www.bea.com/ns/weblogic/10.0"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.bea.com/ns/weblogic/10.0 http://www.bea.com/ns/weblogic/10.0/weblogic-ejb-jar.xsd">
    <weblogic-enterprise-bean>
    <ejb-name>referencetablesession</ejb-name>
    <stateless-session-descriptor>
    <pool>
    <max-beans-in-free-pool>100</max-beans-in-free-pool>
    <initial-beans-in-free-pool>0</initial-beans-in-free-pool>
    </pool>
    <stateless-clustering>
    <stateless-bean-is-clusterable>true</stateless-bean-is-clusterable>
    </stateless-clustering>
    </stateless-session-descriptor>
    <transaction-descriptor>
    <trans-timeout-seconds>150</trans-timeout-seconds>
    </transaction-descriptor>
    <enable-call-by-reference>true</enable-call-by-reference>
    <jndi-name>XYZ_ejb/ProviderMaintSvc/referencetablesession</jndi-name>
    </weblogic-enterprise-bean>
    <transaction-isolation>
    <isolation-level>TransactionReadCommitted</isolation-level>
    <method>
    <ejb-name>referencetablesession</ejb-name>
    <method-name>*</method-name>
    </method>
    </transaction-isolation>
    </weblogic-ejb-jar>
    DEPLOYMENT PLAN:
    <?xml version='1.0' encoding='utf-8'?>
    <deployment-plan xmlns="http://www.bea.com/ns/weblogic/deployment-plan"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.bea.com/ns/weblogic/deployment-plan http://www.bea.com/ns/weblogic/deployment-plan/1.0/deployment-plan.xsd">
    <application-name>application.ear</application-name>
    <variable-definition>
    <variable>
    <name>VAR_referencetablesession</name>
    <value>ejb/ProviderClaimOnlineSvc/referencetablesession</value>
    </variable>
    </variable-definition>
    <module-override>
    <module-name>ReferenceTableSession.jar</module-name>
    <module-type>ejb</module-type>
    <module-descriptor external="false">
    <root-element>weblogic-ejb-jar</root-element>
    <uri>META-INF/weblogic-ejb-jar.xml</uri>
    <variable-assignment>
    <name>VAR_referencetablesession</name>
         <xpath>/weblogic-ejb-jar/weblogic-enterprise-bean/\[ejb-name="referencetablesession"\]/jndi-name</xpath>
    <operation>replace</operation>
    </variable-assignment>
    </module-descriptor>
    </module-override>
    <config-root>servers/AdminServer/upload</config-root>
    </deployment-plan>
    ERROR FROM CONSOLE:
    Message icon - Error An error occurred during activation of changes, please see the log for details.
    Message icon - Error Exception preparing module: EJBModule(ReferenceTableSession.jar) [EJB:011023]An error occurred while reading the deployment descriptor. The error was: Unmarshaller failed.
    Message icon - Error Bean already exists: "[email protected]f2(/WeblogicEnterpriseBeans[referencetablesession])"
    ERROR FROM LOG:
    <Oct 2, 2009 1:08:25 PM CDT> <Error> <J2EE> <BEA-160197> <Unable to load descriptor servers/ms34/tmp/_WL_user/ProviderClaimOnlineSvc/y1anoa/ReferenceTableSession.jar/META-INF/weblogic-ejb-jar.xml of module ReferenceTableSession.jar. The error is weblogic.descriptor.DescriptorException: Unmarshaller failed
    at weblogic.descriptor.internal.MarshallerFactory$1.createDescriptor(MarshallerFactory.java:152)
    at weblogic.descriptor.BasicDescriptorManager.createDescriptor(BasicDescriptorManager.java:306)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.getDescriptorBeanFromReader(AbstractDescriptorLoader2.java:788)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.mergeDescriptorBeanWithPlan(AbstractDescriptorLoader2.java:703)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.loadDescriptorBean(AbstractDescriptorLoader2.java:776)
    at weblogic.ejb.spi.EjbJarDescriptor.parseWeblogicEjbJarBean(EjbJarDescriptor.java:160)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.processWLEjbJarXMLWithSchema(EjbDescriptorReaderImpl.java:719)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.parseWLDD(EjbDescriptorReaderImpl.java:541)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.processWeblogicEjbJarXML(EjbDescriptorReaderImpl.java:380)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.createReadOnlyDescriptorFromJarFile(EjbDescriptorReaderImpl.java:191)
    at weblogic.ejb.spi.EjbDescriptorFactory.createReadOnlyDescriptorFromJarFile(EjbDescriptorFactory.java:93)
    at weblogic.ejb.container.deployer.EJBModule.loadEJBDescriptor(EJBModule.java:1198)
    at weblogic.ejb.container.deployer.EJBModule.prepare(EJBModule.java:380)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:93)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:387)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:37)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:58)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:42)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:615)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:191)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.java:16)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:155)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.RedeployOperation.createAndPrepareContainer(RedeployOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.RedeployOperation.doPrepare(RedeployOperation.java:122)
    Thanks,
    Rob
    Edited by: user11337024 on Oct 5, 2009 2:31 PM

    FILE FOR PLAN TO ACT ON:
    application.ear\ReferenceTableSession.jar\META-INF\weblogic-ejb-jar.xml
    <?xml version="1.0" encoding="utf-8"?>
    <weblogic-ejb-jar xmlns="http://www.bea.com/ns/weblogic/10.0"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.bea.com/ns/weblogic/10.0 http://www.bea.com/ns/weblogic/10.0/weblogic-ejb-jar.xsd">
    <weblogic-enterprise-bean>
    <ejb-name>referencetablesession</ejb-name>
    <stateless-session-descriptor>
    <pool>
    <max-beans-in-free-pool>100</max-beans-in-free-pool>
    <initial-beans-in-free-pool>0</initial-beans-in-free-pool>
    </pool>
    <stateless-clustering>
    <stateless-bean-is-clusterable>true</stateless-bean-is-clusterable>
    </stateless-clustering>
    </stateless-session-descriptor>
    <transaction-descriptor>
    <trans-timeout-seconds>150</trans-timeout-seconds>
    </transaction-descriptor>
    <enable-call-by-reference>true</enable-call-by-reference>
    <jndi-name>XYZ_ejb/ProviderMaintSvc/referencetablesession</jndi-name>
    </weblogic-enterprise-bean>
    <transaction-isolation>
    <isolation-level>TransactionReadCommitted</isolation-level>
    <method>
    <ejb-name>referencetablesession</ejb-name>
    <method-name>*</method-name>
    </method>
    </transaction-isolation>
    </weblogic-ejb-jar>
    DEPLOYMENT PLAN:
    <?xml version='1.0' encoding='utf-8'?>
    <deployment-plan xmlns="http://www.bea.com/ns/weblogic/deployment-plan"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.bea.com/ns/weblogic/deployment-plan http://www.bea.com/ns/weblogic/deployment-plan/1.0/deployment-plan.xsd">
    <application-name>application.ear</application-name>
    <variable-definition>
    <variable>
    <name>VAR_referencetablesession</name>
    <value>ejb/ProviderClaimOnlineSvc/referencetablesession</value>
    </variable>
    </variable-definition>
    <module-override>
    <module-name>ReferenceTableSession.jar</module-name>
    <module-type>ejb</module-type>
    <module-descriptor external="false">
    <root-element>weblogic-ejb-jar</root-element>
    <uri>META-INF/weblogic-ejb-jar.xml</uri>
    <variable-assignment>
    <name>VAR_referencetablesession</name>
         <xpath>/weblogic-ejb-jar/weblogic-enterprise-bean/\[ejb-name="referencetablesession"\]/jndi-name</xpath>
    <operation>replace</operation>
    </variable-assignment>
    </module-descriptor>
    </module-override>
    <config-root>servers/AdminServer/upload</config-root>
    </deployment-plan>
    ERROR FROM CONSOLE:
    Message icon - Error An error occurred during activation of changes, please see the log for details.
    Message icon - Error Exception preparing module: EJBModule(ReferenceTableSession.jar) [EJB:011023]An error occurred while reading the deployment descriptor. The error was: Unmarshaller failed.
    Message icon - Error Bean already exists: "[email protected]f2(/WeblogicEnterpriseBeans[referencetablesession])"
    ERROR FROM LOG:
    <Oct 2, 2009 1:08:25 PM CDT> <Error> <J2EE> <BEA-160197> <Unable to load descriptor servers/ms34/tmp/_WL_user/ProviderClaimOnlineSvc/y1anoa/ReferenceTableSession.jar/META-INF/weblogic-ejb-jar.xml of module ReferenceTableSession.jar. The error is weblogic.descriptor.DescriptorException: Unmarshaller failed
    at weblogic.descriptor.internal.MarshallerFactory$1.createDescriptor(MarshallerFactory.java:152)
    at weblogic.descriptor.BasicDescriptorManager.createDescriptor(BasicDescriptorManager.java:306)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.getDescriptorBeanFromReader(AbstractDescriptorLoader2.java:788)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.mergeDescriptorBeanWithPlan(AbstractDescriptorLoader2.java:703)
    at weblogic.application.descriptor.AbstractDescriptorLoader2.loadDescriptorBean(AbstractDescriptorLoader2.java:776)
    at weblogic.ejb.spi.EjbJarDescriptor.parseWeblogicEjbJarBean(EjbJarDescriptor.java:160)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.processWLEjbJarXMLWithSchema(EjbDescriptorReaderImpl.java:719)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.parseWLDD(EjbDescriptorReaderImpl.java:541)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.processWeblogicEjbJarXML(EjbDescriptorReaderImpl.java:380)
    at weblogic.ejb.container.dd.xml.EjbDescriptorReaderImpl.createReadOnlyDescriptorFromJarFile(EjbDescriptorReaderImpl.java:191)
    at weblogic.ejb.spi.EjbDescriptorFactory.createReadOnlyDescriptorFromJarFile(EjbDescriptorFactory.java:93)
    at weblogic.ejb.container.deployer.EJBModule.loadEJBDescriptor(EJBModule.java:1198)
    at weblogic.ejb.container.deployer.EJBModule.prepare(EJBModule.java:380)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:93)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:387)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:37)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:58)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:42)
    at weblogic.application.internal.BaseDeployment$1.next(BaseDeployment.java:615)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:37)
    at weblogic.application.internal.BaseDeployment.prepare(BaseDeployment.java:191)
    at weblogic.application.internal.EarDeployment.prepare(EarDeployment.java:16)
    at weblogic.application.internal.DeploymentStateChecker.prepare(DeploymentStateChecker.java:155)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.prepare(AppContainerInvoker.java:60)
    at weblogic.deploy.internal.targetserver.operations.RedeployOperation.createAndPrepareContainer(RedeployOperation.java:98)
    at weblogic.deploy.internal.targetserver.operations.RedeployOperation.doPrepare(RedeployOperation.java:122)
    Thanks,
    Rob
    Edited by: user11337024 on Oct 5, 2009 2:31 PM

  • DAC - Can I run more then one Exection Plans with full load option enabled?

    Hello All,
    I have configured for subject Area sales – Orders, Backlog and Invoices (OBIEE)
    Now I want to run DAC to load the underline tables for subject Areas Financials – Revenue Analysis and Financials – GL Budget and Expenses, I have identified the relevant fact tables and built a Execution Plan in DAC.
    DAC - In which mode should I run the execution plan?
    With full load option enabled or disabled because already full load run is done for sales – Orders, Backlog and Invoices and this run should not disturb the existing data
    If I run with full load option enabled do it overrides the existing data for other subject area
    Please help me I am bit confused!
    Regards,
    SMA
    Edited by: SMA on Aug 18, 2009 3:36 PM

    No need of combining all the subject areas into one execution plan. Simply run the second ETL without checking the full load option. It will do the full load for the tables related to the newly added subject areas and incremental for the common tables between old and new subject areas.

  • DAC: Clearing Failed Execution Plans and BAW Tables

    Hi all,
    Thank you for taking the time to review this post.
    Background
    Oracle BI Applications 7.9.6 Financial Analytics
    OLTP Source: E-Business Suite 11.5.10
    Steps Taken
    1. In DAC I have create a New Source Container based on Oracle 11.5.10
    2. I have updated the parameters in the Source System parameters
    3. Then I created a new Execution Plan as a copy of the Financials_Oracle 11.5.10 record and checked Full Load Always
    4. Added new Financials Subject Areas so that they have the new Source System
    5. Updated the Parameters tab with the new Source System and Generated the Parameters
    6. Build a new Execution Plan - Fails
    Confirmation for Rerun
    I want to confirm that the correct steps to Rerun an Execution Plan are as follows. I want to ensure that the OLTP (BAW) tables are truncated. I am experiencing duplicates in the W_GL_SEGMENTS_D (and DS) table even though there are no duplicates on the EBS.
    In DAC under the EXECUTE window do the following:
    - Navigate to the 'Current Run' tab.
    - Highlight the failed execution plan.
    - Right click and seleted 'Mark as completed.'
    - Enter the numbers/text in the box.
    Then:
    - In the top toolbar select Tools --> ETL Management --> Reset Data Sources
    - Enter the numbers/text in the boox.
    Your assistance is greatly appreciated.
    Kind Regards,
    Gary.

    Hi HTH,
    I can confirm that I do not have duplicates on the EBS side.
    I got the SQL Statement by:
    1. Open Mapping SDE_ORA_GL_SegmentDimension in the SDE_ORA11510_Adaptor.
    2. Review the SQL Statement in the Source Qualifier SQ_FND_FLEX_VALUES
    3. Run this SQL command against my EBS 11.5.10 source OLTP and the duplicates that are appearing in the W_GL_SEGMENT_DS table to not exist.
    SELECT
    FND_FLEX_VALUE_SETS.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUE_SETS.FLEX_VALUE_SET_NAME,
    FND_FLEX_VALUES.FLEX_VALUE,
    MAX(FND_FLEX_VALUES_TL.DESCRIPTION),
    MAX(FND_FLEX_VALUES.LAST_UPDATE_DATE),
    MAX(FND_FLEX_VALUES.LAST_UPDATED_BY),
    MAX(FND_FLEX_VALUES.CREATION_DATE),
    MAX(FND_FLEX_VALUES.CREATED_BY),
    MAX(FND_FLEX_VALUES.START_DATE_ACTIVE),
    MAX(FND_FLEX_VALUES.END_DATE_ACTIVE),
    FND_FLEX_VALUE_SETS.LAST_UPDATE_DATE LAST_UPDATE_DATE1
    FROM
    FND_FLEX_VALUES, FND_FLEX_VALUE_SETS, FND_FLEX_VALUES_TL
    WHERE
    FND_FLEX_VALUES.FLEX_VALUE_SET_ID = FND_FLEX_VALUE_SETS.FLEX_VALUE_SET_ID AND FND_FLEX_VALUES.FLEX_VALUE_ID = FND_FLEX_VALUES_TL.FLEX_VALUE_ID AND
    FND_FLEX_VALUES_TL.LANGUAGE = 'US' AND
    (FND_FLEX_VALUES.LAST_UPDATE_DATE > TO_DATE('$$LAST_EXTRACT_DATE', 'MM/DD/YYYY HH24:MI:SS') OR
    FND_FLEX_VALUE_SETS.LAST_UPDATE_DATE > TO_DATE('$$LAST_EXTRACT_DATE', 'MM/DD/YYYY HH24:MI:SS'))
    GROUP BY
    FND_FLEX_VALUE_SETS.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUE_SETS.FLEX_VALUE_SET_NAME,
    FND_FLEX_VALUES.FLEX_VALUE,
    FND_FLEX_VALUE_SETS.LAST_UPDATE_DATEHowever, one thing that I noticed was that I wanted to validate what the value of parameter $$LAST_EXTRACT_DATE is being populated with.
    My investigation took me along the following route:
    Checked what was set up in the DAC (+Dac Build AN 10.1.3.4.1.20090415.0146, Build date: April 15 2009+):
    1. Design View -> Source System Parameters -> $$LAST_EXTRACT_DATE = Runtime Variable =@DAC_SOURCE_PRUNED_REFRESH_TIMESTAMP (+Haven't been able to track this Variable down!)+
    2. Setup View -> DAC System Properties -> InformaticaParameterFileLocation -> $INFA_HOME/server/infa_shared/SrcFiles
    Reviewing one of the log files for my failing Task:
    $INFA_HOME/server/infa_shared/SrcFiles/ORA_11_5_10.DWRLY_OLAP.SDE_ORA11510_Adaptor.SDE_ORA_GLSegmentDimension_Full.txt
    I noticed that several variables near the bottom (including $$LAST_EXTRACT_DATE) have not been populated. This variable gets populated at Runtime but is there a log file that shows the value that gets populated? I would also have expected that there would have been a Subsitution Variable in the place of a Static Value.
    [SDE_ORA11510_Adaptor.SDE_ORA_GLSegmentDimension_Full]
    $$ANALYSIS_END=01/01/2011 12:59:00
    $$ANALYSIS_END_WID=20110101
    $$ANALYSIS_START=12/31/1979 01:00:00
    $$ANALYSIS_START_WID=19791231
    $$COST_TIME_GRAIN=QUARTER
    $$CURRENT_DATE=03/17/2010
    $$CURRENT_DATE_IN_SQL_FORMAT=TO_DATE('2010-03-17', 'YYYY-MM-DD')
    $$CURRENT_DATE_WID=20100317
    $$DATASOURCE_NUM_ID=4
    $$DEFAULT_LOC_RATE_TYPE=Corporate
    $$DFLT_LANG=US
    $$ETL_PROC_WID=21147868
    $$FILTER_BY_SET_OF_BOOKS_ID='N'
    $$FILTER_BY_SET_OF_BOOKS_TYPE='N'
    $$GBL_CALENDAR_ID=WPG_Calendar~Month
    $$GBL_DATASOURCE_NUM_ID=4
    $$GLOBAL1_CURR_CODE=AUD
    $$GLOBAL1_RATE_TYPE=Corporate
    $$GLOBAL2_CURR_CODE=GBP
    $$GLOBAL2_RATE_TYPE=Corporate
    $$GLOBAL3_CURR_CODE=MYR
    $$GLOBAL3_RATE_TYPE=Corporate
    $$HI_DATE=TO_DATE('3714-01-01 00:00:00', 'YYYY-MM-DD HH24:MI:SS')
    $$HI_DT=01/01/3714 12:00:00
    $$HR_ABSNC_EXTRACT_DATE=TO_DATE('1980-01-01 08:19:00', 'YYYY-MM-DD HH24:MI:SS')
    $$HR_WRKFC_ADJ_SERVICE_DATE='N'
    $$HR_WRKFC_EXTRACT_DATE=01/01/1970
    $$HR_WRKFC_SNAPSHOT_DT=TO_DATE('2004-01-01 00:00:00', 'YYYY-MM-DD HH24:MI:SS')
    $$HR_WRKFC_SNAPSHOT_TO_WID=20100317
    $$Hint1=
    $$Hint_Tera_Post_Cast=
    $$Hint_Tera_Pre_Cast=
    $$INITIAL_EXTRACT_DATE=06/27/2009
    $$INVPROD_CAT_SET_ID=27
    $$INV_PROD_CAT_SET_ID10=
    $$INV_PROD_CAT_SET_ID1=27
    $$INV_PROD_CAT_SET_ID2=
    $$INV_PROD_CAT_SET_ID3=
    $$INV_PROD_CAT_SET_ID4=
    $$INV_PROD_CAT_SET_ID5=
    $$INV_PROD_CAT_SET_ID6=
    $$INV_PROD_CAT_SET_ID7=
    $$INV_PROD_CAT_SET_ID8=
    $$INV_PROD_CAT_SET_ID9=
    $$LANGUAGE=
    $$LANGUAGE_CODE=E
    $$LAST_EXTRACT_DATE=
    $$LAST_EXTRACT_DATE_IN_SQL_FORMAT=
    $$LAST_TARGET_EXTRACT_DATE_IN_SQL_FORMAT=
    $$LOAD_DT=TO_DATE('2010-03-17 19:27:10', 'YYYY-MM-DD HH24:MI:SS')
    $$LOW_DATE=TO_DATE('1899-01-01 00:00:00', 'YYYY-MM-DD HH24:MI:SS')
    $$LOW_DT=01/01/1899 00:00:00
    $$MASTER_CODE_NOT_FOUND=
    $$ORA_HI_DATE=TO_DATE('4712-12-31 00:00:00', 'YYYY-MM-DD HH24:MI:SS')
    $$PROD_CAT_SET_ID10=
    $$PROD_CAT_SET_ID1=2
    $$PROD_CAT_SET_ID2=
    $$PROD_CAT_SET_ID3=
    $$PROD_CAT_SET_ID4=
    $$PROD_CAT_SET_ID5=
    $$PROD_CAT_SET_ID6=
    $$PROD_CAT_SET_ID7=
    $$PROD_CAT_SET_ID8=
    $$PROD_CAT_SET_ID9=
    $$PROD_CAT_SET_ID=2
    $$SET_OF_BOOKS_ID_LIST=1
    $$SET_OF_BOOKS_TYPE_LIST='NONE'
    $$SOURCE_CODE_NOT_SUPPLIED=
    $$TENANT_ID=DEFAULT
    $$WH_DATASOURCE_NUM_ID=999
    $DBConnection_OLAP=DWRLY_OLAP
    $DBConnection_OLTP=ORA_11_5_10
    $PMSessionLogFile=ORA_11_5_10.DWRLY_OLAP.SDE_ORA11510_Adaptor.SDE_ORA_GLSegmentDimension_Full.logThe following snippet was discovered in the DAC logs for the failed Task (Log file: SDE_ORA_GLSegmentDimension_Full_DETAIL.log):
    First error code [7004]
    First error message: [TE_7004 Transformation Parse Warning [FLEX_VALUE_SET_ID || '~' || FLEX_VALUE]; transformation continues...]Finally I can confirm that there was a Task Truncate Table W_GL_LINKAGE_INFORMATION_GS that was successfully executed in Task SDE_ORA_GL_LinkageInformation_Extract.
    Any further guidance greatly appreciated.
    Kind Regards,
    Gary.

  • Executing Full load for one task

    Hi Gurus,
    execution plan failed with some errors( unique constraint) errors. now i decided to run a full load for the failed task? is it possible? if possible, can some one give me the exact steps to be followed?
    Regards,
    Nagarjuna

    If you are getting unique constraint error, you can change the index to non unique via the DAC, let it complete and then debug why you are getting duplicate data. By forcing it to full implies that sourece is not providing duplicates but the incremental loads are whoich may be a different issue than source data being bad.

  • DAC Commands for Incremental and Full load

    Hi,
    I'm implementing BIApps 7.9.6.1 for a customer. For R12 container, I noticed for 5 DAC tasks the command for Incremental and Full load starts with "@DAC_" and ends with "_CMD". Due to this, the ETL load fails. Is this a bug..?
    Thanks,
    Seetharam

    You may want to look at Metalink note ID 973191.1:
    Cause
    The 'Load Into Source Dimension', has the following definition:
    -- DAC Client > Design > Tasks > Load Into Source Dimension > Command for Incremental Load = '@DAC_SOURCE_DIMENSION_INCREMENTAL'
    and
    -- - DAC Client > Design > Tasks > Load Into Source Dimension > Command for Full Load = '@DAC_SOURCE_DIMENSION_FULL'
    instead of the actual Informatica workflow names.
    The DAC Parameter is not substituted with appropriate values in Informatica during ETL
    This is caused by the fact that COMMANDS FOR FULL and INCREMENTAL fields in a DAC Task do not allow for database specific texts as described in the following bug:
    Bug 8760212 : COMMANDS FOR FULL AND INCREMENTAL SHOULD ALLOW DB SPECIFIC TEXTS
    Solution
    This issue was resolved after applying Patch 8760212
    The documentation states to apply Patch 8760212 to DAC 10.1.3.4.1 according to the Systems Requirements and Supported Platforms Guide for Oracle Business Intelligence Data Warehouse Administration Console 10.1.3.4.1.
    However, Patch 8760212 has been made obsolete, recently, in this platform and language. Please see the reason stated below on the 'Patches and Update' tab on My Oracle Support.
    Reason for Obsolescence
    Use cumulative Patch 10052370 instead.
    Note: The most recent replacement for this patch is 10052370. If you are downloading Patch 8760212 because it is a prerequisite for another patch or patchset, you should verify whether or not if Patch 10052370 is suitable as a substitute prerequisite before downloading it.

  • Can I use full load for daily loads in LO Cockpit extraction.

    I am using LO cockpit extraction for 2LIS_11_VAITM and i am using delta load for daily load and I have following doubt.
    Can i use full load for daily loads instead of delta in LO Cockpit extraction.
    Because, my understanding is that full load always takes data from setup tables in ECC and set up tables will not have delta data.
    pls reply. Thanks.

    You are right abt the understanding that full load (atleast for the extractor at hand) brings data from setup tables. You are also right that delta records are not loaded into setup tables.
    So if you plan to do full loads every day, you have to load the setup tables everyday. It is at this juncture that I have to remind you that, filling of setup tables requires a downtime of the system to ensure  that no changes / creates happen during the process.
    Hence performing a full load is not a very good approach to go by...especially when SAP made the job easy by doing all the dirty work of identifying changes and writing them to extraction queue.
    Hope this helps!

  • Process Chain Failed for Perticular day, want to take load from R/3 for day

    Hi Experts,
    While monitoring the process chain i found that on 16 Nov2010 the schedule got failed then i checked but that was due to rfc connection to R/3 failed then again i repeated the process in RSPCM. But toady in the morning when i was checking i found that for 17 nov scheduled got run successful but when i drilled down on date for 16th data but it was failed for 16. Now please guid me how to bring the data for 16nov from R/3. My process chain contains both Master Data/ Transaction Data and Deltas / Full loads ???

    Hi,
    As you have mentioned in your post, the process chain for 17th has finished successfully, can you confirm if all the dataload also finished successfully?
    If the dataload also has finished successfully, then ther is not much to be done from your side. Pplease note below points
    1. For Delta Loads- No need to worry about them. For standard datasources, the system would have automatically fetched all the deltas since your last successful delta load. For generic datasources as well, based on previous delta pointer in RSA7, all the delta records would have got fetched.
    2. For full dataloads- Check the selection criteria in your infopackages. If its based on date, then you would have to do dataload again.
    The selection criteria for this dataloads should include the dates for 16th & 17th dataload. Everything else should be fine.
    Else, if your dataload did not finish succesfully on17th, though the process chain has finished. You need to do below
    1. For all the delta loads- Carry out Repeat Delta loads. This would fetch all the deltas missed. After this load, all the deltas would continue to come normally.
    2. For full loads, perform same as mentioned in point 2 above.

Maybe you are looking for

  • EMac hard drive dead?

    My eMac was running slow ... reeeeeeeally slow. Spent ages pondering every mouse click. Just opening folders was agony. I tried the usual round of fixes, some of which helped a little, but after a few hours' use, things would slow down again. Permiss

  • How do I download my entire music library in iTunes to not be in iCloud?

    I have over 10,000 songs in my iTunes library.  18 months ago, I signed up for iTunes Match, but cancelled the service, however, my songs all show as being in iCloud.  How do I download them all from iCloud to my Mac.  Why do they still show as in iC

  • Script Execution in File Adapter

    Hi, I have the following scenario, There are two FTP folders(say FTP_SingleFile and FTP_MultipleFiles). At any point of time, there would be only one file in FTP_SingleFile and n number of files in FTP_MultipleFiles. A sender communication channel  p

  • Creting  object array  with constructor

    Hello all, Here is a query kindly let me know your views. Q) I want to create four objects and if i write a constructor how is it invoked.. kidly help me ... thanks in advance mahesh

  • Can I Open my CS 5 files in CC?

    Can I Open my CS 5 files in CC?