ETL Failing on TASK_GROUP_Extract_BOM

Hello All,
I am also facing an issue in the ETL (OBIA 7.9.6.4) for Supply Chain Management R12.1.3
I can see in DAC that the task "TASK_GROUP_Extract_BOM" is failing on truncate table.
I have excluded this task from ETL container but this task is still the part of ETL and failing.
Further I cannot see the session log of this task in Informatica Monitor.
Do I need to inactive this task from somewhere else as well like from informatica? If yes then how?
Regards
Naeem Akhtar

Hi Naeem,
After excluding this task from your ETL container, you must rebuild your execution plan to exclude the task in the execution plan :
You could see step 8, 9 and 10 in the Oracle documentation : http://docs.oracle.com/cd/E28271_01/fusionapps.1111/e14849/dacexecutionplans.htm#i1061742
Hope it helps,
Benoit

Similar Messages

  • DAC, ETL failed due to Post-Etl Script not exist and table not exist.

    When execute the ETL in DAC, it failed when execute the "Change Capture For Siebel OLTP"
    encounterred the following two type errors.
    how to fix these errors?
    The first is the table not exist:
    ======================================================================
    MESSAGE:::Could not truncate table 'S_ETL_I_IMG_38' because this table does not exist in database 'OLTP_SIA'.
    EXCEPTION CLASS::: com.microsoft.sqlserver.jdbc.SQLServerException
    com.microsoft.sqlserver.jdbc.SQLServerException.makeFromDatabaseError(Unknown Source)
    ======================================================================
    The second is the Post-Etl not exist.
    I have copied the Copy Source Files and Lookup Files. What else should I do? What script it not find? have no idea.
    ======================================================================
    Jan 20, 2009 4:11:06 PM com.siebel.etl.etlmanager.EtlExecutionManager finishEtl
    SEVERE: Incorrectly specified Post-Etl Script/Executable. The file either does n
    ot exist, or is not readable
    UNREGISTERING: com.siebel.etl.engine.bore.FixedFailedNodesFacility
    686 SEVERE Tue Jan 20 16:11:06 CST 2009
    END OF ETL
    687 SEVERE Tue Jan 20 16:11:06 CST 2009 Incorrectly specified Post-Etl Script
    /Executable. The file either does not exist, or is not readable
    UNREGISTERING: com.siebel.etl.engine.core.SessionHandler
    UNREGISTERING: com.siebel.etl.engine.bore.ServerTokenPool
    UNREGISTERING: com.siebel.etl.engine.core.TableIndexHandler
    UNREGISTERING: com.siebel.etl.engine.bore.NewDispatcher
    Stopping EtlParameters 9550256
    UNREGISTERING: com.siebel.etl.engine.bore.EtlParameters
    UNREGISTERING: com.siebel.etl.engine.bore.ExternalTaskTokenPool
    UNREGISTERING: com.siebel.analytics.etl.execution.RefreshDateHandler

    Hi tigger,
    ad scripts:
    http://download.oracle.com/docs/cd/E12102_01/books/AnyInstAdm784/AnyInstAdmFunctionalRef9.html#wp1005877
    http://download.oracle.com/docs/cd/E12102_01/books/AnyInstAdm784/AnyInstAdmFunctionalRef8.html
    ad tables:
    Without data change capture tables nothing will work. Have you looked at the actual documentaiotn? It contains everythign step by step.
    http://www.oracle.com/technology/documentation/bi_apps.html
    --> Oracle Business Intelligence Applications Fusion Edition Installation and Configuration Guide -> 4.15.3 How to Update Siebel Transactional Database Schema Definitions

  • ETL fails

    Getting error
    Connected to Integration Service: [Powercenter_integr_serv].
    ERROR: unable to open local paramfile [u02/oracle/Informatica/PowerCenter/8.0.6/server/infa_shared/ScrFiles/ORA_R12_Flatfile.DataWarehouse.SILOS.SIL_InsertRowInRunTable.txt].
    Disconnecting from Integration Service
    I checked and directory /u02/oracle/Informatica/PowerCenter/8.0.6/server/infa_shared/ScrFiles is good and permissions to write are there.
    Please let me know if you have encountered it

    Hi,
    Is the file actually there though? I think it is trying to read the file not write to it. One of the steps installing the OBIA is copying the source files into that folder.
    Regards,
    Matt

  • Getting error 36331 while running prebuilt  etl for  first  time

    Hi all,
    I have installed BI apps 7.9.6 and OBIEE 10.1.3.4.1, DAC 10.13.4.1 on a Linux machine with Oracle 10gR2 as data base and EBS 11 i as the data source. I am running the prebuilt ETL for the first time for a new execution plan created with Financials-Receivables subject area added to it. Of the total 292 tasks 30 tasks run well and the ETL fails after that with the following error.
    The following error message is from the DAC server log. I have seen the session logs and workflow logs for the SDE_ORA_CodeDimension_Supplier_Payment_Method task, but could figure out what could be the issue. Please let me know if any one has encountered the same issue.
    =====================================
    ERROR OUTPUT
    =====================================
    524 SEVERE Tue Sep 22 06:41:14 EDT 2009 Could not attach to workflow because of errorCode 36331 For workflow SDE_ORA_CodeDimension_Supplier_Payment_Method
    525 SEVERE Tue Sep 22 06:41:14 EDT 2009
    ANOMALY INFO::: Error while executing : INFORMATICA TASK:SDE_ORA11510_Adaptor:SDE_ORA_CodeDimension_Supplier_Payment_Method:(Source : FULL Target : FULL)
    MESSAGE:::
    Irrecoverable Error
    Error while contacting Informatica server for getting workflow status for SDE_ORA_CodeDimension_Supplier_Payment_Method
    Error Code = 36331:Unknown reason for error code 36331
    Pmcmd output :
    Thx
    RM

    Found that the indices on tables are failing to be created while running the workflow. I have updated the status of the task as completed and re queued the task. ETL ran successfully then.

  • ETL Failure - Integration Service Error

    Hi all,
    Wondering if anyone has seen this error before.
    ETL failed with this error message:
    ERROR: Workflow custom workflow  has not run on this Integration
    Service before.
    Nothing has changed with this workflow which has been running without error.  We only have one integration service.
    Thanks for reading.

    2013-07-12 04:10:06.14 INFORMATICA TASK:ABC_Custom_SIL:ABC_SIL_Special_Programs:(Source
    : FULL Target : FULL) has started.
    ANOMALY INFO::: Error while executing : INFORMATICA TASK:ABC_Custom_SIL:ABC_SIL_Special_Programs:(Source
    : FULL Target : FULL)
    MESSAGE:::
    Irrecoverable Error
    Error while contacting Informatica server for getting workflow status for ABC_SIL_Special_Programs
    Error Code = 3:An error occurred in starting or running or stopping the
    workflow or task.
    Pmcmd output :
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0902], LINUX 64-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Fri Jul 12 04:10:28 2013
    Connected to Integration Service: [INT_PROD].
    ERROR: Workflow [ABC_SIL_Special_Programs] has not run on this Integration
    Service before.
    Disconnecting from Integration Service
    Completed at Fri Jul 12 04:10:29 2013
    =====================================
    ERROR OUTPUT
    =====================================
    Re-Queue to attempt to run again or attach to running workflow
    if Execution Plan is still running or re-submit Execution Plan to execute the
    workflow.
    EXCEPTION CLASS::: com.siebel.analytics.etl.etltask.IrrecoverableException
    com.siebel.analytics.etl.etltask.InformaticaTask.doExecute(InformaticaTask.java:183)
    com.siebel.analytics.etl.etltask.GenericTaskImpl.doExecuteWithRetries(GenericTaskImpl.java:411)
    com.siebel.analytics.etl.etltask.GenericTaskImpl.execute(GenericTaskImpl.java:307)
    com.siebel.analytics.etl.etltask.GenericTaskImpl.execute(GenericTaskImpl.java:214)
    com.siebel.analytics.etl.etltask.GenericTaskImpl.run(GenericTaskImpl.java:586)
    com.siebel.analytics.etl.taskmanager.XCallable.call(XCallable.java:63)
    java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
    java.util.concurrent.FutureTask.run(FutureTask.java:138)
    java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
    java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
    java.util.concurrent.FutureTask.run(FutureTask.java:138)
    java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
    java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
    java.lang.Thread.run(Thread.java:662)
    (Number of retries : 1)
    pmcmd startworkflow -sv INT_PROD -d Domain_lx-xxxxx-p02p -u Administrator -p
    **** -f ABC_Custom_SIL -paramfile
    /opt/hrinfo/informatica/powercenter9.1/server/infa_shared/SrcFiles//DataWarehouse.DataWarehouse.ABC_Custom_SIL.ABC_SIL_Special_Programs.txt
    ABC_SIL_Special_Programs
    2013-07-12 04:10:29.207 INFORMATICA TASK:ABC_Custom_SIL:ABC_SIL_Special_Programs:(Source
    : FULL Target : FULL) has finished execution with Failed status.

  • GROUPBY error in OBIEE 10g

    I had a requirment where I need to calculate the duration of the etl runs on daily basis
    So I have calculated the duration using the below formula.
    by grouping the run_wid and using the column in criteria and hiding it(in order to avoid the error 'column in GROUPBY clause not in the projection list.')..
    the reason for grouping the run_wid is when the etl fails it is run multiple times and multiple entries on same run_wid are generated so I have captured them by
    grouping them based on run_wid.
    Everything works here fine.
    But when I want to analyze on the whole year average etl runs using the above process.
    So in order to achieve this by taking the year and duration and grouping by run_wid. I again get the error 'column in GROUPBY clause not in the projection list.').
    I can't use the run_wid in the criteria because it not showing at year level. Please suggest a way/idea on achieving this.
    Edited by: 949675 on May 30, 2013 10:17 PM

    Hi Yadav,
    Thanks for suggesting the idea it worked little bit by not making me to pull the run_wid in both
    criteria and adding in GROUP by in advanced tab.
    The problem what I am facing now is the value is not average values and its only showing the sum
    of all the values
    The formula for calculating Average is Avg(round(sum(TimestampDiff(SQL_TSI_SECOND,Audit."Start Time",Audit."End Time"))/60,2)/60 by run_wid)
    The columns in my criteria are
    Year|Average
    Ex:
    In year 2011 the ETL has run 50 times, I am seeing 50 etls total run time in Average.

  • SNMP TRAP on EvENT

    Hi & Thanks for read it,
    I would like to design SNMP event, traps in order to be managed by SNMP Network management platform, like HP OpenView, or other.
    It's possible to generate a trap in pl/sql or via enterprise manager on procedure event?
    Anyone have test log4plsql for send traps?
    Anyone know if it's possible to send traps via OS COMMAND on ORACLE?
    Thanks

    I don't want to use oracle MIB. I want to use my own
    MIB.
    For example I want to send a trap if process of ETL
    fails.I am not expert of SNMP setup but looks like Oracle SNMP support in this respect is kinda limited. Hope others may have better idea.

  • Informatica 7.1.4:  IDLE Connections Killed

    So I posted this question on Informatica's forum "Communities" but I wanted to post it here as well.
    We're on Informatica 7.1.4 (yeah, I know its old, but we're about to upgrade) and we have been running into issues where one long running ETL fails because the Oracle 11gR1 DB terminated it's IDLE connection. From what I can understand, here's what happens.
    1. We have a max concurrency rate of 10. So suppose 9 other ETLs are running along with our 1 long running ETL (1hr25minutes is the average time for this one).
    2. All the 10 ETLs create DB connections for the Writer thread at the kick off of the ETL
    3. The long running ETL takes an hour before it even gets to the point where it can start writing to the target. So basically it's connection has been IDLE the entire 1 hour.
    4. On the DB side of things, the DBAs have a cleanup script to remove IDLE sessions that haven't done anything in 1+ hours. So their script kills the connection that the Writer thread is about to use.
    5. Our long running ETL finally finishes sourcing and processing the data and it is ready to insert, but now it fails because the connection has been terminated by the DB.
    I know the easy out is to have the DBAs not kill connections made by Informatica, but I was hoping there would be a more elegant solution.
    Is there a way to have Informatica revalidate it's connection prior to writing? Something like, if connection doesn't exist, re-create, else write? It seems like this should be a server level setting.
    Or is there a way to have Informatica send some kind of heartbeat to the Oracle DB to keep the connection alive?
    Right now we're definitely having the DBAs avoid killing connections created by Informatica. But for the long term, we would like to see Informatica be more intelligent about the health of the connection its using.
    Thanks for the advice.
    -Joe

    Are both of the Informatica Servers on the same machine?
    If so, one thing you can try is when connecting to the Remote Server from Windows is setting the IP Addresses to 127.0.0.1 or using the name localhost for the Informatica Server (In the Server Editor screen of the Workflow Manager). You only need to set the valid IP Address for the Repository Server to get the Repository Admin tool to connect. Once that connection is made the Informatica Server uses that information to start itself.
    I used an approach like this to stand up both Informatica Servers on Oracle Enterprise Linux.

  • Issues with W_DAY_D

    Hi All,
    I have 10 years of data for Order Management that was loaded as Full. I am running nightly incremental loads.
    I needed to replace the fiscal month calendar with the one client is using, i.e July is first month and June is month 12.
    I placed the fiscal_month.csv file that client provided.
    Since the year extends beyond 2010, I made the refresh date for the date dimensions NULL.
    I changed the $$END_DATE parameter for SIL_DayDimension task to 2020 and pushed it to references.
    The ETL fails to truncate W_DAY_D after the changes and I get unique constraint errors. Any reason why? Thanks for your responses

    Well if you check the DAC tasks, then you'll see that only the SIL_DayDimension_UpdateFlag runs for incremental loads.
    If you want to reload the W_DAY_D, then you need to reset the refresh date of the TARGET tables which then in turn will propagate to W_DAY_D. I.e.: The target tables for tasks starting with SIL_DayDimension*.
    Alternatively (and that's what I do normally), create a Subject Area explicitely for W_DAY_D and its staging tables. This way you can run a reload of that table as often as you want through its own execution plan.
    HTH.
    Chris

  • DW Tables Creation - Columns Missing

    Hi All,
    When i run the ETL for the execution plan containing tasks related to SCM, some tasks fail, predominantly of which are due to columns missing in the DW tables.
    I dropped and recreated the DW tables ( Tools -> ETLMgmt -> Configure) but it didnt help.
    I inspected the ctl file, oracle_bi_dw.ctl in ( \OracleBI\DAC\conf\sqlgen\ctl-file), which i believe is used to create the DW tables. The CTL file has definitions for all the tables needed but certain columns are missing, and due to this my ETL fails . I tried creating the columns manually after each ETL task failed.
    I need guidance into how to recreate the tables or which configuration step to redo to get the correct CTL file used for the DW table creation.
    Thanks,
    Raghav

    Hi All,
    Re-installing BIApps did not work . I guess the ctl is created while you import the DAC metadata.
    I copied the ctl file from a fellow developer. I dropped and created all the tables. All the DW table definitions referenced from the new ctl file are fine now.
    Now i would always keep a backup of a proper ctl file to face such future problems :)
    Thanks,
    Raghav

  • Loyalty Analytics & missing tables

    We have upgraded our OLTP and OLAP to include Loyalty analytics but the ETL fails due to missing IMG tables due to missing OLTP objects such as S2_NODE, S2_CONTACT and S2_ADDR_ORG. Does anyone know where those tables/views are defined or have the DDL for them? We have done the change capture process in DAC but these objects do not get included.
    Thanks

    966918 wrote:
    Hello !
    There are missing tables in an expdp dump of a database we use in a test environnement.
    I don't know why ...
    The datafile seems not to be corrupted. The table is listed in a tool like TOAD, and is listed in the catalog view DBA_TABLES.
    But the expdp logs doesn't list this table while listing other tables from the same user.
    If i try to expdp only a table of these (a missing table in the dumpfile), i get a 39166 Object not found error.
    which schema owns the "missing" object?

  • Error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.

    Hi,
    Today one my task in the DAC Etl failed and when I opened the DAC log file of it. I saw the error description as below
    Error while contacting Informatica server for getting workflow status for PLP_SRC_WH_AP_Invoice_Reconciliation
    Error Code = 36331:Unknown reason for error code 36331
    Since there is no much information in DAC log file about the failure, I opened the PLP_SRC_WH_AP_Invoice_Reconciliation session log file.
    The error description in the session log file says.
    There is an error in the port [NEXTVAL]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
    ... nl:ERROR(u:'transformation error')).
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11023 : An error occurred converting data in the port [NEXTVAL]. The row was dropped.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11021 : An error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01 : CMN_1086 : RECON_SEQ_GEN: Number of errors exceeded threshold [1].
    2014-01-14 08:00:45 : INFO : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : CMN_1053 : : Rowdata: ( RowType=-4(error but dropped due to abort) Src Rowid=1186281 Targ Rowid=1186281
      NEXTVAL (ROW_WID:Int:): "(NULL)"
    I am not sure of this error and I have never seen before. Can any one share some ideas on how to tackle this error and make the task execute successfully.
    Many Thanks.

    Hi,
    fist check weather any special characteristics if not
    check in data source under this we have fields tab check the format of a particular field format internal/external/check u choose internal format, if any check routine once
    use Semantic Groups in the DTP.
    Try  it
    Thanku
    lokeeshM
    Edited by: lmedaSAP_BI on Oct 20, 2010 6:44 AM

  • DAC ETL tasks failing with Error Code

    Here is a copy of just 1 of the tasks that fail. I did search for this and found reference to increasing DB Processes to 500 but that had no effect.
    I have tried to reduce the number of Maximum Sessions in Setup to but it had no effect.
    Some of the tasks completed. The remaining stay as Failed even after re-running ETL.
    Using Informatica 8.1.1 SP5 32bit
    Build: 7.9.5.1.011209.1448
    Release Version: Oracle Business Intelligence Applications 7.9.5.1
    Package: 011209.1448
    Oracle 32bit 10g client
    Windows Server 2003 R2 64bit
    Details:
    pmcmd startworkflow -u Administrator -p **** -s 10.1.11.69:4006 -f SDE_ORAR12_Adaptor -lpf D:\OracleBI\DAC\Informatica\parameters\SDE_ORAR12_Adaptor.SDE_ORA_Stage_ValueSetHier_Extract_Full.txt SDE_ORA_Stage_ValueSetHier_Extract_Full
    Status Desc : Failed
    WorkFlowMessage :
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [8.1.1 SP5], build [186.0822], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2008
    All Rights Reserved.
    Invoked at Thu Apr 15 14:39:18 2010
    Connected to Integration Service at [10.1.11.69:4006]
    Folder: [SDE_ORAR12_Adaptor]
    Workflow: [SDE_ORA_Stage_ValueSetHier_Extract_Full] version [1].
    Workflow run status: [Failed]
    Workflow run error code: [36331]
    Workflow run error message: [WARNING: Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] failed and its "fail parent if this task fails" setting is turned on. So, Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] will be failed.]
    Start time: [Thu Apr 15 14:38:58 2010]
    End time: [Thu Apr 15 14:39:03 2010]
    Workflow log file: [d:\Informatica\PowerCenter8.1.1\server\infa_shared\WorkflowLogs\SDE_ORA_Stage_ValueSetHier_Extract_Full.log]
    Workflow run type: [User request]
    Run workflow as user: [Administrator]
    Integration Service: [PowerCenter_Integration_Service]
    Disconnecting from Integration Service
    Completed at Thu Apr 15 14:39:18 2010
    =====================================
    ERROR OUTPUT
    =====================================
    Error Message : Unknown reason for error code 36331
    ErrorCode : 36331
    Workflow Log.
    INFO : LM_36435 [Thu Apr 15 14:38:58 2010] : (4716|4956) Starting execution of workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] in folder [SDE_ORAR12_Adaptor] last saved by user [Administrator].
    INFO : LM_44195 [Thu Apr 15 14:38:58 2010] : (4716|4956) Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] service level [SLPriority:5,SLDispatchWaitTime:1800].
    INFO : VAR_27085 [Thu Apr 15 14:38:58 2010] : (4716|4956) Parameter file [d:\Informatica\PowerCenter8.1.1\server\infa_shared\Temp\SDE_ORA_Stage_ValueSetHier_Extract_Full_a04956] is opened for [workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full]].
    INFO : LM_36330 [Thu Apr 15 14:38:58 2010] : (4716|4956) Start task instance [Start]: Execution started.
    INFO : LM_36318 [Thu Apr 15 14:38:58 2010] : (4716|4956) Start task instance [Start]: Execution succeeded.
    INFO : LM_36505 : (4716|4956) Link [Start --> SDE_ORA_Stage_ValueSetHier_Extract_Full]: empty expression string, evaluated to TRUE.
    INFO : LM_36388 [Thu Apr 15 14:38:58 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] is waiting to be started.
    INFO : LM_36330 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution started.
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6793 Fetching initialization properties from the Integration Service. : (Thu Apr 15 14:38:58 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [DISP_20305 The [Preparer] DTM with process id [7972] is running on node [node01_OBIEEDEV01].
    : (Thu Apr 15 14:38:58 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [PETL_24036 Beginning the prepare phase for the session.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Connect to Repository].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Connect to Repository]. It took [0.234368] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6794 Connected to repository [PowerCenter] in domain [Domain_OBIEEDEV01] user [Administrator]]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Fetch Session from Repository].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Fetch Session from Repository]. It took [0.234367] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [ORA_R12] for session parameter:[$DBConnection_OLTP].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [9] for mapping parameter:[$$DATASOURCE_NUM_ID].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_ValueSetHier.$$LAST_EXTRACT_DATE].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Partition Group Formation].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Partition Group Formation]. It took [0.031249] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [PETL_24037 Finished the prepare phase for the session.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6792 Notifying the Integration Service that the prepare phase has been completed. : (Thu Apr 15 14:38:59 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6228 Writing session output to log file [d:\Informatica\PowerCenter8.1.1\server\infa_shared\SessLogs\SDE_ORAR12_Adaptor.SDE_ORA_Stage_ValueSetHier_Extract_Full.log].]
    INFO : LM_36682 [Thu Apr 15 14:38:59 2010] : (4716|4960) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: started a process with pid [7972] on node [node01_OBIEEDEV01].
    ERROR : LM_36320 [Thu Apr 15 14:39:03 2010] : (4716|4964) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution failed.
    WARNING : LM_36331 : (4716|4964) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] failed and its "fail parent if this task fails" setting is turned on. So, Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] will be failed.
    ERROR : LM_36320 [Thu Apr 15 14:39:03 2010] : (4716|4964) Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution failed.
    Thanks

    Can you please post the session log?
    Thanks,
    Austin

  • DAC: failed task during ETL for financial apps

    I am trying  my first ETL on OBIA 7.9.6.4
    i'm using  Oracle EBS 12.1.1 as source system.
    the ETL completes 314 tasks successfully ,but it fails the task named:
    "SDE_ORA_GL_AR_REV_LinkageInformation_Extract"
    DAC Error log:
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0903], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Wed Sep 18 09:46:41 2013
    Connected to Integration Service: [infor_int].
    Folder: [SDE_ORAR1211_Adaptor]
    Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Instance: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Mapping: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract]
    Session log file: [C:\Informatica\server\infa_shared\SessLogs\.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log]
    Source success rows: [0]
    Source failed rows: [0]
    Target success rows: [0]
    Target failed rows: [0]
    Number of transformation errors: [0]
    First error code [4035]
    First error message: [RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATI]
    Task run status: [Failed]
    Integration Service: [infor_int]
    Integration Service Process: [infor_int]
    Integration Service Grid: [infor_int]
    Node Name(s) [node01_AMAZON-9C628AAE]
    Preparation fragment
    Partition: [Partition #1]
    Transformation instance: [SQ_XLA_AE_LINES]
    Transformation: [SQ_XLA_AE_LINES]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [16004], message [ERROR: Prepare failed. : [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_CO]
    Start time: [Wed Sep 18 09:46:13 2013]
    End time: [Wed Sep 18 09:46:13 2013]
    Partition: [Partition #1]
    Transformation instance: [W_GL_LINKAGE_INFORMATION_GS]
    Transformation: [W_GL_LINKAGE_INFORMATION_GS]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [0], message [No errors encountered.]
    Start time: [Wed Sep 18 09:46:14 2013]
    End time: [Wed Sep 18 09:46:14 2013]
    Disconnecting from Integration Service
    Completed at Wed Sep 18 09:46:41 2013
    Informatica session logs:
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1211] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [26] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [01/01/1970] for mapping parameter:[$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [01/01/1990] for mapping parameter:[$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27028 Use override value ['NONE'] for mapping parameter:[$$LEDGER_TYPE_LIST].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] at [Wed Sep 18 09:46:13 2013].
    DIRECTOR> TM_6683 Repository Name: [infor_rep]
    DIRECTOR> TM_6684 Server Name: [infor_int]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1211_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] Run Instance Name: [] Run Id: [2130]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GL_AR_REV_LinkageInformation_Extract [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode ,Yes]
    DIRECTOR> TM_6708 Using configuration property [OraDateToTimestamp ,Yes]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,[APPS]@[ 54.225.65.108:1521:VIS] [DWH_REP2]@[AMAZON-9C628AAE:1521:obiaDW1]]
    DIRECTOR> TM_6703 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] is run by 32-bit Integration Service  [node01_AMAZON-9C628AAE], version [9.1.0 HotFix2], build [0903].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TE_7004 Transformation Parse Warning [IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || >>>>DISTRIBUTION_ID<<<<),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [JE_HEADER_ID || '~' || JE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>JE_HEADER_ID<<<< || '~' || JE_LINE_NUM<<PM Parse Warning>> [JE_LINE_NUM]: operand converted to a string
    ... JE_HEADER_ID || '~' || >>>>JE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [AE_HEADER_ID || '~' || AE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>AE_HEADER_ID<<<< || '~' || AE_LINE_NUM<<PM Parse Warning>> [AE_LINE_NUM]: operand converted to a string
    ... AE_HEADER_ID || '~' || >>>>AE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Sep 18 09:46:13 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Sep 18 09:46:13 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [54.225.65.108:1521/VIS], user [APPS]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [AMAZON-9C628AAE:1521/obiaDW1], user [DWH_REP2], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_LINKAGE_INFORMATION_GS :SQL INSERT statement:
    INSERT INTO W_GL_LINKAGE_INFORMATION_GS(SOURCE_DISTRIBUTION_ID,JOURNAL_LINE_INTEGRATION_ID,LEDGER_ID,LEDGER_TYPE,DISTRIBUTION_SOURCE,JE_BATCH_NAME,JE_HEADER_NAME,JE_LINE_NUM,POSTED_ON_DT,GL_ACCOUNT_ID,SLA_TRX_INTEGRATION_ID,DATASOURCE_NUM_ID)  VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GL_LINKAGE_INFORMATION_GS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [SQ_XLA_AE_LINES] User specified SQL Query [SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Sep 18 09:46:13 2013)
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Wed Sep 18 09:46:13 2013
    Target tables:
         W_GL_LINKAGE_INFORMATION_GS
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GL_LINKAGE_INFORMATION_GS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Wed Sep 18 09:46:13 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GL_LINKAGE_INFORMATION_GS (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
    WRT_8044 No data loaded for this target
    WRITER_1_*_1> WRT_8043 *****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GL_LINKAGE_INFORMATION_GS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Wed Sep 18 09:46:14 2013)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Wed Sep 18 09:46:14 2013)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_XLA_AE_LINES] (Instance Name: [SQ_XLA_AE_LINES])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GL_LINKAGE_INFORMATION_GS] (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] completed at [Wed Sep 18 09:46:14 2013].
    *I did some queries in my source database (Vision) , table "XLA_EVENTS" exists , column "UPG_BATCH_ID" also exists
    *I added "XLA_EVENTS" to the FROM clause and ran it in SQL Developer
    *in the SELECT clause ,i see a column named "AEHEADER.EVENT_TYPE_CODE"
    but there is no table named "AEHEADER" in the FROM clause
    so i added it manually , it's probably refers to "XLA_AE_HEADERS"
    Final query looks like this:
    SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_AE_HEADERS AEHEADER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    *when i run that query,it takes a lot of time executing without returning any results (last time it took 4 hours before i cancel it)
    my questions are:
    -what's wrong with that query?
    -how can i change the query in the workflow?
    could anyone please help?

    thank you very much
    i found SQ_XLA_AE_LINES and checked its SQL query,it's a very healthy query
    SELECT  DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
          AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , XLA_AE_HEADERS AEHEADER
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AEHEADER.APPLICATION_ID = 222
    AND XLA_EVENTS.APPLICATION_ID=222
    AND AEHEADER.AE_HEADER_ID = AELINE.AE_HEADER_ID
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND AEHEADER.EVENT_ID=XLA_EVENTS.EVENT_ID
    AND JHEADER.LAST_UPDATE_DATE >=
              TO_DATE('$$LAST_EXTRACT_DATE'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE($$FILTER_BY_LEDGER_ID, 'Y', T.LEDGER_ID, 1) IN ($$LEDGER_ID_LIST)
    AND DECODE($$FILTER_BY_LEDGER_TYPE, 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ($$LEDGER_TYPE_LIST)
    i compared this query with the query that appears in the Error messages,they are different (the Error message is stated in the first post)
    the query that appears in the Error messages misses a couple of lines,specifically in the "FROM" clause and the "WHERE" clause
    what might cause that issue?

  • Task fails while running Full load ETL

    Hi All,
    I am running full load ETL For Oracle R12(vanila Instance) HR But 4 tasks are failing SDE_ORA_JobDimention, SDE_ORA_HRPositionDimention, SDE_ORA_CodeDimension_Pay_level and SDE_ORA_CodeDimensionJob, I changed the parameter for all these task as mentioned in the Installation guide and rebuilled. Please help me out.
    Log is like this for SDE_ORA_JobDimention
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R12] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [9] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_JobDimension.$$JOBCODE_FLXFLD_SEGMENT_COL].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_JobDimension.$$JOBFAMILYCODE_FLXFLD_SEGMENT_COL].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_JobDimension.$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_JobDimension_Full] at [Fri Sep 26 10:52:05 2008]
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Base_Integration_Service]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR12_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_JobDimension_Full]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_JobDimension [version 1]
    DIRECTOR> TM_6827 [C:\Informatica\PowerCenter8.1.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_JobDimension_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6703 Session [SDE_ORA_JobDimension_Full] is run by 32-bit Integration Service [node01_HSCHBSCGN20031], version [8.1.1], build [0831].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 Session Sort Order: [Binary]
    MAPPING> TM_6156 Using LOW precision decimal arithmetic
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6307 DTM Error Log Disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_JobDimension_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Fri Sep 26 10:52:13 2008)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Fri Sep 26 10:52:14 2008)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 32000000 bytes and Block size is 1280000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [dev], user [apps]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [orcl], user [obia], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_JOB_DS :SQL INSERT statement:
    INSERT INTO W_JOB_DS(JOB_CODE,JOB_NAME,JOB_DESC,JOB_FAMILY_CODE,JOB_FAMILY_NAME,JOB_FAMILY_DESC,JOB_LEVEL,W_FLSA_STAT_CODE,W_FLSA_STAT_DESC,W_EEO_JOB_CAT_CODE,W_EEO_JOB_CAT_DESC,AAP_JOB_CAT_CODE,AAP_JOB_CAT_NAME,ACTIVE_FLG,CREATED_BY_ID,CHANGED_BY_ID,CREATED_ON_DT,CHANGED_ON_DT,AUX1_CHANGED_ON_DT,AUX2_CHANGED_ON_DT,AUX3_CHANGED_ON_DT,AUX4_CHANGED_ON_DT,SRC_EFF_FROM_DT,SRC_EFF_TO_DT,DELETE_FLG,DATASOURCE_NUM_ID,INTEGRATION_ID,TENANT_ID,X_CUSTOM) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_JOB_DS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    WRITER_1_*_1> WRT_8005 Writer run started.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [mplt_BC_ORA_JobDimension.Sq_Jobs] User specified SQL Query [SELECT
    PER_JOBS.JOB_ID,
    PER_JOBS.BUSINESS_GROUP_ID,
    PER_JOBS.JOB_DEFINITION_ID,
    PER_JOBS.DATE_FROM,
    PER_JOBS.DATE_TO,
    PER_JOBS.LAST_UPDATE_DATE AS CHANGED_ON_DT,      PER_JOBS.LAST_UPDATED_BY, PER_JOBS.CREATED_BY, PER_JOBS.CREATION_DATE,
    PER_JOB_DEFINITIONS.LAST_UPDATE_DATE AS AUX1_CHANGED_ON_DT,
    PER_JOB_DEFINITIONS.JOB_DEFINITION_ID,
    PER_JOBS.NAME,
    PER_JOBS.JOB_INFORMATION1, PER_JOBS.JOB_INFORMATION3,
    PER_JOBS. AS JOB_FAMILY_CODE, PER_JOB_DEFINITIONS.  AS JOB_CODE,
      '0' AS X_CUSTOM
    FROM
    PER_JOBS, PER_JOB_DEFINITIONS
    WHERE
    PER_JOBS.JOB_DEFINITION_ID = PER_JOB_DEFINITIONS.JOB_DEFINITION_ID]
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Fri Sep 26 10:53:05 2008
    Target tables:
    W_JOB_DS
    READER_1_1_1> RR_4049 SQL Query issued to database : (Fri Sep 26 10:53:05 2008)
    READER_1_1_1> CMN_1761 Timestamp Event: [Fri Sep 26 10:53:06 2008]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-01747: invalid user.table.column, table.column, or column specification
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    PER_JOBS.JOB_ID,
    PER_JOBS.BUSINESS_GROUP_ID,
    PER_JOBS.JOB_DEFINITION_ID,
    PER_JOBS.DATE_FROM,
    PER_JOBS.DATE_TO,
    PER_JOBS.LAST_UPDATE_DATE AS CHANGED_ON_DT, PER_JOBS.LAST_UPDATED_BY, PER_JOBS.CREATED_BY, PER_JOBS.CREATION_DATE,
    PER_JOB_DEFINITIONS.LAST_UPDATE_DATE AS AUX1_CHANGED_ON_DT,
    PER_JOB_DEFINITIONS.JOB_DEFINITION_ID,
    PER_JOBS.NAME,
    PER_JOBS.JOB_INFORMATION1, PER_JOBS.JOB_INFORMATION3,
    PER_JOBS. AS JOB_FAMILY_CODE, PER_JOB_DEFINITIONS. AS JOB_CODE,
    '0' AS X_CUSTOM
    FROM
    PER_JOBS, PER_JOB_DEFINITIONS
    WHERE
    PER_JOBS.JOB_DEFINITION_ID = PER_JOB_DEFINITIONS.JOB_DEFINITION_ID
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    PER_JOBS.JOB_ID,
    PER_JOBS.BUSINESS_GROUP_ID,
    PER_JOBS.JOB_DEFINITION_ID,
    PER_JOBS.DATE_FROM,
    PER_JOBS.DATE_TO,
    PER_JOBS.LAST_UPDATE_DATE AS CHANGED_ON_DT, PER_JOBS.LAST_UPDATED_BY, PER_JOBS.CREATED_BY, PER_JOBS.CREATION_DATE,
    PER_JOB_DEFINITIONS.LAST_UPDATE_DATE AS AUX1_CHANGED_ON_DT,
    PER_JOB_DEFINITIONS.JOB_DEFINITION_ID,
    PER_JOBS.NAME,
    PER_JOBS.JOB_INFORMATION1, PER_JOBS.JOB_INFORMATION3,
    PER_JOBS. AS JOB_FAMILY_CODE, PER_JOB_DEFINITIONS. AS JOB_CODE,
    '0' AS X_CUSTOM
    FROM
    PER_JOBS, PER_JOB_DEFINITIONS
    WHERE
    PER_JOBS.JOB_DEFINITION_ID = PER_JOB_DEFINITIONS.JOB_DEFINITION_ID
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Fri Sep 26 10:53:06 2008]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_JOB_DS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Fri Sep 26 10:53:06 2008
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_JOB_DS (Instance Name: [W_JOB_DS])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [mplt_BC_ORA_JobDimension.Sq_Jobs] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [mplt_BC_ORA_JobDimension.Sq_Jobs] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_JOB_DS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Fri Sep 26 10:53:06 2008)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Fri Sep 26 10:53:06 2008)
    MAPPING> TM_6018 Session [SDE_ORA_JobDimension_Full] run completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [Sq_Jobs] (Instance Name: [mplt_BC_ORA_JobDimension.Sq_Jobs])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_JOB_DS] (Instance Name: [W_JOB_DS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_JobDimension_Full] completed at [Fri Sep 26 10:53:07 2008]

    To make use of the warehouse you would probably want to connect to an EBS instance in order to populate the warehouse.
    Since the execution plan you intend to run is designed for the EBS data-model. I guess if you really didn't want to connect to the EBS instance to pull data you could build one using the universal adapter. This allows you to load out of flat-files if you wish, but I wouldn't reccomend making this a habit for actual implementation as it does create another potential point of failure (populating the flat-files).
    Thanks,
    Austin

Maybe you are looking for