DAC ORDERED TASK

Hi,
The DAC orders the tasks automaticly according to task's parent group and task's dependencies.
My question is : can i order the task manually or in another word to determine the depth of the task manually?
Thanks

Hi
DAC ordrers the tasks automatically using the source and target information which it gets after task synchronization.
If you want to keep your own dependency either you can add the target table of the preceeding task after which
you want to run this particular task.....or set the phase dependency....accordingly....
Regards

Similar Messages

  • Re-order Tasks

    How do you re-order Tasks on an Iphone 5s?.  They import from Outlook in the order the ywere entered. I cannot fidn a way to re-order tasks.

    Duplicated topic, please, continue here:
    *[https://support.mozilla.org/questions/990678 questions/990678]

  • Exporting work order task relationships to excel

    Hey,
    I am in the process of developing excel spreadsheets to let us copy and paste SAP outputs and generate MS Project schedules from this data. We have been able to do a copy paste of all the information required from SAP except for the relationships between work order tasks. Is there a transaction which will let me view, and copy, these relationships for each task within several work orders at the same time?
    I have had a look through the forum and couldn't find this question anywhere else. Hope someone can help me out!
    Thanks
    Dave

    To produce extracts for Excel with total flexibility in filtering, the Report Editor in P6 should be named the Extract Editor as I use it all the time. Create a one that loops through data source Activities (and apply filter for activity codes etc ), then add two child data sources for Successors and Predecessors and place any filters you need on each. Execute with output in csv format for Excel. This solution also has two advantages in the ability to extract across multiple projects and also if you have the option installed..to get P6 to schedule it to occur automatically .

  • DAC: failed task during ETL for financial apps

    I am trying  my first ETL on OBIA 7.9.6.4
    i'm using  Oracle EBS 12.1.1 as source system.
    the ETL completes 314 tasks successfully ,but it fails the task named:
    "SDE_ORA_GL_AR_REV_LinkageInformation_Extract"
    DAC Error log:
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0903], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Wed Sep 18 09:46:41 2013
    Connected to Integration Service: [infor_int].
    Folder: [SDE_ORAR1211_Adaptor]
    Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Instance: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Mapping: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract]
    Session log file: [C:\Informatica\server\infa_shared\SessLogs\.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log]
    Source success rows: [0]
    Source failed rows: [0]
    Target success rows: [0]
    Target failed rows: [0]
    Number of transformation errors: [0]
    First error code [4035]
    First error message: [RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATI]
    Task run status: [Failed]
    Integration Service: [infor_int]
    Integration Service Process: [infor_int]
    Integration Service Grid: [infor_int]
    Node Name(s) [node01_AMAZON-9C628AAE]
    Preparation fragment
    Partition: [Partition #1]
    Transformation instance: [SQ_XLA_AE_LINES]
    Transformation: [SQ_XLA_AE_LINES]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [16004], message [ERROR: Prepare failed. : [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_CO]
    Start time: [Wed Sep 18 09:46:13 2013]
    End time: [Wed Sep 18 09:46:13 2013]
    Partition: [Partition #1]
    Transformation instance: [W_GL_LINKAGE_INFORMATION_GS]
    Transformation: [W_GL_LINKAGE_INFORMATION_GS]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [0], message [No errors encountered.]
    Start time: [Wed Sep 18 09:46:14 2013]
    End time: [Wed Sep 18 09:46:14 2013]
    Disconnecting from Integration Service
    Completed at Wed Sep 18 09:46:41 2013
    Informatica session logs:
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1211] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [26] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [01/01/1970] for mapping parameter:[$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [01/01/1990] for mapping parameter:[$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27028 Use override value ['NONE'] for mapping parameter:[$$LEDGER_TYPE_LIST].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] at [Wed Sep 18 09:46:13 2013].
    DIRECTOR> TM_6683 Repository Name: [infor_rep]
    DIRECTOR> TM_6684 Server Name: [infor_int]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1211_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] Run Instance Name: [] Run Id: [2130]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GL_AR_REV_LinkageInformation_Extract [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode ,Yes]
    DIRECTOR> TM_6708 Using configuration property [OraDateToTimestamp ,Yes]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,[APPS]@[ 54.225.65.108:1521:VIS] [DWH_REP2]@[AMAZON-9C628AAE:1521:obiaDW1]]
    DIRECTOR> TM_6703 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] is run by 32-bit Integration Service  [node01_AMAZON-9C628AAE], version [9.1.0 HotFix2], build [0903].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TE_7004 Transformation Parse Warning [IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || >>>>DISTRIBUTION_ID<<<<),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [JE_HEADER_ID || '~' || JE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>JE_HEADER_ID<<<< || '~' || JE_LINE_NUM<<PM Parse Warning>> [JE_LINE_NUM]: operand converted to a string
    ... JE_HEADER_ID || '~' || >>>>JE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [AE_HEADER_ID || '~' || AE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>AE_HEADER_ID<<<< || '~' || AE_LINE_NUM<<PM Parse Warning>> [AE_LINE_NUM]: operand converted to a string
    ... AE_HEADER_ID || '~' || >>>>AE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Sep 18 09:46:13 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Sep 18 09:46:13 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [54.225.65.108:1521/VIS], user [APPS]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [AMAZON-9C628AAE:1521/obiaDW1], user [DWH_REP2], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_LINKAGE_INFORMATION_GS :SQL INSERT statement:
    INSERT INTO W_GL_LINKAGE_INFORMATION_GS(SOURCE_DISTRIBUTION_ID,JOURNAL_LINE_INTEGRATION_ID,LEDGER_ID,LEDGER_TYPE,DISTRIBUTION_SOURCE,JE_BATCH_NAME,JE_HEADER_NAME,JE_LINE_NUM,POSTED_ON_DT,GL_ACCOUNT_ID,SLA_TRX_INTEGRATION_ID,DATASOURCE_NUM_ID)  VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GL_LINKAGE_INFORMATION_GS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [SQ_XLA_AE_LINES] User specified SQL Query [SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Sep 18 09:46:13 2013)
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Wed Sep 18 09:46:13 2013
    Target tables:
         W_GL_LINKAGE_INFORMATION_GS
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GL_LINKAGE_INFORMATION_GS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Wed Sep 18 09:46:13 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GL_LINKAGE_INFORMATION_GS (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
    WRT_8044 No data loaded for this target
    WRITER_1_*_1> WRT_8043 *****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GL_LINKAGE_INFORMATION_GS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Wed Sep 18 09:46:14 2013)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Wed Sep 18 09:46:14 2013)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_XLA_AE_LINES] (Instance Name: [SQ_XLA_AE_LINES])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GL_LINKAGE_INFORMATION_GS] (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] completed at [Wed Sep 18 09:46:14 2013].
    *I did some queries in my source database (Vision) , table "XLA_EVENTS" exists , column "UPG_BATCH_ID" also exists
    *I added "XLA_EVENTS" to the FROM clause and ran it in SQL Developer
    *in the SELECT clause ,i see a column named "AEHEADER.EVENT_TYPE_CODE"
    but there is no table named "AEHEADER" in the FROM clause
    so i added it manually , it's probably refers to "XLA_AE_HEADERS"
    Final query looks like this:
    SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_AE_HEADERS AEHEADER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    *when i run that query,it takes a lot of time executing without returning any results (last time it took 4 hours before i cancel it)
    my questions are:
    -what's wrong with that query?
    -how can i change the query in the workflow?
    could anyone please help?

    thank you very much
    i found SQ_XLA_AE_LINES and checked its SQL query,it's a very healthy query
    SELECT  DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
          AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , XLA_AE_HEADERS AEHEADER
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AEHEADER.APPLICATION_ID = 222
    AND XLA_EVENTS.APPLICATION_ID=222
    AND AEHEADER.AE_HEADER_ID = AELINE.AE_HEADER_ID
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND AEHEADER.EVENT_ID=XLA_EVENTS.EVENT_ID
    AND JHEADER.LAST_UPDATE_DATE >=
              TO_DATE('$$LAST_EXTRACT_DATE'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE($$FILTER_BY_LEDGER_ID, 'Y', T.LEDGER_ID, 1) IN ($$LEDGER_ID_LIST)
    AND DECODE($$FILTER_BY_LEDGER_TYPE, 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ($$LEDGER_TYPE_LIST)
    i compared this query with the query that appears in the Error messages,they are different (the Error message is stated in the first post)
    the query that appears in the Error messages misses a couple of lines,specifically in the "FROM" clause and the "WHERE" clause
    what might cause that issue?

  • DAC set tasks depth manually change number

    I have an execution plan with only two tasks and both have depth of 0. One is SDE and the other is SIL. The problem is even thought SDE is set to ExtractFact, and the SIL is set to LoadFact, but when the plan runs, both are run in parallel. Or the Sil runs first then the SDE, which is not correct.
    How to set one tasks depth to 1 and the other to 2, or set parameter to run them sequentially.
    Looked in the Doc and nothing specific.
    Thanks

    As mentioned, the DAC will run the tasks based on dependencies. So if the SDE loads a Staging table that is the source for the SIL..then the DAC will order the SDE before that SIL. However, the DAC also does everything in parallel..which means that its not required to run ALL SDEs before ANY SILs. You can have a SIL run before a SDE (or PLP run at the same time as a SIL)...etc..as long as there is no dependencies. This is why even if a certain task fails, the DAC still continues to run non-dependent tasks before it finally FAILS the load. As long as your targets/sources and Foreign Keys are defined, the DAC will assemble based on dependencies and run all non-dependent tasks in parallel.

  • DAC ETL tasks failing with Error Code

    Here is a copy of just 1 of the tasks that fail. I did search for this and found reference to increasing DB Processes to 500 but that had no effect.
    I have tried to reduce the number of Maximum Sessions in Setup to but it had no effect.
    Some of the tasks completed. The remaining stay as Failed even after re-running ETL.
    Using Informatica 8.1.1 SP5 32bit
    Build: 7.9.5.1.011209.1448
    Release Version: Oracle Business Intelligence Applications 7.9.5.1
    Package: 011209.1448
    Oracle 32bit 10g client
    Windows Server 2003 R2 64bit
    Details:
    pmcmd startworkflow -u Administrator -p **** -s 10.1.11.69:4006 -f SDE_ORAR12_Adaptor -lpf D:\OracleBI\DAC\Informatica\parameters\SDE_ORAR12_Adaptor.SDE_ORA_Stage_ValueSetHier_Extract_Full.txt SDE_ORA_Stage_ValueSetHier_Extract_Full
    Status Desc : Failed
    WorkFlowMessage :
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [8.1.1 SP5], build [186.0822], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2008
    All Rights Reserved.
    Invoked at Thu Apr 15 14:39:18 2010
    Connected to Integration Service at [10.1.11.69:4006]
    Folder: [SDE_ORAR12_Adaptor]
    Workflow: [SDE_ORA_Stage_ValueSetHier_Extract_Full] version [1].
    Workflow run status: [Failed]
    Workflow run error code: [36331]
    Workflow run error message: [WARNING: Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] failed and its "fail parent if this task fails" setting is turned on. So, Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] will be failed.]
    Start time: [Thu Apr 15 14:38:58 2010]
    End time: [Thu Apr 15 14:39:03 2010]
    Workflow log file: [d:\Informatica\PowerCenter8.1.1\server\infa_shared\WorkflowLogs\SDE_ORA_Stage_ValueSetHier_Extract_Full.log]
    Workflow run type: [User request]
    Run workflow as user: [Administrator]
    Integration Service: [PowerCenter_Integration_Service]
    Disconnecting from Integration Service
    Completed at Thu Apr 15 14:39:18 2010
    =====================================
    ERROR OUTPUT
    =====================================
    Error Message : Unknown reason for error code 36331
    ErrorCode : 36331
    Workflow Log.
    INFO : LM_36435 [Thu Apr 15 14:38:58 2010] : (4716|4956) Starting execution of workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] in folder [SDE_ORAR12_Adaptor] last saved by user [Administrator].
    INFO : LM_44195 [Thu Apr 15 14:38:58 2010] : (4716|4956) Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] service level [SLPriority:5,SLDispatchWaitTime:1800].
    INFO : VAR_27085 [Thu Apr 15 14:38:58 2010] : (4716|4956) Parameter file [d:\Informatica\PowerCenter8.1.1\server\infa_shared\Temp\SDE_ORA_Stage_ValueSetHier_Extract_Full_a04956] is opened for [workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full]].
    INFO : LM_36330 [Thu Apr 15 14:38:58 2010] : (4716|4956) Start task instance [Start]: Execution started.
    INFO : LM_36318 [Thu Apr 15 14:38:58 2010] : (4716|4956) Start task instance [Start]: Execution succeeded.
    INFO : LM_36505 : (4716|4956) Link [Start --> SDE_ORA_Stage_ValueSetHier_Extract_Full]: empty expression string, evaluated to TRUE.
    INFO : LM_36388 [Thu Apr 15 14:38:58 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] is waiting to be started.
    INFO : LM_36330 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution started.
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6793 Fetching initialization properties from the Integration Service. : (Thu Apr 15 14:38:58 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [DISP_20305 The [Preparer] DTM with process id [7972] is running on node [node01_OBIEEDEV01].
    : (Thu Apr 15 14:38:58 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [PETL_24036 Beginning the prepare phase for the session.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Connect to Repository].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Connect to Repository]. It took [0.234368] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6794 Connected to repository [PowerCenter] in domain [Domain_OBIEEDEV01] user [Administrator]]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Fetch Session from Repository].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Fetch Session from Repository]. It took [0.234367] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [ORA_R12] for session parameter:[$DBConnection_OLTP].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27028 Use override value [9] for mapping parameter:[$$DATASOURCE_NUM_ID].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_ValueSetHier.$$LAST_EXTRACT_DATE].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6721 Started [Partition Group Formation].]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6722 Finished [Partition Group Formation]. It took [0.031249] seconds.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [PETL_24037 Finished the prepare phase for the session.]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6792 Notifying the Integration Service that the prepare phase has been completed. : (Thu Apr 15 14:38:59 2010)]
    INFO : LM_36488 [Thu Apr 15 14:38:59 2010] : (4716|4956) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] : [TM_6228 Writing session output to log file [d:\Informatica\PowerCenter8.1.1\server\infa_shared\SessLogs\SDE_ORAR12_Adaptor.SDE_ORA_Stage_ValueSetHier_Extract_Full.log].]
    INFO : LM_36682 [Thu Apr 15 14:38:59 2010] : (4716|4960) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: started a process with pid [7972] on node [node01_OBIEEDEV01].
    ERROR : LM_36320 [Thu Apr 15 14:39:03 2010] : (4716|4964) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution failed.
    WARNING : LM_36331 : (4716|4964) Session task instance [SDE_ORA_Stage_ValueSetHier_Extract_Full] failed and its "fail parent if this task fails" setting is turned on. So, Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full] will be failed.
    ERROR : LM_36320 [Thu Apr 15 14:39:03 2010] : (4716|4964) Workflow [SDE_ORA_Stage_ValueSetHier_Extract_Full]: Execution failed.
    Thanks

    Can you please post the session log?
    Thanks,
    Austin

  • How can I find error message according to DAC etl task error code?

    Hi all,
    I run the DAetl tasks accroding to the BI APPS installation guide for Financials - Receivables.
    While I found some task result in errors and restart the run can't resolve them.
    I looked at the error tasks and found in the task detail page some index creating step got error. It indicated error code 1542, howerver I can't find any valueable error message.
    Is't a dac error code or informatica error code? where is the document that list all the error code,error message, root cause and solutions?
    Thanks & Regards,
    Scott
    Edited by: Typhoon on Jun 30, 2009 11:21 AM
    Edited by: Typhoon on Jun 30, 2009 11:37 AM

    Hi
    First try to run the workflow of that mapping individually by hardcoding the parameter values........and running through Informatica monitor itself.......later you can revert the changes.......In this way you can findout whether there is some problem in Informatica mapping or DAC configuration....If the problem is at Informatica mapping I think you need to change the session property for target as "Normal" Instead of "Bulk"(because you have indexes being created for the target which would be set record wise.....). Hope this helps
    With Regards
    Ven
    Edited by: user11254581 on Jun 29, 2009 10:58 PM

  • DAC ETL tasks Completed with Error Code 36331

    Dear Sir,
    I have an execution plan task completed with error code 36331.
    I know it must be somthing wrong with the workflow.
    We are using BI APPS 7.9.5 with Informatica 8.1.1 SP4 on Windows Server 2003.
    I don't understand is why DAC didn't fail the task for this error code?
    How can I get to fail the task for such error.
    Great appreciation for any advise in advance.
    Roger

    Check this blog:
    http://siebel.ittoolbox.com/groups/technical-functional/siebel-analytics-l/complete-etl-ends-with-error-639247
    It might be helpful for you.
    Regards,
    Bhoomi

  • CProjects Internal Order, Tasks in CATS

    I am working on cProjects 4.0
    I want tasks (and project) to appear in CATS against which assigned resources would capture effort.
    Do I have to have the Controlling Method as "Internal Order (Structure Element, Manual)" to have the project and tasks appearing in CATS?
    In case I have the Controlling Method as "Internal Order (Project Definition, Automatic)", don't I get to see the tasks and Project definition in CATS?
    My requirement is that the project definition and tasks assigned to a resource in cProjects 4.0 should appear in CATS. By doing the later, my fear is that only the internal order (created/triggered from cProjects) would appear in CATS.
    Please suggest.

    Hi Bittu,
    I am not 100 percent get what you mean. The note suggested by Subramanian  should give you about the technical settings to configure the CATS.
    From my point of view, it does not matter, which controlling method are used. Certainly you have to check, the settings for the confirmation object, to the task or to the role. The user may get empty work list in CATS due to incorrect settings, but I think internal order will never be showed x in the CAT2 transaction only, there must be a the object of cProject (role or task) as well.
    Please take a look my other post regarding parameter settings for CATS profile as well.
    Kind regards,
    Zhenbo

  • DAC: failed tasks during ETL for financial apps

    I am trying  my first ETL on OBIA 7.9.6.4
    i'm using  Oracle EBS 12.1.1 as source system.
    the ETL completes 200+ tasks successfully ,but it fails the rest of them
    first task that fails is :
    SDE_ORA_GL_AR_REV_LinkageInformation_Extract
    Error message:
    All Task Batches
      Informatica Session Batch
      INFORMATICA TASK:SDE_ORAR1211_Adaptor:SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full:1:(Source : FULL Target : FULL)
      2013-09-03 14:57:14.627 Acquiring Resources
      2013-09-03 14:57:14.643 Acquired Resources
      2013-09-03 14:57:14.658 INFORMATICA TASK:SDE_ORAR1211_Adaptor:SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full:1:(Source : FULL Target : FULL) has started.
      ANOMALY INFO::: Error while executing : INFORMATICA TASK:SDE_ORAR1211_Adaptor:SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full:1:(Source : FULL Target : FULL)
      MESSAGE:::
      Irrecoverable Error
      pmcmd startworkflow -sv infor_int -d Domain_AMAZON-9C628AAE -u Administrator2 -p ****  -f SDE_ORAR1211_Adaptor  -paramfile C:\Informatica\server\infa_shared\SrcFiles\SDE_ORAR1211_Adaptor.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.txt  SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full
      Status Desc : Failed
      WorkFlowMessage :
      Error Message : Unknown reason for error code 36331
      ErrorCode : 36331
      EXCEPTION CLASS::: com.siebel.analytics.etl.etltask.IrrecoverableException
      com.siebel.analytics.etl.etltask.InformaticaTask.doExecute(InformaticaTask.java:254)
      com.siebel.analytics.etl.etltask.GenericTaskImpl.doExecuteWithRetries(GenericTaskImpl.java:477)
      com.siebel.analytics.etl.etltask.GenericTaskImpl.execute(GenericTaskImpl.java:372)
      com.siebel.analytics.etl.etltask.GenericTaskImpl.execute(GenericTaskImpl.java:253)
      com.siebel.analytics.etl.etltask.GenericTaskImpl.run(GenericTaskImpl.java:655)
      com.siebel.analytics.etl.taskmanager.XCallable.call(XCallable.java:63)
      java.util.concurrent.FutureTask$Sync.innerRun(Unknown Source)
      java.util.concurrent.FutureTask.run(Unknown Source)
      java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
      java.util.concurrent.FutureTask$Sync.innerRun(Unknown Source)
      java.util.concurrent.FutureTask.run(Unknown Source)
      java.util.concurrent.ThreadPoolExecutor$Worker.runTask(Unknown Source)
      java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
      java.lang.Thread.run(Unknown Source)
      (Number of retries : 1)
      pmcmd startworkflow -sv infor_int -d Domain_AMAZON-9C628AAE -u Administrator2 -p ****  -f SDE_ORAR1211_Adaptor  -paramfile C:\Informatica\server\infa_shared\SrcFiles\SDE_ORAR1211_Adaptor.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.txt  SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full
      2013-09-03 15:15:01.346 INFORMATICA TASK:SDE_ORAR1211_Adaptor:SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full:1:(Source : FULL Target : FULL) has finished execution with Failed status.
    (Failed)
    after that, many dependent tasks fail
    session logs at "SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full_SESSIONS.log":
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0903], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Tue Sep 03 15:14:24 2013
    Connected to Integration Service: [infor_int].
    Folder: [SDE_ORAR1211_Adaptor]
    Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Instance: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Mapping: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract]
    Session log file: [C:\Informatica\server\infa_shared\SessLogs\.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log]
    Source success rows: [0]
    Source failed rows: [0]
    Target success rows: [0]
    Target failed rows: [0]
    Number of transformation errors: [0]
    First error code [4035]
    First error message: [RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATI]
    Task run status: [Failed]
    Integration Service: [infor_int]
    Integration Service Process: [infor_int]
    Integration Service Grid: [infor_int]
    Node Name(s) [node01_AMAZON-9C628AAE]
    Preparation fragment
    Partition: [Partition #1]
    Transformation instance: [SQ_XLA_AE_LINES]
    Transformation: [SQ_XLA_AE_LINES]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [16004], message [ERROR: Prepare failed. : [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_CO]
    Start time: [Tue Sep 03 15:06:16 2013]
    End time: [Tue Sep 03 15:06:16 2013]
    Partition: [Partition #1]
    Transformation instance: [W_GL_LINKAGE_INFORMATION_GS]
    Transformation: [W_GL_LINKAGE_INFORMATION_GS]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [0], message [No errors encountered.]
    Start time: [Tue Sep 03 15:06:20 2013]
    End time: [Tue Sep 03 15:06:20 2013]
    Disconnecting from Integration Service
    Completed at Tue Sep 03 15:14:59 2013
    =====================================
    ERROR OUTPUT
    =====================================
    next failing task is:
    SDE_ORA_APTransactionFact_Payment_Full
    error log:
    java.lang.Thread.run(Unknown Source)
    305  SEVERE  Tue Sep 03 11:19:43 GMT 2013  Request to start workflow : 'SDE_ORAR1211_Adaptor:SDE_ORA_APTransactionFact_Payment_Full with the instance name SDE_ORA_APTransactionFact_Payment_Full' has completed with error code 0
    306  SEVERE  Tue Sep 03 11:20:16 GMT 2013  Request to start workflow : 'SDE_ORAR1211_Adaptor:SDE_ORA_APTransactionFact_PaymentSchedule_Full with the instance name SDE_ORA_APTransactionFact_PaymentSchedule_Full' has completed with error code 0
    307  SEVERE  Tue Sep 03 11:20:18 GMT 2013  Request to start workflow : 'SDE_ORAR1211_Adaptor:SDE_ORA_APTransactionFact_Distributions_Full with the instance name SDE_ORA_APTransactionFact_Distributions_Full' has completed with error code 0
    308  SEVERE  Tue Sep 03 11:20:24 GMT 2013  Request to start workflow : 'SDE_ORAR1211_Adaptor:SDE_ORA_Stage_ValueSetHier_Flatten with the instance name SDE_ORA_Stage_ValueSetHier_Flatten' has completed with error code 0
    311  SEVERE  Tue Sep 03 11:22:14 GMT 2013  MESSAGE:::No value for @DAC_SOURCE_PRUNED_REFRESH_TIMESTAMP available!
    i searched online but can't find useful information about the errors
    could anyone help , please?
    another question : can i continue the ETL when it fails? without starting it from the beginning?

    i got the query from informatica session logs of the failed task,and executed it in the source DB
    SELECT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    DLINK.ACCOUNTING_LINE_CODE LINE_CODE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
    JBATCH.NAME BATCH_NAME,
    JHEADER.NAME HEADER_NAME,
    PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES GLIMPREF
       , XLA_AE_LINES AELINE
       , XLA_AE_HEADERS AEHEADER
       , GL_JE_HEADERS JHEADER
       , GL_JE_BATCHES JBATCH
       , GL_LEDGERS T
       , GL_PERIODS PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN ('AP_INV_DIST','AP_PMT_DIST','AP_PREPAY')
    AND DLINK.APPLICATION_ID = 200
    AND AELINE.APPLICATION_ID = 200
    AND AEHEADER.APPLICATION_ID = 200
    AND AEHEADER.AE_HEADER_ID = AELINE.AE_HEADER_ID
    AND AEHEADER.UPG_BATCH_ID IS NULL
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID = DLINK.AE_HEADER_ID      
    AND AELINE.AE_LINE_NUM = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID = T.LEDGER_ID
    AND JHEADER.STATUS = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    the result returns 5000 rows ,seems like healthy data
    what do we get out of that?

  • WMS Sales Order Pick Load Task

    On WMS when picking sales order task if the task quantity is equal to the remaining quantity on the locator/LPN the system automatically loads the serials after you enter the locator and LPN details. Is it possible to enforce that users must always scan the serials and prevent the system from automatically loading the serials if its the quantity to pick agrees with the remaining LPN qunatity.
    Navigation
    WMS Mobile - Task - Directed Task - Interleaved Task - Accept Any Task

    Hi Rajiv,
    It would be good to flesh out the details a bit more, but Accellos One Warehouse should be able to accomplish what you are requesting.
    Feel free to send me a text with your email address so that we can correspond privately.
    Regards,
    Carl Marin
    +1 514 516 1861

  • To Integrate materialized view Refresh in DAC Execution Plan

    We have couple of materialized views which we want to get refreshed in DAC execution plan.
    We have referred Technote "BI_Apps796_Perf_Tech_Note_V8" which detailed how materialized view can be refreshed in DAC Execution Plan.
    We have followed below steps but our materized view is not getting refreshed.
    1. Create new Task Actions in DAC using "Tools -> Seed Data -> Actions -> Task Actions"
    with SQL Statement link
    BEGIN
    DBMS_MVIEW.REFRESH('getTableName()', 'C');
    DBMS_STATS.GATHER_TABLE_STATS(ownname => 'getTableOwner()', tabname=> 'getTableName()', cascade => TRUE, estimate_percent => DBMS_STATS.AUTO_SAMPLE_SIZE, method_opt => 'FOR ALL COLUMNS SIZE AUTO', degree => DBMS_STATS.DEFAULT_DEGREE);
    END;
    2. Defined custom MV as a table in DAC
    3. Added custom MV as the related table to the original Fact.
    4. Reassembled corresponding Subject Areas and rebuilt Execution plan
    But the MV is not getting refershed. Neither could locate any task in Ordered Task which will refresh this MV.
    Have we missed any step? Any pointers in this direction would be appericiated?
    Do we have to specify action .. to task "Load Fact table tas" and specify Success action as Refresh MV?
    How can we trace task in All tasks/Ordered tasks in DAC EP to locate task corresponding to MV refresh?

    Thanks Veeravalli for Suggesion!
    Tried adding Task action Refresh MV as action type "Success Action" for Fact task .. It did add a step of Refresh MV for load Fact Task in EP
    But it is passing the Orgininal Factable name as to be refreshed as MV which is giving DB error.
    "Refresh MV" for WC_B_TEST_F failed on "DataWarehouse"
    sql: BEGIN
    DBMS_MVIEW.REFRESH('WC_B_TEST_F', 'C');
    DBMS_STATS.GATHER_TABLE_STATS(ownname => 'olap', tabname=> 'WC_B_TEST_F',cascade => FALSE, estimate_percent => DBMS_STATS.AUTO_SAMPLE_SIZE, method_opt => 'FORALL COLUMNS SIZE AUTO', degree => DBMS_STATS.DEFAULT_DEGREE);
    END;
    Error: MESSAGE:::ORA-23401: materialized view "OLAP"."WC_B_TEST_F" does not exist
    ORA-06512: at "SYS.DBMS_SNAPSHOT", line 2563
    ORA-06512: at "SYS.DBMS_SNAPSHOT", line 2776
    ORA-06512: at "SYS.DBMS_SNAPSHOT", line 2745
    ORA-06512: at line 2
    instead of WC_B_TEST_F, we wanted "OLAP"."WC_B_TEST_MV" to be refreshed
    Do we need to create another task for MV refresh rather than task action? and add Refresh MV as task action there?
    Thanks,

  • GL configuration on DAC

    Hi,
    we are using source system is Oracle EBS 11.5.10 and
    BI Application 7.9.4.
    We have license only on GL and Profitability area. and we needs to configure ETL load for only GL and Profitability.
    In DAC there is seeded execution plan called Financials_Oralce 11.5.10.
    But that is included all Financial subject area (GL , COGS, AR, AP and Revenue).
    Can you please guide us that how to configure ETL load for only on GL & Profitability in DAC.
    and Please confirms us what are the CSV files or any configuration files needs to be configure for GL and Profitability load.
    if possible then send us document which is help us to configure GL & Profitability load.
    We have tried to find out from OBI Application and configuration guide. but we haven't found concrete solution from guide.
    Thanks & Regards,
    Bhavik
    Message was edited by:
    user629194

    Here are the configuration steps:
    http://download.oracle.com/docs/cd/E12127_01/doc/bia.795/e12083/anyimp_configfinance.htm#i1144647
    You will need to create a new Execution Plan in Custom Container to load only GL and Profitability (which I think is Revenue, but I am not sure).
    The EP for Financials_Oracle 11.5.10 contains these subject areas:
    Financials - Cost of Goods Sold
    Financials - General Ledger
    Financials - Payables
    Financials - Receivables
    Financials - Revenue
    For your new one only used GL and Revenue.
    Make sure you update your parameters and build your ordered tasks.

  • DAC as schedular tool for Informatica workflows(Not for OBIApps repository)

    I have to use DAC as administrative / scheduler tool for Informatica workflows. Earlier, I configured the same for OBI Apps. It's running fine. Now I am not refering to Oracle_BI_DW_Base.rep which is the inbuilt repository provided with OBI Apps. I have to run my workflows through DAC. I have gone through the below listed steps to achieve the same,
    1) Created New USER
    2) Used that user to configure DAC connection as well as for DAC repository tables. (am not sure whether it is a best practice or not)
    3) Created a new Source System Container. (no containers were there initially)
    4) I have folder in Informatica as "MyRep", so I created task logical and physical folders with the same name in DAC (Tools => Seed Data =>........)
    5) Created new Subject Area, tables, tasks, performed Synchronize Tasks.
    5) Set up informatica servers, Physical Data Sources. Tested them, no flaws.
    6) Added new Execution Plan.
    7) Assigned Subject Area to it.
    8) Clicked on "generate" button in parameters section of execution plan. No Parameters generated. Now got into confusion.
    *9) I couldn't find my tasks listed in ORDERED TASKS tab.*
    10) Clicked on BUILD. Then got error message as below,
    MESSAGE:::No tasks were found to build this execution plan.
    Please do let me know what is the wrong step here.
    Thanks,

    Thanks Ahsan, it solved my problem.
    I forgot to set Configuration Tags properly, later Assembled the subject area. Cheeeeeeerrrrrrsss... It's working
    Edited by: ABT on Dec 19, 2011 9:49 PM

  • Rg: Issues in Purchase order workflow.

    Hi All,
    I need to configure my workflow for purchase order to recive mail(SAP Inbox). These are the configuration I hvae done but still i am unable to recive mail.
    1. Created Characteristic.
    2. Created Class.
    3. Created Release group, Approval is 2 level , done with release strategy.
    4. The worklfow for release group is 1(Workflow bia plant,group) and agent is user
    The release strategy are wotrking fine.
    Now for configuration of recieving mail.
    1. I have created a role and in the authorization tab i have maintained the release group and release code and in user tab i have maintained the user.
    2. In SWDD i have copied the existing Release purchase order
      in task i have given release pur order task TS20000166 and in agent i have given rule 20000027.
    3. In PFTC i have given task which i have created in swdd  and created the trigerring events.
    Apart from this is there any configuration to be done.
    Please help regarding this.
    Thanks,
    Srivatsan

    Hi,
    Check the settings in t-code SO13
    you need to have an entery in automatig forwarding
    Good luck
    Shachar

Maybe you are looking for

  • How do I put my iphone pics on my computer?

    Help!! Im an Apple Idiot... I did it a while back and do not remember how to put my pictures taken from my iphone and put them on my computer!! Someone Please Help ME!!

  • Grouping issue- LiveCycle Designer 7.0

    Hello, We are trying for the past few days to create a dynamic PDF that uses an XML Data source to parse and display grouped information, We have taken a look at the samples but none of them seem to be helpful in this case. For ex. If we have a XML d

  • When / where to render - workflow question.

    Hello. I did try a couple of searches but did not seem to hit what I am looking for. I am a relativly new convert to Premiere Pro, came over from the Sony world. Purchased a EVGA GTX560 last week (used to be radeon based) and I am in awe. Loving thin

  • User ID Change

    Have Email as the ID when you log in, pass word is below it in the box that says Itunes password and you fill in password and hit ok. How do you change the email because the email it was set up with has gone away and now I need to set a new email but

  • IDOC to HTTP

    I need to send my IDOC's from R/3 to XI and then to our partners over HTTP(s). Could you please let me know what are all the necessary steps that I need to do? Thanks a lot.