Autodisplay based on interval

Hi,
Often we running a query every 30 / 60 / 90 seconds in sql plus.
Every time we used */ & Ente*r button.
Is there any command for autodisplay based on interval?

why dont u use a time stamp to display the output evry 30 sec ?
--Create table script
CREATE    TABLE static_interval (value_id NUMBER, value_date DATE, v_value NUMBER);
-- Data generation script (just for test purpose)
DECLARE
   v_interval   DATE   := TO_DATE ('05/8/2010', 'mm/dd/yyyy');
   v_date       DATE;
   v_value      NUMBER;
BEGIN
   v_date := v_interval;
   WHILE v_interval < v_date + 1
   LOOP
      v_value := ROUND (DBMS_RANDOM.VALUE (1, 5), 0);
      INSERT INTO static_interval
                  (value_id, value_date, v_value
           VALUES (18296, v_interval, v_value
      v_interval := v_interval + 5 / 1440;
   END LOOP;
   COMMIT;
END;
--Above script will generate 5 minute interval for the day with random values(v_value column of table created above) between 1 to 5.
--For example,  when I run the data generation script(though different run will produce different v_value) it generate following data.
--see this example u can get an idea
with t1 as (
            select  value_id,
                    value_date,
                    v_value,
                    case lag(v_value) over(order by value_date)
                      when v_value then 0
                      else 1
                    end start_of_group
              from  static_interval
     t2 as (
            select  value_id,
                    value_date,
                    v_value,
                    sum(start_of_group) over(order by value_date) grp
              from  t1
select  value_id,
        value_date,
        v_value,
        case
          when count(*) over(partition by grp) >= 4 and row_number() over(partition by grp order by value_date desc) = 1
            then min(value_date) over(partition by grp)
        end min_value_date,
        case
          when count(*) over(partition by grp) >= 4 and row_number() over(partition by grp order by value_date desc) = 1
            then max(value_date) over(partition by grp)
        end max_value_date
  from  t2
  order by value_date
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 00:00:00          4
     18296 05/08/2010 00:05:00          3
     18296 05/08/2010 00:10:00          3
     18296 05/08/2010 00:15:00          4
     18296 05/08/2010 00:20:00          3
     18296 05/08/2010 00:25:00          2
     18296 05/08/2010 00:30:00          3
     18296 05/08/2010 00:35:00          4
     18296 05/08/2010 00:40:00          2
     18296 05/08/2010 00:45:00          2
     18296 05/08/2010 00:50:00          5
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 00:55:00          2
     18296 05/08/2010 01:00:00          1
     18296 05/08/2010 01:05:00          4
     18296 05/08/2010 01:10:00          5
     18296 05/08/2010 01:15:00          5
     18296 05/08/2010 01:20:00          4
     18296 05/08/2010 01:25:00          4
     18296 05/08/2010 01:30:00          1
     18296 05/08/2010 01:35:00          2
     18296 05/08/2010 01:40:00          2
     18296 05/08/2010 01:45:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 01:50:00          2
     18296 05/08/2010 01:55:00          1
     18296 05/08/2010 02:00:00          4
     18296 05/08/2010 02:05:00          2
     18296 05/08/2010 02:10:00          4
     18296 05/08/2010 02:15:00          2
     18296 05/08/2010 02:20:00          2
     18296 05/08/2010 02:25:00          2
     18296 05/08/2010 02:30:00          4
     18296 05/08/2010 02:35:00          2
     18296 05/08/2010 02:40:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 02:45:00          2
     18296 05/08/2010 02:50:00          4
     18296 05/08/2010 02:55:00          5
     18296 05/08/2010 03:00:00          2
     18296 05/08/2010 03:05:00          4
     18296 05/08/2010 03:10:00          1
     18296 05/08/2010 03:15:00          3
     18296 05/08/2010 03:20:00          4
     18296 05/08/2010 03:25:00          2
     18296 05/08/2010 03:30:00          2
     18296 05/08/2010 03:35:00          4
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 03:40:00          3
     18296 05/08/2010 03:45:00          4
     18296 05/08/2010 03:50:00          5
     18296 05/08/2010 03:55:00          3
     18296 05/08/2010 04:00:00          2
     18296 05/08/2010 04:05:00          5
     18296 05/08/2010 04:10:00          2
     18296 05/08/2010 04:15:00          2
     18296 05/08/2010 04:20:00          4
     18296 05/08/2010 04:25:00          3
     18296 05/08/2010 04:30:00          1
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 04:35:00          3
     18296 05/08/2010 04:40:00          4
     18296 05/08/2010 04:45:00          2
     18296 05/08/2010 04:50:00          1
     18296 05/08/2010 04:55:00          4
     18296 05/08/2010 05:00:00          3
     18296 05/08/2010 05:05:00          2
     18296 05/08/2010 05:10:00          2
     18296 05/08/2010 05:15:00          3
     18296 05/08/2010 05:20:00          3
     18296 05/08/2010 05:25:00          4
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 05:30:00          2
     18296 05/08/2010 05:35:00          2
     18296 05/08/2010 05:40:00          3
     18296 05/08/2010 05:45:00          1
     18296 05/08/2010 05:50:00          3
     18296 05/08/2010 05:55:00          3
     18296 05/08/2010 06:00:00          1
     18296 05/08/2010 06:05:00          2
     18296 05/08/2010 06:10:00          4
     18296 05/08/2010 06:15:00          4
     18296 05/08/2010 06:20:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 06:25:00          2
     18296 05/08/2010 06:30:00          4
     18296 05/08/2010 06:35:00          5
     18296 05/08/2010 06:40:00          2
     18296 05/08/2010 06:45:00          1
     18296 05/08/2010 06:50:00          4
     18296 05/08/2010 06:55:00          1
     18296 05/08/2010 07:00:00          4
     18296 05/08/2010 07:05:00          4
     18296 05/08/2010 07:10:00          4
     18296 05/08/2010 07:15:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 07:20:00          3
     18296 05/08/2010 07:25:00          4
     18296 05/08/2010 07:30:00          4
     18296 05/08/2010 07:35:00          4
     18296 05/08/2010 07:40:00          1
     18296 05/08/2010 07:45:00          2
     18296 05/08/2010 07:50:00          1
     18296 05/08/2010 07:55:00          3
     18296 05/08/2010 08:00:00          5
     18296 05/08/2010 08:05:00          3
     18296 05/08/2010 08:10:00          5
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 08:15:00          2
     18296 05/08/2010 08:20:00          4
     18296 05/08/2010 08:25:00          3
     18296 05/08/2010 08:30:00          2
     18296 05/08/2010 08:35:00          2
     18296 05/08/2010 08:40:00          2
     18296 05/08/2010 08:45:00          1
     18296 05/08/2010 08:50:00          1
     18296 05/08/2010 08:55:00          5
     18296 05/08/2010 09:00:00          3
     18296 05/08/2010 09:05:00          4
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 09:10:00          3
     18296 05/08/2010 09:15:00          5
     18296 05/08/2010 09:20:00          2
     18296 05/08/2010 09:25:00          4
     18296 05/08/2010 09:30:00          4
     18296 05/08/2010 09:35:00          2
     18296 05/08/2010 09:40:00          1
     18296 05/08/2010 09:45:00          1
     18296 05/08/2010 09:50:00          4
     18296 05/08/2010 09:55:00          1
     18296 05/08/2010 10:00:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 10:05:00          1
     18296 05/08/2010 10:10:00          2
     18296 05/08/2010 10:15:00          4
     18296 05/08/2010 10:20:00          2
     18296 05/08/2010 10:25:00          4
     18296 05/08/2010 10:30:00          1
     18296 05/08/2010 10:35:00          5
     18296 05/08/2010 10:40:00          4
     18296 05/08/2010 10:45:00          4
     18296 05/08/2010 10:50:00          5
     18296 05/08/2010 10:55:00          5
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 11:00:00          3
     18296 05/08/2010 11:05:00          4
     18296 05/08/2010 11:10:00          1
     18296 05/08/2010 11:15:00          2
     18296 05/08/2010 11:20:00          2
     18296 05/08/2010 11:25:00          4
     18296 05/08/2010 11:30:00          1
     18296 05/08/2010 11:35:00          3
     18296 05/08/2010 11:40:00          3
     18296 05/08/2010 11:45:00          2
     18296 05/08/2010 11:50:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 11:55:00          4
     18296 05/08/2010 12:00:00          3
     18296 05/08/2010 12:05:00          3
     18296 05/08/2010 12:10:00          2
     18296 05/08/2010 12:15:00          1
     18296 05/08/2010 12:20:00          2
     18296 05/08/2010 12:25:00          4
     18296 05/08/2010 12:30:00          3
     18296 05/08/2010 12:35:00          2
     18296 05/08/2010 12:40:00          3
     18296 05/08/2010 12:45:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 12:50:00          4
     18296 05/08/2010 12:55:00          5
     18296 05/08/2010 13:00:00          2
     18296 05/08/2010 13:05:00          2
     18296 05/08/2010 13:10:00          1
     18296 05/08/2010 13:15:00          5
     18296 05/08/2010 13:20:00          5
     18296 05/08/2010 13:25:00          1
     18296 05/08/2010 13:30:00          5
     18296 05/08/2010 13:35:00          2
     18296 05/08/2010 13:40:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 13:45:00          3
     18296 05/08/2010 13:50:00          2
     18296 05/08/2010 13:55:00          2
     18296 05/08/2010 14:00:00          3
     18296 05/08/2010 14:05:00          4
     18296 05/08/2010 14:10:00          4
     18296 05/08/2010 14:15:00          3
     18296 05/08/2010 14:20:00          2
     18296 05/08/2010 14:25:00          4
     18296 05/08/2010 14:30:00          4
     18296 05/08/2010 14:35:00          5
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 14:40:00          3
     18296 05/08/2010 14:45:00          3
     18296 05/08/2010 14:50:00          4
     18296 05/08/2010 14:55:00          3
     18296 05/08/2010 15:00:00          3
     18296 05/08/2010 15:05:00          3
     18296 05/08/2010 15:10:00          2
     18296 05/08/2010 15:15:00          2
     18296 05/08/2010 15:20:00          5
     18296 05/08/2010 15:25:00          4
     18296 05/08/2010 15:30:00          5
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 15:35:00          4
     18296 05/08/2010 15:40:00          1
     18296 05/08/2010 15:45:00          2
     18296 05/08/2010 15:50:00          2
     18296 05/08/2010 15:55:00          1
     18296 05/08/2010 16:00:00          4
     18296 05/08/2010 16:05:00          5
     18296 05/08/2010 16:10:00          3
     18296 05/08/2010 16:15:00          5
     18296 05/08/2010 16:20:00          5
     18296 05/08/2010 16:25:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 16:30:00          5
     18296 05/08/2010 16:35:00          5
     18296 05/08/2010 16:40:00          2
     18296 05/08/2010 16:45:00          5
     18296 05/08/2010 16:50:00          3
     18296 05/08/2010 16:55:00          3
     18296 05/08/2010 17:00:00          3
     18296 05/08/2010 17:05:00          2
     18296 05/08/2010 17:10:00          2
     18296 05/08/2010 17:15:00          1
     18296 05/08/2010 17:20:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 17:25:00          3
     18296 05/08/2010 17:30:00          3
     18296 05/08/2010 17:35:00          3 05/08/2010 17:20:00 05/08/2010 17:35:00
     18296 05/08/2010 17:40:00          4
     18296 05/08/2010 17:45:00          2
     18296 05/08/2010 17:50:00          1
     18296 05/08/2010 17:55:00          4
     18296 05/08/2010 18:00:00          4
     18296 05/08/2010 18:05:00          5
     18296 05/08/2010 18:10:00          5
     18296 05/08/2010 18:15:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 18:20:00          4
     18296 05/08/2010 18:25:00          3
     18296 05/08/2010 18:30:00          3
     18296 05/08/2010 18:35:00          2
     18296 05/08/2010 18:40:00          4
     18296 05/08/2010 18:45:00          2
     18296 05/08/2010 18:50:00          2
     18296 05/08/2010 18:55:00          2
     18296 05/08/2010 19:00:00          2 05/08/2010 18:45:00 05/08/2010 19:00:00
     18296 05/08/2010 19:05:00          4
     18296 05/08/2010 19:10:00          4
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 19:15:00          5
     18296 05/08/2010 19:20:00          2
     18296 05/08/2010 19:25:00          2
     18296 05/08/2010 19:30:00          1
     18296 05/08/2010 19:35:00          5
     18296 05/08/2010 19:40:00          3
     18296 05/08/2010 19:45:00          3
     18296 05/08/2010 19:50:00          2
     18296 05/08/2010 19:55:00          3
     18296 05/08/2010 20:00:00          2
     18296 05/08/2010 20:05:00          3
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 20:10:00          2
     18296 05/08/2010 20:15:00          4
     18296 05/08/2010 20:20:00          3
     18296 05/08/2010 20:25:00          2
     18296 05/08/2010 20:30:00          5
     18296 05/08/2010 20:35:00          3
     18296 05/08/2010 20:40:00          5
     18296 05/08/2010 20:45:00          3
     18296 05/08/2010 20:50:00          3
     18296 05/08/2010 20:55:00          4
     18296 05/08/2010 21:00:00          1
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 21:05:00          3
     18296 05/08/2010 21:10:00          2
     18296 05/08/2010 21:15:00          1
     18296 05/08/2010 21:20:00          2
     18296 05/08/2010 21:25:00          4
     18296 05/08/2010 21:30:00          4
     18296 05/08/2010 21:35:00          5
     18296 05/08/2010 21:40:00          4
     18296 05/08/2010 21:45:00          1
     18296 05/08/2010 21:50:00          1
     18296 05/08/2010 21:55:00          4
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 22:00:00          5
     18296 05/08/2010 22:05:00          3
     18296 05/08/2010 22:10:00          3
     18296 05/08/2010 22:15:00          2
     18296 05/08/2010 22:20:00          4
     18296 05/08/2010 22:25:00          3
     18296 05/08/2010 22:30:00          3
     18296 05/08/2010 22:35:00          2
     18296 05/08/2010 22:40:00          2
     18296 05/08/2010 22:45:00          2
     18296 05/08/2010 22:50:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 22:55:00          2 05/08/2010 22:35:00 05/08/2010 22:55:00
     18296 05/08/2010 23:00:00          5
     18296 05/08/2010 23:05:00          2
     18296 05/08/2010 23:10:00          5
     18296 05/08/2010 23:15:00          4
     18296 05/08/2010 23:20:00          5
     18296 05/08/2010 23:25:00          4
     18296 05/08/2010 23:30:00          5
     18296 05/08/2010 23:35:00          3
     18296 05/08/2010 23:40:00          1
     18296 05/08/2010 23:45:00          2
  VALUE_ID VALUE_DATE             V_VALUE MIN_VALUE_DATE      MAX_VALUE_DATE
     18296 05/08/2010 23:50:00          1
     18296 05/08/2010 23:55:00          3
288 rows selected.regards,
friend
Edited by: most wanted!!!! on Oct 18, 2011 10:44 PM
Edited by: most wanted!!!! on Oct 18, 2011 10:53 PM

Similar Messages

  • Dynamic Grouping on Time Interval Parameters

    I have started a report that pulls volumes from our database dependant upon an Interval parameter. This function works just fine using the following code.
    WHERE
    (@Interval = 'Daily (Yesterday)') AND (Date_DateTime >= DATEADD(dd, -1, DATEDIFF(dd, 0, GETDATE())) AND Date_DateTime < DATEADD(dd, 0, DATEDIFF(dd, 0, GETDATE())))
    OR (@Interval = 'Weekly (Sun-Sat)') AND (Date_DateTime >= DATEADD(wk, DATEDIFF(wk, 7, GETDATE()), - 1) AND Date_DateTime <= dateadd(ms,- 3,dateadd(wk,datediff(wk,7,getdate()),6)))
    OR (@Interval = 'Weekly (Mon-Sun)') AND (Date_DateTime >= dateadd(wk,datediff(wk,7,getdate()),0) AND Date_DateTime <= dateadd(ms,- 3,dateadd(wk,datediff(wk,7,getdate()),7)))
    OR (@Interval = 'Weekly (Previous 7 days)') AND (Date_DateTime >= dateadd(day,datediff(day,0,GetDate())- 7,0) AND Date_DateTime <= dateadd(ms,- 3,dateadd(day,datediff(day,0,GetDate())- 0,0)))
    OR (@Interval = '4 previous weeks (Sun-Sat)') AND (Date_DateTime >= dateadd(wk,datediff(wk,7,getdate()),-22) AND Date_DateTime <= dateadd(ms,- 3,dateadd(wk,datediff(wk,7,getdate()),6)))
    OR (@Interval = 'Monthly (Last full Month)') AND (Date_DateTime >= dateadd(mm,-1,dateadd(mm,datediff(mm,0,getdate()),0)) AND Date_DateTime <= dateadd(ms,-3,dateadd(mm,0,dateadd(mm,datediff(mm,0,getdate()),0))))
    OR (@Interval = '3 Months (Last 3 full Months)') AND (Date_DateTime >= dateadd(mm,-3,dateadd(mm,datediff(mm,0,getdate()),0)) AND Date_DateTime <= dateadd(ms,-3,dateadd(mm,0,dateadd(mm,datediff(mm,0,getdate()),0))))
    OR (@Interval = '90 days (Current Date - 90)') AND (Date_DateTime >= DATEADD(dd, -90, DATEDIFF(dd, 0, GETDATE())) AND Date_DateTime <= DATEADD(dd, 0, DATEDIFF(dd, 0, GETDATE())))
    OR (@Interval = 'Yearly (Last full Year)') AND (Date_DateTime >= dateadd(yy,-1,dateadd(yy,datediff(yy,0,getdate()),0)) AND Date_DateTime <= dateadd(ms,-3,dateadd(yy,0,dateadd(yy,datediff(yy,0,getdate()),0))))
    OR (@Interval = 'Yearly (Last 12 full Months)') AND (Date_DateTime >= dateadd(mm,-13,dateadd(mm,datediff(mm,0,getdate()),0)) AND Date_DateTime <= dateadd(ms,-3,dateadd(mm,0,dateadd(mm,datediff(mm,0,getdate()),0))))
    OR (@Interval = 'Custom (Select Start Date and End Date)') AND (Date_DateTime >= @StartDate AND Date_DateTime <= @EndDate)
    Now I would like to be able to have my graph dynamically change dependant upon which Interval is selected. So for example if Daily or any of the Weekly Intervals are selected, the graph will show a Daily bar with the appropriate Date. If any of the Monthly
    or Yearly intervals are selected, it will show a Monthly graph with appropriate date.
    I can get it to work one way or the other, but I would like for it to be dynamic.

    Hi D.Hansen,
    According to your description, you want your chart to display records based on parameter selection. Right?
    In this scenario, since your where clause in query is working fine, so it can already return the records dynamically based on Interval selection. Now we just need to put the Date_DateTime into Category in the chart, it supposed to display records dynamically.
    Based on your information, we still can't figure out what the issue is. Please post some detail information or screenshots of the report design if possible. It may help us test your case in our local environment. Thank you.
    Best Regards,
    Simon Hou

  • SDE_ORAR1213_Adaptor.SDE_ORA_GLBalanceFact_Full

    Dears,
    when im trying run full load in finance analysts via dac, im getting error,
    Log file for your reference.
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1213] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1213.DATAWAREHOUSE.SDE_ORAR1213_Adaptor.SDE_ORA_GLBalanceFact_Full.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [1000] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[mplt_BC_ORA_GL_Balance_Fact.$$LEDGER_TYPE_LIST].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GLBalanceFact_Full] at [Fri Apr 12 22:05:36 2013].
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Base_Integration_Service]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1213_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GLBalanceFact_Full] Run Instance Name: [] Run Id: [2324]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GLBalanceFact [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [E:\Informatica\9.0.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_GLBalanceFact_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode,Yes]
    DIRECTOR> TM_6708 Using configuration property [ServerPort,4006]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,apps@R12PLY baw@orcl]
    DIRECTOR> TM_6708 Using configuration property [overrideMptlVarWithMapVar,Yes]
    DIRECTOR> TM_6703 Session [SDE_ORA_GLBalanceFact_Full] is run by 64-bit Integration Service [node01_OBIEETESTAPP], version [9.0.1 HotFix2], build [1111].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GLBalanceFact_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Fri Apr 12 22:05:37 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Fri Apr 12 22:05:37 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [R12PLY], user [apps]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [orcl], user [baw], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_ACCT_BUDGET_FS :SQL INSERT statement:
    INSERT INTO W_ACCT_BUDGET_FS(GL_ACCOUNT_ID,LEDGER_ID,BUDGET_LEDGER_ID,BUDGET_ID,PERIOD_BEGIN_DT,PERIOD_END_DT,BUDGET_LOC_AMT,LOC_CURR_CODE,CHANGED_BY_ID,CHANGED_ON_DT,AUX1_CHANGED_ON_DT,AUX2_CHANGED_ON_DT,ADJUSTMENT_FLG,DATASOURCE_NUM_ID,INTEGRATION_ID,TENANT_ID,X_CUSTOM) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8146 Writer: Target is database [orcl], user [baw], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_BALANCE_FS :SQL INSERT statement:
    INSERT INTO W_GL_BALANCE_FS(BUSN_AREA_ORG_ID,GL_ACCOUNT_ID,BALANCE_DATE,DB_CR_IND,BALANCE_ACCT_AMT,BALANCE_LOC_AMT,ACTIVITY_ACCT_AMT,ACTIVITY_LOC_AMT,ACCT_CURR_CODE,LOC_CURR_CODE,INTEGRATION_ID,DATASOURCE_NUM_ID,TENANT_ID,X_CUSTOM,ADJUSTMENT_FLAG,BALANCE_TYPE_FLAG,TRANSLATED_FLAG,SUMMARY_ACCOUNT_FLAG,LAST_UPDATE_DATE,AUX1_CHANGED_ON_DT,AUX2_CHANGED_ON_DT,LEDGER_ID) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_ACCT_BUDGET_FS, W_GL_BALANCE_FS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Fri Apr 12 22:05:37 2013
    Target tables:
    W_ACCT_BUDGET_FS
    W_GL_BALANCE_FS
    READER_1_1_1> RR_4029 SQ Instance [mplt_BC_ORA_GL_Balance_Fact.SQ_GL_BALANCES] User specified SQL Query [SELECT
    BAL.LEDGER_ID,
    BAL.CODE_COMBINATION_ID,
    BAL.CURRENCY_CODE,
    LED.CURRENCY_CODE,
    PER.PERIOD_NAME,
    BAL.ACTUAL_FLAG,
    BAL.TRANSLATED_FLAG,
    BAL.TEMPLATE_ID,
    BAL.PERIOD_NET_DR,
    BAL.PERIOD_NET_CR,
    ( BAL.BEGIN_BALANCE_DR + BAL.PERIOD_NET_DR ),
    ( BAL.BEGIN_BALANCE_CR + BAL.PERIOD_NET_CR) ,
    BAL.PERIOD_NET_DR_BEQ,
    BAL.PERIOD_NET_CR_BEQ,
    ( BAL.BEGIN_BALANCE_DR_BEQ + BAL.PERIOD_NET_DR_BEQ ) PERIOD_END_BALANCE_DR_BEQ,
    ( BAL.BEGIN_BALANCE_CR_BEQ + BAL.PERIOD_NET_CR_BEQ ) PERIOD_END_BALANCE_CR_BEQ,
    PER.START_DATE,
    PER.END_DATE,
    BAL.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_BAL,
    BAL.LAST_UPDATED_BY AS LAST_UPDATED_BY_BAL,
    PER.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_PERIODS,
    PER.LAST_UPDATED_BY AS LAST_UPDATED_BY_PERIODS,
    LED.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_SOB,
    LED.LAST_UPDATED_BY AS LAST_UPDATED_BY_SOB,
    BAL.BUDGET_VERSION_ID AS BUDGET_VERSION_ID,
    PER.ADJUSTMENT_PERIOD_FLAG AS ADJUSTMENT_PERIOD_FLAG,
    CASE WHEN
    BAL.TRANSLATED_FLAG = 'Y' THEN 'TRANSLATED'
    WHEN
    BAL.TRANSLATED_FLAG = 'R' THEN 'ENTERED_FOREIGN'
    WHEN
    BAL.CURRENCY_CODE = 'STAT' THEN 'STAT'
    WHEN
    ((BAL.PERIOD_NET_DR_BEQ = 0) OR (BAL.PERIOD_NET_DR_BEQ IS NULL)) AND
    ((BAL.PERIOD_NET_CR_BEQ = 0) OR (BAL.PERIOD_NET_CR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_DR_BEQ = 0) OR (BAL.BEGIN_BALANCE_DR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_CR_BEQ = 0) OR (BAL.BEGIN_BALANCE_CR_BEQ IS NULL))
    THEN 'BASE'
    ELSE 'ENTERED_LEDGER'
    END CURRENCY_BALANCE_TYPE
    FROM
    GL_BALANCES BAL
    , GL_LEDGERS LED
    , GL_PERIODS PER
    WHERE LED.LEDGER_ID = BAL.LEDGER_ID
    AND PER.PERIOD_SET_NAME = LED.PERIOD_SET_NAME
    AND BAL.PERIOD_NAME = PER.PERIOD_NAME
    AND BAL.PERIOD_TYPE = PER.PERIOD_TYPE
    AND NVL(BAL.TRANSLATED_FLAG, 'X') IN ('Y', 'X', 'R')
    AND BAL.ACTUAL_FLAG IN ( 'A','B')
    AND BAL.TEMPLATE_ID IS NULL
    AND
    BAL.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    OR PER.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    AND DECODE(, 'Y', LED.LEDGER_ID, 1) IN ()
    AND DECODE(, 'Y', LED.LEDGER_CATEGORY_CODE, 'NONE') IN ()]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Fri Apr 12 22:05:37 2013)
    READER_1_1_1> CMN_1761 Timestamp Event: [Fri Apr 12 22:05:37 2013]
    *READER_1_1_1> RR_4035 SQL Error [*
    ORA-00936: missing expression
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    BAL.LEDGER_ID,
    BAL.CODE_COMBINATION_ID,
    BAL.CURRENCY_CODE,
    LED.CURRENCY_CODE,
    PER.PERIOD_NAME,
    BAL.ACTUAL_FLAG,
    BAL.TRANSLATED_FLAG,
    BAL.TEMPLATE_ID,
    BAL.PERIOD_NET_DR,
    BAL.PERIOD_NET_CR,
    ( BAL.BEGIN_BALANCE_DR + BAL.PERIOD_NET_DR ),
    ( BAL.BEGIN_BALANCE_CR + BAL.PERIOD_NET_CR) ,
    BAL.PERIOD_NET_DR_BEQ,
    BAL.PERIOD_NET_CR_BEQ,
    ( BAL.BEGIN_BALANCE_DR_BEQ + BAL.PERIOD_NET_DR_BEQ ) PERIOD_END_BALANCE_DR_BEQ,
    ( BAL.BEGIN_BALANCE_CR_BEQ + BAL.PERIOD_NET_CR_BEQ ) PERIOD_END_BALANCE_CR_BEQ,
    PER.START_DATE,
    PER.END_DATE,
    BAL.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_BAL,
    BAL.LAST_UPDATED_BY AS LAST_UPDATED_BY_BAL,
    PER.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_PERIODS,
    PER.LAST_UPDATED_BY AS LAST_UPDATED_BY_PERIODS,
    LED.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_SOB,
    LED.LAST_UPDATED_BY AS LAST_UPDATED_BY_SOB,
    BAL.BUDGET_VERSION_ID AS BUDGET_VERSION_ID,
    PER.ADJUSTMENT_PERIOD_FLAG AS ADJUSTMENT_PERIOD_FLAG,
    CASE WHEN
    BAL.TRANSLATED_FLAG = 'Y' THEN 'TRANSLATED'
    WHEN
    BAL.TRANSLATED_FLAG = 'R' THEN 'ENTERED_FOREIGN'
    WHEN
    BAL.CURRENCY_CODE = 'STAT' THEN 'STAT'
    WHEN
    ((BAL.PERIOD_NET_DR_BEQ = 0) OR (BAL.PERIOD_NET_DR_BEQ IS NULL)) AND
    ((BAL.PERIOD_NET_CR_BEQ = 0) OR (BAL.PERIOD_NET_CR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_DR_BEQ = 0) OR (BAL.BEGIN_BALANCE_DR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_CR_BEQ = 0) OR (BAL.BEGIN_BALANCE_CR_BEQ IS NULL))
    THEN 'BASE'
    ELSE 'ENTERED_LEDGER'
    END CURRENCY_BALANCE_TYPE
    FROM
    GL_BALANCES BAL
    , GL_LEDGERS LED
    , GL_PERIODS PER
    WHERE LED.LEDGER_ID = BAL.LEDGER_ID
    AND PER.PERIOD_SET_NAME = LED.PERIOD_SET_NAME
    AND BAL.PERIOD_NAME = PER.PERIOD_NAME
    AND BAL.PERIOD_TYPE = PER.PERIOD_TYPE
    AND NVL(BAL.TRANSLATED_FLAG, 'X') IN ('Y', 'X', 'R')
    AND BAL.ACTUAL_FLAG IN ( 'A','B')
    AND BAL.TEMPLATE_ID IS NULL
    AND
    BAL.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    OR PER.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    AND DECODE(, 'Y', LED.LEDGER_ID, 1) IN ()
    AND DECODE(, 'Y', LED.LEDGER_CATEGORY_CODE, 'NONE') IN ()
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    BAL.LEDGER_ID,
    BAL.CODE_COMBINATION_ID,
    BAL.CURRENCY_CODE,
    LED.CURRENCY_CODE,
    PER.PERIOD_NAME,
    BAL.ACTUAL_FLAG,
    BAL.TRANSLATED_FLAG,
    BAL.TEMPLATE_ID,
    BAL.PERIOD_NET_DR,
    BAL.PERIOD_NET_CR,
    ( BAL.BEGIN_BALANCE_DR + BAL.PERIOD_NET_DR ),
    ( BAL.BEGIN_BALANCE_CR + BAL.PERIOD_NET_CR) ,
    BAL.PERIOD_NET_DR_BEQ,
    BAL.PERIOD_NET_CR_BEQ,
    ( BAL.BEGIN_BALANCE_DR_BEQ + BAL.PERIOD_NET_DR_BEQ ) PERIOD_END_BALANCE_DR_BEQ,
    ( BAL.BEGIN_BALANCE_CR_BEQ + BAL.PERIOD_NET_CR_BEQ ) PERIOD_END_BALANCE_CR_BEQ,
    PER.START_DATE,
    PER.END_DATE,
    BAL.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_BAL,
    BAL.LAST_UPDATED_BY AS LAST_UPDATED_BY_BAL,
    PER.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_PERIODS,
    PER.LAST_UPDATED_BY AS LAST_UPDATED_BY_PERIODS,
    LED.LAST_UPDATE_DATE AS LAST_UPDATE_DATE_SOB,
    LED.LAST_UPDATED_BY AS LAST_UPDATED_BY_SOB,
    BAL.BUDGET_VERSION_ID AS BUDGET_VERSION_ID,
    PER.ADJUSTMENT_PERIOD_FLAG AS ADJUSTMENT_PERIOD_FLAG,
    CASE WHEN
    BAL.TRANSLATED_FLAG = 'Y' THEN 'TRANSLATED'
    WHEN
    BAL.TRANSLATED_FLAG = 'R' THEN 'ENTERED_FOREIGN'
    WHEN
    BAL.CURRENCY_CODE = 'STAT' THEN 'STAT'
    WHEN
    ((BAL.PERIOD_NET_DR_BEQ = 0) OR (BAL.PERIOD_NET_DR_BEQ IS NULL)) AND
    ((BAL.PERIOD_NET_CR_BEQ = 0) OR (BAL.PERIOD_NET_CR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_DR_BEQ = 0) OR (BAL.BEGIN_BALANCE_DR_BEQ IS NULL)) AND
    ((BAL.BEGIN_BALANCE_CR_BEQ = 0) OR (BAL.BEGIN_BALANCE_CR_BEQ IS NULL))
    THEN 'BASE'
    ELSE 'ENTERED_LEDGER'
    END CURRENCY_BALANCE_TYPE
    FROM
    GL_BALANCES BAL
    , GL_LEDGERS LED
    , GL_PERIODS PER
    WHERE LED.LEDGER_ID = BAL.LEDGER_ID
    AND PER.PERIOD_SET_NAME = LED.PERIOD_SET_NAME
    AND BAL.PERIOD_NAME = PER.PERIOD_NAME
    AND BAL.PERIOD_TYPE = PER.PERIOD_TYPE
    AND NVL(BAL.TRANSLATED_FLAG, 'X') IN ('Y', 'X', 'R')
    AND BAL.ACTUAL_FLAG IN ( 'A','B')
    AND BAL.TEMPLATE_ID IS NULL
    AND
    BAL.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    OR PER.LAST_UPDATE_DATE >=
    TO_DATE('', 'MM/DD/YYYY HH24:MI:SS')
    AND DECODE(, 'Y', LED.LEDGER_ID, 1) IN ()
    AND DECODE(, 'Y', LED.LEDGER_CATEGORY_CODE, 'NONE') IN ()
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Fri Apr 12 22:05:37 2013]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_ACCT_BUDGET_FS, W_GL_BALANCE_FS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Fri Apr 12 22:05:37 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_ACCT_BUDGET_FS (Instance Name: [W_ACCT_BUDGET_FS])
    WRT_8044 No data loaded for this target
    WRT_8036 Target: W_GL_BALANCE_FS (Instance Name: [W_GL_BALANCE_FS])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [mplt_BC_ORA_GL_Balance_Fact.SQ_GL_BALANCES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [mplt_BC_ORA_GL_Balance_Fact.SQ_GL_BALANCES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_ACCT_BUDGET_FS, W_GL_BALANCE_FS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Fri Apr 12 22:05:37 2013)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Fri Apr 12 22:05:38 2013)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_GL_BALANCES] (Instance Name: [mplt_BC_ORA_GL_Balance_Fact.SQ_GL_BALANCES])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_ACCT_BUDGET_FS] (Instance Name: [W_ACCT_BUDGET_FS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> CMN_1740 Table: [W_GL_BALANCE_FS] (Instance Name: [W_GL_BALANCE_FS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GLBalanceFact_Full] completed at [Fri Apr 12 22:05:38 2013].

    Check this..
    BI Apps load - DAC variables not substituted.....
    Its very common error message.
    Ensure that you restart the informatica services and DAC services once you follow above resolution.
    please mark correct if it resolves your issue.
    Regards,
    Veeresh Rayan

  • DAC: failed task during ETL for financial apps

    I am trying  my first ETL on OBIA 7.9.6.4
    i'm using  Oracle EBS 12.1.1 as source system.
    the ETL completes 314 tasks successfully ,but it fails the task named:
    "SDE_ORA_GL_AR_REV_LinkageInformation_Extract"
    DAC Error log:
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0903], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Wed Sep 18 09:46:41 2013
    Connected to Integration Service: [infor_int].
    Folder: [SDE_ORAR1211_Adaptor]
    Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Instance: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Mapping: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract]
    Session log file: [C:\Informatica\server\infa_shared\SessLogs\.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log]
    Source success rows: [0]
    Source failed rows: [0]
    Target success rows: [0]
    Target failed rows: [0]
    Number of transformation errors: [0]
    First error code [4035]
    First error message: [RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATI]
    Task run status: [Failed]
    Integration Service: [infor_int]
    Integration Service Process: [infor_int]
    Integration Service Grid: [infor_int]
    Node Name(s) [node01_AMAZON-9C628AAE]
    Preparation fragment
    Partition: [Partition #1]
    Transformation instance: [SQ_XLA_AE_LINES]
    Transformation: [SQ_XLA_AE_LINES]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [16004], message [ERROR: Prepare failed. : [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_CO]
    Start time: [Wed Sep 18 09:46:13 2013]
    End time: [Wed Sep 18 09:46:13 2013]
    Partition: [Partition #1]
    Transformation instance: [W_GL_LINKAGE_INFORMATION_GS]
    Transformation: [W_GL_LINKAGE_INFORMATION_GS]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [0], message [No errors encountered.]
    Start time: [Wed Sep 18 09:46:14 2013]
    End time: [Wed Sep 18 09:46:14 2013]
    Disconnecting from Integration Service
    Completed at Wed Sep 18 09:46:41 2013
    Informatica session logs:
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1211] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [26] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [01/01/1970] for mapping parameter:[$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [01/01/1990] for mapping parameter:[$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27028 Use override value ['NONE'] for mapping parameter:[$$LEDGER_TYPE_LIST].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] at [Wed Sep 18 09:46:13 2013].
    DIRECTOR> TM_6683 Repository Name: [infor_rep]
    DIRECTOR> TM_6684 Server Name: [infor_int]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1211_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] Run Instance Name: [] Run Id: [2130]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GL_AR_REV_LinkageInformation_Extract [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode ,Yes]
    DIRECTOR> TM_6708 Using configuration property [OraDateToTimestamp ,Yes]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,[APPS]@[ 54.225.65.108:1521:VIS] [DWH_REP2]@[AMAZON-9C628AAE:1521:obiaDW1]]
    DIRECTOR> TM_6703 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] is run by 32-bit Integration Service  [node01_AMAZON-9C628AAE], version [9.1.0 HotFix2], build [0903].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TE_7004 Transformation Parse Warning [IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || >>>>DISTRIBUTION_ID<<<<),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [JE_HEADER_ID || '~' || JE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>JE_HEADER_ID<<<< || '~' || JE_LINE_NUM<<PM Parse Warning>> [JE_LINE_NUM]: operand converted to a string
    ... JE_HEADER_ID || '~' || >>>>JE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [AE_HEADER_ID || '~' || AE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>AE_HEADER_ID<<<< || '~' || AE_LINE_NUM<<PM Parse Warning>> [AE_LINE_NUM]: operand converted to a string
    ... AE_HEADER_ID || '~' || >>>>AE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Sep 18 09:46:13 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Sep 18 09:46:13 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [54.225.65.108:1521/VIS], user [APPS]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [AMAZON-9C628AAE:1521/obiaDW1], user [DWH_REP2], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_LINKAGE_INFORMATION_GS :SQL INSERT statement:
    INSERT INTO W_GL_LINKAGE_INFORMATION_GS(SOURCE_DISTRIBUTION_ID,JOURNAL_LINE_INTEGRATION_ID,LEDGER_ID,LEDGER_TYPE,DISTRIBUTION_SOURCE,JE_BATCH_NAME,JE_HEADER_NAME,JE_LINE_NUM,POSTED_ON_DT,GL_ACCOUNT_ID,SLA_TRX_INTEGRATION_ID,DATASOURCE_NUM_ID)  VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GL_LINKAGE_INFORMATION_GS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [SQ_XLA_AE_LINES] User specified SQL Query [SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Sep 18 09:46:13 2013)
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Wed Sep 18 09:46:13 2013
    Target tables:
         W_GL_LINKAGE_INFORMATION_GS
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GL_LINKAGE_INFORMATION_GS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Wed Sep 18 09:46:13 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GL_LINKAGE_INFORMATION_GS (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
    WRT_8044 No data loaded for this target
    WRITER_1_*_1> WRT_8043 *****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GL_LINKAGE_INFORMATION_GS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Wed Sep 18 09:46:14 2013)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Wed Sep 18 09:46:14 2013)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_XLA_AE_LINES] (Instance Name: [SQ_XLA_AE_LINES])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GL_LINKAGE_INFORMATION_GS] (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] completed at [Wed Sep 18 09:46:14 2013].
    *I did some queries in my source database (Vision) , table "XLA_EVENTS" exists , column "UPG_BATCH_ID" also exists
    *I added "XLA_EVENTS" to the FROM clause and ran it in SQL Developer
    *in the SELECT clause ,i see a column named "AEHEADER.EVENT_TYPE_CODE"
    but there is no table named "AEHEADER" in the FROM clause
    so i added it manually , it's probably refers to "XLA_AE_HEADERS"
    Final query looks like this:
    SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_AE_HEADERS AEHEADER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    *when i run that query,it takes a lot of time executing without returning any results (last time it took 4 hours before i cancel it)
    my questions are:
    -what's wrong with that query?
    -how can i change the query in the workflow?
    could anyone please help?

    thank you very much
    i found SQ_XLA_AE_LINES and checked its SQL query,it's a very healthy query
    SELECT  DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
          AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , XLA_AE_HEADERS AEHEADER
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AEHEADER.APPLICATION_ID = 222
    AND XLA_EVENTS.APPLICATION_ID=222
    AND AEHEADER.AE_HEADER_ID = AELINE.AE_HEADER_ID
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND AEHEADER.EVENT_ID=XLA_EVENTS.EVENT_ID
    AND JHEADER.LAST_UPDATE_DATE >=
              TO_DATE('$$LAST_EXTRACT_DATE'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE($$FILTER_BY_LEDGER_ID, 'Y', T.LEDGER_ID, 1) IN ($$LEDGER_ID_LIST)
    AND DECODE($$FILTER_BY_LEDGER_TYPE, 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ($$LEDGER_TYPE_LIST)
    i compared this query with the query that appears in the Error messages,they are different (the Error message is stated in the first post)
    the query that appears in the Error messages misses a couple of lines,specifically in the "FROM" clause and the "WHERE" clause
    what might cause that issue?

  • Help: I want to auto schedule a load using file watcher but it runs only once for the first time and after that it is not running at all

    Hi All,
    I am trying  to execute the below code as provided from one of the blogs. i am able to run the job only once based on a file watcher object(i.e. for very first time) and after that the job is not running at all and if  i schedule the job to run automatically based on interval of 10 or more minutes it is executing properly). Please let me know or guide me if i have missed any step or configuration.that is needed.
    Version of Oracle 11.2.0.1.0
    OS : Windows 7 Prof
    Given all the necessary privileges
    BEGIN
      DBMS_SCHEDULER.CREATE_CREDENTIAL(
         credential_name => 'cred',
         username        => 'XXXX',
         password        => 'XXXX');
    END;
    CREATE TABLE ZZZZ (WHEN timestamp, file_name varchar2(100),
       file_size number, processed char(1));
    CREATE OR REPLACE PROCEDURE YYYY
      (payload IN sys.scheduler_filewatcher_result) AS
    BEGIN
      INSERT INTO ZZZZ VALUES
         (payload.file_timestamp,
          payload.directory_path || '/' || payload.actual_file_name,
          payload.file_size,
          'N');
    END;
    BEGIN
      DBMS_SCHEDULER.CREATE_PROGRAM(
        program_name        => 'prog1',
        program_type        => 'stored_procedure',
        program_action      => 'YYYY',
        number_of_arguments => 1,
        enabled             => FALSE);
      DBMS_SCHEDULER.DEFINE_METADATA_ARGUMENT(
        program_name        => 'prog1',
        metadata_attribute  => 'event_message',
        argument_position   => 1);
      DBMS_SCHEDULER.ENABLE('prog1');
    END;
    BEGIN
      DBMS_SCHEDULER.CREATE_FILE_WATCHER(
        file_watcher_name => 'file_watcher1',
        directory_path    => 'D:\AAAA',
        file_name         => '*.txt',
        credential_name   => 'cred',
        destination       => NULL,
        enabled           => FALSE);
    END;
    BEGIN
      DBMS_SCHEDULER.CREATE_JOB(
        job_name        => 'job1',
        program_name    => 'prog1',
        queue_spec      => 'file_watcher1',
        auto_drop       => FALSE,
        enabled         => FALSE);
      DBMS_SCHEDULER.SET_ATTRIBUTE('job1','PARALLEL_INSTANCES',TRUE);
    END;
    EXEC DBMS_SCHEDULER.ENABLE('file_watcher1,job1');
    Regards,
    kumar.

    Please post a copy and paste of a complete run of a test case, similar to what I have shown below.
    SCOTT@orcl12c> SELECT banner FROM v$version
      2  /
    BANNER
    Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production
    PL/SQL Release 12.1.0.1.0 - Production
    CORE    12.1.0.1.0    Production
    TNS for 64-bit Windows: Version 12.1.0.1.0 - Production
    NLSRTL Version 12.1.0.1.0 - Production
    5 rows selected.
    SCOTT@orcl12c> CONN / AS SYSDBA
    Connected.
    SYS@orcl12c> -- set file watcher interval to one minute:
    SYS@orcl12c> BEGIN
      2    DBMS_SCHEDULER.SET_ATTRIBUTE
      3       ('file_watcher_schedule',
      4        'repeat_interval',
      5        'freq=minutely; interval=1');
      6  END;
      7  /
    PL/SQL procedure successfully completed.
    SYS@orcl12c> CONNECT scott/tiger
    Connected.
    SCOTT@orcl12c> BEGIN
      2    -- create credential using operating system user and password (fill in your own):
      3    DBMS_SCHEDULER.CREATE_CREDENTIAL
      4       (credential_name     => 'cred',
      5        username          => '...',
      6        password          => '...');
      7  END;
      8  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- create table to insert results into:
    SCOTT@orcl12c> CREATE TABLE ZZZZ
      2    (WHEN      timestamp,
      3      file_name varchar2(100),
      4      file_size number,
      5      processed char(1))
      6  /
    Table created.
    SCOTT@orcl12c> -- create procedure to insert results:
    SCOTT@orcl12c> CREATE OR REPLACE PROCEDURE YYYY
      2    (payload IN sys.scheduler_filewatcher_result)
      3  AS
      4  BEGIN
      5    INSERT INTO ZZZZ VALUES
      6        (payload.file_timestamp,
      7         payload.directory_path || '/' || payload.actual_file_name,
      8         payload.file_size,
      9         'N');
    10  END;
    11  /
    Procedure created.
    SCOTT@orcl12c> -- create program, define metadata, and enable:
    SCOTT@orcl12c> BEGIN
      2    DBMS_SCHEDULER.CREATE_PROGRAM
      3       (program_name          => 'prog1',
      4        program_type          => 'stored_procedure',
      5        program_action      => 'YYYY',
      6        number_of_arguments => 1,
      7        enabled          => FALSE);
      8    DBMS_SCHEDULER.DEFINE_METADATA_ARGUMENT(
      9       program_name         => 'prog1',
    10       metadata_attribute  => 'event_message',
    11       argument_position   => 1);
    12    DBMS_SCHEDULER.ENABLE ('prog1');
    13  END;
    14  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> BEGIN
      2    -- create file watcher:
      3    DBMS_SCHEDULER.CREATE_FILE_WATCHER
      4       (file_watcher_name   => 'file_watcher1',
      5        directory_path      => 'c:\my_oracle_files',
      6        file_name          => 'f*.txt',
      7        credential_name     => 'cred',
      8        destination          => NULL,
      9        enabled          => FALSE);
    10  END;
    11  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> BEGIN
      2    -- create job:
      3    DBMS_SCHEDULER.CREATE_JOB
      4       (job_name          => 'job1',
      5        program_name          => 'prog1',
      6        queue_spec          => 'file_watcher1',
      7        auto_drop          => FALSE,
      8        enabled          => FALSE);
      9    -- set attributes:
    10    DBMS_SCHEDULER.SET_ATTRIBUTE ('job1', 'PARALLEL_INSTANCES', TRUE);
    11  END;
    12  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- enable:
    SCOTT@orcl12c> EXEC DBMS_SCHEDULER.enable ('file_watcher1, job1');
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- write file (file must not exist previously):
    SCOTT@orcl12c> CREATE OR REPLACE DIRECTORY upncommon_dir AS 'c:\my_oracle_files'
      2  /
    Directory created.
    SCOTT@orcl12c> declare
      2    filtyp utl_file.file_type;
      3  begin
      4    filtyp := utl_file.fopen ('UPNCOMMON_DIR', 'file1.txt', 'W', NULL);
      5    utl_file.put_line (filtyp, 'File has arrived ' || SYSTIMESTAMP, TRUE);
      6    utl_file.fclose (filtyp);
      7  end;
      8  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- wait long enough (may take more than one minute) for job to run:
    SCOTT@orcl12c> EXEC DBMS_LOCK.SLEEP (100)
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- check for results:
    SCOTT@orcl12c> SELECT * FROM zzzz
      2  /
    WHEN
    FILE_NAME
    FILE_SIZE P
    22-OCT-13 10.12.28.309000 PM
    c:\my_oracle_files/file1.txt
            57 N
    1 row selected.
    SCOTT@orcl12c> declare
      2    filtyp utl_file.file_type;
      3  begin
      4    filtyp := utl_file.fopen ('UPNCOMMON_DIR', 'file2.txt', 'W', NULL);
      5    utl_file.put_line (filtyp, 'File has arrived ' || SYSTIMESTAMP, TRUE);
      6    utl_file.fclose (filtyp);
      7  end;
      8  /
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- wait long enough (may take more than one minute) for job to run:
    SCOTT@orcl12c> EXEC DBMS_LOCK.SLEEP (100)
    PL/SQL procedure successfully completed.
    SCOTT@orcl12c> -- check for results:
    SCOTT@orcl12c> SELECT * FROM zzzz
      2  /
    WHEN
    FILE_NAME
    FILE_SIZE P
    22-OCT-13 10.12.28.309000 PM
    c:\my_oracle_files/file1.txt
            57 N
    22-OCT-13 10.14.08.580000 PM
    c:\my_oracle_files/file2.txt
            57 N
    2 rows selected.

  • Pulling in a variable into TestStand from an externally running LabView Vi ?

    Hello,
    My coworker has a problem and I'm tyring to help him.
    We have a functional test written by an outside company, using TestStand 4.1
    He has a LabView Vi with an Thermotron driver that monitors  a chamber temp and outputs an INTERVAL number ( 1 intervel every x minutes)
    That interval number is written to text file.
    We would like to loop the Main function of the TestStand sequence based on that Interval number, but I'm unsure how to read that variable from within Test Stand.
    I can either pull interval numbers from the text file, or set up logic in the VI to make a true / false flag based on interval numbers
    i.e. interval 30 - 60 run the test....61 to 75 pause (dont run test)....76 to 125 run test...etc
    Can you tell me how I might pull that interval into TestStand ?
    Is there a way without incorporating the stand alone VI in question into the Test Seq ?
    Thank you for any advice in advance !
    Solved!
    Go to Solution.

    Hi,
    Are you saying that you want to read this value stored in your text file and use this value to set the number if iteration of your test sequence loop.
    You will need a piece of software that will obtain the value from your file and then pass this back into TestStand to store in a teststand variable, probably a local. TestStand doesn't have the built in function to directly read the data from your file.
    Therefore you will need a code module attached to say an Action Step Type. The code module can be written in any one of the supported language, as you mentioned LabVIEW then maybe this is your perferred language. So your VI need to obtain the data from your file and return the value back to TestStand and you do this by sending your data out of your VI via its connector pane and then in teststand assigning a Teststand variable to your output connection of your VI.
    I am assuming you have the basic knowledge of setting up a Step in TestStand.
    Regards
    Ray Farmer

  • Log file

    Dears,
    Running finance module using dac. this task has been failed.
    log file for your reference.
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1213] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1213.DATAWAREHOUSE.SDE_ORAR1213_Adaptor.SDE_ORA_CodeDimension_Bank_Cat.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[MPLT_ADI_CODES.$$CATEGORY].
    DIRECTOR> VAR_27028 Use override value [1000] for mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[MPLT_SA_ORA_CODES.$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_CodeDimension_Bank_Cat] at [Sat Apr 13 20:10:18 2013].
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Base_Integration_Service]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1213_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_CodeDimension_Bank_Cat] Run Instance Name: [] Run Id: [2791]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_CodeDimension_Ap_Lookup [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [E:\Informatica\9.0.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_CodeDimension_Bank_Cat].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode,Yes]
    DIRECTOR> TM_6708 Using configuration property [ServerPort,4006]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,apps@R12PLY baw@orcl]
    DIRECTOR> TM_6708 Using configuration property [overrideMptlVarWithMapVar,Yes]
    DIRECTOR> TM_6703 Session [SDE_ORA_CodeDimension_Bank_Cat] is run by 64-bit Integration Service [node01_OBIEETESTAPP], version [9.0.1 HotFix2], build [1111].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6155 Using HIGH precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [orcl], user [baw]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_CodeDimension_Bank_Cat]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Sat Apr 13 20:10:18 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Sat Apr 13 20:10:18 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 32000000 bytes and Block size is 128000 bytes.
    LKPDP_1> DBG_21097 Lookup Transformation [mplt_ADI_Codes.Lkp_Master_Map]: Default sql to create lookup cache: SELECT MASTER_CODE,DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE FROM W_MASTER_MAP_D ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,MASTER_CODE
    LKPDP_3> DBG_21312 Lookup Transformation [mplt_ADI_Codes.Lkp_W_CODE_D]: Lookup override sql to create cache: SELECT W_CODE_D.SOURCE_NAME_1 AS SOURCE_NAME_1, W_CODE_D.SOURCE_NAME_2 AS SOURCE_NAME_2, W_CODE_D.MASTER_DATASOURCE_NUM_ID AS MASTER_DATASOURCE_NUM_ID, W_CODE_D.MASTER_CODE AS MASTER_CODE, W_CODE_D.MASTER_VALUE AS MASTER_VALUE, W_CODE_D.W_INSERT_DT AS W_INSERT_DT, W_CODE_D.TENANT_ID AS TENANT_ID, W_CODE_D.DATASOURCE_NUM_ID AS DATASOURCE_NUM_ID, W_CODE_D.SOURCE_CODE AS SOURCE_CODE, W_CODE_D.CATEGORY AS CATEGORY, W_CODE_D.LANGUAGE_CODE AS LANGUAGE_CODE FROM W_CODE_D
    WHERE
    W_CODE_D.CATEGORY IN () ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,SOURCE_NAME_1,SOURCE_NAME_2,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,TENANT_ID
    LKPDP_2> DBG_21097 Lookup Transformation [mplt_ADI_Codes.Lkp_Master_Code]: Default sql to create lookup cache: SELECT MASTER_VALUE,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,CATEGORY,LANGUAGE_CODE FROM W_MASTER_CODE_D ORDER BY MASTER_DATASOURCE_NUM_ID,MASTER_CODE,CATEGORY,LANGUAGE_CODE,MASTER_VALUE
    LKPDP_1> TE_7212 Increasing [Index Cache] size for transformation [mplt_ADI_Codes.Lkp_Master_Map] from [1000000] to [2611200].
    LKPDP_1> TE_7212 Increasing [Data Cache] size for transformation [mplt_ADI_Codes.Lkp_Master_Map] from [2000000] to [2007040].
    LKPDP_3> TE_7212 Increasing [Index Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [1000000] to [2611200].
    LKPDP_3> TE_7212 Increasing [Data Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [2000000] to [2007040].
    LKPDP_2> TE_7212 Increasing [Index Cache] size for transformation [mplt_ADI_Codes.Lkp_Master_Code] from [1000000] to [2611200].
    LKPDP_2> TE_7212 Increasing [Data Cache] size for transformation [mplt_ADI_Codes.Lkp_Master_Code] from [2000000] to [2007040].
    READER_1_1_1> DBG_21438 Reader: Source is [R12PLY], user [apps]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8147 Writer: Target is database [orcl], user [baw], bulk mode [OFF]
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL INSERT statement:
    INSERT INTO W_CODE_D(DATASOURCE_NUM_ID,SOURCE_CODE,SOURCE_CODE_1,SOURCE_CODE_2,SOURCE_CODE_3,SOURCE_NAME_1,SOURCE_NAME_2,CATEGORY,LANGUAGE_CODE,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,W_UPDATE_DT,TENANT_ID) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL UPDATE statement:
    UPDATE W_CODE_D SET SOURCE_CODE_1 = ?, SOURCE_CODE_2 = ?, SOURCE_CODE_3 = ?, SOURCE_NAME_1 = ?, SOURCE_NAME_2 = ?, MASTER_DATASOURCE_NUM_ID = ?, MASTER_CODE = ?, MASTER_VALUE = ?, W_INSERT_DT = ?, W_UPDATE_DT = ?, TENANT_ID = ? WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL DELETE statement:
    DELETE FROM W_CODE_D WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_CODE_D]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes] User specified SQL Query [SELECT AP_LOOKUP_CODES.LOOKUP_CODE, AP_LOOKUP_CODES.LOOKUP_TYPE, AP_LOOKUP_CODES.DESCRIPTION
    FROM
    AP_LOOKUP_CODES
    WHERE
    LOOKUP_TYPE =   'ACCOUNT TYPE']
    READER_1_1_1> RR_4049 SQL Query issued to database : (Sat Apr 13 20:10:19 2013)
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Sat Apr 13 20:10:19 2013
    Target tables:
    W_CODE_D
    READER_1_1_1> RR_4050 First row returned from database to reader : (Sat Apr 13 20:10:19 2013)
    READER_1_1_1> BLKR_16019 Read [4] rows, read [0] error rows for source table [AP_LOOKUP_CODES] instance name [mplt_BC_ORA_Codes_Ap_Lookup.AP_LOOKUP_CODES]
    READER_1_1_1> BLKR_16008 Reader run completed.
    LKPDP_3> TM_6660 Total Buffer Pool size is 609824 bytes and Block size is 65536 bytes.
    LKPDP_3:READER_1_1> DBG_21438 Reader: Source is [orcl], user [baw]
    LKPDP_3:READER_1_1> BLKR_16003 Initialization completed successfully.
    LKPDP_3:READER_1_1> BLKR_16007 Reader run started.
    LKPDP_3:READER_1_1> RR_4049 SQL Query issued to database : (Sat Apr 13 20:10:19 2013)
    LKPDP_3:READER_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    LKPDP_3:READER_1_1> RR_4035 SQL Error [
    ORA-00936: missing expression
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT W_CODE_D.SOURCE_NAME_1 AS SOURCE_NAME_1, W_CODE_D.SOURCE_NAME_2 AS SOURCE_NAME_2, W_CODE_D.MASTER_DATASOURCE_NUM_ID AS MASTER_DATASOURCE_NUM_ID, W_CODE_D.MASTER_CODE AS MASTER_CODE, W_CODE_D.MASTER_VALUE AS MASTER_VALUE, W_CODE_D.W_INSERT_DT AS W_INSERT_DT, W_CODE_D.TENANT_ID AS TENANT_ID, W_CODE_D.DATASOURCE_NUM_ID AS DATASOURCE_NUM_ID, W_CODE_D.SOURCE_CODE AS SOURCE_CODE, W_CODE_D.CATEGORY AS CATEGORY, W_CODE_D.LANGUAGE_CODE AS LANGUAGE_CODE FROM W_CODE_D
    WHERE
    W_CODE_D.CATEGORY IN () ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,SOURCE_NAME_1,SOURCE_NAME_2,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,TENANT_ID
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT W_CODE_D.SOURCE_NAME_1 AS SOURCE_NAME_1, W_CODE_D.SOURCE_NAME_2 AS SOURCE_NAME_2, W_CODE_D.MASTER_DATASOURCE_NUM_ID AS MASTER_DATASOURCE_NUM_ID, W_CODE_D.MASTER_CODE AS MASTER_CODE, W_CODE_D.MASTER_VALUE AS MASTER_VALUE, W_CODE_D.W_INSERT_DT AS W_INSERT_DT, W_CODE_D.TENANT_ID AS TENANT_ID, W_CODE_D.DATASOURCE_NUM_ID AS DATASOURCE_NUM_ID, W_CODE_D.SOURCE_CODE AS SOURCE_CODE, W_CODE_D.CATEGORY AS CATEGORY, W_CODE_D.LANGUAGE_CODE AS LANGUAGE_CODE FROM W_CODE_D
    WHERE
    W_CODE_D.CATEGORY IN () ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,SOURCE_NAME_1,SOURCE_NAME_2,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,TENANT_ID
    Oracle Fatal Error].
    LKPDP_3:READER_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    LKPDP_3:READER_1_1> BLKR_16004 ERROR: Prepare failed.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_ADI_Codes.Lkp_W_CODE_D], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_ADI_Codes.Exp_Master_Code_Lookup], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_ADI_Codes.Exp_Master_Code_Lookup], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_ADI_Codes.Exp_Master_Map_Lookup], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_ADI_Codes.Exp_Master_Map_Lookup], and the session is terminating.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_CODE_D] at end of load
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Fil_Code_Valid], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    WRITER_1_*_1> WRT_8035 Load complete time: Sat Apr 13 20:10:19 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_CODE_D (Instance Name: [W_CODE_D])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Fil_Code_Valid], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Exp_Code_Cleanse], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Exp_Code_Cleanse], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    MANAGER> PETL_24007 Received request to stop session run. Attempting to stop worker threads.
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Exp_Code_Name_Resolution], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_SA_ORA_Codes.Exp_Code_Name_Resolution], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_BC_ORA_Codes_Ap_Lookup.Exp_Ap_Lookup_Codes], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_BC_ORA_Codes_Ap_Lookup.Exp_Ap_Lookup_Codes], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes], and the session is terminating.
    TRANSF_1_1_1> CMN_1761 Timestamp Event: [Sat Apr 13 20:10:19 2013]
    TRANSF_1_1_1> TM_6085 A fatal error occurred at transformation [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes], and the session is terminating.
    TRANSF_1_1_1> DBG_21511 TE: Fatal Transformation Error.
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_CODE_D] has completed. The total run time was insufficient for any meaningful statistics.
    MAPPING> CMN_1793 The index cache size that would hold [0] rows in the lookup table for [mplt_ADI_Codes.Lkp_W_CODE_D], in memory, is [0] bytes
    MAPPING> CMN_1792 The data cache size that would hold [0] rows in the lookup table for [mplt_ADI_Codes.Lkp_W_CODE_D], in memory, is [0] bytes
    MANAGER> PETL_24005 Starting post-session tasks. : (Sat Apr 13 20:10:19 2013)
    MANAGER> PETL_24007 Received request to stop session run. Attempting to stop worker threads.
    MANAGER> PETL_24007 Received request to stop session run. Attempting to stop worker threads.
    MANAGER> PETL_24029 Post-session task completed successfully. : (Sat Apr 13 20:10:19 2013)
    MAPPING> TE_7216 Deleting cache files [PMLKUP13765_131084_0_2791W32] for transformation [mplt_ADI_Codes.Lkp_Master_Map].
    MAPPING> TE_7216 Deleting cache files [PMLKUP13765_131081_0_2791W32] for transformation [mplt_ADI_Codes.Lkp_W_CODE_D].
    MAPPING> TE_7216 Deleting cache files [PMLKUP13765_131083_0_2791W32] for transformation [mplt_ADI_Codes.Lkp_Master_Code].
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [Sq_Ap_Lookup_Codes] (Instance Name: [mplt_BC_ORA_Codes_Ap_Lookup.Sq_Ap_Lookup_Codes])
         Output Rows [4], Affected Rows [4], Applied Rows [4], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_CODE_D] (Instance Name: [W_CODE_D])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_CodeDimension_Bank_Cat] completed at [Sat Apr 13 20:10:20 2013].

    Hi 966148,
    All your code combination tasks refer Category as the parameter which DAC passes.
    If you did not follow the instruction what I gave in
    SDE_ORAR1213_Adaptor.SDE_ORA_GLBalanceFact_Full
    then all your code tasks fails with Database driver error.
    If you followed then Close the thread by saying it fixed your issue.
    Mark helpfull or correct.
    Regards,
    Veeresh Rayan

  • SQL Statement error when running ETL from DAC

    Dear all,
    I installed and configured biapps 7.9.6.3, then I run the full load of Subject CRM - Loyalty in DAC. And I got the error of task fail, I check the Session log files in Informatica server.
    $ view SEBL_VERT_811.DATAWAREHOUSE.SDE_SBL_Vert_811_Adaptor.SDE_GeographyDimension_Business.log
    DIRECTOR> VAR_27028 Use override value [0] for user-defined workflow/worklet variable:[$$passInStatus].
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [SEBL_VERT_811] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [SEBL_VERT_811.DATAWAREHOUSE.SDE_SBL_Vert_811_Adaptor.SDE_GeographyDimension_Business.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[MPLT_LOAD_W_GEO_DS.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint2].
    DIRECTOR> TM_6014 Initializing session [SDE_GeographyDimension_Business] at [Mon Jul 25 17:29:47 2011].
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Server]
    DIRECTOR> TM_6686 Folder: [SDE_SBL_Vert_811_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_GeographyDimension_Business] Run Instance Name: [] Run Id: [260]
    DIRECTOR> TM_6101 Mapping name: SDE_GeographyDimension_Business [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Storage] will be used as storage directory for session [SDE_GeographyDimension_Business].
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode,Yes]
    DIRECTOR> TM_6708 Using configuration property [ServerPort,6325]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,SIEBEL@ANSDEV dwhadmin@ANBDEV]
    DIRECTOR> TM_6703 Session [SDE_GeographyDimension_Business] is run by 64-bit Integration Service [node01_hkhgc01dvapp01], version [9.0.1 HotFix2], build [1111].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [UNICODE]
    MAPPING> CMN_1570 Server Code page: [UTF-8 encoding of Unicode]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6185 Warning. Code page validation is disabled in this session.
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [ANBDEV], user [dwhadmin]
    MAPPING> CMN_1716 Lookup [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G] uses database connection [Relational:DataWarehouse] in code page [UTF-8 encoding of Unicode]
    MAPPING> CMN_1716 Lookup [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] uses database connection [Relational:DataWarehouse] in code page [UTF-8 encoding of Unicode]
    MAPPING> DBG_21694 AGG_COUNTRY_CITY_ZIPCODE Partition [0]: Index cache size = [1048576], Data cache size = [2097152]
    MAPPING> TE_7212 Increasing [Index Cache] size for transformation [AGG_COUNTRY_CITY_ZIPCODE] from [1048576] to [2402304].
    MAPPING> TE_7212 Increasing [Data Cache] size for transformation [AGG_COUNTRY_CITY_ZIPCODE] from [2097152] to [2097528].
    MAPPING> TE_7029 Aggregate Information: Creating New Index and Data Files
    MAPPING> TE_7034 Aggregate Information: Index file is [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.idx]
    MAPPING> TE_7035 Aggregate Information: Data file is [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.dat]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_GeographyDimension_Business]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Mon Jul 25 17:29:47 2011)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Mon Jul 25 17:29:47 2011)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 36000000 bytes and Block size is 128000 bytes.
    LKPDP_2> DBG_21097 Lookup Transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS]: Default sql to create lookup cache: SELECT CITY,COUNTRY,ZIPCODE,STATE_PROV FROM W_GEO_DS ORDER BY CITY,COUNTRY,ZIPCODE,STATE_PROV
    LKPDP_1> DBG_21312 Lookup Transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G]: Lookup override sql to create cache: SELECT W_LST_OF_VAL_G.VAL AS VAL, W_LST_OF_VAL_G.R_TYPE AS R_TYPE FROM W_LST_OF_VAL_G
    WHERE
         W_LST_OF_VAL_G.R_TYPE LIKE 'ETL%' ORDER BY R_TYPE,VAL
    LKPDP_1> TE_7212 Increasing [Index Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G] from [1048576] to [1050000].
    LKPDP_2> TE_7212 Increasing [Index Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] from [20000000] to [20006400].
    LKPDP_2> TE_7212 Increasing [Data Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] from [20000000] to [20004864].
    READER_1_1_1> DBG_21438 Reader: Source is [ANSDEV], user [SIEBEL]
    READER_1_1_1> BLKR_16051 Source database connection [SEBL_VERT_811] code page: [UTF-8 encoding of Unicode]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [ANBDEV], user [dwhadmin], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8221 Target database connection [DataWarehouse] code page: [UTF-8 encoding of Unicode]
    WRITER_1_*_1> WRT_8124 Target Table W_GEO_DS :SQL INSERT statement:
    INSERT INTO W_GEO_DS(CITY,CONTINENT,COUNTRY,COUNTY,STATE_PROV,ZIPCODE,DATASOURCE_NUM_ID,X_CUSTOM) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8020 No column marked as primary key for table [W_GEO_DS]. UPDATEs Not Supported.
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GEO_DS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Mon Jul 25 17:29:47 2011
    Target tables:
    W_GEO_DS
    READER_1_1_1> RR_4029 SQ Instance [SQ_S_ADDR_ORG] User specified SQL Query [SELECT  DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
          '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    READER_1_1_1> RR_4049 SQL Query issued to database : (Mon Jul 25 17:29:47 2011)
    READER_1_1_1> CMN_1761 Timestamp Event: [Mon Jul 25 17:29:47 2011]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00942: table or view does not exist
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Mon Jul 25 17:29:47 2011]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GEO_DS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Mon Jul 25 17:29:47 2011
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GEO_DS (Instance Name: [W_GEO_DS])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_S_ADDR_ORG] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_S_ADDR_ORG] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_2_1] created for [the transformation stage] of partition point [AGG_COUNTRY_CITY_ZIPCODE] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GEO_DS] has completed. The total run time was insufficient for any meaningful statistics.
    MAPPING> CMN_1791 The index cache size that would hold [0] aggregate groups of input rows for [AGG_COUNTRY_CITY_ZIPCODE], in memory, is [0] bytes
    MAPPING> CMN_1790 The data cache size that would hold [0] aggregate groups of input rows for [AGG_COUNTRY_CITY_ZIPCODE], in memory, is [0] bytes
    MAPPING> CMN_1793 The index cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G], in memory, is [0] bytes
    MAPPING> CMN_1792 The data cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G], in memory, is [0] bytes
    MAPPING> CMN_1793 The index cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS], in memory, is [0] bytes
    MAPPING> CMN_1792 The data cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS], in memory, is [0] bytes
    MANAGER> PETL_24005 Starting post-session tasks. : (Mon Jul 25 17:29:47 2011)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Mon Jul 25 17:29:47 2011)
    MAPPING> TE_7216 Deleting cache files [PMLKUP14527_524289_0_260L64] for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G].
    MAPPING> TE_7216 Deleting cache files [PMLKUP14527_524293_0_260L64] for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS].
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> TE_7216 Deleting cache files [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.idx] for transformation [AGG_COUNTRY_CITY_ZIPCODE].
    MANAGER> TE_7216 Deleting cache files [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.dat] for transformation [AGG_COUNTRY_CITY_ZIPCODE].
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_S_ADDR_ORG] (Instance Name: [SQ_S_ADDR_ORG])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GEO_DS] (Instance Name: [W_GEO_DS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_GeographyDimension_Business] completed at [Mon Jul 25 17:29:48 2011].
    After review the log, I found the select statement was fail, the SQL below was wrong:
    SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    There is no table V_ADDR_ORG but S_ADDR_ORG in the Siebel transaction database. So I don't know why it generate this sql when transfering data to OBAW. This is the out of box bi application.
    Experts! How could I fix this problem?? Please help, thank you very much!!
    Best regards,
    Ryan

    Ryan,
    Yes you missed to create views for source tables. so in dac design>> tables tab, select any table and right clieck on it then click on change capture scripts>> generate view scripts. it will ask you whether it can generate script for all the tables. so it will generate view script for you. now run the entire esript in source database. then your problem will be solved.
    if this answered your question . make my answer correct.
    Thanks
    Jay.

  • Reg: internal table?

    hello friends?
    i have one basic doubts,
    please clear this things,
    what are all the ways we can create the internal table?
    how we create internal table using data statement?
    how we create internal table using types statement?
    what is the difference between these two methods?

    FYI..
    DATA - Defining an Internal Table
    Variants:
    1. DATA itab TYPE itabtype [WITH HEADER LINE].
    2. DATA itab {TYPE tabkind OF linetype|
                  LIKE tabkind OF lineobj}
              WITH [UNIQUE|NON-UNIQUE] keydef
              [INITIAL SIZE n] [WITH HEADER LINE].
    3. DATA itab {TYPE TABLE OF linetype|LIKE TABLE OF lineobj}.
    4. DATA itab TYPE RANGE OF type.
    DATA itab LIKE RANGE OF f.
    5. DATA itab [TYPE linetype|LIKE lineobj] OCCURS n
              [WITH HEADER LINE].
    6. DATA: BEGIN OF itab OCCURS n,
          END   OF itab [VALID BETWEEN f1 AND f2].
    In an ABAP Objects context, a more severe syntax check is performed that in other ABAP areas. See New naming conventions and LIKE references to Dictionary Types not allowed.
    Effect
    Defines an internal table.
    To fill and process internal tables, use the statements INSERT, APPEND, READ TABLE, LOOP, SORT, and so on.
    The OCCURS or INITIAL SIZE parameter (OCCURS value) determines the number of lines that are created when the table itself is created. However, the table is extended dynamically on demand. For details, refer to Performance Notes for Internal Tables. The OCCURS value, has no other semantic meaning (apart from one exception in the APPEND SORTED BY statement). If you do not specify an INIT IAL SIZE , the system uses the default value 0.
    If you specify WITH HEADER LINE, the table is created with a header line, that is, a field with the same name. It has the same type as the line type of the table.
    This addition is not allowed in an ABAP Objects context. See Tables with header line not allowed.
    Variant 1
    DATA itab TYPE itabtype [WITH HEADER LINE].
    Effect
    itabtype must be an internal table type that you have already defined using TYPES. The statement creates an internal table in the program with this type.
    In general, the type specification for the table object must be complete. The exception to this is a standard table, in which the key definition may be missing. In this case, the system automatically uses a default key.
    Example
    Creating a hashed table by referring to an existing table type:
    TYPES: BEGIN OF STRUC, NAME(10), AGE TYPE I, END OF STRUC,
           HTAB TYPE HASHED TABLE OF STRUC WITH UNIQUE KEY NAME.
    DATA : PERSONS TYPE HTAB.
    Variant 2
    DATA itab {TYPE tabkind OF linetype|LIKE tabkind OF lineobj}           WITH [UNIQUE|NON-UNIQUE] keydef
              [INITIAL SIZE n] [WITH HEADER LINE].
    Effect
    Creates an internal table in the program with the type tabkind. Since there are no generic field definitions, you cannot use the table types ANY TABLE or INDEX TABLE.
    The structure of the table lines is defined by the type linetype if you use a TYPE reference) or by the type of the referred object lineobj (when you use a LIKE reference).
    The same rules apply to the UNIQUE and NON-UNIQUE additions in the DATA statement as in a TYPES definition. You may only omit the definition when defining a standard table.
    If you do not specify the INITIAL SIZE the system uses a default initial size of 0.
    Variant 3
    DATA itab {TYPE TABLE OF linetype|LIKE TABLE OF lineobj}.
    Effect
    This is a shortened form of the definition of a standard table. It corresponds to
       DATA itab {TYPE STANDARD TABLE OF linetype|
                  LIKE STANDARD TABLE OF lineobj} WITH DEFAULT KEY.
    or the old definition (compare variant 4)
       DATA itab {TYPE linetype|LIKE lineobj} OCCURS 0.
    Variant 4
    DATA itab TYPE RANGE OF type. DATA itab LIKE RANGE OF f.
    Additions:
    1. ... INITIAL SIZE n
    2. ... WITH HEADER LINE
    Effect
    Creates an internal table itab with table type STANDARD. The line type is a structure with the following components:
      SIGN(1)   TYPE C
      OPTION(2) TYPE C
      LOW       TYPE type bzw. LIKE f
      HIGH      TYPE type bzw. LIKE f
    Addition 1
    ...INITIAL SIZE n
    Effect
    The INITIAL SIZE specification determines how many table lines are created when the table itself is created. The table is also dynamically expanded as required. For further information, refer to Performance Notes for Internal Tables. The INITIAL SIZE value has no semantic meaning (apart from one exception in the ei APPEND SORTED BY statement). If you do not specify the INITIAL SIZE, the system uses the default value 0.
    Addition 2
    ... WITH HEADER LINE
    This addition is not allowed in an ABAP Objects context. See Tables with Header Lines Not Allowed.
    Effect
    Creates an internal table and a header line for it, that is, a field with the same name as the internal table and the same type as the line type of the internal table.
    Variant 5
    DATA itab [TYPE linetype|LIKE lineobj] OCCURS n                                        [WITH HEADER LINE].
    This variant is not allowed in an ABAP Objects context. See Declaration with OCCURS not allowed.
    Effect
    This variant exists to ensure compatibility with Release 3.x. If you do not specify a line type, the system uses type C with length 1. Otherwise, the variant is the same as
       DATA itab {TYPE STANDARD TABLE OF linetype|
                  LIKE STANDARD TABLE OF lineobj}
                 INITIAL SIZE n [WITH HEADER LINE].
    Example
    TYPES: BEGIN OF LINE_TYPE,
             NAME(20) TYPE C,
             AGE      TYPE I,
           END   OF LINE_TYPE.
    DATA:  PERSONS    TYPE LINE_TYPE OCCURS 20,
           PERSONS_WA TYPE LINE_TYPE.
    PERSONS_WA-NAME = 'Michael'.  PERSONS_WA-AGE  = 25.
    APPEND PERSONS_WA TO PERSONS.
    PERSONS_WA-NAME = 'Gabriela'. PERSONS_WA-AGE  = 22.
    APPEND PERSONS_WA TO PERSONS.
    The internal table PERSONS now contains two entries.
    Variant 6
    DATA: BEGIN OF itab OCCURS n,         ...
          END   OF itab [VALID BETWEEN f1 AND f2].
    This variant is not allowed in an ABAP Objects context. See Declaration with OCCURS not allowed.
    Effect
    Creates an internal table itab with type STANDARD and a header line. The line type consists of the fields between "BEGIN OF itab OCCURS n" and " END OF itab".
    Use the VALID BETWEEN f1 AND f2 addition to specify that the components f1 and f2 of the internal table itab contain a line-based validity interval. You can only use this addition in conjunction with the PROVIDE statement.
    Example
    DATA: BEGIN OF PERSONS OCCURS 20,
            NAME(20),
            AGE TYPE I,
          END   OF PERSONS.
    PERSONS-NAME = 'Michael'.
    PERSONS-AGE  = 25.
    APPEND PERSONS.
    PERSONS-NAME = 'Gabriela'.
    PERSONS-AGE  = 22.
    APPEND PERSONS.
    The internal table consists of two entries. PERSONS also has a header line (work area), which is an interface between the program and the actual table contents.
    Additional help
    Internal Table Objects
    Ramesh

  • Informatica Workflow Manager ODBC Relational Connection for ETL in DAC

    In Informatica Workflow Manager, I have created a Relational Connection of type ODBC and specified Connect String as "DSN=BIEEDW" where "BIEEDW" is the System ODBC DSN already set pointing to a SQL Server 2008 database.
    However, when the ETL run in DAC, the following error occurs in Session log files showing that the database and driver cannot be located:
    MAPPING> CMN_1569 Server Mode: [UNICODE]
    MAPPING> CMN_1570 Server Code page: [MS Windows Traditional Chinese, superset of Big 5]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [DSN=BIEEDW], user [bieedw02]
    MAPPING> CMN_1761 Timestamp Event: [Wed May 22 01:29:17 2013]
    MAPPING> CMN_1022 Database driver error...
    CMN_1022 [
    [Microsoft][ODBC Driver Manager] Data source name not found and no default driver specified
    Database driver error...
    Function Name : Connect
    Database driver error...
    Function Name : Connect
    Database Error: Failed to connect to database using user [bieedw02] and connection string [DSN=BIEEDW].]
    Any hint in setting the Connect String for ODBC Relational Connection?

    Hi,
    Let me tell you the real story:
    Our server architecture consists of two servers:
    Windows Server 2008 R2 (64-bit) platform with the following installed:
    - SQL Server 2008
    - DAC 10.1.3.4.1
    - OBIEE 11g
    - BI Apps (Financial Analytics) 7.9.6.3
    - Informatica Server 9.1.0 HotFix 2
    Windows Server 2003 Enterprise Edition SP2 (32-bit) platform with the following installed:
    - Informatica Clients (i.e. Workflow Manager, Repository Manager, Designer and Workflow Monitor)
    And thus the ODBC Relational Connection is configured in Informatica Client machine (Workflow Manager) which is a 32-bit platform.
    Any idea?

  • Distributing a .MUP to MSA clients using Upgrade Console?

    Hello,
    We need to distribute a sap note needed for MSA containing a .MUP file. I have created a new upgrade package in the Upgrade Console. In the package I have added the .Mup file under the tab 'Files'.
    I have tried to send out the package only containing tile file, and it gets distributed to the client.
    The next thing is to start the installation of the .Mup file. How is this normally done? I tried to create a bat file with a reference to the mup file, but that didnu2019t work. Is it really necessary to create a bat file in order to be able to launch the MUP file on the client u2013 or can this be done in another way?
    BR AC
    Edited by: Andy Cooper on Jan 31, 2009 2:07 PM

    Hello
    You do not have to execute an MUP from clcient machine manually.You just hvae to run the contrans which will get the MUP in the form of BDOC, and client upgrade target will be invoked to execute the MUP implicitly.
    I would suggest you to check the following step by step.
    1. If you have SAPGUI 7.1 is installed in the machine. Kindly implement the note 1287556.
    2. Open TransportInstaller.exe from Mobile\Bin.Net folder, Login to ARS, go to 'Service' tab page and check that transport service is installed, go to 'CRm Server Connection page' and provide the CRM connection details, test the conenction and save the data.
    NOTE: For the step 2, you can get more info in theinstallation guide ( Appendix 17). It has detailed explanation for installing the transport service.
    3. Go to service Control manager ( Run command --> Services.msc), Start the transport service. when you start the service, you will be able to see user inyerface window where in you can check the status of the changelist for your upgrade MUP. Please note that change list will be created implicitly.
    The following options need to be changed from Transport service user interface window.
    1. Go to Tools -> Options in the Transport service UI window. It will open ' generals ettings' window.
    2. Go to Options tab page, Change the polling interval according to you.
    3. Go to Services tab page, select the following options.
    - Run time object change lists
    - Pack the released changelists
    - Send the packed changelists to the server
    - Push run time objects from server to mobile clients
    4. Open Upgrade Console from Mobile\Bin.Net folder, Do the send upgrade of a particular u[pgrade.
    Now Come to Transport service UI window. It will start polling based polling interval which was set.
    so you will be able to see the changelist ( for the MUP upgrade) in the user interface window with in the polling interval time and will be automatically sent to the CRM server.
    5. Now from the client machine, Ensure that site is subscribed for the following subscription
            - WORK BENCH LANGUAGE INDEPENDENT
            - WORKBENCH LANGUAGE DEPENDENT EN
    Run the contrans from client and and verify that MUP is executed implicitly.
    Hope I have explained you clearly.
    Best Regards
    Shankar

  • Execution Plan Failed for Full Load

    Hi Team,
    When i run the full Load the execution plan failed and verified the log and found below information from SEBL_VERT_8_1_1_FLATFILE.DATAWAREHOUSE.SIL_Vert.SIL_InsertRowInRunTable
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [SEBL_VERT_8_1_1_FLATFILE.DATAWAREHOUSE.SIL_Vert.SIL_InsertRowInRunTable.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[mplt_SIL_InsertRowInRunTable.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [21950495] for mapping parameter:[MPLT_GET_ETL_PROC_WID.$$ETL_PROC_WID].
    DIRECTOR> TM_6014 Initializing session [SIL_InsertRowInRunTable] at [Mon Sep 26 15:53:45 2011].
    DIRECTOR> TM_6683 Repository Name: [infa_rep]
    DIRECTOR> TM_6684 Server Name: [infa_service]
    DIRECTOR> TM_6686 Folder: [SIL_Vert]
    DIRECTOR> TM_6685 Workflow: [SIL_InsertRowInRunTable] Run Instance Name: [] Run Id: [8]
    DIRECTOR> TM_6101 Mapping name: SIL_InsertRowInRunTable [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [H:\Informatica901\server\infa_shared\Storage] will be used as storage directory for session [SIL_InsertRowInRunTable].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6703 Session [SIL_InsertRowInRunTable] is run by 32-bit Integration Service [node01_eblnhif-czc80685], version [9.0.1 HotFix2], build [1111].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [Connect_to_OLAP], user [OBAW]
    MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
    MAPPING> CMN_1022 Database driver error...
    CMN_1022 [
    Database driver error...
    Function Name : Logon
    ORA-12154: TNS:could not resolve service name
    Database driver error...
    Function Name : Connect
    Database Error: Failed to connect to database using user [OBAW] and connection string [Connect_to_OLAP].]
    MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
    MAPPING> CMN_1076 ERROR creating database connection.
    MAPPING> DBG_21520 Transform : LKP_W_PARAM_G_Get_ETL_PROC_WID, connect string : Relational:DataWarehouse
    MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
    MAPPING> TE_7017 Internal error. Failed to initialize transformation [MPLT_GET_ETL_PROC_WID.LKP_ETL_PROC_WID]. Contact Informatica Global Customer Support.
    MAPPING> CMN_1761 Timestamp Event: [Mon Sep 26 15:53:45 2011]
    MAPPING> TM_6006 Error initializing DTM for session [SIL_InsertRowInRunTable].
    MANAGER> PETL_24005 Starting post-session tasks. : (Mon Sep 26 15:53:45 2011)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Mon Sep 26 15:53:45 2011)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_FILE_DUAL] (Instance Name: [SQ_FILE_DUAL])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SIL_InsertRowInRunTable] completed at [Mon Sep 26 15:53:46 2011].
    I checked physical datasource connection in DAC and workflow manager and tested and verified but still facing same issue. I connected through oracle merant ODBC driver.
    Pls. let me know the solution for this error
    Regards,
    VSR

    Hi,
    Did you try using Oracle 10g/11g drivers as datasource? If not I think you should try that but before that ensure that from your DAC server box you are able to tnsping the OLAP database. Hope this helps
    If this is helpful mark is helpful
    Regards,
    BI Learner

  • Function module to give a date based on a particular date and interval.

    Hi there,
    I am writing a code where I need to to find another date
    based on a key date and and an interval.
    Is there any function module where I pass the date and the interval
    (say 30) and it gives me a new date after subtracting 30 from it.
    Or is there any piece of code which does it?
    Thanks in advance.
    Regards,
    Kate

    hi Kate,
    you can try ?
    data : date2 like sy-datum.
    date2 = yourdate - 30.
    hope this helps.

  • Interval paritioning based on Integer

    Oracle 11.0.1.7:
    Can I create interval based partitioning on integer column. We have a integer column that basically is a sequence. I need to create partition dynamically every 10Million rows, how can I do that in interval based partitioning scheme?

    Hi,
    Example
    create table test
               (sno number(6),
               last_name varchar2(30),
               salary number(6))
               partition by range(salary)
               Interval  (5000)
           partition p1 values less than (5000),
           partition p2 values less than (10000),
         partition p3 values less than (15000),
         partition p4 values less than (20000));SS

  • Charging Porposnate dep to the cost centes based on the Interval Assigned

    Hi,
    I have scenario were asste is being transferred from one cost center to another and depreciation needs to be charged on proposnate basis between the old and new cost center.
    I have changed the interval in the aaset master (T-Code AS02) for transfer, but when I am posting depreciation system is chaging 100% dep. to new cost center irrespective of the date of changes.
    Regads,
    AB

    Hi,
    I tested it, made some research and found note 29947:
    When posting depreciation with RABUCH00, with account assignment to cost
    center, depreciation is always posted to the cost center assigned at the
    end of the posting period in which depreciation is being posted.        
    Depreciation is not distributed for a proportionate period of time (by  
    date) to the time-dependent cost centers assigned in table ANLZ.        
    This note speaks about RABUCH00 but is is release independend and still valid.
    Have you made another experience?
    ( Time independent organization unit is not relevant in that case)
    regards
    Bernhard
    Edited by: Bernhard Kirchner on Nov 16, 2010 12:31 PM

Maybe you are looking for

  • The font on my internet browser and mail has changed.  How do I change it back to a standard font?

    The font on my internet browser and mail has changed.  How do I change it back to a standard font? 

  • Flash animation to html conversion hairline problem

    Hi. I am Suzi. I am currently working in Flash CC and working with some animations. I am needing to turn the Flash animations in to HTML and Java Script which are being viewed on iPads, iPhone and on the net.  My issue is that when they are eventuall

  • Menu keeps appearing in the swf file

    I made a flash for my website and used the fscommand to hide the menu, and it works well when i open the swf file. But when i insert it inside a web page the menu still there when any one click the mouse right button :S I think i did this before and

  • HT1338 could not find intallation information for this machine

    i erased my hard drive and was trying to install the new mac ox lion version, but i have this error message:" could find installation information for this machine, contact applecare" i am connected to the internet, i could not find any answer via app

  • BW 3.5 courses

    I am thinking about doing the delta courses and possibly obtain the 3.5 delta certification. However, a good chunk of the course is on BPS, a new component of BW. I want to get some feedback from those whom have done the courses. Thanks in advance.