HR Configuration for BI Apps

We are lookng for feedback on how other people have dealt with configuration changes to HR spreadsheets. For example, we had to make configuration changes to event types and it appears we have to reprocess that table. We make a special subject area and execution plan in DAC and updated the refresh date for the D and DS table to be way in the past. Is this the correct approach and will this have to occur every time a configuration change is made?

Can anyone who has done a vanilla implementation of BI Apps (7.9.4) check the files in /opt/oracle/informatica/server/SrcFiles and see +how many files/sessions have the TYPE2_FLG=Y vs. TYPE2_FLG=N.                                                                                                                                                                                                                                                                                                                                                                                                           

Similar Messages

  • Profile configuration for my app

    Is it possible to set preference values for my app in the profile configuration file? If so, an example would be greatly appreciated. Thanks.

    Hi,
    Have you ever figured out a way to use iPhone configuration utility for your own app? I'm looking into mechanisms for this purpose as well.

  • Project Analytics - PA - Configurations for Oracle Apps 7.9.6

    Guys - I have gone through the docs and cannot find what configuration steps are necessary for us to run OBAW / DAC loads for PA (Project Accounting / Project Analytics).
    Any pointers are greatly appreciated.
    Regards,

    Here are the main steps:
    Install OBIEE
    Install OBIA (e.g. 7.9.6) use Project analytics option while install
    Create Target Database ( e.g. 10.2.x or 11g)
    Install Informatica (e.g. 8.6, SP4)
    Client
    Server (Server / Repository)
    Integration Service and Repository Service
    Install DAC (e.g. 10g)
    Install DAC client
    Install DAC Server
    Restore Informatica Repository
    Restore DAC Repository
    Register Informatica Repository services with DAC
    Connect DAC to DW
    Create Custom Container in Informatica for ETL loads

  • Php update/configure for Server 4.0

    I had working Apache2 + PHP 5.6 running perfectly locally for development.
    Web Docs were located in /Sites folder in user directory.
    After Installing Server 4.0 all my setup is messed up. I can configure website again but it not uses PHP I have upgraded to.
    I feel lack of control/configuration for web apps with Server 4.0. How I can get it working back as perviously or uninstall Server 4.0?
    Filip

    I have found solution to fix Apache after uninstallation of Server 4.0
    http://remonpel.nl/2014/08/restore-apache2-service-after-removal-of-server-app/# more-554
    I found also more articles that maybe would help me running my websites with OS X Server.... but for now I have decided to not come back to try this.
    http://remonpel.nl/

  • Configuring socket policy for flex apps(with blocked port 843)?

    We have built several flex-based ecommerce apps for a fortune 500 customer of ours, that for various reasons, we need to use sockets to a different domain and requires a socket policy file, but were having trouble configuring our flex apps for deployment in thier enviornment where they are blocking virtually everything except port 80 . The current documentation in in regards to socket policy files and crossdomain files in a non-standard configuration not using port 843 is not providing any useful help to us.
    Here is the scenario:
    Flex apps are served from domain www.a.com in  to users browsers via http. The apps then make socket connections to domain www.b.com:80 where there are php scripts serving json data to the flex apps via port 80 using http(we use sockets because we need to set and read back http headers). The problem is the flex apps cannot make socket connections to the www.b.com domain without errors like below(unless we setup a socket policy server on port 843 of www.b.com, in which case everything works):
    Warning: Timeout on xmlsocket://www.b.com:80 (at 3 seconds) while waiting for socket policy file.  This should not cause any problems, but see http://www.adobe.com/go/strict_policy_files for an explanation.
    Error: Request for resource at xmlsocket://www.b.com:80 by requestor from http://www.a.com/bin-debug/DownloadManagerFlex.swf is denied due to lack of policy file permissions.
    Error: Request for resource at xmlsocket://www.b.com:80 by requestor from http://www.a.com.us/bin-debug/DownloadManagerFlex.swf is denied due to lack of policy file permissions.
    Since we cannot use port  843 for the socket policy file server, we setup the socket policy server on a different ip in the same domain: spf.b.com:80 (using the sample perl code Adobe provides), and per the docs(cited below), use Security.loadPolicyFile("xmlsocket://spf.b.com:80") before we invoke "socket.connect", to supposedly tell the flash player to check there for the socket policy file. The problem, as you can see from the error log, is that the  loadPolicyFile("xmlsocket://spf.b.com:80") is ignored.
    No matter what we do or how we set things up, we cannot get the flash player to recognize the loadPolicyFile(), it always wants to go to the port were making the socket connection on. It is unclear how to properly configure the flex app, socket policy file and crossdomain file for the above scenario. The docs allude to being able to serve  the socket policy file from a different port 80 in the same domain as the socket connection were trying to make, but were having no luck with that.
    ->Can anyone shed some light on how to make this work or what are we  missing/doing wrong? Also, if we can get this to work, are we  stuck with a 3 second delay because this(very large) customer is blocking port 843?
    As an aside,  the documentation for all this is a bit scattered, unclear and contrdictory:
    One document says:(http://www.adobe.com/devnet/flashplayer/articles/fplayer9_security_07.html)
    "This warning usually means one of two things: first, that you need to set up a
                socket policy file server on port 843, which is the first location that Flash
                Player checks by default; or second, that you need to provide more explicit
                guidance to Flash Player from ActionScript by calling loadPolicyFile to indicate the location
                of a socket policy file. When you call loadPolicyFile rather than allowing Flash Player to check
                locations by default, Flash Player will wait as long as necessary for a
                response from a socket policy file server, rather than timing out after 3
                seconds."
    Another document says(http://www.adobe.com/devnet/flashplayer/articles/socket_policy_files.html):
    "If an ActionScript Security.loadPolicyFile() command exists within               the SWF file, then the Flash Player runtime checks that location. Flash Player checks               the destination of the loadPolicyFile() only after it has checked the               master policy file on port 843 for permission to acknowledge other policy               files. If the developer has not specified a loadPolicyFile() command,               then Flash Player checks the destination port of the connection."

    I found the reason why the Flex application was ignoring the socket policy (crossdomain.XML). I have a policy server that listens to port 843 and submits the policy to the Flex client. My policy was getting ignored by the Flex application and I was getting the sandbox security error you were getting. The solution to this problem isto write a null byte right after the policy server sends the policy. I'm using Apache Mina that is wrtten is Java and the null byte is written as follows:
    public void sessionCreated (IoSession session)
            throws Exception
            session.write(_policy);  -- > policy string
            session.write("\u0000"); --> null byte
             //session.close(true); ---> No need to close the session because it is closed by the Flex client after it receives the null byte.
    Now my Flex application can read and accept the policy from port 843 and I'm not getting more security violations.
    Thanks for your reply,
    Alberto

  • HT2404 my hard drive crashed and nothing recoverable.  Apple installed a new drive.  How can I go back and download the apps I had purchased for my previous configuration, for instance Lion 10.7, and Photoshop Elements 10?

    My hard drive crashed and nothing recoverable, local Apple Store installed a new drive.  How can I go back and download the apps I had purchased for my previous configuration, for instance Lion 10.7, and Photoshop Elements 10?

    A blank drive needs to be formatted first, you do that hold c boot (or hold option key boot) from the 10.6 install disk (or Lion USB)
    Read how to format the drive in Disk Utility here
    https://discussions.apple.com/docs/DOC-3044
    How to install OS X
    https://discussions.apple.com/docs/DOC-3251
    Software Update fully,
    Hold the option key and click on Purchases when in MAS to download your 10.7 etc.

  • When importing a spreadsheet with redeem codes into apple configurator, it says it's not the right codes for the app.  But it is the right codes. I used one of the codes to purchase the app in itunes.

    I have a cart of 25 ipads with a macbook air.  I have logged into my VPP account and purchased the needed number of copies.  I downloaded the spreadsheet.  I took the first code from the spreadsheet, and logged into itunes using a regular account (not my VPP account), and used the code to download the app.  I put the app into configurator and then tried to add the spreadsheet with the codes in it.  It comes up and tells me that it is not the correct spreadsheet for that app. But it is the correct spreadsheet!!  I tried re-downloading the spreadsheet again and even tried re-downloading the app.  So I've now used two of my redeem codes and still can't get it to accept it into apple configurator.  I removed the app from apple configurator and re-added it.  The error message says it's not the right codes for the app "Grammar Games by Tap To Learn" and to please get the codes for "Grammar Games by Tap To Learn."  It's the same one!

    It won't let me edit my post, so I'm typing this reply to add...
    I tried taking the version number out of the title. It looked like there were two spaces in the title "Grammar Games  by Tap to Learn". So I even made sure that matched. I still can't import my codes. Any other things to try would be really appreciated.

  • I have one apple ID for multiple devices in my family.  I'd like to keep it that way for itunes/app purchases.  I would like a simple step 1, step 2, step 3 response on what I need to do to separate all other features like imessage, contacts, emails, etc.

    I have one apple ID for multiple devices in my family.  I'd like to keep it that way for itunes/app purchases.  I would like a simple step 1, step 2, step 3 response on what I need to do to separate all other features like imessage, contacts, emails, etc.
    I have been reasearching how to do this on the internet, but I haven't found an easy explanation yet.  My family is going crazy over each others imessages being sent to others in the family and not being able to use FaceTime because of conflicting email addresses.  I have read that if each person gets their own iCloud account, this would work.  However, I need to know what to do after I set everyone up with their own iCloud account.  Do I make that the default email address to be contacted or can they still use their hotmail email addresses.  Any help- with easy explanation- would be much appreciated!!

    We do this in my family now.  We have one account for purchases, so it is used to share music and apps (I think that is in Settings/iTunes & App Stores).  Each iDevice has this configured.
    Then, each of us has our own iCloud account that is configured under Settings/iCloud.  That then allows us to have our own Mail/Contacts/Calendars/Reminders/Safari Bookmarks/Notes/Passbook/Photo Stream/Documents & Data/Find My iPhone/and Backup.  That Backup piece is pretty sweet and comes in handly if you replace your iDevice.  You can just restore from it.
    So we all share the Apple Store account but we all have our own iCloud accounts to keep the rest seperate or things like you mentioned are a nightmare.
    In answer to what iCloud does for you: http://www.apple.com/icloud/features/
    Think of it as an internet based ("cloud") area for all of those items listed in my response.  What you need to remember is photo stream only maintans the last 1000 pictures so don't count it as a complete backup solution for your pictures.  Even though I rarely sync with a computer these days, I do still try to sync my phone with iPhoto (I have an iMac) so that I have copies of all of my pictures.  1000 may not stretch as far as it sounds.
    Message was edited by: Michael Pardee

  • How to have Multiple Configurations for a Flex Application

    Hi All,
    I've seen this question asked in various forms multiple times, but never answered: How do you set up a FLEX configuration to have different configurations for development (local), test server, staging server, beta server, and production server? I cannot imagine that Adobe has not accounted for this, something that is very standard in most development teams. Zend Framework has a very elegant solution for this: in the configuration file you define all aspects that can be configured, as well as the various environments (development, staging, beta, live, etc.). Then, in the bootstrap file you specify which environment you want to work in, and you're all set.
    How can something like this be achieved in FLEX?
    Many thanks!
    -Mike

    Maybe my setup is unconventional, I doubt it, though:
    1) SVN repo stores FLEX and PHP codebases in separate projects.
    2) Developer gets code from repo, and runs on local machine. -> need to configure FLEX to point to local PHP code here for Remote Objects.
    3) Developer then publishes Flex app to staging server. -> need to configure FLEX to point to staging PHP code here.
    4) Flex app then gets published to live server. -> need to configure FLEX to point to live PHP code here.
    Now, using ANT, the scenario would look like this:
    1) SVN repo stores FLEX and PHP codebases in separate projects.
    2) Developer gets code from repo, and runs on local machine. -> need to configure FLEX to point to local PHP code here for Remote Objects.
    3) Code is good and developer checks it back into SVN.
    4) ANT extracts latest build from SVN and builds it, then publishes to staging server. -> need to configure FLEX to point to staging PHP code.
    I am using PHPUnderControl as my build server for PHP.
    I am also unclear as to how my goal could be achieved using ANT?
    Thanks!
    Mike

  • Problem with roaming for specific apps

    Hi :)
    I have a problem with my iPhone 4s:
    When i use roaming on my iPhone 4s i try to save some money by turning off roaming for specific apps, but one day for ca. 3 weeks ago it stopped working!
    Some apps can use roaming but some can't ;(
    There are like 10 apps always turned on, but if i try turning roaming off for an app that is on already it wont work!
    If i close settings and reopen it, the app is now denied using roaming!
    If i try allowing an app it changes but still wont work and after some time it gets in the same possission i cant change!
    I have never experienced this before and i could't find an answer on google ;(
    If you know how to fix it please respond to me here :)
    ~somonekilledme
    P.S: Im really sorry if there are bad english here, but im mot english sooooo..... yeah

    You are on the right track suspecting DHCP conflicts.
    To configure your "roaming network" connect your "slave" Extreme to your main Internet router - either directly or through a switch - and place it in "bridge mode". That will work.
    A router will be distributing IP addresses using DHCP. Anything connected to it should be configured as a bridge, the essential difference being that a bridge does not distribute IP addresses.
    I do not believe your "slave" Extreme will function connected as it is to the "main" Extreme, but may work if you assign it a static IP address. It is not unusual for it to retain this static IP after having been changed to DHCP if its lease is still valid.
    The OK article you referenced must be terribly outdated if it is referring to AirPort Admin Utility. That has not been used in about a decade. Read this one:
    Wi-Fi base stations: Extending the range of your wireless network by adding additional Wi-Fi base stations
    Unfortunately its screenshots refer to AirPort Utility 5.6, which has been replaced with the prettier, though less useful 6.2.

  • The wifi at my school isn't very good. And I can't send iMessages through it. It works for some apps like score center and my radar app, but doesn't work for other apps like clash of clans and iMessage. How can I get it to work on that  wifi network

    The wifi at my school isn't very good. And I can't send iMessages through it. It works for some apps like score center and my radar app, but doesn't work for other apps like clash of clans and iMessage. How can I get it to work on my school wifi. I have an iPhone 4S on AT&T and iOS 7.04

    In a roaming network, your "main" router is the device that would require port mapping/forwarding to be configured in order to access the IP camera from the Internet. This router is also the one that would be provide the private IP address for the camera which you will want to be a static one.
    So as you described your network, the IP cameras should be getting an IP address or you assigned it a static one and this is the address that you would enter in the Private IP address (or equivalent depending on the router used) field when setting up port mapping.
    If you are not able to access this camera from the local network, then this should be troubleshot first.

  • JBO-33001: Configuration file /oracle/apps/crmCommon/activities/activityManagementService/applicationModule/common/bc4j.xcfg is not found in the classpath.

    I am trying to consume a base model which is available in the Customer.jws (AppCmmnCompActivitiesPublicModelActivityManagement) to my Mobile UI project available in Sales.jws (SalesMobilePublicUi). We have added proper Databindings.cpx entry for accessing the applicationModule present in the model.
    On deploying the sales mobile app, I get the following ConfigException:
    JBO-33001: Configuration file /oracle/apps/crmCommon/activities/activityManagementService/applicationModule/common/bc4j.xcfg is not found in the classpath.
    The bc4j.xcfg mentioned in the above path actually contains the entry. On checking about this issue, I found out that the prominent solution would be model project should be added as Dependencies to the UI project. But the UI is in different workspace (Sales.jws) and so the list of dependencies shown are those of this workspace only, whereas the model is present in Customer.jws and it does not come under the dependencies list.
    How can this issue be resolved? Please let me know any pointers on solving this issue.

    Build an adfLibrary from the model in the Customers.jws and add this adfLibrary to the other ui project. This will get you the dependency.
    Timo

  • Is it possible to run TWO parellel configurations for one Object??

    SAP Gurus,
    We have a strange requirement from the client to have to configurations established in the system for one configuration object.
    Without taking much of your time let me paraphrase the requirements.
    The client is having two different types of putaway strategies in their current business process. And they toggle between these putaway strategies by reaching out to their production support group and they make configuration changes to have the other put away strategy
    One Step Putaway: GR area --> High Rack Storage
    Two Step Putaway: GR area --> Interim storage (area at the end of each aisle of high rack storage) --> High Rack
    Client wants to have two parallel configurations for both of these put away strategies and needs an ability to switch from one to another. This switching ability should be accessed from SAP screen without any IT intervention (no configuration)
    -  Switching ability should allow the warehouse manager to change from one step to two step put away strategies without any IT intervention.
    -  Switching ability should be accessed my SAP Screen (Z-transaction or something with a radio button to toggle around)
    Is it possible to have two configuration made for Putaway Strategies and switch them on/off depending on the warehouse needs??
    Your invaluable feedback on this will be highly appreciated.
    Thanks again folks,
    Dhaval
    Edited by: Dhaval Joshi on Mar 4, 2008 12:09 PM

    Q1--I don't believe it is possible. An organization can have multiple licenses tied to its main email address, for example, but then they are assigned to the individual users.
    Q2--Yes, you would just need to log out of your account within the CC desktop app prior to logging in on the other account.
    If you have any questions, feel free to reach out.

  • DAC: failed task during ETL for financial apps

    I am trying  my first ETL on OBIA 7.9.6.4
    i'm using  Oracle EBS 12.1.1 as source system.
    the ETL completes 314 tasks successfully ,but it fails the task named:
    "SDE_ORA_GL_AR_REV_LinkageInformation_Extract"
    DAC Error log:
    =====================================
    STD OUTPUT
    =====================================
    Informatica(r) PMCMD, version [9.1.0 HotFix2], build [357.0903], Windows 32-bit
    Copyright (c) Informatica Corporation 1994 - 2011
    All Rights Reserved.
    Invoked at Wed Sep 18 09:46:41 2013
    Connected to Integration Service: [infor_int].
    Folder: [SDE_ORAR1211_Adaptor]
    Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Instance: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    Mapping: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract]
    Session log file: [C:\Informatica\server\infa_shared\SessLogs\.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log]
    Source success rows: [0]
    Source failed rows: [0]
    Target success rows: [0]
    Target failed rows: [0]
    Number of transformation errors: [0]
    First error code [4035]
    First error message: [RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATI]
    Task run status: [Failed]
    Integration Service: [infor_int]
    Integration Service Process: [infor_int]
    Integration Service Grid: [infor_int]
    Node Name(s) [node01_AMAZON-9C628AAE]
    Preparation fragment
    Partition: [Partition #1]
    Transformation instance: [SQ_XLA_AE_LINES]
    Transformation: [SQ_XLA_AE_LINES]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [16004], message [ERROR: Prepare failed. : [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_CO]
    Start time: [Wed Sep 18 09:46:13 2013]
    End time: [Wed Sep 18 09:46:13 2013]
    Partition: [Partition #1]
    Transformation instance: [W_GL_LINKAGE_INFORMATION_GS]
    Transformation: [W_GL_LINKAGE_INFORMATION_GS]
    Applied rows: [0]
    Affected rows: [0]
    Rejected rows: [0]
    Throughput(Rows/Sec): [0]
    Throughput(Bytes/Sec): [0]
    Last error code [0], message [No errors encountered.]
    Start time: [Wed Sep 18 09:46:14 2013]
    End time: [Wed Sep 18 09:46:14 2013]
    Disconnecting from Integration Service
    Completed at Wed Sep 18 09:46:41 2013
    Informatica session logs:
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R1211] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [.SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full.ORA_R1211.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [26] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [01/01/1970] for mapping parameter:[$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [01/01/1990] for mapping parameter:[$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27028 Use override value ['NONE'] for mapping parameter:[$$LEDGER_TYPE_LIST].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] at [Wed Sep 18 09:46:13 2013].
    DIRECTOR> TM_6683 Repository Name: [infor_rep]
    DIRECTOR> TM_6684 Server Name: [infor_int]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR1211_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] Run Instance Name: [] Run Id: [2130]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GL_AR_REV_LinkageInformation_Extract [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode ,Yes]
    DIRECTOR> TM_6708 Using configuration property [OraDateToTimestamp ,Yes]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,[APPS]@[ 54.225.65.108:1521:VIS] [DWH_REP2]@[AMAZON-9C628AAE:1521:obiaDW1]]
    DIRECTOR> TM_6703 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] is run by 32-bit Integration Service  [node01_AMAZON-9C628AAE], version [9.1.0 HotFix2], build [0903].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TE_7004 Transformation Parse Warning [IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || >>>>DISTRIBUTION_ID<<<<),
    SOURCE_TABLE || '~' || DISTRIBUTION_ID)
    <<PM Parse Warning>> [||]: operand converted to a string
    ... IIF(EVENT_TYPE_CODE='RECP_REVERSE',
    IIF(UPG_BATCH_ID>0,
    SOURCE_TABLE || '~' || DISTRIBUTION_ID,
    SOURCE_TABLE || '~RECEIPTREVERSE~' || DISTRIBUTION_ID),
    SOURCE_TABLE || '~' || >>>>DISTRIBUTION_ID<<<<)
    ]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [JE_HEADER_ID || '~' || JE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>JE_HEADER_ID<<<< || '~' || JE_LINE_NUM<<PM Parse Warning>> [JE_LINE_NUM]: operand converted to a string
    ... JE_HEADER_ID || '~' || >>>>JE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [AE_HEADER_ID || '~' || AE_LINE_NUM]; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning [<<PM Parse Warning>> [||]: operand converted to a string
    ... >>>>AE_HEADER_ID<<<< || '~' || AE_LINE_NUM<<PM Parse Warning>> [AE_LINE_NUM]: operand converted to a string
    ... AE_HEADER_ID || '~' || >>>>AE_LINE_NUM<<<<]; transformation continues...
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Sep 18 09:46:13 2013)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Sep 18 09:46:13 2013)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [54.225.65.108:1521/VIS], user [APPS]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [AMAZON-9C628AAE:1521/obiaDW1], user [DWH_REP2], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_LINKAGE_INFORMATION_GS :SQL INSERT statement:
    INSERT INTO W_GL_LINKAGE_INFORMATION_GS(SOURCE_DISTRIBUTION_ID,JOURNAL_LINE_INTEGRATION_ID,LEDGER_ID,LEDGER_TYPE,DISTRIBUTION_SOURCE,JE_BATCH_NAME,JE_HEADER_NAME,JE_LINE_NUM,POSTED_ON_DT,GL_ACCOUNT_ID,SLA_TRX_INTEGRATION_ID,DATASOURCE_NUM_ID)  VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GL_LINKAGE_INFORMATION_GS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    READER_1_1_1> RR_4029 SQ Instance [SQ_XLA_AE_LINES] User specified SQL Query [SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Sep 18 09:46:13 2013)
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Wed Sep 18 09:46:13 2013
    Target tables:
         W_GL_LINKAGE_INFORMATION_GS
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00904: "XLA_EVENTS"."UPG_BATCH_ID": invalid identifier
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Wed Sep 18 09:46:13 2013]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GL_LINKAGE_INFORMATION_GS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Wed Sep 18 09:46:13 2013
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GL_LINKAGE_INFORMATION_GS (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
    WRT_8044 No data loaded for this target
    WRITER_1_*_1> WRT_8043 *****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GL_LINKAGE_INFORMATION_GS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Wed Sep 18 09:46:14 2013)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Wed Sep 18 09:46:14 2013)
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_XLA_AE_LINES] (Instance Name: [SQ_XLA_AE_LINES])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GL_LINKAGE_INFORMATION_GS] (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
      Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GL_AR_REV_LinkageInformation_Extract_Full] completed at [Wed Sep 18 09:46:14 2013].
    *I did some queries in my source database (Vision) , table "XLA_EVENTS" exists , column "UPG_BATCH_ID" also exists
    *I added "XLA_EVENTS" to the FROM clause and ran it in SQL Developer
    *in the SELECT clause ,i see a column named "AEHEADER.EVENT_TYPE_CODE"
    but there is no table named "AEHEADER" in the FROM clause
    so i added it manually , it's probably refers to "XLA_AE_HEADERS"
    Final query looks like this:
    SELECT DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_AE_HEADERS AEHEADER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('01/01/1970 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('N', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    *when i run that query,it takes a lot of time executing without returning any results (last time it took 4 hours before i cancel it)
    my questions are:
    -what's wrong with that query?
    -how can i change the query in the workflow?
    could anyone please help?

    thank you very much
    i found SQ_XLA_AE_LINES and checked its SQL query,it's a very healthy query
    SELECT  DISTINCT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
          AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE,
    AELINE.CODE_COMBINATION_ID,
    AEHEADER.EVENT_TYPE_CODE,
    NVL(XLA_EVENTS.UPG_BATCH_ID,0) UPG_BATCH_ID
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , XLA_AE_HEADERS AEHEADER
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
       , XLA_EVENTS
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AR_DISTRIBUTIONS_ALL'
              , 'RA_CUST_TRX_LINE_GL_DIST_ALL')
    AND DLINK.APPLICATION_ID = 222
    AND AELINE.APPLICATION_ID = 222
    AND AEHEADER.APPLICATION_ID = 222
    AND XLA_EVENTS.APPLICATION_ID=222
    AND AEHEADER.AE_HEADER_ID = AELINE.AE_HEADER_ID
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND AEHEADER.EVENT_ID=XLA_EVENTS.EVENT_ID
    AND JHEADER.LAST_UPDATE_DATE >=
              TO_DATE('$$LAST_EXTRACT_DATE'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE($$FILTER_BY_LEDGER_ID, 'Y', T.LEDGER_ID, 1) IN ($$LEDGER_ID_LIST)
    AND DECODE($$FILTER_BY_LEDGER_TYPE, 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ($$LEDGER_TYPE_LIST)
    i compared this query with the query that appears in the Error messages,they are different (the Error message is stated in the first post)
    the query that appears in the Error messages misses a couple of lines,specifically in the "FROM" clause and the "WHERE" clause
    what might cause that issue?

  • How to create a single login for multiple apps on tomcat server?

    Hello,
    I am running the most recent versions of apache and tomcat on several dells with XP pro.
    When I login to an app I have created, a session variable is set, but when I browse to one of the other apps on the same computer, it does not recognize that the session variable has been set and I have to login again.
    Can somebody please suggest how, or where I can find docs, to configure tomcat to pass the session variables to other apps?
    Thank you very much.
    Or, can anybody suggest the most effective strategy for creating single-logon functionality for multiple apps with tomcat and apache running on one computer? and with tomcat and apache running on multiple computers? (i.e., each computer has tomcat and apache integrated).
    The database is Oracle 10g running on each computer.
    Thank you again.

    A good place to start is http://www.javaworld.com/javaworld/javatips/jw-javatip34.html
    Also, do a search in this forum on HttpURLConnection. That class allows you to use POST method to send form data to a web server.
    "Hidden" variables are only hidden in HTML. The HTTP that gets POSTed to the web server doesn't distinguish between hidden and not hidden. That is, the content you would write to the HttpURLConnection.getOutputStream() would be something like:
    hidden=1&submit=ok(Of course, the variable names would depend on what the web server was expecting from the form.)
    Also, be sure to set the Content-Type request parameter to "application/x-www-form-urlencoded"

Maybe you are looking for

  • Video output to full screen?

    I have a situation where I need to synch multiple separate recorded backing tracks with video in a live performance situation with a live band. I was told that Logic 8 allows for outputting full screen video synched to music tracks. I have tested thi

  • Intermedia - ORDWEBUTL package NOT found Error.

    Hi, The intermdeia Clipboard gives the following error message : The ORDWEBUTL package was not found in the requested database. Please contact your database administrator. How do I install this package in the database ? My OS : Windows 2000 Pro. Web

  • Does dreamweaver cs4 offer html email templates by any chance?

    I am trying to create an html email which does not fall apart when sent via email. I am currently working on cs3. just wondering if cs4 has this new functionality? cheers! Angus 

  • EIC Authentication error in ESS

    Hi SDN, when I open the Employee Interaction Center(EIC) in Personal informaion (ESS), I am getting the following error, 'An error occurred while data was read from your ERP system. Contact your system administrator.' How can I resolve this error? Th

  • XML validation using Oracle XML parser v2

    Not sure if this is the right forum for this question as I didn't found any. I had a tough time trying to build an XMLSchema (Oracle XML Parser V2) object from the input schema having include/import. Actually I am working on a Java program to validat