Executing Full load for one task

Hi Gurus,
execution plan failed with some errors( unique constraint) errors. now i decided to run a full load for the failed task? is it possible? if possible, can some one give me the exact steps to be followed?
Regards,
Nagarjuna

If you are getting unique constraint error, you can change the index to non unique via the DAC, let it complete and then debug why you are getting duplicate data. By forcing it to full implies that sourece is not providing duplicates but the incremental loads are whoich may be a different issue than source data being bad.

Similar Messages

  • FULL LOAD FOR A SINGLE INFOBJECTS IN ODS

    Hi,
    I am in a scenario, where I need to do a repair full load for one of my ODS object which is extracting data from R/3 system.
    But the problems lies in extraction time, the estimated time to complete the full load is 90hour....which quite impossible to afford in production environment. Hence looking for a solution through which I can load only the modified info object of the ODS.
    Can anybody share some thoughts on this please?
    Regards,
    Kironmoy Banerjee.

    It seems in our case, modifying the transfer structure is troublesome as it involves lot of routines in many info object..
    Can someone suggest me what method I should proceed with to bypass this too long running full load, but need to have a data refresh to that particular infobjects (at least) to make the changes available on old data.
    Any suggestion will be highly appreciated.
    B.R,
    Kironmoy Banerjee

  • Multiple email notifications for one task

    Hi Al,,
    I am getting more than one email notifications for one task in less than an hour. Please let me know if any one has ocem across this situation and got the solution.
    your remarks are highly appreciated with points!
    regards,
    theja

    Hi,
    Are you using this RSWUWFML2 program or Extended notification to receive the email notification.
    If Yes Choose 'One Message per Work item'.
    If no, tell me from where you getting the email.
    Regards,
    Surjith

  • Can I use full load for daily loads in LO Cockpit extraction.

    I am using LO cockpit extraction for 2LIS_11_VAITM and i am using delta load for daily load and I have following doubt.
    Can i use full load for daily loads instead of delta in LO Cockpit extraction.
    Because, my understanding is that full load always takes data from setup tables in ECC and set up tables will not have delta data.
    pls reply. Thanks.

    You are right abt the understanding that full load (atleast for the extractor at hand) brings data from setup tables. You are also right that delta records are not loaded into setup tables.
    So if you plan to do full loads every day, you have to load the setup tables everyday. It is at this juncture that I have to remind you that, filling of setup tables requires a downtime of the system to ensure  that no changes / creates happen during the process.
    Hence performing a full load is not a very good approach to go by...especially when SAP made the job easy by doing all the dirty work of identifying changes and writing them to extraction queue.
    Hope this helps!

  • How to do a  Delta after a Full load for an ODS

    Hi all,
    How to do set up a delta after a full load for an ODS ?
    Thanks
    S N

    Dear SN,
    When ever you already have some Init Requests and Full Requests , your Delta will not be activated for the ODS.
    Now Just use RSSM_SET_REPAIR_FULL_FLAG to make all your Full requests in to Repair requests. then go to IP Delete your Init Req, Then load a Init Req without data transfer if you feel that all your history till date is available or else, you can always load Repair requests, by setting the same in the Infopackage -->Scheduler in the menu , for all your history data .
    Hope this helps.
    Thanks
    Krish
    *Assigning points is way of sayin Thanks in SDN

  • [Macbok Pro 15,4"] Nominal full load for battery

    Hi,
    I've bought a macbook pro 15,4" 2.2GHz two weeks ago
    at start, the health of my battery never grow more than 98%
    now it is not more than 97%
    the nominal full load seem to be 5600mAh and now I have 5439mAh max
    I did 6 full battery unload load, reinstall leopard, and all operation to calibrate the battery but I never got 100% of the battery
    is some have the same behavior ?
    Yoann

    Welcome to the Forums!
    Nominal full charge capacity for a new 15" MB battery is usually stated as 5500 mAh (technically, it's 5555.5 mAh, but programs like iStat and CoconutBattery use the 5500 mAh value).
    But "nominal" is just that. (Try this tip - move your cursor so it hovers over the word 'nominal' and press Apple-Ctrl-D, then read the 3rd definition.) A nominal value is a stated value that the 'real' value does not necessarily correspond exactly to. So, some new batteries have full charge capacities of ~5400 mAh, some ~5700 mAh, etc. Nominal full charge capacity for a 17" battery is 6300 mAh, and my replacement battery (for one that failed) started off at 6650 mAh.
    The full charge capacity will change over the life of the battery, usually going slowly down, but sometimes up slightly.
    First off, be sure you properly calibrated your battery.
    http://docs.info.apple.com/article.html?artnum=86284
    You should do that every couple of months.
    Keep an eye on 'health' but don't obsess. Apple spec is >80% health over 300 cycles.
    Hope this helps...

  • JSP page doesnt load for one user on machine1, but does 4 all on same box

    Hi,
    I have a Citrix application on a box to enable many users to connect remotely to different applications on differnt boxes using the browser on this box. Just like terminal Services or remote desktop
    I tested 15 users and all of them are able to login to the machine and able to open a jsp page of an application located on a differernt box using the browser on this box.
    But for one user I am facing an issue. For this user, I am not able to get the page and instead I only see a message as "page loading" and nothing more than that, Its getting stuck there itself with no progress.
    Please help me,
    Thanks a lot in advance,
    Srinivas.

    Hi,
    I suspect there is something missing as far as the JRE for this guy is concerned. Sothing to do with permissions. java policy, security etc.
    I am not sure if that helps. I am not in the development of the application, and will check with those guys. I only neeed to provode support on this machine and be able to give access to that application which is on Linux. The browser is in Windows machine,
    Thanks,
    Srinivas.

  • Auto  filter settings taking when executed the query for one user id......

    Hi All,
    We are facing one strange issue for  one particular user id  ( i.e for top management) ,when this user executes any of the report  by in putting his own selection parameters say date 15.09.2009 and plant 1111 the output of the report is coming as 15.06.2009 to 19.06.2009 and the plant is also coming as 2222 instead of 1111.
    I have checked it there are variants for this user id .
    When i checked in  the filters in portal it is showing for the cal day15.06.2009 to 19.06.2009,plant 2222,Materials Xyz,Keyfigures say Net amount only, nd more over its getting filtered onalso..... something there is a problem its taking automatic filter values.
    What could be the issue does any one faced this kind of problem?
    Pl help me out by providing your valuable inputs.
    Rgds,
    Rakesh

    closed

  • Need to skip the data load for one of the sub chain.

    Hi All,
    In our project, we have one meta chain and have many sub chains inside that. Today we don't want to run one of the sub chain. As we are in BW7.0, Skip options are not available. Can you please help how to stop the data load for the particular sub-chain.Please suggest.
    Thanks.

    Hi Jalina,
    If this is a frequent request then you can create a custom ABAP Program and then use the below simple logic to skip/run the meta chain, but you will also have to change the link between the sub chains as "Always" instead of "Successful" so that no matter if the dependent chain is successful OR not the nex chain will proceed. If you are comfortable with event then you can also use event to achieve this
    Program Logic- Create a new table where you maintain meaningful values (indicator/description) in it and the program should read the data from the table, Based on the values in the table the program will be successful/fail
    Thanks
    Abhishek Shanbhogue

  • Select multiple channels for one task - control and shift keys don't work

    Hi,
    Following advice from several other posts, I'm trying to select multiple physical channels  from the dropdown list so that I can then use the "unflatten channel string" function later on. It looks like if I hold down either shift or cntrl I should be able to select multiple channels from the list, however, this doesn't work for me. I tried going into the "key navigations" property for the task channel control but it doesn't have any associations listed. Has anyone else come across this and found a fix?  I have previously got around this by stating multiple channels with a colon (eg. ai1:3) but this doesn't work for the unflatten function.
    Thanks, 
    Claire.
    Solved!
    Go to Solution.

    Thanks for your reply Dennis, I see what you mean, unfortunately that doesn't solve my problem so I'm missing something else. I've attached my VIs. I expected that the output from the unflattenstring block would be an array of either the channel references or the names I had given them, and that therefore the "array size" in the 4th pane would be "9" (equal to the number of channels).  Instead the output is a an "unnamed task#(some number)" when I look at it with a probe. Perhaps I'm misunderstanding how the task naming works?
    THanks, 
    Claire.
    Attachments:
    Analog SW Trigger_v3.vi ‏33 KB
    Cont Acq&Graph Voltage-Analog SW Trigger_cfjv3.vi ‏64 KB

  • Memory and CPU load for one Oracle instance

    I need to design a database server with 5 oracle instances.
    Each instance have only 2 or 3 connected application users and the rate of transactions is about 10 transactions/hour.
    How can I estimate to total memory and CPU needed for my server ?
    The minimum esimation : 2x2Ghz and 2Go of RAM per instance is it pertinent ?
    Dominik

    Dominik wrote:
    I need to design a database server with 5 oracle instances.
    Each instance have only 2 or 3 connected application users and the rate of transactions is about 10 transactions/hour.
    How can I estimate to total memory and CPU needed for my server ?
    The minimum esimation : 2x2Ghz and 2Go of RAM per instance is it pertinent ?
    DominikThough the picked specs look okay but the better answer, as being mentioned already, would be "it depends" . At this moment, almost all the machines are coming with very good specs so 2gb probably would be the least . You can even have a 4gb RAM system with x86_64 architecture , leaving room for future growth if needed. About setting the memory for the SGA, if the release is 11g, setting the Memory_Target would take care of the the memory allocation for both SGA and PGA and you can start by the default that Oracle picks which I guess is about 40% of the total RAM. Afterwards, you can monitor the performance and can bring down the memory settings if you feel its overly allocated.
    That said, I think that this is a question better answered by the capacity planning and management people. I shall see if I can ask my friend, who is in this domain to comment on it.
    Just my 2 cents.
    Aman....

  • Delta load to Full load

    Hello Gurus,
    As per my knowledge while creating <b>delta load</b> form <b>R/3 to BW</b> we need to <i>declare the type of load in R/3 and in the infopackage</i> ....
    Where <b>my doubt</b> come on,
    i need to <b>convert delta load to full load for one info object</b>  ,is their <i>any steps to do in R/3 for changing the type of load </i>.
    where <i>i had just changed the load type in only Info package</i> ..<b>.is it runs fine</b>.
    Plz assist me
    Thanks in advance
    SAP

    Hi,
    There is nothing specific needs to be done in R/3.
    A data source is delta enabled but still you can load it through full load and in that case you need not do an initialization just schedule a full load package.
    i need to convert delta load to full load for one info object ,is their any steps to do in R/3 for changing the type of load .
    For this just delete the init which is present there and create the new package with option full update.Just make sure that the last delta is loaded for precaution.
    Since infoobject is overwrite so no need to delete the data also every day there will be full load and it will bring the new data along with.
    This wil start loading the full load into your object.
    Hope it clears
    Thanks
    Message was edited by:
            Ajeet Singh

  • How to do a full load in DAC for a particular Module?

    Hi,
    We have one Execution plan loading all the BI Analytics module as of now.
    1)Financials,
    2)Procurement and Spend,
    3)Supply Chain and Order Management and
    4)EAM
    Issue is if  i go to Tools-->ETL Management-->Reset Data Sources  in DAC, it refreshes dates for all the Tables in the warehouse. ( and hence full load for all the Modules happens when i start the execution plan)
    I dont want full load for other modules, just want to do full load for a particular Module lets say EAM Module ( and also want to make sure that this particular full load doesnt create issues with the  data for other modules)
    Let me know how to achieve this.
    Any help in this regard would be highly appreciated.
    Thanks
    Ashish

    I can get a list of all the Facts and dimensions for a particular module from BI Apps Content guide... and then i can go and make refresh dates as Null for those Tables.
    Now should i run the execution plan for one particular module or an Execution plan containing all the modules?
    Cause i got this from someone..
    "The Data models for the different modules have many common dimensions, so a reload of one module could leave the facts in another in an invalid state as the dimension keys would be moved.
    Therefore it should not be expected that you can reload a module in isolation.
    However, you can selectively do a reload of any table by resetting it's refresh date in the DAC console.
    Then the DAC will determine the dependent objects that will also need a full load in order to maintain data consistency."
    Thanks
    Ashish

  • DAC - Can I run more then one Exection Plans with full load option enabled?

    Hello All,
    I have configured for subject Area sales – Orders, Backlog and Invoices (OBIEE)
    Now I want to run DAC to load the underline tables for subject Areas Financials – Revenue Analysis and Financials – GL Budget and Expenses, I have identified the relevant fact tables and built a Execution Plan in DAC.
    DAC - In which mode should I run the execution plan?
    With full load option enabled or disabled because already full load run is done for sales – Orders, Backlog and Invoices and this run should not disturb the existing data
    If I run with full load option enabled do it overrides the existing data for other subject area
    Please help me I am bit confused!
    Regards,
    SMA
    Edited by: SMA on Aug 18, 2009 3:36 PM

    No need of combining all the subject areas into one execution plan. Simply run the second ETL without checking the full load option. It will do the full load for the tables related to the newly added subject areas and incremental for the common tables between old and new subject areas.

  • Project Analytics 7.9.6.1 - Error while running a full load

    Hi All,
    I am performing a full load for Projects Analytics and get the following error,
    =====================================
    ERROR OUTPUT
    =====================================
    1103 SEVERE Wed Nov 18 02:49:36 WST 2009 Could not attach to workflow because of errorCode 36331 For workflow SDE_ORA_CodeDimension_Gl_Account
    1104 SEVERE Wed Nov 18 02:49:36 WST 2009
    ANOMALY INFO::: Error while executing : INFORMATICA TASK:SDE_ORA11510_Adaptor:SDE_ORA_CodeDimension_Gl_Account:(Source : FULL Target : FULL)
    MESSAGE:::
    Irrecoverable Error
    Error while contacting Informatica server for getting workflow status for SDE_ORA_CodeDimension_Gl_Account
    Error Code = 36331:Unknown reason for error code 36331
    Pmcmd output :
    Session log initialises NULL value to mapping parameter MPLT_ADI_CODES.$$CATEGORY. This is then used insupsequent SQL and results in ORA-00936: missing expression error following are the initialization section and the load section containing the error in the log
    Initialisation
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_11_5_10] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [ORA_11_5_10.DATAWAREHOUSE.SDE_ORA11510_Adaptor.SDE_ORA_CodeDimension_Gl_Account_Segments.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[MPLT_ADI_CODES.$$CATEGORY].
    DIRECTOR> VAR_27028 Use override value [4] for mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[MPLT_SA_ORA_CODES.$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_CodeDimension_Gl_Account_Segments] at [Wed Nov 18 02:49:11 2009].
    DIRECTOR> TM_6683 Repository Name: [repo_service]
    DIRECTOR> TM_6684 Server Name: [int_service]
    DIRECTOR> TM_6686 Folder: [SDE_ORA11510_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_CodeDimension_Gl_Account_Segments] Run Instance Name: [] Run Id: [17]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_CodeDimension_GL_Account_Segments [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\PowerCenter8.6.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_CodeDimension_Gl_Account_Segments].
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6703 Session [SDE_ORA_CodeDimension_Gl_Account_Segments] is run by 32-bit Integration Service [node01_ASG596138], version [8.6.1], build [1218].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [UNICODE]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6185 Warning. Code page validation is disabled in this session.
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [orcl], user [DAC_REP]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_Master_Map] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_Master_Code] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_W_CODE_D] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_CodeDimension_Gl_Account_Segments]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Nov 18 02:49:14 2009)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Nov 18 02:49:14 2009)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 32000000 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [asgdev], user [APPS]
    READER_1_1_1> BLKR_16051 Source database connection [ORA_11_5_10] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8147 Writer: Target is database [orcl], user [DAC_REP], bulk mode [OFF]
    WRITER_1_*_1> WRT_8221 Target database connection [DataWarehouse] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL INSERT statement:
    INSERT INTO W_CODE_D(DATASOURCE_NUM_ID,SOURCE_CODE,SOURCE_CODE_1,SOURCE_CODE_2,SOURCE_CODE_3,SOURCE_NAME_1,SOURCE_NAME_2,CATEGORY,LANGUAGE_CODE,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,W_UPDATE_DT,TENANT_ID) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL UPDATE statement:
    UPDATE W_CODE_D SET SOURCE_CODE_1 = ?, SOURCE_CODE_2 = ?, SOURCE_CODE_3 = ?, SOURCE_NAME_1 = ?, SOURCE_NAME_2 = ?, MASTER_DATASOURCE_NUM_ID = ?, MASTER_CODE = ?, MASTER_VALUE = ?, W_INSERT_DT = ?, W_UPDATE_DT = ?, TENANT_ID = ? WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL DELETE statement:
    DELETE FROM W_CODE_D WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_CODE_D]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    Load section
    *****START LOAD SESSION*****
    Load Start Time: Wed Nov 18 02:49:16 2009
    Target tables:
    W_CODE_D
    READER_1_1_1> RR_4029 SQ Instance [mplt_BC_ORA_Codes_GL_Account_Segments.Sq_Fnd_Flex_Values] User specified SQL Query [SELECT
    FND_FLEX_VALUES.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUES.FLEX_VALUE,
    MAX(FND_FLEX_VALUES_TL.DESCRIPTION),
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM,
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME
    FROM
    FND_FLEX_VALUES,
    FND_FLEX_VALUES_TL,
    FND_ID_FLEX_SEGMENTS,
    FND_SEGMENT_ATTRIBUTE_VALUES
    WHERE
    FND_FLEX_VALUES.FLEX_VALUE_ID = FND_FLEX_VALUES_TL.FLEX_VALUE_ID AND FND_FLEX_VALUES_TL.LANGUAGE ='US' AND
    FND_ID_FLEX_SEGMENTS.FLEX_VALUE_SET_ID =FND_FLEX_VALUES.FLEX_VALUE_SET_ID AND
    FND_ID_FLEX_SEGMENTS.APPLICATION_ID = 101 AND
    FND_ID_FLEX_SEGMENTS.ID_FLEX_CODE ='GL#' AND
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM =FND_SEGMENT_ATTRIBUTE_VALUES.ID_FLEX_NUM AND
    FND_SEGMENT_ATTRIBUTE_VALUES.APPLICATION_ID =101 AND
    FND_SEGMENT_ATTRIBUTE_VALUES.ID_FLEX_CODE = 'GL#' AND
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME=FND_SEGMENT_ATTRIBUTE_VALUES.APPLICATION_COLUMN_NAME AND
    FND_SEGMENT_ATTRIBUTE_VALUES.ATTRIBUTE_VALUE ='Y'
    GROUP BY
    FND_FLEX_VALUES.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUES.FLEX_VALUE,
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM,
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Nov 18 02:49:17 2009)
    READER_1_1_1> RR_4050 First row returned from database to reader : (Wed Nov 18 02:49:17 2009)
    LKPDP_3> DBG_21312 Lookup Transformation [mplt_ADI_Codes.Lkp_W_CODE_D]: Lookup override sql to create cache: SELECT W_CODE_D.SOURCE_NAME_1 AS SOURCE_NAME_1, W_CODE_D.SOURCE_NAME_2 AS SOURCE_NAME_2, W_CODE_D.MASTER_DATASOURCE_NUM_ID AS MASTER_DATASOURCE_NUM_ID, W_CODE_D.MASTER_CODE AS MASTER_CODE, W_CODE_D.MASTER_VALUE AS MASTER_VALUE, W_CODE_D.W_INSERT_DT AS W_INSERT_DT, W_CODE_D.TENANT_ID AS TENANT_ID, W_CODE_D.DATASOURCE_NUM_ID AS DATASOURCE_NUM_ID, W_CODE_D.SOURCE_CODE AS SOURCE_CODE, W_CODE_D.CATEGORY AS CATEGORY, W_CODE_D.LANGUAGE_CODE AS LANGUAGE_CODE FROM W_CODE_D
    WHERE
    W_CODE_D.CATEGORY IN () ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,SOURCE_NAME_1,SOURCE_NAME_2,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,TENANT_ID
    LKPDP_3> TE_7212 Increasing [Index Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [1000000] to [4734976].
    LKPDP_3> TE_7212 Increasing [Data Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [2000000] to [2007040].
    READER_1_1_1> BLKR_16019 Read [625] rows, read [0] error rows for source table [FND_ID_FLEX_SEGMENTS] instance name [mplt_BC_ORA_Codes_GL_Account_Segments.FND_ID_FLEX_SEGMENTS]
    READER_1_1_1> BLKR_16008 Reader run completed.
    LKPDP_3> TM_6660 Total Buffer Pool size is 609824 bytes and Block size is 65536 bytes.
    LKPDP_3:READER_1_1> DBG_21438 Reader: Source is [orcl], user [DAC_REP]
    LKPDP_3:READER_1_1> BLKR_16051 Source database connection [DataWarehouse] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    LKPDP_3:READER_1_1> BLKR_16003 Initialization completed successfully.
    LKPDP_3:READER_1_1> BLKR_16007 Reader run started.
    LKPDP_3:READER_1_1> RR_4049 SQL Query issued to database : (Wed Nov 18 02:49:18 2009)
    LKPDP_3:READER_1_1> CMN_1761 Timestamp Event: [Wed Nov 18 02:49:18 2009]
    LKPDP_3:READER_1_1> RR_4035 SQL Error [
    ORA-00936: missing expression
    Could you please suggest what the issue might be and how it can be fixed?
    Many thanks,
    Kiran

    I have continued related detains in the following thread,
    Mapping Parameter  $$CATEGORY not included in the parameter file (7.9.6.1)
    Apologies for the inconvenience.
    Thanks,
    Kiran

Maybe you are looking for