Error while running loading full load

Hi all,
I'm getting below error while running the mappings in DAC
INFO : LM_36488 [Fri Sep 11 02:47:04 2009] : (580|4952) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [VAR_27028 Use override value [4,0] for mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].]
Regards
Rama

Hi all,
Detail Below Error,
INFO : LM_36435 [Fri Sep 11 03:11:49 2009] : (580|4952) Starting execution of workflow [SDE_ORA_CodeDimension_CostCenter] in folder [SDE_ORA11510_Adaptor] last saved by user [Administrator].
INFO : LM_44195 [Fri Sep 11 03:11:49 2009] : (580|4952) Workflow [SDE_ORA_CodeDimension_CostCenter] service level [SLPriority:5,SLDispatchWaitTime:1800].
INFO : LM_36330 [Fri Sep 11 03:11:49 2009] : (580|4952) Start task instance [Start]: Execution started.
INFO : LM_36318 [Fri Sep 11 03:11:49 2009] : (580|4952) Start task instance [Start]: Execution succeeded.
INFO : LM_36505 : (580|4952) Link [Start --> SDE_ORA_CodeDimension_CostCenter]: empty expression string, evaluated to TRUE.
INFO : LM_36388 [Fri Sep 11 03:11:49 2009] : (580|4952) Session task instance [SDE_ORA_CodeDimension_CostCenter] is waiting to be started.
INFO : LM_36682 [Fri Sep 11 03:11:49 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter]: started a process with pid [5644] on node [node01_obidevapp1].
INFO : LM_36330 [Fri Sep 11 03:11:49 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter]: Execution started.
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6793 Fetching initialization properties from the Integration Service. : (Fri Sep 11 03:11:49 2009)]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [DISP_20305 The [Preparer] DTM with process id [5644] is running on node [node01_obidevapp1].
: (Fri Sep 11 03:11:49 2009)]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [PETL_24036 Beginning the prepare phase for the session.]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6721 Started [Connect to Repository].]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6722 Finished [Connect to Repository]. It took [0.312494] seconds.]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6794 Connected to repository [Oracle_BI_DW_Base] in domain [Domain_obidevapp1] user [Administrator]]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6721 Started [Fetch Session from Repository].]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6722 Finished [Fetch Session from Repository]. It took [0.406242] seconds.]
INFO : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [VAR_27028 Use override value [(2,4,5,9,0)] for mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [CMN_1761 Timestamp Event: [Fri Sep 11 03:11:50 2009]]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [VAR_27056 Data conversion error in converting [(2,4,5,9,0)].]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [CMN_1761 Timestamp Event: [Fri Sep 11 03:11:50 2009]]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [VAR_27054 Error in assigning initial data value to mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [CMN_1761 Timestamp Event: [Fri Sep 11 03:11:50 2009]]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6270 Error: Variable parameter expansion error.]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [CMN_1761 Timestamp Event: [Fri Sep 11 03:11:50 2009]]
ERROR : LM_36488 [Fri Sep 11 03:11:50 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] : [TM_6163 Error initializing variables and parameters for the partition.]
ERROR : LM_36320 [Fri Sep 11 03:11:52 2009] : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter]: Execution failed.
WARNING : LM_36331 : (580|4972) Session task instance [SDE_ORA_CodeDimension_CostCenter] failed and its "fail parent if this task fails" setting is turned on. So, Workflow [SDE_ORA_CodeDimension_CostCenter] will be failed.
ERROR : LM_36320 [Fri Sep 11 03:11:52 2009] : (580|4972) Workflow [SDE_ORA_CodeDimension_CostCenter]: Execution failed.
Regards,
Rama

Similar Messages

  • Project Analytics 7.9.6.1 - Error while running a full load

    Hi All,
    I am performing a full load for Projects Analytics and get the following error,
    =====================================
    ERROR OUTPUT
    =====================================
    1103 SEVERE Wed Nov 18 02:49:36 WST 2009 Could not attach to workflow because of errorCode 36331 For workflow SDE_ORA_CodeDimension_Gl_Account
    1104 SEVERE Wed Nov 18 02:49:36 WST 2009
    ANOMALY INFO::: Error while executing : INFORMATICA TASK:SDE_ORA11510_Adaptor:SDE_ORA_CodeDimension_Gl_Account:(Source : FULL Target : FULL)
    MESSAGE:::
    Irrecoverable Error
    Error while contacting Informatica server for getting workflow status for SDE_ORA_CodeDimension_Gl_Account
    Error Code = 36331:Unknown reason for error code 36331
    Pmcmd output :
    Session log initialises NULL value to mapping parameter MPLT_ADI_CODES.$$CATEGORY. This is then used insupsequent SQL and results in ORA-00936: missing expression error following are the initialization section and the load section containing the error in the log
    Initialisation
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_11_5_10] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [ORA_11_5_10.DATAWAREHOUSE.SDE_ORA11510_Adaptor.SDE_ORA_CodeDimension_Gl_Account_Segments.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27027 Use default value [] for mapping parameter:[MPLT_ADI_CODES.$$CATEGORY].
    DIRECTOR> VAR_27028 Use override value [4] for mapping parameter:[MPLT_SA_ORA_CODES.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [DEFAULT] for mapping parameter:[MPLT_SA_ORA_CODES.$$TENANT_ID].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_CodeDimension_Gl_Account_Segments] at [Wed Nov 18 02:49:11 2009].
    DIRECTOR> TM_6683 Repository Name: [repo_service]
    DIRECTOR> TM_6684 Server Name: [int_service]
    DIRECTOR> TM_6686 Folder: [SDE_ORA11510_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_CodeDimension_Gl_Account_Segments] Run Instance Name: [] Run Id: [17]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_CodeDimension_GL_Account_Segments [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [C:\Informatica\PowerCenter8.6.1\server\infa_shared\Storage] will be used as storage directory for session [SDE_ORA_CodeDimension_Gl_Account_Segments].
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6703 Session [SDE_ORA_CodeDimension_Gl_Account_Segments] is run by 32-bit Integration Service [node01_ASG596138], version [8.6.1], build [1218].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [UNICODE]
    MAPPING> CMN_1570 Server Code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6185 Warning. Code page validation is disabled in this session.
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [orcl], user [DAC_REP]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_Master_Map] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_Master_Code] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> CMN_1716 Lookup [mplt_ADI_Codes.Lkp_W_CODE_D] uses database connection [Relational:DataWarehouse] in code page [MS Windows Latin 1 (ANSI), superset of Latin1]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_CodeDimension_Gl_Account_Segments]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Wed Nov 18 02:49:14 2009)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Wed Nov 18 02:49:14 2009)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 32000000 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [asgdev], user [APPS]
    READER_1_1_1> BLKR_16051 Source database connection [ORA_11_5_10] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8147 Writer: Target is database [orcl], user [DAC_REP], bulk mode [OFF]
    WRITER_1_*_1> WRT_8221 Target database connection [DataWarehouse] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL INSERT statement:
    INSERT INTO W_CODE_D(DATASOURCE_NUM_ID,SOURCE_CODE,SOURCE_CODE_1,SOURCE_CODE_2,SOURCE_CODE_3,SOURCE_NAME_1,SOURCE_NAME_2,CATEGORY,LANGUAGE_CODE,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,W_UPDATE_DT,TENANT_ID) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL UPDATE statement:
    UPDATE W_CODE_D SET SOURCE_CODE_1 = ?, SOURCE_CODE_2 = ?, SOURCE_CODE_3 = ?, SOURCE_NAME_1 = ?, SOURCE_NAME_2 = ?, MASTER_DATASOURCE_NUM_ID = ?, MASTER_CODE = ?, MASTER_VALUE = ?, W_INSERT_DT = ?, W_UPDATE_DT = ?, TENANT_ID = ? WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8124 Target Table W_CODE_D :SQL DELETE statement:
    DELETE FROM W_CODE_D WHERE DATASOURCE_NUM_ID = ? AND SOURCE_CODE = ? AND CATEGORY = ? AND LANGUAGE_CODE = ?
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_CODE_D]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    Load section
    *****START LOAD SESSION*****
    Load Start Time: Wed Nov 18 02:49:16 2009
    Target tables:
    W_CODE_D
    READER_1_1_1> RR_4029 SQ Instance [mplt_BC_ORA_Codes_GL_Account_Segments.Sq_Fnd_Flex_Values] User specified SQL Query [SELECT
    FND_FLEX_VALUES.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUES.FLEX_VALUE,
    MAX(FND_FLEX_VALUES_TL.DESCRIPTION),
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM,
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME
    FROM
    FND_FLEX_VALUES,
    FND_FLEX_VALUES_TL,
    FND_ID_FLEX_SEGMENTS,
    FND_SEGMENT_ATTRIBUTE_VALUES
    WHERE
    FND_FLEX_VALUES.FLEX_VALUE_ID = FND_FLEX_VALUES_TL.FLEX_VALUE_ID AND FND_FLEX_VALUES_TL.LANGUAGE ='US' AND
    FND_ID_FLEX_SEGMENTS.FLEX_VALUE_SET_ID =FND_FLEX_VALUES.FLEX_VALUE_SET_ID AND
    FND_ID_FLEX_SEGMENTS.APPLICATION_ID = 101 AND
    FND_ID_FLEX_SEGMENTS.ID_FLEX_CODE ='GL#' AND
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM =FND_SEGMENT_ATTRIBUTE_VALUES.ID_FLEX_NUM AND
    FND_SEGMENT_ATTRIBUTE_VALUES.APPLICATION_ID =101 AND
    FND_SEGMENT_ATTRIBUTE_VALUES.ID_FLEX_CODE = 'GL#' AND
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME=FND_SEGMENT_ATTRIBUTE_VALUES.APPLICATION_COLUMN_NAME AND
    FND_SEGMENT_ATTRIBUTE_VALUES.ATTRIBUTE_VALUE ='Y'
    GROUP BY
    FND_FLEX_VALUES.FLEX_VALUE_SET_ID,
    FND_FLEX_VALUES.FLEX_VALUE,
    FND_ID_FLEX_SEGMENTS.ID_FLEX_NUM,
    FND_ID_FLEX_SEGMENTS.APPLICATION_COLUMN_NAME]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Wed Nov 18 02:49:17 2009)
    READER_1_1_1> RR_4050 First row returned from database to reader : (Wed Nov 18 02:49:17 2009)
    LKPDP_3> DBG_21312 Lookup Transformation [mplt_ADI_Codes.Lkp_W_CODE_D]: Lookup override sql to create cache: SELECT W_CODE_D.SOURCE_NAME_1 AS SOURCE_NAME_1, W_CODE_D.SOURCE_NAME_2 AS SOURCE_NAME_2, W_CODE_D.MASTER_DATASOURCE_NUM_ID AS MASTER_DATASOURCE_NUM_ID, W_CODE_D.MASTER_CODE AS MASTER_CODE, W_CODE_D.MASTER_VALUE AS MASTER_VALUE, W_CODE_D.W_INSERT_DT AS W_INSERT_DT, W_CODE_D.TENANT_ID AS TENANT_ID, W_CODE_D.DATASOURCE_NUM_ID AS DATASOURCE_NUM_ID, W_CODE_D.SOURCE_CODE AS SOURCE_CODE, W_CODE_D.CATEGORY AS CATEGORY, W_CODE_D.LANGUAGE_CODE AS LANGUAGE_CODE FROM W_CODE_D
    WHERE
    W_CODE_D.CATEGORY IN () ORDER BY DATASOURCE_NUM_ID,SOURCE_CODE,CATEGORY,LANGUAGE_CODE,SOURCE_NAME_1,SOURCE_NAME_2,MASTER_DATASOURCE_NUM_ID,MASTER_CODE,MASTER_VALUE,W_INSERT_DT,TENANT_ID
    LKPDP_3> TE_7212 Increasing [Index Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [1000000] to [4734976].
    LKPDP_3> TE_7212 Increasing [Data Cache] size for transformation [mplt_ADI_Codes.Lkp_W_CODE_D] from [2000000] to [2007040].
    READER_1_1_1> BLKR_16019 Read [625] rows, read [0] error rows for source table [FND_ID_FLEX_SEGMENTS] instance name [mplt_BC_ORA_Codes_GL_Account_Segments.FND_ID_FLEX_SEGMENTS]
    READER_1_1_1> BLKR_16008 Reader run completed.
    LKPDP_3> TM_6660 Total Buffer Pool size is 609824 bytes and Block size is 65536 bytes.
    LKPDP_3:READER_1_1> DBG_21438 Reader: Source is [orcl], user [DAC_REP]
    LKPDP_3:READER_1_1> BLKR_16051 Source database connection [DataWarehouse] code page: [MS Windows Latin 1 (ANSI), superset of Latin1]
    LKPDP_3:READER_1_1> BLKR_16003 Initialization completed successfully.
    LKPDP_3:READER_1_1> BLKR_16007 Reader run started.
    LKPDP_3:READER_1_1> RR_4049 SQL Query issued to database : (Wed Nov 18 02:49:18 2009)
    LKPDP_3:READER_1_1> CMN_1761 Timestamp Event: [Wed Nov 18 02:49:18 2009]
    LKPDP_3:READER_1_1> RR_4035 SQL Error [
    ORA-00936: missing expression
    Could you please suggest what the issue might be and how it can be fixed?
    Many thanks,
    Kiran

    I have continued related detains in the following thread,
    Mapping Parameter  $$CATEGORY not included in the parameter file (7.9.6.1)
    Apologies for the inconvenience.
    Thanks,
    Kiran

  • Error in Source System while running the Full Load

    HI Experts,
    We are using 0CRM_UT_SRV_CONT_I datasource from CRM to extract data.
    It is giving error while running the full load as 'Error Occured in Source system' with messaage Number : RSM340.
    I  want to mention certain points here:
    1) In RSA3 its running fine.
    2) Connection between BW and CRM system is OK.
    3) We have replicated datasource and send active veriosn from Dev system.
    4) When I run Init without data, Init set sucessful with entry in RSA7 in CRM.
    5) Delta run fine with green status.
    6) As source system is static client, hence do data for delta.
    The Main issue happens when we run either the Full load or Init With Data transfer.
    Kindly suggest.
    Thanks
    Mayank

    Hi Mayank,
    see in source system (CRM) with transaction SM21 and ST22 the error log and and then you let us know.
    Charly

  • Errors: ORA-00054 & ORA-01452 while running DAC Full Load

    Hi Friends,
    Previously, I ran full load...it went well. And, I did some sample reports also in BI APPS 7.9.6.2
    Now, I modified few parameters as per the Business and I try to run Full Load again...But I struck with few similar errors. I cleared couple of DB Errors.
    Please, help me out to solve the below errors.
    1. ANOMALY INFO::: Error while executing : TRUNCATE TABLE:W_SALES_BOOKING_LINE_F
    MESSAGE:::com.siebel.etl.database.IllegalSQLQueryException: DataWarehouse:TRUNCATE TABLE W_SALES_BOOKING_LINE_F
    ORA-00054: resource busy and acquire with NOWAIT specified or timeout expired
    --I checked W_SALES_BOOKING_LINE_F, it contain s data.
    2. ANOMALY INFO::: Error while executing : CREATE INDEX:W_GL_REVN_F:W_GL_REVN_F_U1
    MESSAGE:::java.lang.Exception: Error while execution : CREATE UNIQUE INDEX
         W_GL_REVN_F_U1
    ON
         W_GL_REVN_F
         INTEGRATION_ID ASC
         ,DATASOURCE_NUM_ID ASC
    NOLOGGING
    with error DataWarehouse:CREATE UNIQUE INDEX
         W_GL_REVN_F_U1
    ON
         W_GL_REVN_F
         INTEGRATION_ID ASC
         ,DATASOURCE_NUM_ID ASC
    NOLOGGING
    ORA-01452: cannot CREATE UNIQUE INDEX; duplicate keys found
    -- Yes, I found duplicate values in this table W_GL_REVN_F. But, how can I rectify it. I did some engineering, but failed.
    please tell me the steps to acheive....
    Thanks in advance..
    Stone

    Hi, Please see the answers (in bold) below.
    1. ANOMALY INFO::: Error while executing : TRUNCATE TABLE:W_SALES_BOOKING_LINE_F
    MESSAGE:::com.siebel.etl.database.IllegalSQLQueryException: DataWarehouse:TRUNCATE TABLE W_SALES_BOOKING_LINE_F
    ORA-00054: resource busy and acquire with NOWAIT specified or timeout expired
    --I checked W_SALES_BOOKING_LINE_F, it contain s data.
    Just restart the load, It seems like your DB processes are busy and the table still has a  lock on it which means something is not yet Commited/Rolled Back.
    If this issue repeats you can mail your DBA and ask him to look in to the issue
    2. ANOMALY INFO::: Error while executing : CREATE INDEX:W_GL_REVN_F:W_GL_REVN_F_U1
    MESSAGE:::java.lang.Exception: Error while execution : CREATE UNIQUE INDEX
    W_GL_REVN_F_U1
    ON
    W_GL_REVN_F
    INTEGRATION_ID ASC
         ,DATASOURCE_NUM_ID ASC
         NOLOGGING
         with error DataWarehouse:CREATE UNIQUE INDEX
         W_GL_REVN_F_U1
         ON
         W_GL_REVN_F
         INTEGRATION_ID ASC
         ,DATASOURCE_NUM_ID ASC
         NOLOGGING
         ORA-01452: cannot CREATE UNIQUE INDEX; duplicate keys found
         -- Yes, I found duplicate values in this table W_GL_REVN_F. But, how can I rectify it. I did some engineering, but failed.
         please tell me the steps to achieve....
    please execute this sql and get the duplicate values. If the count is less you can delete the records based on ROW_WID
    How  many duplicates do you have in total?
    *1. SELECT INTEGRATION_ID,DATASOURCE_NUM_ID,count(*) FROM W_GL_REVN_F*
    GROUP BY INTEGRATION_ID, DATASOURCE_NUM_ID
    HAVING COUNT()>1*
    *2. SELECT ROW_WID,DATASOURCE_NUM_ID,INTEGRATION_ID FROM W_GL_REVN_F*
    WHERE INTEGRATION_ID= (from 1st query)
    *3. DELETE from W_GL_REVN_F where ROW_WID=( from 2nd query)*
    Hope this helps !!

  • Error While running the ETL Load in DAC (BI Financial Analytics)

    Hi All,
    I have Installed and Configured BI Applictions 7.9.5 and Informatic8.1.1. For the first time when we run the ETL Load in DAC it has failed.for us every Test Connection was sucess.and getting the error message as below.
    The log file which I pasted below is from the path
    /u01/app/oracle/product/Informatica/PowerCenter8.1.1/server/infa_shared
    /SessLogs
    SDE_ORAR12_Adaptor.SDE_ORA_GL_AP_LinkageInformation_Extract_Full.log
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [ORA_R12] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [9] for mapping parameter:[$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value ['Y'] for mapping parameter:[$$FILTER_BY_LEDGER_ID].
    DIRECTOR> VAR_27028 Use override value ['N'] for mapping parameter:[$$FILTER_BY_LEDGER_TYPE].
    DIRECTOR> VAR_27028 Use override value [04/02/2007] for mapping parameter:[$$INITIAL_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$LAST_EXTRACT_DATE].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[$$LEDGER_ID_LIST].
    DIRECTOR> VAR_27028 Use override value ['NONE'] for mapping parameter:[$$LEDGER_TYPE_LIST].
    DIRECTOR> TM_6014 Initializing session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full] at [Thu Feb 12 12:49:33 2009]
    DIRECTOR> TM_6683 Repository Name: [DEV_Oracle_BI_DW_Rep]
    DIRECTOR> TM_6684 Server Name: [DEV_Oracle_BI_DW_Rep_Integration_Service]
    DIRECTOR> TM_6686 Folder: [SDE_ORAR12_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_ORA_GL_AP_LinkageInformation_Extract_Full]
    DIRECTOR> TM_6101 Mapping name: SDE_ORA_GL_AP_LinkageInformation_Extract [version 1]
    DIRECTOR> TM_6827 [u01/app/oracle/product/Informatica/PowerCenter8.1.1/server/infa_shared/Storage] will be used as storage directory for session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full].
    DIRECTOR> CMN_1805 Recovery cache will be deleted when running in normal mode.
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,apps@devr12 bawdev@devbi]
    DIRECTOR> TM_6703 Session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full] is run by 64-bit Integration Service [node01_oratestbi], version [8.1.1 SP4], build [0817].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [ASCII]
    MAPPING> CMN_1570 Server Code page: [ISO 8859-1 Western European]
    MAPPING> TM_6151 Session Sort Order: [Binary]
    MAPPING> TM_6156 Using LOW precision decimal arithmetic
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6307 DTM Error Log Disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TE_7004 Transformation Parse Warning; transformation continues...
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Thu Feb 12 12:49:34 2009)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Thu Feb 12 12:49:34 2009)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 12582912 bytes and Block size is 128000 bytes.
    READER_1_1_1> DBG_21438 Reader: Source is [devr12.tessco.com], user [apps]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [DEVBI], user [bawdev], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8124 Target Table W_GL_LINKAGE_INFORMATION_GS :SQL INSERT statement:
    INSERT INTO W_GL_LINKAGE_INFORMATION_GS(SOURCE_DISTRIBUTION_ID,JOURNAL_LINE_INTEGRATION_ID,LEDGER_ID,LEDGER_TYPE,DISTRIBUTION_SOURCE,JE_BATCH_NAME,JE_HEADER_NAME,JE_LINE_NUM,POSTED_ON_DT,SLA_TRX_INTEGRATION_ID,DATASOURCE_NUM_ID) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GL_LINKAGE_INFORMATION_GS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Thu Feb 12 12:49:34 2009
    Target tables:
    W_GL_LINKAGE_INFORMATION_GS
    READER_1_1_1> RR_4029 SQ Instance [SQ_XLA_AE_LINES] User specified SQL Query [SELECT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    DLINK.ACCOUNTING_LINE_CODE LINE_CODE,
          AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
        JBATCH.NAME BATCH_NAME,
       JHEADER.NAME HEADER_NAME,
          PER.END_DATE
    FROM XLA_DISTRIBUTION_LINKS DLINK
       , GL_IMPORT_REFERENCES        GLIMPREF
       , XLA_AE_LINES                              AELINE
       , GL_JE_HEADERS                         JHEADER
       , GL_JE_BATCHES                         JBATCH
       , GL_LEDGERS                                 T
       , GL_PERIODS   PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
             (  'AP_INV_DIST', 'AP_PMT_DIST'
              , 'AP_PREPAY')
    AND DLINK.APPLICATION_ID = 200
    AND AELINE.APPLICATION_ID = 200
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID         = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID         = DLINK.AE_HEADER_ID        
    AND AELINE.AE_LINE_NUM           = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID   = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID       = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID                   = T.LEDGER_ID
    AND JHEADER.STATUS                         = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
              TO_DATE('04/02/2007 00:00:00'
                    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('Y', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')]
    READER_1_1_1> RR_4049 SQL Query issued to database : (Thu Feb 12 12:49:34 2009)
    READER_1_1_1> CMN_1761 Timestamp Event: [Thu Feb 12 12:49:34 2009]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-01114: IO error writing block to file 513 (block # 328465)
    ORA-27072: File I/O error
    Linux-x86_64 Error: 28: No space left on device
    Additional information: 4
    Additional information: 328465
    Additional information: -1
    ORA-01114: IO error writing block to file 513 (block # 328465)
    ORA-27072: File I/O error
    Linux-x86_64 Error: 28: No space left on device
    Additional information: 4
    Additional information: 328465
    Additional information: -1
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    DLINK.ACCOUNTING_LINE_CODE LINE_CODE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
    JBATCH.NAME BATCH_NAME,
    JHEADER.NAME HEADER_NAME,
    PER.END_DATE
    FROM XLA_DISTRIBUTION_LINKS DLINK
    , GL_IMPORT_REFERENCES GLIMPREF
    , XLA_AE_LINES AELINE
    , GL_JE_HEADERS JHEADER
    , GL_JE_BATCHES JBATCH
    , GL_LEDGERS T
    , GL_PERIODS PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
    ( 'AP_INV_DIST', 'AP_PMT_DIST'
    , 'AP_PREPAY')
    AND DLINK.APPLICATION_ID = 200
    AND AELINE.APPLICATION_ID = 200
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID = DLINK.AE_HEADER_ID
    AND AELINE.AE_LINE_NUM = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID = T.LEDGER_ID
    AND JHEADER.STATUS = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
    TO_DATE('04/02/2007 00:00:00'
    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('Y', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT
    DLINK.SOURCE_DISTRIBUTION_ID_NUM_1 DISTRIBUTION_ID,
    DLINK.SOURCE_DISTRIBUTION_TYPE SOURCE_TABLE,
    DLINK.ACCOUNTING_LINE_CODE LINE_CODE,
    AELINE.ACCOUNTING_CLASS_CODE,
    GLIMPREF.JE_HEADER_ID JE_HEADER_ID,
    GLIMPREF.JE_LINE_NUM JE_LINE_NUM,
    AELINE.AE_HEADER_ID AE_HEADER_ID,
    AELINE.AE_LINE_NUM AE_LINE_NUM,
    T.LEDGER_ID LEDGER_ID,
    T.LEDGER_CATEGORY_CODE LEDGER_TYPE,
    JBATCH.NAME BATCH_NAME,
    JHEADER.NAME HEADER_NAME,
    PER.END_DATE
    FROM XLA_DISTRIBUTION_LINKS DLINK
    , GL_IMPORT_REFERENCES GLIMPREF
    , XLA_AE_LINES AELINE
    , GL_JE_HEADERS JHEADER
    , GL_JE_BATCHES JBATCH
    , GL_LEDGERS T
    , GL_PERIODS PER
    WHERE DLINK.SOURCE_DISTRIBUTION_TYPE IN
    ( 'AP_INV_DIST', 'AP_PMT_DIST'
    , 'AP_PREPAY')
    AND DLINK.APPLICATION_ID = 200
    AND AELINE.APPLICATION_ID = 200
    AND AELINE.GL_SL_LINK_TABLE = GLIMPREF.GL_SL_LINK_TABLE
    AND AELINE.GL_SL_LINK_ID = GLIMPREF.GL_SL_LINK_ID
    AND AELINE.AE_HEADER_ID = DLINK.AE_HEADER_ID
    AND AELINE.AE_LINE_NUM = DLINK.AE_LINE_NUM
    AND GLIMPREF.JE_HEADER_ID = JHEADER.JE_HEADER_ID
    AND JHEADER.JE_BATCH_ID = JBATCH.JE_BATCH_ID
    AND JHEADER.LEDGER_ID = T.LEDGER_ID
    AND JHEADER.STATUS = 'P'
    AND T.PERIOD_SET_NAME = PER.PERIOD_SET_NAME
    AND JHEADER.PERIOD_NAME = PER.PERIOD_NAME
    AND JHEADER.CREATION_DATE >=
    TO_DATE('04/02/2007 00:00:00'
    , 'MM/DD/YYYY HH24:MI:SS' )
    AND DECODE('Y', 'Y', T.LEDGER_ID, 1) IN (1)
    AND DECODE('N', 'Y', T.LEDGER_CATEGORY_CODE, 'NONE') IN ('NONE')
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Thu Feb 12 12:49:34 2009]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GL_LINKAGE_INFORMATION_GS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Thu Feb 12 12:49:34 2009
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GL_LINKAGE_INFORMATION_GS (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_XLA_AE_LINES] has completed: Total Run Time = [0.673295] secs, Total Idle Time = [0.000000] secs, Busy Percentage = [100.000000].
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_XLA_AE_LINES] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GL_LINKAGE_INFORMATION_GS] has completed. The total run time was insufficient for any meaningful statistics.
    MANAGER> PETL_24005 Starting post-session tasks. : (Thu Feb 12 12:49:35 2009)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Thu Feb 12 12:49:35 2009)
    MAPPING> TM_6018 Session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full] run completed with [0] row transformation errors.
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_XLA_AE_LINES] (Instance Name: [SQ_XLA_AE_LINES])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GL_LINKAGE_INFORMATION_GS] (Instance Name: [W_GL_LINKAGE_INFORMATION_GS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_ORA_GL_AP_LinkageInformation_Extract_Full] completed at [Thu Feb 12 12:49:36 2009]
    Thanks in Advance,
    Prashanth
    Edited by: user10719430 on Feb 11, 2009 7:33 AM
    Edited by: user10719430 on Feb 12, 2009 11:31 AM

    Need to increase temp tablespace.

  • Error while running ebs/forms load test scripts

    Hi, I´m new using the oracle application testing suite tools, i need to know the reasons of the following errors generated when running load test scripts in some of the virtual users:
    "ThinBrowserException", "VariableNotFoundException".
    how can i solved that?

    Tell me How many VUsers are given?

  • Error in Transaction Data - Full Load

    Hello All,
        This is the current scenario that I am working on:
    There is a process chain which has two transaction data load (FULL LOADS) processes to the same cube.In the process monitor everything seems okay (data loads seem fine) but overall status for both loads failed due to 'Error in source system/extractor' and it says 'error in data selection'.
    Processing is set to data targets only.
    On doing a manage on the cube, I found 3 old requests that were red and NOT set to QM status red. So I set them to QM status red and Deleted them and the difference I saw was that the subsequent requests became available for Reporting.
    Now this data load which is a full load takes for ever - I dont even know why I do not see a initialize delta update option there - can Anyone tell me why I dont see that.
    And, coming to the main question, how do I get the process chain completed - will I have to repeat the data loads or what options do I have to have a succesfully running process chain or at least these 2 full loads of transaction data.
    Thank you - points will be assigned for helpful answers
    - DB
    Edited by: Darshana on Jun 6, 2008 12:01 AM
    Edited by: Darshana on Jun 6, 2008 12:05 AM

    One interesting discovery I just found in R/3, was this job log with respect to the above process chain:
    it says that the job was cancelled in R/3 because the material ledger currencies were changed.
    the process chain is for inventory management and the data load process that get cancelled are for  the job gets cancelled in the source system:
    1. Material Valuation: period ending inventories
    2. Material Valuation: prices
    The performance assistant says this but I am not sure how far can I work on the R/3 side to rectify this:
    Material ledger currencies were changed
    Diagnosis
    The currencies currently set for the material ledger and the currency types set for valuation area 6205 differ from those set at conversion of the data (production startup).
    System Response
    The system does not allow you to post transactions after changing the currency settings to ensure consistency.
    Procedure
    Replace the current settings with the those entered at production
    start-up.
    If you wish to change the currency settings, you must use programs to convert data from the old to the new currencies.
    Inform your system administrator
    Anyone knowledgable in this area please give your inputs.
    - DB

  • Error while activating DSO after loading from another DSO..Production issue

    Hello Gurus,
    I got an error while activating DSO2 after loading from another DSO1.
    DSO1 load and activation was sucessful but while loading from DSO1 to DSO2 load was fine but got error while activating...
      >Error when assigning SID: Action VAL_SID_CONVERT table ZWEEK_NUM
      >Value u201809u2019 of characteristic ZWEEK_NUM is not a number with 000004 spaces
    this theard was posted after searching for related post in SDN.
    and i tried Function Module RSODS_CHECK_RSODSACTUPDTYPE but it is not available .. but I used this in BW 3.10u2026..
    is their any similar Function Module in BI 7.0 please advice.
    We are using SAP NetWeaver BI 7.0
    Release- 700    Service level- 017
    Thanks in advance
    Sandy

    sloved the issue......
    it was due to wrong data for the related field...
    it  must be "0009 "
    Thanks for ur support. closing the ticket
    Regards
    Sandy

  • Error while running AD User Target Recon

    Hi,
    We are getting the below error while running AD User Target Recon:
    [2012-09-04T10:07:32.262-04:00] [oim_server2] [NOTIFICATION] [] [oracle.iam.features.scheduler.agentry.operations] [tid: [ACTIVE].ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: xelsysadm] [ecid: 98fe4d9aa175090d:7101298f:13991840195:-8000-0000000000000235,0] [APP: oim#11.1.2.0.0] ADP ClassLoader failed to load: Script1[[
    java.lang.ClassNotFoundException: ADP ClassLoader failed to load: Script1
    at com.thortech.xl.dataobj.tcADPClassLoader.findClass(tcADPClassLoader.java:229)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
    at java.lang.Class.forName0(Native Method)
    at java.lang.Class.forName(Class.java:247)
    at oracle.iam.scheduler.vo.ClassLoaderObjectInputStream.resolveClass(ClassLoaderObjectInputStream.java:72)
    at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1574)
    at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1495)
    at java.io.ObjectInputStream.readClass(ObjectInputStream.java:1461)
    at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1311)
    at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1946)
    at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1870)
    at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1752)
    at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1328)
    at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1946)
    at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1870)
    at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1752)
    at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1328)
    at java.io.ObjectInputStream.readObject(ObjectInputStream.java:350)
    at oracle.iam.scheduler.vo.JobHistory.getExceptionObject(JobHistory.java:78)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.prepare(LookupActor.java:1282)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.refresh(LookupActor.java:3074)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.perform(LookupActor.java:2495)
    at oracle.iam.consoles.faces.mvc.canonic.Model.perform(Model.java:579)
    at oracle.iam.consoles.faces.mvc.admin.Model.perform(Model.java:326)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doPerform(Controller.java:257)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doSelectAction(Controller.java:179)
    at oracle.iam.consoles.faces.event.NavigationListener.processAction(NavigationListener.java:99)
    at javax.faces.event.ActionEvent.processListener(ActionEvent.java:88)
    at org.apache.myfaces.trinidad.component.UIXComponentBase.broadcast(UIXComponentBase.java:675)
    at org.apache.myfaces.trinidad.component.UIXCommand.broadcast(UIXCommand.java:179)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.run(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._processPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.broadcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclude.java:102)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.run(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._processPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.broadcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclude.java:96)
    at javax.faces.component.UIViewRoot.broadcastEvents(UIViewRoot.java:475)
    at javax.faces.component.UIViewRoot.processApplication(UIViewRoot.java:756)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl._invokeApplication(LifecycleImpl.java:889)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl._executePhase(LifecycleImpl.java:379)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl.execute(LifecycleImpl.java:194)
    at javax.faces.webapp.FacesServlet.service(FacesServlet.java:265)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adf.model.servlet.ADFBindingFilter.doFilter(ADFBindingFilter.java:205)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adfinternal.view.faces.webapp.rich.RegistrationFilter.doFilter(RegistrationFilter.java:106)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterListChain.doFilter(TrinidadFilterImpl.java:446)
    at oracle.adfinternal.view.faces.activedata.AdsFilter.doFilter(AdsFilter.java:60)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterListChain.doFilter(TrinidadFilterImpl.java:446)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl._doFilterImpl(TrinidadFilterImpl.java:271)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl.doFilter(TrinidadFilterImpl.java:177)
    at org.apache.myfaces.trinidad.webapp.TrinidadFilter.doFilter(TrinidadFilter.java:92)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.iam.platform.auth.web.OIMAuthContextFilter.doFilter(OIMAuthContextFilter.java:107)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adf.library.webapp.LibraryFilter.doFilter(LibraryFilter.java:179)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
    at java.security.AccessController.doPrivileged(Native Method)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:315)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:442)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:139)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3730)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3696)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)
    [2012-09-04T10:07:32.314-04:00] [oim_server2] [NOTIFICATION] [] [oracle.iam.features.scheduler.agentry.operations] [tid: [ACTIVE].ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: xelsysadm] [ecid: 98fe4d9aa175090d:7101298f:13991840195:-8000-0000000000000235,0] [APP: oim#11.1.2.0.0] [[
    java.lang.NullPointerException
    at java.io.ByteArrayInputStream.<init>(ByteArrayInputStream.java:89)
    at oracle.iam.scheduler.vo.JobHistory.getExceptionObject(JobHistory.java:76)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.prepare(LookupActor.java:1282)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.refresh(LookupActor.java:3074)
    at oracle.iam.features.scheduler.agentry.operations.LookupActor.perform(LookupActor.java:2495)
    at oracle.iam.consoles.faces.mvc.canonic.Model.perform(Model.java:579)
    at oracle.iam.consoles.faces.mvc.admin.Model.perform(Model.java:326)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doPerform(Controller.java:257)
    at oracle.iam.consoles.faces.mvc.canonic.Controller.doSelectAction(Controller.java:179)
    at oracle.iam.consoles.faces.event.NavigationListener.processAction(NavigationListener.java:99)
    at javax.faces.event.ActionEvent.processListener(ActionEvent.java:88)
    at org.apache.myfaces.trinidad.component.UIXComponentBase.broadcast(UIXComponentBase.java:675)
    at org.apache.myfaces.trinidad.component.UIXCommand.broadcast(UIXCommand.java:179)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.run(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._processPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.broadcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclude.java:102)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent$1.run(ContextSwitchingComponent.java:92)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent._processPhase(ContextSwitchingComponent.java:361)
    at oracle.adf.view.rich.component.fragment.ContextSwitchingComponent.broadcast(ContextSwitchingComponent.java:96)
    at oracle.adf.view.rich.component.fragment.UIXInclude.broadcast(UIXInclude.java:96)
    at javax.faces.component.UIViewRoot.broadcastEvents(UIViewRoot.java:475)
    at javax.faces.component.UIViewRoot.processApplication(UIViewRoot.java:756)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl._invokeApplication(LifecycleImpl.java:889)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl._executePhase(LifecycleImpl.java:379)
    at oracle.adfinternal.view.faces.lifecycle.LifecycleImpl.execute(LifecycleImpl.java:194)
    at javax.faces.webapp.FacesServlet.service(FacesServlet.java:265)
    at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adf.model.servlet.ADFBindingFilter.doFilter(ADFBindingFilter.java:205)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adfinternal.view.faces.webapp.rich.RegistrationFilter.doFilter(RegistrationFilter.java:106)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterListChain.doFilter(TrinidadFilterImpl.java:446)
    at oracle.adfinternal.view.faces.activedata.AdsFilter.doFilter(AdsFilter.java:60)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl$FilterListChain.doFilter(TrinidadFilterImpl.java:446)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl._doFilterImpl(TrinidadFilterImpl.java:271)
    at org.apache.myfaces.trinidadinternal.webapp.TrinidadFilterImpl.doFilter(TrinidadFilterImpl.java:177)
    at org.apache.myfaces.trinidad.webapp.TrinidadFilter.doFilter(TrinidadFilter.java:92)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.iam.platform.auth.web.OIMAuthContextFilter.doFilter(OIMAuthContextFilter.java:107)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.adf.library.webapp.LibraryFilter.doFilter(LibraryFilter.java:179)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:119)
    at java.security.AccessController.doPrivileged(Native Method)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:315)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:442)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:139)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3730)
    at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3696)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)
    Have anyone faced this issue ?? Any idea how to resolve this ??
    Thanks,
    Hrushi

    920194 wrote:
    Hi Bikash,
    Thank you that helped a bit!! I see the same behaviour i.e., recon events are created when using displayName or givenName attribute and it does not work using "samAccountName". Were you able to find the solution for this ??The solution would have to be provided by Oracle as this looks like a bug, since the connector doc gives example of filter as samAccountName.
    Also, when we reconcile using displayName/givenName attribute, even though the recon event is created, the status if the scheduler is Failed with "oracle.iam.connectors.icfcommon.exceptions.OIMException: Thor.API.Exceptions.tcAPIException: Row index out of bounds" Any idea on this ??Full stacktrace??
    Thanks,
    Hrushi

  • Error while running web application through JDEV (10.1.3.0.3) in OC4J

    Error while running web application through JDEV (10.1.3.0.3) in OC4J.
    Here is the error message.
    07/10/02 14:45:28 Exception in thread "OC4J Launcher" oracle.classloader.util.AnnotatedNoClassDefFoundError:
         Missing class: javax.xml.bind.JAXBContext
         Dependent class: com.oracle.corba.ee.impl.orb.config.InternalSettingsORBConfigImpl
         Loader: oc4j:10.1.3
         Code-Source: /C:/jdev/j2ee/home/lib/oc4j-internal.jar
         Configuration: <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar
    The missing class is not available from any code-source or loader in the server.
    07/10/02 14:45:28      at oracle.classloader.PolicyClassLoader.handleClassNotFound (PolicyClassLoader.java:2073) [C:/jdev/j2ee/home/lib/pcl.jar (from system property java.class.path), by sun.misc.Launcher$AppClassLoader@7]
         at oracle.classloader.PolicyClassLoader.internalLoadClass (PolicyClassLoader.java:1681) [C:/jdev/j2ee/home/lib/pcl.jar (from system property java.class.path), by sun.misc.Launcher$AppClassLoader@7]
         at oracle.classloader.PolicyClassLoader.loadClass (PolicyClassLoader.java:1633) [C:/jdev/j2ee/home/lib/pcl.jar (from system property java.class.path), by sun.misc.Launcher$AppClassLoader@7]
         at oracle.classloader.PolicyClassLoader.loadClass (PolicyClassLoader.java:1618) [C:/jdev/j2ee/home/lib/pcl.jar (from system property java.class.path), by sun.misc.Launcher$AppClassLoader@7]
         at java.lang.ClassLoader.loadClassInternal (ClassLoader.java:319) [jre bootstrap, by jre.bootstrap]
         at com.oracle.corba.ee.impl.orb.config.InternalSettingsORBConfigImpl.init (InternalSettingsORBConfigImpl.java:46) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.oracle.corba.ee.impl.orb.config.SunRIORBConfigImpl.init (SunRIORBConfigImpl.java:97) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.oracle.iiop.server.IIOPServerExtensionProvider.configureOrb (IIOPServerExtensionProvider.java:26) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.oracle.corba.ee.impl.orb.ORBServerExtensionProviderImpl.preInitApplicationServer (ORBServerExtensionProviderImpl.java:45) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.evermind.server.ApplicationServer.serverExtensionPreInit (ApplicationServer.java:1031) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.evermind.server.ApplicationServer.setConfig (ApplicationServer.java:861) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at com.evermind.server.ApplicationServerLauncher.run (ApplicationServerLauncher.java:98) [C:/jdev/j2ee/home/lib/oc4j-internal.jar (from <code-source> in boot.xml in C:\jdev\j2ee\home\oc4j.jar), by oc4j:10.1.3]
         at java.lang.Thread.run (Thread.java:595) [jre bootstrap, by jre.bootstrap]

    Hi,
    The guide you were refering was pointing to 10.1.2 wizards.
    For the latest 10.1.3 tutorial, please follow the below tutorial link :
    http://www.oracle.com/technology/products/jdev/101/tutorials/WS/WSandAScontrol.htm
    Hope this helps,
    Sunil..

  • VSTS loadtest error while running Loadtest with Runsettings as No of test iterations , ERROR is : could not read result repository: unknown transaction id in results:

    Subject: VSTS loadtest error while running Loadtest with Runsettings as  No of test iterations ,  ERROR is :  could not read result repository: unknown transaction
    id in results:
    Hello All,
    I am facing the following error while running a loadtest (when i choose the scenario to be executed for No of test iterations). But the same loadtest with the Run duration for X minutes is running fine without any issues. 
    Any suggestions on the issue i am facing
    error is :
    could not read result repository: unknown transaction id in results:
    or
    could not read result repository: unknown request id in results: 
    Thanks
    Satish

    Hi Kotapati,
    >>Subject: VSTS loadtest error while running Loadtest with Runsettings as  No of test iterations ,  ERROR is :  could not read result repository: unknown transaction
    id in results:
    According to your description, you mean that when you run your load test using the Test Iterations way and then get the error message, am I right?
    If yes, please you try to close the VS, and then open this VS again. Then clean your load test project/solution and then re-build it, run your load test again check this issue again.
    In addition, I suggest you could try to create a new load test and then run it using Test Iterations way check if you get the issue for the specified load test or all load tests.
    If possible, I suggest you could share me simple load test for us so that we will further help you.
    You could upload your load test to OneDrive and then copy link here.
    Best Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Error while running the tutorial posted by Steve Muench

    Hi Guys,
    I have been trying to run the tutorial published by Steve Muench on October 9 2006. I've followed all the steps as specified in the tutorial.
    Chapter 3.4:Run the Application, should supposed to compile and run my application using stand-alone embedded OC4J. It was compiled witout any error. But gave me the following error while running the application.
    [Starting OC4J using the following ports: HTTP=8988, RMI=23891, JMS=9227.]
    C:\oracledevday\jdeveloper\jdev\system\oracle.j2ee.10.1.3.39.14\embedded-oc4j\config>
    C:\oracledevday\jdeveloper\jdk\bin\javaw.exe -client -classpath C:\oracledevday\jdeveloper\j2ee\home\oc4j.jar;C:\oracledevday\jdeveloper\jdev\lib\jdev-oc4j-embedded.jar -Xverify:none -DcheckForUpdates=adminClientOnly -Doracle.application.environment=development -Doracle.j2ee.dont.use.memory.archive=true -Doracle.j2ee.http.socket.timeout=500 -Doc4j.jms.usePersistenceLockFiles=false oracle.oc4j.loader.boot.BootStrap -config C:\oracledevday\jdeveloper\jdev\system\oracle.j2ee.10.1.3.39.14\embedded-oc4j\config\server.xml
    [waiting for the server to complete its initialization...]
    2006-10-12 22:47:17.250 NOTIFICATION JMSServer[]: OC4J JMS server recovering transactions (commit 0) (rollback 0) (prepared 0).
    2006-10-12 22:47:17.265 NOTIFICATION JMSServer[]: OC4J JMS server recovering local transactions Queue[jms/Oc4jJmsExceptionQueue].
    2006-10-12 22:47:22.609 ERROR J2EE HTTP-00004 Internal error raised tyring to instantiate web-application: webapp defined in web site OC4J 10g (10.1.3) Default Web Site. Application: datatags does not exist. Error creating Web application: webapp
    Ready message received from Oc4jNotifier.
    Embedded OC4J startup time: 14015 ms.
    Target URL -- http://10.10.10.10:8988/MyDemo/faces/pages/EmployeesTable.jspx
    06/10/12 22:47:22 Oracle Containers for J2EE 10g (10.1.3.1.0) initialized
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Short,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Short)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Byte,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Byte)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Integer,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Integer)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Long,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Long)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Float,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Float)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Double,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(null,java.lang.Double)
    22:47:26 WARN (Digester) -[ValidatorRule]{faces-config/validator} Merge(javax.faces.LongRange)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.DateTime,null)
    22:47:26 WARN (Digester) -[ConverterRule]{faces-config/converter} Merge(javax.faces.Number,null)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(*)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(*)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/EmployeesTable.jspx)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/DepartmentsTable.jspx)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/LocationsTable.jspx)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/CountriesTable.jspx)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/RegionsTable.jspx)
    22:47:27 WARN (Digester) -[NavigationRuleRule]{faces-config/navigation-rule} Merge(/pages/JobsTable.jspx)
    2006-10-12 22:47:29.546 TRACE Setting JAZN Config property ...
    2006-10-12 22:47:29.625 NOTIFICATION ---- JAZNSecurityContext.getUserPrincipal(): NULL
    2006-10-12 22:47:32.015 NOTIFICATION ADF Faces is running with time-stamp checking enabled. This should not be used in a production environment. See the oracle.adf.view.faces.CHECK_FILE_MODIFICATION property in WEB-INF/web.xml
    22:47:32 DEBUG (JhsPageLifecycle) -Executing prepareModel, page=/pages/EmployeesTable.jspx, pagedef=EmployeesPageDef
    22:47:32 DEBUG (JhsNavigationHandlerImpl) -Executing checkRoles
    22:47:32 DEBUG (BreadcrumbStack) -Adding breadcrumb to stack: "Employees" (/pages/EmployeesTable.jspx)
    22:47:32 DEBUG (JhsPageLifecycle) -Executing prepareRender, page=/pages/EmployeesTable.jspx, pagedef=EmployeesPageDef
    22:47:32 DEBUG (BreadcrumbStack) -Breadcrumb already on stack; rolling back the stack
    22:47:32 DEBUG (BreadcrumbStack) -Adding breadcrumb to stack: "Employees" (/pages/EmployeesTable.jspx)
    2006-10-12 22:47:34.468 ERROR
    Please advise and let me know if you need any other information from me.
    Thanks,
    Magesh.

    Thanks Steven.
    That makes sense. Is there anything I can do to make this demo work without de-installing 10.1.3.1 Preview? Or If I also install 10.1.3.0.4 with SU5 on another folder, does that affect my 10.1.3.1 Preview version?
    I'm using 10.1.3.1 Preview version to learn Oracle's SOA suite.
    Thanks again,
    Magesh.

  • Error while running the process chain

    Hi, I am getting following error while running the process chain.
    "Lock NOT set for: Deleting the data completely from a data target"
    Please suggest.
    amit

    hi,
    This is due to a lock on the objects which is being used by ur load.
    Even if master data used by this would be in a lock by an attribute change run. wait till that attribute change run finishes and then repeat.
    If this target itself is locked by someother process like rollup , compression , this error would appear. In this case also repeat the step after those locking process finish.
    some general locks during loading:
    Sometimes parallel processing in DTP might lead to this, its better set the DTP and infopackage in serial processing. In this case change the settings and then repeat the process.
    There is a specific case in which the Serial DTP gets locked by itself, in this case repeat it untill it succeeds.
    You can identify the above scenario as below. This is applicable only if the DTPs processing mode is serial.
    if u see in SM 37 this load would have three DTP processes running.
    it should have only two.
    In this case cancel the process and repeat it untill u get it right.
    hope this helps u
    regard,
    ranjith
    Edited by: Ranjith  Kumar Murugesan on Sep 12, 2008 10:44 AM
    Edited by: Ranjith  Kumar Murugesan on Sep 12, 2008 10:46 AM

  • Error while running  BDB tcl testsuite commands

    Hi,
    I build berkeley db (4.7.25) libraries with tcl test suite is enabled in the solaris 9 machine using below configuration
    ../dist/configure --enable-dynamic ${BUILD_TAG} enable-tcl with-tcl=/export/tcl8.5/lib enable-test enable-java -prefix ${BUILD_DIR}
    I used tcl 8.5, jdk1.6.0_18 and gcc-3.4.6 for building libraries.Libraries are build successfully.
    When I run tcl command "run_std" in the tcl prompt, test is completed with below error logs in ALL.OUT file.
    ==========================================================================
    cat ALL.OUT | grep FAIL
    FAIL: error message in .out file: ./TESTDIR/lock003.1.out:FAIL:14:53:06 (00:00:01) number of other locks on 5: expected 0, got 1
    FAIL: error message in .out file: ./TESTDIR/lock003.1.out:FAIL:14:53:06 (00:00:00) ../dist/../test/lockscript.tcl run: FAIL:14:53:06 (00:00:01) number of other locks on 5: expected 0, got 1: pid 28332: expected 0, got 1
    FAIL: error message in .out file: ./TESTDIR/lock003.1.out:FAIL:18:24:50 (00:00:00) number of other locks on 44: expected 0, got 1
    FAIL: error message in .out file: ./TESTDIR/lock003.1.out:FAIL:18:24:50 (00:00:00) ../dist/../test/lockscript.tcl run: FAIL:18:24:50 (00:00:00) number of other locks on 44: expected 0, got 1: pid 28378: expected 0, got 1
    FAIL:03:09:45 (00:00:42) db_dump/db_load(./TESTDIR/primary007.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:09:45 (00:00:00) Dump/load of ./TESTDIR/primary007.db failed.
    FAIL:03:09:45 (00:00:00) db_dump/db_load(./TESTDIR/secondary007.0.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:09:45 (00:00:00) Dump/load of ./TESTDIR/secondary007.0.db failed.
    FAIL:03:09:45 (00:00:00) db_dump/db_load(./TESTDIR/secondary007.1.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:09:45 (00:00:00) Dump/load of ./TESTDIR/secondary007.1.db failed.
    FAIL:03:10:31 (00:00:46) db_dump/db_load(./TESTDIR/primary008.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:10:31 (00:00:00) Dump/load of ./TESTDIR/primary008.db failed.
    FAIL:03:10:31 (00:00:00) db_dump/db_load(./TESTDIR/secondary008.0.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:10:31 (00:00:00) Dump/load of ./TESTDIR/secondary008.0.db failed.
    FAIL:03:10:31 (00:00:00) db_dump/db_load(./TESTDIR/secondary008.1.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:10:31 (00:00:00) Dump/load of ./TESTDIR/secondary008.1.db failed.
    FAIL:03:12:06 (00:01:35) db_dump/db_load(./TESTDIR/primary001.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:06 (00:00:00) Dump/load of ./TESTDIR/primary001.db failed.
    FAIL:03:12:06 (00:00:00) db_dump/db_load(./TESTDIR/secondary001.0.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:06 (00:00:00) Dump/load of ./TESTDIR/secondary001.0.db failed.
    FAIL:03:12:06 (00:00:00) db_dump/db_load(./TESTDIR/secondary001.1.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:06 (00:00:00) Dump/load of ./TESTDIR/secondary001.1.db failed.
    FAIL:03:12:34 (00:00:28) db_dump/db_load(./TESTDIR/primary002.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:34 (00:00:00) Dump/load of ./TESTDIR/primary002.db failed.
    FAIL:03:12:34 (00:00:00) db_dump/db_load(./TESTDIR/secondary002.0.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:34 (00:00:00) Dump/load of ./TESTDIR/secondary002.0.db failed.
    FAIL:03:12:34 (00:00:00) db_dump/db_load(./TESTDIR/secondary002.1.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:34 (00:00:00) Dump/load of ./TESTDIR/secondary002.1.db failed.
    FAIL:03:12:47 (00:00:00) db_dump/db_load(./TESTDIR/primary006.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:47 (00:00:00) Dump/load of ./TESTDIR/primary006.db failed.
    FAIL:03:12:47 (00:00:00) db_dump/db_load(./TESTDIR/secondary006.1.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:47 (00:00:00) Dump/load of ./TESTDIR/secondary006.1.db failed.
    FAIL:03:12:47 (00:00:00) db_dump/db_load(./TESTDIR/secondary006.2.db:couldn't create pipe: too many open files): expected 0, got 1
    FAIL:03:12:47 (00:00:00) Dump/load of ./TESTDIR/secondary006.2.db failed.
    FAIL:03:13:10 (00:00:23) r: sindex: db open:too many open files
    FAIL:06:38:47 (02:19:40) rep_process_message returned DB_NOTFOUND: No matching key/data pair found
    ====================================================================================
    What could be the reason for the above errors while running the tcl testsuite ?
    Is there any problem in berkeley db libraries that are build with above configuration?
    Could you please clarify.
    Thanks in advance.
    Regards,
    Madhu

    Hi Sandra,
    I have executed lock003 in the tclsh prompt and error occurred while executing this lock test.
    ========================================================
    [exec] Lock003: Multi-process random lock test
    [exec] Lock003.a: Create environment
    [exec] Script watcher process 11571 launching lockscript.tcl process 11574.
    [exec] Script watcher process 11570 launching lockscript.tcl process 11573.
    [exec] Script watcher process 11572 launching lockscript.tcl process 11575.
    [exec] Script watcher process 11576 launching lockscript.tcl process 11578.
    [exec] Lock003.b: 5 independent processes now running
    [exec] Script watcher process 11577 launching lockscript.tcl process 11597.
    [exec] 06:35:39 (00:00:00) processes running: 11570 11571 11574 11572 11573 11575 11576 11578 11577 11597
    [exec] 06:36:15 (00:00:36) processes running: 11570 11571 11574 11572 11573 11575 11576 11578 11577 11597
    [exec] 06:36:51 (00:00:36) processes running: 11570 11571 11574 11572 11573 11575 11576 11578 11577 11597
    [exec] 06:37:27 (00:00:36) processes running: 11570 11571 11574 11572 11573 11575 11576 11578 11577 11597
    [exec] 06:38:03 (00:00:36) processes running: 11570 11571 11574 11572 11573 11575 11576 11578 11577 11597
    [exec] 06:38:38 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:39:12 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:39:48 (00:00:36) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:40:22 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:40:57 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:41:32 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:42:07 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 06:42:41 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:48:46 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:49:20 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:49:55 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:50:29 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:51:04 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:51:39 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:52:13 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:52:48 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:53:23 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:53:58 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:54:32 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:55:07 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:55:41 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:56:16 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:56:51 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:57:25 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:58:00 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:58:34 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:59:09 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 09:59:46 (00:00:37) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:00:21 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:00:56 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:01:30 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:02:05 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:02:39 (00:00:34) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] 10:03:14 (00:00:35) processes running: 11571 11574 11572 11575 11576 11578 11577 11597
    [exec] All processes have exited.
    [exec] FAIL: error message in .out file: ./TESTDIR/lock003.0.out:FAIL:06:38:00 (00:00:00) number of other locks on 52: expected 0, got 1
    [exec] FAIL: error message in .out file: ./TESTDIR/lock003.0.out:FAIL:06:38:00 (00:00:00) ../dist/../test/lockscript.tcl run: FAIL:06:38:00 (00:00:00) number of other locks on 52: expected 0, got 1: pid 11573: expected 0, got 1
    ===================================================================
    Is this problem occur if berkeley db jars having any issues ?Why because, when berkeleydb jars (build without tcl test flags in the same environment) are used in application under high load,application is hanged because of berkeley db native api.
    Sorry,I could not change ulimit value in my system because of lack of permissions.
    Can you please explain what could be the reason for the above error and how to rectify it.
    Thanks
    Madhu

  • Error while running SSIS package from Integration service or via Job

    Hi All,
    I encounter the below error while running SSIS Package from Job or integration service. However on execution completes success fully while running from data tools. The issue occurs after migration to 2012 from 2oo5 dtsx. PFB the error.
    SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on OLE DB Source returned error code 0xC02020C4.  The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by
    the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    A buffer failed while allocating 10485760 bytes.
    The system reports 26 percent memory load. There are 206110765056 bytes of physical memory with 150503776256 bytes.
    free. There are 2147352576 bytes of virtual memory with 185106432 bytes free. The paging file has 208256339968 bytes with 145642921984 bytes free.
    The package also runs successfully from other servers. This happens only in one of our server.seems like issue with some sql configuration.

    Hi ,
    Are you running using SQL Agent Job and Data tools on same server or different?
    If it is executing fine using Data tools and failing with Job it might be User credentials issue.Try
    to run Job with your credentials by using proxy .
    Regards,
    Prathy
    Prathy K

Maybe you are looking for