Errors in Data Profiling job

Anyone seen anything like this before ?
INFORMATIONAL Aggregation and Data Type analysis for object FEEDER_CNT_LUT
Error Error loading object: FEEDER_CNT_LUT
Error ORA-00942: table or view does not exist
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 672
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 974
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 1796
ORA-06512: at line 1
INFORMATIONAL Aggregation and Data Type analysis for object FEEDER_EQUIPMENT_LIST
INFORMATIONAL Data rule analysis for object TRACK_CMT complete in 0.016 s.
INFORMATIONAL Referential analysis initialization.
Error ORA-06502: PL/SQL: numeric or value error
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 1677
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 1706
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 1264
ORA-06512: at line 1
INFORMATIONAL Referential analysis for objects.
Error ORA-01722: invalid number
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 2018
ORA-06512: at "OWB.WB_RT_PROFILE_SERVICE", line 1327
ORA-06512: at line 1
INFORMATIONAL Data profiling operations complete.
INFORMATIONAL Functional dependency and unique key discovery for object STREETNAMES
...

Have you profiled this data before? I looks from the error messages as if data rules are being applied to the tables and those rules are failing due to datatype mismatches?No, this is the first time that we have run profiling.
The first error implies the table/view is no longer present in your schema. The error implying the table/view is no longer there is odd because the tables were not manually deleted and in fact I checked after the failure and all the tables are still there...
Are you creating a completely new data profile module? Yes, this is a completely new data profile module.
Are your re-running the profiling having made changes to the data rules or profiling options? No, this is the first run.

Similar Messages

  • ERROR: Invalid data pump job state; one of: DEFINING or EXECUTING.

    Hi all.
    I'm using Oracle 10g (10.2.0.3) I had some scheduled jobs for export the DB with datapump. Some days ago they began to fail with the message:
    ERROR: Invalid data pump job state; one of: DEFINING or EXECUTING.
    Wait a few moments and check the log file (if one was used) for more details.
    The EM says that I can't monitor the job because there's no master tables. How can I find the log?
    I also founded in the bdump that exceptions occurred:
    *** 2008-08-22 09:04:54.671
    *** ACTION NAME:(EXPORT_ITATRD_DIARIO) 2008-08-22 09:04:54.671
    *** MODULE NAME:(Data Pump Master) 2008-08-22 09:04:54.671
    *** SERVICE NAME:(SYS$USERS) 2008-08-22 09:04:54.671
    *** SESSION ID:(132.3726) 2008-08-22 09:04:54.671
    kswsdlaqsub: unexpected exception err=604, err2=24010
    kswsdlaqsub: unexpected exception err=604, err2=24010
    * kjdrpkey2hv: called with pkey 114178, options x8
    * kjdrpkey2hv: called with pkey 114176, options x8
    * kjdrpkey2hv: called with pkey 114177, options x8
    Some help would be appreciated.
    Thanks.

    Delete any uncompleted or failed jobs and try again.

  • Error - No date profile assigned to the transaction/item category

    Dear Expterts,
    I am getting the error while trying to create a complaint (In-house repair) transaction.
    No date profile assigned to the transaction/item category
    Message no. CRM_BILLING015
    System Response
    Billing dates cannot be maintained.
    Procedure
    Assign a date profile to the transaction type or item category in Customizing.
    Make sure that this date profile contains all the necessary date types:
    BILL_DATE
    INVCR_DATE
    date profiles are assigned to both transaction type and Item category. mentioned date rules are in place in standard configuration itself. still the problem persist.
    Please help me to resolve the issue. Your help will be highly appreciated.
    Best regards
    Raghu ram

    Hi Pepe,
    Thanks for the response, these transaction types are already there. and 'Today' date rules are assigned to both date types.
    Still the problem exist.
    Best regards,
    Raghu ram

  • Background job finished but Error sending data

    Dear all,
    report is running in my prod server . once this report completed sapuser get mail . when i check the job status it shows
    Job started
    Step 001 started (program GP4CF9U420G825I4P7DU8XY49WH, variant CUSTOMEROUTSTD, user ID S2K_BASIS)
    Spool request (number 0000012259) created without immediate output
    Error sending data (4)
    Job finished
    one more clarity i required
    this report should goes to 4 SAP users . how can i configure this report to 4 sapusers
    kindly suggest me

    > Step 001 started (program GP4CF9U420G825I4P7DU8XY49WH, variant CUSTOMEROUTSTD, user ID S2K_BASIS)
    > Spool request (number 0000012259) created without immediate output
    > Error sending data (4)
    I'd check SCOT/SOST for any errors. Also check the syslog (SM21).
    > this report should goes to 4 SAP users . how can i configure this report to 4 sapusers
    Create a distribution list (SAPoffice-Inbox - environment - distribution list) and enter that as addresse for the output.
    Markus

  • Data Profiling - Value too large for column error

    I am running a data profile which completes with errors. The error being reported is an ORA 12899 Value too large for column actual (41 maximum 40).
    I have checked the actual data in the table and the maximum is only 40 characters.
    Any ideas on how to solve this. Even though it completes no actual profile is done on the data due to the error.
    OWB version 11.2.0.1
    Log file below.
    Job     Rows Selected     Rows Inserted     Rows Updated     Rows Deleted     Errors     Warnings     Start Time     Elapsed Time     
    Profile_1306385940099                                   2011-05-26 14:59:00.0     106     
    Data profiling operations complete.                                             
    Redundant column analysis for objects complete in 0 s.                                              
    Redundant column analysis for objects.                                              
    Referential analysis for objects complete in 0.405 s.                                              
    Referential analysis for objects.                                              
    Referential analysis initialization complete in 8.128 s.                                             
    Referential analysis initialization.                                             
    Data rule analysis for object TABLE_NAME complete in 0 s.                                             
    Data rule analysis for object TABLE_NAME                                             
    Functional dependency and unique key discovery for object TABLE_NAME complete in 0 s.                                             
    Functional dependency and unique key discovery for object TABLE_NAME                                             
    Domain analysis for object TABLE_NAME complete in 0.858 s.                                             
    Domain analysis for object TABLE_NAME                                             
    Pattern analysis for object TABLE_NAME complete in 0.202 s.                                             
    Pattern analysis for object TABLE_NAME                                             
    Aggregation and Data Type analysis for object TABLE_NAME complete in 9.236 s.                                             
    Aggregation and Data Type analysis for object TABLE_NAME                                             
    Functional dependency and unique key discovery for object TABLE_NAME complete in 0 s.                                             
    Functional dependency and unique key discovery for object TABLE_NAME                                             
    Domain analysis for object TABLE_NAME complete in 0.842 s.                                             
    Domain analysis for object TABLE_NAME                                             
    Pattern analysis for object TABLE_NAME complete in 0.187 s.                                             
    Pattern analysis for object TABLE_NAME                                             
    Aggregation and Data Type analysis for object TABLE_NAME complete in 9.501 s.                                             
    Aggregation and Data Type analysis for object TABLE_NAME                                             
    Functional dependency and unique key discovery for object TABLE_NAME complete in 0 s.                                             
    Functional dependency and unique key discovery for object TABLE_NAME                                             
    Domain analysis for object TABLE_NAME complete in 0.717 s.                                             
    Domain analysis for object TABLE_NAME                                             
    Pattern analysis for object TABLE_NAME complete in 0.156 s.                                             
    Pattern analysis for object TABLE_NAME                                             
    Aggregation and Data Type analysis for object TABLE_NAME complete in 9.906 s.                                             
    Aggregation and Data Type analysis for object TABLE_NAME                                             
    Functional dependency and unique key discovery for object TABLE_NAME complete in 0 s.                                             
    Functional dependency and unique key discovery for object TABLE_NAME                                             
    Domain analysis for object TABLE_NAME complete in 0.827 s.                                             
    Domain analysis for object TABLE_NAME                                             
    Pattern analysis for object TABLE_NAME complete in 0.187 s.                                             
    Pattern analysis for object TABLE_NAME                                             
    Aggregation and Data Type analysis for object TABLE_NAME complete in 9.172 s.                                             
    Aggregation and Data Type analysis for object TABLE_NAME                                             
    Functional dependency and unique key discovery for object TABLE_NAME complete in 0 s.                                             
    Functional dependency and unique key discovery for object TABLE_NAME                                             
    Domain analysis for object TABLE_NAME complete in 0.889 s.                                             
    Domain analysis for object TABLE_NAME                                             
    Pattern analysis for object TABLE_NAME complete in 0.202 s.                                             
    Pattern analysis for object TABLE_NAME                                             
    Aggregation and Data Type analysis for object TABLE_NAME complete in 9.313 s.                                             
    Aggregation and Data Type analysis for object TABLE_NAME                                             
    Execute data prepare map for object TABLE_NAME complete in 9.267 s.                                             
    Execute data prepare map for object TABLE_NAME                                             
    Execute data prepare map for object TABLE_NAME complete in 10.187 s.                                             
    Execute data prepare map for object TABLE_NAME                                             
    Execute data prepare map for object TABLE_NAME complete in 8.019 s.                                             
    Execute data prepare map for object TABLE_NAME                                             
    Execute data prepare map for object TABLE_NAME complete in 5.507 s.                                             
    Execute data prepare map for object TABLE_NAME                                             
    Execute data prepare map for object TABLE_NAME complete in 10.857 s.                                             
    Execute data prepare map for object TABLE_NAME                                             
    Parameters                                             
    O82647310CF4D425C8AED9AAE_MAP_ProfileLoader                              1     2011-05-26 14:59:00.0     11     
    ORA-12899: value too large for column "SCHEMA"."O90239B0C1105447EB6495C903678"."ITEM_NAME_1" (actual: 41, maximum: 40)                                             
    Parameters                                             
    O68A16A57F2054A13B8761BDC_MAP_ProfileLoader                              1     2011-05-26 14:59:11.0     5     
    ORA-12899: value too large for column "SCHEMA"."O0D9332A164E649F3B4D05D045521"."ITEM_NAME_1" (actual: 41, maximum: 40)                                             
    Parameters                                             
    O78AD6B482FC44D8BB7AF8357_MAP_ProfileLoader                              1     2011-05-26 14:59:16.0     9     
    ORA-12899: value too large for column "SCHEMA"."OBF77A8BA8E6847B8AAE4522F98D6"."ITEM_NAME_2" (actual: 41, maximum: 40)                                             
    Parameters                                             
    OA79DF482D74847CF8EA05807_MAP_ProfileLoader                              1     2011-05-26 14:59:25.0     10     
    ORA-12899: value too large for column "SCHEMA"."OB0052CBCA5784DAD935F9FCF2E28"."ITEM_NAME_1" (actual: 41, maximum: 40)                                             
    Parameters                                             
    OFFE486BBDB884307B668F670_MAP_ProfileLoader                              1     2011-05-26 14:59:35.0     9     
    ORA-12899: value too large for column "SCHEMA"."O9943284818BB413E867F8DB57A5B"."ITEM_NAME_1" (actual: 42, maximum: 40)                                             
    Parameters

    Found the answer. It was the database character set for multi byte character sets.

  • ORA-39097: Data Pump job encountered unexpected error -12801

    Hallo!I am running Oracle RAC 11.2.0.3.0 database on IBM-AIX 7.1 OS platform.
    We normally do data pump expdp backups and we created a OS authenticated user and been having non-DBA users use this user (instead of / as sysdba which is only used by DBAs) to run expdp.This OS authenticated user has been working fine until it statrd gigin use error below
    Export: Release 11.2.0.3.0 - Production on Fri Apr 5 23:08:22 2013
    Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, Real Application Clusters, Automatic Storage Management, Oracle Label Security,
    OLAP, Data Mining, Oracle Database Vault and Real Application Testing optio
    FLASHBACK automatically enabled to preserve database integrity.
    Starting "OPS$COPBKPMG"."SYS_EXPORT_SCHEMA_16": /******** DIRECTORY=COPKBFUB_DIR dumpfile=COPKBFUB_Patch35_PreEOD_2013-04-05-23-08_%U.dmp logfile=COPKBFUB_Patch35_PreEOD_2013-04-05-23-08.log cluster=n parallel=4 schemas=BANKFUSION,CBS,UBINTERFACE,WASADMIN,CBSAUDIT,ACCTOPIC,BFBANKFUSION,PARTY,BFPARTY,WSREGISTRY,COPK
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 130.5 GB
    Processing object type SCHEMA_EXPORT/USER
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Processing object type SCHEMA_EXPORT/SYNONYM/SYNONYM
    Processing object type SCHEMA_EXPORT/DB_LINK
    Processing object type SCHEMA_EXPORT/SEQUENCE/SEQUENCE
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/TABLE/COMMENT
    Processing object type SCHEMA_EXPORT/PACKAGE/PACKAGE_SPEC
    Processing object type SCHEMA_EXPORT/PACKAGE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/FUNCTION/FUNCTION
    Processing object type SCHEMA_EXPORT/FUNCTION/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/PROCEDURE/PROCEDURE
    Processing object type SCHEMA_EXPORT/PROCEDURE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/PACKAGE/COMPILE_PACKAGE/PACKAGE_SPEC/ALTER_PACKAGE_SPEC
    Processing object type SCHEMA_EXPORT/FUNCTION/ALTER_FUNCTION
    Processing object type SCHEMA_EXPORT/PROCEDURE/ALTER_PROCEDURE
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/FUNCTIONAL_INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/FUNCTIONAL_INDEX/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/VIEW/VIEW
    Processing object type SCHEMA_EXPORT/VIEW/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/VIEW/GRANT/CROSS_SCHEMA/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/VIEW/COMMENT
    Processing object type SCHEMA_EXPORT/PACKAGE/PACKAGE_BODY
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Processing object type SCHEMA_EXPORT/MATERIALIZED_VIEW
    Processing object type SCHEMA_EXPORT/POST_SCHEMA/PROCACT_SCHEMA
    . . exported "WASADMIN"."BATCHGATEWAYLOGDETAIL" 2.244 GB 9379850 rows
    . . exported "WASADMIN"."UBTB_TRANSACTION" 13.71 GB 46299982 rows
    . . exported "WASADMIN"."INTERESTHISTORY" 2.094 GB 13479801 rows
    . . exported "WASADMIN"."MOVEMENTSHISTORY" 1.627 GB 13003451 rows
    . . exported "WASADMIN"."ACCRUALSREPORT" 1.455 GB 18765315 rows
    ORA-39097: Data Pump job encountered unexpected error -12801
    ORA-39065: unexpected master process exception in MAIN
    ORA-12801: error signaled in parallel query server PZ99, instance copubdb02dc:COPKBFUB2 (2)
    ORA-01460: unimplemented or unreasonable conversion requested
    Job "OPS$COPBKPMG"."SYS_EXPORT_SCHEMA_16" stopped due to fatal error at 23:13:37
    Please assist.

    have you seen this?
    *Bug 13099577 - ORA-1460 with parallel query [ID 13099577.8]*

  • Error ORA-14400 generated Data Profile

    Hello People,
    When I generated data profile with 5 tables:
    Profile_1224766956846
    INFORMATIONAL
    Data profiling operations complete.
    Profile_1224766956846
    Error
    ORA-14400: inserted partition key does not map to any partition
    ORA-06512: at "OWBREP.WB_RT_PROFILE_SERVICE", line 1113
    ORA-06512: at "OWBREP.WB_RT_PROFILE_SERVICE", line 1199
    ORA-06512: at line 1
    Profile_1224766956846
    INFORMATIONAL
    Referential analysis for objects.
    Profile_1224766956846
    INFORMATIONAL
    Referential analysis initialization complete in 53.978 s.
    Profile_1224766956846
    INFORMATIONAL
    Referential analysis initialization.
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object CONTRATO complete in 0.253 s.
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object CONTRATO
    Profile_1224766956846
    INFORMATIONAL
    Functional dependency and unique key discovery for object CONTRATO complete in 2.279 s.
    Profile_1224766956846
    INFORMATIONAL
    Functional dependency and unique key discovery for object CONTRATO
    Profile_1224766956846
    INFORMATIONAL
    Domain analysis for object CONTRATO complete in 0.208 s.
    Profile_1224766956846
    INFORMATIONAL
    Domain analysis for object CONTRATO
    Profile_1224766956846
    INFORMATIONAL
    Aggregation and Data Type analysis for object CONTRATO complete in 47.028 s.
    Profile_1224766956846
    INFORMATIONAL
    Aggregation and Data Type analysis for object CONTRATO
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object IMPUESTO complete in 0.007 s.
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object IMPUESTO
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object CUENTA complete in 0.746 s.
    Profile_1224766956846
    INFORMATIONAL
    Data rule analysis for object CUENTA
    Profile_1224766956846
    INFORMATIONAL
    Execute data prepare map for object CONTRATO complete in 1.747 s.
    Profile_1224766956846
    INFORMATIONAL
    Execute data prepare map for object CONTRATO
      O59DC772DA581112EE044001A_MAP_ProfileLoader
    Success
    It only happens when I try to do a foreing key between CONTRATO -> CUENTA. I have chosen data type discovery, common format discovery, copy data in workspace, pattern discovery, domain discovery, unique key discovery, functional dependency, relationship discovery and data rule profiling in the property inspector.
    Sometimes I am able to profile without any error, but sometimes I cannot. I am not able to identify the steps when the profiling works correctly.
    I am using 10g R2 database and OWB repository 10.2.0.4
    Thanks for your help!

    ORA-14400:     inserted partition key does not map to any partition
    Cause:     An attempt was made to insert a record into, a Range or Composite Range object, with a concatenated partition key that is beyond the concatenated partition bound list of the last partition -OR- An attempt was made to insert a record into a List object with a partition key that did not match the literal values specified for any of the partitions.
    Action:     Do not insert the key. Or, add a partition capable of accepting the key, Or add values matching the key to a partition specification
    Having a partition is not the issue. The table should have a partition with a high value (on the partition key) which is higher than the value you are trying to insert.
    If you are partitioning by date and the last partition has a high value of 01-aug-2008 you will get this error if one of the rows you are inserting has the date 02-aug-2008.
    also you might want to check:
    [http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:1392403659992]
    Edited by: Darthvader-647181 on Oct 23, 2008 7:41 AM

  • Oracle Data Profiling and Quality Microsoft Visual C++ Runtime Error

    Hi All,
    I have installed Oracle Data Profiling and Quality on a Windows 7 64-bit machine but every time I try to run the Profiler I get a C++ runtime error which when dismissed reveals an ODQ alert box that says a ReadProcessMemory or WriteProcessMemory did not complete.
    Has anyone encountered similar or does anyone have any ideas? I do not know where to start. Have tried a reinstall from a fresh download but get the same problem. The Metabase Server Admin tool works OK.
    Happy to supply any other info.
    If there's an ODQ forum please let me know and I will move.
    Thanks
    Dicksam

    Change the compatibility mode to windows XP service pack 2. I have this installed on a windows 2008 64bit server had to do that in order to get the metabase admin and the control center to work.

  • Data Profiler error DPF-0009

    Hi,
    Has anybody encountered this error in the Data Profiler before?
    DPF-0009: table_name data loading map could not be deployed because of compilation errors.
    ORA-06550: line 140, column 19:
    PL/SQL: ORA-00942 table or view does not exist
    I am connected as the repository administrator.
    Any help much appreciated.
    Regards,
    Brad
    Message was edited by:
    scrivb

    I had this when changing a password on database level only, instead of changing it with the OWB tools.
    (I wanted to change the Profiler-User password)
    Changed passwords again and it worked.

  • ORA-39097: Data Pump job encountered unexpected error -39076

    Hi Everyone,
    Today i tried to take a export dump pump(table specific) from my test database, version is 10.2.0.4 on Solaris10(64-bit) and i got the following error message,
    Job "SYSTEM"."SYS_EXPORT_TABLE_23" successfully completed at 09:51:36
    ORA-39097: Data Pump job encountered unexpected error -39076
    ORA-39065: unexpected master process exception in KUPV$FT_INT.DELETE_JOB
    ORA-39076: cannot delete job SYS_EXPORT_TABLE_23 for user SYSTEM
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 934
    ORA-31632: master table "SYSTEM.SYS_EXPORT_TABLE_23" not found, invalid, or inaccessible
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 1079
    ORA-20000: Unable to send e-mail message from pl/sql because of:
    ORA-29260: network error: Connect failed because target host or object does not exist
    ORA-39097: Data Pump job encountered unexpected error -39076
    ORA-39065: unexpected master process exception in MAIN
    ORA-39076: cannot delete job SYS_EXPORT_TABLE_23 for user SYSTEM
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 934
    ORA-31632: master table "SYSTEM.SYS_EXPORT_TABLE_23" not found, invalid, or inaccessible
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 1079
    ORA-20000: Unable to send e-mail message from pl/sql because of:
    ORA-29260: network error: Connect failed because target host or object does not exist
    i hope the export dumpfile is valid one but i don't know why i am getting this error message. Does any one have faced this kind of problem. please advice me
    Thanks
    Shan

    Once you see this:
    Job "SYSTEM"."SYS_EXPORT_TABLE_23" successfully completed at 09:51:36The Data Pump job is done with the dumpfile. There is some clean up that is needed and it looks like something in the cleanup failed. Not sure what it was, but you dumpfile should be fine. One easy way to test it is to run impdp with sqlfile. This will do everything import will do, but instead of creating objects, it writes the ddl to the sql file.
    impdp user/password sqlfile=my_test.sql directory=your_dir dupmfile=your_dump.dmp ...
    If that works, then your dumpfile should be fine. The last thing the export does is write the Data Pump master table to the dumpfile. The first thing that import does is read that table in. So, if you can read it in (which impdp sqlfile does) your dump is good.
    Dean

  • Error while scheduling Background Job for User/Role Full Synchronization

    Hi all,
    We have installed RAR 5.3 Component and uploaded the authorization data & established the connectors to the backend system.
    We have performed all the post installation activities and everything is complete.
    When we have scheduled User -Full Synchronization with the Back End system as  a part of Post Installation Activity we are receiving the below error message
    "Error while executing the Job:Cannot assign an empty string to host variable 2."
    Also the VIEW LOG/ Terminate Job buttons are disabled  in this screen.
    Can somebody please help us in resolving the above issue
    Thanks and Best Regards,
    Srihari.K

    Hi,
    We are copy pasting the error log (Part as it is huge) below here. We could able to do Full Synch for Roles and also for Profiles. Only for User Synch we are getting this error and none of the users are sychronized to RAR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user WILSONA of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user WINDC of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user WLADICHJ of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user WUK of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user ZENGS of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: Update user ZHENGL of HL2-QAHR
    Jan 13, 2009 12:34:27 AM com.virsa.cscext.dao.CSCDAO populateGenObjUser
    INFO: All System Flag:false=====Last Batch Flag:true
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.bg.BatchRiskAnalysis loadUserData
    INFO: @@@ User sync completed for params true: Syskey List is [HL2-QAHR]
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.bg.BgJob run
    WARNING: *** Job Exception: Cannot assign an empty string to host variable 2.
    com.sap.sql.log.OpenSQLException: Cannot assign an empty string to host variable 2.
         at com.sap.sql.log.Syslog.createAndLogOpenSQLException(Syslog.java:85)
         at com.sap.sql.log.Syslog.createAndLogOpenSQLException(Syslog.java:124)
         at com.sap.sql.types.VarcharResultColumn.setString(VarcharResultColumn.java:57)
         at com.sap.sql.jdbc.common.CommonPreparedStatement.setString(CommonPreparedStatement.java:511)
         at com.sap.engine.services.dbpool.wrappers.PreparedStatementWrapper.setString(PreparedStatementWrapper.java:355)
         at com.virsa.cscext.dao.CSCDAO.updateIgnoredUserData(CSCDAO.java:1388)
         at com.virsa.cscext.dao.CSCDAO.populateGenObjUser(CSCDAO.java:1169)
         at com.virsa.cc.xsys.bg.BatchRiskAnalysis.populateGenObj(BatchRiskAnalysis.java:868)
         at com.virsa.cc.xsys.bg.BatchRiskAnalysis.insertBAPIUserData(BatchRiskAnalysis.java:142)
         at com.virsa.cc.xsys.bg.BatchRiskAnalysis.loadUserData(BatchRiskAnalysis.java:390)
         at com.virsa.cc.xsys.bg.BatchRiskAnalysis.performBatchSyncAndAnalysis(BatchRiskAnalysis.java:1275)
         at com.virsa.cc.xsys.bg.BgJob.runJob(BgJob.java:402)
         at com.virsa.cc.xsys.bg.BgJob.run(BgJob.java:264)
         at com.virsa.cc.xsys.riskanalysis.AnalysisDaemonBgJob.scheduleJob(AnalysisDaemonBgJob.java:240)
         at com.virsa.cc.xsys.riskanalysis.AnalysisDaemonBgJob.start(AnalysisDaemonBgJob.java:80)
         at com.virsa.cc.comp.BgJobInvokerView.wdDoModifyView(BgJobInvokerView.java:436)
         at com.virsa.cc.comp.wdp.InternalBgJobInvokerView.wdDoModifyView(InternalBgJobInvokerView.java:1225)
         at com.sap.tc.webdynpro.progmodel.generation.DelegatingView.doModifyView(DelegatingView.java:78)
         at com.sap.tc.webdynpro.progmodel.view.View.modifyView(View.java:337)
         at com.sap.tc.webdynpro.clientserver.cal.ClientComponent.doModifyView(ClientComponent.java:481)
         at com.sap.tc.webdynpro.clientserver.window.WindowPhaseModel.doModifyView(WindowPhaseModel.java:551)
         at com.sap.tc.webdynpro.clientserver.window.WindowPhaseModel.processRequest(WindowPhaseModel.java:148)
         at com.sap.tc.webdynpro.clientserver.window.WebDynproWindow.processRequest(WebDynproWindow.java:335)
         at com.sap.tc.webdynpro.clientserver.cal.AbstractClient.executeTasks(AbstractClient.java:143)
         at com.sap.tc.webdynpro.clientserver.session.ApplicationSession.doProcessing(ApplicationSession.java:319)
         at com.sap.tc.webdynpro.clientserver.session.ClientSession.doApplicationProcessingStandalone(ClientSession.java:713)
         at com.sap.tc.webdynpro.clientserver.session.ClientSession.doApplicationProcessing(ClientSession.java:666)
         at com.sap.tc.webdynpro.clientserver.session.ClientSession.doProcessing(ClientSession.java:250)
         at com.sap.tc.webdynpro.clientserver.session.RequestManager.doProcessing(RequestManager.java:149)
         at com.sap.tc.webdynpro.serverimpl.defaultimpl.DispatcherServlet.doContent(DispatcherServlet.java:62)
         at com.sap.tc.webdynpro.serverimpl.defaultimpl.DispatcherServlet.doGet(DispatcherServlet.java:46)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:740)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.runServlet(HttpHandlerImpl.java:401)
         at com.sap.engine.services.servlets_jsp.server.HttpHandlerImpl.handleRequest(HttpHandlerImpl.java:266)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:386)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.startServlet(RequestAnalizer.java:364)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.invokeWebContainer(RequestAnalizer.java:1039)
         at com.sap.engine.services.httpserver.server.RequestAnalizer.handle(RequestAnalizer.java:265)
         at com.sap.engine.services.httpserver.server.Client.handle(Client.java:95)
         at com.sap.engine.services.httpserver.server.Processor.request(Processor.java:175)
         at com.sap.engine.core.service630.context.cluster.session.ApplicationSessionMessageListener.process(ApplicationSessionMessageListener.java:33)
         at com.sap.engine.core.cluster.impl6.session.MessageRunner.run(MessageRunner.java:41)
         at com.sap.engine.core.thread.impl3.ActionObject.run(ActionObject.java:37)
         at java.security.AccessController.doPrivileged(Native Method)
         at com.sap.engine.core.thread.impl3.SingleThread.execute(SingleThread.java:102)
         at com.sap.engine.core.thread.impl3.SingleThread.run(SingleThread.java:172)
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.bg.BgJob setStatus
    INFO: Job ID: 13 Status: Error
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.bg.BgJob updateJobHistory
    FINEST: --- @@@@@@@@@@@ Updating the Job History -
    2@@Msg is Error while executing the Job:Cannot assign an empty string to host variable 2.
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.bg.dao.BgJobHistoryDAO insert
    INFO: -
    Background Job History: job id=13, status=2, message=Error while executing the Job:Cannot assign an empty string to host variable 2.
    Jan 13, 2009 12:34:27 AM com.virsa.cc.xsys.riskanalysis.AnalysisDaemonBgJob scheduleJob
    INFO: -
    Complted Job =>13----
    Please let us know how to resolve this error
    Thanks and Best Regards,
    Srihari.K

  • Error while data loading

    Hi Gurus,
    I am getting error while data loading. At BI side when I  check the error it says Background Job Cancelled and when I check in R3 side I got the following error
    Job started
    Step 001 started (program SBIE0001, variant &0000000065503, user ID R3REMOTE)
    Asynchronous transmission of info IDoc 2 in task 0001 (0 parallel tasks)
    DATASOURCE = 2LIS_11_V_ITM
             Current Values for Selected Profile Parameters               *
    abap/heap_area_nondia......... 2000683008                              *
    abap/heap_area_total.......... 4000317440                              *
    abap/heaplimit................ 40894464                                *
    zcsa/installed_languages...... ED                                      *
    zcsa/system_language.......... E                                       *
    ztta/max_memreq_MB............ 2047                                    *
    ztta/roll_area................ 6500352                                 *
    ztta/roll_extension........... 3001024512                              *
    4 LUWs confirmed and 4 LUWs to be deleted with function module RSC2_QOUT_CONFIRM_DATA
    ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    Job cancelled
    Please help me out what should I do.
    Regards,
    Mayank

    Hi Mayank,
    The log says it went to short dump due to temp space issue.as its the source system job ,check in the source system side for temp table space and also check at the BI side as well.
    Check with your basis regarding the TEMP PSA table space - if its out of space ask them to increase the table space and try to repeat the load.
    Check the below note
    Note 796422 - DBIF_RSQL_SQL_ERROR during deletion of table BWFI_AEDAT
    Regards
    KP
    Edited by: prashanthk on Jul 19, 2010 10:42 AM

  • Errors while running Sync Jobs in GRC 10

    Dear Experts,
    I am trying to configure RAR in GRC 10. I have all required config steps and now I am trying to run Sync Jobs and I am getting below errors.
    Authorization Synch
    Program for Authorization data Synchronization
    Starting authorization sync for connector RC1 and language EN
    Error in RC1; Reason Error in RFC; 'Function module "/GRCPI/GRIA_AUTH_G
    PFCG authorization sync failed with errors
    Repository Object Synch
    Program for Repository Profile Syncronization
    Processing for connector RC1
    Error in RFC; 'Function module "/GRCPI/GRIA_PROF_GET_RANGES" not'
    Profile sync failed with errors
    Program for Repository Role Synchronization
    Processing for connector RC1
    Error in RFC; 'Function module "/GRCPI/GRIA_ROLE_GET_RANGES" not'
    Role sync failed with errors
    Program for Repository User Synchronization
    Processing for connector RC1
    Error in RFC; 'Function module "/GRCPI/GRIA_USR_GET_RANGES" not f'
    User sync failed with errors
    Repository Object sync job failed with errors
    Please check SLG1 for further details
    I appreciate your help.
    Thanks,
    Raj

    Hi Raj,
    I have a similar error for which I have put up a separate post, your input would be highly valuable.
    My RFC connection works via the test in SM59 but when executing GRAC_AUTH_SYNC I get a simiar error to your above.
    Please let me know what steps you took.
    Best regards,
    Paul

  • An unexpected error occurred while the job was running. (ID 104 Details: The parameter is incorrect (0x80070057))

    We are using DPM Version: 4.1.3465.0 with System Center 2012 Service Pack 1 to back up VMs in our Hyper V Cluster. We have a total of 37 VMs with a combination of 2008r2 servers and 2012 servers. One of our 2012 servers will not backup, and we consistently
    get the message "An unexpected error occurred while the job was running. (ID 104 Details: The parameter is incorrect (0x80070057))". I have seen this error message when I Googled it, but cannot get past it. I have deleted the Protection Group and
    data and tried to re-create, but get the same errors. I also thought the the server might be mis-reporting the data size, but I expanded that to cover the replication data. I have also tried to perform consistancy checks multiple times with no luck. I have
    not seen a difinitive answer from my Google searches, so was hopeful someone has seen this also, and have had success in resolving. Currently this server has no backup recovery points, so I am worried if it does crash. I think I have the rollups installed
    as well.

    Hi Kelly,
    does the issue still persists?

  • Error on Data Flow Task MSSQL 2012 Clustered "Description: The version of Lookup is not compatible with this version of the DataFlow. "

    We have an SSIS package that runs on clustered MSSQL 2012 Enterprise Nodes that is failing.  We use a job to executer the package.
    Environmental information:
    Product - Microsoft SQL Server Enterprise: Core-based Licensing (64-bit)
    Operating System - Microsoft Windows NT 6.1 (7601)
    Patform - NT x64
    Version - MSSQL Version 11.0.3349.0
    Package is set to 32 -bit.  All permissions verified.  Runs in lower environments, same MSSQL version.  All environments are clustered.  In the failing environment, all nodes are at the same service pack.  I have not verified if all
    nodes in the failing environment have SSIS installed.  Data access is installed.  We have other simpler packages that run in this environment, just not this one.  Time to ask the community for help!
    Error:
    Source: Data Flow Task - Data Flow Task (SSIS.Pipeline)     Description: The version of Lookup is not compatible with this version of the DataFlow.  End Error  Error:  Code: 0xC0048020    
    Description: Component "Conditional Split, clsid {7F88F654-4E20-4D14-84F4-AF9C925D3087}" could not be created and returned error code 0x80070005 "Access is denied.". Make sure that the component is registered correctly.  End Error 
    Description: The component is missing, not registered, not upgradeable, or missing required interfaces. The contact information for this component is "Conditional Split;Microsoft Corporation; Microsoft SQL Server; (C) Microsoft Corporation; All Rights
    Reserved; http://www.microsoft.com/sql/support;0".  End Error 
    (Left out shop specific information.  This is the first error in the errors returns by the job history for this package. )
    Thanks in advance.

    Hi DeveloperMax,
    According to your description, the error occurs when you execute the package with Agent job on clustered MSSQL 2012 Enterprise Nodes.
    As per my understanding, I think this issue can be caused by you use SQL Server Agent to schedule a SQL Server Integration Services package in a 64-bit environment. And the SSIS package is referencing some 32-Bit DLL or 32-Bit drivers which are available
    only in 32-bit versions, so the job failed.
    To fix this issue, we should use the 32-bit version of the DTExec.exe utility to schedule the 64-bit SQL Server Agent to run a package. To run a package in 32-bit mode from a 64-bit version of SQL Server Agent, we can go to the Job Step dialog box, then
    select “32 bit runtime” in the Advanced tab.
    Besides, we should make sure that SQL Server Integration Services is installed on the failing environment.
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

Maybe you are looking for

  • MDS persistence for column within a loop

    Hello, I am trying out basic setup of MDS. I was able to make it work for most of my columns except for a column within a loop. I also tried using the ADF Read Only Dynamic table and it is not persisting any of the columns of the table, since it is i

  • Direct input from a key board.

    Can you get java to read directly from the keyboard, like in C?

  • How to read web data in OBIEE

    Experts, need your guidnace in reading data on cloud / web withing OBIEE. Say for example, I want to read facebook application (searching for certain text string) in OBIEE and create reports against the search result. Please advise how to proceed on

  • Backup on T520i

    Using Backup and recovery tool: 1. I have tried to backup to a NAS. Connection parameters where accepted and backup starts, but after a few minutes a message pops up that backup was aborted. 2. First time I opened the Single Instance Storage window t

  • Where is iPhone crash log for Windows?

    Mac keeps them in ~/Library/Logs/CrashReporter/MobileDevice. What is the corresponding location for Windows? Does the user have to sync the iPhone to get the latest crash log for an iPhone app?