Error when using dbms_pumpdata to export data

       l_dp_handle   NUMBER; 
       l_dp_handle := 
 (operation => 'EXPORT', job_mode => 'TABLE'); 
        add_file (handle      => l_dp_handle 
                , filename    => 'scott_tb_emp.dmp' 
                , directory   => 'DB_DUMP_DIR' 
                , filetype    => DBMS_DATAPUMP.KU$_file_type_dump_file); 
        add_file (handle      => l_dp_handle 
                , filename    => 'scott_tb_emp.log' 
                , directory   => 'DB_DUMP_DIR' 
                , filetype    => DBMS_DATAPUMP.KU$_file_type_log_file); 
       DBMS_DATAPUMP.metadata_filter (handle        => l_dp_handle 
                                    , name          => 'NAME_EXPR' 
                                    , VALUE         => ' =''MYTABLE'' ' 
                                    , object_type   => 'TABLE'); 
     DBMS_DATAPUMP.data_filter( handle       => l_dp_handle 
                                , name         => 'SUBQUERY' 
                                , VALUE        => 'WHERE DT < ''11/01/2010''' 
                                , table_name   => 'MYTABLE' );                                 
       DBMS_DATAPUMP.start_job (l_dp_handle); 
       DBMS_DATAPUMP.detach (l_dp_handle); 
I want to export a subset of data in MYTABLE, it gave error
11:17:54  [DECLARE - 0 row(s), 0.000 secs]  [Error Code: 39001, SQL State: 99999]  ORA-39001: invalid argument value
ORA-06512: at "SYS.DBMS_SYS_ERROR", line 79
ORA-06512: at "SYS.DBMS_DATAPUMP", line 3444
ORA-06512: at "SYS.DBMS_DATAPUMP", line 3693
ORA-06512: at line 6
... 1 statement(s) executed, 0 row(s) affected, exec/fetch time: 0.000/0.000 sec  [0 successful, 0 warnings, 1 errors]

Check this test, i can get the successful dump file
       l_dp_handle   NUMBER;
       job_status VARCHAR2(30);
       l_dp_handle :=
 (operation => 'EXPORT', job_mode => 'TABLE');
        add_file (handle      => l_dp_handle
                , filename    => 'scott_tb_emp.dmp'
                , directory   => 'DATA_DD_DIR'
                , filetype    => DBMS_DATAPUMP.KU$_file_type_dump_file);
        add_file (handle      => l_dp_handle
                , filename    => 'scott_tb_emp.log'
                , directory   => 'DATA_DD_DIR'
                , filetype    => DBMS_DATAPUMP.KU$_file_type_log_file);
       DBMS_DATAPUMP.metadata_filter (handle        => l_dp_handle
                                    , name          => 'NAME_EXPR'
                                    , VALUE         => ' =''EMP'' '
                                    , object_type   => 'TABLE');
     DBMS_DATAPUMP.data_filter( handle       => l_dp_handle
                                , name         => 'SUBQUERY'
                                , VALUE        => 'WHERE SAL > 1000'
                                , table_name   => 'EMP' );                                
       DBMS_DATAPUMP.start_job (l_dp_handle);
       dbms_output.put_line ('DataPump Export - '||to_char(sysdate,'DD/MM/YYYY HH24:MI:SS')||' Status '||job_status);
       DBMS_DATAPUMP.detach (l_dp_handle);
   2    3    4    5    6    7    8    9   10   11   12   13   14   15   16   17   18   19   20   21   22   23   24   25   26   27   28   29   30   31 
32  /
PL/SQL procedure successfully completed.
SQL> !ls -lrt|tail
total 144
-rw-rw-r-- 1 oracle oracle    858 Sep 13 11:08 scott_tb_emp.log
-rw-rw---- 1 oracle oracle 110592 Sep 13 11:08 scott_tb_emp.dmp
SQL> !cat scott_tb_emp.log
Starting "ME"."SYS_EXPORT_TABLE_01": 
Estimate in progress using BLOCKS method...
Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
Total estimation using BLOCKS method: 64 KB
Processing object type TABLE_EXPORT/TABLE/TABLE
Processing object type TABLE_EXPORT/TABLE/AUDIT_OBJ
. . exported "ME"."EMP"                                  8.515 KB      13 rows
Master table "ME"."SYS_EXPORT_TABLE_01" successfully loaded/unloaded
Dump file set for ME.SYS_EXPORT_TABLE_01 is:
Job "ME"."SYS_EXPORT_TABLE_01" successfully completed at 11:08:47
SQL> select count(*) from emp where sal>1000;
Hope this help

Similar Messages

  • Teradata database error when using a Universe-defined Date condition

    I have encountered an odd problem when connecting Crystal to a Universe.  There is a pre-defined Date condition built into the Universe that is defined as follows:
    @Select(Date\Week End Date)  >  @Select(Conditions\User Response Select Date week -28)  AND @Select(Date\Week End Date)  <=  @Select(Conditions\User Response Select Date)
    The filter is designed to retrieve a rolling 4 weeks of data.
    When I try and create and refresh a Crystal Report that includes this Date prompt, I get the following error:
    Failed to retrieve data from the database.
    Details: 22003:(ODBC Teradata Driver)(Teradata Database) A character string failed conversion to a numeric value.
    If I create the same query in a WebI report it works fine, but in Crystal the error is produced.  I have also tried constructing the same prompt and filter within Crystal rather than use the pre-built prompt but that did not work either.
    Does anyone have any experience or advice on this error?  I can't seem to find many related postings on this specific error.
    BO Enterprise XI 3.1 SP2 running on Windows 2008 Server R2
    Crystal Reports 2008 SP2 running on Windows XP SP3
    Universe is built on a Teradata 12 source

    One item I found I had to do with Teradata was cast my date/time values to dates, otherwise I got weird errors like this. You say it's working in Web Intelligence, so this might not be the issue, but it is an easy thing to test.
    In the universe, change your end date definition to
    cast(table.end_date as date)
    and see if that helps.

  • Transfer(IDocs and TRFC) error when use infopackage to load data from R/3

    I create an infopackage to load data from R/3, when start immediately ,there is no data transfered. The return error is :
      transfer(IDocs and TRFC): Errors occurred
         request IDoc: Application document not posted
    Does anybody know how can correct this error?
    Thanks great.

    Hi Yimeng,
    Check in SM58 transaction code in R/3 and see if any LUWs are stuck there. If so you can execute them manually and complete the data load.

  • Getting an error when using adobe photoshop CC, "Could not move the data because of a program error". Any idea on what this means?

    Getting an error when using adobe photoshop CC, "Could not move the data because of a program error". Any idea on what this means?

    Have no idea of what you were even doing in Photoshop my crystal ball is out for repairs so I can not see.  

  • Oracle giving Block corruption errors when using CDC for sending the data to SQL Server 2012

    Hello Friends,
    We are facing an error while using CDC with Oracle. It is a "Block corruption" error, which indicates at some level of data corruption. We ran RMAN validate command to scan the database for corruption but it returned with no errors, however he
    Alert Log in Oracle is still coming up with the following error. Has anyone experienced this error when using Oracle Standard Edition and SQL 2012 ?
    Trace file e:\app\pulse-ad\diag\rdbms\orcl\orcl\trace\orcl_ora_5992.trc
    Oracle Database 11g Release - 64bit Production
    Windows Server 2003 Version V5.2 Service Pack 2
    CPU                 : 4 - type 8664, 4 Physical Cores
    Process Affinity    : 0x0000000000000000
    Memory (Avail/Total): Ph:6782M/24575M, Ph+PgF:12203M/30844M
    Instance name: orcl
    Redo thread mounted by this instance: 1
    Oracle process number: 151
    Windows thread id: 5992, image: ORACLE.EXE (SHAD)
    *** 2013-12-12 03:04:33.655
    *** SESSION ID:(1281.3832) 2013-12-12 03:04:33.655
    *** CLIENT ID:() 2013-12-12 03:04:33.655
    *** SERVICE NAME:(orcl) 2013-12-12 03:04:33.655
    *** MODULE NAME:(xdbcdcsvc.exe) 2013-12-12 03:04:33.655
    *** ACTION NAME:() 2013-12-12 03:04:33.655
    Lost-write detected for sequence 70856. The lost-write starts occurring in block 11193. The current block being validating is 12930.
    Block dump of the first lost-write block:
    Flag: 0x1 Format: 0x22 Block: 0x00002bb9 Seq: 0x000114bf Beg: 0x94 Cks:0x68ee
    Dump of memory from 0x0000000598D06C00 to 0x0000000598D06E00
    598D06C00 00002201 00002BB9 000114BF 68EE8094  [."...+.........h]
    598D06C10 00085BF1 0023BDA1 000DE19C 000DE19C  [.[....#.........]
    598D06C20 0000000C 00000000 2209160A 5BF10000  [..........."...[]
    598D06C30 3EB10502 00C0F5CA 0031BDA1 00010205  [...>......1.....]
    598D06C40 02B22C6A 038A6D69 00000001 00000000  [j,]
    598D06C50 4D554407 30373230 35BB0206 001100AE  [.DUM0270...5....]
    598D06C60 0001040A 000D000E 038A6D69 56B25735  []
    598D06C70 729C0003 E19C0001 000C0006 000D0006  [...r............]
    598D06C80 02BB0502 00C0F5CD 0023BDA1 000A0002  [..........#.....]
    598D06C90 00C00013 000000D0 00030201 56B25736  [............6W.V]
    598D06CA0 03890001 00000000 00000000 002E0105  [................]
    598D06CB0 FFFF0003 00C0F5CD 56B25736 3EB10003  [........6W.V...>]
    598D06CC0 FFFF0024 0014000C 000C0018 00120014  [$...............]
    598D06CD0 09CC0058 E75B0022 0009000F 00085BF1  [X...".[......[..]
    598D06CE0 0024BDA1 000DE19D 000DE19D 0000000C  [..$.............]
    598D06CF0 00000000 2309160A 5BF10000 3EB10502  [.......#...[...>]
    598D06D00 00C0F5CD 0020BDA1 00010205 02B22C72  [...... .....r,..]
    598D06D10 03900974 00000019 00000000 3030300A  [t............000]
    598D06D20 33303030 06323132 AE35BB02 0B441100  [0003212...5...D.]
    598D06D30 0001040A 000D000E 03900974 56B25736  [........t...6W.V]
    598D06D40 729C0003 E19D0011 000C0006 000D0006  [...r............]
    598D06D50 02BB0502 00C0F5CD 0024BDA1 00EA0002  [..........$.....]
    598D06D60 00270016 000001FC 00032C01 56B25736  [..'......,..6W.V]
    598D06D70 00000001 00000000 30393007 002E0105  [.........090....]
    598D06D80 FFFF0003 00C0F5CD 56B25736 00000003  [........6W.V....]
    598D06D90 FFFF0025 00140052 000C0018 00070035  [%...R.......5...]
    598D06DA0 0003000A 00070003 0001001D 00030001  [................]
    598D06DB0 00010001 00010001 00010001 00010001  [................]
    598D06DC0 00010001 00010001 00010001 00010001  [................]
    598D06DD0 00010001 00000001 00010001 00010001  [................]
    598D06DE0 00010001 00000014 09720174 00000022  [........t.r."...]
    598D06DF0 0009000F 00085BF1 0025BDA1 000DE19A  [.....[....%.....]
    Block dump of the current block being validating:
    Flag: 0x1 Format: 0x22 Block: 0x00003282 Seq: 0x000114c8 Beg: 0x0 Cks:0x312a
    Dump of memory from 0x0000000598DDFE00 to 0x0000000598DE0000
    598DDFE00 00002201 00003282 000114C8 312A8000  [."...2........*1]
    598DDFE10 50424703 31303607 34353335 69745319  [.GBP.6015354.Sti]
    598DDFE20 6E696C72 72502067 6375646F 4C207374  [rling Products L]
    598DDFE30 4E206474 C3025650 0380013D 0457454E  [td NPV..=...NEW.]
    598DDFE40 4E1E09C2 1E09C204 10C2024E 1E09C204  [...N....N.......]
    598DDFE50 09C2044E C2024E1E 03C30510 021B0929  [N....N......)...]
    598DDFE60 C3053DC3 0F192602 2602C305 C3050F19  [.=...&.....&....]
    598DDFE70 0C1A6203 5102C105 C2041F4E 044E1E09  [.b.....QN.....N.]
    598DDFE80 4E1E09C2 0410C202 4E1E09C2 1E09C204  [...N.......N....]
    598DDFE90 10C2024E 2903C305 78071B09 011D0B71  [N......)...xq...]
    598DDFEA0 BF020101 1FBF0215 4E018001 53014E01  [...........N.N.S]
    598DDFEB0 0723002C 0B0C7178 0A3C3C18 30303030  [,.#.xq...<<.0000]
    598DDFEC0 33373030 4D033337 47034255 36075042  [007373.MUB.GBP.6]
    598DDFED0 38333936 4E113331 2065776B 74616C50  [693813.Nkwe Plat]
    598DDFEE0 6D756E69 56504E20 0B0AC303 4E038001  [inum NPV.......N]
    598DDFEF0 C2045745 0459512E 59512EC2 5253C203  [EW...QY...QY..SR]
    598DDFF00 512EC204 2EC20459 C2035951 C3055253  [...QY...QY..SR..]
    598DDFF10 1B092903 0B0AC303 3C04C305 C3053239  [.).........<92..]
    598DDFF20 32393C04 4F08C305 C105114F 1F4E5102  [.<92...OO....QN.]
    598DDFF30 512EC204 2EC20459 C2035951 C2045253  [...QY...QY..SR..]
    598DDFF40 0459512E 59512EC2 5253C203 2903C305  [.QY...QY..SR...)]
    598DDFF50 78071B09 01190A71 C0030101 C0034709  [...xq........G..]
    598DDFF60 8001330A 4E014E01 002C5301 71780723  [.3...N.N.S,.#.xq]
    598DDFF70 3C180B0C 30300A3C 30303030 33373337  [...<<.0000007373]
    598DDFF80 42554D03 50424703 31304207 344C5131  [.MUB.GBP.B011QL4]
    598DDFF90 6F725020 63657073 614A2074 206E6170  [ Prospect Japan ]
    598DDFFA0 646E7546 64724F20 44535520 30302E30  [Fund Ord USD0.00]
    598DDFFB0 04C30331 03800133 0557454E 5B1603C3  [1...3...NEW....[]
    598DDFFC0 03C30521 04215B16 1F4004C3 1603C305  [!....[!...@.....]
    598DDFFD0 C305215B 215B1603 4004C304 03C3051F  [[!....[!...@....]
    598DDFFE0 031B0929 043304C3 4D245AC2 245AC204  [).....3..Z$M..Z$]
    598DDFFF0 02C3054D 040A1A18 494002C1 1603C305  [M.........@I....]
    *** 2013-12-12 03:05:07.984
    ** LOGMINER WARNING - Invalidated 6 LCRs **
    Complete dump of first invalid START LCR follows:
    ++  LCR Dump Begin: 0x0000000532C004E0 - CANNOT_SUPPORT
         op: 255, Original op: 3, baseobjn: 0, objn: 233316, objv: 1
         DF: 0x00000002, DF2: 0x00000000, MF: 0x00000000, MF2: 0x00000000
         PF: 0x40000001, PF2: 0x00002000
         MergeFlag: 0x00, FilterFlag: 0x00
         Id: 0, iotPrimaryKeyCount: 3, numChgRec: 4
         NumCrSpilled: 0
         RedoThread#: 1, rba: 0x0114c8.0001c6ce.00d4
         scn: 0x0003.56b593be, xid: 0x0008.00c.00100d85, pxid: 0x0008.00c.00100d85
         ncol: 0newcount: 0, oldcount: 0
         LUBA: 0x3.c109c0.c.15.38f64

    Hi Dee,
    Thank you for your question.
    I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated.
    Thank you for your understanding and support.
    Mike Yin
    TechNet Community Support

  • Error when using filter on date column in interactive report

    I'm getting the following error when using the filter on a date column. None of the criteria in the filter list work. All give me the same error but when I use the filter on a column feature (in the search bar) and use ">" and "<" those work.
    Can someone please help me????
    " Settledate is in the last 2 years
    ORA-30175: invalid type given for an argument ORA-02063: preceding line from MAINDATA "

    I have a similar problem . I have an interactive report based on resultset stored in a collection; since all the columns of a collection are of string datatype, I convert the columns to appropriate datatypes before its returned to the user; everything works as desired except when I try to filter on a data column , I get a ORA-01858: a non-numeric character was found where a numeric was expected .
    I turned on debug mode and am copying the output here , column c009 which is aliased to 'C' is the one that I am trying to apply the filter on
    ROWID as apxws_row_pk,
    count(*) over () as apxws_row_cnt
    from (
    select * from (
    Select c001,c002,c003,c004,c005,c006,c007,c008,to_date(c009,'dd-MON-yy HH24.MI') checkouttimestamp,
    cast(to_date(c009,'dd-MON-yy HH24.MI') as timestamp) c,
    --TO_TIMESTAMP(c009,'dd-MON-yy HH24.MI') c1 ,
    trunc(to_date(c009,'dd-MON-yy HH24.MI')) checkoutdate,
    c010,c011,to_number(c012),to_number(c013), to_number(c014),c015,c016,c017,to_number(c018) from apex_collections where collection_name = 'IR_BOOKSALES'
    ) r
    where ("C" between systimestamp - ((1/24) :APXWS_EXPR_1) and systimestamp)*
    ) r where rownum <= to_number(:APXWS_MAX_ROW_CNT)
    order by ROWID
    0.04: IR binding: ":APXWS_EXPR_1"="APXWS_EXPR_1" value="1"
    0.04: IR binding: ":APXWS_MAX_ROW_CNT"="APXWS_MAX_ROW_CNT" value="2000"
    ORA-01858: a non-numeric character was found where a numeric was expected
    I have tried to convert c009 to date, timestamp and timestamp with time zone datatypes ; they all return the same error ,but all 3 datatypes work fine with systimestamp when I run them in sqlplus. Any help would be appreciated.

  • Error when using data postion delay

    I obtained this error when using data position delay. Can anyone please
    tell me the reason why I can not configure the delay to be between 0.7
    and 0.8.
    Possible reason(s):
    Driver Status:  (Hex 0xBFFA403F) Unable to configure requested delay property given the current clock rate.
    Ensure that the delay value specified is within the valid range for the
    frequency specified or do not configure property for delay operation.
    Device: PXI1Slot2
    Property: Data Position
    Sample Rate: 40000000.0
    Requested Value: 0.71875
    Supported Values: 0.000 through 0.200, 0.300 through 0.700, 0.800 through 1.000,
    Status Code: -219401

    The 656x devices have a limit on the available data delay values based
    on frequency.  Between 25MHz and 50MHz there are "holes" in the
    available delay range.  Above 50MHz, the full 0 to 1 data delay is
    valid.  If you look in the specifications installed with NI HSDIO
    for the NI 656x, on page 7, Figure 1, there is an illustration showing
    how to compute what the valid range is for the frequency of interest.
    For example, at 40MHz, valid values are:
    0 to 0.2
    0.3 to 0.7
    0.8 to 1
    In the error message posted below, you tried to input a value of 0.71875 which is not within these ranges.

  • Error when using PPR in the OA page developed.

    We are encountering an error when using PPR for the custom OA page being developed. I am using JDeveloper version 1437) and Oracle Applications version 11.5.10.CU2. The scenario we are using PPR and the steps to reproduce the issue is mentioned below,
    1) PPR is enabled for a custom OA page that is developed.
    2) The PPR is being used to enable or disable a particular field based on the
    selection of a value in a poplist in the same OA Page.
    Error replication steps:
    1) The user enters a value in the primary mandatory field which has to have
    unique value.
    2) The user selects a value from the poplist of a field which has PPR associated.
    3) The user enters data in all other mandatory fields in the OA page except one
    and tries to save the form and is shown a error message saying that all
    mandatory fields are not entered(This is correct). Now the page gets
    automatically refreshed and so the entered values are wiped off.
    4) Now if the user again enters a value in the primary mandatory field (as in
    step1) and if that value happens to be already existing value and then select
    a value from the poplist which has PPR enabled then immediately the error
    "Please enter a unique value" is entered, though we did not save the record.
    The unique record validation would happen only at the time of saving the
    record as per out business logic.
    5) But if in Step 4 if i had entered a unique value again then the above said error
    does not happen. There are few more similar kind of issues happening in the
    same page because PPR is enabled.
    Please note that this issue happens only when the page is deployed in a oracle
    applications instance and not in JDeveloper.
    Would this issue because of some cacheing problem? If Yes then how do we resolve it? Or would it be some other problem caused by our internal coding.
    Any inputs for resolving this issue would be highly appreciated.
    Thanks, Raja.

    What works in Jdeveloper should work when deployed in iAS as well.
    And PPR is an old feature given quite enough time to mature, so before we suspect OA we should investigate the code. What I do not understand is how the page refreshes in Step 3 and why does the user have to re-enter the fields just because he missed entering one mandatory field the last time.
    Mandatory field validations are client validations set by making the fields 'Required'. You obviously have client validations turned off on your PPR, otherwise you would be stuck with a javascript popup on PPR. The only way you can get the "Please enter unique..." error on PPR is Server Validation. So where have you placed this validation? If it is not in the EO, is it in the AM? And are you unconditionally calling the AM validation method in your processFormRequest? The PPR will also go through processFormRequest so I hope the PPR event is in a different if-else block.

  • "Index: 0, Size: 0" error when using lexical references

    I got "Index: 0, Size: 0" when using lexical references in data template. Can any body tell me what happen? The detail error is:
    [042308_084608187][][EXCEPTION] java.lang.IndexOutOfBoundsException: Index: 0, S
    ize: 0
    at oracle.apps.xdo.dataengine.DataTemplateParser.getParentDataSource(Dat
    at oracle.apps.xdo.dataengine.XMLPGEN.writeDefaultGroup(
    at oracle.apps.xdo.dataengine.XMLPGEN.writeGroupStructure(
    at oracle.apps.xdo.dataengine.XMLPGEN.processData(
    at oracle.apps.xdo.dataengine.XMLPGEN.processXML(
    at oracle.apps.xdo.dataengine.XMLPGEN.writeXML(
    at oracle.apps.xdo.dataengine.DataProcessor.processDataStructre(DataProc
    at oracle.apps.xdo.dataengine.DataProcessor.processData(DataProcessor.ja
    at oracle.apps.xdo.servlet.ReportContextImplV11.getReportXMLData(ReportC
    at oracle.apps.xdo.servlet.CoreProcessor.process(

    Here is the data template, the lexical reference is &where_clause, thanks.
    <dataTemplate name="Tamplate" dataSourceRef="dmdb" defaultPackage="RRBY_PKG">
              <property name="debug_mode" value="on"/>
                   <parameter name="BUName" dataType="character"/>
              <parameter name="StartDate" dataType="date"/>
              <parameter name="EndDate" dataType="date"/>
    <dataTrigger name="beforeReport" source="RRBY_PKG.dynamic_where(BUName =&gt; :BUName)"/>
         <sqlstatement name="GroupByAgent">
      NVL(, 'Unassigned') UNITNAME, QUEUENAME,
      NVL2(, 'Unassigned', || ': ' || FULLQUEUENAME,
      COUNT(queueEvent.requestNumber) REQNUM
    FROM((((queueevent queueevent
    INNER JOIN assignment assignment ON queueevent.assignmentkey = assignment.assignmentkey)
    INNER JOIN datedim datedim ON queueevent.datekey = datedim.datekey)
    INNER JOIN businessunit businessunit ON queueevent.businessunitkey = businessunit.businessunitkey)
    INNER JOIN queue queue ON queueevent.forwardedqueuekey = queue.queuekey)
    INNER JOIN timedim timedim ON queueevent.timekey = timedim.timekey
    WHERE dateDim.FULLDATE >= TO_DATE(:StartDate, 'yyyy-mm-dd')
    AND BW_COMMON_DATEADD('s', timeDim.SECONDOFDAY, BW_COMMON_DATEADD('s', queueEvent.EVENTDURATION, dateDim.FULLDATE)) > TO_DATE(:StartDate, 'yyyy-mm-dd')
    AND BW_COMMON_DATEADD('s', timeDim.SECONDOFDAY, BW_COMMON_DATEADD('s', queueEvent.EVENTDURATION, dateDim.FULLDATE)) &lt; TO_DATE(:EndDate, 'yyyy-mm-dd') AND &where_clause
    group by,]]>

  • Errors when using source2wsdd

    Any help will be greatly appreciated.
    I got the following errors when using the ant task source2wsdd:
    [source2wsdd] classpath C:\bea\weblogic81\server\lib\weblogic.jar;C:\bea\weblogi
    [source2wsdd] source2wsdd: Cannot find doclet class
    [source2wsdd] 1 error
    [ant] Exiting C:\project\DCII\ant\batch-build.xml.
    C:\project\ant\build.xml:67: The following error occurred while executing this l
    C:\project\DCII\ant\batch-build.xml:326: javadoc execution failed
    Caused by: C:\project\DCII\ant\batch-build.xml:326: javadoc execution failed
    at weblogic.ant.taskdefs.webservices.autotype.JavaSource2DD.runJavadoc(J
    at weblogic.ant.taskdefs.webservices.autotype.JavaSource2DD.execute(Java
    at ise.antelope.tasks.Call.execute(
    at ise.antelope.tasks.IfTask.doIf(
    at ise.antelope.tasks.IfTask.execute(
    ... 10 more
    --- Nested Exception ---
    C:\project\DCII\ant\batch-build.xml:326: javadoc execution failed
    at weblogic.ant.taskdefs.webservices.autotype.JavaSource2DD.runJavadoc(J
    at weblogic.ant.taskdefs.webservices.autotype.JavaSource2DD.execute(Java
    at ise.antelope.tasks.Call.execute(
    at ise.antelope.tasks.IfTask.doIf(
    at ise.antelope.tasks.IfTask.execute(
    Total time: 8 seconds
    The following is the major parts from my ant build file:
    <path id="webservice.classpath">
    <pathelement location="${wls.home.dir}/server/lib/weblogic.jar"/>
    <pathelement location="${wls.home.dir}/server/lib/webservices.jar"/>
    <target name = "initialize">
    <mkdir dir="${build.dir}"/>
    <mkdir dir="${class.dir}"/>
    <mkdir dir="${dist.dir}/${deploy.env}"/>
    <mkdir dir="${jdocs.dir}"/>
    <mkdir dir="${weblogic.application.dir}"/>
    <taskdef name="call"
    <taskdef name="if"
    <taskdef name="servicegen"
    <taskdef name="clientgen"
    <taskdef name="source2wsdd"
    <taskdef name="wldeploy"
    <target name="buildEar"
    depends="buildCommonJar, ejbc">
    <echo message ="[${}] Checking to see if ${ear.file} is up to date...[${appc.classpath}]" />
    <classpath refid="webservice.classpath" />
    <echo message ="[Done with buildEar]" />

    Did you ever resolve this?

  • Error When Using Class :(

        import flash.display.MovieClip;
        public class Main extends MovieClip
            public var enemyA:Array = [];
            public var birdA:Array = [];
            public var rowNum:int = 2;
            private var gap:int = 100;
            private var obj_no = 2;
            public var enemy1:mychar = new mychar();
            public var TheBird:BirdChar = new BirdChar();
            public function Main()
                // constructor code
            public function createEnemyF()
                for (var i:int = 0; i < rowNum; i++)
                    for (var j:int = 0; j < obj_no; j++)
                        enemy1.x = Math.random() * stage.stageWidth - enemy1.width;
                        enemy1.y = - i * (gap + enemy1.height) - 30.65;
            public function createBirdF() {
                TheBird.x = 270.95;
                TheBird.y = 350.95;
            public function loopF(event:Event) {
            public function updateEnemyPositionsF() {
                enemy1.y +=  2;
            public function updateBirdPositionsF() {
                TheBird.x = mouseX;
            public function hitTestF() {
                    trace('The Bird Hit Enemy 1');
    This conflicts and causes this error:
    1046: Type was not found or was not a compile-time constant: MouseEvent
    ^ ^ Code repeats to all of my event listeners
    Thanks for your time.

    Thanks, i thought it might be this because of previous problems but i seen i already had : import; so i thought that would be OK!
    Second Error:
    1180: Call to an undefined method Timer.
    I think this is the same sort of thing but what to import to fix this?
    So far these are my imports:
        import flash.display.MovieClip;
    Thanks for helping.
    Date: Thu, 3 Nov 2011 05:36:22 -0600
    From: [email protected]
    To: [email protected]
    Subject: Error When Using Class
        Re: Error When Using Class
        created by markerline in Flash Pro - General - View the full discussion
    Looks like you imported events.Event but not events.MouseEvent (or some similar syntax) basically you must import MouseEvents separately from other Events.
         Replies to this message go to everyone subscribed to this thread, not directly to the person who posted the message. To post a reply, either reply to this email or visit the message page:
         To unsubscribe from this thread, please visit the message page at In the Actions box on the right, click the Stop Email Notifications link.
         Start a new discussion in Flash Pro - General by email or at Adobe Forums
      For more information about maintaining your forum email notifications please go to

  • Dbms_crypto - avoid error when using different key in lower environment

    Hello Experts,
    We are using Oracle We are planning to implement dbms_crypto to encrypt few columns. We clone the data from production to lower environment ( DEV, QC).
    For the lower environments, we do not want to get the sensitive data from production and do not plan to use same key. Rather than getting an error when using differnt key, is it possible to get a different resultset back.
    In other words, we want the implementation to be same across environments but want to use a diffent key in lower environment and get different result (or garbage).
    Any suggestions would be greatly appreciated.
    While testing this logic, I am getting following error when using differnt key to decrypt. It works fine if I use same key.
    Error at line 1
    ORA-28817: PL/SQL function returned an error.
    ORA-06512: at "SYS.DBMS_CRYPTO_FFI", line 67
    ORA-06512: at "SYS.DBMS_CRYPTO", line 44
    ORA-06512: at line 19
      l_credit_card_no    VARCHAR2(19) := '1234 5678 9012 3456';
      l_ccn_raw           RAW(128) := UTL_RAW.cast_to_raw(l_credit_card_no);
    l_key               RAW(128) := UTL_RAW.cast_to_raw('abcdefgh');
       l2_key               RAW(128) := UTL_RAW.cast_to_raw('12345678');
      l_encrypted_raw     RAW(2048);
      l_decrypted_raw     RAW(2048);
      DBMS_OUTPUT.put_line('Original  : ' || l_credit_card_no);
      l_encrypted_raw := DBMS_CRYPTO.encrypt(src => l_ccn_raw,
                                             typ => DBMS_CRYPTO.des_cbc_pkcs5,
                                             key => l_key);
      DBMS_OUTPUT.put_line('Encrypted : ' || RAWTOHEX(UTL_RAW.cast_to_raw(l_encrypted_raw)));
      l_decrypted_raw := DBMS_CRYPTO.decrypt(src => l_encrypted_raw,
                                             typ => DBMS_CRYPTO.des_cbc_pkcs5,
                                             key => l2_key); --**Using different key to decrypt
      DBMS_OUTPUT.put_line('Decrypted : ' || UTL_RAW.cast_to_varchar2(l_decrypted_raw));
    END;Thank you.

    If I understand what you are trying to do ... and I may not ... it is not going to work.
      2   l_credit_card_no VARCHAR2(19) := '1612-1791-1809-2605';
      3   l_ccn_raw RAW(128) := utl_raw.cast_to_raw(l_credit_card_no);
      4   l_key1     RAW(128) := utl_raw.cast_to_raw('abcdefgh');
      5   l_key2     RAW(128) := utl_raw.cast_to_raw('zyxwvuts');  -- alternate key used to attempt a different decryption
      7   l_encrypted_raw RAW(2048);
      8   l_decrypted_raw RAW(2048);
      9  BEGIN
    10    dbms_output.put_line('Original : ' || l_credit_card_no);
    12    l_encrypted_raw := dbms_crypto.encrypt(l_ccn_raw, dbms_crypto.des_cbc_pkcs5, l_key1);
    14    dbms_output.put_line('Encrypted : ' || RAWTOHEX(utl_raw.cast_to_raw(l_encrypted_raw)));
    16    l_decrypted_raw := dbms_crypto.decrypt(src => l_encrypted_raw, typ => dbms_crypto.des_cbc_pkc
    s5, key => l_key1);
    18    dbms_output.put_line('Key1 : ' || utl_raw.cast_to_varchar2(l_decrypted_raw));
    20    l_decrypted_raw := dbms_crypto.decrypt(src => l_encrypted_raw, typ => dbms_crypto.des_cbc_pkc
    s5, key => l_key2);
    22    dbms_output.put_line('Key2 : ' || utl_raw.cast_to_varchar2(l_decrypted_raw));
    23  END;
    24  /
    Original : 1612-1791-1809-2605
    Encrypted : 3534443342333642353141363846384237463732384636373943374630364234323243334539383042323135
    Key1 : 1612-1791-1809-2605
    ERROR at line 1:
    ORA-28817: PL/SQL function returned an error.
    ORA-06512: at "SYS.DBMS_CRYPTO_FFI", line 67
    ORA-06512: at "SYS.DBMS_CRYPTO", line 44
    ORA-06512: at line 20

  • Error when running package to load data from InfoProvider

    I encounter the following error when using the standard package to load data from a InfoProvider to BPC. I tried loading the data from both DSO and InfoCube, but I get the same error.
    The log message where the error occurs is as follows:
    Task name CLEAR CUBE DATA:
    Could not perform write (INHERITED ERROR)
    The "convert" step of the process chain completed successfully. This error is happening when the "clear cube according data" step of the process chain gets executed.
    Any ideas on why this is happening?

    Hi Laeral:
    You should be able to just open one VISA session then write and read both commands then close the session at the end. This sounds to me like an error that comes from trying to open two VISA sessions to the instrument at the same time. I have attached some very basic LabVIEW VISA code that writes and reads two different commands. Hopefully it will get you started.
    Emilie S.
    Applications Engineer
    National Instruments
    Basic VISA Example Two ‏39 KB

  • Error when using Oracle Migration WorkBench (Sql Server 2k to Oracle 8i)

    I'm trying to migrate a Sql Server 2k database to an Oracle 8i schema using OMWB. After using BCP to collect metadata information on the source database, I tried to capture the source data model and got a "Capture aborted" message on the progress screen (even though then it appears a dialog window reporting "0 errors and 0 warnings", funny!). Down there is the error log OMWB generated. After reading some posts, I changed some columns of the OMWB dictionary that holds the source database name, but still got the same error. If someone can enlighten me on this toppic, I'd greatly appreciated it.
    Here's the error log of OMWB:
    ** Oracle Migration Workbench
    ** Production
    ** ( Build 20050629 )
    ** OMWB_HOME: D:\Datos\Download\Oracle\Oracle Migration Workbench\omwb
    ** user language: es
    ** user region: null
    ** user timezone:
    ** file encoding: Cp1252
    ** java version: 1.4.2_04
    ** java vendor: Sun Microsystems Inc.
    ** o.s. arch: x86
    ** o.s. name: Windows 2000
    ** o.s. version: 5.0
    ** Classpath:
    ** Started : Fri Feb 23 17:18:48 GMT-03:00 2007
    ** Workbench Repository : Oracle8i Enterprise Edition Release - Production
    With the Partitioning option
    JServer Release - Production
    Repository Connection URL: jdbc:oracle:thin:@ServerCms:1521:sigcewbc
    ** The following plugins are installed:
    ** Microsoft Access 2.0/95/97/2000/2002/2003 Plugin, Production Release
    ** Microsoft SQLServer 2000 Plugin, Production Release
    ** Microsoft SQLServer 6.5 Plugin, Production Release
    ** Microsoft SQLServer 7.0 Plugin, Production Release
    ** Active Plugin : SQLServer2K
    EXCEPTION :SQLServer2KDisconnSourceModelLoad.loadSourceModel(): oracle.mtg.migration.MigrationStopException: java.lang.IndexOutOfBoundsException: Index: 15, Size: 15
    ** Shutdown : Fri Feb 23 18:00:50 GMT-03:00 2007

    Duplicate thread
    Error when using Oracle Migration WorkBench (Sql Server 2k to Oracle 8i)

  • Error when converting time stamp in date time time zone CST  in IC

    Hi Experts,
    We have recently upgraded from CRM 5.0 to CRM 2007. We are using IC WEB Employee interaction center.
    so when agent create service ticket in interaction center and close the ticket on same day we are getting an error saying Error when converting time stamp in date time time zone CST.
    we thought there may be problem while installing SAP CRM by basis team in time zone, but we did not find any problem.
    So please advice how to resolve this issue.

    Hi Teja,
    I saw a similar message, and it was resolved by adding a timezone to the user profile defaults tab in SU3. I believe the system has been configured to convert between users' timezone and the system timezone, therefore it needs to know each users' timezone.

Maybe you are looking for