Errors importing TTS using impdp

Hello All,
I am seeing following errors when importing TTS using impdp. Please let me know if you know any work around.
Thanks,
Sunny boy
Corp_DEODWPRD$ impdp parfile=impdp_tts.par
Import: Release 11.2.0.3.0 - Production on Mon Jul 30 13:23:50 2012
Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
With the Partitioning, Automatic Storage Management, OLAP, Data Mining
and Real Application Testing options
Master table "SYS"."SYS_IMPORT_TRANSPORTABLE_02" successfully loaded/unloaded
Starting "SYS"."SYS_IMPORT_TRANSPORTABLE_02": /******** AS SYSDBA parfile=impdp_tts.par
Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
Processing object type TRANSPORTABLE_EXPORT/TABLE
Processing object type TRANSPORTABLE_EXPORT/GRANT/OWNER_GRANT/OBJECT_GRANT
ORA-39126: Worker unexpected fatal error in KUPW$WORKER.LOAD_METADATA [SELECT process_order, flags, xml_clob, NVL(dump_fileid, :1), NVL(dump_position, :2), dump_length, dump_allocation, NVL(value_n, 0), grantor, object_row, object_schema, object_long_name, partition_name, subpartition_name, processing_status, processing_state, base_object_type, base_object_schema, base_object_name, base_process_order, property, size_estimate, in_progress, original_object_schema, original_object_name, creation_level, object_int_oid FROM "SYS"."SYS_IMPORT_TRANSPORTABLE_02" WHERE process_order between :3 AND :4 AND duplicate = 0 AND processing_state NOT IN (:5, :6, :7) ORDER BY process_order]
ORA-39183: internal error -19 ocurred during decompression phase 2
ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
ORA-06512: at "SYS.KUPW$WORKER", line 9001
----- PL/SQL Call Stack -----
object line object
handle number name
0x5d9562370 20462 package body SYS.KUPW$WORKER
0x5d9562370 9028 package body SYS.KUPW$WORKER
0x5d9562370 4185 package body SYS.KUPW$WORKER
0x5d9562370 9725 package body SYS.KUPW$WORKER
0x5d9562370 1775 package body SYS.KUPW$WORKER
0x59c4fbb90 2 anonymous block
Job "SYS"."SYS_IMPORT_TRANSPORTABLE_02" stopped due to fatal error at 13:52:57
PAR file:
USERID='/ as sysdba'
directory=ODW_MIG_DIR
dumpfile=exp_full_ODW_tts_20120719_01.dmp
logfile=imp_FULL_ODW_TTS_20120730.log
exclude=PROCACT_INSTANCE
TRANSPORT_DATAFILES=(
'+DEODW_DATA_GRP/deodwprd/data/ts_appsds_d_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/ts_appsds_i_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_aud_d_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_capone_d01_02.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_capone_d01_03.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_capone_d01_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_capone_d_p2012_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_capone_i01_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_ts_ccr_d01_3.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_ts_ccr_d01_2.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_ts_ccr_d01_1.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_cst_stage_d_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_cst_stage_i_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_dsdw_doc_stg_d01_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_x_etl_d01_02.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_x_report_work_d01_01.dbf',
'+DEODW_DATA_GRP/deodwprd/data/deodw_x_sas_d01_01.dbf')

Still the same error with system account. Working with oracle about this iissue.
Corp_DEODWPRD$ impdp parfile=impdp_tts.par
Import: Release 11.2.0.3.0 - Production on Tue Jul 31 11:47:32 2012
Copyright (c) 1982, 2011, Oracle and/or its affiliates. All rights reserved.
UDI-28002: operation generated ORACLE error 28002
ORA-28002: the password will expire within 6 days
Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
With the Partitioning, Automatic Storage Management, OLAP, Data Mining
and Real Application Testing options
Master table "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" successfully loaded/unloaded
Starting "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01": system/******** parfile=impdp_tts.par
Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
Processing object type TRANSPORTABLE_EXPORT/TABLE
Processing object type TRANSPORTABLE_EXPORT/GRANT/OWNER_GRANT/OBJECT_GRANT
ORA-39126: Worker unexpected fatal error in KUPW$WORKER.LOAD_METADATA [SELECT process_order, flags, xml_clob, NVL(dump_fileid, :1), NVL(dump_position, :2), dump_length, dump_allocation, NVL(value_n, 0), grantor, object_row, object_schema, object_long_name, partition_name, subpartition_name, processing_status, processing_state, base_object_type, base_object_schema, base_object_name, base_process_order, property, size_estimate, in_progress, original_object_schema, original_object_name, creation_level, object_int_oid FROM "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" WHERE  process_order between :3 AND :4 AND duplicate = 0 AND processing_state NOT IN (:5, :6, :7) ORDER BY process_order]
ORA-39183: internal error -19 ocurred during decompression phase 2
ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
ORA-06512: at "SYS.KUPW$WORKER", line 9001
----- PL/SQL Call Stack -----
object line object
handle number name
0x5dac10b98 20462 package body SYS.KUPW$WORKER
0x5dac10b98 9028 package body SYS.KUPW$WORKER
0x5dac10b98 4185 package body SYS.KUPW$WORKER
0x5dac10b98 9725 package body SYS.KUPW$WORKER
0x5dac10b98 1775 package body SYS.KUPW$WORKER
0x59c554178 2 anonymous block
Job "SYSTEM"."SYS_IMPORT_TRANSPORTABLE_01" stopped due to fatal error at 12:13:04
Edited by: Sunny boy on Jul 31, 2012 12:16 PM

Similar Messages

  • Importing tablespace using impdp

    Hi,
    I am trying to transport tablespace using expdp/impdp like:
    source side: expdp dumpfile=tbs1.dmp logfile=tbs1.log directory=datapump transport_tablespaces=tbs1
    transformed to target side
    target side: impdp dumpfile=tbs1.dmp directory=datapump transport_datafiles='/u01/app/oracle/sales/tbs1.dbf'
    it is giving:
    ORA-39123: Data Pump transportable tablespace job aborted
    ORA-29345: cannot plug a tablespace into a database using an incompatible character set
    when i searched in google some suggestions are given like do transport using rman.
    But when we have the option in impdp why i need to go to rman tool.
    i want to use impdp only. but i ma unable to use it properly while importing.
    can you please alter me.
    thank you!

    Hello,
    ORA-29345: cannot plug a tablespace into a database using an incompatible character setIt cannot work due to the error ORA-29345 above:
    ORA-29345: cannot plug a tablespace into a database using an incompatible character set
    Cause: Oracle does not support plugging a tablespace into a database using an
    incompatible character set.
    Action: Use import/export or unload/load to move data instead.
    i want to use impdp only. but i ma unable to use it properly while importing.So, as suggested, move the data by import/export. But, be careful, as you have different character set you may have conversions.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Import database using impdp

    Hello,
    I wanna import a database using impdp. Can i import while i'm logged with system/<password> ? On which database should i be logged? I have a single database, 'orcl'
    Thanks

    I wanna import a database using impdp.
    Can i import while i'm logged with system/<password> ?
    On which database should i be logged? I have a single database, 'orcl'You don't have to login to any database to perform the import using datapump import.
    You just have to set the oracle database environment where you want to import into.
    suppose, you have an export dump generated already,using datapump export(expdp) and if you want to import into orcl database,
    set the oracle environment to orcl
    You have to have the directory object created with read,write access to the user performing export/import(expdp or impdp)
    use impdp help=y at the command line to see the list of available options.
    http://download.oracle.com/docs/cd/B13789_01/server.101/b10825/dp_import.htm
    -Anantha

  • Unable to import database using impdp via network_link

    Dear All,
    I am trying to import database from remote location's dump to my local database after creating dblink from local db to remote db.
    At Remote Server:
    - Directory created and granted the privileges. Dir Name: PREMLIVE_EXPDP
    At Local server:
    Username: expuser with dba, imp and exp full database privileges.
    DB Link name: to_premiatest14
    I faced an error of directory invalid when I run the below command in local server to import schema from remote database.
    C:\>impdp expuser/expuser network_link=to_premiatest14 directory=PREMLIVE_EXPDP remap_schema=PREMIATEST:PREMIATEST15 logfile=2013-09-02_premiatest15.log
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39002: invalid operation
    ORA-39070: Unable to open the log file.
    ORA-39087: directory name PREMLIVE_EXPDP is invalid
    Any help or advice's would be really highly appreciated.
    Regards,
    Syed

    Some more details follows.
    Server 1: Remote Server
    BANNER
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bi
    PL/SQL Release 10.2.0.5.0 - Production
    CORE    10.2.0.5.0      Production
    TNS for 64-bit Windows: Version 10.2.0.5.0 - Production
    NLSRTL Version 10.2.0.5.0 - Production
    TNS-Entry:
    PREMIATEST14=
      (DESCRIPTION =
        (ADDRESS = (PROTOCOL = TCP)(HOST = XX.XX.XX.14)(PORT = 1521))
        (CONNECT_DATA =
          (SERVER = DEDICATED)
          (SERVICE_NAME = malath)
    DIRECTORY:
    PREMLIVE_EXPDP AS
    'C:\Dump\PREMLIVE_EXPDP'; ---->>>> This location/directory holds the dump file.
    Server 2: Local Server
    BANNER
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bi
    PL/SQL Release 10.2.0.5.0 - Production
    CORE    10.2.0.5.0      Production
    TNS for 64-bit Windows: Version 10.2.0.5.0 - Production
    NLSRTL Version 10.2.0.5.0 - Production
    TNS-Entry:
    PREMIATEST15=
      (DESCRIPTION=
        (ADDRESS=
          (PROTOCOL=TCP)
          (HOST=XX.XX.XX.15)
          (PORT=1521)
        (CONNECT_DATA=
          (SERVICE_NAME=PREMIA15)
    PREMIATEST14 =
      (DESCRIPTION =
        (ADDRESS = (PROTOCOL = TCP)(HOST = XX.XX.XX.14)(PORT = 1521))
        (CONNECT_DATA =
          (SERVER = DEDICATED)
          (SERVICE_NAME = malath)
    Network/DBLink:
    CREATE PUBLIC DATABASE LINK TO_PREMIATEST14 CONNECT TO EXPUSER IDENTIFIED BY <PWD> USING 'PREMIATEST14';
    DBLink tested :
    SQL> select count(*) from premiatest.brkdivion@to_premiatest14;
      COUNT(*)
            94
    Error on importing:
    C:\>impdp expuser/expuser network_link=to_premiatest14 directory=PREMLIVE_EXPDP
    remap_schema=PREMIATEST:PREMIATEST15 logfile=2013-09-02_premiatest15.log
    Import: Release 10.2.0.5.0 - 64bit Production on Tuesday, 03 September, 2013 1:2
    3:10
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bit
    Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39002: invalid operation
    ORA-39070: Unable to open the log file.
    ORA-39087: directory name PREMLIVE_EXPDP is invalid
    I don't know where am i going wrong . Valuable advice's/assistance will be highly appreciated.
    Thanks & Regards,
    Syed

  • [MFC] Error Importing logs using SCP - "Error while downloading file. The remote host has terminated the connection"

    Background: 
    In order to transfer logs to MFC, we had to use an intermediate logging server in our OOB network.  When this logging server crashed we had to rebuild the server with new hardware and SCP no longer worked.
    Issue: 
    The host key changed on the new server and had to be manually updated.  We suspected it was related to the hosts key but had difficulty finding where the known hosts info was stored.
    Solution: 
    Go to your install location of MFC and remove the known_hosts file.  In our case the file was located at:  "D:\Program Files\IronPort Systems\Mail Flow Central\mailFC\tmp\known_hosts".  Instead of removing the file, we renamed it to known_hosts.old and restarted the MFC service.  Afterwards we could see all the old logs importing.
    The issue itself was not difficult to resolve, it just took more time than expected for something that would seem straightforward.  To complicate things, we even raised a query to customercare who came back saying that they do not support the server on which MFC is running.  But clearly the source of the issue was related to the application rather than the server itself.

    Thanks for your comment qetzacoatl, however I don't this this will work for me, I am on a team, and we need to be able to check-in/out files and make sure we don't override eachothers work. I also don't want to have to use 2 programs to accomplish the task one should be able to do. Its now 3 weeks going and I can't get any work done, it seems like its getting worse. Nobody from Adobe seems to want to comment on my thread at all....So maybe I should just find a completely new solution and get rid of DW all together, Aptana is looking VERY nice right about now.

  • Getting an ORA-01003 error importing images using Apex import options

    I took an export of the images from my application (would be nice if this was included in the applicatoin export options)
    When I try to import, I get the following error
    Execution of the statement was unsuccessful. ORA-01003: no statement parsed
    begin wwv_flow.g_import_in_progress := true; end;
    ORA-01003: no statement parsed
    This worked for me on apex.oracle.com.. can't do it on any of my apex instances.
    I am able to import applications without a problem
    Any suggestions?

    Anyone have any ideas?

  • ORA-02374: conversion error loading table during import using IMPDP

    HI All,
    We are trying to migrate the data from one database to an other database.
    The source database is having character set
    SQL> select value from nls_database_parameters where parameter='NLS_CHARACTERSET';
    VALUE
    US7ASCII
    The destination database is having character set
    SQL> select value from nls_database_parameters where parameter='NLS_CHARACTERSET';
    VALUE
    AL32UTF8
    We took an export of the whole database using expdp and when we try to import to the destination database using impdp. We are getting the following error.
    ORA-02374: conversion error loading table <TABLE_NAME>
    ORA-12899: value too large for column <COLUMN NAME> (actual: 42, maximum: 40)
    ORA-02372: data for row:<COLUMN NAME> : 0X'4944454E5449464943414349E44E204445204C4C414D414441'
    Kindly let me know how to overcome this issue in destination.
    Thanks & Regards,
    Vikas Krishna

    Hi,
    You can overcome this issue by increasing the column width in the target database for the max value required for all data to be imported successfully in the table.
    Regards

  • Error when using IMPDP command on import

    I'm following an Oracle document guiding me thru the creation of schemas and importing of data for a Sample OBIEE App. But I'm getting an error using the IMPDP command. The command I'm issuing at the Command Prompt is as follows:
    impdp sys/password@Eagle as sysdba directory=datapumpdir dumpfile=SASchemas.dmp version=11.2.0.0.0 schemas=BISAMPLE,OBIEE_NAVTEQ,BIFOD,BIBPM,DEV_ODI_REPO,ODI_STAGING LOGFILE=SASchemas_imp.log
    (Note that I'm using my actual password in place of 'password')
    Unfortunately, I keep getting the error: unknown command beginning "impdp sys/..." rest of line ignored'
    Why is this not working? I've checked the syntax very carefully.
    Thanks, John

    Thanks for the replies. Yep it's a duplicate post, which I rarely do, but the subject matter related more to the database than OBIEE and I got no response in the OBIEE forum.
    Yes, this is a single machine installation so I am running the command on the machine where the database is up and running. I tried leaving version= 11.2.0.0 out of the command as Fiedi Z suggested, but I still get the same error. I'm scratching my head as to why the error handler doesn't even recognize that I'm using the IMPDP command. Any other ideas?
    Thanks, John

  • Getting Errors while using IMPDP command in UNIX

    Hi All,
    My work involves Refreshing of DB with latest Data. For this i am using IMPDP command to import the data. But Every time during refresh i will get errors saying like "oracle Generated Errors Oracle Not available".
    I don't know what is the reason for getting this error, and that too my db goes down after this error. Again i will startup the db , and i will try my luck to complete the process.
    so please tell me what are the actual problems that might cause the IMPDP to be failed.
    with one advise , i have stopped my listener and done the importing, Interestingly this time IMPDP works without errors. But this is also not always working.
    But i feel that every time starting and stopping the listener is not the correct solution.
    so please help me in this regard.
    Regards
    Naveen R.

    The error i am getting is...
    UDI-03113: operation generated ORACLE error 3113
    ORA-03113: end-of-file on communication channel
    Process ID: 6208
    Session ID: 346 Serial number: 298
    UDI-03114: operation generated ORACLE error 3114
    ORA-03114: not connected to ORACLE
    [user1@node1 cpt]$
    DB version is 11.1.0.6.0
    os is.. Red Hat Linux

  • Only to import  JAVA CLASS  and JAVA SOURCE  from .DMP file using IMPDP

    hi,
    I have a schema X, In that schema some of the *"JAVA CLASS" and "JAVA SOURCE"* objects are missing ..
    The procedures ,functions..etc objects were updated at X schema..
    I have 1 dmp file of Y schema , containing the missing "JAVA CLASS" and "JAVA SOURCE" s.. Can I import the the same to the schema X using IMPDP
    i tried using INCLUDE clause but it is not working
    eg:
    impdp john1/john1@me11g22 directory=datadump dumpfile=DEVF2WAR.DMP remap_schema=devf2war:john INCLUDE="JAVA CLASS","JAVA CLASS"but error..
    ORA-39001: invalid argument value
    ORA-39041: Filter  "INCLUDE" either identifies all object types or no object types.regards,
    jp

    Double post: IMPDP to import  JAVA CLASS  and JAVA SOURCE from .DMP file
    Already replied.
    Regards
    Gokhan

  • Import  TTS error

    Hi all,
    i am trying to import some TTS (Transaportable Table Space) but I'm getting a strange error.
    I tried first time and it worked fine but I forgot a TTS to import.
    So I dropped the TS, the schema and the tables to restart in a clear situation.
    Then I tried again but I get this error:
    Import: Release 11.1.0.6.0 - 64bit Production on Thursday, 30 July, 2009 18:49:25
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "MYDWH"."SYS_IMPORT_TRANSPORTABLE_01" successfully loaded/unloaded
    Starting "MYDWH"."SYS_IMPORT_TRANSPORTABLE_01": MYDWH/******** parfile=import.par
    Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK
    ORA-39123: Data Pump transportable tablespace job aborted
    ORA-00721: changes by release 10.2.0.3.0 cannot be used by release 10.2.0.1.0
    Job "MYDWH"."SYS_IMPORT_TRANSPORTABLE_01" stopped due to fatal error at 18:49:28
    ERROR:
    ORA-00604: error occurred at recursive SQL level 1
    ORA-20102: Error in Trigger IPFILT.IP_FILTERING
    ORA-06512: at line 119
    ORA-20001: User non found
    some help?
    which error is the fatal error?
    ORA-00721 or ORA-00604/ORA-20102?
    where do it find the version 10.2.0.3.0 and 10.2.0.1.0?
    My DB is 11g release 11.1.0.6.0!?!
    thanks,
    Sandrokan

    This usually happens when you are trying to downgrade using tts. Your compatibility for your target system which is running 11.1.0.6 looks to be set to 10.2.0.1. The dumpfile was created from a system with compatibility set to 10.2.0.3. Since there could be new things in the 10.2.0.3 dumpfile that 10.2.0.1 does not know about, the dumpfile can't be imported. The only way around this that I know of is to up the compatibility level on your 11.1.0.6 system to 10.2.0.3 or later.
    Dean

  • Can we use impdp to import the data from an normal exp dump?

    Hi All,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?
    Please suggest
    thanks and Regards
    Arun

    Hi,
    I have a export dump taken from a 9i database. Can i use impdp to import the data into 10g database from that 9i exp dump?Yes, it can be.
    Refer:
    http://wiki.oracle.com/thread/3734722/can+a+9i+dump+file+be+imported+into+10%3F
    thanks,
    X A H E E R

  • Getting error while importing metadata using View Objects

    Hi All,
    I am trying to create a repository using View Object in OBIEE 11.1.5.1 but getting error while viewing the data, after importing the metadata in the repository "[nQSError: 77031] Error occurs while calling remote service ADFService11G. Details: Runtime error for service -- ADFService11G - oracle/apps/fnd/applcore/common/ApplSession".
    I am also getting error "žADFException-2015: The BI Server is incompatible with the BI-ADF Broker Servlet: BI Server protocol version = null, BI-ADF Broker Servlet protocol version = 1" during testing my sample which is deployed to Admin server. I followed BI Adminstrator help file guide in order to create the sample for creating repository using view object.
    Admin server log says
    [2011-09-27T02:59:03.646-05:00] [AdminServer] [NOTIFICATION] [] [oracle.bi.integration.adf] [tid: [ACTIVE].ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: <anonymous>] [ecid: b260b57746aa92d3:-1f9ca26:1328fcfd3e6:-8000-0000000000006744,0] [APP: BIEEOrders] [[
    QUERY:
    <?xml version="1.0" encoding="UTF-8" ?><ADFQuery><Parameters></Parameters><Projection><Attribute><Name><![CDATA[Deptno]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute><Attribute><Name><![CDATA[Dname]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute><Attribute><Name><![CDATA[Loc]]></Name><ViewObject><![CDATA[AppModule.DeptViewObj1]]></ViewObject></Attribute></Projection><JoinSpec><ViewObject><Name><![CDATA[AppModule.DeptViewObj1]]></Name></ViewObject></JoinSpec></ADFQuery>
    [2011-09-27T02:59:04.199-05:00] [AdminServer] [ERROR] [] [oracle.bi.integration.adf.v11g.obieebroker] [tid: [ACTIVE].ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] [userId: <anonymous>] [ecid: b260b57746aa92d3:-1f9ca26:1328fcfd3e6:-8000-0000000000006744,0] [APP: BIEEOrders] java.lang.NoClassDefFoundError: oracle/apps/fnd/applcore/common/ApplSession[[
         at oracle.bi.integration.adf.ADFDataQuery.makeQueryBuilder(ADFDataQuery.java:81)
         at oracle.bi.integration.adf.ADFDataQuery.<init>(ADFDataQuery.java:70)
         at oracle.bi.integration.adf.ADFReadQuery.<init>(ADFReadQuery.java:15)
         at oracle.bi.integration.adf.ADFService.makeADFQuery(ADFService.java:227)
         at oracle.bi.integration.adf.ADFService.execute(ADFService.java:136)
         at oracle.bi.integration.adf.v11g.obieebroker.ADFServiceExecutor.execute(ADFServiceExecutor.java:185)
         at oracle.bi.integration.adf.v11g.obieebroker.OBIEEBroker.doGet(OBIEEBroker.java:89)
         at oracle.bi.integration.adf.v11g.obieebroker.OBIEEBroker.doPost(OBIEEBroker.java:106)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:300)
         at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.adf.share.http.ServletADFFilter.doFilter(ServletADFFilter.java:62)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.security.jps.ee.http.JpsAbsFilter$1.run(JpsAbsFilter.java:111)
         at java.security.AccessController.doPrivileged(Native Method)
         at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:313)
         at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:413)
         at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:94)
         at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:161)
         at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:136)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at weblogic.servlet.internal.RequestEventsFilter.doFilter(RequestEventsFilter.java:27)
         at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.wrapRun(WebAppServletContext.java:3715)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3681)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2277)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2183)
         at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1454)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    Caused by: java.lang.ClassNotFoundException: oracle.apps.fnd.applcore.common.ApplSession
         at weblogic.utils.classloaders.GenericClassLoader.findLocalClass(GenericClassLoader.java:297)
         at weblogic.utils.classloaders.GenericClassLoader.findClass(GenericClassLoader.java:270)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:305)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:246)
         at weblogic.utils.classloaders.GenericClassLoader.loadClass(GenericClassLoader.java:179)
         ... 38 more
    Please suggest how to make it work.

    Hi,
    Thanks for providing the online help URL, i have already completed the steps mentioned in URL
    I am able to import metadata using view object but getting above error("[nQSError: 77031] Error occurs while calling remote service ADFService11G. Details: Runtime error for service -- ADFService11G - oracle/apps/fnd/applcore/common/ApplSession".") while validating the data.
    It fails at step 5 of "To import metadata from an ADF Business Component data source:" section of above URL.

  • Error message when importing data using Import and export wizard

    Getting below error message when importing data using IMPORT and EXPORT WIZARD
    Error 0xc0202009: Data Flow Task 1: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.
    <dir>
    <dir>
    Messages
    Error 0xc0202009: Data Flow Task 1: SSIS Error Code DTS_E_OLEDBERROR.  An OLE DB error has occurred. Error code: 0x80004005.
    An OLE DB record is available.  Source: "Microsoft SQL Server Native Client 11.0"  Hresult: 0x80004005  Description: "Could not allocate a new page for database REPORTING' because of insufficient disk space in filegroup 'PRIMARY'.
    Create the necessary space by dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.".
    (SQL Server Import and Export Wizard)
    Error 0xc0209029: Data Flow Task 1: SSIS Error Code DTS_E_INDUCEDTRANSFORMFAILUREONERROR.  The "Destination - Buyer_.Inputs[Destination Input]" failed because error code 0xC020907B occurred, and the error row disposition on "Destination
    - Buyer_First_Qtr.Inputs[Destination Input]" specifies failure on error. An error occurred on the specified object of the specified component.  There may be error messages posted before this with more information about the failure.
    (SQL Server Import and Export Wizard)
    Error 0xc0047022: Data Flow Task 1: SSIS Error Code DTS_E_PROCESSINPUTFAILED.  The ProcessInput method on component "Destination - Buyer" (28) failed with error code 0xC0209029 while processing input "Destination Input" (41). The
    identified component returned an error from the ProcessInput method. The error is specific to the component, but the error is fatal and will cause the Data Flow task to stop running.  There may be error messages posted before this with more information
    about the failure.
    (SQL Server Import and Export Wizard)
    Error 0xc02020c4: Data Flow Task 1: The attempt to add a row to the Data Flow task buffer failed with error code 0xC0047020.
    (SQL Server Import and Export Wizard)
    </dir>
    </dir>
    Error 0xc0047038: Data Flow Task 1: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED.  The PrimeOutput method on Source - Buyer_First_Qtr returned error code 0xC02020C4.  The component returned a failure code when the pipeline engine called PrimeOutput().
    The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing.  There may be error messages posted before this with more information about the failure.
    (SQL Server Import and Export Wizard)
    Smash126

    Hi Smash126,
    Based on the error message” Could not allocate a new page for database REPORTING' because of insufficient disk space in filegroup 'PRIMARY'. Create the necessary space by dropping objects in the filegroup, adding additional files to the filegroup, or setting
    autogrowth on for existing files in the filegroup”, we can know that the issue is caused by the there is no sufficient disk space in filegroup 'PRIMARY' for the ‘REPORTING’ database.
    To fix this issue, we can add additional files to the filegroup by add a new file to the PRIMARY filegroup on Files page, or setting Autogrowth on for existing files in the filegroup to increase the necessary space.
    The following document about Add Data or Log Files to a Database is for your reference:
    http://msdn.microsoft.com/en-us/library/ms189253.aspx
    If there are any other questions, please feel free to ask.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • ORA:31696 error occured while imporitng using impdp utility

    Hi Forum Team,
    I found below exception while importing export dump file into one of my schema user using impdp utility.
    ORA-31696: unable to export/import TABLE_DATA:"PMM"."IMGCTLEE" using client specified AUTOMATIC method
    ORA-31696: unable to export/import TABLE_DATA:"PMM"."PRDSIMEE" using client specified AUTOMATIC method
    PMM is my schema user. But, the import process is not erminating after this exception. its continuing.
    import command given as below.
    impdp userid=pmm_css/pmm_css@pmmtrnglocal dumpfile='pmm750.dmp' logfile='pmm750_1_log.log' schemas=pmm
    PLease can some one reply with answer to my query?
    Thanks in advance.

    Hi.
    Your tables has columns of which type.. ??? maybe LONG, RAW, others..
    There are some metalink's bugs related... you can check that is your case...
    1. 4239903
    Regards.

Maybe you are looking for