Importin dump file to a partition table

Import: Release 10.2.0.1.0 - Production on Wed Feb 28 16:25:13 2007
Copyright (c) 1982, 2005, Oracle. All rights reserved.
Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bit Production
With the Partitioning and Data Mining options
Export file created by EXPORT:V09.02.00 via conventional path
Warning: the objects were exported by EXP_FULL_USER, not by you
import done in UTF8 character set and AL16UTF16 NCHAR character set
export client uses WE8ISO8859P1 character set (possible charset conversion)
IMP-00055: Warning: partition or subpartition "CALL_DETAIL_2005":"CD_200501" not found in export file
Import terminated successfully with warnings.
BUT PARTITIONS ARE THERE , AND I CAN SEE THEM THROUGH TOAD
Message was edited by:
shakil_zubair

Hi,
When I am using the second method by creating the
partition by subquery I am getting these errors:
Table created.
SP2-0734: unknown command beginning "PARTITION ..." -
rest of line ignored.
SP2-0734: unknown command beginning "VALUES LES..." -
rest of line ignored.
SP2-0734: unknown command beginning "STORAGE ( ..." -
rest of line ignored.
SP2-0734: unknown command beginning "PARTITION ..." -
rest of line ignored.
SP2-0044: For a list of known commands enter HELP
and to leave enter EXIT.
SP2-0734: unknown command beginning "VALUES LES..." -
rest of line ignored.
SP2-0734: unknown command beginning "STORAGE ( ..." -
rest of line ignored.What is the statement you're executing? I believe Search-Net's example inadvertently left off a close paren, i.e.
create table mytbpart(id number(10), price number(10) )
partition by range(price)
(partition p1 values less than (501),
partition p2 values less than (maxvalue));but I'm not sure that would cause the particular errors you're seeing...
Justin

Similar Messages

  • Dump File Size vs Database Table Size

    Hi all!
    Hope you're all well. If Datapump estimates that 18 million records will produce a 2.5GB dumpfile, does this mean that 2.5GB will also be consumed on the database table when this dump file is imported into a database?
    Many thanks in advance!
    Regards
    AC

    does this mean that 2.5GB will also be consumed on the database table when this dump file is imported into a database?No since the size after import depends on various factors like block size, block storage parameters etc.

  • Importing Some Partion of a Partitioned Table

    Hi Guru's,
    I import a partitioned table using original import option, during import i am successful to import 80% of data but after that import discontinued due to n/w error. I just want to know can i continue my import from where it is discontinued or i need to import the complete partitioned table again. The export dump file is complete partitioned table export not partition wise export.
    Any help will be greatly appreciated.
    Thank in advance.

    >
    I import a partitioned table using original import option, during import i am successful to import 80% of data but after that import discontinued due to n/w error. I just want to know can i continue my import from where it is discontinued or i need to import the complete partitioned table again. The export dump file is complete partitioned table export not partition wise export.
    >
    Use the TABLES clause in the import statement.
    See the Utilities doc
    http://docs.oracle.com/cd/B28359_01/server.111/b28319/dp_import.htm
    >
    TABLES
    Default: There is no default
    Purpose
    Specifies that you want to perform a table-mode import.
    Syntax and Description
    TABLES=[schema_name.]table_name[:partition_name]
    In a table-mode import, you can filter the data that is imported from the source by specifying a comma-delimited list of tables and partitions or subpartitions.
    If you do not supply a schema_name, it defaults to that of the current user. To specify a schema other than your own, you must either have the IMP_FULL_DATABASE role or remap the schema to the current user.
    The use of filtering can restrict what is imported using this import mode. See Filtering During Import Operations.
    If a partition_name is specified, it must be the name of a partition or subpartition in the associated table.
    >
    The doc has an example
    >
    The following example shows the use of the TABLES parameter to import partitions:
    impdp hr DIRECTORY=dpump_dir1 DUMPFILE=expdat.dmp TABLES=sh.sales:sales_Q1_2000,sh.sales:sales_Q2_2000

  • How to import external table, which exist in export dump file.

    My export dump file has one external table. While i stated importing into my developement instance , I am getting the error "ORA-00911: invalid character".
    The original definition of the extenal table is as given below
    CREATE TABLE EXT_TABLE_EV02_PRICEMARTDATA
    EGORDERNUMBER VARCHAR2(255 BYTE),
    EGINVOICENUMBER VARCHAR2(255 BYTE),
    EGLINEITEMNUMBER VARCHAR2(255 BYTE),
    EGUID VARCHAR2(255 BYTE),
    EGBRAND VARCHAR2(255 BYTE),
    EGPRODUCTLINE VARCHAR2(255 BYTE),
    EGPRODUCTGROUP VARCHAR2(255 BYTE),
    EGPRODUCTSUBGROUP VARCHAR2(255 BYTE),
    EGMARKETCLASS VARCHAR2(255 BYTE),
    EGSKU VARCHAR2(255 BYTE),
    EGDISCOUNTGROUP VARCHAR2(255 BYTE),
    EGREGION VARCHAR2(255 BYTE),
    EGAREA VARCHAR2(255 BYTE),
    EGSALESREP VARCHAR2(255 BYTE),
    EGDISTRIBUTORCODE VARCHAR2(255 BYTE),
    EGDISTRIBUTOR VARCHAR2(255 BYTE),
    EGECMTIER VARCHAR2(255 BYTE),
    EGECM VARCHAR2(255 BYTE),
    EGSOLATIER VARCHAR2(255 BYTE),
    EGSOLA VARCHAR2(255 BYTE),
    EGTRANSACTIONTYPE VARCHAR2(255 BYTE),
    EGQUOTENUMBER VARCHAR2(255 BYTE),
    EGACCOUNTTYPE VARCHAR2(255 BYTE),
    EGFINANCIALENTITY VARCHAR2(255 BYTE),
    C25 VARCHAR2(255 BYTE),
    EGFINANCIALENTITYCODE VARCHAR2(255 BYTE),
    C27 VARCHAR2(255 BYTE),
    EGBUYINGGROUP VARCHAR2(255 BYTE),
    QTY NUMBER,
    EGTRXDATE DATE,
    EGLISTPRICE NUMBER,
    EGUOM NUMBER,
    EGUNITLISTPRICE NUMBER,
    EGMULTIPLIER NUMBER,
    EGUNITDISCOUNT NUMBER,
    EGCUSTOMERNETPRICE NUMBER,
    EGFREIGHTOUTBOUNDCHARGES NUMBER,
    EGMINIMUMORDERCHARGES NUMBER,
    EGRESTOCKINGCHARGES NUMBER,
    EGINVOICEPRICE NUMBER,
    EGCOMMISSIONS NUMBER,
    EGCASHDISCOUNTS NUMBER,
    EGBUYINGGROUPREBATES NUMBER,
    EGINCENTIVEREBATES NUMBER,
    EGRETURNS NUMBER,
    EGOTHERCREDITS NUMBER,
    EGCOOP NUMBER,
    EGPOCKETPRICE NUMBER,
    EGFREIGHTCOSTS NUMBER,
    EGJOURNALBILLINGCOSTS NUMBER,
    EGMINIMUMORDERCOSTS NUMBER,
    EGORDERENTRYCOSTS NUMBER,
    EGRESTOCKINGCOSTSWAREHOUSE NUMBER,
    EGRETURNSCOSTADMIN NUMBER,
    EGMATERIALCOSTS NUMBER,
    EGLABORCOSTS NUMBER,
    EGOVERHEADCOSTS NUMBER,
    EGPRICEADMINISTRATIONCOSTS NUMBER,
    EGSHORTPAYMENTCOSTS NUMBER,
    EGTERMCOSTS NUMBER,
    EGPOCKETMARGIN NUMBER,
    EGPOCKETMARGINGP NUMBER,
    EGWEIGHTEDAVEMULTIPLIER NUMBER
    ORGANIZATION EXTERNAL
    ( TYPE ORACLE_LOADER
    DEFAULT DIRECTORY EV02_PRICEMARTDATA_CSV_CON
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv')
    REJECT LIMIT UNLIMITED
    NOPARALLEL
    NOMONITORING;
    While importing , when i seen the log file , it is failing the create the external table. Getting the error "ORA-00911: invalid character".
    Can some one suggest how to import external tables
    Addressing this issue will be highly appriciated.
    Naveen

    Hi Srinath,
    When i observed the create table syntax of external table from import dump log file, it show few lines as below. I could not understand these special characters. And create table definationis failing with special character viz ORA-00911: invalid character
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv').
    I even observed the create table DDL from TOAD. It is same as i mentioned earlier
    Naveen

  • Archiving old data from a partitioned table

    Hi,
    While sifting through all the options for archiving the old data from a table which is also indexed, i came across a few methods which could be used:
    1. Use a CTAS to create new tables on a different tablespace from the partitions on the exisitng table. Then swap these new tables with the data from the partitions, drop the partitions on the exisiting table (or truncate them although i m not sure which one is the recommended method),offline this tablespace and keep it archived. In case you require it in the future, again swap these partitions wih the data from the archived tables. I am not sure if i got the method correctly.
    2. Keep an export of all the partitions which need to be archived and keep that .dmp file on a storage media. Once they are exported, truncate the corresponding partitions in the original table. If required in the future, import these partitions.
    But i have one constraint on my Db which is I cannot create a new archive tablespace for holding the tables containing the partitioned data into then as I have only 1 tablespace allocated to my appplication on that DB as there are multiple apps residing on it together. Kindly suggest what option is the best suited for me now. Should I go with option 2?
    Thanks in advance.

    Hi,
    Thanks a bunch for all your replies. Yeah, I am planning to go ahead with the option 2. Below is the method I have decided upon. Kindly verify is my line of understanding is correct:
    1. export the partition using the clause:
    exp ID/passwd file=abc.dmp log=abc.log tables=schemaname.tablename:partition_name rows=yes indexes=yes
    2. Then drop this partition on the original table using the statement:
    ALTER TABLE tablename drop PARTITION partition_name UPDATE GLOBAL INDEXES;
    If I now want to import this dump file into my original table again, I will first have to create this partition on my table using the statement:
    3. ALTER TABLE tablename ADD PARTITION partition_name VALUES LESS THAN ( '<<>>' ) ;
    4. Then import the data into that partition using:
    imp ID/passwd FILE=abc.dmp log=xyz.log TABLES=schemaname.tablename:partition_name IGNORE=y
    Now my query here is that this partitioned table has a global index associated with it. So once i create this partition and import the data into it, do i need to drop and recreate the global indexes on this table or is there any aother method to update the indexes. Kindly suggest.
    Thanks again!! :)

  • Sql* Loader syntax to load data into a partitioned table

    Hi All,
    I was trying to load data from a csv file to a partitioned table.
    The table name is countries_info
    columns :
    country_code ,
    country_name,
    country_language
    The column country_code is list partitioned with partition p1 for values 'EN','DE','IN' etc
    and partition p2 for 'KR','AR','IT' etc.
    I tried to load data into this table, but I was getting some error Message (not mapping to partitioned key).
    I tried syntax
    load data
    infile 'countries.csv'
    append
    into table countries_info
    partition(p1) ,
    partition(p2)
    fields terminated by ','
    country_code ,
    country_name,
    country_language
    What is the correct syntax- I searched a lot but have not been able to find out.

    Peeush_Rediff wrote:
    Hi All,
    I tried to load data into this table, but I was getting some error Message (not mapping to partitioned key).It's not some error message, it's relevant information for resolving problems you encounter while using Oracle.
    In your case, although you didn't specifiy the exact ORA you recived, it sounds like [ORA-14400|http://forums.oracle.com/forums/search.jspa?threadID=&q=ORA-14400&objID=f61&dateRange=all&userID=&numResults=15&rankBy=10001]
    What is the correct syntax- I searched a lot but have not been able to find out.It's not about correct syntax , it's about understanding message that was trying to tell you that something went wrong.
    That message was (i guess) ORA-14400.
    So, refering to that ORA message , you will need to add new partition ,
    where the data from your csv file that currently don't map to any of the specified partition would fit ,
    or add partition that would be able to accept data range in which value/s for which you received that ORA message could fit into.

  • Can I use gpt to recreate/unerase a partition table? (Rebuild the GPT/GUID partition table?)  I don't want to do FILE recovery.

    (Yes, I've  googled a bunch and read threads like this one already.)
    Can I use gpt or some other app to recreate/unerase a partition table?  That is, how can I rebuild a disk's GPT/GUID partition table?)  I don't want to do FILE recovery.
    What happened: Instead of erasing a single partition off a disk with many partitions, the entire partition table was erased (using Disk Utility, w/o deleting the underlying files).  Somehow the "Erasing a disk deletes all data on all its partitions." warning message was missed.
    I have a copy of the output of df, with the number of blocks in each partition, from just prior to the erasure, so I should be able to recreate the GPT/GUID partition table.  Editing the GPT with a hex editor is not feasible.  Simply recreating the partitions with Disk Utility will overwrite the key filesystem tables on each partition, and I don't want to do that, plus Disk Utility doesn't allow me to specify exact partition sizes anyway.
    Surely there's an app for rebuilding the partition table (other than emacs' hexl-mode!) for recreating/unerasing a partition table when the partition sizes and orders are known?  I've looked at the advertising for a bunch of recovery software and none of them clearly indicate that they will do what I want. 
    I guess I can try using gpt on a copy of the reformatted drive I've made with dd, and see what happens.  But perhaps someone knows of a tool that should do what I need, or knows if gpt is that tool or not.
    There are answers and tools that will do FILE recovery - search for files and recover the ones that aren't fragmented or deleted.  As far as I can find, they just look for files on the disk, and don't pay much, if any attention to the filesystem info or directory heirarchy, which in this case is valuable.  Of course I could send it in to DriveSavers, or the like.  But none of that seems necessary, and the scavenging file recovery apps won't do the job well,
    E.g. some are mentioned here:
    I don't want to do FILE recovery.
    Thanks for any help.
    The links in this post are to pages describing the underlined term, e.g. the man pages for df and gpt.
    dd output includes:
    Filesystem
    512-blocks 
    Used Available Capacity  Mounted on

    Aperture has the ability to work with files in their existing location. They are called "referenced masters." When you import images, you should select the "In their current location" in the "Store Files:" drop down box. Have a read of the documentation for full specifics. Unsure how you can resolve your duplication; might be some work but next time have a read of the manual first
    Information for versions is stored in the Aperture database (library file). The masters can be inside the library file itself, or they can be somewhere else.

  • Is it possible to view/get the table data from a dump file

    I have dmp files generated using expdp on oracle 11g..
    expdp_schemas_18MAY2013_1.dmp
    expdp_schemas_18MAY2013_2.dmp
    expdp_schemas_18MAY2013_3.dmp
    Can I use a parameter file given below to get the table data in to the sql file or impdp the only option to load the table data in to database.
    vi test1.par
    USERID="/ as sysdba"
    DIRECTORY=DATA
    dumpfile=expdp_schemas_18MAY2013%S.dmp
    SCHEMAS=USER1,USER2
    LOGFILE=user_dump_data.log
    SQLFILE=user_dump_data.sql
    and impdp parfile=test1.par.

    Hi,
    To explain more about my situation.
    Target database has the dump files, where as the source database (cloud) doesn't have access to the target database.
    However, I can request the target DB DBA team to run the par files and provide me the output like a SQL file which I can take and run it in my source database.
    I got the metadata the same way, but is there any way I could get the data from the dump files on the target DB without actually accessing it? My question might sound weird, but please let me know.
    <
    1. export from the source into a dumpfile. You can do this on the source database and then copy the file over to your local database or you can do this from a local database if you have a database link defined on the local database that points to the source database.  In the second case, your dumpfile will be created on your local database.
    >
    How can i access the dump using the database link?

  • Can we find the number of rows in table from the dump file

    Hi All,
    Can we find the number of rows in table from the dump file with out importing the table in to the database?
    Please let me know ,if any option is there.
    Thanks,
    Kumar.

    <s>Try to import with option SHOW=Y, that should skip the number of rows which are into a table from a dump file.</s><br>
    <br>
    Nicolas.<br>
    Oops, sorry, that doesn't show the number of lines...<br>
    Message was edited by: <br>
    N. Gasparotto

  • How to import only rows of tables from from a dump file

    Good day,
    I have impoted a schema from a dump files with option rows=n,(accordingly it should import only structure of the tables)
    now i want to import rows for these tables in that schema. Kindly guide me how to do that. I have used below syntax for import.
    imp userid=system/manager@testdb fromuser=abc touser=abc file=d:\abctest\
    abc.dmp log=e:\abc15.log buffer=100000000 ignore=y constraints=n grants=n statistics=none rows=n;
    Any help willbe much appriciated.
    Regards
    Ansh

    Hi Sybrand,
    Thanx for the suggestion. Actually it is my fault that I had not explained the whole scenarion.
    Let me explain it one by one...
    (1) First i was importing a schema with rows=y option and I was getting some constraints error(can not insert NULL value in 'not null' column of some particular table) and import was unsuccesful.
    (2) Then i desided to import only structure of the tables in that schema, and later disabled that constraint which was creating the problem.
    (3) Now I m willing to import the data in these tables using that .dmp file, using rows=y, and this was my question if i m doing it in a right way or do i need to do some further step.
    Hope this time i have made picture more clear.
    Thanx a lot Hans and Sybrand
    btw no wall around so can't bang my head :)

  • Where is my EXPORT dump file ? How to do BRTools off line table reorgan ?

    Hello,
    Here was my BRTools Export output message :
    About to export specified tables via Direct Path ...
    Current user changed to SAPPRD
    . . exporting table                          CDHDR   60357362 rows exported
    Export terminated successfully without warnings.
    BR0280I BRSPACE time stamp: 2011-08-09 16.53.04
    BR1160I 1 table exported by EXP utility
    BR0280I BRSPACE time stamp: 2011-08-09 16.53.04
    BR0670I Enter 'c[ont]' to continue, 'b[ack]' to go back, 's[top]' to abort:
    But where is my Export dump file ?
    Besides, could you tell me the procedures of using BRTools Export & Import to do off-line table reorgan ?
    Due to some reasons, I can't do online table reorgan.

    Hi Ming,
    I noted the OSS note 646681 - Reorganizing tables with BRSPACE in my previous message, already. You don't need any additional effort such as creating script, if you use brspace as I noted in my last message.
    Stop the SAP system, but not database then execute the command, below;
    1) brspace -u / -f tbreorg -t "CDHDR"
    2) At the incoming menu steps, select --> (8 - Table reorganization mode (mode)) .. offline
    3) Then continue with "c"
    At the end of this steps, table and related indexes will be reorganized.
    You can update statistics by using the command, below after you complete the steps, above respectfully;
    brconnect -u / -c -f stats -t all -f collect -p 4
    Best regards,
    Orkun Gedik

  • Using exp dump file, how to know schema/tablespace/database/table....

    Hi
    I have export dump file. I don't have any information other then that about the dumpfile.
    1. How do I know what version of exp was used(or database version)?
    2. what does the exp contain, whether it is table/schema/tablespace/database level export?
    3. Is it possible that I can used impdp command to export a file which was exported using "exp" command?

    1. Sorry for my ignorance but in first case how do I know if i have to use "imp or impdp".
    As discussed in the above mentioned thread strings might reveal something but not sure. Otherwise you can always confirm by trying to run it.
    And assuming that if I imported the dump(using trial and error rule), how do I know what level of export was the dump file created... ie
    how do I know the present object(object imported in target) was belonging to which tablespace and schema in source database(question 2 of "Sidhu")?
    When you will import it back in a new database, everything will be same as the original database. All objects will belong to their respective tablespaces,schemas blah blah...(where they were in the original database).
    Sidhu

  • Error while importing a dump file in my local database

    Hi,
    I have a dump file named *"system_21Sep2010_DVSD.DMP"* which i'm trying to import in my local database named *"nandita1"*.
    The dump file is found under the following path *"C:\DevSuiteHome_1\BIN"* and i executed the following command for importing the dump file:
    C:\DevSuiteHome_1\BIN>imp system/system@nandita1 FILE=system_21Sep2010_DVSD.DMP FULL=YES
    I'm getting the following error while importing the dump file:
    Import: Release 10.1.0.4.2 - Production on Wed Oct 6 16:25:37 2010
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Produc
    tion
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.01.00 via conventional path
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    IMP-00015: following statement failed because the object already exists:
    "CREATE UNDO TABLESPACE "UNDOTBS1" BLOCKSIZE 8192 DATAFILE '/oradata_dev/DV"
    "SD/undotbs01.dbf' SIZE 104857600 AUTOEXTEND ON NEXT 5242880 MAXSIZE "
    "32767M EXTENT MANAGEMENT LOCAL "
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "SYSAUX" BLOCKSIZE 8192 DATAFILE '/oradata_dev/DVSD/sysa"
    "ux01.dbf' SIZE 367001600 AUTOEXTEND ON NEXT 10485760 MAXSIZE 32767M "
    "EXTENT MANAGEMENT LOCAL AUTOALLOCATE ONLINE PERMANENT SEGMENT SPACE MANA"
    "GEMENT AUTO"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TEMPORARY TABLESPACE "TEMP" BLOCKSIZE 8192 TEMPFILE '/oradata_dev/D"
    "VSD/temp01.dbf' SIZE 20971520 AUTOEXTEND ON NEXT 655360 MAXSIZE 3276"
    "7M EXTENT MANAGEMENT LOCAL UNIFORM SIZE 1048576"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "USERS" BLOCKSIZE 8192 DATAFILE '/oradata_dev/DVSD/users"
    "01.dbf' SIZE 5242880 AUTOEXTEND ON NEXT 1310720 MAXSIZE 32767M EXTEN"
    "T MANAGEMENT LOCAL AUTOALLOCATE ONLINE PERMANENT SEGMENT SPACE MANAGEMEN"
    "T AUTO"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "XX_DATA" BLOCKSIZE 8192 DATAFILE '/oradata_dev/DVSD/XX_"
    "DATA01.dbf' SIZE 140509184 AUTOEXTEND ON NEXT 1048576 MAXSIZE 32767M"
    " EXTENT MANAGEMENT LOCAL AUTOALLOCATE ONLINE PERMANENT "
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "XX_INDEX" BLOCKSIZE 8192 DATAFILE '/oradata_dev/DVSD/XX"
    "_INDEX.dbf' SIZE 93323264 AUTOEXTEND ON NEXT 1048576 MAXSIZE 32767M "
    "EXTENT MANAGEMENT LOCAL AUTOALLOCATE ONLINE PERMANENT "
    IMP-00015: following statement failed because the object already exists:
    "CREATE PROFILE "MONITORING_PROFILE" LIMIT COMPOSITE_LIMIT DEFAULT SESSIONS_"
    "PER_USER DEFAULT CPU_PER_SESSION DEFAULT CPU_PER_CALL DEFAULT LOGICAL_READS"
    "_PER_SESSION DEFAULT LOGICAL_READS_PER_CALL DEFAULT IDLE_TIME DEFAULT CONNE"
    "CT_TIME DEFAULT PRIVATE_SGA DEFAULT FAILED_LOGIN_ATTEMPTS UNLIMITED PASSWOR"
    "D_LIFE_TIME DEFAULT PASSWORD_REUSE_TIME DEFAULT PASSWORD_REUSE_MAX DEFAULT "
    "PASSWORD_LOCK_TIME DEFAULT PASSWORD_GRACE_TIME DEFAULT"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "OUTLN" IDENTIFIED BY VALUES '4A3BA55E08595C81' TEMPORARY TABLE"
    "SPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "TSMSYS" IDENTIFIED BY VALUES '3DF26A8B17D0F29F' DEFAULT TABLES"
    "PACE "USERS" TEMPORARY TABLESPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "ANONYMOUS" IDENTIFIED BY VALUES 'anonymous' DEFAULT TABLESPACE"
    " "SYSAUX" TEMPORARY TABLESPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "OLAPSYS" IDENTIFIED BY VALUES '3FB8EF9DB538647C' DEFAULT TABLE"
    "SPACE "SYSAUX" TEMPORARY TABLESPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "MDDATA" IDENTIFIED BY VALUES 'DF02A496267DEE66' DEFAULT TABLES"
    "PACE "USERS" TEMPORARY TABLESPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "SYSMAN" IDENTIFIED BY VALUES '28F72A3C2D75FDE9' DEFAULT TABLES"
    "PACE "SYSAUX" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "MGMT_VIEW" IDENTIFIED BY VALUES '442167C25FAC883C' TEMPORARY T"
    "ABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "SCOTT" IDENTIFIED BY VALUES 'F894844C34402B67' DEFAULT TABLESP"
    "ACE "USERS" TEMPORARY TABLESPACE "TEMP" PASSWORD EXPIRE ACCOUNT LOCK"
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "BUXXDDWA" IDENTIFIED BY VALUES '2672292A792DB64B' DEFAULT TABL"
    "ESPACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "XXDDWA" IDENTIFIED BY VALUES 'F5C178B2796496F4' DEFAULT TABLES"
    "PACE "XX_INDEX" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "BUXXDDWS" IDENTIFIED BY VALUES '06D47E0665132890' DEFAULT TABL"
    "ESPACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "BUXXSUC" IDENTIFIED BY VALUES '1705C688518ADCDC' DEFAULT TABLE"
    "SPACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "JIRAUSER" IDENTIFIED BY VALUES 'F0E1C662FCD1E09F' DEFAULT TABL"
    "ESPACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "XXSFA" IDENTIFIED BY VALUES 'C3DFA37A29E072B6' DEFAULT TABLESP"
    "ACE "XX_INDEX" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "XXDDW" IDENTIFIED BY VALUES 'CFF3D9B1B6C20DCC' DEFAULT TABLESP"
    "ACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "XXDDWS" IDENTIFIED BY VALUES '9EF0EE1BAC5855C4' DEFAULT TABLES"
    "PACE "XX_DATA" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE USER "XXSUC" IDENTIFIED BY VALUES '39594B53A6F1CED0' DEFAULT TABLESP"
    "ACE "XX_INDEX" TEMPORARY TABLESPACE "TEMP""
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "SELECT_CATALOG_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "SELECT_CATALOG_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "EXECUTE_CATALOG_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "EXECUTE_CATALOG_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "DELETE_CATALOG_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "DELETE_CATALOG_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "RECOVERY_CATALOG_OWNER""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "RECOVERY_CATALOG_OWNER" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "GATHER_SYSTEM_STATISTICS""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "GATHER_SYSTEM_STATISTICS" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "LOGSTDBY_ADMINISTRATOR""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "LOGSTDBY_ADMINISTRATOR" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "AQ_ADMINISTRATOR_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "AQ_USER_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "AQ_USER_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "GLOBAL_AQ_USER_ROLE" IDENTIFIED GLOBALLY "
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "SCHEDULER_ADMIN""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "SCHEDULER_ADMIN" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "HS_ADMIN_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "HS_ADMIN_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "AUTHENTICATEDUSER""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "AUTHENTICATEDUSER" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "OEM_ADVISOR""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "OEM_ADVISOR" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "OEM_MONITOR""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "OEM_MONITOR" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "WM_ADMIN_ROLE""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "WM_ADMIN_ROLE" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVAUSERPRIV""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVAUSERPRIV" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVAIDPRIV""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVAIDPRIV" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVASYSPRIV""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVASYSPRIV" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVADEBUGPRIV""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVADEBUGPRIV" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "EJBCLIENT""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "EJBCLIENT" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVA_ADMIN""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVA_ADMIN" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "JAVA_DEPLOY""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "JAVA_DEPLOY" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "CTXAPP""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "CTXAPP" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "XDBADMIN""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "XDBADMIN" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "XDBWEBSERVICES""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "XDBWEBSERVICES" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "OLAP_DBA""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "OLAP_DBA" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "OLAP_USER""
    IMP-00015: following statement failed because the object already exists:
    "REVOKE "OLAP_USER" FROM SYSTEM"
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLE "MGMT_USER""
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "MVIEW$_ADVSEQ_GENERIC" MINVALUE 1 MAXVALUE 4294967295 INCR"
    "EMENT BY 1 START WITH 1 CACHE 50 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "MVIEW$_ADVSEQ_ID" MINVALUE 1 MAXVALUE 4294967295 INCREMENT"
    " BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "LOGMNR_EVOLVE_SEQ$" MINVALUE 1 MAXVALUE 999999999999999999"
    "999999999 INCREMENT BY 1 START WITH 1 CACHE 20 ORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "LOGMNR_SEQ$" MINVALUE 1 MAXVALUE 9999999999999999999999999"
    "99 INCREMENT BY 1 START WITH 1 CACHE 20 ORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "LOGMNR_UIDS$" MINVALUE 1 MAXVALUE 999999999999999999999999"
    "999 INCREMENT BY 1 START WITH 100 CACHE 20 ORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_FLAVORS_S" MINVALUE -2147483647 MAXVALUE 214748364"
    "7 INCREMENT BY 1 START WITH 1 NOCACHE NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_FLAVOR_NAME_S" MINVALUE 1 MAXVALUE 999999999999999"
    "999999999999 INCREMENT BY 1 START WITH 1 NOCACHE NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_REPPROP_KEY" MINVALUE 1 MAXVALUE 99999999999999999"
    "9999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT_LOG_SEQUENCE" MINVALUE 1 MAXVALUE 99999999999999999"
    "9999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_REFRESH_TEMPLATES_S" MINVALUE 1 MAXVALUE 999999999"
    "999999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_USER_AUTHORIZATIONS_S" MINVALUE 1 MAXVALUE 9999999"
    "99999999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_TEMPLATE_REFGROUPS_S" MINVALUE 1 MAXVALUE 99999999"
    "9999999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_TEMPLATE_OBJECTS_S" MINVALUE 1 MAXVALUE 9999999999"
    "99999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_TEMPLATE_PARMS_S" MINVALUE 1 MAXVALUE 999999999999"
    "999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_USER_PARM_VALUES_S" MINVALUE 1 MAXVALUE 9999999999"
    "99999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_TEMPLATE_SITES_S" MINVALUE 1 MAXVALUE 999999999999"
    "999999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_TEMP_OUTPUT_S" MINVALUE 1 MAXVALUE 999999999999999"
    "999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_RUNTIME_PARMS_S" MINVALUE 1 MAXVALUE 9999999999999"
    "99999999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "TEMPLATE$_TARGETS_S" MINVALUE 1 MAXVALUE 99999999999999999"
    "9999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "REPCAT$_EXCEPTIONS_S" MINVALUE 1 MAXVALUE 9999999999999999"
    "99999999999 INCREMENT BY 1 START WITH 1 CACHE 20 NOORDER NOCYCLE"
    . importing OLAPSYS's objects into OLAPSYS
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "OLAP_ID_SEQ" MINVALUE 1 MAXVALUE 9999999999999999999999999"
    "99 INCREMENT BY 1 START WITH 122 NOCACHE NOORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "OLAP_IRID" MINVALUE 1 MAXVALUE 999999999999999999999999999"
    " INCREMENT BY 1 START WITH 24 NOCACHE ORDER NOCYCLE"
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "CWM2_OLAP_ENABLESEQ" MINVALUE 1 MAXVALUE 9999 INCREMENT BY"
    " 1 START WITH 1 NOCACHE ORDER CYCLE"
    I've been waiting for hours, but still it is blocked at the same place, that is:
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "CWM2_OLAP_ENABLESEQ" MINVALUE 1 MAXVALUE 9999 INCREMENT BY"
    " 1 START WITH 1 NOCACHE ORDER CYCLE"
    I've tried recreating the database, but still the error is the same.
    I've tried including the ignore=Y, that is:
    C:\DevSuiteHome_1\BIN>imp system/system@nandita1 FILE=system_21Sep2010_DVSD.DMP
    FULL=YES IGNORE=YES
    The error generated is as displayed below:
    Import: Release 10.1.0.4.2 - Production on Wed Oct 6 16:34:04 2010
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Produc
    tion
    With the Partitioning, OLAP and Data Mining options
    Export file created by EXPORT:V10.01.00 via conventional path
    import done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    . importing OLAPSYS's objects into OLAPSYS
    Over here also, i had to kill the process since it was blocked at the following statement: ". importing OLAPSYS's objects into OLAPSYS"
    Please can someone help me out.
    Thanking in advanced

    plzzzzzzzzzzzzzzzzz.............sme1 help me out

  • Error while importing a container dump file

    Hi,
    I am trying to import a container from one Oracle Server to another server which are located at two different locations. I have exported the container from the first server and FTPed it to the second server and tried to import the same into the second server. While doing this, i got the following message:
    Connected to: Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    Export file created by EXPORT:V08.01.07 via conventional path
    import done in WE8ISO8859P1 character set and WE8ISO8859P1 NCHAR character set
    . importing REPOS_MANAGER's objects into REPOS_MANAGER
    . . importing table "XTSYS_EXPORT_OBJECTS" 1 rows imported
    . . importing table "XTSYS_IMPORT_IRID_MAPPING"
    IMP-00009: abnormal end of export file
    IMP-00018: partial import of previous table completed: 1281 rows imported
    IMP-00033: Warning: Table "XTSYS_IMPORT_IVID_MAPPING" not found in export file
    IMP-00033: Warning: Table "XTSYS_TABS_EXPORTED" not found in export file
    IMP-00033: Warning: Table "XTSYS_RM$REPOSITORIES" not found in export file
    Import terminated successfully with warnings.
    I am able to import the same dump file into another server located in the same location as the Server One without any issues.
    One thing that i have observed is: On the Server 2, if a container is already present and the XTSYS_ tables are already populated, when i tried to import this dump file, it asked me "Temporary tables are already populated. Do you want to overwrite them. Otherwise the existing temporaty tables will be used for this purpose". If i clicked No, then it imported the container without any problems, BUT, the new container contents are not the same as found on Server 1's container contents.
    Can somebody suggest what could be the problem and how can i import that dump file.
    thanks a lot in advance
    regards,
    Vijay

    Vijay,
    There is a note <Note:160378.1> on the Oracle Support site: http://metalink.oracle.com.
    There are a number of possibilities, the most likely of these is that you are using the wrong import/export utilities for an older or newer database version.
    You don't mention which release of Designer you are using, Designer 9i uses the 9.0.1 export/import utilities by default. You cannot use the 9.0.1 export/import utilities against an 8.1.7 database.
    If not already installed, install the 8.1.7 export/import utilities from the Oracle 8i Release 3 (8.1.7) Client Cd into a new Oracle Home.
    In the Registry, change [hklm]\software\oracle\HOMEx\repos61\EXECUTE_EXPORT
    and EXECUTE_IMPORT to point to the 8.1.7 export/import utitilites.
    (HOMEx corresponds to your Designer 9i Oracle Home)
    A typical value for them would be:
    EXECUTE_EXPORT=d:\des_817\bin\exp.exe
    EXECUTE_IMPORT=d:\des_817\bin\imp.exe
    or
    EXECUTE_EXPORT=d:\V817\bin\exp.exe
    EXECUTE_IMPORT=d:\V817\bin\imp.exe
    After changing the parameters in the Registry, verify if they are active by going into the Repository Administration Utility, Check Requirements,
    Parameter Settings and look for EXECUTE_EXPORT and EXECUTE_IMPORT.
    Regards
    Sue

  • How to import partitioned tables in different tablespace

    Hi everyone,
    I try to import the partitioned tables in different tablespace.
    Consider the following situation:
    I have a dump file which is created by using "Export" utility. Some data are in partitioned tables, some of them are in non-partitioned tables. Also, all tables are located in "MYTBS" tablesapce. I try to import all data from this dump file to another database. I didn't get error messages when importing the data from non-partitioned tables. However, I got error message when importing the data from partitioned tables. The error message is: tablespace 'MYTBS' does not exist.
    I just want to how I can solve this problem other than create 'MYTBS' tablespace for my new database.
    Thanks in advance.
    Angel

    Hi,
    I got the following error message:
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE TABLE "FACILITYCONNECTION",....., "CONNECTIONTYPE" "
    "NUMBER(1, 0) NOT NULL ENABLE) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 25"
    "5 PARALLEL ( DEGREE DEFAULT INSTANCES 1) NOLOGGING STORAGE( PCTINCREASE 0) "
    "TABLESPACE "MYTBS" PARTITION BY RANGE ("CONNECTIONTYPE" ) (PARTITION "
    ""EXT" VALUES LESS THAN (1) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 S"
    "TORAGE(INITIAL 65536) TABLESPACE "MYTBS" NOLOGGING, PARTITION "FAC" VA"
    "LUES LESS THAN (2) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 STORAGE(I"
    "NITIAL 65536) TABLESPACE "MYTBS" NOLOGGING )"
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MYTBS' does not exist
    Thanks.
    Angel

Maybe you are looking for

  • How do I make new tabs from a first window go behind instead of in front of the first window?

    I use the same version of Firefox at work & new tabs open "behind" the initial tab, I can't figure out how to make this happen on my home computer! At home, new tabs open "in front" of the initial tab & it's really frustrating!

  • How to view gif files...

    Hi I have launched swing application in webstart. I have used some gif file in my application. While creating Jar file I have added all the gif files also. But I couln't view that gif files while running the application on the webstart. Thanks in adv

  • Adobe Acrobat 8 unable to load plug-in in Window Vista

    I have developed a plug in for Adobe Acrobat, but Abode Acrobat unable to load the plug in. I try to troubleshoot the problem by putting a messagebox in the PIHandshake function. Noticed that, the function is not being called by Adobe Acrobat when Ac

  • RFC Adapter Trace!

    Hi Everybody, I'm currently trying to debug a problem with the RFC adapter. Is there a flag that can be set that would allow me to see the actual JCO code being executed by the RFC adapter? I'm in a situation where the RFC is being executed and it's

  • Java.lang.StringIndexOutOfBoundsException: String index out of range: 3

    I have just downloaded Jdev (10.1.3.1) and am trying out the tutorial "Introduction to JavaServer Faces using JDeveloper". When I try to run the Welcome.jsp I get the following : java.lang.StringIndexOutOfBoundsException: String index out of range: 3