Sequence nextval after an schema level export import

If I export a schema that has some sequences and then truncate tables and then import the schema, do I get theh old sequence values?
I guess my question is do sequences get stored at the schema level or the database level.
I noticed that sequences are exported at the schema level when I do an export so that may be the answer but your confirmation would be appreciated.

Hi,
Nothing to worry,imp/exp does not change the value of Nextval.you can use truncate table after exp, then u can import it.
Regards
Vinay Agarwal
OCP

Similar Messages

  • Full list of schemas not exported / imported when FULL=Y

    Hi Guys,
    Verison: 10.2.0.5
    Can advise on the full list of schemas not exported / imported when FULL=Y?
    thanks!

    dbaing wrote:
    Hi Guys,
    Verison: 10.2.0.5
    Can advise on the full list of schemas not exported / imported when FULL=Y?
    thanks!System schemas. Please see the following:
    http://docs.oracle.com/cd/B19306_01/server.102/b14215/dp_export.htm#i1006790
    http://docs.oracle.com/cd/E11882_01/server.112/e22490/dp_export.htm#i1006790
    You may also be interested in the following:
    -How to Perform A Full Database Export Import During Upgrade, Migrate, Copy Or Move Of A Database [ID 286775.1]
    -Oracle DataPump Quick Start [ID 413965.1]
    -How To move Or Copy A Database Using DataPump [ID 855268.1]

  • SAP is not coming up after system refresh with Export/Import option

    Hi,
    I have exported Java 6.40 and finished the refresh successfully. And I imported the Java successfully. After importing Java, I restarted SAP but is not coming up. When I checked the dev_w0 trace file, I got the following information.
    trc file: "dev_w0", trc level: 1, release: "640"
    ACTIVE TRACE LEVEL           1
    ACTIVE TRACE COMPONENTS      all, M

    B Wed Aug 05 10:48:17 2009
    B  create_con (con_name=R/3)
    B  Loading DB library 'F:\usr\sap\IDS\SYS\exe\run\dboraslib.dll' ...
    B  Library 'F:\usr\sap\IDS\SYS\exe\run\dboraslib.dll' loaded
    B  Version of 'F:\usr\sap\IDS\SYS\exe\run\dboraslib.dll' is "640.00", patchlevel (0.220)
    B  New connection 0 created
    M sysno      00
    M sid        IDS
    M systemid   560 (PC with Windows NT)
    M relno      6400
    M patchlevel 0
    M patchno    247
    M intno      20020600
    M make:      multithreaded, Unicode
    M pid        844
    M
    M  ***LOG Q0Q=> tskh_init, WPStart (Workproc 0 844) [dpxxdisp.c   1170]
    I  MtxInit: -2 0 0
    M  DpSysAdmExtCreate: ABAP is active
    M  DpShMCreate: sizeof(wp_adm)          20328     (1452)
    M  DpShMCreate: sizeof(tm_adm)          2969176     (14772)
    M  DpShMCreate: sizeof(wp_ca_adm)          24000     (80)
    M  DpShMCreate: sizeof(appc_ca_adm)     8000     (80)
    M  DpShMCreate: sizeof(comm_adm)          290000     (580)
    M  DpShMCreate: sizeof(vmc_adm)          0     (424)
    M  DpShMCreate: sizeof(wall_adm)          (38456/34360/64/184)
    M  DpShMCreate: SHM_DP_ADM_KEY          (addr: 05FE0040, size: 3391480)
    M  DpShMCreate: allocated sys_adm at 05FE0040
    M  DpShMCreate: allocated wp_adm at 05FE1B88
    M  DpShMCreate: allocated tm_adm_list at 05FE6AF0
    M  DpShMCreate: allocated tm_adm at 05FE6B18
    M  DpShMCreate: allocated wp_ca_adm at 062BB970
    M  DpShMCreate: allocated appc_ca_adm at 062C1730
    M  DpShMCreate: allocated comm_adm_list at 062C3670
    M  DpShMCreate: allocated comm_adm at 062C3688
    M  DpShMCreate: allocated vmc_adm_list at 0630A358
    M  DpShMCreate: system runs without vmc_adm
    M  DpShMCreate: allocated ca_info at 0630A380
    M  DpShMCreate: allocated wall_adm at 0630A388
    M  ThTaskStatus: rdisp/reset_online_during_debug 0
    X  EmInit: MmSetImplementation( 2 ).
    X  <ES> client 0 initializing ....
    X  Using implementation flat
    M  <EsNT> Memory Reset disabled as NT default
    X  ES initialized.

    M Wed Aug 05 10:48:18 2009
    M  calling db_connect ...
    C  Prepending e:\oracle\IDS to Path.
    C  got NLS_LANG='AMERICAN_AMERICA.UTF8' from environment
    C  Client NLS settings: AMERICAN_AMERICA.UTF8
    C  Logon as OPS$-user to get SAPIDS's password
    C  Connecting as /@IDS on connection 0 (nls_hdl 0) ... (dbsl 640 070308)
    C  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    C    0 UTF8                                                      1   05B17398   05B1C458   05B1BEC0
    C  Attaching to DB Server IDS (con_hdl=0,svchp=05B1BE14,srvhp=05B1CA84)
    C  Starting user session (con_hdl=0,svchp=05B1BE14,srvhp=05B1CA84,usrhp=05B256B8)
    C  *** ERROR => OCI-call 'OCISessionBegin' failed: rc = 1017
    [dboci.c      4508]
    C  *** ERROR => CONNECT failed with sql error '1017'
    [dbsloci.c    11395]
    C  Try to connect with default password
    C  Connecting as SAPIDS/<pwd>@IDS on connection 0 (nls_hdl 0) ... (dbsl 640 070308)
    C  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    C    0 UTF8                                                      1   05B17398   05B1C458   05B1BEC0
    C  Starting user session (con_hdl=0,svchp=05B1BE14,srvhp=05B1CA84,usrhp=05B256B8)
    C  *** ERROR => OCI-call 'OCISessionBegin' failed: rc = 1017
    [dboci.c      4508]
    C  *** ERROR => CONNECT failed with sql error '1017'
    [dbsloci.c    11395]
    B  ***LOG BY2=> sql error 1017   performing CON [dbsh#3 @ 1204] [dbsh    1204 ]
    B  ***LOG BY0=> ORA-01017: invalid username/password; logon denied [dbsh#3 @ 1204] [dbsh    1204 ]
    B  ***LOG BY2=> sql error 1017   performing CON [dblink#3 @ 428] [dblink  0428 ]
    B  ***LOG BY0=> ORA-01017: invalid username/password; logon denied [dblink#3 @ 428] [dblink  0428 ]
    M  ***LOG R19=> tskh_init, db_connect ( DB-Connect 000256) [thxxhead.c   1283]
    M  in_ThErrHandle: 1
    M  *** ERROR => tskh_init: db_connect (step 1, th_errno 13, action 3, level 1) [thxxhead.c   9708]

    M  Info for wp 0

    M    stat = 4
    M    reqtype = 1
    M    act_reqtype = -1
    M    rq_info = 0
    M    tid = -1
    M    mode = 255
    M    len = -1
    M    rq_id = 65535
    M    rq_source = 255
    M    last_tid = 0
    M    last_mode = 0
    M    int_checked_resource(RFC) = 0
    M    ext_checked_resource(RFC) = 0
    M    int_checked_resource(HTTP) = 0
    M    ext_checked_resource(HTTP) = 0
    M    report = >                                        <
    M    action = 0
    M    tab_name = >                              <

    M  *****************************************************************************
    M  *
    M  *  LOCATION    SAP-Server ord-sapproj_IDS_00 on host ord-sapproj (wp 0)
    M  *  ERROR       tskh_init: db_connect
    M  *
    M  *  TIME        Wed Aug 05 10:48:18 2009
    M  *  RELEASE     640
    M  *  COMPONENT   Taskhandler
    M  *  VERSION     1
    M  *  RC          13
    M  *  MODULE      thxxhead.c
    M  *  LINE        9893
    M  *  COUNTER     1
    M  *
    M  *****************************************************************************

    M  PfStatDisconnect: disconnect statistics
    M  Entering TH_CALLHOOKS
    M  ThCallHooks: call hook >ThrSaveSPAFields< for event BEFORE_DUMP
    M  *** ERROR => ThrSaveSPAFields: no valid thr_wpadm [thxxrun1.c   730]
    M  *** ERROR => ThCallHooks: event handler ThrSaveSPAFields for event BEFORE_DUMP failed [thxxtool3.c  255]
    M  Entering ThSetStatError
    M  Entering ThReadDetachMode
    M  call ThrShutDown (1)...
    M  ***LOG Q02=> wp_halt, WPStop (Workproc 0 844) [dpnttool.c   357]
    I have checked database connectivity with the command 'r3trans -d', it got finished with the return code 0012 and the trace.log file  is showing the following information.
    =======================================
    4 ETW000 r3trans version 6.13 (release 640 - 17.07.08 - 08:11:00).
    4 ETW000 unicode enabled version
    4 ETW000 ===============================================
    4 ETW000
    4 ETW000 date&time   : 05.08.2009 - 10:52:31
    4 ETW000 control file: <no ctrlfile>
    4 ETW000 R3trans was called as follows: r3trans -d
    4 ETW000  trace at level 2 opened for a given file pointer
    4 ETW000  [dev trc     ,00000]  Wed Aug 05 10:52:34 2009                                  0.000000
    4 ETW000  [dev trc     ,00000]  db_con_init called                                        0.000000
    4 ETW000  [dev trc     ,00000]  create_con (con_name=R/3)                                 0.000000
    4 ETW000  [dev trc     ,00000]  Loading DB library 'dboraslib.dll' ...                    0.000000
    4 ETW000  [dev trc     ,00000]  load shared library (dboraslib.dll), hdl 0          8449  0.008449
    4 ETW000  [dev trc     ,00000]      using "f:\usr\sap\IDS\SYS\exe\run\dboraslib.dll"
    4 ETW000                                                                              37  0.008486
    4 ETW000  [dev trc     ,00000]  Library 'dboraslib.dll' loaded                        18  0.008504
    4 ETW000  [dev trc     ,00000]  function DbSlExpFuns loaded from library dboraslib.dll
    4 ETW000                                                                              27  0.008531
    4 ETW000  [dev trc     ,00000]  Version of 'dboraslib.dll' is "640.00", patchlevel (0.220)
    4 ETW000                                                                             219  0.008750
    4 ETW000  [dev trc     ,00000]  function dsql_db_init loaded from library dboraslib.dll
    4 ETW000                                                                              27  0.008777
    4 ETW000  [dev trc     ,00000]  function dbdd_exp_funs loaded from library dboraslib.dll
    4 ETW000                                                                              36  0.008813
    4 ETW000  [dev trc     ,00000]  New connection 0 created                              26  0.008839
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = -000000001 state = DISCONNECTED, perm = YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              34  0.008873
    4 ETW000  [dev trc     ,00000]  db_con_connect (con_name=R/3)                         22  0.008895
    4 ETW000  [dev trc     ,00000]  find_con_by_name found the following connection for reuse:
    4 ETW000                                                                              25  0.008920
    4 ETW000  [dev trc     ,00000]  0: name = R/3, con_id = 000000000 state = DISCONNECTED, perm = YES, reco = NO , timeout = 000, con_max = 255, con_opt = 255, occ = NO
    4 ETW000                                                                              32  0.008952
    4 ETW000  [dev trc     ,00000]  Setting DIR_CLIENT_ORAHOME is not set as environment variable,
    4 ETW000                        assuming using instant client with unspecified location.
    4 ETW000                                                                             227  0.009179
    4 ETW000  [dev trc     ,00000]  -->oci_initialize (con_hdl=0)                        130  0.009309
    4 ETW000  [dev trc     ,00000]  got NLS_LANG='AMERICAN_AMERICA.UTF8' from environment
    4 ETW000                                                                              32  0.009341
    4 ETW000  [dev trc     ,00000]  Client NLS settings: AMERICAN_AMERICA.UTF8         23827  0.033168
    4 ETW000  [dev trc     ,00000]  Logon as OPS$-user to get SAPIDS's password           34  0.033202
    4 ETW000  [dev trc     ,00000]  Connecting as /@IDS on connection 0 (nls_hdl 0) ... (dbsl 640 070308)
    4 ETW000                                                                              35  0.033237
    4 ETW000  [dev trc     ,00000]  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    4 ETW000                                                                              32  0.033269
    4 ETW000  [dev trc     ,00000]    0 UTF8                                                      1   0253CDB0   02541E18   02541880
    4 ETW000                                                                             112  0.033381
    4 ETW000  [dev trc     ,00000]  Allocating service context handle for con_hdl=0       29  0.033410
    4 ETW000  [dev trc     ,00000]  Allocating server context handle                      25  0.033435
    4 ETW000  [dev trc     ,00000]  Attaching to DB Server IDS (con_hdl=0,svchp=025417D4,srvhp=025423DC)
    4 ETW000                                                                              51  0.033486
    4 ETW000  [dev trc     ,00000]  Assigning server context 025423DC to service context 025417D4
    4 ETW000                                                                           41860  0.075346
    4 ETW000  [dev trc     ,00000]  Allocating user session handle                        47  0.075393
    4 ETW000  [dev trc     ,00000]  Starting user session (con_hdl=0,svchp=025417D4,srvhp=025423DC,usrhp=0254AFA8)
    4 ETW000                                                                              50  0.075443
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCISessionBegin' failed: rc = 1017
    4 ETW000                                                                            4381  0.079824
    4 ETW000  [dbsloci.    ,00000]  *** ERROR => CONNECT failed with sql error '1017'
    4 ETW000                                                                              31  0.079855
    4 ETW000  [dev trc     ,00000]     set_ocica() -> OCI or SQL return code 1017         23  0.079878
    4 ETW000  [dev trc     ,00000]  Try to connect with default password                  80  0.079958
    4 ETW000  [dev trc     ,00000]  Connecting as SAPIDS/<pwd>@IDS on connection 0 (nls_hdl 0) ... (dbsl 640 070308)
    4 ETW000                                                                              28  0.079986
    4 ETW000  [dev trc     ,00000]  Nls CharacterSet                 NationalCharSet              C      EnvHp      ErrHp ErrHpBatch
    4 ETW000                                                                              31  0.080017
    4 ETW000  [dev trc     ,00000]    0 UTF8                                                      1   0253CDB0   02541E18   02541880
    4 ETW000                                                                              32  0.080049
    4 ETW000  [dev trc     ,00000]  Assigning username to user session 0254AFA8           19  0.080068
    4 ETW000  [dev trc     ,00000]  Assigning password to user session 0254AFA8           25  0.080093
    4 ETW000  [dev trc     ,00000]  Starting user session (con_hdl=0,svchp=025417D4,srvhp=025423DC,usrhp=0254AFA8)
    4 ETW000                                                                              32  0.080125
    4 ETW000  [dboci.c     ,00000]  *** ERROR => OCI-call 'OCISessionBegin' failed: rc = 1017
    4 ETW000                                                                            1691  0.081816
    4 ETW000  [dbsloci.    ,00000]  *** ERROR => CONNECT failed with sql error '1017'
    4 ETW000                                                                              28  0.081844
    4 ETW000  [dev trc     ,00000]     set_ocica() -> OCI or SQL return code 1017         18  0.081862
    4 ETW000  [dblink      ,00428]  ***LOG BY2=>sql error 1017   performing CON [dblink#3 @ 428]
    4 ETW000                                                                             119  0.081981
    4 ETW000  [dblink      ,00428]  ***LOG BY0=>ORA-01017: invalid username/password; logon denied [dblink#3 @ 428]
    4 ETW000                                                                              30  0.082011
    2EETW169 no connect possible: "DBMS = ORACLE                           --- dbs_ora_tnsname = 'IDS'"
    =================================================================================
    Kindly suggest me the solution to resolve the issue.
    Thanks & regards,
    Mogileeswar

    Please check the logs it is clrerly mentioned thet your sapsystem is not able to connect to database.
    ERROR => CONNECT failed with sql error '1017' [dbsloci.c 11395] B ***LOG BY2=> sql error 1017 performing CON [dbsh#3 @ 1204] [dbsh 1204 ] B ***LOG BY0=> ORA-01017: invalid username/password; logon denied [dbsh#3 @ 1204] [dbsh 1204 ] B ***LOG BY2=> sql error 1017 performing CON [dblink#3 @ 428] [dblink 0428 ] B ***LOG BY0=> ORA-01017: invalid username/password
    Check the Note------Note 713685 - Hom./Het.System Copy SAP Web AS 6.40
    run these commands:
    brconnect -u / -c -l E -f chpass -o SAPSR3 -password whatever_you_want
    sqlplus /nolog @sapdba_role.sql SID NT
    Also check the parameter DBS_ORA_SCHEMA is set to SAPSR3.
    Also read note 400241.
    Thanks
    Rishi Abrol

  • Steps to Perform before schema level import

    Hi,
    We are planning to migrate Oracle 8i database to Oracle 10g Database.
    Approach that we have decided is export/Import.
    Can anyone tell me what all steps we have to perform before importing dmp to new database?
    We are planning to go for schema level export/import.
    Thanks in Advance
    AT

    1. Get a list of users to be exported
    select distinct owner from dba_segments where owner NOT LIKE '%SYS%'
    2. exp parfile=gen8i.prf
    vi gen8i.prf
    system/sys
    owern=( <list generated above>)
    rest of the exp paramters such as statistics, consistent
    3. imp the dump file
    4. recompile the packages
    5. done.

  • How to take  schema level  Encrypted export  dump

    Hi ,
    Pls. let me know how to take schema level export dump with encryption .
    regards
    chandrasekar.v

    This article on how to setup and maintain TDE may help you: Transparent Data Encryption
    additionaly, you may find this interesting > Transparent Data Encryption (TDE) in Oracle 10g Database Release 2
    ~ Madrid.

  • Migrate Database- Export/Import

    Hi,
    I need to migrate an Oracle database 9i from Sun Solaris to Linux. The final target database version would be 11g.
    Since this is a 9i database, I see that we have only option of export and database. We have around 15 schemas.
    I have some queries related to it.
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?
    Since there is a database version different I dont want to touch sys, system schema objects.
    Appreciate your thoughts.
    Regards
    Cherrish Vaidiyan

    Hi,
    Let me try to answer some of these questions you queried for:
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?Export won't export sys objects. For example, there are tables in sys, like obj$ that contain information for other metadata objects, like scott.emp, etc. This is not exported becuase when scott.emp is exported, the data from obj$ is essentially exported that way. When the dumpfile is imported and scott.emp is recreated, the data in sys.obj$ will be restored through the create table statement. As far as the SYSTEM schema is concerned, some objects are exported and some are not. There are tables in system that contain information about queues, jobs, etc. These will probably not make any sense on the target system so those types of tables are excluded from the export job. Other objects make sense to export/import so those are done. This is all figured out in the internals of export/import. Thre are other schemas that are not exproted. Some that I can think of are DMSYS, ORDSYS, etc. This would be for the same reason as SYS.
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?No, the dumpfiles are formatted differently. If you use exp, then you must use imp. If you use expdp, then you must use impdp. You can do exp on 9i and imp on 11g with the dumfile that was created on 9i.
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?This is case by case decision. It depends on what you want. If you want the complete database moved, then I would personally think that a full=y is what you would want to do. If you just did schema exports, then you would never export the tablespaces. This would mean that you would have to create the tablespaces on the source system before you ran imp. There are other objects that are not exported when a schema level export is performed that are exproted when a full is performed. This information can be seen in the utilities guide. Look to see what is exported when in user/schema mode vs full/database mode.
    Since there is a database version different I dont want to touch sys, system schema objects.This is all done for you with the internal workings of exp/imp.
    Dean
    Edited by: Dean Gagne on Jul 29, 2009 8:38 AM

  • Reduce export import time

    Hl All,
    I want to reduce the total time spent during export-import.
    Restrictions:
    =========
    1)Cross Platform
    2)exp and imp from lower version to upper version
    3)No 10G database
    Basically i want to do exp and imp in parallel so it reduces the total time spent on this activity. I thought of doing schema level exp-imp in parallel but i am afraid of the dependencies.
    Is there any other way to achieve the same or if i go with the above specified approach, can anyone provide some valuable inputs to that.
    I am trying to automate the above so that it becomes one time effort and rest all the times it(script) should do at his own.
    Thanks and regards
    Neeraj

    Hi All,
    Data volume is not less than 60GB and not more than 150GB.
    If i use a pipe on unix in between exp-imp what if my exp is slower at any
    point of time and import is faster(Any reason)?
    is import going to wait for the contents coming into the pipe through export or import will fail.
    I can consider creating indexes using the flat file.
    is there any way to get only the indexes in the flat file, i mean if i use indexfile
    option for import it gives me "CREATE TABLE..." statements too which means import utility is reading full
    dumpfile and giving me the output, i want only the "CREATE INDEX..." statement in the flat file.
    What about the schema level export and import?
    Any valuable inputs or proper steps from anyone out there
    Any restrictions while importing the schema's.
    Thanks and Regards

  • How do you produce both Customs Export & Import documentation in an STO?

    Hello All -
    How do you produce both Customs Export & Import documentation in an STO scenario? For instance, if a company would like to create a STO in SAP to move materials from one plant to another and would like to produce both the Export & Import paperwork in GTS, what documents would be needed in R/3 to make this possible?
    Would the following be needed, or is there some other way?
    - STO
    - Outbound Delivery & F8 Invoice to generate the Customs Shipment in GTS for the Export documentation
    - Inbound Delivery to generate the Customs Declaration in GTS for the Import documentation

    Thanks Sameer -
    Yes, for the Import documentation, I think we will be declaring prior to Goods Receipt, so we will first use the Inbound delivery, and then the GR document later on once the materials have been received. (I believe you and I were talking about some of this in another thread actually.) Thank you for confirming this.
    As for the Note, I have also seen this - but thought that this was to more or less give visibility to the sending plant up-front at STO of what export licenses might be needed. (According to the note, currently they would only be able to check if licenses were needed at the outbound delivery, and not on the PO)
    Per the note: "Given a Stock Transport Purchase Order scenario: Plant A in Germany issues a STO PO against vendor plant US10 in the US. GTS looks at the PO in terms of importing from the US (departure country) coming into Plant A in Germany and performs compliance checks. Meanwhile, exporting plant US10 only has visibility at outbound delivery creation to see if am Export License is needed. Exporting plant US10 may want to know if they need a license to the country of destination, in this case Germany, at STO PO creation."
    However with the proper country-level export/import activation of the legal regulations, I believe that you can still have the STO check export-related licenses even though it is mapped to GTS as IMPORD. I have done some testing on this and it appears to work (you just need to ensure that you are activating the relevant countries for export/import properly in the legal regulation, and you can have both export & import documents check the same legal regulation and deliver the correct license determination). I was just wondering if you had any experience with this that was contrary to what I have found, or if there is something you think I am missing..
    Thanks

  • Sequence behavior after importing via DataPump

    Hi Friends,
    I'm running Oracle DB 11.2.0.3 on Windows 2008 R2 SP1 Servers and I faced a strange sequences behavior after importing a schema via Data Pump.
    The export is done this way:
    EXPDP userid/password dumpfile= logfile= directory= remap_dumpfile=y (no news)
    The import is done this way
    IMPDP userid/password dumpfile= logfile= directory= remap_schema=(old_one:new_one) remap_tablespace=(old_ones:new_ones, so on...)
    The import works fine. There are no errors and the sequences are as well imported with no warnings.
    The strange behavior is that the sequences seems to "reset". When we call a sequence the NEXTVAL is just lower than the values already stored in the Database, and we get ORA-00001 a lot. The sequence should know that vale. I don't have this problem when using exp/imp, just via DataPump.
    So that when we create an order that should receive the value of 100, as an example, because we have 99 orders on the system, Oracle suggest a value lower than 99 or even the number one value (01).
    We then wrote a script to check the CURVAL of the sequences on the base schema to recreate the sequences using this initial value on the new imported schema.
    Does anyone faced this problem before?
    Any suggestions?
    Tks a lot

    Richard
    I've tried what you just said.
    Adding the parameter consistent=y makes Oracle to show a message like that at the beginning of the export
    "flashback_time=TO_TIMESTAMP('2013-09-03 12:18:12', 'YYYY-MM-DD HH24:MI:SS')"
    It warns me: Legacy Parameter CONSISTENT=TRUE, and replaces with flashback_time.
    Really, I did not know about this behavior with this "old" parameter. I'm very appreciated about your help.
    I was almost thinking it was a DataPump bug or something.
    Thanks a lot Richard. I'll now update my scripts and make lots of test.
    If you have more advices using this parameter please share us.
    Cheers

  • Oracle 10g - Data Pump: Export / Import of Sequences ?

    Hello,
    I'm new to this forum and also to Oracle (Version 10g). Since I could not find an answer to my question, I open this post in hoping to get some help from the experienced users.
    My question concerns the Data Pump Utility and what happens to sequences which were defined in the source database:
    I have exported a schema with the following command:
    "expdp <user>/<pass> DIRECTORY=DATA_PUMP_DIR DUMPFILE=dumpfile.dmp LOGFILE=logfile.log"
    This worked fine and also the import seemed to work fine with the command:
    "impdp <user>/<pass> DIRECTORY=DATA_PUMP_DIR DUMPFILE=dumpfile.dmp"
    It loaded the exported objects directly into the schema of the target database.
    BUT:
    Something has happened to my sequences. :-(
    When I want to use them, all sequences start again with value "1". Since I have already included data with higher values in my tables, I get into trouble with the PK of these tables because I used sequences sometimes as primary key.
    My question go in direction to:
    1. Did I something wrong with Data Pump Utility?
    2. How is the correct way to export and import sequences that they keep their actual values?
    3. When the behaviour described here is correct, how can I correct the values that start again from the last value that was used in the source database?
    Thanks a lot in advance for any help concerning this topic!
    Best regards
    FireFighter
    P.S.
    It might be that my english sounds not perfect since it is not my native language. Sorry for that! ;-)
    But I hope that someone can understand nevertheless. ;-)

    My question go in direction to:
    1. Did I something wrong with Data Pump Utility?I do not think so. But may be with the existing schema :-(
    2. How is the correct way to export and import
    sequences that they keep their actual values?If the Sequences exist in the target before the import, oracle does not drop and recreate it. So you need to ensure that the sequences do not already exist in the target or the existing ones are dropped before the import.
    3. When the behaviour described here is correct, how
    can I correct the values that start again from the
    last value that was used in the source database?You can either refresh with the import after the above correction or drop and manually recreate the sequences to START WITH the NEXT VALUE of the source sequences.
    The easier way is to generate a script from the source if you know how to do it

  • How can we take the incremental export using the datapump at schema level.

    Hi,
    How can we take the incremental export using the datapump at schema level. Example, today i had taken the full export of one schema.
    After 7 days from now , how can i take the export of change data , from the data of my full export.
    using the export datapump parameter - FLASHBACK_TIME ,can we mention the date range. Example sysdate - (sysdate -7)
    Please advice
    thanks
    Naveen.

    Think of the Data Pump Export/Import tools as taking a "picture" or "snapshot."
    When you use these utilities it exports the data as it appears (by default) while it's doing the export operation. There is no way to export a delta in comparison to a previous export.
    The FLASHBACK_TIME parameter allows you to get a consistent export as of a particular point in time.
    I recommend you go to http://tahiti.oracle.com. Click on your version and go to the Utilities Guide. This is where all the information on Data Pump is located and should answer a lot of your questions.

  • Export/Import with apps user or EUL schema owner ?

    Hi,
    I am working with a migration plan to move Discoverer 9 to Discoverer 10.
    It is an apps mode EUL, where business areas and workbooks have been granted to Apps Users & Responsibilities.
    Customer has only maintained the EUL with the database user of the EUL owner.
    What is recommended for exporting/importing the EUL from 9 to 10.
    To connect to Administrator as the EUL owner, or to connect as an apps user ( SYSADMIN ? )
    Will all the grants to apps users/responsibilities work when importing as schema owner ?
    I tried to create a entire EUL export file connected as EUL owner, and for the workbooks the export log contained :
    1234#.Workbook name1
    2345#.Workbook name 2
    1234 is the user_id for the workbook owner ( All workbooks have been created by an apps user )
    Will the import process manage to convert this user_id, and set the correct owner for the imported workbooks ?
    ( I will use the option Only take ownership if original owner not found )
    I am not sure, but I think I have seen the following syntax in other projects when exporting an apps mode EUL
    SYSADMIN.Workbook name 1
    SYSADMIN.Workbook name 2

    >
    What is recommended for exporting/importing the EUL from 9 to 10.
    To connect to Administrator as the EUL owner, or to connect as an apps user ( SYSADMIN ? )
    Will all the grants to apps users/responsibilities work when importing as schema owner ?
    Hi,
    The best that i know is that you should export using the DB user so that you will not have problem with workbooks or BA that you are not granted for (or any other grants or privileges).
    The import should be done using the APPS user (the super user you use).
    That way if you import a workbook that the owner will not be found then you will get the ownership and you can after migration deal with it.
    If you'll do that with the DB (EUL owner) user after migration you will not have access to those workbooks.
    Any way about the workbooks i suggest you'll save them as DIS files for cases that the import of the workbooks or the owner association for them fail.

  • Can i do client export import between systems with different patch levels

    we want to do a client export import between two systems. As I understand the patch levels should be the same in both for this to happen..
    But right now I am in a situation where time is a big constrant. The basis patch level is on par. But the rest are not. Can I do  client import successfully at this stage.

    The only thing that may stop it working could be additional functionality between the systems.
    I work on XI and there is a huge difference between SP12 and SP15.  We were able to do the transports between systems, but as Dev was patched before QA, and the transport was reliant on standard settings not being available, it still went though, but we had to transport it again after applying SP15 to QA.
    Again, it probably depends on what System you have running on Netweaver...

  • EXPORT at schema level, but exclude some tables within the export

    I have been searching, but had no luck in finding the correct syntax for my situation.
    I'm simply trying to export at the schema level, but I want to omit certain tables from the export.
    exp cltest/cltest01@clprod file=exp_CLPROD092508.dmp log=exp_CLPROD092508.log statistics=none compress=N
    Thanks!

    Hi,
    Think in simple first.. you use the TABLES Clause..
    Example.
    exp scott/tiger file=empdept.expdat tables=(EMP,DEPT) log=empdept.log
    In case if you scehma contains less number of tables.. !!
    Logically if you have large number of tables, I say this solutuion might work ...all around... alternative solutions to solve the problems.. If you have hundered of tables... in your schema....
    Try to Create a New Schema and using CTAS create a tables which are skippable in the Current Scehma.
    Do an Export and once the Job Done.. you recreate the backup fom New schema
    and Import to DB (Destinaiton)
    - Pavan Kumar N

  • Import specific schema from exported full database dump...

    In my application if I take backup expdp using full=y and during impdp if I want some specific schemas only then it will fine? or It recommended to specify full=y also in impdp.
    suggest me..

    In my application if I take backup expdp using full=y and during impdp if I want some specific schemas only then it will fine? or It recommended to specify full=y also in impdp.
    Hi,
    Yes it is fine.
    you have three option.
    1. full
    2. schema level
    3. table level
    it is up to you how you want to import your database.
    please refer: http://tahiti.oracle.com === import/export chapter
    Regards,
    Taj

Maybe you are looking for

  • 2lis_13_vdkon

    We are getting the following error when transporting to bi qlty. Further I find the info objects 0KAPPL, 0KINAK, 0CONDRECNO, 0kstat, 0KINAK  are missing in BI QAS system. Do we need to do activate KBETR, KNUMH, MWSK1 in R/3 as I find the standard dat

  • AOL mail error.. Firefox has detected that the server is redirecting the request for this address in a way that will never complete.

    I have cleared all cookies and cache and reset firefox to original settings.. EMAIL is still not working due to this error on AOL.. Is it aol or Firefox?

  • Adding new field in Database view

    Hi,         I am adding a new field to existing database view so only adding is enough or shall I use database utility. Please let me know. Thanks & Regards, Joseph

  • Unable to install printer driver

    I am settting up my new printer and can't complete the task. I attempt to download the software and keep getting a message that the device is not detected. Please help Thank you

  • NetStorage Single sign-on

    We've been working on trying to be able to SSO into NetStorage from uPortal for a while, and after a lot of trial, error, and Google searches, we were finally able to get it to work yesterday! Basically, we have a JSP page that you dump into your Tom