Size of database dump using DBMCLI

Hi,
Anybody know of a one liner (or more) to get the size of a database backup dump to disk using DBMCLI?
Thanks

Hi Mark,
you should be able to use the following DBMCLI syntax for both liveCache/Unix and CS/Windows:
<b>dbmcli -d <SID> -n <host> -u <dbm-user>,<dbm-pwd> -uSQL <sap-user>,<sap-pwd> sql_execute select serverdbsize from serverdbstatistics</b> [followed by Enter]
example:
dbmcli -d LC1 -u control,control -uSQL sapr3,sap sql_execute select serverdbsize from serverdbstatistics
(in the above example I omitted the '-n <host>' part, because the database resides locally)
I hope that helps you out.
Kind regards,
Roland

Similar Messages

  • Database refresh using dump and archive logs

    Hi all,
    I have a full datapump dump taken from oracle 11g R2 database (PROD) running in HP-UX server.The dump was taken on March,1,2012.
    Also i have all the archive logs till today (March 13,2012).
    I want to clone it to a new database (TEST) in windows machine with this dump, and i want to refresh(restore) the database (TEST) using this archive logs to make the database sync with the (PROD) till today.
    I need your suggestions.

    raoofdba wrote:
    Hi all,
    I have a full datapump dump taken from oracle 11g R2 database (PROD) running in HP-UX server.The dump was taken on March,1,2012.
    Also i have all the archive logs till today (March 13,2012).
    I want to clone it to a new database (TEST) in windows machine with this dump, and i want to refresh(restore) the database (TEST) using this archive logs to make the database sync with the (PROD) till today.
    I need your suggestions.I suggest you to perform transport tablespace method.below is the link: -
    http://neeraj-dba.blogspot.in/2012/01/cross-platform-transportable.html

  • Database is using too much size of the disk

    Hi all,
    I am using Oracle AS 10g and ITs database is using too much size of the disk. Because there is no space on the disk. Can anyone help me out to resolve this problem.
    Saqib

    Be sure you have enought disk space before installing (see install guide). Add disk space if you need more.

  • Dump database dbname using config

    Hi,
    By using sp_config_dump stored procedure have taken backup of database:
    sp_config_dump @config_name='sampleDB',
    @stripe_dir = '/sybase/backup' ,
    @compression = '101' ,
    @verify = 'header'
    Have dump the database using following syntax: dump database sample using config = 'sampleDB'
    Now what is the syntax for loading database as following syntax is Incorrect:
    1)load database from "dump_path" with compression
    OR
    2)load database from "compress::dump_path"

    Hi,
    I guess you are running SAP application on ASE (this is the comunity for SAP applications running on ASE), therefore the 'enable dump history' parameter is set by default, so you might find the answer about which would be the correct load command to execute using the "listonly='LOAD_SQL' " option on your load database command. This is specified on SAP Note "1611715 - SYB: How to restore an SAP ASE database server (Windows) " or SAP Note "1618817 - SYB: How to restore an SAP ASE database server (UNIX) ":
    On the "Load user databases" part (existent on both notes), you have this item:
    >>>>
    2. Load the database and transaction logs.
    Log in to ASE with isql as user 'sapsa' and execute the command:
    1> load database <DB_NAME> with listonly = 'LOAD_SQL'
    2> go
    Save the output to an SQL command file and execute it in isql.
    <<<<
    Regards,
    Victoria.

  • Exporting whole database (10GB) using Data Pump export utility

    Hi,
    I have a requirement that we have to export the whole database (10GB) using Data Pump export utility because it is not possible to send the 10GB dump in a CD/DVD to the system vendor of our application (to analyze few issues we have).
    Now when i checked online full export is available but not able to understand how it works, as we never used this data pump utility, we use normal export method. Also, will data pump reduce the size of the dump file so it can fit in a DVD or can we use Parallel Full DB export utility to split the files and include them in a DVD, is it possible.
    Please correct me if i am wrong and kindly help.
    Thanks for your help in advance.

    You need to create a directory object.
    sqlplus user/password
    create directory foo as '/path_here';
    grant all on directory foo to public;
    exit;
    then run you expdp command.
    Data Pump can compress the dumpfile if you are on 11.1 and have the appropriate options. The reason for saying filesize is to limit the size of the dumpfile. If you have 10G and are not compressing and the total dumpfiles are 10G, then by specifying 600MB, you will just have 10G/600MB = 17 dumpfiles that are 600MB. You will have to send them 17 cds. (probably a few more if dumpfiles don't get filled up 100% due to parallel.
    Data Pump dumpfiles are written by the server, not the client, so the dumpfiles don't get created in the directory where the job is run.
    Dean

  • 10g Full Database Dump Takes more than 2 Hours Still not finished

    Hi all,
    Full Database Dump Takes more than 2 Hours, Still not finished.
    Version - Oracle 10g 1.0.2.0
    Database Size is Around 160GB.
    Used Below Query to take Full Database Dump.
    expdp user/pwd@10gdb full=y directory=test_dir dumpfile=curent10g.dmp logfile=expdpcurent10g.log;
    It takes more than 1 hour in processing the functions.i.e)
    Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/FUNCTION
    And the Log File:
    Export: Release 10.1.0.2.0 - Production on Friday, 04 May, 2012 10:17
    Copyright (c) 2003, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    FLASHBACK automatically enabled to preserve database integrity.
    Starting "EXPTEST"."SYS_EXPORT_FULL_02": exptest/********@curentdb full=Y directory=test_dir dumpfile=curent10g.dmp logfile=expdpcurent10g.log;
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 4.574 GB
    Processing object type DATABASE_EXPORT/TABLESPACE
    Processing object type DATABASE_EXPORT/DE_SYS_USER/USER
    Processing object type DATABASE_EXPORT/SCHEMA/USER
    Processing object type DATABASE_EXPORT/ROLE
    Processing object type DATABASE_EXPORT/GRANT/SYSTEM_GRANT/PROC_SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/GRANT/SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/ROLE_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/DEFAULT_ROLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLESPACE_QUOTA
    Processing object type DATABASE_EXPORT/RESOURCE_COST
    Processing object type DATABASE_EXPORT/SCHEMA/DB_LINK
    Processing object type DATABASE_EXPORT/TRUSTED_DB_LINK
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/SEQUENCE
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/GRANT/DE_S_SEQ_OWNER_OBJGRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/DIRECTORY/DIRECTORY
    Processing object type DATABASE_EXPORT/DIRECTORY/GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/CONTEXT
    Processing object type DATABASE_EXPORT/SCHEMA/DE_PUBLIC_SYNONYM/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/TYPE/TYPE_SPEC
    Processing object type DATABASE_EXPORT/DE_SYSTEM_PROCOBJACT/DE_PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    Processing object type DATABASE_EXPORT/DE_SYSTEM_PROCOBJACT/DE_POST_SYSTEM_ACTIONS/PROCACT_SYSTEM
    Processing object type DATABASE_EXPORT/SCHEMA/PROCACT_SCHEMA
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/PRE_TABLE_ACTION
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/GRANT/DE_S_TABLE_OWNER_OBJGRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/GRANT/DE_S_TABLE_NOTWGO_OBJGRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/INDEX
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/STATISTICS/TABLE_STATISTICS
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/COMMENT
    Processing object type DATABASE_EXPORT/SCHEMA/PACKAGE/PACKAGE_SPEC
    Processing object type DATABASE_EXPORT/SCHEMA/PACKAGE/GRANT/DE_S_PACKAGE_OWNER_OBJGRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/FUNCTION
    Help me.

    Export is Still in progress
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Job: SYS_EXPORT_FULL_02
    Owner: EXPTEST
    Operation: EXPORT
    Creator Privs: FALSE
    GUID: 94AB0C8AA824462AA96C8481CF0F676F
    Start Time: Friday, 04 May, 2012 10:18
    Mode: FULL
    Instance: curentdb
    Max Parallelism: 1
    EXPORT Job Parameters:
    Parameter Name Parameter Value:
    CLIENT_COMMAND userexp/********@curentdb full=Y directory=test_dir dumpfile=curent10g.dmp logfile=expdpcurent10g.log;
    DATA_ACCESS_METHOD AUTOMATIC
    ESTIMATE BLOCKS
    INCLUDE_METADATA 1
    LOG_FILE_DIRECTORY TEST_DIR
    LOG_FILE_NAME expdpcurent10g.log;
    TABLE_CONSISTENCY 0
    State: EXECUTING
    Bytes Processed: 0
    Current Parallelism: 1
    Job Error Count: 0
    Dump File: C:\ORACLE DMP\CURENT10G.DMP
    bytes written: 4,096
    Worker 1 Status:
    State: EXECUTING
    Object Schema: PUBLIC
    Object Type: DATABASE_EXPORT/SCHEMA/PACKAGE/GRANT/DE_S_PACKAGE_OWNER_OBJGRANT/OBJECT_GRANT
    Completed Objects: 1
    Total Objects: 1
    What Does it mean, Suggest.

  • Cannot complete dump using db1_dump185

    We have an older BerkeleyDB database file of 22810624 bytes. When we dump using:
    ./db1_dump185 -p dbfile
    it ends with the following error
    db1_dump185: seq: Cannot allocate memory
    We were using DB_File perl module to read/write the database but either it became corrupted or reached some size limit and now we can't access all the data.
    Can anyone help with suggesting as to how to retrieve all the information in the database?

    We have an older BerkeleyDB database file of 22810624 bytes. When we dump using:
    ./db1_dump185 -p dbfile
    it ends with the following error
    db1_dump185: seq: Cannot allocate memory
    We were using DB_File perl module to read/write the database but either it became corrupted or reached some size limit and now we can't access all the data.
    Can anyone help with suggesting as to how to retrieve all the information in the database?

  • Error while taking dump using datapump

    getting following error -
    Export: Release 10.2.0.1.0 - Production on Friday, 15 September, 2006 10:31:41
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Starting "XX"."SYS_EXPORT_SCHEMA_02": XX/********@XXX directory=dpdump dumpfile=XXX150906.dmp logfile=XXX150906.log
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    ORA-39125: Worker unexpected fatal error in KUPW$WORKER.GET_TABLE_DATA_OBJECTS while calling DBMS_METADATA.FETCH_XML_CLOB []
    ORA-31642: the following SQL statement fails:
    BEGIN "DMSYS"."DBMS_DM_MODEL_EXP".SCHEMA_CALLOUT(:1,0,0,'10.02.00.01.00'); END;
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_METADATA", line 907
    ORA-06550: line 1, column 7:
    PLS-00201: identifier 'DMSYS.DBMS_DM_MODEL_EXP' must be declared
    ORA-06550: line 1, column 7:
    PL/SQL: Statement ignored
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 6235
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    2A68E610 14916 package body SYS.KUPW$WORKER
    2A68E610 6300 package body SYS.KUPW$WORKER
    2A68E610 9120 package body SYS.KUPW$WORKER
    2A68E610 1880 package body SYS.KUPW$WORKER
    2A68E610 6861 package body SYS.KUPW$WORKER
    2A68E610 1262 package body SYS.KUPW$WORKER
    255541A8 2 anonymous block
    Job "XX"."SYS_EXPORT_SCHEMA_02" stopped due to fatal error at 10:33:12
    Action required is contact customer support. And on metalink found a link that states it a bug in 10g release 1 that was suppose to be fixed in 10g release 1 version 4.
    some of the default schemas were purposely dropped from the database. The only default schema available now are -
    DBSNMP, DIP, OUTLN, PUBLIC, SCOTT, SYS, SYSMAN, SYSTEM, TSMSYS.
    DIP, OUTLN, TSMSYS were created again.
    Could this be a cause of problem??
    Thanks in adv.

    Hi,
    Below is the DDL taken from different database. Will this be enough ? One more thing please, what shall be the password should it be DMSYS.....since this will not be used by me but system.
    CREATE USER "DMSYS" PROFILE "DEFAULT" IDENTIFIED BY "*******" PASSWORD EXPIRE DEFAULT TABLESPACE "SYSAUX" TEMPORARY TABLESPACE "TEMP" QUOTA 204800 K ON "SYSAUX" ACCOUNT LOCK
    GRANT ALTER SESSION TO "DMSYS"
    GRANT ALTER SYSTEM TO "DMSYS"
    GRANT CREATE JOB TO "DMSYS"
    GRANT CREATE LIBRARY TO "DMSYS"
    GRANT CREATE PROCEDURE TO "DMSYS"
    GRANT CREATE PUBLIC SYNONYM TO "DMSYS"
    GRANT CREATE SEQUENCE TO "DMSYS"
    GRANT CREATE SESSION TO "DMSYS"
    GRANT CREATE SYNONYM TO "DMSYS"
    GRANT CREATE TABLE TO "DMSYS"
    GRANT CREATE TRIGGER TO "DMSYS"
    GRANT CREATE TYPE TO "DMSYS"
    GRANT CREATE VIEW TO "DMSYS"
    GRANT DROP PUBLIC SYNONYM TO "DMSYS"
    GRANT QUERY REWRITE TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_JOBS_RUNNING" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_REGISTRY" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_SYS_PRIVS" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_TAB_PRIVS" TO "DMSYS"
    GRANT SELECT ON "SYS"."DBA_TEMP_FILES" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_LOCK" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_REGISTRY" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_SYSTEM" TO "DMSYS"
    GRANT EXECUTE ON "SYS"."DBMS_SYS_ERROR" TO "DMSYS"
    GRANT DELETE ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT INSERT ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT SELECT ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT UPDATE ON "SYS"."EXPDEPACT$" TO "DMSYS"
    GRANT SELECT ON "SYS"."V_$PARAMETER" TO "DMSYS"
    GRANT SELECT ON "SYS"."V_$SESSION" TO "DMSYS"
    The other database has the DMSYS and the status is EXPIRED & LOCKED but I'm still able to take the dump using datapump??

  • You want to know the amount of space the transaction log for the Customer database is using. Which T-SQL command would you use?

    You want to know the amount of space the transaction log for the Customer database is using. Which T-SQL command would you use?

    Forced me to do a little research.
    DBCC SQLPERF(logspace)
    See also
    http://stackoverflow.com/questions/198343/how-can-i-get-the-size-of-the-transaction-log-in-sql-2005-programmatically
    For every expert, there is an equal and opposite expert. - Becker's Law
    My blog
    My TechNet articles

  • Database Instance Using Swap Space even When There is Plenty of Free RAM

    Hi,
    I am facinga strange issue in oracle on solaris 5.10
    Platform:solaris 5.10
    Database version:11.2.0.1.0
    Our database Instance Using Swap Space even When There is Plenty of Free RAM.
    We have 80GB of physical RAM, 32GB is set to oracle SGA_MAX_SIZE
    memory parameters:
    memory_max_target big integer 32G
    memory_target big integer 32G
    sga_max_size big integer 32G
    sga_target big integer 32G
    but still oracle is taking memory from swap even though we have enough physical memory.
    please find the top output
    load averages: 0.09, 0.18, 0.84 01:07:49
    134 processes: 133 sleeping, 1 on cpu
    CPU states: 99.4% idle, 0.4% user, 0.3% kernel, 0.0% iowait, 0.0% swap
    Memory: 80G real, 69G free, 35G swap in use, 49G swap free
    PID USERNAME LWP PRI NICE SIZE RES STATE TIME CPU COMMAND
    1052 oracle 13 59 0 95M 66M sleep 13:16 0.07% oraagent.bin
    10256 oracle 1 59 0 32G 835M sleep 0:00 0.03% oracle
    1227 oracle 15 59 0 78M 56M sleep 4:27 0.02% ocssd.bin
    9524 vector1 112 59 0 1704M 776M sleep 45:04 0.02% java
    988 oracle 38 59 0 119M 92M sleep 3:00 0.02% ohasd.bin
    1297 oracle 1 101 -20 453M 390M sleep 2:47 0.02% oracle
    1765 oracle 1 101 -20 32G 329M sleep 2:45 0.02% oracle
    10258 oracle 1 59 0 32G 345M sleep 0:00 0.02% oracle
    10259 oracle 1 59 0 3400K 2024K cpu/20 0:00 0.01% top
    1777 oracle 1 59 0 32G 342M sleep 1:14 0.01% oracle
    1246 oracle 6 59 0 56M 44M sleep 1:22 0.01% diskmon.bin
    1803 oracle 1 59 0 32G 514M sleep 0:30 0.01% oracle
    1215 oracle 15 59 0 71M 46M sleep 1:12 0.00% cssdagent
    1307 oracle 1 59 0 453M 390M sleep 0:45 0.00% oracle
    1217 oracle 11 59 0 81M 49M sleep 0:35 0.00% orarootagent.bi
    Vmstat output:
    $vmstat 5 5
    kthr memory page disk faults cpu
    r b w swap free re mf pi po fr de sr s3 s3 s3 s3 in sy cs us sy id
    0 0 0 53698448 73900568 20 81 32 6 5 0 1 -0 0 1 -0 2235 1447 1984 1 0 99
    0 0 0 51196160 72037552 1 13 0 0 0 0 0 0 0 0 0 2505 1403 2204 0 0 100
    0 0 0 51193488 72035864 55 380 0 0 0 0 0 0 0 0 0 2487 2143 2203 0 1 99
    0 0 0 51183856 72030176 0 0 0 3 3 0 0 0 0 0 0 2496 1370 2182 0 0 99
    0 0 0 51182648 72029112 22 117 0 0 0 0 0 0 0 0 0 2503 1408 2193 1 0 99
    $cat /etc/release
    Solaris 10 10/09 s10s_u8wos_08a SPARC
    Copyright 2009 Sun Microsystems, Inc. All Rights Reserved.
    Use is subject to license terms.
    Assembled 16 September 2009
    I refered the metalink document id: [ID 761960.1]
    Database Instance Using Swap Space When There is Plenty of Free RAM metalink note [ID 761960.1]
    and tried the disabling the sga_max_size and memory_max_target and memory target with all different combination but no use.
    We have one more box where same setup was configured in solaris platform but there the above memory parameters are not disabled but it is not using swap space it takes memry only from physical RAM
    Can any one throw some lights on this issue. It wopuld be great if I come to know where the exact problem
    Thanks in advance
    With Regards,
    Boobathi P

    Hi,
    I didn't say I am not using 11gr2 memory parameters I am using all the automatic memory parameters but the thinks is as per the metalink note I tried by disabling the automatic memory parameters. but that didn't help me in preventing unnecessary swap usage. So I reverted. now it is using automatic memory parameters.
    so What is the solution for this?
    With Regards,
    Boo

  • How to determine who performed the last full database dump before cumulative dump

    Hello,
    ASE 15.7 SP100 allows cumulative backups, and if cumulative dump is tried before a full dump, ASE shows this error:
    You cannot execute an incremental dump before a full database dump.
    From my experiments, it seems that it does not matter how the full dump is performed (by native isql, or by a 3rd party API). As long as a full dump is done on a database, ASE seems to be keeping track of all the changed pages since the last full dump. So this means that you can perform a full dump using a 3rd party API, and then on isql, if you run a cumulative dump command, the cumulative dump succeeds, which is based on the last full dump by another library!
    So, my question is: is there any way to programmatically determine how (by native isql or 3rd party) the last full backup was performed? I believe $SYBASE_HOME/ASE-15_0/dumphist contains this info, but it requires 'enable dump history' to be set first, and I am looking for a solution which does not involve checking a disk file.
    Thanks,
    Ali

    Dear Mr Razib,
    I have not explored the feature but ASE autiding might provide you the possibility to access information on past database dumps via SQL.
    Apart from that - I am not aware of an SQL interface to the dumphist file (would be nice to have, I agree)
    Enabling dump history is definitley highly recommended  (in my opinion) .
    There is yet another feature which might help you to prevent an DBA from dumping databases and transactions to various locations.
    When you create a DUMP CONFIGURATION and additionally set parameter
            enforce dump configuration
    ASE will prevent normal (free style) DUMP commands but enforce the use of an existing dump configuration. The mechanism is not fool proof (nothing prevents from creating yet another dump configuration on the fly) - but at least something.
    With kind regards
    Tilman Model-Bosch

  • Export/Import full database dump fails to recreate the workspaces

    Hi,
    We are studying the possibility of using Oracle Workspace to maintain multiple versions of our business data. I recently did some tests with the import/export of a dump that contains multiple workspaces. I'm not able to import the dump successfully because it doesn't recreate the workspaces on the target database instance, which is an empty Oracle database freshly created. Before to launch the import with the Oracle Data Pump utility, I have the "LIVE" workspace that exists in the WMSYS.WM$WORKSPACES_TABLE table. After the import is done, there is nothing in that table...
    The versions of the Oracle database and Oracle Workspace Manager are the same on source and target database:
    Database version (from the V$VERSION view):
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    CORE     10.2.0.4.0     Production
    TNS for 64-bit Windows: Version 10.2.0.4.0 - Production
    NLSRTL Version 10.2.0.4.0 - Production
    Workspace Manager version (from the WM_INSTALLATION view):
    OWM_VERSION: 10.2.0.4.3
    In order to recreate the tablespaces successfully during the full import of the dump, the directory structure for the tablespaces is the same on the target database's computer. I used the instructions given in this document, section "1.6 Import and Export Considerations" at page 1-19:
    http://www.wyswert.com/documentation/oracle/database/11.2/appdev.112/e11826.pdf
    Considering that the release of Oracle database used is version 10.2.0.4, I use the following command to import the dump since it doesn't contain the WMSYS schema:
    impdp system/<password>@<database> DIRECTORY=data_pump_dir DUMPFILE=expfull.dmp FULL=y TABLE_EXISTS_ACTION=append LOGFILE=impfull.log
    The only hints I have in the import log file are the following block of lines:
    1st one:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Echec de la création du type d'objet PROCACT_SYSTEM avec erreur :
    ORA-01843: ce n'est pas un mois valide
    SQL en échec :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Echec de la création du type d'objet PROCACT_SCHEMA avec erreur :
    ORA-20000: Workspace Manager Not Properly Installed
    SQL en échec :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    The target operating system is in french... here is a raw translation:
    1st one:
    ==============================
    Treatment of object type DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Object type creation failed PROCACT_SYSTEM with error :
    ORA-01843: invalid month
    failed SQL :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Treatment of object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Object type creation failed PROCACT_SCHEMA with error :
    ORA-20000: Workspace Manager Not Properly Installed
    failed SQL :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    By the way, the computer of the source database is Vista 64 in english and the destination computer is Vista 64 in french, do you think this has anything to do with these error messages? The parameters of the NLS_SESSION_PARAMETERS view are the same in the source and target database...
    Thanks in advance for your help!
    Joel Autotte
    Lead Developer
    Edited by: 871007 on Jul 13, 2011 7:31 AM

    I tried to import the full database dump with the "imp" command and I had the same error message with more detail:
    . Import d'objets SYSTEM dans SYSTEM
    IMP-00017: Echec de l'instruction suivante avec erreur ORACLE 1843 :
    "BEGIN "
    "if (system.wm$_check_install) then"
    " return ;"
    " end if ;"
    " begin execute immediate 'insert into wmsys.wm$workspaces_ta"
    "ble"
    " values(''LIVE'',"
    " ''194'',"
    " ''-1'',"
    " ''SYS'',"
    " *to_date(''09-DEC-2009 00:00:00'', ''DD-MON-YYYY HH"*
    *"24:MI:SS''),"*
    " ''0'',"
    " ''UNLOCKED'',"
    " ''0'',"
    " ''0'',"
    " ''37'',"
    " ''CRS_ALLNONCR'',"
    " to_date(''06-APR-2011 14:24:57'', ''DD-MON-YYYY HH"
    "24:MI:SS''),"
    " ''0'',"
    " '''')'; end;"
    "COMMIT; END;"
    IMP-00003: Erreur ORACLE 1843 rencontrée
    ORA-01843: ce n'est pas un mois valide
    ORA-06512: à ligne 5
    The call to the TO_DATE function with the string of the date format is incompatible with the format and language set in the NLS parameters of my database. Since I know that the "imp" and "exp" commands work client side (unlike "impdp" and "expdp"), I did a full export of the source database from the computer on which I try to import the dump to have the same language exported in the dump and I was able to import the dump successfully.
    But, I would like to use the Data Pump tool instead of the old import and export tools. How would it be possible to set the NLS parameters that "impdp" must use before to launch the import of the dump?
    The NLS parameters are set to the "AMERICAN" settings in both source and destination database when I validate that with the "NLS_DATABASE_PARAMETERS" view. The dump I did with "expdp" on the source database exported the data with the "AMERICAN" settings as well...
    On the destination database, the computer language is in french and I guess it is considered by the Data Pump import tool since it outputs the log in the language of the computer and it tries to do a TO_DATE with the wrong date format because it works with the "CANADIAN FRENCH" settings...
    Any suggestions? I'll continue to read on my side and re-post if I find the solution.
    Thanks!
    Edited by: 871007 on Jul 19, 2011 8:42 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM

  • Import  a database dump

    Hi all,
    I have tow database in my oracle server 10g , mydb1 and mudb2 , and have tow database dumps ,mydb1.dmp and mydb2.dmp. I learned from other forums that to be able to import the dump file i need that the specific database should be started and up . My question is how to start one database mydb1 and in the same time the other database mydb2 is down because by default they are started at the same time .
    BR

    Hi,
    Wat is the OS Are you using?
    I am just guessing it should be windows, Whenever windows startup there is a registry entry
    HKEY_LOCAL_MACHINE\SOFTWARE\ORACLE\oracle_home_name.
    There will be a key
    ORA_SID_AUTOSTART. SID is your database SID.
    If it is set to true, the databases will start autoamtically, you can change it to false if you dont need it.
    Or if you navigate to the services.msc you can find the ORACLESERVICESID where you have two types of start parameters 1) Automatic (whenver windows starts it will automatically start the db)
    2) Manual (You can start the DB whenever you need it)
    To start database connect to the sql prompt and issue STARTUP to shutdown the db issue SHUTDOWN ON sql prompt.
    Regards,
    Vijayaraghavan K

  • From a complete database dump, I need to import 8 schemas

    I have a complete database dump made with datapump on oracle 10g on Linux.
    I need to import only 8 schemas.
    In my parfile I have to use the INCLUDE option?
    How can I tell oracle to import only these 4 schemas?

    Hi,
    What is the version you are using.. ??
    In my parfile I have to use the INCLUDE option?
    As per yu requirment it might not fit, I suppose
    The INCLUDE and EXCLUDE parameters can be used to limit the export/import to specific objects.
    Second thing is that when you use INCLUDE parameter only those objects specified by it will be included and EXCLUDE will make sure that all objects except those specified by it will be included in the export.
    How can I tell oracle to import only these 4 schemas?
    Better I suggest that import the full dB in to some test DB, then do an schema level export with respective your requirement.
    As, sybrand had pointed the correct point of SCHEMAS, which I missed out..
    - Pavan Kumar N
    Edited by: Pavan Kumar on Nov 13, 2008 4:26 PM

  • Increase the size of the cache using the cache.size= number of pages ?

    Hi All,
    I am getting this error when I do load testing.
    I have Connection pool for Sybase database that I am using in my JPD. I am using Database control of weblogic to call the Sybase Stored procedure.
    I got following exception when I was doing load testing with 30 concurrent users.
    Any idea why this exception is coming ?
    thanks in advance
    Hitesh
    javax.ejb.EJBException: [WLI-Core:484047]Tracking MDB failed to acquire resources.
    java.sql.SQLException: Cache Full. Current size is 2069 pages. Increase the size of the cache using the cache.size=<number of pages>
         at com.pointbase.net.netJDBCPrimitives.handleResponse(Unknown Source)
         at com.pointbase.net.netJDBCPrimitives.handleJDBCObjectResponse(Unknown Source)
         at com.pointbase.net.netJDBCConnection.prepareStatement(Unknown Source)
         at weblogic.jdbc.common.internal.ConnectionEnv.makeStatement(ConnectionEnv.java:1133)
         at weblogic.jdbc.common.internal.ConnectionEnv.getCachedStatement(ConnectionEnv.java:917)
         at weblogic.jdbc.common.internal.ConnectionEnv.getCachedStatement(ConnectionEnv.java:905)
         at weblogic.jdbc.wrapper.Connection.prepareStatement(Connection.java:350)
         at weblogic.jdbc.wrapper.JTSConnection.prepareStatement(JTSConnection.java:479)
         at com.bea.wli.management.tracking.TrackingMDB.getResources(TrackingMDB.java:86)
         at com.bea.wli.management.tracking.TrackingMDB.onMessage(TrackingMDB.java:141)
         at com.bea.wli.management.tracking.TrackingMDB.onMessage(TrackingMDB.java:115)
         at weblogic.ejb20.internal.MDListener.execute(MDListener.java:370)
         at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:262)
         at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2678)
         at weblogic.jms.client.JMSSession.execute(JMSSession.java:2598)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:219)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:178)

    hitesh Chauhan wrote:
    Hi All,
    I am getting this error when I do load testing.
    I have Connection pool for Sybase database that I am using in my JPD. I am using Database control of weblogic to call the Sybase Stored procedure.
    I got following exception when I was doing load testing with 30 concurrent users.
    Any idea why this exception is coming ?
    thanks in advance
    Hitesh Hi. Please note below, the stacktrace and exception is coming from the
    Pointbase DBMS, nothing to do with Sybase. It seems to be an issue
    with a configurable limit for PointBase, that you are exceeding.
    Please read the PointBase configuration documents, and/or configure
    your MDBs to use Sybase.
    Joe
    >
    javax.ejb.EJBException: [WLI-Core:484047]Tracking MDB failed to acquire resources.
    java.sql.SQLException: Cache Full. Current size is 2069 pages. Increase the size of the cache using the cache.size=<number of pages>
         at com.pointbase.net.netJDBCPrimitives.handleResponse(Unknown Source)
         at com.pointbase.net.netJDBCPrimitives.handleJDBCObjectResponse(Unknown Source)
         at com.pointbase.net.netJDBCConnection.prepareStatement(Unknown Source)
         at weblogic.jdbc.common.internal.ConnectionEnv.makeStatement(ConnectionEnv.java:1133)
         at weblogic.jdbc.common.internal.ConnectionEnv.getCachedStatement(ConnectionEnv.java:917)
         at weblogic.jdbc.common.internal.ConnectionEnv.getCachedStatement(ConnectionEnv.java:905)
         at weblogic.jdbc.wrapper.Connection.prepareStatement(Connection.java:350)
         at weblogic.jdbc.wrapper.JTSConnection.prepareStatement(JTSConnection.java:479)
         at com.bea.wli.management.tracking.TrackingMDB.getResources(TrackingMDB.java:86)
         at com.bea.wli.management.tracking.TrackingMDB.onMessage(TrackingMDB.java:141)
         at com.bea.wli.management.tracking.TrackingMDB.onMessage(TrackingMDB.java:115)
         at weblogic.ejb20.internal.MDListener.execute(MDListener.java:370)
         at weblogic.ejb20.internal.MDListener.onMessage(MDListener.java:262)
         at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2678)
         at weblogic.jms.client.JMSSession.execute(JMSSession.java:2598)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:219)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:178)

Maybe you are looking for

  • How can I get my iPad screen back on?

    I cant get my iPad screen to turn on. I have tried holding the home and sleep button down and tried 'restoring' the iPad but it still not working and nothing appears on the screen. It is charged. I was having trouble before wtih watching a video. Whe

  • IMac G5 doesn't go past grey screen with apple and spinning gear

    I have an iMac G5 (Ambient Light Sensor) running Mac OS X 10.5.2 Leopard. Last night my computer froze while using QuickTime. I had to force shutdown the computer by holding the power button for 5 seconds after everything else failed. Now it won't st

  • Change logs for query...........

    Hi, How to check the logs for query changes. Means which user has done what change.In properties we get the last user only. Is there any way to check the full history of changes? Can we change that in some transport request? Bye, Jeetu

  • How is it best to develope a multiple window (front panel) application in LabVIEW?

    I have only been using LabVIEW for a couple months so keep this in mind while I ask this question. I have been testing out multiple ways of developing a application using multiple windows and passing data between them, in LabVIEW. I am not sure how t

  • Abap function module http_post and umlaut characters

    I am using abap function module HTTP_POST to frame xml and send the data to our middleware system 'CASTIRON'. Everything works fine but when there are any umlaut characters in xml, the function HTTP_POPST is unable to transmit the data and is creatin