Delete dump file
Hi experts,
Can i delete dump files through SQL command
Yes, to be more precise, "host" is a SQL*Plus command, not a SQL command.
Similar Messages
-
Hello,
i am taking backups everyday on windows and that backup is generating dump file.
so i have manually delete that dump file everyday.
is there any way so that i can set a job which will automatically delete dumpfiles??
thanks in advanceI'm not sure what you are referring to as a dump file is this possibly a transcript or log file created by OSB? If this is something created outside of OSB you could invoke an after script inside the dataset used for the backup.
-
User can't delete Dump file (data pump)
when I do expdp in 10.2.0.4, the dump file is created but the user don't have permission to delete those dmp files. Every time i need to login as 'oracle' user then only I can delete it. But I want that user should do it by itself.
Please let me know if there is any way to create a dump file by user who can delete it by his own.Either you may need to follow as I suggested above or as below.
Create the specific directory for each OS user who is using expdp utility and create the directories in Oracel (SQL), assign this directory to the database user.
i.e.
OS User - Scott
Database User - Scott
login as super user
$ cd /u01/oradata/expdp
$ mkdir scott
$ chmod 700 scott
Login to sqlplus as sysdba
SQL> conn /as sysdba
SQL> create or replace directory scott_dmp as '/u01/oradata/expdp/scott';
SQL> grant read,write on directory scott_dmp to scott;
Regards,
Sabdar Syed. -
Re: User can't delete Dump file (data pump)
Hi,
You could possibly set the gid bit on the datapump directory you are using?
http://en.wikipedia.org/wiki/Setuid
so basically chmod g+s your_unix_path
That should mean that the files are created with the same groupas the parent directory not the user who created them - would that work?
Regards,
Harry
http://dbaharrison.blogspot.comEither you may need to follow as I suggested above or as below.
Create the specific directory for each OS user who is using expdp utility and create the directories in Oracel (SQL), assign this directory to the database user.
i.e.
OS User - Scott
Database User - Scott
login as super user
$ cd /u01/oradata/expdp
$ mkdir scott
$ chmod 700 scott
Login to sqlplus as sysdba
SQL> conn /as sysdba
SQL> create or replace directory scott_dmp as '/u01/oradata/expdp/scott';
SQL> grant read,write on directory scott_dmp to scott;
Regards,
Sabdar Syed. -
DUMP FILES UNABLE TO DELETE IN LINUX
Hi
Our company currently runs the databases on RAC (Oracle 11gR2). We generally take the backup of SCHEMAS everyday using the datapump. My issues here is I am unable to delete the dump files older that 3 days using the below script
00 14 * * * find /nbds2/orabkp/*.dmp.gz -mtime +3 -exec rm -f {} \;
but I am able to delete the files using the below command
rm -f abcb.dmp.gz . I verified the file permissions etc on the dump file that is generated everything looks fine.
Even I tried using the below script
#################cut here - rmdumpfiles #############
#!/bin/bash
echo "starting at `date`"
for i in `find /nbds2/orabkp -mtime +3 -name "*.dmp.gz" `
do
echo "deleting $i "
rm -f $i
done
echo "done at `date` "
chmod 750 rmdumpfiles
Crontab entry:
00 14 * * * /home/oracle/rmdumpfiles >> /home/oracle/rmdump.log
But the files didn't get deleted and below is the information I got from the log file
"starting at Mon Feb 18 17:59:01 PST 2013
done at Mon Feb 18 17:59:01 PST 2013"
Can someone help me please
Thank you
KarrenHi Dude,
Thank you very much for your quick response. Actually I had to delete a lot of files manually due to storage issue( I am left with only the dump files from of 17,18,19). But my results are weird. I am not able to find the issue. Can you help me
stat /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_17_2013.dmp.gz
File: `/nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_17_2013.dmp.gz'
Size: 7013910756 Blocks: 13753000 IO Block: 65536 regular file
Device: 16h/22d Inode: 21604 Links: 1
Access: (0640/-rw-r-----) Uid: ( 1001/ oracle) Gid: ( 1001/oinstall)
Access: 2013-02-17 18:00:04.000000000 -0800
Modify: 2013-02-17 18:21:11.000000000 -0800
Change: 2013-02-17 18:26:41.988556000 -0800
find /nbds2/orabkp/* -iname "*.dmp.gz" -mtime +3
Ans: None
find /nbds2/orabkp/* -iname "*.dmp.gz" -mtime +2
Ans: None
find /nbds2/orabkp/* -iname "*.dmp.gz" -mtime +1
Ans: /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_17_2013.dmp.gz
ls -l /nbds2/orabkp/FULL_EXP_FSDEV_*
-rw-r----- 1 oracle oinstall 7013910756 Feb 17 18:21 /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_17_2013.dmp.gz
-rw-r--r-- 1 oracle asmadmin 75805 Feb 17 18:21 /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_17_2013.log
-rw-r--r-- 1 oracle asmadmin 75805 Feb 18 18:21 /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_18_2013.log
-rw-r----- 1 oracle oinstall 7014081297 Feb 19 18:20 /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_19_2013.dmp.gz
-rw-r--r-- 1 oracle asmadmin 75805 Feb 19 18:20 /nbds2/orabkp/FULL_EXP_FSDEV_expdp_Feb_19_2013.log
date
Ans: Wed Feb 20 09:27:02 PST 2013
Now I am able to delete the files thank you very much for your help.This is the first time I am having this issue I am not sure the reason behind it "date" and "mtime" are conflicting and that is causing the issue. Can you please enlighten me regarding this
Edited by: 988802 on Feb 20, 2013 9:42 AM
Edited by: 988802 on Feb 20, 2013 10:04 AM -
System exception while deleting the file from app server in background job
Hi All,
I have a issue while the deleting the file from application server.
I am using the statement DELETE DATASET in my program to delete the file from app server.
I am able to delete the file from the app server when i run the program from app server.
When i run the same report from background job i am getting the message called System exception.
Is there any secuirity which i need to get the issue.
Thank You,
TaraginiHi All,
I get all the authorization sto delete the file from application serever.
Thing is i am able to run the program sucessfully in foreground but not in the background .
It i snot giving any short dump also just JOB is cancelled with the exception 'Job cancelled after system exception ERROR_MESSAGE'.
Can anybody please give me suggestion
Thanks,
Taragini -
How can I delete PDF files from my iMac desktop?
How can I delete PDF files from my iMac desktop?
Not on my system,When I try to dump if off in the trash it will not go. I get a message that says this file cannot be removed. I also tried going to File and clicking on Move to Trash. No dice.
-
Problem in importing dump file in Oracle 10g
Hi all,
I have a database dump with size of 1.8 GB. One of the tables in dump (TABLEPHOTO) is around 1.5 GB and has a BLOB type column.
The import process stops (gets frozen) after importing 4294 rows of TABLEPHOTO.
I do not receive any error and when checking database resources in EM, do not see anything unususl.
The dump file can not be exported again since the schema is deleted!
Can anybody help with this issue?
Regards,
AmirWhat does Oracle wait interface say when the session appears to be frozen ?
select sid,username,status,state,event,p1,p2 from v$session_wait where sid=<sid>;You may also want to use Tanel Poder's Snapper to see the session activity at that time:
@snapper ash,stats 5 1 <sid>http://www.tanelpoder.com/files/scripts/snapper.sql
http://tech.e2sn.com/oracle-scripts-and-tools/session-snapper -
*** DUMP FILE SIZE IS LIMITED TO 0 BYTES ***
I always get a trace file with
*** DUMP FILE SIZE IS LIMITED TO 0 BYTES ***Do you have a cleardown script on your dump directory (probably user dump) that deletes dumps while sessions are still running. You get a trace in background dump directory (pmon, I think) that gives your message when that happens.
OTOH it might be something completely different. -
Export Dump File without defined VPD policies
Hi there,
Can some one help me on this problem ?
I've some tables with vpd policies defined on them. When I export these tables the dump file puts some instructions related to these policies such as:
EXECUTE DBMS_RLS.ADD_GROUPED_POLICY(sys_context('userenv','current_schema'),'"TABLE1"','SYS_DEFAULT','POLICY1','MD','ACESS1','SELECT,UPDATE,DELETE',FALSE,TRUE,TRUE);
Then when I try to import those tables, I'm getting some problems with that if I don't have those policies/functions created (which is the case).
So,
how can I export tables without those instructions ?
Or
how can I prevent the execution of those lines during import task ?
Thanks in advance,
HelenaIn both (export and import) cases the policy could also be disabled for the duration export and import.
-
Dbms_datapupm, dump file permissions
Oracle 11.2.0.2
Using dbms_datapipe APIs I can successfully create the dump file. However, that file does not have read or write permissions, so the UNIX user can not zip and ftp the file as required, not can it chmod the permissions. Please, advise. Thanks.Use ACLs. For example:
hpux > # whoami
hpux > whoami
oradba
hpux > # Create directory /tmp/acl_test for datapump files
hpux > mkdir /tmp/acl_test
hpux > # set directory /tmp/acl_test access to rwx for owner (Unix user oradba) and no access to group and other
hpux > chmod 700 /tmp/acl_test
hpux > # set ACL access to directory /tmp/acl_test file itself to rwx for user oracle
hpux > setacl -m u:oracle:rwx /tmp/acl_test
hpux > # set ACL access to any file created in directory /tmp/acl_test to rwx for user oracle
hpux > setacl -m d:u:oracle:rwx /tmp/acl_test
hpux > # set ACL access to any file created in directory /tmp/acl_test to rwx for user oradba
hpux > setacl -m d:u:oradba:rwx /tmp/acl_test
hpux > # show directory /tmp/acl_test ACLs
hpux > getacl /tmp/acl_test
# file: /tmp/acl_test
# owner: oradba
# group: appdba
user::rwx
user:oracle:rwx
group::---
class:rwx
other:---
default:user:oracle:rwx
default:user:oradba:rwx
hpux > # create Oracle directory object
hpux > sqlplus / << EOF
hpuxcreate directory acl_test as '/tmp/acl_test';
hpuxexit
hpuxEOF
SQL*Plus: Release 11.1.0.7.0 - Production on Mon Aug 22 15:27:56 2011
Copyright (c) 1982, 2008, Oracle. All rights reserved.
Connected to:
Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
With the Partitioning, OLAP and Data Mining options
SQL>
Directory created.
SQL> Disconnected from Oracle Database 11g Enterprise Edition Release 11.1.0.7.0
- 64bit Production
With the Partitioning, OLAP and Data Mining options
hpux > # datapump export
hpux > expdp / JOB_NAME=acl_test TABLES=T_INDEX_USAGE PARALLEL=1 COMPRESSION=ALL
REUSE_DUMPFILES=Y DIRECTORY=ACL_TEST dumpfile=acl_test_%U.dmp logfile=acl_test.
log
Export: Release 11.1.0.7.0 - 64bit Production on Monday, 22 August, 2011 15:28:0
7
Copyright (c) 2003, 2007, Oracle. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit
Production
With the Partitioning, OLAP and Data Mining options
Starting "OPS$ORADBA"."ACL_TEST": /******** JOB_NAME=acl_test TABLES=T_INDEX_USAGE PARALLEL=1 COMPRESSION=ALL REUSE_DUMPFILES=Y DIRECTORY=ACL_TEST dumpfile=acl_test_%U.dmp logfile=acl_test.log
Estimate in progress using BLOCKS method...
Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
Total estimation using BLOCKS method: 0 KB
Processing object type TABLE_EXPORT/TABLE/TABLE
. . exported "OPS$ORADBA"."T_INDEX_USAGE" 0 KB 0 rows
Master table "OPS$ORADBA"."ACL_TEST" successfully loaded/unloaded
Dump file set for OPS$ORADBA.ACL_TEST is:
/tmp/acl_test/acl_test_01.dmp
Job "OPS$ORADBA"."ACL_TEST" successfully completed at 15:28:40
hpux > # directory /tmp/acl_test listing
hpux > ls -l /tmp/acl_test
total 64
-rw-r-----+ 1 oracle dba 1036 Aug 22 15:28 acl_test.log
-rw-r-----+ 1 oracle dba 20480 Aug 22 15:28 acl_test_01.dmp
hpux > # copy datapump files (to prove we can read them)
hpux > cp /tmp/acl_test/acl_test_01.dmp /tmp/acl_test/acl_test_01_copy.dmp
hpux > cp /tmp/acl_test/acl_test.log /tmp/acl_test/acl_test_copy.log
hpux > # delete files
hpux > rm /tmp/acl_test/*
/tmp/acl_test/acl_test.log: 640+ mode ? (y/n) y
/tmp/acl_test/acl_test_01.dmp: 640+ mode ? (y/n) y
hpux > # delete directory
hpux > rmdir /tmp/acl_test
hpux > But based on "Oracle does have rights for the directory" and "UNIX user does have rights for the directory too" all you need is ACL for non-oracle UNIX user:
setacl -m d:u:unix_user:rwx directory_pathSY. -
How to Delete crashreporter files on iPad 1 on iOs 5.1.1
I would like to know what is supposed to do to delete old crashreporter dump files on the iPad to make room for new incidents crash reports
I am getting now
..... ReportCrash[114] <Error>: Failed to save memory report to /Library/Logs/CrashReporter/LowMemory-2014-03-07-135429.plist using uid: 0 gid: 0, euid: 0 egid: 0. 100 logs already present.
and unable to save new reports
I saw some information about deleting the files through iFile or ssh but this would need the iPad to be jailbroken - so not an option for me.
I have been unable to find any posts or information to advise on how to purge old CrashReporter filesAnybody from Apple here on these forums?
Surely the answer can't be to jail-break the iPad or reset to factory defaults ! -
Can delete trace files in bdump
Hi all,
i have lot of trace file is user dump directory i deleted all trace file.but i have confusion that whether to delete background trace files which are being generated by backgroung process in bdump directory.please any one suggest me what should i do. i know these kind of files ar only daignostic file.thanks alot in advance.Hi,
What I usually do is scheduling a process that deletes old files (let's say older than 1 week or so).
As you said, this is for diagnostic only, and we usually don't need old files.
Decide how much time you want to keep, and delete older files (make sure to delete only .trc files, you should keep .log files more).
Liron Amitzi
Senior DBA consultant
[www.dbsnaps.com]
[www.orbiumsoftware.com] -
hi guys,
I have disk space problem on my database server , 98% of my partition disk is used...checking the files I see that in my udump diectory there is a new 3GB trace file created by Oracle ...my questiions are :
1 - can i delete this file?
2- how can i know if oracle is still using it?
my os system is Linux and I' working on Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bi
thanks a lot,thank you... so these files are not necessary for Oracle?
below some lines from the big udump file
/u01/app/oracle/admin/aed01/udump/aed01_ora_11397.trc
Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bit Production
With the Partitioning, OLAP and Data Mining options
ORACLE_HOME = /u01/app/oracle/oracle/product/10.2.0/aed01
System name: Linux
Node name: aed01.es6.foree.lan
Release: 2.6.18-53.1.14.el5xen
Version: #1 SMP Tue Feb 19 07:33:17 EST 2008
Machine: x86_64
Instance name: aed01
Redo thread mounted by this instance: 1
Oracle process number: 27
Unix process pid: 11397, image: [email protected] (TNS V1-V3)
*** 2009-11-13 00:05:54.880
*** ACTION NAME:() 2009-11-13 00:05:54.490
*** MODULE NAME:(SQL*Plus) 2009-11-13 00:05:54.490
*** SERVICE NAME:(SYS$USERS) 2009-11-13 00:05:54.490
*** SESSION ID:(913.14607) 2009-11-13 00:05:54.478
*********START PLSQL RUNTIME DUMP************
***Got internal error Exception caught in pfrrun() while running PLSQL***
***Got ORA-4030 while running PLSQL***
FUNCTION FREE_CLIENT.DB_FUNC_SEARCH_PARENT:
library unit=b9776680 line=1 opcode=84 static link=2aabac7768e8 scope=0
FP=2aabac776b88 PC=cea6769a Page=0 AP=2aabac776b88 ST=2aabac776cd8
DL0=2aaaacccd6a0 GF=2aaaacccd708 DL1=2aaaacccd6c8 DPF=2aaaacccd6f8 DS=cea67740
DON library unit variable list instantiation
0 b9776680 2aaaacccd708 2aaaaca4b3b8
1
2
3
4
scope frame
2 0
1 2aabac776b88
package variable address size
0 2aaaacccd710 328
version=43123476 instantiation size=432
exception id error DON offset begin end DID
0 1403 0 0 72 20 71
1 0 0 1 114 14 109
aed01_ora_11397.trc
Edited by: OracleJavaLinux on Nov 13, 2009 2:18 AM -
i have my Developer Suite 9i application running under UNIX. In the path where the exceutable files are there, some files like dump ones are getting created with the format:
f90webm_dump_<pid> where pid is the process identifier
Can i remove these files because these are occupying space on the UNIX server. Please help in solving my doubt as i need it resolved urgently.1. which user i should use to search and destroy?! root-user or SYSID-User?!
Always use the user with the least permissions to do the job, don't use root if your sidadm can do it. If you want to find / delete SAP files use sidadm.
2. I found no core files and the harddisk is still 100% full, what files might also cause this problem
In your first post you wrote that /usr/sap/SID/INST/work dir is full, it is most likely that some trace files got to large. Check for files like dev_*, dew_w0 for example is the trace file of work process 0, dev_disp is the trace of the dispatcher and so on. You either have to increase the size of the filesystem, or find the cause for the growing file. It can be due to an increased trace level.
3. What on database-side could cause this problems?! can i search for sth here.
This does not look like a database issue so far.
4. i was unable to use the given scripts (noob!), what can i do else?!
Which one, please post what you typed and the error you got.
Best regards, Michael
Maybe you are looking for
-
My iphone wont show up in itunes
iphone wont connect to itunes. windows computer new itunes update done
-
Problem deleting pdf files on network ressources when preview window of the explorer is enabled
Some of our customers have the Preview Function enabled in Windows Explorer. When they delete a pdf File on a network ressource (\\server1\share1\ for example) in Windows Explorer the file disappears, but will not be deleted immediately. After a refr
-
Why can't i email a document from pages since upgrading to yosemite
since upgrading to yosemite i cannot email a document from pages. the "send via email" won't highlight and nothing happens
-
My Mac is slower since I install Lion... what can i do?
What can I do???? Is very slow the start up... it takes over a minute tu start, with Snow Leopard used to take 29 seconds...
-
How to watch movie on slixster on the Z10
I have some movies saved on Flixster. I have downloaded the Flixster app for the Z10 but the movie will not load to watch. I can see the movie I own, click on it to see details of the movie but when I click on the "Watch Now" button, nothing happen