Temp data file

hello
As read somewhere that if we delete temp data file in oracle 10g, on restarting the database it automatically creates the new temp file. So deleted that file from oradata folder. Restarted the database but not creating new temp file.
can anyone tell why this is so?

But for XE it works:
SQL> select name from v$tempfile;
NAME
C:\ORACLEXE\ORADATA\XE\TEMP.DBF
SQL> shutdown abort
Instance ORACLE arrÛtÚe.
SQL> exit
DÚconnectÚ de Oracle Database 10g Express Edition Release 10.2.0.1.0 - Productio
n
C:>del c:\oraclexe\oradata\xe\temp.dbf
C:>dir c:\oraclexe\oradata\xe\temp.dbf
Le volume dans le lecteur C s'appelle OS
Le numéro de série du volume est D298-E605
Répertoire de c:\oraclexe\oradata\xe
Fichier introuvable
C:>sqlplus / as sysdba
SQL*Plus: Release 10.2.0.1.0 - Production on Jeu. Ao¹t 26 15:56:30 2010
Copyright (c) 1982, 2005, Oracle.  All rights reserved.
ConnectÚ Ó une instance inactive.
SQL> startup
Instance ORACLE lancÚe.
Total System Global Area  805306368 bytes
Fixed Size                  1289996 bytes
Variable Size             218104052 bytes
Database Buffers          583008256 bytes
Redo Buffers                2904064 bytes
Base de donnÚes montÚe.
Base de donnÚes ouverte.
SQL> select name from v$tempfile;
NAME
C:\ORACLEXE\ORADATA\XE\TEMP.DBF
SQL>and in alert log at restart:
Thu Aug 26 15:57:05 2010
Re-creating tempfile C:\ORACLEXE\ORADATA\XE\TEMP.DBF

Similar Messages

  • Should We consider Temp Data Files While Estimating The Database Size

    Hi,
    The Database Size is sum of physical files like
    Control file
    redo log file
    datafiles
    temp files
    so i want to know why are we considering the temp files..Because it's temporary. At one stage of database, temp size it could me more and at one stage it could be less.
    So why consider the temp file???
    Please share your views on it..
    Thanks
    Umesh

    So, in essence the size of your datafiles is the size of your tablespaces?No. The size of the tablespace is the sum of the sizes of the datafiles in the tablespace --- i.e. the datafiles determine the tablespace size, not the other way round.
    (Although when you CREATE or ALTER TABLESPACE, you specify the sizes of the datafiles that you want to belong to the tablespace).
    the temporary tablespace has space allocated to it regardless of whether there are temporary tables in that tablespace or not.Two points here :
    1. On most OSs the temporary tablespace tempfile is created as a "sparse" file. So, if you issue a CREATE TEMPORARY TABLESPACE TEMP TEMPFILE 'xyz.dbf' SIZE 1000M; and then did an "ls -l" at the OS level, "xyz.dbf' would appear to be only a few tens of KBs in size. The OS "grows" the file to 1000M as necessary.
    When talking to your OS administrator ensure that you get 1000M (or the AUTOEXTEND MAXSIZE !!) space allocated even though he might "see" only a few 10s of KBs used on the first day.
    2. The temporary tablespace does not have objects (other than "global temporary tables" that overflow from memory to disk). It is really temporary space for joins, sorts, order bys etc.
    So, your datafile size is not affected regardless of your temporary tables coming and going.Yes, your datafile sizes and tempfile sizes are independent. Yet, when "sizing" disk space for the database you must include the tempfile size. However, when reporting to IT Management with a statement "our database size is ".. you might want to break it up into components like Data Dictionary, Tables, Indexes, TemporarySpace, RedoLogs and ArchiveLogs. You could also differentiate between OS-allocated space (sizes of datafiles) and Oracle-allocated space (sizes of segments) and actual used space (which you'd have to compute !) .
    Hemant K Chitale
    Edited by: Hemant K Chitale on Feb 17, 2011 10:42 PM
    Added (Although ....) paragraph to first point.

  • Temp data file corrupted.

    Hi.
    2 nodes RAC , Db version :  11.2.0.1
    One of temp file  is Temp table space is corrupted on node 1 .
    So we want to drop the corrupted temp file and create new one.
    Can some one please provide the  Steps to drop and create temp file on RAC>
    Do it requires any down time ?
    -Thanks

    Hi,
    - We will need to create a new temporary tablespace
    - ALTER DATABASE DEFAULT TEMPORARY TABLESPACE <new temp tablespace>; - This make sure all your users start using the new temp tablespace. You cannot drop a default temporary tablespace unless you assign another temporary tablespace.
    - Make sure that there are no other users using the old temp tablespace by querying v$sort_usage.
    - Once confirmed, you can drop this old tablespace(tbs) and start using the new temp tbs you created. If you want to use the old temp tablespaces, recreate it with the same name and repeat the procedure.
    Regards,
    Suntrupth

  • How to read the data file and write into the same file without a temp table

    Hi,
    I have a requirement as below:
    We are running lockbox process for several business, but for a few businesses we have requirement where in we receive a flat file in different format other than how the transmission format is defined.
    This is a 10.7 to 11.10 migration. In 10.7 the users are using a custom table into which they are first loading the raw data and writing a pl/sql validation on that and loading it into a new flat file and then running the lockbox process.
    But in 11.10 we want to restrict using temp table how can we achieve this.
    Can we read the file first and then do validations accordingly and then write to the same file and process the lockbox.
    Any inputs are highly appreciated.
    Thanks & Regards,
    Lakshmi Kalyan Vara Prasad.

    Hello Gurus,
    Let me tell you about my requirement clearly with an example.
    Problem:
    i am receiving a dat file from bank in below format
    105A371273020563007 07030415509174REF3178503 001367423860020015E129045
    in this detail 1 record starting from 38th character to next 15 characters is merchant reference number
    REF3178503 --- REF denotes it as Sales Order
    ACC denotes it as Customer No
    INV denotes it as Transaction Number
    based on this 15 characters......my validation comes.
    If i see REF i need to pick that complete record and then fill that record with the SO details as per my system and then submit the file for lockbox processing.
    In 10.7 they created a temporary table into which they are loading the data using a control file....once the data is loaded into the temporary table then they are doing a validation and updating the record exactly as required and then creating one another file and then submitting the file for lockbox processing.
    Where as in 11.10 they want to bypass these temporary tables and writing it into a different file.
    Can this be handled by writing a pl/sql procedure ??
    My findings:
    May be i am wrong.......but i think .......if we first get the data into ar_payments_interface_all table and then do the validations and then complete the lockbox process may help.
    Any suggestions from Oracle GURUS is highly appreciated.
    Thanks & Regards,
    Lakshmi Kalyan Vara Prasad.

  • Hi, my Ipod Schuffle (4 gen) dissapears from windows vista right after I execute Itunes, I've change the drive letter, erase temps in App data file, verified USB drivers, etc... Can someone guide on how to solve this?

    Hi, my Ipod Schuffle (4 gen) dissapears from windows vista right after I execute Itunes, I've change the drive letter, erase temps in App data file, verified USB drivers, etc... Can someone guide on how to solve this?

    If you want the shuffle to be always visible in Windows (outside of iTunes), you need to Enable disk use.
    Select the shuffle in the iTunes sidebar, under DEVICES.  Over to the right, select the Summary tab.  You should see a checkbox that says Enable disk use.  Check it and click Apply.
    NOTE:  If disk use is enabled, you have to Eject the shuffle in iTunes, BEFORE you physically disconnect it.  If disk use is NOT enabled, you can disconnect the shuffle at any time, as long as iTunes is not doing something with it at that moment (such as syncing); the shuffle's light will then be blinking (to indicate DO NOT DISCONNECT).
    So, unless you want to use the shuffle like a USB flash drive (carry random files on it), you may want to NOT Enable disk use.

  • Repl. Partioning BSO to ASO: increase of size of .dat file in temp-folder

    Hello,
    we are shifting data from a BSO Cube to an ASO cube via replicated partitioning. The partitioning takes about 50 minutes to execute.
    Size of .dat in metadata-folder: 8 mb
    Size of .dat in default-folder: 150 mb
    Size of .dat in temp-folder: 38 gb
    Does anyone have an explanation for the enormous size of the .dat file in temp-folder?
    Many thanks in advance!
    Michael

    I am doing the same BSO to ASO. My ess00001.dat in default is 1.9GB, in metadata it is 8.2MB, the OTL file in <db> is 18MB and the outline has about 10,000 members (rough guess). Our partition replication script looks like this:
    login <user> identified by <password> on <server>;
    spool on to <logfile>;
    refresh replicated partition <srcBSO_App>.<srcBSO_DB> to <tgtASO_App>.<tgtASO_db> at <server> updated data;
    Exit;
    I have a second process running in a task scheduler that is continuously updating the aggregates in the ASO cube. Perhaps that is cleaning out my temp .dat. The MaxL command it calls is:
    execute aggregate selection on database <tgtASO_App>.<tgtASO_db> based on query_data;
    Please check out the post I put on the other thread about how we run MaxL from a calc script and other thoughts on "round tripping" Planning-ASO-Planning. Another trick: Retrieve speed is dramatically improved by disabling and working around the @XREFs.

  • RMAN duplicate target using set until telling me it can not find data files

    I have RMAN scripts that I use freqently to clone a database from DB (production) to another (non-prod). These work just fine. I use the set until so that I can tell it at what point I want the new DB to be created from the backups of the source.
    I had a request to go back a few weeks and the backup files (I do compressed backups to disk) were on tape. I had my backup person restore my backup directory for the source DB as it looked on a certain day (May 28). I have an 8 day retention policy and so the backup files that were restored showed files all the way back to 5/20 ( includes the point I want to use in my set until clause).
    However, whenever I try to execute the rman clone, it tells me for each datafile:
    RMAN-06023: no backup or copy of datafile 1 found to restore
    Like I said, I am able to do this for a current backup that is on disk. I moved the backup files to the source db server and it works fine. However, from these files restored from tape it errors.
    Here is the RMAN script:
    spool log to c:\temp\clone_CSPROD_CSPRSUM1.log;
    #connect to catalog <catalog info>
    # target is the source and auxiliary is destination
    #connect target <put in source info here>
    #connect auxiliary /
    run {
    allocate auxiliary channel d1 type disk format 'F:\backups\CSPROD\d1\CSPROD_DATA_%s';
    allocate auxiliary channel d2 type disk format 'F:\backups\CSPROD\d2\CSPROD_DATA_%s';
    allocate auxiliary channel d3 type disk format 'F:\backups\CSPROD\d3\CSPROD_DATA_%s';
    allocate auxiliary channel d4 type disk format 'F:\backups\CSPROD\d4\CSPROD_DATA_%s';
    allocate auxiliary channel a1 type disk format 'F:\backups\CSPROD\a1\CSPROD_arch_%s';
    ##Archivelog number get from sql archive log list command
    #set until sequence 831;
    set until time "to_date('2011-05-25 08:00:00', 'YYYY-MM-DD HH24:MI:SS')";
    duplicate target database to CSPRSUM1 nofilenamecheck
    logfile
    group 1('+DATA/CSPRSUM1/onlinelog/redo1a.log', '+FRA/CSPRSUM1/onlinelog/redo1b.log') size 50m,
    group 2('+DATA/CSPRSUM1/onlinelog/redo2a.log', '+FRA/CSPRSUM1/onlinelog/redo2b.log') size 50m,
    group 3('+DATA/CSPRSUM1/onlinelog/redo3a.log', '+FRA/CSPRSUM1/onlinelog/redo3b.log') size 50m;
    Exit
    Here is the output:
    Spooling started in log file: c:\temp\clone_CSPROD_CSPRSUM1.log
    Recovery Manager11.1.0.7.0
    RMAN> #connect catalog <redacted info>>
    2> # target is the source and auxiliary is destination
    3> #connect target <redacted info>
    4> #connect auxiliary /
    5>
    6> run {
    7> allocate auxiliary channel d1 type disk format 'F:\backups\CSPROD\d1\CSPROD_DATA_%s';
    8> allocate auxiliary channel d2 type disk format 'F:\backups\CSPROD\d2\CSPROD_DATA_%s';
    9> allocate auxiliary channel d3 type disk format 'F:\backups\CSPROD\d3\CSPROD_DATA_%s';
    10> allocate auxiliary channel d4 type disk format 'F:\backups\CSPROD\d4\CSPROD_DATA_%s';
    11> allocate auxiliary channel a1 type disk format 'F:\backups\CSPROD\a1\CSPROD_arch_%s';
    12> ##Archivelog number get from sql archive log list command
    13> #set until sequence 831;
    14> set until time "to_date('2011-05-25 08:00:00', 'YYYY-MM-DD HH24:MI:SS')";
    15> duplicate target database to CSPRSUM1 nofilenamecheck
    16> logfile
    17> group 1('+DATA/CSPRSUM1/onlinelog/redo1a.log', '+FRA/CSPRSUM1/onlinelog/redo1b.log') size 50m,
    18> group 2('+DATA/CSPRSUM1/onlinelog/redo2a.log', '+FRA/CSPRSUM1/onlinelog/redo2b.log') size 50m,
    19> group 3('+DATA/CSPRSUM1/onlinelog/redo3a.log', '+FRA/CSPRSUM1/onlinelog/redo3b.log') size 50m;
    20> }
    starting full resync of recovery catalog
    full resync complete
    allocated channel: d1
    channel d1: SID=534 device type=DISK
    allocated channel: d2
    channel d2: SID=533 device type=DISK
    allocated channel: d3
    channel d3: SID=532 device type=DISK
    allocated channel: d4
    channel d4: SID=531 device type=DISK
    allocated channel: a1
    channel a1: SID=530 device type=DISK
    executing command: SET until clause
    Starting Duplicate Db at 15-JUN-11
    RMAN-05529: WARNING: DB_FILE_NAME_CONVERT resulted in invalid ASM names; names changed to disk group only.
    contents of Memory Script:
    set until scn 260398799;
    set newname for datafile 1 to
    "+data";
    set newname for datafile 2 to
    "+data";
    set newname for datafile 3 to
    "+data";
    set newname for datafile 4 to
    "+data";
    set newname for datafile 5 to
    "+data";
    set newname for datafile 6 to
    "+data";
    set newname for datafile 7 to
    "+data";
    set newname for datafile 8 to
    "+data";
    set newname for datafile 9 to
    "+data";
    set newname for datafile 10 to
    "+data";
    set newname for datafile 11 to
    "+data";
    set newname for datafile 12 to
    "+data";
    set newname for datafile 13 to
    "+data";
    set newname for datafile 14 to
    "+data";
    set newname for datafile 15 to
    "+data";
    set newname for datafile 16 to
    "+data";
    set newname for datafile 17 to
    "+data";
    set newname for datafile 18 to
    "+data";
    set newname for datafile 19 to
    "+data";
    set newname for datafile 20 to
    "+data";
    set newname for datafile 21 to
    "+data";
    set newname for datafile 22 to
    "+data";
    set newname for datafile 23 to
    "+data";
    set newname for datafile 24 to
    "+data";
    set newname for datafile 25 to
    "+data";
    set newname for datafile 26 to
    "+data";
    set newname for datafile 27 to
    "+data";
    set newname for datafile 28 to
    "+data";
    set newname for datafile 29 to
    "+data";
    set newname for datafile 30 to
    "+data";
    set newname for datafile 31 to
    "+data";
    set newname for datafile 32 to
    "+data";
    set newname for datafile 33 to
    "+data";
    set newname for datafile 34 to
    "+data";
    set newname for datafile 35 to
    "+data";
    set newname for datafile 36 to
    "+data";
    set newname for datafile 37 to
    "+data";
    set newname for datafile 38 to
    "+data";
    set newname for datafile 39 to
    "+data";
    set newname for datafile 40 to
    "+data";
    set newname for datafile 41 to
    "+data";
    set newname for datafile 42 to
    "+data";
    set newname for datafile 43 to
    "+data";
    set newname for datafile 44 to
    "+data";
    set newname for datafile 45 to
    "+data";
    set newname for datafile 46 to
    "+data";
    set newname for datafile 47 to
    "+data";
    set newname for datafile 48 to
    "+data";
    set newname for datafile 49 to
    "+data";
    set newname for datafile 50 to
    "+data";
    set newname for datafile 51 to
    "+data";
    set newname for datafile 52 to
    "+data";
    set newname for datafile 53 to
    "+data";
    set newname for datafile 54 to
    "+data";
    set newname for datafile 55 to
    "+data";
    set newname for datafile 56 to
    "+data";
    set newname for datafile 57 to
    "+data";
    set newname for datafile 58 to
    "+data";
    set newname for datafile 59 to
    "+data";
    set newname for datafile 60 to
    "+data";
    set newname for datafile 61 to
    "+data";
    set newname for datafile 62 to
    "+data";
    set newname for datafile 63 to
    "+data";
    set newname for datafile 64 to
    "+data";
    set newname for datafile 65 to
    "+data";
    set newname for datafile 66 to
    "+data";
    set newname for datafile 67 to
    "+data";
    set newname for datafile 68 to
    "+data";
    set newname for datafile 69 to
    "+data";
    set newname for datafile 70 to
    "+data";
    set newname for datafile 71 to
    "+data";
    set newname for datafile 72 to
    "+data";
    set newname for datafile 73 to
    "+data";
    set newname for datafile 74 to
    "+data";
    set newname for datafile 75 to
    "+data";
    set newname for datafile 76 to
    "+data";
    set newname for datafile 77 to
    "+data";
    set newname for datafile 78 to
    "+data";
    set newname for datafile 79 to
    "+data";
    set newname for datafile 80 to
    "+data";
    set newname for datafile 81 to
    "+data";
    set newname for datafile 82 to
    "+data";
    set newname for datafile 83 to
    "+data";
    set newname for datafile 84 to
    "+data";
    set newname for datafile 85 to
    "+data";
    set newname for datafile 86 to
    "+data";
    set newname for datafile 87 to
    "+data";
    set newname for datafile 88 to
    "+data";
    set newname for datafile 89 to
    "+data";
    set newname for datafile 90 to
    "+data";
    set newname for datafile 91 to
    "+data";
    set newname for datafile 92 to
    "+data";
    set newname for datafile 93 to
    "+data";
    set newname for datafile 94 to
    "+data";
    set newname for datafile 95 to
    "+data";
    set newname for datafile 96 to
    "+data";
    set newname for datafile 97 to
    "+data";
    set newname for datafile 98 to
    "+data";
    set newname for datafile 99 to
    "+data";
    set newname for datafile 100 to
    "+data";
    set newname for datafile 101 to
    "+data";
    set newname for datafile 102 to
    "+data";
    set newname for datafile 103 to
    "+data";
    set newname for datafile 104 to
    "+data";
    set newname for datafile 105 to
    "+data";
    set newname for datafile 106 to
    "+data";
    set newname for datafile 107 to
    "+data";
    set newname for datafile 108 to
    "+data";
    set newname for datafile 109 to
    "+data";
    set newname for datafile 110 to
    "+data";
    set newname for datafile 111 to
    "+data";
    set newname for datafile 112 to
    "+data";
    set newname for datafile 113 to
    "+data";
    set newname for datafile 114 to
    "+data";
    set newname for datafile 115 to
    "+data";
    set newname for datafile 116 to
    "+data";
    set newname for datafile 117 to
    "+data";
    set newname for datafile 118 to
    "+data";
    set newname for datafile 119 to
    "+data";
    set newname for datafile 120 to
    "+data";
    set newname for datafile 121 to
    "+data";
    set newname for datafile 122 to
    "+data";
    restore
    clone database
    executing Memory Script
    executing command: SET until clause
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    executing command: SET NEWNAME
    Starting restore at 15-JUN-11
    released channel: d1
    released channel: d2
    released channel: d3
    released channel: d4
    released channel: a1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 06/15/2011 11:21:42
    RMAN-01005: not all datafiles have backups that can be recovered to scn 260398799
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-06026: some targets not found - aborting restore
    RMAN-06023: no backup or copy of datafile 122 found to restore
    RMAN-06023: no backup or copy of datafile 121 found to restore
    RMAN-06023: no backup or copy of datafile 120 found to restore
    RMAN-06023: no backup or copy of datafile 119 found to restore
    RMAN-06023: no backup or copy of datafile 118 found to restore
    RMAN-06023: no backup or copy of datafile 117 found to restore
    RMAN-06023: no backup or copy of datafile 116 found to restore
    RMAN-06023: no backup or copy of datafile 115 found to restore
    RMAN-06023: no backup or copy of datafile 114 found to restore
    RMAN-06023: no backup or copy of datafile 113 found to restore
    RMAN-06023: no backup or copy of datafile 112 found to restore
    RMAN-06023: no backup or copy of datafile 111 found to restore
    RMAN-06023: no backup or copy of datafile 110 found to restore
    RMAN-06023: no backup or copy of datafile 109 found to restore
    RMAN-06023: no backup or copy of datafile 108 found to restore
    RMAN-06023: no backup or copy of datafile 107 found to restore
    RMAN-06023: no backup or copy of datafile 106 found to restore
    RMAN-06023: no backup or copy of datafile 105 found to restore
    RMAN-06023: no backup or copy of datafile 104 found to restore
    RMAN-06023: no backup or copy of datafile 103 found to restore
    RMAN-06023: no backup or copy of datafile 102 found to restore
    RMAN-06023: no backup or copy of datafile 101 found to restore
    RMAN-06023: no backup or copy of datafile 100 found to restore
    RMAN-06023: no backup or copy of datafile 99 found to restore
    RMAN-06023: no backup or copy of datafile 98 found to restore
    RMAN-06023: no backup or copy of datafile 97 found to restore
    RMAN-06023: no backup or copy of datafile 96 found to restore
    RMAN-06023: no backup or copy of datafile 95 found to restore
    RMAN-06023: no backup or copy of datafile 94 found to restore
    RMAN-06023: no backup or copy of datafile 93 found to restore
    RMAN-06023: no backup or copy of datafile 92 found to restore
    RMAN-06023: no backup or copy of datafile 91 found to restore
    RMAN-06023: no backup or copy of datafile 90 found to restore
    RMAN-06023: no backup or copy of datafile 89 found to restore
    RMAN-06023: no backup or copy of datafile 88 found to restore
    RMAN-06023: no backup or copy of datafile 87 found to restore
    RMAN-06023: no backup or copy of datafile 86 found to restore
    RMAN-06023: no backup or copy of datafile 85 found to restore
    RMAN-06023: no backup or copy of datafile 84 found to restore
    RMAN-06023: no backup or copy of datafile 83 found to restore
    RMAN-06023: no backup or copy of datafile 82 found to restore
    RMAN-06023: no backup or copy of datafile 81 found to restore
    RMAN-06023: no backup or copy of datafile 80 found to restore
    RMAN-06023: no backup or copy of datafile 79 found to restore
    RMAN-06023: no backup or copy of datafile 78 found to restore
    RMAN-06023: no backup or copy of datafile 77 found to restore
    RMAN-06023: no backup or copy of datafile 76 found to restore
    RMAN-06023: no backup or copy of datafile 75 found to restore
    RMAN-06023: no backup or copy of datafile 74 found to restore
    RMAN-06023: no backup or copy of datafile 73 found to restore
    RMAN-06023: no backup or copy of datafile 72 found to restore
    RMAN-06023: no backup or copy of datafile 71 found to restore
    RMAN-06023: no backup or copy of datafile 70 found to restore
    RMAN-06023: no backup or copy of datafile 69 found to restore
    RMAN-06023: no backup or copy of datafile 68 found to restore
    RMAN-06023: no backup or copy of datafile 67 found to restore
    RMAN-06023: no backup or copy of datafile 66 found to restore
    RMAN-06023: no backup or copy of datafile 65 found to restore
    RMAN-06023: no backup or copy of datafile 64 found to restore
    RMAN-06023: no backup or copy of datafile 63 found to restore
    RMAN-06023: no backup or copy of datafile 62 found to restore
    RMAN-06023: no backup or copy of datafile 61 found to restore
    RMAN-06023: no backup or copy of datafile 60 found to restore
    RMAN-06023: no backup or copy of datafile 59 found to restore
    RMAN-06023: no backup or copy of datafile 58 found to restore
    RMAN-06023: no backup or copy of datafile 57 found to restore
    RMAN-06023: no backup or copy of datafile 56 found to restore
    RMAN-06023: no backup or copy of datafile 55 found to restore
    RMAN-06023: no backup or copy of datafile 54 found to restore
    RMAN-06023: no backup or copy of datafile 53 found to restore
    RMAN-06023: no backup or copy of datafile 52 found to restore
    RMAN-06023: no backup or copy of datafile 51 found to restore
    RMAN-06023: no backup or copy of datafile 50 found to restore
    RMAN-06023: no backup or copy of datafile 49 found to restore
    RMAN-06023: no backup or copy of datafile 48 found to restore
    RMAN-06023: no backup or copy of datafile 47 found to restore
    RMAN-06023: no backup or copy of datafile 46 found to restore
    RMAN-06023: no backup or copy of datafile 45 found to restore
    RMAN-06023: no backup or copy of datafile 44 found to restore
    RMAN-06023: no backup or copy of datafile 43 found to restore
    RMAN-06023: no backup or copy of datafile 42 found to restore
    RMAN-06023: no backup or copy of datafile 41 found to restore
    RMAN-06023: no backup or copy of datafile 40 found to restore
    RMAN-06023: no backup or copy of datafile 39 found to restore
    RMAN-06023: no backup or copy of datafile 38 found to restore
    RMAN-06023: no backup or copy of datafile 37 found to restore
    RMAN-06023: no backup or copy of datafile 36 found to restore
    RMAN-06023: no backup or copy of datafile 35 found to restore
    RMAN-06023: no backup or copy of datafile 34 found to restore
    RMAN-06023: no backup or copy of datafile 33 found to restore
    RMAN-06023: no backup or copy of datafile 32 found to restore
    RMAN-06023: no backup or copy of datafile 31 found to restore
    RMAN-06023: no backup or copy of datafile 30 found to restore
    RMAN-06023: no backup or copy of datafile 29 found to restore
    RMAN-06023: no backup or copy of datafile 28 found to restore
    RMAN-06023: no backup or copy of datafile 27 found to restore
    RMAN-06023: no backup or copy of datafile 26 found to restore
    RMAN-06023: no backup or copy of datafile 25 found to restore
    RMAN-06023: no backup or copy of datafile 24 found to restore
    RMAN-06023: no backup or copy of datafile 23 found to restore
    RMAN-06023: no backup or copy of datafile 22 found to restore
    RMAN-06023: no backup or copy of datafile 21 found to restore
    RMAN-06023: no backup or copy of datafile 20 found to restore
    RMAN-06023: no backup or copy of datafile 19 found to restore
    RMAN-06023: no backup or copy of datafile 18 found to restore
    RMAN-06023: no backup or copy of datafile 17 found to restore
    RMAN-06023: no backup or copy of datafile 16 found to restore
    RMAN-06023: no backup or copy of datafile 15 found to restore
    RMAN-06023: no backup or copy of datafile 14 found to restore
    RMAN-06023: no backup or copy of datafile 13 found to restore
    RMAN-06023: no backup or copy of datafile 12 found to restore
    RMAN-06023: no backup or copy of datafile 11 found to restore
    RMAN-06023: no backup or copy of datafile 10 found to restore
    RMAN-06023: no backup or copy of datafile 9 found to restore
    RMAN-06023: no backup or copy of datafile 8 found to restore
    RMAN-06023: no backup or copy of datafile 7 found to restore
    RMAN-06023: no backup or copy of datafile 6 found to restore
    RMAN-06023: no backup or copy of datafile 5 found to restore
    RMAN-06023: no backup or copy of datafile 4 found to restore
    RMAN-06023: no backup or copy of datafile 3 found to restore
    RMAN-06023: no backup or copy of datafile 2 found to restore
    RMAN-06023: no backup or copy of datafile 1 found to restore
    RMAN> Exit
    Recovery Manager complete.
    Edited by: kerrygm on Jun 15, 2011 11:36 AM

    Maybe making progress.
    I got into RMAN and did a List backup command. In looking at the output, I see that it does have the 5/20 backup showing. Here is part of the output:
    BS Key Type LV Size Device Type Elapsed Time Completion Time
    965689 Full 19.86M DISK 00:00:01 16-JUN-11
    BP Key: 967942 Status: AVAILABLE Compressed: NO Tag: TAG20110616T231223
    Piece Name: F:\BACKUPS\CSPROD\C-368871413-20110616-01
    SPFILE Included: Modification time: 16-JUN-11
    SPFILE db_unique_name: CSPROD
    Control File Included: Ckp SCN: 369596068 Ckp time: 16-JUN-11
    BS Key Size Device Type Elapsed Time Completion Time
    966218 403.84M DISK 00:00:00 20-MAY-11
    BP Key: 967842 Status: AVAILABLE Compressed: YES Tag: TAG20110520T231012
    Piece Name: F:\BACKUPS\CSPROD\A1\CSPROD_ARCH_5844
    List of Archived Logs in backup set 966218
    Thrd Seq Low SCN Low Time Next SCN Next Time
    1 2112 237027300 20-MAY-11 237046947 20-MAY-11
    1 2113 *237046947* 20-MAY-11 237059284 20-MAY-11
    1 2114 237059284 20-MAY-11 237216514 20-MAY-11
    1 2115 237216514 20-MAY-11 237709545 20-MAY-11
    1 2116 237709545 20-MAY-11 237722825 20-MAY-11
    1 2117 237722825 20-MAY-11 237730431 20-MAY-11
    1 2118 237730431 20-MAY-11 237860199 20-MAY-11
    I then changed my rman script to use the *237046947* SCN number that says it is from 5/20. I them kicked off my rman script and got the following result.
    Spooling started in log file: c:\temp\clone_CSPROD_CSPRSUM1.log
    Recovery Manager11.1.0.7.0
    RMAN> #connect catalog 'rmancat/rmancat@rmancat';
    2> # target is the source and auxiliary is destination
    3> #connect target sys/<pwd>@csprod
    4> #connect auxiliary /
    5>
    6> run {
    7> allocate auxiliary channel d1 type disk format 'F:\backups\CSPROD\d1\CSPROD_DATA_%s';
    8> allocate auxiliary channel d2 type disk format 'F:\backups\CSPROD\d2\CSPROD_DATA_%s';
    9> allocate auxiliary channel d3 type disk format 'F:\backups\CSPROD\d3\CSPROD_DATA_%s';
    10> allocate auxiliary channel d4 type disk format 'F:\backups\CSPROD\d4\CSPROD_DATA_%s';
    11> allocate auxiliary channel a1 type disk format 'F:\backups\CSPROD\a1\CSPROD_arch_%s';
    12> ##Archivelog number get from sql archive log list command
    13> set until sequence 237046947;
    14> ##set until time "to_date('2011-05-20 08:00:00', 'YYYY-MM-DD HH24:MI:SS')";
    15> duplicate target database to CSPRSUM1 nofilenamecheck
    16> logfile
    17> group 1('+DATA/CSPRSUM1/onlinelog/redo1a.log', '+FRA/CSPRSUM1/onlinelog/redo1b.log') size 50m,
    18> group 2('+DATA/CSPRSUM1/onlinelog/redo2a.log', '+FRA/CSPRSUM1/onlinelog/redo2b.log') size 50m,
    19> group 3('+DATA/CSPRSUM1/onlinelog/redo3a.log', '+FRA/CSPRSUM1/onlinelog/redo3b.log') size 50m;
    20> }
    allocated channel: d1
    channel d1: SID=532 device type=DISK
    allocated channel: d2
    channel d2: SID=533 device type=DISK
    allocated channel: d3
    channel d3: SID=531 device type=DISK
    allocated channel: d4
    channel d4: SID=555 device type=DISK
    allocated channel: a1
    channel a1: SID=534 device type=DISK
    executing command: SET until clause
    Starting Duplicate Db at 17-JUN-11
    RMAN-05529: WARNING: DB_FILE_NAME_CONVERT resulted in invalid ASM names; names changed to disk group only.
    contents of Memory Script:
    set until scn 373021170;
    set newname for datafile 1 to
    "+data";
    set newname for datafile 2 to
    Starting restore at 17-JUN-11
    channel d1: starting datafile backup set restore
    channel d1: specifying datafile(s) to restore from backup set
    channel d1: restoring datafile 00003 to +DATA
    channel d2: restoring datafile 00122 to +DATA
    channel d2: reading from backup piece F:\BACKUPS\CSPROD\D4\CSPROD_DATA_6149
    channel d3: starting datafile backup set restore
    channel d3: specifying datafile(s) to restore from backup set
    channel d3: restoring datafile 00002 to +DATA
    channel d3: restoring datafile 00020 to +DATA
    channel d4: restoring datafile 00120 to +DATA
    channel d4: reading from backup piece F:\BACKUPS\CSPROD\D2\CSPROD_DATA_6147
    channel d1: piece handle=F:\BACKUPS\CSPROD\D3\CSPROD_DATA_6148 tag=TAG20110616T230013
    channel d1: restored backup piece 1
    channel d1: restore complete, elapsed time: 00:01:15
    channel d4: piece handle=F:\BACKUPS\CSPROD\D2\CSPROD_DATA_6147 tag=TAG20110616T230013
    channel d4: restored backup piece 1
    channel d4: restore complete, elapsed time: 00:03:15
    channel d2: piece handle=F:\BACKUPS\CSPROD\D4\CSPROD_DATA_6149 tag=TAG20110616T230013
    channel d2: restored backup piece 1
    channel d2: restore complete, elapsed time: 00:03:35
    channel d3: piece handle=F:\BACKUPS\CSPROD\D1\CSPROD_DATA_6146 tag=TAG20110616T230013
    channel d3: restored backup piece 1
    channel d3: restore complete, elapsed time: 00:04:55
    Finished restore at 17-JUN-11
    sql statement: CREATE CONTROLFILE REUSE SET DATABASE "CSPRSUM1" RESETLOGS ARCHIVELOG
    MAXLOGFILES 23
    MAXLOGMEMBERS 3
    MAXDATAFILES 1021
    MAXINSTANCES 8
    MAXLOGHISTORY 584
    LOGFILE
    GROUP 1 ( '+DATA/CSPRSUM1/onlinelog/redo1a.log', '+FRA/CSPRSUM1/onlinelog/redo1b.log' ) SIZE 50 M ,
    GROUP 2 ( '+DATA/CSPRSUM1/onlinelog/redo2a.log', '+FRA/CSPRSUM1/onlinelog/redo2b.log' ) SIZE 50 M ,
    GROUP 3 ( '+DATA/CSPRSUM1/onlinelog/redo3a.log', '+FRA/CSPRSUM1/onlinelog/redo3b.log' ) SIZE 50 M
    DATAFILE
    '+DATA/csprsum1/datafile/system.1535.754067379'
    CHARACTER SET WE8MSWIN1252
    contents of Memory Script:
    switch clone datafile all;
    executing Memory Script
    datafile 2 switched to datafile copy
    input datafile copy RECID=1 STAMP=754067676 file name=+DATA/csprsum1/datafile/sysaux.1536.754067379
    datafile 121 switched to datafile copy
    input datafile copy RECID=120 STAMP=754067677 file name=+DATA/csprsum1/datafile/waapp.1653.754067403
    datafile 122 switched to datafile copy
    input datafile copy RECID=121 STAMP=754067677 file name=+DATA/csprsum1/datafile/cu_custom.1549.754067381
    contents of Memory Script:
    set until scn 373021170;
    recover
    clone database
    delete archivelog
    executing Memory Script
    executing command: SET until clause
    Starting recover at 17-JUN-11
    starting media recovery
    Oracle Error:
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: '+DATA/csprsum1/datafile/system.1535.754067379'
    released channel: d1
    released channel: d2
    released channel: d3
    released channel: d4
    released channel: a1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 06/17/2011 15:14:54
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-06053: unable to perform media recovery because of missing log
    RMAN-06025: no backup of archived log for thread 1 with sequence 3818 and starting SCN of 372868482 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3817 and starting SCN of 372685133 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3816 and starting SCN of 372593528 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3815 and starting SCN of 372374385 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3814 and starting SCN of 372325053 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3813 and starting SCN of 372316138 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3812 and starting SCN of 372249619 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3811 and starting SCN of 371775981 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3810 and starting SCN of 371643855 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3809 and starting SCN of 371614442 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3808 and starting SCN of 371432892 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3807 and starting SCN of 371121955 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3806 and starting SCN of 371047786 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3805 and starting SCN of 371029095 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3804 and starting SCN of 371018252 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3803 and starting SCN of 370947755 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3802 and starting SCN of 370857440 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3801 and starting SCN of 370814417 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3800 and starting SCN of 370797061 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3799 and starting SCN of 370756569 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3798 and starting SCN of 370746833 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3797 and starting SCN of 370746693 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3796 and starting SCN of 370746568 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3795 and starting SCN of 370746068 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3794 and starting SCN of 370745451 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3793 and starting SCN of 370733767 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3792 and starting SCN of 370674629 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3791 and starting SCN of 370501026 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3790 and starting SCN of 370498513 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3789 and starting SCN of 370497977 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3788 and starting SCN of 370497635 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3787 and starting SCN of 370497319 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3786 and starting SCN of 370496938 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3785 and starting SCN of 370495428 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3784 and starting SCN of 370491173 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3783 and starting SCN of 370390074 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3782 and starting SCN of 370385574 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3781 and starting SCN of 370385310 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3780 and starting SCN of 370385044 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3779 and starting SCN of 370368486 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3778 and starting SCN of 370333960 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3777 and starting SCN of 370330980 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3776 and starting SCN of 370327876 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3775 and starting SCN of 370174433 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3774 and starting SCN of 370148373 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3773 and starting SCN of 370147821 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3772 and starting SCN of 370147623 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3771 and starting SCN of 370141528 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3770 and starting SCN of 369711271 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3769 and starting SCN of 369621694 found to restore
    RMAN-06025: no backup of archived log for thread 1 with sequence 3768 and starting SCN of 369595614 found to restore
    RMAN> Exit
    Recovery Manager complete.
    What I am not sure is if the SCN I used in my set until sequence command was from 5/20, when I look at the backup pieces that it is reading to restore from, they appear to be from backup files taken from June.
    (e.g channel d1: reading from backup piece F:\BACKUPS\CSPROD\D3\CSPROD_DATA_6148 -- this is a 6/16 backup file).
    Also, as you can see it was not able to do a media recovery.
    Thanks in advance for your help.

  • SAP data files

    Hi
    My SAP version is ECC 5.0 with oracle 9.2.0.7 on SOLARIS 10.
    My current DB size is 300GB where i have 23 data files on my oracle/<SID>/sapdata1...4
    I have implemented MM,FICO,PP,QM modules.My monthly DB growth rate is 30-35 GB.
    My Questions
    01.Is this growing rate is acceptable.If not how can i find the error.
    02.PSAP<SID> table space is the place where my data is written.I have to keep on monitoring the space on this table space and i have to manually extend that once it is filled.Why do i need to do this whwre i have enable "AUTOEXTEND ON" and why it is not automatically extend.
    03.Table reqorganization/Index Rebuilding  is not giving me any gain and why?
    Pls give me your feed back.
    With the current growing rate i can go for another 1.5 yrs but after that what should i do..
    Roshantha

    Hi,
    Including the brtools,
    >> we are using Oracle 10g on RHEL5.4 64Bit i want to move all my sap data files along with Temp are there any specific notes that i need to follow apart form 868750 and there is no mention of Temp file in this note.
    You can move all the datafiles including TEMP tablespace, by the statement, below;
    ALTER DATABASE RENAME FILE '<old_full_path>' TO '<new_full_path>';
    >> Apart from this i want to know what changes need to be done in the control file for the information of the new file system,so that the database can be started using changed control file having information of the renamed and moved file system.
    You will not change anything on control files, manually. It will be updated after you executed the statement, respectfully
    >> Are there any oracle parameters or profile parameters that needs to be changed and how much downtime is required.
    No you don't need to change any database profile parameter after you move the datafile(s). If you prepare a script and execute it, it will be done in a few seconds. This is because only configuration will be updated. The datafiles will not be created from the scratch.
    Best regards,
    Orkun Gedik

  • Error in running Extract Definition Upload from Data File concurrent.

    Hi all,
    Am trying to upload the 834 Extract Layout from the data file by running the concurrent program, Extract Definition Upload from Data File.
    After running this concurrent program am getting the Extract Layout definition with Record layout and Data elements within it.
    But some record layouts has the changes at their repeating level in that.
    Please suggest me how do i get the same repeating levels for the record layouts when i move the 834 benefit extract layout definition
    from one instance to another instance.

    Hi,
    We have exactly the same error in IBolt. This error happens sometime and we don't find the reason.
    IBolt has been upgraded to Version 3.1 SP1 and a fix has been developed by the support. After installation, we've work hardly during 2 days for testing all the flows by the customer without having this error.
    Ibolt works as a service and when this error occurs, the service stopped and must be restarted.
    In our flows, we have put a delete of the observer.dll file just before the rest of the flow but it doesn't solve the problem.
    Deleting the directory can be temporarly a solution but the error come back another day.
    Each time the SBO client or DTW is started, the %temp%\SMS_OBJ_DLL directory is created...
    When you just delete the observer.dll file and leave the rest of the directory like it is, the file is re-created when you start and log to SBO. I would say that this file is a copy of another file observer_800178.dll (800178 depending of the version of SBO yoi have). 800178 corresponds to SBO 2007 A PL 42.
    Do you have more information since your post has been send ?
    Thanks in advance for your help.
    Best regards.

  • Error while trying to import a repository data file in the B2Bconsole

    I am trying to import a repository data file with the delivery-channel element
    and attributes as below. <delivery-channel name="GM_DC" syncReplyMode="none" nonrepudiation-of-origin="false"
    nonrepudiation-of-receipt="false" secure-transport="true" confidentiality="false"
    routing-proxy="false" document-exchange-name="GM_DE" transport-name="GM"/> I have
    modified the WLC.dtd present in the lib\dtd subdirectory of the WebLogic Integration
    installation directory to add the syncReplyMode attribute. When i import the xml
    file,i get the following error com.bea.b2b.B2BException: ERROR: Invalid document,
    file 'C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\B2B_COV_Export.xml'. at com.bea.b2b.bulkloader.BulkLoader.processDataFile(Unknown
    Source) at com.bea.b2b.bulkloader.BulkLoader.processDataFile(Unknown Source) at
    com.bea.b2b.bulkloader.BulkLoader.load(Unknown Source)
    Please advise,
    Thanks in advance

    Hello Esther,
    the problem seems to be that the temp folder of the target Integration Builder system can't be found:
    'The system cannot find the path specified
    at [..] FileAccess.getTempDirectory([..])'.
    You'll experience the same problem if you try a file based import or export within the Integration Builder directly.
    I would recommend to continue the search there. You could check if the environment variables (for Windows: TEMP and TMP) of the OS of the system with the target Integration Builder to an existing path. Check also if the WebAs can access this path.
    Good luck
    Frank

  • Impact of data file size on DB performance

    Hi,
    I have a general query regarding size of data files.
    Considering DB performance, which of the below 2 options are better?
    1. Bigger data file size but less number of files (ex. 2 files with size 8G each)
    2. Smaller data file size but more number of files (ex. 8 files with size 2G each)
    I am working on a DB where I have noticed where very high I/O.
    I understand there might be many reasons for this.
    However, I am checking for possibility to improve DB performance though optimizing data file sizes. (Including TEMP/UNDO table spaces)
    Kindly share your experiences with determining optimal file size.
    Please let me know in case you need any DB statistics.
    Few details are as follows:
    OS: Solaris 10
    Oracle: 10gR2
    DB Size: 80G (Approx)
    Data Files: UserData - 6 (15G each), UNDO - 2 (8G each), TEMP - 2 (4G each)
    Thanks,
    Ullhas

    Ullhas wrote:
    I have a general query regarding size of data files.
    Considering DB performance, which of the below 2 options are better?Size or number really does not matter assuming other variables constant. More files results in more open file handles, but in your size db, it matters not.
    I am working on a DB where I have noticed where very high I/O.
    I understand there might be many reasons for this.
    However, I am checking for possibility to improve DB performance though optimizing data file sizes. (Including TEMP/UNDO table spaces)Remember this when tuning I/O: The fastest I/O is the one that never takes place! High I/O may very well be a symptom of unnecessary FTS or poor execution plans. Validate this first before tuning I/O and you will be much better off.
    Regards,
    Greg Rahn
    http://structureddata.org

  • Oracle 11g unreasonable large data file that can't be shrunk

    I got this Oracle 11g installed on my win7 just for running my local application. In the Enterprise Manager the "USERS" tablespaces shows over 12g for allocated size, about 1g for Space Used(also see the table snippet below), only about 8% of space is used. However, when I tried to shrink the data file/space even to half of its current size, both from command line(resize) and Enterprise Manager, i got the error that says "Failed to commit: ORA-03297: file contains used data beyond requested RESIZE value ". I was able to resize my TEMP tablespace successfully with the same command.
    Any insight on this? thanks a lot. I'm about to run out my hard drive space.
    Name / Allocated Size(MB) /Space Used(MB) /Allocated Space Used(%) /Auto Extend /Allocated Free Space(MB) /Status Datafiles /Type /Extent Management /Segment Management
    USERS / 12,288.0 /1,026.7 /8.4 /YES /11,261.3 / /1 /PERMANENT /LOCAL /AUTO

    Jonathan Lewis wrote:
    user1340127 wrote:
    However, when I tried to shrink the data file/space even to half of its current size, both from command line(resize) and Enterprise Manager, i got the error that says "Failed to commit: ORA-03297: file contains used data beyond requested RESIZE value ". I was able to resize my TEMP tablespace successfully with the same command.
    Any insight on this? thanks a lot. I'm about to run out my hard drive space.
    You have an object stored at the end of the datafile, you will need to move it down the file (or to another tablespace) before you can shrink the file. I though OEM had a "tablespace map" feature to help you see what objects were located where but if not, see:In EM (or 10g dbconsole, at any rate), you select a tablespace, drop down menu Show Tablespace Contents, then there is a tablespace map icon to expand the map. It can be slow. Then you can scroll down to the bottom, scroll up and look for where the last non-green extents are. Also, you can find the beginning of datafiles (if you have multiple data files in the tablespace) by watching for purple header blocks. Hovering over the segments gives information, and clicking on them (or selecting segments in the contents list up above) shows where all the extents are in yellow. I haven't tried the reorganize link...

  • Help required for importing a repository data file in the B2B console

    I am trying to import a repository data file with the delivery-channel element
    and attributes as below.
    <delivery-channel
    name="GM_DC"
    syncReplyMode="none"
    nonrepudiation-of-origin="false"
    nonrepudiation-of-receipt="false"
    secure-transport="true"
    confidentiality="false"
    routing-proxy="false"
    document-exchange-name="GM_DE"
    transport-name="GM"/>
    I have modified the WLC.dtd present in the lib\dtd subdirectory of the WebLogic
    Integration installation directory to add the syncReplyMode attribute.
    When i import the xml file,i get the following error
    com.bea.b2b.B2BException: ERROR: Invalid document, file 'C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\B2B_COV_Export.xml'.
         at com.bea.b2b.bulkloader.BulkLoader.processDataFile(Unknown Source)
         at com.bea.b2b.bulkloader.BulkLoader.processDataFile(Unknown Source)
         at com.bea.b2b.bulkloader.BulkLoader.load(Unknown Source)
    Please advise,
    Aruna

    Here are a few Links to a helpful Knowledge Base article and a White Paper that should help you out: http://digital.ni.com/public.nsf/allkb/BBCAD1AB08F1B6BB8625741F0082C2AF and http://www.ni.com/white-paper/10435/en/ . The methods for File IO in Real Time are the same for all of the Real Time Targets. The White Paper has best practices for the File IO and goes over how to do it. 
    Alex D
    Applications Engineer
    National Instruments

  • Intra date file processing in EBS

    Hi,
    While processing Intra date file, we are unable to post the opening balance (BAI code 040) in EBS processing.I have configured 040 External transaction and activated the 21 process type but still we are unable to post the opening balance with 040 value.

    HI
    Bhavesh, I am using NFS with By date.
    As Lalit suggested, I put the processing mode as Archieve, and gave queue name as "Temp". In the archieve folder I could able to see the file sequence how they are getting processed and it is up to my requirement.
    Now can I view the same file processing in "Temp" Queue. For this I think I need to configure the queues.
    Please guide me how to configure the queues and view the files in this queue "Temp".
    Thanks,
    anil.

  • Sql loader:loading data file name to one of the tabe column.

    Hi,
    I want to load the data file name also to one of the table column.
    Table structure:
    File_Name varchar2
    Name varchar2
    Age number
    Data File Format:
    Vijay|24
    Anandh|24
    Senthil|26
    In the above eg how to write a control file to load the data file name to the File_Name column of the table.
    Note:
    The shell script will first pick up the data files to be loaded and calls the control file.
    Thanks,
    Vijay.

    From a unix load perspective this is what we have done :
    Create a control file (say load_data.ctl) something like
    LOAD DATA
    append
    Into table file_name_age
    File_Name constant '<FILENAME>',
    Name CHAR TERMINATED BY '|'
    Age INTEGER EXTERNAL TERMINATED BY '|' )
    in a unix loop do the following
    for i in `ls $LOADDIR` ; do
    #Replace <FILENAME> with actual filename and create a temp ctl file
    sed s/\<FILENAME\>/${1}/g < ${CTLDIR}/load_data.ctl > ${TMPDIR}/load_data_${1}.ctl
    #load using temp file
    sqlldr / CONTROL=${TMPDIR}/load_data_${1}.ctl
    DATA=$i
    #remove temp
    rm ${TMPDIR}/load_data_${1}.ctl
    done
    Where
    ${TMPDIR}=your temp directory
    ${CTLDIR}=where your main directory for ctl files
    $LOADDIR=Your data directory

Maybe you are looking for