Oracle binary and control files

Hi All,
I want know whether the oracle binary and control files are they related in anyway.
I have my physical files on a SAN storage and my oracle binary files on a local disk.
In case if I delete my oracle binaries and restore it from a backup, will I be able to start my database without any issues.
Since all my oracle datafiles,controlfiles and redofiles are located in SAN storage.

Oracle binaries and control files are related in some way because Oracle version is recorded in control files:
oerr ora 201
00201, 00000, "control file version %s incompatible with ORACLE version %s"
// *Cause:  The control file was created by incompatible software.
// *Action: Either restart with a compatible software release or use
//          CREATE CONTROLFILE to create a new control file that is
//          compatible with this release.When restoring Oracle binaries on UNIX you should take care about setuid bits on oracle executable to avoid local connection issues by non oracle Unix accounts.

Similar Messages

  • How to create redlog and control file at ASM in linux RAC

    Hi Experts,
    I will to maintance a oracle 10g database at ASM as RAC iin linux red hat.
    i am a new person with some question.
    nornally speaking, oracle recommadition for oracle 10g database as
    create 3 copy fills for control file
    create at least 2 redo log with mirror files in system.
    However, I checked find
    redlog file is at FRA place +FLSdisk1 and no mirror
    control file is at FRA place--+FLSDISK1/
    datebase file at ‘+DATA1/
    There are no mirror for relog.
    Go to EM, I also could not find place to enter file name.
    We use ASM to hold database to support RAC.
    Do i need to create redlog file as
    ALTER DATABASE ADD LOGFILE GROUP 1 ('+FLSdisk1/sale/onlinelog/REDO01.LOG','+FLSdisk1/sale/onlinelog/REDO01_mirror.LOG') SIZE 1000M reuse;
    my boss told me that ASM is reliable.
    Do you need to creat more directory to arrange redlog and control files in ASM for RAC system?
    FRA is a good place to store control file and redlog file ?
    Thanks
    JIM
    Edited by: user589812 on Jul 3, 2009 3:03 PM

    ASM is reliable but a smart DBA is very careful. If ASM is doing mirroring this is like RAID doing mirroring. What happens if you accidentally delete one copy ... the other one disappears instantly. Not a good idea.
    With respect to redo logs you need a minimum of three groups, two members, and one thread per instance. So a 2 node cluster should, at a minimum have 12 physical files.
    Not mirroring the redo logs, assuming multiple members, is not as critical.

  • If it is not specified,Oracle creates one control file in a default locatio

    Hi
    The CONTROL_FILES parameter is not required when you create a database. If it is not specified,Oracle creates one control file in a default location,.
    Where oracle create contol file default in this condition
    Thanks

    Can you try to create a database and see
    FYI
    It creates in $ORACLE_HOME\database folder..if not specified but in UNIX i think it will be under $ORACLE_HOME/dbs
    Because i dont have UNIX right now to test it
    Message was edited by:
    Maran Viswarayar

  • Very high log file sequential read and control file sequential read waits?

    I have a 10.2.0.4 database and have 5 streams capture processes running to replicate data to another database. However I am seeing very high
    log file sequential read and control file sequential read by the capture procesess. This is causing slowness in the database as the databass is wasting so much time on these wait events. From AWR report
    Elapsed: 20.12 (mins)
    DB Time: 67.04 (mins)
    and From top 5 wait events
    Event Waits Time(s) Avg Wait(ms) % Total Call Time Wait Class
    CPU time 1,712 42.6
    log file sequential read 99,909 683 7 17.0 System I/O
    log file sync 49,702 426 9 10.6 Commit
    control file sequential read262,625 384 1 9.6 System I/O
    db file sequential read 41,528 378 9 9.4 User I/O
    Oracle support hasn't been of much help, other than wasting my 10 days and telling me to try this and try that.
    Do you have streams running in your environment, are you experiencing this wait. Have you done anything to resolve these waits..
    Thanks

    Welcome to the forums.
    There is insufficient information in what you have posted to know that your analysis of the situation is correct or anything about your Streams environment.
    We don't know what you are replicating. Not size, not volume, not type of capture, not rules, etc.
    We don't know the distance over which it is being replicated ... 10 ft. or 10 light years.
    We don't have any AWR or ASH data to look at.
    etc. etc. etc. If this is what you provided Oracle Support it is no wonder they were unable to help you.
    To diagnose this problem, if one exists, requires someone on-site or with a very substantial body of data which you have not provided. The first step is to fill in the answers to all of the obvious first level questions. Then we will likely come back with a second level of questioning.
    But when you do ... do not post here. Your questions are not "Database General" they are specific to Streams and there is a Streams forum specifically for them.
    Thank you.

  • How to create parameter and control file like filename + date

    Hello there
    I am trying to create parameter and control file with following command
    in SQLPLUS
    create pfile='/u03/oradata/WEBDB/backup/initWEBDB.ora' from spfile;
    In RMAN
    copy current controlfile to '/u03/oradata/WEBDB/backup/cf_longterm.cpy';
    how can I put date at the end of filename like
    initWEBDB8jan06.ora and cf_longterm8jan06.cpy
    Thanks in advance
    Lionel

    ASM is reliable but a smart DBA is very careful. If ASM is doing mirroring this is like RAID doing mirroring. What happens if you accidentally delete one copy ... the other one disappears instantly. Not a good idea.
    With respect to redo logs you need a minimum of three groups, two members, and one thread per instance. So a 2 node cluster should, at a minimum have 12 physical files.
    Not mirroring the redo logs, assuming multiple members, is not as critical.

  • Location of Redo log and control files?

    Dear all,
    I am checking the location of redo log and control files, but found that the redo log file (like log02a.dbf ....) in the same directory of data files. However, I couldn't find any control files in the data files directries.
    What could be the location of control files?
    Amy

    select name
      from v$controlfile
    or
    show parameter control_filesKhurram

  • SQL Loader and control file changes for different users

    In the front end of my application I can select a data file and a control file, and load data to the table mentioned in .ctl file. Every user who logs in uses the same .ctl file and so loads onto the same table. Now I want the user to load data onto the table in his own schema. I can get the username of the user currently logged in and i want to insert it into that username.table. So can i copy the contents of the .ctl file into a variable, modify it into username.table in that string and pass that variable as a parameter to the sqlldr command instead of the .ctl file.
    Or is there a better way how I can modify the same control file everytime to change tablename to username.tablename in .ctl file and pass to sqlldr to load data to table in local user schema table.
    Thanks and Regards

    Thanks for the reply .. user do have their user credentials but only for the application ... but all users use a common loader and control file once they log into the application. So irrespective of which user is logged in he selects the same control file and loads to the same table mentioned in the control file .. i instead want user to be able to load to the table in control file but into his schema like username.tablename instead of just the tablename mentioned in .ctl file.

  • Load data with SQL Loader link field between CSV file and Control File

    Hi all,
    in a SQL Loader control file, how do you specify link with field in CSV file and Control file?
    E.g. if I wat to import the record in table TEST (col1, col2, col3) with data in csv file BUT in different position. How to do this?
    FILE CSV (with variable position):
    test1;prova;pippo;Ferrari;
    xx;yy;hello;by;
    In the table TEST i want that col1 = 'prova' (xx),
    col2 = 'Ferrari' (yy)
    col3 = default N
    the others data in CSV file are ignored.
    so:
    load data
    infile 'TEST.CSV'
    into table TEST
    fields terminated by ';'
    col1 ?????,
    col2 ?????,
    col3 CONSTANT "N"
    Thanks,
    Attilio

    With '?' mark i mean " How i can link this COL1 with column in csv file ? "
    Attilio

  • Multiplexing redo logs and control files to a separate diskgroup

    General question this one...
    I've been using ASM for a few years now and have always installed a new system with 3 diskgroups
    +DATA - for datafiles, control files, redo logs
    +FRA - for achive logs, flash recovery. RMAN backup
    Those I guess are the standards, but I've always created an extra (very small) diskgroup, called +ONLINE where I keep multiplexed copies of the redo logs and control files.
    My reasoning behind this is that if there are any issues with the +DATA diskgroup, the redo logs and control files can still be accessed.
    In the olden days (all those 5 years ago!), on local storage, this was important, but is it still important now? With all the striping and mirroring going on (both at ASM and RAID level), am I just being overtly paranoid? Does this additional +ONLINE diskgroup actually hamper performance? (with dual write overheads that are not necessary)
    Thoughts?

    Some of the decision will probably depend on your specific environment's data activity, volume, and throughput.
    Something to remember is that redo logs are sequential write, which benefit from a lower RAID overhead (RAID-10, 2 writes per IOP vs RAID-5, 4 writes per IOP). RAID-10 is often not cost-effective for the data portion of a database. If your database is OLTP with a high volume of random reads/writes, you're potentially hurting redo throughput by creating contention on the disks sharing data and redo. Again, that depends entirely on what you're seeing in terms of wait events. A low volume database would probably not experience any noticeable degraded performance.
    In my environment, I have RAID-5 and RAID-10 available, and since the RAID-10 requirement from a capacity perspective for redo is very low, it makes sense to create 2 diskgroups for online redo, separate from DATA, and separate from each other. This way, we don't need to be concerned with DATA transactions impacting REDO performance, and vice versa, and we still maintain redo redundancy.
    In my opinion, you can't be too paranoid. :)
    Good luck!
    K

  • Database restore without temp, undo and control files.

    Hi All,
    You might found this question silly but I don't know so asking this question here.
    I have cold back up of the database. Now, I want create clone of that database, but I have some different paths for the DBFs so I will create new control file after restoring the database.
    Now, I know that I don't need control files and tempfiles to be restored. I have 10 undo files in backup but on the new clone database I don't need all 10. I want only 5. So can I do the restoration without undo , temp and control file and later on add undo and temp?? and if yes then tell me that can I add them at mount level??
    This is my first restore, Please guide me its very urgent

    Nitin Joshi wrote:
    f the COLD Backup does not include the Online Redo Logs, an ALTER DATABASE OPEN RESETLOGS is requireed >>to create these Online Redo Logs. Unfortunately, an OPEN RESETLOGS can only be done after an Incomplete >>Recovery or when using a Backup Control file.
    Therefore, we do a RECOVER with a CANCEL to simulate an Incomplete Recovery.Completely agree with you Hemant. And the links you've provided,i've gone through many times. Excellent description.
    I just wanted to know in above(OP's) scenario if he has complete cold backup(includes online redo logs), does he really need open reset logs or any recovery?
    Regards!no , if you have cold backup with online redo log files then i don't think so you need to open database in resetlogs.Resetlog is always after incomplete recovery or recovery using backup controlfile or you dont have redo logs.
    I am completely agree with you that with given scenario for the cold backup undo tablespace would not be part of recovery and you can
    -offline drop undo tablespace file
    -create another one undo tablespace and its undo datafile
    -point spfile to that newly undo tablespace
    I think Aman is saying in the context of restore and recover online database where undo tablespace create a vital role in database recovery, the undo blocks roll back the effects of uncommitted transactions previously applied by the rolling forward phase.
    Khurram

  • Where is the location of tablespace file and control file

    Hi, all
    where is the location of tablespace file and control file? tks

    For DataFiles, query DBA_DATA_FILES or V$DATAFILE
    For TempFiles, query DBA_TEMP_FILES or V$TEMPFILE
    For Online Redo Logs, query v$LOGFILE
    For Archived Redo Logs, query v$ARCHIVED_LOG
    for Controlfiles, query v$CONTROLFILE
    Hemant K Chitale
    http://hemantoracledba.blogspot.com

  • Using archivelog and control file from other Oracle server

    I am still bothered with my backup process.
    I have 2 AIX boxes (same model, say A and B); both have BAAN 5 and Oracle 10g R2 on. Right now my colleague insists to use the export pump (cold backup) from Prod Oracle server (A) to restore the Oracle server on Box B. The Prod server has the archivelog mode turn on. But it will miss any transaction data from import pump till the crash point of Box A. So this is my confusion.
    Can I pass the control files and archivelog files form Box A (prod server) to Box B and use them to restore the Box B as the latest Prod server? How?
    I tried to convince them to use the RMAN backup? But not successful?
    I think the best way is probably to use the Oracle Data guard. However, there is always one concern to my manager and colleague, that such process will cause the data on the restored server (failover, Box B) not recognizable by the BAAN, which define the objects (tables).
    Thanks

    Performing a logical backup is not useful to restore to the point of failure. The only valid and available option is a hot backup/archivelog mode. Your recovery manager backup perform a controlfile and redologfile backup, so those can be restored at the destination. You must take care of the way you perform the backup, and ensure the paths where your backup is being deposited are visible by the second node. A shared storage with same mount points is suitable in this case. A tape robot configured at both nodes is also a suitable solution.
    Recovery manager perfoms a controlfile and spfile restore, too. This rman command perform the action:
    SET DBID <DBID of the database,for which you want to restore the controlfile>;
    RESTORE CONTROLFILE FROM <name_of_backupiece_which_contains_the_controlfile backup>';
    I don't see any problem on the recovery manager side, and technically speaking, on the Oracle side it is perfectly possible to restore your database at a remote location. I don't know what happens on the BAAN side, if you are required to have it configured to be operative on the target node. You could try to clone your database at the node B, configure BAAN and prepare the proceduere in case of failure.
    Configuring a dataguard is also a recomended action. So it is to think about Cold Failed Over clusters. I have recently performed a CFC configuration with BAAN. No problem it works smoothly.
    ~ Madrid.

  • Flat File and Control Files Questions

    Greetings,
    I've worked with Oracle for about 10 years, but have little experience with using sql-loader.
    I have data from Visual FoxPro tables going into Oracle 10g via a Perl script. I am having issues and therefore have a couple questions.
    1) If the data from my foxpro table is basically everything in the table as in 'Select * from table-name', does the control file have to list every column that is in the FoxPro table?
    -- I have a case where a FoxPro table has 15 columns but we are trying to upload only 10 columns. The script is dynamic. It selects * from each FoxPro table and creates a Flat File for each on the fly. Then sql-loader uploads the data to Oracle. The Flat File for this one table has data from all 15 columns, but the Control File only lists 10 of the columns to be uploaded into Oracle.
    2) Do the column names in the control file 'have' to match both the column names in the FoxPro table and the Oracle table, or only the Oracle table?

    YankeeFan wrote:
    Greetings,
    I've worked with Oracle for about 10 years, but have little experience with using sql-loader.
    I have data from Visual FoxPro tables going into Oracle 10g via a Perl script. I am having issues and therefore have a couple questions.
    1) If the data from my foxpro table is basically everything in the table as in 'Select * from table-name', does the control file have to list every column that is in the FoxPro table?
    -- I have a case where a FoxPro table has 15 columns but we are trying to upload only 10 columns. The script is dynamic. It selects * from each FoxPro table and creates a Flat File for each on the fly. Then sql-loader uploads the data to Oracle. The Flat File for this one table has data from all 15 columns, but the Control File only lists 10 of the columns to be uploaded into Oracle.
    Yes - use the FILLER spec to ignore columns you do not care about - http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/ldr_field_list.htm#sthref946
    2) Do the column names in the control file 'have' to match both the column names in the FoxPro table and the Oracle table, or only the Oracle table?Only the Oracle table.
    HTH
    Srini

  • Datafiles and control file corrupted

    I have oracle 9i installed on windows 2003 server. A couple of days before our windows crashed and the system administrator took the backup of all files and reinstalled windows. After reinstallation of windows I restored the files backed up by system administrator and tried to start the database. I got the following error:
    ORA-00227: corrupt block detected in controlfile: (block 1, # blocks 1)
    ORA-00202: controlfile: 'D:\ORACLE\ORADATA\ORCL\CONTROL03.CTL'
    All the multiplexed copies of control files are also corrupted. We do not have the backup since last few days and the database is in noarchivelog mode. Simple restoration to the old backup would be a great loss. Kindly help me to recover the control file and datafiles.

    All the multiplexed copies of control files are also corrupted. We do not have the backup since last few days and the database is in noarchivelog mode. Simple restoration to the old backup would be a great loss. Kindly help me to recover the control file and datafiles.You could try to re-create the control files assuming the database itself was closed properly.
    However you should only do this if you know what you are doing.
    In any case you should backup the database in its current state in case things getting worse by trying to recover. CHECK this backup twice!
    In any case: Open a support ticket at Oracle. You will most probably need their help.
    In addition to that - it look quite bad for your data. You should prepare for data loss (just to be sure; i dont want to scare you)....
    Ronny Egner
    My Blog: http://blog.ronnyegner-consulting.de

  • Sqlldr and control file

    Howdy:
    I'm running Oracle 9i (9.0.2) on Redhat Linux 7.3.
    I am trying to append data from a flat file into a
    table where one of the field is timestamp(6).
    The data has fractional seconds and looks like so:
    [snip data]
    2002-12-08 22:29:38.216712
    [snip data]
    I want to use sqlldr to load the data, but when
    I try to create the control file, I am getting
    errors telling me my method is wrong.
    In my control file, I am trying to do things like:
    [snip versions of the 'UPDATED' timestamp field]
    * UPDATED TIMESTAMP "YYYY-MM-DD HH24:MI:SS"
    * UPDATED TIMESTAMP "YYYY-MM-DD HH24:MI:SS.FFF"
    * UPDATED POSITION(132:151)"TO_DATE(:UPDATED, 'YYYY-MM-DD HH24:MI:SS')"
    [snip versions ...]
    I am having no success.
    Can someone give an example of how to create
    a control file for sqlldr when defining a
    timestamp field with fractional seconds six
    digits past the decimal?
    Thanks!
    -X

    --update:
    --this is pretty much for my notes (as i'm
    --sure i will need this info again).
    --i have figured out how to fix my problem.
    --re-cap:
    --so, i'm running Oracle 9i (9.2.0) on Redhat Linux 7.2 as
    --you know.  I want to append my data from a flat file
    --into a table that i've created.  i also want to use SQLLDR
    --to do it.  i  have create a bunch of scripts to do this
    --and i was close - the problem was that in one of the
    --fields (table) was a TIMESTAMP field and i couldn't
    --figure out how to create a control file to say 'hey, there
    --is a timestamp field in there and it has fractions in
    --the seconds area ... 6 decimal places over'.
    --so, this is what my table looks like:
    [snip table]
    SQL> describe hmp_dates
    Name Null? Type
    CONTRACT CHAR(12)
    MBR_NUM CHAR(2)
    DATE DATE
    EVENT_ID NUMBER(38)
    VALUE VARCHAR2(50)
    RESULT VARCHAR2(50)
    UPDATED TIMESTAMP(6) WITH TIME ZONE
    USERNAME VARCHAR2(50)
    CLAIM VARCHAR2(50)
    [snip table]
    --this is what my data looks like:
    [snip data]
    xxxxxxxxx |02|2000-05-31|38|\N|\N|2002-12-08 22:29:38.216712|ADMIN|\N
    xxxxxxxxx | |1999-06-06|38|\N|\N|2002-12-08 22:29:38.216712|ADMIN|\N
    xxxxxxxxx | |1998-12-28|38|\N|\N|2002-12-08 22:29:38.216712|ADMIN|\N
    xxxxxxxxx | |1999-12-03|38|\N|\N|2002-12-08 22:29:38.216712|ADMIN|\N
    [snip data]
    --this is what my control file looks like:
    [snip control]
    LOAD DATA
    INFILE '/var/local/table_backup/hmp_dates_03July2003.txt'
    APPEND
    INTO TABLE HMP_DATES
    FIELDS TERMINATED BY '|' OPTIONALLY ENCLOSED BY '"'
    CONTRACT,
    MBR_NUM,
    "DATE" DATE "YYYY-MM-DD",
    EVENT_ID,
    VALUE,
    RESULT,
    UPDATED TIMESTAMP WITH TIME ZONE 'YYYY-MM-DD HH24:MI:SSXFF', *****
    USERNAME,
    CLAIM
    [snip control]
    --please note the astrisks. that is the elfin goodness. (SSXFF to denote precision
    --seconds).
    --hopes this helps someone besides me.
    -X

Maybe you are looking for

  • Disable or bypass SSL on Sharepoint Foundation 2013

    My SSL certificate is expired and I am going to purchase the new certificate for the server. But at the mean time, I want to disable SSL (https) login, I've tried to remove the certificate and binding from IIS and remove the certificate from SharePoi

  • How can I sync my devices to new PC without erasing files?

    Ok, I've seen similar posts, so I'm guessing that this may be a stupid question, but I want to ask again according to my specific situation. I have an iPhone 4S, an iPad 2, and an iPad Air.   I had the first two synced on my laptop (which died last y

  • Windows Vista Home Basic 64-bit

    I'm looking at the customisable configurations, and I see that you offer 8GB RAM but that a 64-bit system is required to support it. At the same time, the only option that explicity says it's 64-bit is the Business edition of Vista, which I definitel

  • Oracle 8i for solaris(intel based machine) software download

    hi sir, i wnat oracle 8i software on intel(solaris) platform, i am not abel to find out in oraclr site, plz give me replay , where can i get regds jani So that we may better diagnose DOWNLOAD problems, please provide the following information. - Serv

  • Network Two Printer in Different Rooms

    I currently have a airport extreme with a shared printer connected to it. My my two printers can print to this shared printer like it is supposed to. I would like to add another shared printer, however I want it in a different room than where my airp