Moving the database to different mount point.

Hi,
I have R12 app tier on one server and 10gR2 database on separate server.
I want to move the database files from /u01/oradata/PROD to /u02/oradata/PROD on the same server.
I know how to do this for the database using alter database rename file .....
But what do I have to do with the application and context files.?
Do I just need to edit the values s_dbhome1, s_dbhome2, s_dbhome3 and s_dbhome4 to point to /u02/oradata/PROD and then run adconfig?
Many thanks
Skulls.

I am not cloning, just moving the databaseI know, but Rapid Clone does also help (use the same hostname, domain name, port pool, ..etc except for the directory paths) -- Please note you do not need to use "alter database rename file" as Rapid Clone will take care of that.
so I think manual changing datfile names is needed, followed by edit context file and run autoconfig?This is also valid.
Thanks,
Hussein

Similar Messages

  • Is it possible in 9i to take export backup in two different mount point

    Hello Team,
    Is it possible in 9i to take export in two different mount point with file size 22 gb.
    exp owner=PERFSTAT FILE =/global/nvishome5/oradata/jlrvista/PERFSTAT_exp01.dmp,/global/nvishome4/oradata/jlrvista/export/PERFSTAT_exp02.dmp FILESIZE=22528
    I tried with above but no luck..so later killed session
    prs72919-oracle:/global/nvishome5/oradata/jlrvista$ exp owner=SLENTON FILE =/global/nvishome5/oradata/jlrvista/PERFSTAT_exp01.dmp,/global/nvishome4/oradata/jlrvista/export/PERFSTAT_exp02.dmp FILESIZE=2048
    Export: Release 9.2.0.8.0 - Production on Thu Nov 14 13:25:54 2013
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Username: / as sysdba
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.8.0 - 64bit Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.8.0 - Production
    Export done in US7ASCII character set and UTF8 NCHAR character set
    server uses UTF8 character set (possible charset conversion)
    About to export specified users ...
    . exporting pre-schema procedural objects and actions
    . exporting foreign function library names for user SLENTON
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions for user SLENTON
    continuing export into file /global/nvishome4/oradata/jlrvista/export/PERFSTAT_exp02.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    About to export SLENTON's objects ...
    . exporting database links
    . exporting sequence numbers
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    . exporting cluster definitions
    . about to export SLENTON's tables via Conventional Path ...
    . . exporting table                      G_AUTHORS
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp >
    continuing export into file expdat.dmp
    Export file: expdat.dmp > ps -ef | grep exp
    continuing export into file ps -ef | grep exp.dmp
    Export file: expdat.dmp > ^C
    continuing export into file expdat.dmp
    EXP-00056: ORACLE error 1013 encountered
    ORA-01013: user requested cancel of current operation
    . . exporting table                        G_BOOKS
    Export file: expdat.dmp > ^C
    continuing export into file expdat.dmp
    EXP-00056: ORACLE error 1013 encountered
    ORA-01013: user requested cancel of current operation
    . . exporting table                 G_BOOK_AUTHORS
    Export file: expdat.dmp > ^C
    continuing export into file expdat.dmp
    Export file: expdat.dmp > Killed

    See the text in BOLD , if you do not specify the sufficient export file names, export will prompt you to provide additional file names. So either for your 22 GB you need to give 11 different file names or provide the filename when its prompted.
    FILE
    Default: expdat.dmp
    Specifies the names of the export files. The default extension is .dmp, but you can specify any extension. Since Export supports multiple export files , you can specify multiple filenames to be used.
    When Export reaches the value you have specified for the maximum FILESIZE, Export stops writing to the current file, opens another export file with the next name specified by the parameter FILE and continues until complete or the maximum value of FILESIZE is again reached. If you do not specify sufficient export filenames to complete the export, Export will prompt you to provide additional filenames.

  • The database is in mount or nomount state.

    I see similar posts, but I have not understood the solutions. This is my first installation of a DB. I'm taking a class that requires db 10g, and I'm trying to load it on my home computer. Windows XP pro. Everything seemed to be going well, until the point where it is suppose (by the instructions I'm following) to configure a new user account. When the URL:http://localhost:5500/em is suppose to open to the login page it gives me the following error page
    The database status is currently unavailable. It is possible that the database is in mount or nomount state.
    If I hit the "Startup" button I get the following page...
    Startup/Shutdown:Specify Host and Target Database Credentials
    Specify the following credentials in order to change the status of the database.
    It wants "Host Credentials" with username and password that no matter what I put in, it doesn't like. What are the "Host" Credentials?
    The installation I'm following is off of this site...
    http://www.course.com/cdkit/installinstructions_dbIII.cfm
    I can log in to SQL with the default Scott/Tiger - but (and I don't know anything about this really), if the db won't start a lot of good it does me.
    Anybody have any suggestions? I can't seem to establish a new user account for 10g, as the db is 'currently unavailable.''
    thanks,
    Dan

    Taj,
    You tried to help me before, and I have the same problem with a differnet computer. The installation goes great until I get to step 13 (listed below). When the Oracle Enterprise Manager 10g Database Control web page opens it tells me that the db is not online. I did what you said in previous posts - except the last post. Am I suppose to do that?
    I can't seem to create a new user. Actually I just tried to log on to http://localhost:5500/em and it said the page can't load.
    13.     When the End of Installation page displays and confirms that the installation was successful, click Exit, and then click Yes to confirm exiting. The Oracle Enterprise Manager 10g Database Control web page opens.
    Configuring a New User Account for the Oracle10g Database
    When you start your computer, the Oracle10g database starts automatically, and is ready to accept data requests. Before you can use the Oracle10g database, you must perform some configuration tasks. First you create a user account, which is identified by a user name and password and is associated with an area in the database in which you create your database tables. You create a user account named dbuser, with password oracle.
    To create the user account:
    1.     On the Oracle Enterprise Manager Database Control Login page, type SYSMAN as the User Name and enter the password you specified in step 6 above, and then click Login.
    TIP: If you are not on the Oracle Enterprise Manager Database Control Login page, then open a Web Browser and type the following URL: http://localhost:5500/em
    2.     The Oracle Database Licensing Information 10g page will appear. Read the page and scroll to the bottom and then click I Agree. This page appears the first time a new user logs in to the Oracle Enterprise Manager Database Control application.
    3.     The Oracle Enterprise Manager Database Control Home page will appear. To create a new user for the Oracle database: orcl, click the Administration link.
    4.     Under the Security heading click the Users link.
    5.     On the Oracle Enterprise Manager Database Control Users page, click the Create button located on the right hand side of the page.
    6.     Type DBUSER in the Name field, oracle in the Enter Password field, and oracle in the Confirm Password field.
    7.     Type USERS in the Default Tablespace field and Temp in the Temporary Tablespace field.
    8.     Click the Roles link, click the Modify button, select DBA in the Available Roles list, and then click the move arrow so that DBA appears in the Selected Roles list.
    9.     Click OK on the Modify Roles page, and then click OK on the Create User page. A message confirms that the object was successfully created.
    10.     Click the Logout link at the bottom of the page.
    11.     Close the Oracle Enterprise Manager 10g Database Control web page by clicking the X icon in the upper right corner of the web page.
    Starting SQL*Plus and Logging On to the Oracle Database 10g
    To confirm that you can connect to your Oracle10g database, you now start SQL*Plus and log on to the database using the DBUSER account.
    To start SQL*Plus and log on to the database:
    1.     Click Start on the Windows taskbar, point to Programs or All Programs, point to Oracle – OraDb10g_home1, point to Application Development, and then click SQL Plus. The Log On dialog box opens and requests your user name, password, and host string.
    2.     Type DBUSER in the User Name field, press Tab, type oracle in the Password field, leave the Host String field blank, then click OK. The SQL*Plus program window opens, and the SQL prompt appears.
    3.     Close the Oracle SQL*Plus window.

  • Btrfs with different file systems for different mount points?

    Hey,
    I finally bought a SSD, and I want to format it to f2fs (and fat to boot on UEFI) and install Arch on it, and in my old HDD I intend to have /home and /var and  try btrfs on them, but I saw in Arch wiki that btrfs "Cannot use different file systems for different mount points.", it means I cannot have a / in f2fs and a /home in btrfs? What can I do? Better use XFS, ZFS or ext4 (I want the faster one)?
    Thanks in advance and sorry my english.

    pedrofleck wrote:Gosh, what was I thinking, thank you! (I still have a doubt: is btrfs the best option?)
    Just a few weeks ago many of us were worrying about massive data loss due to a bug introduced in kernel 3.17 that caused corruption when using snapshots. Because btrfs is under heavy developement, this sort of thing can be expected. That said, I have my entire system running with btrfs. I have 4 volumes, two raid1, a raid0 and a jbod. I also run rsync to an ext4 partition and ntfs. Furthermore I make offline backups as well.
    If you use btrfs make sure you have backups and make sure you are ready to use them. Also, make sure you check sum your backups. rsync has the option to use checksums in place of access times to determine what to sync.

  • Unexpected disconnection external disk and different mount points

    Dear community,
    I have an application that needs to read and write data from an external disk called "external"
    If the volume is accidentally unmounted by unproperly pluggin it off), it will remount as "external_1" in the terminal,
    and my app won't see it as the original valid destination.
    According to this documentation:
    https://support.apple.com/en-us/HT203258
    it needs a reboot to be solved, optionally removing the wrong unused mount points before rebooting.
    Would there be a way to force OSX remounting the volume on the original mount point automatically?
    or checking the disk UUID and bypassing the different mount point name (app or os level?)
    Thanks for any clue on that.

    See DUMPFILE

  • Cannot open the database, stuck in mount mode

    Someone please shed some lights on this problem.
    I ran into some media failure, and ended up restoring the database from the offline whole backup.
    The database is in the nonarchivelog mode (I know, it is not good), so there is no archive log files and
    restore database is the only solution.
    . startup mount
    . restore database <== ran ok
    . ALTER DATABASE CLEAR LOGFILE GROUP 1 <== ran ok
    . ALTER DATABASE CLEAR LOGFILE GROUP 2 <== alter failed
    ORA-01624: log 2 needed for crash recovery of instance orcl (thread 1)
    Alter failed with group 2 and 3 because one was CURRENT,
    and one was ACTIVE.
    SQL> select group#, sequence#, members, status from v$log;
    GROUP# SEQUENCE# MEMBERS STATUS
    1 0 2 UNUSED
    3 189 2 CURRENT
    2 188 2 ACTIVE
    I also tried 'alter database open resetlogs', and got:
    ORA-01139: RESETLOGS option only valid after an incomplete database recovery
    I don't think I can recover the datafile since there is no archive log file here.
    Besides, if the restore database ran fine, why did I get SYSTEM01.DBF datafile needs media recovery message?
    Right now, the database is in mount mode, and I cannot open it. How do I solve the problem with
    ALTER DATABASE CLEAR LOGFILE GROUP for group 2 and 3?
    Any suggestion? Thank you for the input.
    The database is 11G running on Window.

    I would think the cold backup ran successfully. Here is the backup record:
    RMAN> run {
    2> shutdown immediate;
    3> startup mount;
    4> allocate channel c1 type disk;
    5> backup as backupset database;
    6> alter database open;
    7> }
    8>
    using target database control file instead of recovery catalog
    database closed
    database dismounted
    Oracle instance shut down
    connected to target database (not started)
    Oracle instance started
    database mounted
    Total System Global Area 535662592 bytes
    Fixed Size 1348508 bytes
    Variable Size 331353188 bytes
    Database Buffers 197132288 bytes
    Redo Buffers 5828608 bytes
    allocated channel: c1
    channel c1: SID=156 device type=DISK
    Starting backup at 29-NOV-09
    channel c1: starting full datafile backup set
    channel c1: specifying datafile(s) in backup set
    input datafile file number=00002 name=C:\APP\DAVECW\ORADATA\ORCL\SYSAUX01.DBF
    input datafile file number=00001 name=C:\APP\DAVECW\ORADATA\ORCL\SYSTEM01.DBF
    input datafile file number=00003 name=C:\APP\DAVECW\ORADATA\ORCL\UNDOTBS01.DBF
    input datafile file number=00005 name=C:\APP\DAVECW\ORADATA\ORCL\EXAMPLE01.DBF
    input datafile file number=00006 name=C:\APP\DAVECW\ORADATA\ORCL\INVENTORY01.DBF
    input datafile file number=00004 name=C:\APP\DAVECW\ORADATA\ORCL\USERS01.DBF
    channel c1: starting piece 1 at 29-NOV-09
    channel c1: finished piece 1 at 29-NOV-09
    piece handle=C:\APP\DAVECW\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_11_29\O1_MF_NNNDF_TAG20091129T230
    632_5K6VQY0W_.BKP tag=TAG20091129T230632 comment=NONE
    channel c1: backup set complete, elapsed time: 00:01:55
    channel c1: starting full datafile backup set
    channel c1: specifying datafile(s) in backup set
    including current control file in backup set
    including current SPFILE in backup set
    channel c1: starting piece 1 at 29-NOV-09
    channel c1: finished piece 1 at 29-NOV-09
    piece handle=C:\APP\DAVECW\FLASH_RECOVERY_AREA\ORCL\BACKUPSET\2009_11_29\O1_MF_NCSNF_TAG20091129T230
    632_5K6VVNJR_.BKP tag=TAG20091129T230632 comment=NONE
    channel c1: backup set complete, elapsed time: 00:00:03
    Finished backup at 29-NOV-09
    database opened
    released channel: c1
    Recovery Manager complete.

  • Clone A Database Instance on a different mount point.

    Hello Gurs,
    I need you help as I need to clone a 11.1.0.7 database instance to a NEW mount point on the same host. The host is a HP-UX box and my question is do I need to install oracle database software in this new mount point and then clone ?? or cloning to the NEW MOUNT point itself will create all the necessary software?. Please provide me any documents that will be helpful for the process.
    Thanks In Advance.

    882065 wrote:
    Hello Gurs,
    my question is do I need to install oracle database software in this new mount point and then clone ??No.
    or cloning to the NEW MOUNT point itself will create all the necessary software?.No: cloning a database on same host means cloning database files : it does not mean cloning Oracle executables. You don't need to clone ORACLE_HOME on same host.
    Please provide me any documents that will be helpful for the process.
    Try to use : http://www.oracle-base.com/articles/11g/DuplicateDatabaseUsingRMAN_11gR2.php
    Thanks In Advance.Edited by: P. Forstmann on 29 nov. 2011 19:53

  • How datafile being extended over different mount point automatically

    Hello,
    I would like to understand if I have like 20 datafiles created over 2 mount point. All of it being setup with auto extend of 1GB and maxfile of 10GB. All files are not a max size yet.
    10 datafiles at /mountpoint1 with free space of 50GB
    10 datafiles at /mountpoint2 with free space of 200MB
    Since mountpoint2 have absolutely no space for auto extend, will it keep extending datafiles at mountpoint1 until it hit the maxsize of each file?
    Will it cause any issue of having mountpoint could not be extended due to mountpoint2?

    Girish Sharma wrote:
    In general, extents are allocated in a round-robin fashionNot necessarily true. I used to believe that, and even published a 'proof demo'. But then someone (may have been Jonothan Lewis) pointed out that there were other variables I didn't control for that can cause oracle to completely fill one file before moving to the next. Sorry, I don't have a link to that converstation, but it occurred in this forum, probably some time in 2007-2008.Ed,
    I guess you are looking for below thread(s)... ?
    Re: tablespaces or datafile
    or
    Re: tablespace with multiple files , how is space consumed?
    Regards
    Girish SharmaYes,but even those weren't the first 'publication' of my test results, as you see in those threads I refer to an earlier demo. That may have been on usenet in comp.database.oracle.server.

  • Creating database in other mount point

    i had installed oracle 10g in mount point u01 but i want to create database in mount point u02
    how can i do this??
    i had to do this by using dbca

    This is the same as yesterday's question:
    creating mountpoint after installation
    Whatever you did yesterday when you typed "u01", just do the same but type "u02".
    There was no need to create a new post just for this.
    Perhaps you need to spend more time reading what others have already posted into various forum web sites instead of splattering your questions and expecting others to tell you every minuscule step to do.
    This thread is locked because you are just repeating yourself.

  • Checking the space for /archlog mount point script

    I have the below shell script which is checking /archlog mount point space on cappire(solaris 10) server. When the space usage is above 80% it should e-mail. When i tested this script it is working as expected.
    #!/usr/bin/ksh
    export MAIL_LIST="[email protected]"
    export ARCH_STATUS=`df -k /archlog | awk '{ print $5 }' | grep -v Use%`
    echo $ARCH_STATUS
    if [[ $ARCH_STATUS > 80% ]]
    then echo "archive destination is $ARCH_STATUS full please contact DBA"
    echo "archive destination /archlog is $ARCH_STATUS full on Cappire." | mailx -s "archive destination on cappire is $ARCH_STATUS full" $MAIL_LIST
    else
    exit 1
    fi
    exit
    When i scheduled a cron job it is giving different result. Right now /archlog is 6%, it should exit without e-mailing anything. But, i am getting the below e-mail from cappire server which is strange.
    subject:archive destination on cappire is capacity
    below is the e-mail content.
    6% full
    Content-Length: 62
    archive destination /archlog is capacity 6% full on Cappire.
    Please help me in resolving this issue - why i am getting the above e-mail, i should not get any e-mail with the logic.
    Is there any issue with the cron. Please let me know.

    user01 wrote:
    I have the below shell script which is checking /archlog mount point space on cappire(solaris 10) server. When the space usage is above 80% it should e-mail. When i tested this script it is working as expected.
    #!/usr/bin/ksh
    export MAIL_LIST="[email protected]"
    export ARCH_STATUS=`df -k /archlog | awk '{ print $5 }' | grep -v Use%`
    echo $ARCH_STATUS
    if [[ $ARCH_STATUS > 80% ]]
    then echo "archive destination is $ARCH_STATUS full please contact DBA"
    echo "archive destination /archlog is $ARCH_STATUS full on Cappire." | mailx -s "archive destination on cappire is $ARCH_STATUS full" $MAIL_LIST
    else
    exit 1
    fi
    exit
    When i scheduled a cron job it is giving different result. Right now /archlog is 6%, it should exit without e-mailing anything. But, i am getting the below e-mail from cappire server which is strange.
    subject:archive destination on cappire is capacity
    below is the e-mail content.
    6% full
    Content-Length: 62
    archive destination /archlog is capacity 6% full on Cappire.
    Please help me in resolving this issue - why i am getting the above e-mail, i should not get any e-mail with the logic.
    Is there any issue with the cron. Please let me know.Not a problem with cron, but possibly an issue with the fact that you are doing a string comparison on something that you are thinking of as a number.
    Also, when I'm piping a bunch of stuff together and get unexpected results, I find it useful to break it down at a command line to confirm that each step is returning what I expect.
    df -k /archlog
    df -k /archlog | awk '{ print $5 }'
    df -k /archlog | awk '{ print $5 }' | grep -v Use%
    A common mistake is to forget that jobs submitted from cron don't source the owning user's .profile. You need to make sure the script takes care of setting its environment, but that doesn't look to be the issue for this particular problem.

  • Having issue restarting database after switching mount point

    We needed to change the mount point of the disk that has our mount points, so instead of going to /opt, it going to /opt/san/mdb4. I have created a control file to use to start up the database, but I'm still having issues. Here is part of the control file:
    CREATE CONTROLFILE REUSE SET DATABASE "INVEST" RESETLOGS NOARCHIVELOG
    MAXLOGFILES 64
    MAXLOGMEMBERS 2
    MAXDATAFILES 254
    MAXINSTANCES 16
    MAXLOGHISTORY 1815
    LOGFILE
    GROUP 1 (
    '/opt/san/mdb4/db02/oradata/INVEST/redo01a.log',
    '/opt/san/mdb4/db03/oradata/INVEST/redo01b.log'
    ) SIZE 50M,
    etc....
    DATAFILE
    '/opt/san/mdb4/db02/oradata/INVEST/XXX.dbf',
    '/opt/san/mdb4/db03/oradata/INVEST/XXX01.dbf',
    '/opt/san/mdb4/db07/oradata/INVEST/XXX02.dbf',
    '/opt/san/mdb4/db07/oradata/INVEST/XXX03.dbf',
    '/opt/san/mdb4/db03/oradata/INVEST/XXX04.dbf',
    '/opt/san/mdb4/db03/oradata/INVEST/XXX05.dbf',
    '/opt/san/mdb4/db03/oradata/INVEST/XXX06.dbf',
    etc...
    CHARACTER SET UTF8
    All of the files are in the correct spot. I have also edited the init.ora file so that the directories are correct. This is the error that I get every time:
    CREATE CONTROLFILE REUSE SET DATABASE "INVEST" RESETLOGS NOARCHIVELOG
    ORA-01503: CREATE CONTROLFILE failed
    ORA-00449: background process 'DBW0' unexpectedly terminated with error 7446
    ORA-07446: sdnfy: bad value '' for parameter .
    I am running Oracle 8.1.7 on Sun Solaris 8 (SunOS 5.8).
    Does anyone have any idea what is going on? Let me know.

    According to error message you have syntax error in your create statement. double check it.
    ORA-07446: sdnfy: bad value 'string' for parameter string.
    Cause: The directory specified as the value for the stated parameter could not be used.
    Action: Make sure the directory you have specified is a valid directory/file specification.

  • Taking export on different mount point.

    We have two databases in one server. I want to  take export of one schema from one database and store it directly on other database's mount point due to space crunch. How can i do that. We are using solaris 5.10 as OS and database version is 11.2.0.3.

    Thanks for your quick reply. Here is what i tried:
    Server Name - unixbox02
    Source database name - DV01
    Target database name - DV02
    I want to take export of test schema from DV01 to "/orabkup01/DV02/data_dump". The test schema is 100gig+ in size and i dont have enough space on /orabkup01/DV01.
    I have created directory on DV01 named datadir1 as 'unixbox02:/orabkup01/DV02/data_dump'.
    Then granted read and write privilege to system.
    (Not sure to who else i need to grant this privilege)
    After that I ran the below script:
    expdp "'/ as sysdba'"  schemas=test directory=datadir1 dumpfile=a1.dmp logfile=a2.log grants=y indexes=y rows=y constraints=y
    But I have received the below error:
    ORA-39002: invalid operation
    ORA-39070: Unable to open the log file.
    ORA-29283: invalid file operation
    ORA-06512: at "SYS.UTL_FILE", line 536
    ORA-29283: invalid file operation
    I am new to oracle dba, hence, I am trying to explain as much as possible.

  • Moving the database in the same server

    Your replay:-
    for WIN NT: take a look at note 61747.1 Moving a 7.3 or 8.x database. And take a look at Note:99275.1, UNIX/NT: Moving a Database Within the Same Machine.
    Clarification: Where can i get the help for Note:61747.1 & 99275.1

    Sorry Sridhar.
    Take a look in MetaLink (www.metalink.com) and use the note ids to look up the information.
    Thanks.

  • Moving the database of SAP J2EE Adapter Engine on another host

    Hello all,
    I'd like to ask SAP/Oracle experts if there is any obstacle, why would not be able to move Oracle database of SAP J2EE Adapter Engine on other/different host after the installation.
    As I describe int he thread [Installation of J2EE Adapter Engine with DB on different host|Installation of J2EE Adapter Engine with DB on different host;, the SAPINST program doesn't offer the possibility to install J2EE Adapter Engine with DB on other host than the central instance.
    Only thing I'm not sure about is if this is done by any purpose for some database requirement, or it's simply not there.
    SAP J2EE AE is normal JAVA AS with some more functionalities and is also installed within SAP PI/XI installation, which can run in "distributed model" system.
    What do you think ?
    Thank you for your opinion in advance.
    Regards
    Tomas

    No opinion anyone ?
    Tom

  • SCCM 2012 Moving the Database to Another Drive

    Due to space constraints I need to move my SCCM 2012 R2 database to another drive on the same system(server). I will run preinst/stopsite command to stop all sccm services, but wanted to know if I could copy and move the DB and Log files to their
    new drive location, instead of performing a SQL backup and restore? We are currently running SCCM 2012 R2 with SQL 2012 SP1 CU4 as our backend database.
    Thanks,
    -Bern
    

    you may have already seen these posts... worth a quick look:
    http://social.technet.microsoft.com/Forums/systemcenter/en-US/3462715c-d1b7-4f1b-b5f2-1f35235e4179/move-sccm-database-to-another-drive?forum=configmgradminconsole
    http://support.microsoft.com/kb/2709082
    http://damatisystemcenter.com/2013/05/23/moving-sccm-2012-sp1-database-to-new-folder/

Maybe you are looking for