SAPDB : Archive the Log and Data files

Hi,
I am running a SAP WAS JAVA on my laptop with SAPDB as the database. Its used for my local developments.
I have 3 log files and 3 data files of 1 GB each. I want to reduce this to 1 log file and 1 data file to free some space on my Laptop.
How can this be done.
Regards,
Shubham

Hi,
well, you should check the <b>current filling level</b> of your database instance to see how large your single data volume should become.
Then, I would recommend you put your db in <b>admin</b> state and make a full data <b>backup</b>.
Then, you would need to do do a <b>recovery with initialization</b>. This is most conveniently done using the newest available version of the Database Manager GUI.
During this process, you should be able to specify your desired new volume setup (please keep the filling level from your the beginning in mind). For the backup to recover, choose the one you just made.
Hope that helps.
Kind regards,
Roland

Similar Messages

  • How to transport the CO and DATA file to SAP system

    Hello Experts ,
                          I want to know the procedure to import a transport request into an isolated SAP system ( which is not in the transport route ) using the CO and DATA files .
    I know that user should have adminstrator role to do this but i dont know how to do this .
    Regards
    Vivek

    Simply copy the source cofile and datafile from the source system to the target system and then import.
    Data file is located at /usr/sap/trans/<sid>/data with format R<transport_number>.<sid> and cofile is located at /usr/sap/trans/<sid>/cofile with format K<transport_number>.<sid>
    Copy them to the same location on the target system and then go to STMS -> Extras -> Add -> insert the name of the transport in format <sid>K<transport_number>, this will add the transport to the import queue then just simply import the transport.
    Regards
    Juan

  • SQL 2000 Log and data file free space

    HI ~ In SQL 2000, are there any tsql I can get the free space of the mdf and ldf file ?

    I would like to look into EACH ldf too like SQL 2005 onward using sysfiles
    in SQL 2000
    DBCC sqlperf(logspace) will just sum all the ldf size..... 
    Hello,
    Below would work in SQL2000.I did not tested it on 2000 ,I dont have 2000 anymore.This will give information for data and log file free space for all databases
    exec sp_msforeachdb 'SELECT name ,size/128.0 - CAST(FILEPROPERTY(name, ''SpaceUsed'') AS int)/128.0 AS AvailableSpaceInMB
    FROM sysfiles';
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Do iOS apps that are ONLINE translators (e.g requires Wifi or internet, such as Google Translate) store every translation you type in the Documents and Data files?

    I have several apps that are translators. Many of them are for offline use, (especially ones that cost money) but they keep eating away at my memory with every translation I type. What about the ONLINE translators, like GOOGLE TRANSLATE? Do they store the translations inthe app's Documents and Data, or not at all? Please help me! I'm dying to know! Also, does Google Translate charge you money for unlimited translations, or is it entirely free?

    No no no, thank you. However, I learned just now that "flash storage" does not equal "Documents and Data" because I noticed that the Documents and Data for Google Translate increases very slighty with every translation (faster if speaking the translation). But this doesn't bother me at all. Why? Becuz the app size is oh so small (2.3 MB) & every translator (all 50+ of them) is included for free (how dope is that?) without having to download any additional memory for each language. So when the time comes to delete and redownload it due to memory capacity shortage, it still requires only 2 MB to redownload it. I bet that won't harm my device at all. Hey, I can live with that!
    Thanks for your support for my question, Deggie, kudos!!!

  • APIs for accessing the ESS log and output files from UCM

    Hi,
    As I understand, the output and log files of an ESS job will be uploaded to UCM.
    We have a requirement where in we are building a simple error handling framework, which gets triggered after a series of ESS jobs are run. In the error handling framework based on some processing logic we need to e-mail the error or log files to the users. I am not able to find any information related to retrieving the log and output files for and ESS job from UCM. Any inputs on this will be appreciated.
    Thanks for your help,
    Thiru

    For accessing content in UCM you can use the RIDC APIs or UCM Web Services. Note that the UCM also provides features for workflow etc. that could potentially be used for notifications refer to the developer guide for details. For UCM related questions you may want to use this forum instead
    Jani Rautiainen
    Fusion Applications Developer Relations
    https://blogs.oracle.com/fadevrel/

  • Win 2008  WL 10.3.3 stdout appearing in .log and .out files

    Recently noticed a ballooning [ServerName].out file in the logs directory. In weblogic management console I do have it configured to redirect stdout and stderr to weblogic logging (.log file). Both the .log and .out file contain the same stdout/stderr information. I would like to eliminate the .out file if possible (since WL only rotates the .log), but cannot find where it is configured. The managed servers are NOT windows services (no -log option).
    Did not find any logging parameters in JAVA_OPTIONS or paramters in the startManagedSvc.cmd file.
    Is this something needing to be corrected at the application level? (log4j)

    opie wrote:
    Recently noticed a ballooning [ServerName].out file in the logs directory. In weblogic management console I do have it configured to redirect stdout and stderr to weblogic logging (.log file). Both the .log and .out file contain the same stdout/stderr information. I would like to eliminate the .out file if possible (since WL only rotates the .log), but cannot find where it is configured. The managed servers are NOT windows services (no -log option).
    Did not find any logging parameters in JAVA_OPTIONS or paramters in the startManagedSvc.cmd file.
    Is this something needing to be corrected at the application level? (log4j)Depends on what you are actually seeing in those files. Are you outputting log4j messages to a log file AND the console?
    Here is a snippet of the log4j configuration file that denotes writing to the console:
        <appender name="ConsoleAppender" class="org.apache.log4j.ConsoleAppender">
            <layout class="org.apache.log4j.PatternLayout">
                <param name="ConversionPattern" value="%d{yyyy-MM-dd hh:mm:ss} %-5p [%t] - %C{1}.%M -> %m%n" />
            </layout>
          </appender>
        <root>
            <level value="ALL" />
            <appender-ref ref="ConsoleAppender" />
        </root>Edited by: ForumKid2 on Dec 29, 2010 11:36 AM

  • How to clear the alertlog and trace file?

    since the database was created,the log and trace file have't been cleared.how to clear the alertlog and trace file?3tx!!

    Hi Friend.
    You can eliminate all the files ".TRC" (trace files) to purify the directory BDUMP. These are not necessary files in order that the Oracle Server works.
    The file Alert.log is a file that the Oracle Server can recreate if you eliminate it. When Oracle Server's process realizes certain action (for example ARCH), Oracle creates again the file (if it does not exist), or it adds text (if it exists) with the new income.
    It can happen, that appears some Bug if the file Alert.log does not exist. Though this is slightly possible.
    Anyhow I recommend to you in UNIX to use from the prompt: $> filename in order to take the size of the file to 0 bytes, without need to eliminate it. Is the same thing when you want to purify the listener.log, the sqlnet.log or the sbtio.log.
    I wait for my commentaries you be of great help.
    Bye Friend...

  • The file structure online redo log, archived redo log and standby redo log

    I have read some Oracle documentation for file structure and settings in Data Guard environment. But I still have some doubts. What is the best file structure or settings in Oracle 10.2.0.4 on UNIX for a data guard environment with 4 primary databases and 4 physical standby databases. Based on Oracle documents, there are 3 redo logs. They are: online redo logs, archived redo logs and standby redo logs. The basic settings are:
    1. Online redo logs --- This redo log must be on Primary database and logical standby database. But it is not necessary to be on physical standby database because physical standby is not open. It doesn't generate redo log. However, if don't set up online redo log on physical standby, when primary failover and switch standby as primary. How can standby perform without online redo logs? In my standby databases, online redo logs have been set up.
    2. Archived redo logs --- It is obviously that primary database, logical and physical standby database all need to have this log file being set up. Primary use it to archive log files and ship to standby. Standby use it to receive data from archived log and apply to database.
    3. Standby redo logs --- In the document, it says A standby redo log is similar to an online redo log, except that a standby redo log is used to store redo data received from another database. A standby redo log is required if you want to implement: The maximum protection and maximum availability levels of data protection and Real-time apply as well as Cascaded destinations. So it seems that this standby redo log only should be set up on standby database, not on primary database. Am my understanding correct? Because I review current redo log settings on my environment, I have found that Standby redo log directory and files have been set up on both primary and standby databases. I would like to get more information and education from experts. What is the best setting or structure on primary and standby database?

    FZheng:
    Thanks for your input. It is clear that we need 3 type of redo logs on both databases. You answer my question.
    But I have another one. In oracle ducument, it says If you have configured a standby redo log on one or more standby databases in the configuration, ensure the size of the current standby redo log file on each standby database exactly matches the size of the current online redo log file on the primary database. It says: At log switch time, if there are no available standby redo log files that match the size of the new current online redo log file on the primary database. The primary database will shut down
    My current one data gurard envirnment setting is: On primary DB, online redo log group size is 512M and standby redo log group size is 500M. On the standby DB, online redo log group size is 500M and standby redo log group size is 750M.
    This was setup by someone I don't know. Is this setting OK? or I should change Standby Redo Log on standby DB to 512M to exactly meatch with redo log size on primary?
    Edited by: 853153 on Jun 22, 2011 9:42 AM

  • How to extract the size and date from a given file

    Hi,
    I want to extract the size and date of a file (it can be either a video, audio or text file) that the user points to the location of. But I am not sure how. Does Java have an api that can do this? If not is there some other way of doing this? Can anyone help? Thanks in advance.

    Have a look at java.io.File, specifically
    public long lastModified()
    This format returned (I find) is nasty, so then use java.util.Date (or java.sql.Date, look the same on the surface to me) to format it.
    Cheers,
    Radish21

  • Block size in tt for writing data to transaction log and checkpoint files

    Hello,
    what block size is TimesTen using when its writing data to transaction log and checkpoint files? Does it use some fixed block size during filesystem writes?

    Although in theory logging can write 2 KB blocks in almost all circumstances it will write 4 KB or larger so yes a filesystem with a 4 KB block size is fine for both checkpointing and logging.
    Chris

  • To read COMTRADE file in Labview there is an example provided. Can somebody provide the sample .cfg and .dat files required for its working?

    To read COMTRADE file in Labview there is an example provided. Can somebody provide the sample .cfg and .dat files required for its working?

    Thanks for the reply.
    But this library doesn't contain any sample .cfg and .dat files which can be read and understood. Can you please provide the same?

  • Deleted files from the Original and Data folders - help?

    Hello,
    I imported my fiance's iphoto library to try to recover photos that I lost from when my hard drive crashed. They first were not importing into iphoto (while the files were on the harddrive), so I placed them in my Originals/Data iphoto folders. Now I have 6-8 versions of files that I don't even necessarily need, mostly thumbnails and a few of the original files.
    I know I have made a lot of mistakes... so at this point what do I do- I've stopped manually deleting folders from these categories, but I do not want to lose what precious pictures I had from before the harddrive failure, and how do I get back to a iphoto library without duplicates?
    Please advise!!! Thank you!

    My library still opens within iPhoto- will it continue to do so? Do I need to do anything now (since I have already messed with the Originals and Data folders?)
    In addition, iPhoto now shows me blanks for some pictures/events, is there a way to have iPhoto scan the entire computer and realize files are missing and to ignore them?
    Therefore it is not a fully functioning iPhoto library - it is corrupted
    3 - if you do not have a fully functioning iPhoto library then quit iPhoto and drag the iPhoto library to the desktop, launch iPhoto and create a new library and then open the bad iPhoto library on the desktop and open the originals folder and drag each folder from it onto the iPhoto icon in the dock (or the iPhoto event window) to create events with all of your photos, then back up and run duplicate annihilator
    Since you have corrupted your library you can do anything you want to it before you create a new one from the remaining photos
    LN

  • Basic question on log and temp files

    Hi,
    I need to create a new database in this way:
    - 2 GB for the data file
    - 1 GB for log and temp files
    This is the command I'm issuing to create the db:
    create tablespace xyz datafile '<path>\xyz.dbf' size 2000M reuse;
    How can I say the size of log and temp files?
    Thank you
    Nicola
    Ps-I'm on windows platform

    user575754 wrote:
    - 1 GB for log and temp filesWhat's the link between log file size and tempfile ?
    This is the command I'm issuing to create the db:
    create tablespace xyz datafile '<path>\xyz.dbf' size 2000M reuse;This command does not actually create a database, but a tablespace. If you ran it, that means your db aleady exists.
    So, not sure to follow you on this thread.
    Nicolas.

  • Location of Redo log and control files?

    Dear all,
    I am checking the location of redo log and control files, but found that the redo log file (like log02a.dbf ....) in the same directory of data files. However, I couldn't find any control files in the data files directries.
    What could be the location of control files?
    Amy

    select name
      from v$controlfile
    or
    show parameter control_filesKhurram

  • External portal capturing internal portal URL in Log and trace file

    Hi,
    We are facing one issue in portal like we have two portals for internal (Intranet) and external (Internet) users.
    Once users logged in the application and try to get the information about mylink from the external portal link (internet) they should not get any information about the internal portal.
    But in log and trace file we can see the external portal link capturing the internal portal URL.
    We need to find, from where system capturing the internal portal URL.
    Thanks.

    The tkproffed trace file is in seconds.
    "set timing" is in hh:mi:ss.uu format. So 00:00:01.01 is 1.01 seconds.
    You have to remember that most of these measurements are rounded. While your trace file says it contains one second of trace data, you know it's more.
    One excellent resource for trace files is "Optimizing Oracle Performance" by Cary Millsap & Jeff Holt. (http://www.amazon.com/Optimizing-Oracle-Performance-Cary-Millsap/dp/059600527X ) I thought I knew trace files before, but this book brings your knowledge to a whole new level.
    There is also an excellent WP by Cary Millsap ( http://method-r.com/downloads/doc_details/10-for-developers-making-friends-with-the-oracle-database-cary-millsap ) that gives you some insight.

Maybe you are looking for