Impact on changing DATE

Gurus,
Our production database server currently is set to a wrong time (7 mts ahead). We are planning to sync it up with the network time server. Can you tell what are the implications in doing this? I am assuming this has to be done when DB is down.
Can someone point me to a note reference/websites etc?
Thanks,
Arun

Check this Note from Oracle
Because Oracle tracks the sequence of events in the database using the System
Commit Number (SCN), changing the system clock for daylight savings time will
have no effect on database operation. The only point where the time change
can have potentially harmful effect is during time-based recovery.
Time based recovery requires checking of the actual time the transaction was
recorded in the logfile. Every log record has a time stamp associated with
it. If the system manager for some reason changes the system clock, Oracle
Support recommends shutting down the database and taking a cold backup ( or a
hot backup if preferred). If for some reason a dba has to go back to a backup
which was taken prior to the system clock change and rollforward, recovery
works just fine except for time based recovery (Note that time based recovery
works fine if the system clock is moved forward in time). When the system
clock is changed backwards, its possible that there could be two redo records
with the same time stamp. If time based recovery is done in this scenario,
since ORACLE applies only redo entries that were written prior to a specified
time, ecovery stops when it finds the first redo record which has that
specified time.

Similar Messages

  • Impact of Changing Data Package Size with DTP

    Hi All,
    We have delta dtp to load data from DSO to infocube. Default data package size with dtp is 50,000 records.
    Due to huge no of data, internal table memory space is used and data loading get fails.
    Then we changed the data package size to 10,000, which executes the data load successfully.
    DTP with package size of 50,000 took 40 minutes to execute and failed, but DTP with package size of 10,000 took 15 minutes (for same amount of data).
    Please find below my questions:
    Why a DTP with bigger size of packet runs longer than a DTP with lower packet size ?
    Also by reducing the standard data package size 50,000 to 10,000, will it impact any other data loading?
    Thanks

    Hi Sri,
    If your DTP is taking more time then check your transformation .
    1.Transformation with Routines always take more time so you if you want to reduce the time of execution then routine should be optimized for good performance .
    2.Also check if you have filter at DTP level .Due to filters DTP takes long time .If same data get filtered at routine level it take much lesser time .
    3.If you cannot change routine then you can set semantic keys at your DTP .The package data will be sorted as per semantic keys and thus it may be helpful at routine level for fast processing.
    4.Your routine is getting failed due to  internal table memory space so check if you have select statement in routine without FOR ALL ENTRIES IN RESULT_PACKAGE or SOURCE_PACKAGE line .if you will use this It will reduce record count .
    5.Wherever possible delete duplicate records and if possible filter useless data at start routine itself .
    6.Refresh internal table if data no longer needed .If your tables are global then data will be present at every routine level so refreshing will help to reduce size.
    7.The maximum memory that can be occupied by an internal table (including its internal administration) is 2 gigabytes. A more realistic figure is up to 500 megabytes.
    8.Also check no of jobs running that time .May be you have lots of jobs active at the same time so memory availability will be less and DTP may get failed .
    Why a DTP with bigger size of packet runs longer than a DTP with lower packet size ?
    *Start and end routine works at package level so routine run for each package one by one .By default package have sorted data based on keys (non unique keys (characteristics )of source or target) and by setting semantic keys you can change this order.So Package having more data will take more time in processing then package have lesser data .
    by reducing the standard data package size 50,000 to 10,000, will it impact any other data loading?
    It will only impact running of that load .but yes if lots of other loads are running simultaneously then server can allocate more space to them .So better before reducing package size just check whether it is helpful in routine performance (start and end ) or increasing overhead .
    Hope these points will be helpful .
    Regards,
    Jaya Tiwari

  • Impact of changing data type of a field in Database table

    Hi All,
    I need to change the data type of a field(which is not a key field and it has no dependency with any other field) from NUMC to CHAR by maintaining the same length. Please let me know if there will be any impact in doing this. Hope the following things need to be taken care:
    - Take backup of entire data in the table before doing the change
    - After changing the data type, I need to set the option 'Save Data' and then 'Activate and Adjust Database' in 'Database Utility'.
    - Use 'Where Used List' to check the related objects.
    Please let me know if there is any impact or any thing else need to be taken care apart from the above things.
    Thanks in advance.
    Regards
    Vidhya.

    Hi,
    even if the length is same there would be no impact.
    just v need to adjust the database in the database utility.
    check the table maintenence generator also. if not reflected there u need to delete it and create it again.
    reward if useful,
    teja

  • Impact of changing/enhance an existing master data object

    Hi Experts,
    I have a new requirement to enhance 0comp_code in order to add 2 extra fields in the master data.
    As 0comp_code is an infoobject used in lots of cubes, I would like to know which is the impact of changing it.
    Which actions should I take into account before doing any changes.
    Many thanks in advance!

    Hi,
    If you are adding any new fields to infoobject, you will not have any impact to existing info providers which uses that infoobject.
    If the attr u add is a navigational attr and if you want ur info provider to have it, then you will have to enable it and activate the info provider and the update rule as well.
    is your data source modified to bring in those new fields? or is it going to be manually maintained?
    Regards,
    Pradhiba

  • Upgrade of Database with Oracle Change Data Capture

    Hello,
    We are upgrading and moving our database to a different server.
    The move is from 10G R1 database on Solaris to 11G R2 on Linux.
    Our database uses Oracle Change Data Capture.
    What is the best way to perform this migration? Unlike in the approach below, ideally, it would be without any manual steps to drop and recreate CDC subscriptions, change tables, etc.
    Thanks.
    Considerations for Exporting and Importing Change Data Capture Objects
    http://docs.oracle.com/cd/B13789_01/server.101/b10736/cdc.htm#i1027532
    Starting in Oracle Databse 10g, Oracle Data Pump is the supported export and import utility for Change Data Capture. Change Data Capture change sources, change sets, change tables, and subscriptions are exported and imported by the Oracle Data Pump expdp and impdp commands with the following restrictions:
    After a Data Pump full database import operation completes for a database containing AutoLog Change Data Capture objects, the following steps must be performed to restore these objects:
    1. The publisher must manually drop the change tables with the SQL DROP TABLE command. This is needed because the tables are imported without the accompanying Change Data Capture metadata.
    2. The publisher must re-create the AutoLog change sources, change sets, and change tables using the appropriate DBMS_CDC_PUBLISH procedures.
    3. Subscribers must re-create their subscriptions to the AutoLog change sets.

    Hello,
    I opened SR with Oracle Support, they are suggesting to perform full database import/export
    Change Data Capture change sources, change sets, change tables, and subscriptions are exported and imported by the Oracle Data Pump expdp and impdp commands with the following restrictions.
    Change Data Capture objects are exported and imported only as part of full database export and import operations (those in which the expdp and impdb commands specify the FULL=y parameter). Schema-level import and export operations include some underlying objects (for example, the table underlying a change table), but not the Change Data Capture metadata needed for change data capture to occur."
    CDC has different implementation methods:
    You may use the below query to determine-
    select SOURCE_NAME, SOURCE_DESCRIPTION, CREATED, SOURCE_TYPE, SOURCE_DATABASE, SOURCE_ENABLED from change_sources;
    – Synchronous CDC: with this implementation method you capture changes
    synchronously on the source database into change tables. This method uses
    internal database triggers to enable CDC. Capturing the change is part of the
    original transaction that introduces the change thus impacting the performance
    of the transaction.
    – Asynchronous Autolog CDC: this implementation method requires a staging
    database separate from the source database. Asynchronous Autolog CDC uses
    the database's redo transport services to transport redo log information from
    the source database to the staging database1. Changes are captured at the
    staging database. The impact to the source system is minimal, but there is some
    latency between the original transaction and the change being captured.
    As suggested in the document-
    Change Data Capture objects are exported and imported only as part of full database export and import operations (those in which the expdp and impdb commands specify the FULL=y parameter). Schema-level import and export
    operations include some underlying objects (for example, the table underlying a change table), but not the Change Data Capture metadata needed for change data capture to occur.
    ■ AutoLog change sources, change sets, and change tables are not supported.
    Starting in Oracle Database 10g, Oracle Data Pump is the supported export and import utility for Change Data Capture.
    Re-Creating AutoLog Change Data Capture Objects After an Import Operation
    http://docs.oracle.com/cd/B19306_01/server.102/b14223/cdc.htm#i1027532
    After a Data Pump full database import operation completes for a database containing AutoLog Change Data Capture objects, the following steps must be performed to restore these objects:
    a. The publisher must manually drop the database objects underlying AutoLog Change Data Capture objects.
    b. The publisher must re-create the AutoLog change sources, change sets, and change tables using the appropriate DBMS_CDC_PUBLISH procedures.
    c. Subscribers must re-create their subscriptions to the AutoLog change sets.

  • What are the impacts of changing material group of a Material ?

    Hi,
    Could some one tell me in detail about the impacts of changing material group of a material that is active in procrement?
    Thanks,
    Veera

    Hi,
    If you are changing it from one valid group to another valid group and both will still be there afterwards, then the impact is small, but important.
    The data in the Info systems and BW etc. will have been updated with the the "old" material group and so if you tried to compare them you would find differences.
    It is something that does happen from time to time and as long as the two groups will still be ther afterwards there should be no reason why this would cause any technical or system problems. The effect is purely on the historic data.
    I have seen this done several times in many implementations and so far it has not caused any major problems apart from some reports "appearing" wrong if you compare them against other reports that involve historical movements.
    Steve B

  • Impact of changing a Project Name on PWA 2013

    Hi Ladies and Gents
    I know that if I change the project name of a current project on the Name field on the PDP it will update the Project Plan name. I also know that I  need to change the project site name and I know what the workaround is to ensure that the project plan still
    associates with the project site.
    What I want to ask, is there any other impact to historic data if I make a change to the above
    Have an awesome day
    Willem

    Hello Willem,
    Changing a project name will be fine as the system works off the GUIDs. Test this process on a test project before so that you are satisfied this works for you.
    Paul
    Paul Mather | Twitter |
    http://pwmather.wordpress.com | CPS

  • Impact of change of field Country of origin in material master

    HI EXPERTS,
    what is the impact of change of field Country of origin in material master(Foreign trade data ) on sales order ,delivery and invoice ?
    also will it impact the existing document if we had changed the country of origin in material master?
    Appreciate your early response!
    Thanks
    Manish gupta

    hi
    this is to inform you that
    you are telling about FOREIGN TRADE & CUSTOMS.
    it will be having impact on sales delvery and billing as well as sales document is optional and the
    delivery document will be stopped with GTS error with LEGAL CONTROL ISSUE.
    if you over comes that then you will be caught in billing
    in billing document will be stopped with ERROR called FOREIGN TRADE DATA INCOMPLETE.
    because it deals with FTC.
    regards
    balajia

  • Impact on Master data Infoobject

    Hi Bwers,
    Currently we are making changes to exsiting master data infoobject (Business partner) in the system through adding one more Attribute. And we are including 2 new info objects to exsiting DSO also.
    I was asked to investigate impact of these chanes in systems. Pls provide your valuable suggestions about this activity.
    1.  What about the impact on other info objects (say 0CUSTMOER) which is reference to Bussiness partner.
    2. If any infoset incluing this infoobject, what is the impact on this Infoset / Multiprovider.
    3. If Business partner has been used as a navigational attribute for any other Infoobjects.
    Impact on DSO:-
    1. Impact on above data targets.
    2. And transformation between DSO to above all the infoproviders.
    How to investigate this things in the system.
    Regards,
    Venkat

    1. What about the impact on other info objects (say 0CUSTMOER) which is reference to Bussiness partner.
    > Adding an Attribute to wont affwct 0Customer.
    2. If any infoset incluing this infoobject, what is the impact on this Infoset / Multiprovider.
    > You need to adjust your inoset/MP. In case you want this attribute in the report, you have to select this object in infoset/MP.
    3. If Business partner has been used as a navigational attribute for any other Infoobjects.
    > It can still be used like that.
    Impact on DSO:-
    1. Impact on above data targets.
    ---> need to activate the DSO and load the data again. Need to change the transformation as well.
    2. And transformation between DSO to above all the infoproviders.
    > Those will get inactive. You have to activate it again.
    Thanks...
    Shambhu

  • How to get changed data in ALV in Web Dynpro for ABAP

    METHOD on_data_check .
    DATA:
        node_spfli                          TYPE REF TO if_wd_context_node,
        node_sflight                        TYPE REF TO if_wd_context_node,
        itab_sflight2                        TYPE if_display_view=>elements_sflight.
      node_spfli = wd_context->get_child_node( name = if_display_view=>wdctx_spfli ).
      node_sflight = node_spfli->get_child_node( name = if_display_view=>wdctx_sflight ).
      CALL METHOD node_sflight->get_static_attributes_table
        IMPORTING
          table = itab_sflight2.
    this code is ..get all data(changed and not changed)
    but i want get changed data only, not all data.
    how to get changed data?
    Edited by: Ki-Joon Seo on Dec 27, 2007 6:04 AM

    Hi,
    To get only the changed data in the ALV grid of a WD, you need to capture the "ON_DATA_CHECK" of the ALV grid.
    To this please do the following in the ALV initialization of the ALV table settings :
        lr_table_settings->set_data_check(
                IF_SALV_WD_C_TABLE_SETTINGS=>DATA_CHECK_ON_CELL_EVENT ).
    You may also do this:
        lr_table_settings->set_data_check(            IF_SALV_WD_C_TABLE_SETTINGS=>DATA_CHECK_ON_CHECK_EVENT)
    The above two ways would depend on when do you need to check for the changed data. If you want to check the data as soon as it is entered, then use the first method. Else, use the second method.
    You need to register an EVENT HANDLER for this event.(You may do this in your VIEW or Component Controller).
    In this Event handler, you would find an importing parameter R_PARAM which is a ref type of      IF_SALV_WD_TABLE_DATA_CHECK.
    The attribute T_MODIFIED_CELLS of this interface IF_SALV_WD_TABLE_DATA_CHECK will contain the modified cells of the ALV with the old & new values.

  • Can not change data in Input Ready Query

    All,
    I have created an input ready query on an aggreagation level which has one transactional cube underlying it however, after run the query I am unable to change data in the cells even though I get a warning message saying query is ready for input
    Note:
    1. The underlying realtime cube is in "Planning Mode:Loading not allowed" state in rsa1
    2. I tried including all characteristics so that the record can be uniquely identified at that aggregation level).
    3. I am not using any characteristic restrictions
    4. I have not set any data slices
    5. I have set the property for Planning "Start query in Change Mode"
    6. I have set the 'Key Figure' Planning Properties to "Can be changed Manually or via Planning Functions" for the key figures
    7. I do not have currency/unit in row/column.
    8. There are any locks occurring
    Any pointers welcome.
    Thanks,
    Milind

    Hi Milind,
    Was wondering how was your problem solved. I get the same message when I run the query in web from query designer and in bex analyzer the cells are not updated with the changes that I make.
    I am hitting my head against the wall to figure what could be the problem as my plan query is simply based on aggregation level created directlty upon the real time info cube.
    Have cross checked all the points that you mentioned in this post but I cells do not change.
    There should be some way around to solve this problem.
    Thanks

  • Can we change data in string object.

    Can we change data in string object.

    Saw this hack to access the char[]'s in a String in another thread. Beware that the effects of doing this is possible errors, like incorrect hashCode etc.
    import java.lang.reflect.*;
    public class SharedString {
            public static Constructor stringWrap = null;
            public static String wrap(char[] value, int offset, int length) {
                    try {
                            if (stringWrap == null) {
                                    stringWrap = String.class.getDeclaredConstructor(new Class[] { Integer.TYPE, Integer.TYPE, char[].class });
                                    stringWrap.setAccessible(true);
                            return (String)stringWrap.newInstance(new Object[] { new Integer(offset), new Integer(length), value });
                    catch (java.lang.NoSuchMethodException e) {
                            System.err.println ("NoMethod exception caught: " + e);
                    catch (java.lang.IllegalAccessException e) {
                            System.err.println ("Access exception caught: " + e);
                    catch (java.lang.InstantiationException e) {
                            System.err.println ("Instantiation exception caught: " + e);
                    catch (java.lang.reflect.InvocationTargetException e) {
                            System.err.println ("Invocation exception caught: " + e);
                    return null;
            public static void main(String[] args) {
                    char[] chars = new char[] { 'l', 'e', 'v', 'i', '_', 'h' };
                    String test = SharedString.wrap(chars, 0, chars.length);
                    System.out.println("String test = " + test);
                    chars[0] = 'k';
                    chars[1] = 'a';
                    chars[2] = 'l';
                    chars[3] = 'l';
                    chars[4] = 'a';
                    chars[5] = 'n';
                    System.out.println("String test = " + test);
    } Gil

  • Can't view changed data in journal data

    Hi,
    I have implemented JKM Oracle 10g Consistent Logminer on Oracle 10g with the following option.
    - Asynchronous_mode : yes
    - Auto_configuration : yes
    1. Change Data Capture -> Add to CDC, 2.Subscriber->subscribe (sunopsis),
    3. Start Journal
    The journal has been started correctly wothout errors. The journalized table has always the symbol "green clock". All is gook working.
    And then i inserted 1 record in source table, but i can't view changed data in journal data. I can't understand why journal data was generated.
    There are no errors.
    Help me !!!

    Did your designer was on the good context ?
    Look the list box at the top right of the Designer interface.
    You must have the same as the one where you define your journalization.

  • I can not map field after changing data source location

    Hi
    I have a small problem that I got a report file and database from my customer, after that I setup database, open the file and change data source to my setting. but some filed can not map. The field mapping widonw does not display all field in the table. Of course I have checked the missing fields are existing in the table.
    OS:Windows7
    DB:Oracle11
    CR:XI Release 2
    Does anyone have an idea?

    hi,
    In Map Fields window, there is an option "Match Type".
    Please Unchek that option, so that you will be able to see all the fields from that table.
    Also, while mapping please verify the datatypes of source and target fields.
    Regards,
    Vamsee

  • Can you help me about change data captures in 10.2.0.3

    Hi,
    I made research about Change Data Capture and I try to implement it between two databases for two small tables in 10g release 2.MY CDC implementation uses archive logs to replicate data.
    Change Data Capture Mode Asynchronous autolog archive mode..It works correctly( except for ddl).Now I have some questions about CDC implementation for large tables.
    I have one senario to implement but I do not find exactly how can I do it correctly.
    I have one table (name test) that consists of 100 000 000 rows , everyday 1 000 000 transections occurs on this table and I archive the old
    data more than one year manually.This table is in the source db.I want to replicate this table by using Change Data Capture to other stage database.
    There are some questions about my senario in the following.
    1.How can I make the first load operations? (test table has 100 000 000 rows in the source db)
    2.In CDC, it uses change table (name test_ch) it consists of extra rows related to opearations for stage table.But, I need the orjinal table (name test) for applicaton works in stage database.How can I move the data from change table (test_ch) to orjinal table (name test) in stage database? (I don't prefer to use view for test table)
    3.How can I remove some data from change table(name test_ch) in stage db?It cause problem or not?
    4.There is a way to replicate ddl operations between two database?
    5. How can I find the last applied log on stage db in CDC?How can I find archive gap between source db and stage db?
    6.How can I make the maintanence of change tables in stage db?

    Asynchronous CDC uses Streams to generate the change records. Basically, it is a pre-packaged DML Handler that converts the changes into inserts into the change table. You indicated that you want the changes to be written to the original table, which is the default behavior of Streams replication. That is why I recommended that you use Streams directly.
    <p>
    Yes, it is possible to capture changes from a production redo/archive log at another database. This capability is called "downstream" capture in the Streams manuals. You can configure this capability using the MAINTAIN_* procedures in DBMS_STREAMS_ADM package (where * is one of TABLES, SCHEMAS, or GLOBAL depending on the granularity of change capture).
    <p>
    A couple of tips for using these procedures for downstream capture:
    <br>1) Don't forget to set up log shipping to the downstream capture database. Log shipping is setup exactly the same way for Streams as for Data Guard. Instructions can be found in the Streams Replication Administrator's Guide. This configuration has probably already been done as part of your initial CDC setup.
    <br>2) Run the command at the database that will perform the downstream capture. This database can also be the destination (or target) database where the changes are to be applied.
    <br>3) Explicitly define the parameters capture_queue_name and apply_queue_name to be the same queue name. Example:
    <br>capture_queue_name=>'STRMADMIN.STREAMS_QUEUE'
    <br>apply_queue_name=>'STRMADMIN.STREAMS_QUEUE'

Maybe you are looking for

  • Updating data in a JLabel

    i'm trying to update the contents of a JLabel for my program, such that when the user executes a certain task a certain number changes to another number referenced from an array. I know that the contents of the tour3seat[0] = something other than 0 w

  • Commandlink and checkbox in a region(.jsff).

    Hi all, I have a commandlink and checkbox in a region(.jsff). when i tried to call a method using propertychangeListener or Actionlistener .. methosd is not calling ... Plase give any solution...

  • Swf Not getting focus in Opera

    Hi All            I did embed an swf file in html.The file is working fine in all browsers except opera(all versions). I am facing problem that flash didn't get focus when application initially load.User has to click on page. Then the application wor

  • Can't figure out how to read from .txt file...

    It looks so simple: I have a txt file with one line in it: 1038It is actually a highscore for a little game. But how do get this out of the file into an int? Thank you.

  • VLAN ESW-540

    Hi, I have a new switch ESW-540 24-port, I upgrade firmware to 2.1.19. I build 2 VLAN for 2 networks: VLAN 2 and VLAN 3 . I don't use the default VLAN 1. This is my configuration: port e1 and e2 : ACCESS PORT on UNTTAGGED VLAN 2 port e3 and e4 : ACCE