DTP got failes due to overlapping time intervals of data records

Hi,
Can any one suggest me how to solve this issue....
the problem is when iam loding Master Data DTP, the DTP is showing error message that Overlapping time intervals of data records.I mean the start data and end data for these error records are same.
Eg: For some employees the start date and end data are same.
Please advice , what can be done...
Immediate reply is realy appreciated.

Hi Shaik,
By checking this check box ...
You won't get any duplicate records, before loading the data DTP removes the duplicate records and loads into Master data.
So, you can avoid the duplicate records by checking this.
Regards,
Ravi Kanth

Similar Messages

  • Overlapping time intervals of data record 33 with 34 ,46 with 47 etc..

    Hi Freinds,
    when i was trying to load master data getting the below error message in DTP/Info package.
    The time interval ['20070920 '/' 20070702] & (from/to) for the data records 33 and 34 & overlaps in charecterstics 0HRPOSITION &.
    Can you please explain me how to correct this error.
    Your input 'll be much appreciated and points 'll be assgin.
    Regards,
    Syam

    Hi Gurus,
    Please see the below error message which i got when i was trying to load master data attributes to HR infoobject.
    Diagnosis:
        The time interval ['20070920'/'20070702'] & (from/to) for the data
        records 33 & and 34 & overlaps in characteristic 0HRPOSITION &.
    System Response
    Procedure
        If this message appears during a data load, maintain the attribute in
        the PSA maintenance screens. If this message appears in the master data
        maintenance screens, leave the transaction and call it again. This
        allows you to maintain your master data.
    Procedure for System Administration
    Out 1000 records getting error for 12 records.
    I could n't find in infopackage that Ingnore Duplicate Records and that info object is time dependent.
    Please put some inputs to solve this issue.
    Points 'll awarded for helpul answers.
    Thnaks,
    Syam

  • Overlapping Time intervals of data record 33 with 44,46 with 47 so on...

    Hi Friends,
    when i was trying to load the master data attributes to info object getting the bellow error message:
    Diagnosis
        The time interval ['20070920'/'20070702'] & (from/to) for the data
        records 33 & and 34 & overlaps in characteristic 0HRPOSITION &.
    System Response
    Procedure
        If this message appears during a data load, maintain the attribute in
        the PSA maintenance screens. If this message appears in the master data
        maintenance screens, leave the transaction and call it again. This
        allows you to maintain your master data.
    Procedure for System Administration
    Geeting error for 12 records out of 1200.
    Already my info object is time dependent master data.
    to ignore the dulpicate records i couldn't find any option in that info package..
    please let me know the error ...
    points 'll be assigned for helpul answers..
    Regards,
    Syam

    Hi ,
    You can go for PSA maintainance   or
    If u r using DTP then in DTP Ececution screen under Update tab there is a check box
    Handel the duplicate records
    click it and activate the DTP and Execute.
    Thanks
    RK

  • 'Overlapping Time Intervals' error while loading the Master Data

    Guys,
              Need you help with this. I am trying to load the Master Data (attribute) for InfoObject 0FMD_FRSITM using DTP from DataSource 0FMD_FRSITM_ATTR and am getting an error 'Overlapping Time Intervals'.
    This is happening due to overlapping Valid From/Valid To date's for the same Financial Statement Item (0FMD_FRSITM) with different Editions.
    If I select the option 'Ignore Duplicate Records' while loading the data, it is successful. But not all the records are loads and Editions are missing.
    Below is an example of how the data looks like.
    FRSITM         Valid To                  Valid From                 Edition
    1001             12/31/9999             04/01/2010                2010.1.1
    1001             01/31/2010             03/01/2010                2010.1.2
    1001             12/31/9999             02/01/2010                2010.1.3
    In the above scenario, the Valid To Data for record 1 and record 3 is the same (12/31/9999) and the record 3 is being ignore when I select the option 'Ignore Duplicate Records'.
    And also I see that the Edition is not the Key Field.
    Any suggestions or ideas? Appreciate your help.
    Thanks.

    FRSITM Valid To Valid From Edition
    1001 12/31/9999 04/01/2010 2010.1.1
    1001 01/31/2010 03/01/2010 2010.1.2
    1001 12/31/9999 02/01/2010 2010.1.3
    1. While selecting ignore duplicate, make sure to-date is not part of symantic key.
    2. if system comes across two same records it picks up the latest one.
    3. It shall merge the timelines is overlap is found.
    4. I hope Edition is time dependent attribute
    I believe infoobjet key is FRSITM Valid To.
    In your case because first record has more period and may be the latets record system ignore the earlier one with the same time interval.
    have a look at Q table and you shall be able to figure out for your self.
    Hope it helps
    Vikash

  • Transport Request got failed due to Source system does not Exist

    Dear All,
    I Have collected All data soruces which are related to R3 data sources and Flat File data sources...
    After import the transport request it got failed due to Flata file source system does not exist in BI quality system.
    But all data sources related to R3 and R3 Source system has been moved which are available in BI Quality sytem also.
    why Flat File source and data source are moving.. why i am not sure it was throuing error message called Source System does not exist..
    Thanks,
    Ven

    Dear Venkat ,
    Go to Quality system .
    Use RSA1 Transaction --> go to Tools --> Manage Source System :
    Here you can map the entry  of system ..
    Second : go to Development system under RSA1> Transport Connection and click on Source system button first --> select your source system ---you  can select multiple entry also ..collect your data source ( with before and after option ) and deselect which object you do not need ..And select collection mode to automatically ...
    use the transport button to create  transport request and release from development system .
    I hope you will not get error in Q system .
    If still facing error , Please let me know.
    Regards
    Vikas Sharma
    Robert Bosch
    Edited by: Vikas Sharma-Bosch on May 5, 2011 5:34 PM

  • The Job Failed Due to  MAX-Time at ODS Activation Step

    Hi
            I'm getting these errors "The Job Failed Due to  MAX-Time at ODS Activation Step" and
             "Max-time Failure"
    How to resolve this failure?

    Hi ,
    You can check the ODs activation logs in ODs batch montior >> click on taht job >>> job log
    for this first you check in sm37 how many jobs are running if there are many otehr long running jobs then . the ODs activation will happen very slowly as the system is overloaded .. due to the long running jobs ..this causes the poor performance of system ..
    for checking the performance of system .. you first check the lock waits in ST04..and check wether they are progressing or nt ..
    check sm66 .. for the no of process running on system ..
    check st22 >> for short dumps ..
    check Os07> to check the cpu idle tiem if its less tehn 20% . then it means the cpu is overlaoded .
    check sm21 ...
    check the table space avilable in st04..
    see if the system is overlaoded then the ODS wont get enough work process to create the backgorund jobs for that like(Bi_BCTL*)..jobs .the updation will happen but very slowly ..
    in this case you can kill few long running jobs which are not important .. and kill few ODs activations also ..
    Dont run 23 ODs activation all at a time .. run some of them at a time ..
    And as for checking the key points for data loading is check st22,cehck job in R/3 , check sm58 for trfc,check sm59 for Rfc connections
    Regards,
    Shikha

  • SCOM not generating alerts & The request to synchronize state for OpsMgr Health Service identified by "bdb1712b-0a9a-bbe3-9630-41f2da9f64a1" failed due to the following exception "System.Data.SqlClient.SqlException: Cannot open database "OperationsManager

    Hi Anyone can assisting on this. The SCOM 2007r2 i managed suddenly do not pump out alerts.
    I have perform full scom service restart(data access, agent, config). it no load up related datawarehouse like(stored, write)
    I also restarted my sql db services. It show me the error like this. Anyhow i can fix it
    The request to synchronize state for OpsMgr Health Service identified by "bdb1712b-0a9a-bbe3-9630-41f2da9f64a1" failed due to the following exception "System.Data.SqlClient.SqlException: Cannot open database "OperationsManager" requested
    by the login. The login failed.
    Login failed for user 'HK\svc-scomsdk'.
       at System.Data.ProviderBase.DbConnectionPool.GetConnection(DbConnection owningObject)
       at System.Data.ProviderBase.DbConnectionClosed.OpenConnection(DbConnection outerConnection, DbConnectionFactory connectionFactory)
       at System.Data.ProviderBase.DbConnectionFactory.GetConnection(DbConnection owningConnection)
       at System.Data.ProviderBase.DbConnectionClosed.OpenConnection(DbConnection outerConnection, DbConnectionFactory connectionFactory)
       at System.Data.SqlClient.SqlConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at System.Data.SqlClient.SqlConnection.Open()
       at System.Data.ProviderBase.DbConnectionClosed.OpenConnection(DbConnection outerConnection, DbConnectionFactory connectionFactory)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at System.Data.SqlClient.SqlConnection.Open()
       at System.Data.ProviderBase.DbConnectionClosed.OpenConnection(DbConnection outerConnection, DbConnectionFactory connectionFactory)
       at System.Data.SqlClient.SqlConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Run(Guid source, String cookie, Managers managers, IConfigurationDataAccessor dataAccessor, Stream stream, IConnection connection)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.SqlConnectionHandler.Open(SqlConnection sqlConnection, RetryPolicy retryPolicy, DatabaseConnection databaseConnection)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.EnterpriseManagement.Mom.DataAccess.DatabaseConnection.Open()
       at Microsoft.EnterpriseManagement.Mom.DataAccess.QueryRequest.Execute(SqlNotificationRequest sqlNotificationRequest)
       at Microsoft.Mom.ConfigService.OpsMgrDataAccess.ConfigurationDataAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.DataAccess.DatabaseAccessor.QueryInstanceProperties(ReadOnlyCollection`1 instances)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.Instances.CollectPublicProperties(ReadOnlyCollection`1 identities, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems.ConfigurationItemCollection`2.CollectPublicProperties(IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.ConfigurationItems..ctor(StateContext stateContext, IConfigurationDataAccessor dataAccessor)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Execute(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.CreateResponse(Managers managers)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.Managers.Synchronize(OnDoSynchronizedWork onDoSynchronizedWork)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Run(Guid source, String cookie, Managers managers, IConfigurationDataAccessor dataAccessor, Stream stream, IConnection connection)
       at Microsoft.Mom.ConfigService.Engine.ConfigurationEngine.CommunicationHelper.StateSyncRequestTask.Run(Guid source, String cookie, Managers managers, IConfigurationDataAccessor dataAccessor, Stream stream, IConnection connection)".

    Hi,
    I would like to suggest you reset the SDK account password (in this case, it seems like that the account is 'scomsdk'). And change the password for the account in service configuration.
    Please go through the below article for details:
    How to change the credentials for the SDK Service and for the Config Service in Microsoft System Center Operations Manager
    http://support.microsoft.com/kb/936220/en-us
    Regards,
    Yan Li
    Regards, Yan Li

  • Denormalization with overlapping time intervals

    The table keeps statuses for objects, along with times each status started and ended. Like:
    ID   From   To   Status
    ===  ====   ==   ======
    23   T1     T2    S1
    23   T3     T4    S2
    23   T5     T6    S3The statuses may start and end at arbitrary moments of time, and each object may have several statuses at the same time. So the time intervals in the above table may look like this:
    T1------------T2
        T3----------------------T4
             T5----------T6So, for example, from T! to T3 the object has only S1, while from T5 to T2 it has all 3: S1, S2, and S3.
    I need to denormalize the table and show in the same line all statuses the object is having at different time intervals, that is
    ID   From   To   Status1  Status2  Status3
    ===  ====   ==   ======   =======  =======
    23   T1     T3    S1
    23   T3     T5    S1       S2
    23   T5     T2    S1       S2       S3
    23   T2     T6    S2       S3
    23   T6     T4    S2

    SQL> with t as (select trunc(sysdate) dt from dual),
      2       t1 as (select 23 ID, dt From_dt, dt+10 To_dt, 'S1' Status from t union all
      3              select 23 ID, dt+3 From_dt, dt+18 To_dt, 'S2' Status from t union all
      4              select 23 ID, dt+7 From_dt, dt+14 To_dt, 'S3' Status from t)
      5              select * from t1
      6  /
            ID FROM_DT     TO_DT       STATUS
            23 04.08.2007  14.08.2007  S1
            23 07.08.2007  22.08.2007  S2
            23 11.08.2007  18.08.2007  S3
    SQL>
    SQL> with t as (select trunc(sysdate) dt from dual),
      2       t1 as (select 23 ID, dt From_dt, dt+10 To_dt, 'S1' Status from t union all
      3              select 23 ID, dt+3 From_dt, dt+18 To_dt, 'S2' Status from t union all
      4              select 23 ID, dt+7 From_dt, dt+14 To_dt, 'S3' Status from t),
      5       t2 as (select id,from_dt begdt from t1
      6               union all
      7              select id,to_dt from t1),
      8       t3 as  (select t2.*, lead(begdt) over(partition by id order by begdt) enddt from t2)
      9       select t3.*,
    10              (select status from t1 where t3.id=t1.id and greatest(t1.from_dt,t3.begdt)<least(t1.to_dt, t3.enddt) and t1.status='S1') Status1,
    11              (select status from t1 where t3.id=t1.id and greatest(t1.from_dt,t3.begdt)<least(t1.to_dt, t3.enddt) and t1.status='S2') Status2,
    12              (select status from t1 where t3.id=t1.id and greatest(t1.from_dt,t3.begdt)<least(t1.to_dt, t3.enddt) and t1.status='S3') Ststus3
    13               from t3
    14        where enddt is not null
    15        order by 2
    16  /
            ID BEGDT       ENDDT       STATUS1 STATUS2 STSTUS3
            23 04.08.2007  07.08.2007  S1             
            23 07.08.2007  11.08.2007  S1      S2     
            23 11.08.2007  14.08.2007  S1      S2      S3
            23 14.08.2007  18.08.2007          S2      S3
            23 18.08.2007  22.08.2007          S2     
    SQL>PS
    there's a problem in vizualization on the forum.
    all GREATEST() < LEAST()

  • Deferred transaction fails due to a "ORA-01403: no data found" error

    I recently acticated replication between two Oracle instances.
    (one being the master definition site and the other one being
    a second master site).
    Everything was going well until two weeks ago, when i realized
    that some datas were different between the two Oracle instances.
    I tried to figure out why and found that some deferred transation
    failed to propagate due to a "ORA-01403: no data found" error.
    Not all transaction fail, only a few.
    All the errors are stored in the DEFERROR table of the second
    master site, the DEFERROR table is empty in the master definition
    site.
    For every error, the source is the master definition site and the
    destination is the second master site, which is OK because the
    second master site is supposed to be "read-only". But why are the
    errors stored in the DEFERROR table of the second master site
    (and not in the master definition site (the source)).
    Most errors occur on UPDATE statements but a few of them occur
    on DELETE statements. Given the fact that the record actually is
    in the second master site, how is it possible to get a
    "ORA-01403: no data found" error on a delete statement ???????
    I can't figure out what data cannot be found,
    Thanks for your help,
    Didier.

    What if i tell you the first transaction in error is an UPDATE
    statement on a record that actually is in the destination table
    in the second master site. So the data is there, it should be
    found ...
    I've tried an EXECUTE_ERROR on that very first transaction
    in error but it doesn't work. Here's the call (as repadmin) :
    begin
    DBMS_DEFER_SYS.EXECUTE_ERROR (
    DEFERRED_TRAN_ID => '6.42.271',
    DESTINATION => 'AMELIPI.SENAT.FR'
    end ;
    And here's the full message :
    ORA-01403: no data found
    ORA-06512: at "SYS.DBMS_DEFER_SYS_PART1" line 430
    ORA-06512: at "SYS.DBMS_DEFER_SYS" line 1632
    ORA-06512: at "SYS.DBMS_DEFER_SYS" line 1678
    ORA-06512: at line 2
    I thought it could be because the primary key constraint on the
    destination table wasn't activated but it was. Could it
    be a problem with the corresponding index ?
    Thanks for your help,
    Didier.

  • Changing time dependent master data records

    Hello All,
    I have a time dependent master data infoobject that has 39 million entries.
    Now we decided that we no longer need it time dependent.
    Hence we want to maintain constant value ( 01.01.1000 to 31.12.9999) going forward.(as changing the object's time dependency will have huge impact)
    But for the historical data, we want to fix this and maintain only one record per value which will be the last record.
    Can you please provide inputs on how this can be achieved as i tried various combinations of load( load to a copy object taking only last record, changing its valid to to 01.01.1000 and loading back) and its not working.
    e.g the value XXX has following entries currently:
    Object
    Valid To
    Valid From
    XXX
    15.06.2007
    01.01.1000
    XXX
    19.09.2007
    16.06.2007
    XXX
    28.01.2008
    20.09.2007
    XXX
    30.03.2009
    29.01.2008
    XXX
    31.03.2009
    31.03.2009
    XXX
    15.04.2009
    01.04.2009
    XXX
    31.12.9999
    16.04.2009
    we need to get to the below format:
    Object
    Valid To
    Valid From
    XXX
    31.12.9999
    01.01.1000

    Hi Dhanya,
             The approach u did (load to a copy object taking only last record, changing its valid to to 01.01.1000 and loading back) will not work on Master data because it is not a DSO, with the Proper keys in place, to overwrite all the previously existing data records. This approach will only work if the infoobject is not time dependent and in that case the question u posed will not arise at all .
    SOLUTION:
    One thing is Inevitable as Prashanth has prescribed, u have to delete the transaction data first at all the places, where and all it was used.
    Then just delete the entire master data from the info object and just do a FULL DTP load from the underlying datasource putting a FILTER like "VALID_TO = 31.12.9999" so that only the last record of a particular Value will be loaded. (No need to do a full load from the source because even in the source it might had been the same old Values and might not have been revamped.)
             Even for further delta's, it will be safer to maintain this filter at the DTP level if u r not sure of the  constant value maintenance at the source.
    Thanks & Regards
    Sasidhar

  • Discrete time range over date range  (using CR X1)

    Post Author: sjr
    CA Forum: Formula
    Can anyone show me how to select data over a discrete time range i.e. 6 p.m. till 11 p.m. per day, over a period of a month ?
    Using the data time parameter gives me data from 6p.m. on startdate right through till 11p.m. on endate.

    Here's a more detailed description...
    Create two formula fields to split the date and time from the data records (ds prefixed):
    dsDate
    dsTime
    Create two formula fields to split the date and time from the parameter fields (pm prefixed):
    pmStartDate
    pmStartTime
    pmStartDate
    pmStartTime
    Use your new formula fields in the record selector. Get to this off the Report menu button (in CR 2008):
    Report | Select Expert... | Record.
    You can pick the formula fields in the Forumula Editor to create something like:
    ({@dsDate} > {@pmStartDate} and
    {@dsDate} < {@pmEndDate}) and
    ({@dsTime} > {@pmStartTime} and
    {@dsTime} < {@pmEndTime)
    Tim

  • DTP failed due to exception error

    Hi there,
    DTP failed due to an exception error message An exception with the type CX_SY_OPEN_SQL_DB occurred, but was neither handled locally, nor declared in a RAISING clause
    Please Help
    Thanks
    J

    Hi.....
    I think the full Error is :
    Runtime Errors DBIF_RSQL_SQL_ERROR
    Exception CX_SY_OPEN_SQL_DB
    Actually this is a Database related error............the only thing that u can do is increase the number of process assigned for your system .............and just make sure that no other loads are going on at that time..........if going on also.......it should be very limited.........
    Also u can check the SAP Note : 631668..............I don't know u have access to SAP Notes or not.........So I am copying it here..........
    Symptom
    This note applies only to BW systems under ORACLE.
    Loading data into InfoCubes terminates with an error: ORA 60, "deadlock detected". The monitor log displays the entry "CALLER 70 missing" and the request is flagged as red.
    Other terms
    Caller 70, DEADLOCK, ora0060, bitmap, indexes, DBIF_RSQL_SQL_ERROR
    CX_SY_OPEN_SQL_DB
    Reason and Prerequisites
    In SAP BW, there are two fact tables for including transaction data for Basis InfoCubes: The F and the E fact table.
    Unlike the E fact table, the F fact table contains the information about the request from where the transaction data originates.
    Therefore, transaction data is always written to the F fact table and the package dimension contains the request to which the loaded data belongs.
    Using 'Request compression', the system reads the data of a request from the F fact table and writes it to the E fact table without request ID and compresses.
    For good reporting performance, you must keep only the data of a few requests in the F fact table since the F fact table is partitioned according to the key of the package dimension and the data of all requests is always read. Therefore, you should compress the request if you are sure that the loaded data is correct and that you no longer have to delete the request.
    When you install BW on an ORACLE database, BITMAP indexes are created on the fact tables to improve the reporting performance of the system.
    These have a negative effect on the performance when loading data and the deadlock mentioned above (ORA 60) may occur as a result. The DEADLOCKs occur during parallel insert operations because ORACLE does not support a blocking concept for BITMAP indexes at data record level.
    Solution
    Before loading movement data, delete secondary indexes from the F fact table and create them again after the loading process. If the F fact table is small, no performance problems occur.
    Alternatively, you can also load requests serially into InfoCubes by first loading the data only into the Persistent Staging Area (PSA) and then serially into the InfoCube. If you use the data transfer process (DTP) for loading in BW 7.0, this option is no longer available.
               NOTE
               This avoids the deadlock problem. However, you should note that after some loading processes, it is necessary to reorganize the BITMAP indexes because they degenerate very heavily and, therefore, the read and writing performance deteriorates dramatically.
    Hope this helps........
    Regards,
    Debjani.......

  • My ipad3 is at iOs 7, i'm trying to download update 7.0.2 I keep getting failed due to error like 9 times now. My mini, my two iphone 4s, no problems. What can I do about my ipad3?

    ipad3 running iOs 7 trying to update to 7.0.2 keep tting failed due to error like 9 times. Updated iPad mini and 2 iPhone 4s, no problem.
    Whats up? And what can I do to update?

    Try and update with iTune (computer)

  • HT3275 My Time Capsule keeps saying "backup failed" due to the network password not being correct. I have not changed my password and every day I have to type the same one in and then it works. But then the same thing happens the next day. Can anyone help

    My Time Capsule keeps saying "backup failed" due to the network password not being correct. I have not changed my password and every day I have to type the same one in and then it works. But then the same thing happens the next day. Can anyone help? This only started doing this about 2 weeks ago.

    Hello,
    Mac OS X 10.4 Help, I forgot a password in my Keychain
    http://docs.info.apple.com/article.html?path=Mac/10.4/en/mh1960.html
    Resetting your keychain in Mac OS X...
    If Keychain First Aid finds an issue that it cannot repair, or if you do not know your keychain password, you may need to reset your keychain.
    http://support.apple.com/kb/TS1544
    Open Keychain Access in Utilities, use Keychain First Aid under the Menu item, then either check the Password under that item, change it, or delete it and start over.

  • JDBC initialize failed due to long idling time

    Hi gurus
    Anyone having similar problem and would like to share their solution for the problem stated below.
    Normally the integration for IDOC-XI-JDBC scenarios is working fine, but when the R3 halt sending IDOC to XI, the JDBC connection to MS SQL 2000 server failed to process and initialise due to long idling time. Currently the temporary solution is to go to XI change the Integration Directory : Communication Channel for the JDBC - Activate - Clear Cache. Then go to R3, retrigger the IDOC again. And the above problem will be resolved. But after a while (long idling time again), this problem arise once again.
    2008-10-21 15:18:45     Success     Message successfully received by messaging system. Profile: XI URL: http://locahost:50600/MessagingSystem/receive/AFW/XI Credential (User): PIISUSER
    2008-10-21 15:18:45     Success     Using connection JDBC_http://sap.com/xi/XI/System. Trying to put the message into the receive queue.
    2008-10-21 15:18:45     Success     Message successfully put into the queue.
    2008-10-21 15:18:45     Success     The message was successfully retrieved from the receive queue.
    2008-10-21 15:18:45     Success     The message status set to DLNG.
    2008-10-21 15:18:45     Success     Delivering to channel: CC_IDOC_TO_DBMS_ESMS_588
    2008-10-21 15:18:45     Success     MP: Entering module processor
    2008-10-21 15:18:45     Success     MP: Processing local module localejbs/CallSapAdapter
    2008-10-21 15:18:45     Success     Receiver JDBC adapter: processing started; QoS required: ExactlyOnce
    2008-10-21 15:18:45     Error     JDBC adapter receiver channel CC_IDOC_TO_DBMS_ESMS_588 not initialized. Unable to proceed due to null
    2008-10-21 15:18:46     Error     MP: Exception caught with cause com.sap.aii.af.ra.ms.api.RecoverableException: JDBC Adapter configuration not initialized: null
    2008-10-21 15:18:46     Error     Exception caught by adapter framework: JDBC Adapter configuration not initialized: null
    2008-10-21 15:18:46     Error     Delivery of the message to the application using connection JDBC_http://sap.com/xi/XI/System failed, due to: com.sap.aii.af.ra.ms.api.RecoverableException: JDBC Adapter configuration not initialized: null.
    2008-10-21 15:18:46     Success     The message status set to WAIT.
    2008-10-21 15:18:46     Success     The asynchronous message was successfully scheduled to be delivered at Tue Oct 21 15:23:46 GMT+08:00 2008.
    2008-10-21 15:23:46     Success     The message status set to TBDL.
    2008-10-21 15:23:46     Success     Retrying to deliver message to the application. Retry: 1
    2008-10-21 15:23:46     Success     The message was successfully retrieved from the receive queue.
    2008-10-21 15:23:46     Success     The message status set to DLNG.
    2008-10-21 15:23:46     Success     Delivering to channel: CC_IDOC_TO_DBMS_ESMS_588
    2008-10-21 15:23:46     Success     MP: Entering module processor
    2008-10-21 15:23:46     Success     MP: Processing local module localejbs/CallSapAdapter
    2008-10-21 15:23:46     Success     Receiver JDBC adapter: processing started; QoS required: ExactlyOnce
    2008-10-21 15:23:46     Error     JDBC adapter receiver channel CC_IDOC_TO_DBMS_ESMS_588 not initialized. Unable to proceed due to null
    2008-10-21 15:23:46     Error     MP: Exception caught with cause com.sap.aii.af.ra.ms.api.RecoverableException: JDBC Adapter configuration not initialized: null
    2008-10-21 15:23:46     Error     Exception caught by adapter framework: JDBC Adapter configuration not initialized: null
    2008-10-21 15:23:46     Error     Delivery of the message to the application using connection JDBC_http://sap.com/xi/XI/System failed, due to: com.sap.aii.af.ra.ms.api.RecoverableException: JDBC Adapter configuration not initialized: null.
    2008-10-21 15:23:46     Success     The message status set to WAIT.
    2008-10-21 15:23:46     Success     The asynchronous message was successfully scheduled to be delivered at Tue Oct 21 15:28:46 GMT+08:00 2008.
    2008-10-21 15:28:46     Success     The message status set to TBDL.
    Edited by: Malcolm Pang on Oct 23, 2008 11:08 AM

    In the JDBC adapter , select option Disconnect From Database after Each message processing.
    This will solve the process as your JDBC Database seems to be closing the connection after a particular idle time.
    Regards
    Bhavesh

Maybe you are looking for

  • WEBGUI  is not working in portal

    Hi I had conected to production server and created Transactional Iview with SAPGUI for WINDOWS and running it successfully. My aim is to make use of SAPGUI for HTML in the place of SAPGUI for WINDOWS. So i had activated ITS services successfully and

  • Hi, if IPhone 5 was manufactured in 2013 (after release of IOS 6.1.4) what would release date of be iphone written in IMEI info? 2012 or 2013?

    Hi, if IPhone 5 was manufactured in 2013 (after release of IOS 6.1.4) what would release date of be iphone written in IMEI info? 2012 or 2013? Please answer. Thanks.

  • Cannot print pdf files

    operating system vista, printer: all-in-one photosmart 2575 i have tried (i think) everything, re-loaded my drivers, went through all possible settings, such as Hit PRINT, then hit ADVANCED, and then check the box that says PRINT AS IMAGE. nothing wo

  • Released Version: NI 5640R Installer; Please update

    Released Version: NI 5640R Installer; Please update! The software for the NI PCI-54640R IF-RIO is now released.  Before asking about any issues in programming the NI 5640R, please upgrade to the released version.  This will also make it easier to loo

  • Configuring ssh fails

    Hi, My ssh is failing from Node B to A, whereas it is working fine from A to B I followed the note 300548.1 From node A $ ssh nodeB date , this doesn't prompt for p/w From node B $ ssh nodeA date , this prompt's for password every time. Can somebody