Time dependent data load to BW using Fixed Time Interval selection

Hi,
I'm facing a problem in extracting time-dependent master data to BW. I would really appreciate if anyone can help in this regard.
We have a table in R/3 which has time-dependent data. For example, for a single OBJID we could have multiple records with different BEGDA/ENDDA and different attributes. When we try to load this data into BW InfoObject, we see a time-dependent data selection in update tab of infopackage. We need to select BEGDA/ENDDA in datasource since otherwise during loads we would get a "DUPLICATE RECORDS" load error in BW.
The problem is when we enter a Fixed-time interval Start Data and End Date in the InfoPackage, only the records with that StartDate and EndDate are loaded. Ideally I would want to see all the records whose BEGDA/ENDDA falls in the range of StartDate and EndDate.
I tried modifying the ROOSFIELD table for the data source. I made the BEGDA SELOPTS as GE(16) and ENNDA SELOPTS as LE(64). But still it doesn't seem to work. I checked if the change to ROOSFIELD has any affect in RSA3. Unless we give a range for BEGDA it doesn't seem to behave as expected. Could someone let me know how I can load all records to BW whose BEGDA/ENDDA fall in the range of StarteDate and EndDate.
Thanks,
Anil.

Anil,
If you are using a custom dataSource, make sure the dataSource fields for begda and ennda are labeled "datefrom" and "dateto".
Cheers,
T-

Similar Messages

  • Fixed time interval selection in Infopackage

    Hi Guys,
    I have a generic data source from PA0315 table that I am using to supply master data 0EMPLOYEE. In the update tab of the infopackage it has a fixed time interval selection with Start Date and End Date. I enter 01.01.1900 to 31.12.9999 so that I can capture the entire data for every load.
    But when ever I enter that selection option it always returns with '0' data. But if I enter specific date ranges it returns with data. But being specific does not help me b/c I want the entire date range.
    Has anybody come across this before and have an idea on a solution to capture all the data.
    Any response will be appreciated.
    K.

    Hi,
    I'm facing a similar problem. I would really appreciate if anyone can help in this regard.
    We have a table in R/3 which has time-dependent data. For example, for a single OBJID we could have multiple records with different BEGDA/ENDDA. When we try to load this data into BW InfoObject, we see a time-dependent data selection in update tab of infopackage. We need to select BEGDA/ENDDA since otherwise during loads we would get a "DUPLICATE RECORDS" load error in BW.
    The problem is when we enter a Fixed-time interval Start Data and End Date in the InfoPackage, only the records with that StartDate and EndDate are loaded. Ideally I would want to see all the records whose BEGDA/ENDDA falls in the range of StartDate and EndDate.
    I tried modifying the ROOSFIELD table for the data source. I made the BEGDA SELOPTS as GE(16) and ENNDA SELOPTS as LE(64). But still it doesn't seem to work. I tried the same in RSA3. Unless we give a range for BEGDA it doesn't seem to behave as expected. Could someone let me know how I can load all records to BW whose BEGDA/ENDDA fall in the range of StarteDate and EndDate.
    Thanks,
    Anil

  • After loading the internet using Firefox AND then selecting NEWS and a delay the Error "The MSN toolbar has encountered an error. etc" occurs. Also, when browsing some searches are "REDIRECTED"!

    After loading the internet using Firefox AND then selecting NEWS and a delay the Error "The MSN toolbar has encountered an error. etc" occurs. Also, when browsing some searches are "REDIRECTED"!
    == This happened ==
    Every time Firefox opened
    == Hard to say but have been trying to diagnose for at least a week.

    Original message wouldnt post with attachments - trying again....
    Attachments:
    secondclickheredialog.jpg ‏40 KB
    FirstDialog.jpg ‏19 KB
    Firstclickheredialog.jpg ‏16 KB

  • Time interval selection in IP30

    Hi, gurus!
    We need to use IP30 transaction but indicating a time interval selection.
    That is, the standard transaction allows to indicate an interval for call objects (for instance, 30 days), but we need to indicate a start and end of interval (for example, from 01.01.2011 to 15.01.2011).
    Is there any way of doing this (field exit, ...)?
    Thanks in advance.
    Best regards,
    J Madariaga

    Hi,
    It can be achieved by backgrounfd scheduling
    sm36
    check this for background scheduling details
    Re: Deadline Monitoring
    Regards,
    Jitender

  • Reg data loading into essbase using text files

    Can we load data in parallel from 2 files into the same cube using 2 different rules files? Or do we have to load one file at a time?
    Could someone clarify this?

    I do not believe that by selecting two data files and two load rules in AAS you are getting parrallel data loading. If you look at the log, you will find them to be sequential. For ASO cubes, AAS loads the data into a buffer then applies it. The only real parrallel data loading is using multiple threads for one file. Othere than that is it sequential

  • Problem in data loading in 0PM_DS05 using data source 0PM_MEASUREMENT

    Hi Gurus,
    I have to load data in OPM_DS05 ODS using 0PM_MEASUREMENT data source.
    Now the problem what I am facing is, there is a field equipment number which was there in data source but not mapped to infoobject then I mapped that field
    EQUNR(R3 field) to 0EQUIPMENT(BI infobject) after this  when I try to load data error is comming and when I check the dump I understand that there is some routine written in update rules which refers master data of 0EQUIPMENT thus what I got is I have to load master data first then the ODS but when I load master data then request is not comming in the ATTRIBUTES 0EQUIPMENT.
    I am using 0EQUIPMENT_ATTR data source to load data to 0EQUIPMENT.
    Please help me out from the situation what should I do and also suggest whether it is necessary to load master data for 0EQUIPMENT.
    Your efforts will be appreciated.

    Hi,
    in generally it is always a good idea to load the master data first. In your case there are still some questions.
    a) did you activate the master data after loading?
    b) what kind of short dump are you getting while loading the transactional data?
    c) what is the routine doing?
    I guess your loading issues are more related to the routine than anything else.
    regards
    Siggi

  • Preferred way of Data load in Essbase Using EAS or EIS

    Hi
    I am into designing new Essbase Cubes (ASO). Currently the process being followed is that we make a Metadata Model with Dummy Fact table and then create a MetaOutline and then finally outline from EIS and then Load data using MaxL and rules files through EAS,
    But I am wondering why don't I do all this from EIS.
    Can any of you provide me with some pros and Cons of loading data using EIS??
    Thanks in advance!!
    DJ

    One advantage of using MAxL with load rules is you have more control over the process and SQL that is created.I will agree with what Glenn said that we will have more control over the process if we use EAS as we are provided with the MaxL which is a fully equipped automation language with effective error capturing mechanism when tightly integrated with shell or vbscript.
    At the same time EIS is also effective for the ASO applications which are full refersh.
    If we design and develop the custom relational star schema by defining the joins and relations to feed the EIS OLAP Model, we can eliminate the unnecesary errors and can provide efficiency also.
    Main advantage of EIS is HOLAP solution.

  • Sql loader - Data loading issue with no fixed record length

    Hi All,
    I am trying to load the following data through sql loader. However the records # 1, 3 & 4 are only loading succesfully into the table and rest of the records showing as BAD. What is missing in my syntax?
    .ctl file:
    LOAD DATA
    INFILE 'C:\data.txt'
    BADFILE 'c:\data.BAD'
    DISCARDFILE 'c:\data.DSC' DISCARDMAX 50000
    INTO TABLE icap_gcims
    TRAILING NULLCOLS
         CUST_NBR_MAIN          POSITION(1:9) CHAR NULLIF (CUST_NBR_MAIN=BLANKS),
         CONTACT_TYPE          POSITION(10:11) CHAR NULLIF (CONTACT_TYPE=BLANKS),
         INQUIRY_TYPE          POSITION(12:13) CHAR NULLIF (INQUIRY_TYPE=BLANKS),
         INQUIRY_MODEL          POSITION(14:20) CHAR NULLIF (INQUIRY_MODEL=BLANKS),
         INQUIRY_COMMENTS     POSITION(21:60) CHAR NULLIF (INQUIRY_COMMENTS=BLANKS),
         OTHER_COLOUR POSITION(61:75) CHAR NULLIF (OTHER_COLOUR=BLANKS),
         OTHER_MAKE          POSITION(76:89) CHAR NULLIF (OTHER_MAKE=BLANKS),
         OTHER_MODEL_DESCRIPTION POSITION(90:109) CHAR NULLIF (OTHER_MODEL_DESCRIPTION=BLANKS),
         OTHER_MODEL_YEAR POSITION(110:111) CHAR NULLIF (OTHER_MODEL_YEAR=BLANKS)
    data.txt file:
    000000831KHAN
    000000900UHFA WANTS NEW WARRANTY ID 000001017OHAL
    000001110KHAP
    000001812NHDE231291COST OF SERVICE INSPECTIONS TOO HIGH MAXIMA 92 MK
    000002015TPFA910115CUST UPSET WITH AIRPORT DLR. $200 FOR PLUGS,OIL,FILTER CHANGE. FW
    Thanks,

    Hi,
    Better if you have given the table structure, I check your script it was fine
    11:39:01 pavan_Real>create table test1(
    11:39:02   2  CUST_NBR_MAIN  varchar2(50),
    11:39:02   3  CONTACT_TYPE varchar2(50),
    11:39:02   4  INQUIRY_TYPE varchar2(50),
    11:39:02   5  INQUIRY_MODEL varchar2(50),
    11:39:02   6  INQUIRY_COMMENTS varchar2(50),
    11:39:02   7  OTHER_COLOUR varchar2(50),
    11:39:02   8  OTHER_MAKE varchar2(50),
    11:39:02   9  OTHER_MODEL_DESCRIPTION varchar2(50),
    11:39:02  10  OTHER_MODEL_YEAR varchar2(50)
    11:39:02  11  );
    Table created.
    11:39:13 pavan_Real>select  * from test1;
    no rows selected
    C:\Documents and Settings\ivy3905>sqlldr ara/ara@pavan_real
    control = C:\control.ctl
    SQL*Loader: Release 9.2.0.1.0 - Production on Sat Sep 12 11:41:27 2009
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Commit point reached - logical record count 5
    11:42:20 pavan_Real>select count(*) from test1;
      COUNT(*)                                                                     
             5    control.ctl
    LOAD DATA
    INFILE 'C:\data.txt'
    BADFILE 'c:\data.BAD'
    DISCARDFILE 'c:\data.DSC' DISCARDMAX 50000
    INTO TABLE test1
    TRAILING NULLCOLS
    CUST_NBR_MAIN POSITION(1:9) CHAR NULLIF (CUST_NBR_MAIN=BLANKS),
    CONTACT_TYPE POSITION(10:11) CHAR NULLIF (CONTACT_TYPE=BLANKS),
    INQUIRY_TYPE POSITION(12:13) CHAR NULLIF (INQUIRY_TYPE=BLANKS),
    INQUIRY_MODEL POSITION(14:20) CHAR NULLIF (INQUIRY_MODEL=BLANKS),
    INQUIRY_COMMENTS POSITION(21:60) CHAR NULLIF (INQUIRY_COMMENTS=BLANKS),
    OTHER_COLOUR POSITION(61:75) CHAR NULLIF (OTHER_COLOUR=BLANKS),
    OTHER_MAKE POSITION(76:89) CHAR NULLIF (OTHER_MAKE=BLANKS),
    OTHER_MODEL_DESCRIPTION POSITION(90:109) CHAR NULLIF (OTHER_MODEL_DESCRIPTION=BLANKS),
    OTHER_MODEL_YEAR POSITION(110:111) CHAR NULLIF (OTHER_MODEL_YEAR=BLANKS)
    data.txt
    000000831KHAN
    000000900UHFA WANTS NEW WARRANTY ID 000001017OHAL
    000001110KHAP
    000001812NHDE231291COST OF SERVICE INSPECTIONS TOO HIGH MAXIMA 92 MK
    000002015TPFA910115CUST UPSET WITH AIRPORT DLR. $200 FOR PLUGS,OIL,FILTER CHANGE. FW
    CUST_NBR_MAIN     CONTACT_TYPE     INQUIRY_TYPE     INQUIRY_MODEL     INQUIRY_COMMENTS     OTHER_COLOUR     OTHER_MAKE     OTHER_MODEL_DESCRIPTION     OTHER_MODEL_YEAR
    000000831     KH     AN     NULL     NULL     NULL     NULL     NULL     NULL
    000000900     UH     FA      WANTS     NEW WARRANTY ID 000001017OHAL     NULL     NULL     NULL     NULL
    000001110     KH     AP     NULL     NULL     NULL     NULL     NULL     NULL
    000001812     NH     DE     231291C     OST OF SERVICE INSPECTIONS TOO HIGH MAXI     MA 92 MK     NULL     NULL     NULL
    000002015     TP     FA     910115C     UST UPSET WITH AIRPORT DLR. $200 FOR PLU     GS,OIL,FILTER C     HANGE. FW     NULL     NULL- Pavan Kumar N
    Edited by: Pavan Kumar on Sep 12, 2009 11:46 AM

  • Can we use 0INFOPROV as a selection in Load from Data Stream

    Hi,
    We have implemented BW-SEM BPS and BCS (SEM-BW - 602 and BI 7 ) in our company.
    We have two BPS cubes for Cost Center and Revenue Planning and we have Actuals Data staging cube, we use 0SEM_BCS_10 to load actuals.
    We created a MultiProvider on BPS cubes and Staging cube as a Source Data Basis for BCS.
    Issue:
    When loading plan data or Actuals data into BCS (0BCS_C11) cube using Load from Data Stream method, we have performance issue, We automated load process in a Process Chain. Some times it take about 20 hrs for only Plan data load for 3 group currencies and then elimination tasks.
    What I noticed is, (for example/) when loading Plan data, system is also reading Actuals Cube which is not required, there is no selection available in Mapping or selection tab where I can restrict data load from a particular cube.
    I tried to add 0INFOPROV into databasis but then it doen't show up as selection option in the data collection tasks.
    Is there a way where I can restrict data load into BCS using this load option and able to restrict on cube I will be reading data from ?
    I know that there is a filter Badi available, but not sure how it works.
    Thanks !!
    Naveen Rao Kattela

    Thanks Eugene,
    We do have other characteristics like Value Type (10 = Actual and 20 = Plan) and Version (100 = USD Actual and 200 = USD Plan), but when I am loading data into BCS using Load from Data Stream method, the request goes to all the underlying cubes, which in my case are Planning cubes and Actual Cube, but I don't want request to goto Actual Cube when I am running only Plan load. I think its causing some performance issue.
    For this reason I am thinking if I can use 0INFOPROV as we use in Bex queries to filter the InfoProvider so that the data load performance will improve. 
    I was able to to bring in 0INFOPROV into DataBasis by adding 0INFOPROV  in the characteristics folder used by the Data Basis.
    I am able to see this InfoObject Data Stream Fileds tab. I checked marked it to use it in the selection and regenerated the databasis.
    I was expecting that now this field would be available for selection in data collection method, but its not.
    So If its confirmed that there is no way we can use 0INFOPROV as a selection then I would suggest my client for a redesign of the DataBasis it self.
    Thanks,
    Naveen Rao Kattela

  • Error in 0EMPLOYEE Master Data Load Issue

    Hi,
    We have 0EMPLOYEE Master Data. Due to our new development changes ralted to 0EMPLOYEE, we have scheduled 2 new info packages with Personnel number range. While creation of infopackages, we forget to main time interval from 01.01.1900 to 31.12.9999. Instead of this, the default range was selected as 24.04.2009 to 31.12.9999. Because of this selection in InfoPackage, the Employee Master Data Valid from date was changed to 24.04.2009 for all the employees in the master data after the data load.
    Even after i change this selection properly and loading the data also, its not correcting with correct valid from dates.
    Can you please advice, how can we fix this issue ASAP as its a production issue?
    Thanks!
    Best regards,
    Venkata

    > Even after i change this selection properly and loading the data also, its not correcting with correct valid from dates.
    May be for this you have the ONLY option to delete 0Employee master data and reload it again. For this you need to delete dependent transaction data also.
    Cheers,
    Sree

  • Data Loader On Demand Proxy Usage for Resume operation

    Hi,
    My project required me to use the proxy feature available in Data Loader R19 release.
    I could use the proxy at command line for insert /update operations.
    However, the same doesnt work for RESUME operation in Data loader.
    Tried using proxy settings from command line as well as property file but to no use.
    Any suggestions...
    Regards,
    Sumeet

    Its a Java application so it may run on your Linux/Unix system, you would have to test to see if works. Last time I checked Oracle only supports the application running on windows.

  • EIS Data load

    Hi all,
    I am bit confused about loading data via EIS. We have two process one is to load data via rule file hisdld created by EIS and second is using user defined sql query. Are both these methods independent of each other or not?
    When i say independent i mean if we make some changes in user defined sql and load data are those changes first refelected in the rule file and then data load process occurs using this rule file or data load occurs just via sql quey we run
    It would be great if somebody could clarify the same.
    Thanks in advance!!

    Thanks for the reply.
    Yes i am talking about sql query within EIS so if i understood it right by your post if i make changes to that user defined sql code(within EIS ) it will make the changes in the data load rule file created by EIS i.e "hisdld". So If we talk of a data load process within EIS the User defined sql is used to create data load rule "hisdld" and hence they are dependant on each other.
    Edited by: user10859098 on Jan 29, 2009 10:54 AM
    Edited by: user10859098 on Jan 29, 2009 10:54 AM
    Edited by: user10859098 on Jan 29, 2009 10:56 AM

  • BW Data Load failure

    Hi,
       I created a BW Data Source in R/3 and checked in RSA3 (R/3) it is extracting the required record...
    But when i started data load in BW using that extractor i am geting error message
    <b>Request IDoc : Application document not posted</b>
    How can i proceed.
    Thanks

    Hi,
    Refer the link:
    https://forums.sdn.sap.com/click.jspa?searchID=937834&messageID=2379005
    With rgds,
    Anil Kumar Sharma .P

  • Data Load from DS-- DSO---- Cube

    Hello Guys,
    I have an issue with the data load.
    I use the OFI_GL_14 extractor.
    When I did he innit with data transfer I have around 450 K records.
    When I loaded to DSO from DS by DTP  it loaded the same 450 K .
    When I tried to load the data from DSO to Cube it selects only 14.5K records.
    It says transfered records 14.5 K and added records 14.5 K records.
    Even if it aggregatesat the cube level- it should say transfered 450 K records and 14.5 K records added. is it correct.
    I don't have any filters in the DTP.I checked it. I deleted the DTP and again created another new one. The same result.
    From DS to DSO is delta-DTP and again from DSO to Cube is delta - DTP.
    I am not sure why it filters and select only 14.5 records from 450 K records.
    Do you guys encountered this type of situation. Let me know if I am doing something wrong.
    Thanks again for your input. It is really appreciated .
    Senthil

    Are you using infosource in between? If you do, the infosource is aggregating the the data before it gets to the transformation, in that case you will see the transferred record will be less. If you check all the fields as a key field in the infosource, you will get the same record transferred to the transformation and data will be aggregated going to the cube.
    thanks.
    Wond

  • EIS data load via rules file

    Hi,
    PLease let me know the process to control the EIS data load into essbase using a rules file. I did not find the option in EIS. please help me.
    Thanks,
    pr

    In EIS
    1) You have to define the Logical OLAP Model connecting to the relational source.It defines the joins between fact table and dimension tables.
    2) Based on the OLAP Model You have to create meta outline which defines the rules for loading members and data into essbase.

Maybe you are looking for

  • MULTIPLE DISPLAYS with Tecra A11

    We are upgrading our classrooms with laptops & smart boards.  We have purchased the Tecra A11 laptop with the PA3680U port replicator and the D19W12B monitor. To utilize the smart board (ENO) we have a VGA splitter. When connected conventionally (we

  • Java.lang.UnsupportedOperationException in DefaultContext.loadEnvironment()

    Getting below exception inconsistently, when i run the page. Could someone let me know, why this comes? and how to resolve it? 08/06/22 14:23:29 [1551] Failed to create BindingContainer from reference :jsps_serviceprovidernegotiationPageDef 08/06/22

  • [SOLVED] System hangs at random, suspect X is the culprit

    For a bit over a week now, my system has been hanging seemingly at random. When watching youtube videos, when web browsing, when watching stuff in mplayer, probably more times too. The only way to get back to normal is hitting the physical reset butt

  • Marker command for N8975A Noise Figure Analyzer

    hi, I am new to labVIEW and have a simple question to ask. May I know if anyone have try to place a marker in N8975A Noise Figure Analyzer using labVIEW? If so, may I know what is the command for that? Thanks.

  • E-mail imported camera video clip

    I downloaded into iPhoto a .59 second video clip from my cannon camera. How do I attach it to an e-mail that can be opened by a windows user? when I attach it is says it is not a complete file?