Sizing for extraction of 2LIS_02_ITM / 2LIS_03_BF (data volume)

Hello Experts,
can anybody help me with this issue:
I am extracting data for the datasources 2lis_02_itm and 2lis_03_bf. The bw scenario is a pilot project and this is the first transactional data ever loaded into the bw system. At the moment i am facing the problem, that our database seems to be too small (50 GB). I need to know exactly how much data will be extracted from these two datasources.
The main problem is, that i can not figure out the exact amount of datasets for both extractors. My bw system dumped after i loaded around 60 million datasets only for the datasource 2lis_02_itm (2years). That means that there is almost 50 GB purchase order data in the system. Of course figured out hoch much data is in ekpo and ekko, but there is much more data extracted. I found out, that the tables cdpos and cdhdr are also relevant, because they keep all changes on purchase orders. I think the extractor delievers all these changes. How can i find out how big the tables cdpos and cdhdr are (i think they are strucutres)?
For the extractor 2lis_03_bf i did not find tables which keep all changes.
To sum up: How can i find out how much data will be extracted to bw for the datasources 2lis_02_itm and 2lis_03_bf to bw. I need to know that in GB size.
Can you provide a way to figure that out?
Best regards
Janos Bugler

Hi,
We can't measure the size manually, we have tool QUICK SIZER in service.sap.com there you give your inputs and then it will suggest the Hardware sizing.
Thanks
REeddy

Similar Messages

  • Usage of 2LIS_03_BF DataSource for extraction of MM outside SAP enviroment

    Hi All,
    We currently use Business Content 2LIS_03_BF Data Source to extract Material Movements data into BW Data Warehouse.
    Business plans to extend Data Warehouse landscape and build non-SAP BI environment. One of the assumptions is that we can extract data from R/3 to the non-SAP environment in the same way as we did in case of BW (business expects the same structure and granularity of data). To fulfill that we are trying to “deceive” SAP and extract data outside in the same way as it is done for BW.
    We built ABAP program that calls Function Module 'MCEX_BW_LO_API' with corresponding parameters (this module is used when extracting data with 2LIS_03_BF). Our ABAP calls Function Module and inserts data exported by Function Module into custom-built table. Unfortunately we stuck since it looks that no matter what kind of UPDATE MODE we use – system can only extract data from Setup Tables (Update Mode = ‘C’ or ‘I’), or extract nothing (Update Mode = ‘D’).
    What we cannot achieve is:
    1)     force program to create corresponding delta queue (in RSA7)
    2)     extract delta (which is obvious consequence of the point 1)
    Here are main parameters that we set when calling Function Module:
    I_ISOURCE           = P_SOURCE "'2LIS_03_BF'
    I_UPDMODE         = ‘C’ or ‘I’ or ‘D’
    We suppose that BW InfoPackage sends some more ‘hidden’ parameters to R/3 when it is run for 2LIS_03_BF initialization (parameters that are not explicitly put in Function Module definition). These ‘hidden’ parameters probably allow creating data queue and subsequently run the whole delta process.
    Does anyone tried to make similar approach when extracting data outside SAP environment?
    Do you maybe know what parameters/elements we are missing?
    Maybe someone tried similar way but found out that it is not possible?
    I will be grateful for any advice,
    Andrzej Bobula

    Hi Andrez,
    Welcome to SDN. This is the standard way LO Cockpit works. The data source will be able to extract the load only from setup tables and then the deltas are based on the init based on the setup tables. I have not tried this. I am sure that you can extract data from R/3 into other datawarehouse systems (other than BW).
    Instead of this standard LO way of extraction, what I would do is to write an interface program (RFC FM or BAPI) from R/3 which would read the data directly from the database tables and put them in a queue or put it in a common area as a file. The other DW system would pick up this and post them. If you have a middleware between your R/3 and the third part (dw) system, this becomes even simpler as the data transfer will be handled by the middleware. The deltas can be handled programatically based on the date and the time stamp. Think about this solution. Hope it helps.
    Thanks and Regards
    Subray Hegde

  • How to create a RFC destination for extracting data to HANA

    Hello All,
    Could someone help me in providing a document or note on how to create a RFC destination for extracting data from SAP data source to HANA using SAP LT replication server ?
    I am able to create a data base connection while transforming data from non SAP data source,but wasnt able to transform data which is from SAP abap tables .

    Hi Venkatesh,
    In SM59 t.code we create RFC destinations.
    Go thru the video link for creating RFC destinations step by step
    How to setup a trusted RFC connection between SAP systems: a step-by-step guide - YouTube

  • BO DI for extracting data from legacy system to SAP BW 7.0

    Hi,
    Has anyone worked on scenario where BO Data Integrator is being used as ETL tool for extracting data from legacy systems to SAP BW 7.0 ?
    Please share your experience on such a landscpae or related documents any of you had come across.
    Regards,
    Pritesh

    There are couple of them - Lawson ERP, Kalido, AS/400.  ECC 5.0 might be there as well.
    Regards,
    Pritesh.
    Edited by: pritesh prakash on Jun 14, 2010 2:00 PM

  • Query taking long time for EXTRACTING the data more than 24 hours

    Hi ,
    Query taking long time for EXTRACTING the data more than 24 hours please find the query and explain plan details below even indexes avilable on table's goe's to FULL TABLE SCAN. please suggest me.......
    SQL> explain plan for select a.account_id,round(a.account_balance,2) account_balance,
    2 nvl(ah.invoice_id,ah.adjustment_id) transaction_id,
    to_char(ah.effective_start_date,'DD-MON-YYYY') transaction_date,
    to_char(nvl(i.payment_due_date,
    to_date('30-12-9999','dd-mm-yyyy')),'DD-MON-YYYY')
    due_date, ah.current_balance-ah.previous_balance amount,
    decode(ah.invoice_id,null,'A','I') transaction_type
    3 4 5 6 7 8 from account a,account_history ah,invoice i_+
    where a.account_id=ah.account_id
    and a.account_type_id=1000002
    and round(a.account_balance,2) > 0
    and (ah.invoice_id is not null or ah.adjustment_id is not null)
    and ah.CURRENT_BALANCE > ah.previous_balance
    and ah.invoice_id=i.invoice_id(+)
    AND a.account_balance > 0
    order by a.account_id,ah.effective_start_date desc; 9 10 11 12 13 14 15 16
    Explained.
    SQL> select * from table(dbms_xplan.display);
    PLAN_TABLE_OUTPUT
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)|
    | 0 | SELECT STATEMENT | | 544K| 30M| | 693K (20)|
    | 1 | SORT ORDER BY | | 544K| 30M| 75M| 693K (20)|
    |* 2 | HASH JOIN | | 544K| 30M| | 689K (20)|
    |* 3 | TABLE ACCESS FULL | ACCOUNT | 20080 | 294K| | 6220 (18)|
    |* 4 | HASH JOIN OUTER | | 131M| 5532M| 5155M| 678K (20)|
    |* 5 | TABLE ACCESS FULL| ACCOUNT_HISTORY | 131M| 3646M| | 197K (25)|
    | 6 | TABLE ACCESS FULL| INVOICE | 262M| 3758M| | 306K (18)|
    Predicate Information (identified by operation id):
    2 - access("A"."ACCOUNT_ID"="AH"."ACCOUNT_ID")
    3 - filter("A"."ACCOUNT_TYPE_ID"=1000002 AND "A"."ACCOUNT_BALANCE">0 AND
    ROUND("A"."ACCOUNT_BALANCE",2)>0)
    4 - access("AH"."INVOICE_ID"="I"."INVOICE_ID"(+))
    5 - filter("AH"."CURRENT_BALANCE">"AH"."PREVIOUS_BALANCE" AND ("AH"."INVOICE_ID"
    IS NOT NULL OR "AH"."ADJUSTMENT_ID" IS NOT NULL))
    22 rows selected.
    Index Details:+_
    SQL> select INDEX_OWNER,INDEX_NAME,COLUMN_NAME,TABLE_NAME from dba_ind_columns where
    2 table_name in ('INVOICE','ACCOUNT','ACCOUNT_HISTORY') order by 4;
    INDEX_OWNER INDEX_NAME COLUMN_NAME TABLE_NAME
    OPS$SVM_SRV4 P_ACCOUNT ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT CUSTOMER_NODE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_ACCOUNT_TYPE ACCOUNT_TYPE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_PREVIOUS_INVOICE PREVIOUS_INVOICE_ID ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_NAME ACCOUNT
    OPS$SVM_SRV4 U_ACCOUNT_NAME_ID ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_LAST_MODIFIED_ACCOUNT LAST_MODIFIED ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_INVOICE_ACCOUNT INVOICE_ACCOUNT_ID ACCOUNT
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ACCOUNT SEQNR ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_INVOICE INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA CURRENT_BALANCE ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA INVOICE_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_CIA ACCOUNT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_LMOD LAST_MODIFIED ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADINV ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_PAYMENT PAYMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_ADJUSTMENT ADJUSTMENT_ID ACCOUNT_HISTORY
    OPS$SVM_SRV4 I_ACCOUNT_HISTORY_APPLIED_DT APPLIED_DATE ACCOUNT_HISTORY
    OPS$SVM_SRV4 P_INVOICE INVOICE_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE CUSTOMER_INVOICE_STR INVOICE
    OPS$SVM_SRV4 I_LAST_MODIFIED_INVOICE LAST_MODIFIED INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT ACCOUNT_ID INVOICE
    OPS$SVM_SRV4 U_INVOICE_ACCOUNT BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_BILL_RUN BILL_RUN_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_INVOICE_TYPE INVOICE_TYPE_ID INVOICE
    OPS$SVM_SRV4 I_INVOICE_CUSTOMER_NODE CUSTOMER_NODE_ID INVOICE
    32 rows selected.
    Regards,
    Bathula
    Oracle-DBA

    I have some suggestions. But first, you realize that you have some redundant indexes, right? You have an index on account(account_name) and also account(account_name, account_id), and also account_history(invoice_id) and account_history(invoice_id, adjustment_id). No matter, I will suggest some new composite indexes.
    Also, you do not need two lines for these conditions:
    and round(a.account_balance, 2) > 0
    AND a.account_balance > 0
    You can just use: and a.account_balance >= 0.005
    So the formatted query isselect a.account_id,
           round(a.account_balance, 2) account_balance,
           nvl(ah.invoice_id, ah.adjustment_id) transaction_id,
           to_char(ah.effective_start_date, 'DD-MON-YYYY') transaction_date,
           to_char(nvl(i.payment_due_date, to_date('30-12-9999', 'dd-mm-yyyy')),
                   'DD-MON-YYYY') due_date,
           ah.current_balance - ah.previous_balance amount,
           decode(ah.invoice_id, null, 'A', 'I') transaction_type
      from account a, account_history ah, invoice i
    where a.account_id = ah.account_id
       and a.account_type_id = 1000002
       and (ah.invoice_id is not null or ah.adjustment_id is not null)
       and ah.CURRENT_BALANCE > ah.previous_balance
       and ah.invoice_id = i.invoice_id(+)
       AND a.account_balance >= .005
    order by a.account_id, ah.effective_start_date desc;You will probably want to select:
    1. From ACCOUNT first (your smaller table), for which you supply a literal on account_type_id. That should limit the accounts retrieved from ACCOUNT_HISTORY
    2. From ACCOUNT_HISTORY. We want to limit the records as much as possible on this table because of the outer join.
    3. INVOICE we want to access last because it seems to be least restricted, it is the biggest, and it has the outer join condition so it will manufacture rows to match as many rows as come back from account_history.
    Try the query above after creating the following composite indexes. The order of the columns is important:create index account_composite_i on account(account_type_id, account_balance, account_id);
    create index acct_history_comp_i on account_history(account_id, invoice_id, adjustment_id, current_balance, previous_balance, effective_start_date);
    create index invoice_composite_i on invoice(invoice_id, payment_due_date);All the columns used in the where clause will be indexed, in a logical order suited to the needs of the query. Plus each selected column is indexed as well so that we should not need to touch the tables at all to satisfy the query.
    Try the query after creating these indexes.
    A final suggestion is to try larger sort and hash area sizes and a manual workarea policy.alter session set workarea_size_policy = manual;
    alter session set sort_area_size = 2147483647;
    alter session set hash_area_size = 2147483647;

  • BO DI for extracting data from legacy to SAP BW 7.0

    Hi,
    Has anyone worked on scenario where BO Data Integrator is being used as ETL tool for extracting data from legacy systems to SAP BW 7.0 ?
    Please share your experience on such a landscpae or related documents any of you had come across.
    Regards,
    Pritesh

    Hi,
    BO Integrator connects with other system (Database) and you will be creating your reports on the top of that DB or system. It never saw it being used to pull data from legacy to BW.
    Regards,
    Bashir Awan

  • TDMS for reduce data volume

    Hi Experts,
    We have implemented TDMS and installed in all systems.
    Our system landscape is like - DEV, QUAL and PRDN (ECC system)
    we would like to use TDMS for reduce data volume in non production systems.
    We have one more project system which is in the transport line.
    Could any one please inform which is the best scenario to use TDMS for reducing data volume in non-production systems without much distrubance in production system.
    Also how it is taking care of development system, since the customization will be overwritten in DEV system.
    Thank you in advance.
    Shylesh

    Hi Shylesh,
    you can use several packages to reduce your data. If you need application data you can use a time slice (package TDTIM), if you don't need it you can use the master data and customizing option (TDMDC).
    I suppose you are going to transfer from one client to another so best thing to do is to create a new client on dev if you still need the old customizing settings.
    Concerning the impact of the productive environment. Can you give me an idea on the database size ?
    Regards,
    Eddy

  • Configurations for extracting portal activity data from netweaver portal.

    Hi,
    I need to fetch  the portal activity data from the portal into Solution Manager. I would like to know the configurations on the portal that will be needed in order to extract the portal actvity data . How will i proceed extracting the portal activity data?
    thanks and regards,
    Divya.

    Hi Divya,
    We can not say that Early Watch Alerts is same as the  the portal activity report.
    Its a diagnosis service, for the solution monitoring of SAP and non-SAP systems
    in the SAP Solution Manager. Its gives you the complete details of your System.
    If we talk about Portal then you can get the details like :
    1. Java Heap details
    2. Memory Uses
    3.Portal Objects access details (Like Activity report )
    4.Server Nodes details
       and many more things.................
    The Underlying concept is to ensure smooth operation of individual SAP Systems
    by keeping you informed of their status and allowing you to take action,
    before severe technical problem occurs.
    Check the below link for ewa:
    [SAP EarlyWatch Alert overview|http://www.mediafire.com/?gilijcnnwet]
    [SAP EarlyWatch Alert|http://help.sap.com/saphelp_sm40/helpdata/en/87/50cce344ae495bae240add66a28ed7/content.htm]
    [Using SAP EarlyWatch Alert|http://help.sap.com/saphelp_sm40/helpdata/en/5a/b27f17edb04a6882c59f6619a267b9/content.htm]
    Regards,
    Piyush

  • How to Extract Huge SAP HR Data into Excel?

    Dear Experts & Gurus,
    Need your Valuable suggestions regarding this scenario.
    I need to extract entire SAP HR data, modules are OM,PA,TM,PY and PD into Excel sheets because our clients are migrating to other ERP. The data volume is very huge, as they are using SAP since 14years. Can we use se16 to extract the data , will it able to handle in extracting the data from HR Tables or any other standard ways to extract this huge data without any dumps or errors .
    Appreciate if you provide some useful suggestions or tips  for this scenario.
    Thank you.
    Regards
    Vicky

    Dear Liran,
    Thanks a lot for your valuable suggestions.its really given an brief idea now.
    First:
    What's the purpose of this data download?
    Do you wish to transfer the data from SAP into the new ERP system? Or do you need the download as a backup?
    Yes , Transfer the data from SAP to New ERP system, it is not for Backup.
    IDOCS/ALE from SAP to New ERP?
    Second
    I think Payroll results stored in cluster tables, so here we need Abaper to write program. For OM , PA and TM can we extract via Se16/ Se16n .
    Thank you very much in advance.
    Regards
    Vicky

  • Using RSCRM_BAPI with Huge Data Volume

    Hi,
    I am using RSCRM_BAPI to extract a query output into a database table. But, the query returns a large volume of data. I am not sure whether RSCRM_BAPI works fine when data volume is huge. Please suggest whether using this will be a good design or any other method is available to take care of such a scenario.
    Regards,
    Dibyendu

    I have used RSCRM_BAPI when the records were exceeding 65000(limitations of excel) and it worked for me...
    I think should work for you also....
    But there are some limitations....
    like u cannot see texts etc,,,,
    Assign point if it helps,
    Ajay

  • MaxDB data volumes usage

    Hi all,
    I have the following configuration for my MaxDB (which is a part of a DMS system).
    Version is MaxDB 7.6
    Data volumes are:
    1) 2.GB
    2) 2.GB
    3) 10GB
    4) 10GB
    5) 10GB
    6) 3GB
    7) 3GB
    8) 3GB
    It's seems that for volumes 3, 4 & 5 (10GB size) only 4.5GB are used.
    My question is: why the free 5.5GB of the 10GB volumes is not used?
    Is it a configuration issue or does MaxDB determine where to write automatically?
    I don't see any performance problems.
    Thanks,
    Omri

    > I have the following configuration for my MaxDB (which is a part of a DMS system).
    > Version is MaxDB 7.6
    >
    > Data volumes are:
    > 1) 2.GB
    > 2) 2.GB
    > 3) 10GB
    > 4) 10GB
    > 5) 10GB
    > 6) 3GB
    > 7) 3GB
    > 8) 3GB
    >
    > It's seems that for volumes 3, 4 & 5 (10GB size) only 4.5GB are used.
    They will be used - just put more data into your content server (?).
    > My question is: why the free 5.5GB of the 10GB volumes is not used?
    > Is it a configuration issue or does MaxDB determine where to write automatically?
    No, there is nothing wrong here.
    MaxDB choses the datavolumes to save changed/new pages in a order based on their relative filling.
    Nevertheless as long as there is freespace in a volume it is possible that this will get used as well.
    Maybe you're thinking of a feature that existed until SAP 7.4 where the database would even out the filling degree in times of low activity. This feature is not present anymore - instead changed pages are relocated during savepoints.
    > I don't see any performance problems.
    Why should you? Content server usually does not put the same level of traffic to the database as a OLTP system would. For that you can use different sized data volumes without any performance drops.
    regards,
    Lars

  • Recommended sizing for ZLM server/infrastructure

    Hi,
    I hope this is the right category to place this question.
    I was wondering if any of you does have some recommendation according to the correct sizing for a ZLM infrastructure meaning how large would be a typical management zone when using the postgres db shipped with ZLM?
    What type of hardware are you running your ZLM server software on?
    Field reports welcome :)
    Best regards
    --Daniel

    Daniel ,
    we have just one management zone with a primary and a secondary at the backup data center. We managed sles 9/10 (soon 11) and RH4 and a bunch of third party software.
    The package repository is currently at 35 GB and the database is at 3 GB. The object store is just about 100mb so no need to think about that. that will not grow very much when adding new systems or patches.
    What I would take care of is that the database and package repository are located on a logical volume that you can resize online in case you need to expand it. We use reiser for that. The datbase as you can see is also not so big. But when you add new distributions (SLES 11 or so) you might need another3 GB for 32 and 3GB for the 64bit version. Then when then number of updates increase you need even more. Then later on for SP1 again about 5-6GB for both architectures ....
    Rainer

  • Tables/FM for opening/closing stock for material on given posting date

    Hi All,
    I had requirement of find the opening/closing stock for material on given posting date. The values we can find out using TCode - MB5B.
    But I cannt use MB5B to extract data because I need to find detials for thousands of material.
    Can you please adivce if their is any FM or tables through which i can find this detial on any given date apart from current date.
    Thanks & Regards,

    Hi,
    [https://forums.sdn.sap.com/click.jspa?searchID=19631709&messageID=6593903]
    [https://forums.sdn.sap.com/click.jspa?searchID=19631709&messageID=6630975]

  • Demo - Extraction of Customer Master Data from SAP R/3 ECC6 to MDM5.5

    Hello,
    I am new to MDM . Would appreciate if someone can give a solution to the below request.
    We are doing a demo project in our Company to extract all the master data related to SALES cycle. We are doing an extraction of Customer Master Data. Currently we have the unlicensed version of MDM and very soon we will get the licensed version.
    We have the Customer Master Repository created in MDM Console. The IDOC type we are using is DEBMDM05.
    Is there a easy way to map the fields of IDOC to the MDM Repository. Right now we are doing a manual process of comparing the IDOC fields and MDM Repo in an excel but this is tedious and time consuming.
    Can anyone suggest an easier way to identify so that we know which fields match correctly.
    Thanks
    Leena

    Hi Leena,
    First of all you have 2 options to generate IDOCs from SAP R/3 system
    1.mdmgx>one IDoc per record
    2.mdm_clnt_extr>for generating single IDoc having entire set of records.
    Secondly,if you want to get the data mapped automatically,you should use standard maps provided by SAP(Business Content).
    If at all you have created new fields then you should manually map those fields.
    If any other queries pl ask
    Reward points if found useful
    Regards,
    Pramod.

  • Extract Month & Year  From Date

    Hi All,
    I have a key figure quantity and i want the data on date,for the month and for the year.
    i create the variables for the date,for the month(ie from date to date) and for the year(ie, from date to date) seperatly that means i have three input variables which may sometimes create confusion.
    My problem is that I want to enter only date and the variable itself calculate the month and year from the date so that it returns the data as desired above.
    for eg. : Now i to enter date : 3/14/2009
              enter MTD (FROM/TO) : 3/01/2009 - 3/14/2009
              enter YTD (FROM/TO) : 4/01/2008 - 3/14/2009
    I want to enter only date : 3/14/2009
    and the variable itself extract the month and year till date.
    and also the same variable calculates number of days so that need not to take the formula variable of date difference to calculate the average qty.
    Neha..

    Hi,
    1. Create User Entry Variable on 0CALDAY : Name = ZCDAY.
    2. Craete a Customer Exit Variable on 0CALDAY: Name  = ZMTD.
    3. Craete a Customer Exit Variable on 0CALDAY: Name  = YMTD.
    Properties of Customer Exit Variables.
    Variable reporesents = Single
    Variable Entry = Mandatory
    Processing By = Customer Exit.
    Character = Calender Day
    Here I'm thinking that ZKF is your key figure.
    In columns you create two selections one is for MTD and other is for YTD.
    In MTD selection, drag and drop ZKF and Drag and Drop 0CALDAY and then Right Clcik and Restrict-->
    Selection = Value Range. (In Between) and restrict with the follwoing Variables
    []ZMTD; ZCDAY.
    In YTD selection, drag and drop ZKF and Drag and Drop 0CALDAY and then Right Clcik and Restrict-->
    Selection = Value Range. (In Between) and restrict with the follwoing Variables
    []YTD; ZCDAY.
    Then write the following code in I_STEP = 2.
    DATA: ZT_DT1 TYPE SY-DATUM,
              ZT_DT2 TYPE SY-DATUM,
              ZT_SDT TYPE SY-DATUM,
              ZT_YR(4) TYPE N,
              ZT_DY(2) TYPE N,
              ZT_MT(2) TYPE N,
              ZE_TT(2) TYPE N,
              ZPOPER TYPE POPER,
             ZRELJR TYPE RELJR.
    WHEN 'ZMTD_A'.
      LOOP AT i_t_var_range INTO loc_var_range WHERE vnam = 'ZCDAY'.
        IF sy-subrc = 0.
          CLEAR: l_s_range.
          ZT_DY = '01'.
          ZT_SDT = loc_var_range-low.
          ZT_MT = ZT_SDT+4(2).
          ZT_YR = ZT_SDT+0(4).
          CONCATENATE ZT_YR ZT_MT ZT_DY INTO ZT_DT1.
          l_s_range-low = ZT_DT1.
          l_s_range-sign = 'I'.
          l_s_range-opt = 'EQ'.
          APPEND l_s_range TO e_t_range.
        ENDIF.
      ENDLOOP.
    WHEN 'ZYTD'.
      LOOP AT i_t_var_range INTO loc_var_range WHERE vnam = 'ZCDAY'.
        IF sy-subrc = 0.
          CLEAR: l_s_range.
          ZT_DY = '01'.
          ZT_SDT = loc_var_range-low.
          CALL FUNCTION 'DATE_TO_PERIOD_CONVERT'
            EXPORTING
              I_DATE               = ZT_SDT
    *         I_MONMIT             = 00
              I_PERIV              = 'V3'
           IMPORTING
             E_BUPER              = zbuper
             E_GJAHR              = zbdatj
           EXCEPTIONS
             INPUT_FALSE          = 1
             T009_NOTFOUND        = 2
             T009B_NOTFOUND       = 3
             OTHERS               = 4
          IF SY-SUBRC <> 0.
             MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
             WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
          ENDIF.
          CALL FUNCTION 'FIRST_AND_LAST_DAY_IN_YEAR_GET'
            EXPORTING
              I_GJAHR              = zbdatj
              I_PERIV              = 'V3'
           IMPORTING
             E_FIRST_DAY          = ZT_DT2
    *         E_LAST_DAY           =
           EXCEPTIONS
             INPUT_FALSE          = 1
             T009_NOTFOUND        = 2
             T009B_NOTFOUND       = 3
             OTHERS               = 4
          IF SY-SUBRC <> 0.
             MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
             WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
          ENDIF.
          l_s_range-low = ZT_DT2.
          l_s_range-sign = 'I'.
          l_s_range-opt = 'EQ'.
          APPEND l_s_range TO e_t_range.
        ENDIF.
      ENDLOOP.
    Thanks
    Reddy

Maybe you are looking for

  • Need help getting my itunes library to work!!

    Recently my computer CRASHED.... it wouldn't let me into windows, it was pulling all kinds of error messages. So to fix the problem I wiped the system and reinstalled everything. However, now the Itunes that came with my Ipod doesnt work... So I came

  • 10.5.2 won't send signal to external projector

    Help! I do speaking engagements for a living and just updated to 10.5.2. When I hook an external projector to the computer (DVI to VGA) my MacBook Pro does see the external projector, but will not send a signal to it. It worked yesterday, but as soon

  • Notifications in Portal UWL and external email Id

    When we run the workflow, the Tasks showup in the Universal worklist but the notifications don't (even though they appear in R3 Business workplace). Is this correct? DO we need to do some settings to take care of this. We also need to send the notifi

  • Starting Oracle Application Server

    I've installed OAS 408 on an NT box running service pack 5. When I go to the server manager page I select "All" and then select the start button. I get the following. Any ideas about the "BADFRAME"? Thanks! Please wait while the command is being proc

  • [ANN] Run Oracle Forms and Reports on the Web for $5,000 only

    A new time limited price promotion for customers looking to upgrade Oracle Forms to the Web is now offered. Get a restricted license to use Oracle9iAS Forms Services and Oracle9iAS Reports Services only for $5,000 per CPU (or $100 per named user). Or