Planning area dp apo

hi
guru
so i have created PA. and also time series created for it.. and also PB DV created for it..
now i want to add a KEY figure to that PA??? is that POssible? how i should i add a kf to PA????
the same senerio to MPOS and PB????

Hi dallyanusha,
I think there are series of questions.  I will try to cover few of them.
1) Transfer of data from R/3 to APO.
In SAP, data transfer from R/3 to APO will be through Core interface.
2) Transfer of data from R/3 to BI
This will be done by building infoobjects like infocubes in BW and
store R/3 data in cubes
3) Fixed key figure value
Contains the key figure, in which the fixed value of the basic key figure is saved.
4) Use of Fixed key figure value
While planning a key figure, you can wish to protect the planned value against any overwriting. The value can be fixed in the planning table. The fixed key figure is required to store this information. The fixed basic key figure value is copied down to the finest detail into this key figure. The information with this is available on an aggregated level as regards what share of the key figure is fixed and must not be changed. If the share is 100% then the key figure must not be changed by planning functions. If the share is between 0& and 100% then the value may only be changed in a restricted way. The change is only disaggregated on those details that are not fixed.
Regards
R. Senthil Mareeswaran.

Similar Messages

  • Copy Data from Non CBF Planning Area to CBF Planning Area in APO DP

    Hi Experts,
    I am working in APO implementation in an automative client, here we have material type Vehi and configurable material.
    I have done my forecasting in Non CBF planning area . Now I want to copy Key figure of final conenses forecast which is in non - CBF planning area to a CBF planning area .
    Can anybody has worked on similar scenario or worked in copying data from a non - CBF planning area to CBF planning area.
    Looking for your valuable inputs
    Thanks and regards
    Kapil

    Hi Kapil!
    We have the same need mentioned by you "Copy Data from Non CBF Planning Area to CBF Planning Areau201C.
    We obtained the source data, but can not do to write the same in CBF area.
    You managed to make the recording of such data in CBF area? How did they do it?
    Thanks and regards.
    Clelidon

  • RE : BI APO Question Reg Data feeding from cube to Planning area.

    Hi BW Experts,
    iam working in an Implementation project for SCM in BW prcisely working with APO Bw..
    For that I have taken historical data as a flat file and loaded it in to the external BW Infocube and its fine...
    Second step I have created generate export datasource on topr of BW infocube and replicated in to Bw and used this export datasource as datasource for APO BW Infocube which is inbulit BW System from External Bw..
    also I have created tranformations and data is loaded in the BW cube in APO system.Also Included Version charecterstics..
    When I try to fed the APO Cube data to planning area Iam gettinga the following  warning itsnot an error:
    1.Key figure copy: InfoCube - planning area (DP) 01.01.2010 to 31.12.2010-- Successful
    *2.No data exists for the selection made (see long text*
      Diagnosis:Data could not be loaded from the Cube for the selection you made. Check whether the Cube actually contains data that is relevant for your selection.
    For the second point I have time charecterstics filled in the infocube which Iam feding to a Planning area like 0CALMONTH,0CALWEEK,FiscVarnt,0CALMONTH
    3.Characteristic assignment: No data copied --- Message
    Can you please help me with your thoughts so that i wll try to corner the issue I will be highly obliged

    Hi,
    As I understood, you have loaded data from external BW cube to APO BW cube and now loading planning area from APO BW cube.
    I hope your settings in /SAPAPO/TSCUBE transaction code would be correct and you have selected the correct planning version with correct cube.
    Check if Data in APO BW cube is available for reporting or not and data is avilable for given selction (if any but I guess you are not giving any).
    Thanks,
    S

  • Data source for a planning area

    Hi Experts,
    I have created a data source 9AZDP_APAP for my planning area ZDP_APAP. When I'm creating this data source, SAP itself prefixing my datasource name with '9A'.
    I'm wondering is this ( Starting the data sorce name with '9A') with in our system only or with every system.
    If it is with every system then is there anyway I can override this.
    The reason why I'm asking this question is my data source is giving error " M version is different from A version' in RSA1, when I'm trying to activate my datsource I'm getting authorization error. When I contacted my security people on this they are telling I'm not following the naming standards. They are telling that the data source name should start with Z.
    All the inputs on this will be highly appreciated.
    Regards,
    Brahma Reddy

    Hi,
    Generating a data source on a planning area in APO is really very helpful function we have. this will help us to save a lot lot time to generate cube/rules/source.
    Definately, when system generates the data source it will follow its own logic "9A" which we can not change.
    I guess you  create a DataSource  manually if required, where you can follow your own logic, but that is not really consistent as system generates. becuase when it generates, it will check for time char and all based on storage bucket profile and generates.
    Hope my information will help you in understanding the importance of auto generation of data source
    Yarala

  • Time Series Objects for a Planning Area

    Hi all,
    Can anyone let me know why do we create Time Series Objects for a Planning Area.
    What is its Role & significance..
    Regards,.
    Vishal.S.Pandya

    time series is usually a Demand planning concept(and used in SNP as well but in SNP its predominantly the order series that plays a main role)
    time series is a general concept in statistics(and forecasting) wherein the value of a key figure is represented in a time bucket
    Time series gives you an idea of the gradual change in the values by time and the relation of the future based on the past
    planning area in APO (and other tools) tries to represent this as a 2 dimensional model with time on the columns and key figures in the rows. the value that you load into the cells that are formed by the above are based on the characterisitic values you choose from your MPOS which is linked to the way the values are stored in the planning  area
    The planning area stores data for each key figure in the smallest unit of time(technical storage buckets) and the lowest level of characterisitc value combination

  • Request shows in Red status after extraction of Data from Planning area

    Hi Experts,
    I am extracting data from DP planning area to APO BW Info Cube. It was working fine but from last few days after extraction we are getting Red status of that request. However whatever data packages we are extracting we are getting same in to Info Cube also. But only status of that request is in Red.
    I gone through step by step analysis I found "Data selection succesfully finished?" statement in Red status & rest are in green. When I gone through the details of this statement then I found Missing "Message:Selection Completed" in yellow status.
    Any Idea? How do i resolve this issue?
    Please advise..
    Regards
    Sujay

    Hi Sujay,
    Check the logs obtained after the extraction.
    This only will tell you fare idea for you to proceed further
    to resolve the error that has happened.
    Also check whether the extracted data is correct by
    random sampling of extracted data
    Regards
    R. Senthil Mareeswaran.

  • Planning area data not updated in infocube in APO system

    Hi all,
    User is uploading a flat file using addon program to an infocube 1, from this infocube data for  sales forecast KF is getting transferred to a planning area KF using standard program RTS_INPUT_CUBE.  I can see the updated data in Planning book (which refer the same planning area) data views. In planning book the sales forecast data is also getting copyied to second KF 'Arrangement FC' and its correct.
    Now there is a infocube 2 (second) which is getting data directly from this planning area (infocube also contains both KFs). But When i checked this infocube 2 , the updated data is availabe in Sales forecast KF , but arrangement forecast KF is having old data only and user is running query on second KF , so he is getting wrong report.
    Since there is no data flow for this infocube 2, and it is getting data from planning area, I feel its remote infocube.
    I have also found the datasource for this planning area but don't know how to move further to find why data is not updating properly? Please help on this.
    I have information that 2 weeks before data was properly updated, but this time data is not updated,
    system version  is SAP SCM 4.0

    Hi Vivek
    it is advisable to run the background jobs when the planning books are not being accesses by the users to prevent such inconsistencis. Hence i would advise you to run the jobs durng non-working hours. and if you have a global system, then you may restrict to run the jobs based on regional levels.
    in addition, it is also a good practice to run consistency jobs before and after your have completed the background jobs. i hope you are using process chains to execute the sequeuce of jobs. if yes, then you can add consistency check jobs in the process chains.
    some of the consistency check reports are :
    /SAPAPO/OM17 - liveCache Consistency Check
    /SAPAPO/TSCONS - Consistency Check for Time Series Network
    /SAPAPO/CONSCHK - Model Consistency Check
    and so and so forth..
    you can find these conssistency jobs under APO Adiminstration --> Consistency checks in APO.
    let me know if this helps.
    Rgds, Sandeep

  • Help Required for Mapping Key figures from Cube to APO Planning area.

    Hello Experts,
    We have created cube in APO BW and now we want to map it to Planning area how we can map.
    Can any body explain how we can map keyfigures?
    What is the use of livechache how it will be updated?
    Regards
    Ram

    Hi,
    I am not very sure about the 9ARE aggregate (haven't used it in backups), but RTSCUBE is used to copy time Series (TS) KF data from cube to planning area (SNP or DP).
    Are you trying to restore some time series data from your backup cube to the planning area? If yes, then do a mapping of characteristic from cube to planning area in RTSCUBE, and also map the TS KF between cube and planning area.
    If your KF is not a time series KF, then you can't copy it from cube to planning area. You could get data to cube for some reporting, otherwise I am not sure what use the backup is for you. For SNP, most of the data would be received from R/3, so there's not much point in having a backup.
    Hope this helps.
    Thanks - Pawan

  • APO- BI Datasource from Planning Area

    Hi All,
    I need help with APO-BI datasource generated from Planning Area.
    In the Dev environment we had two clients:
    DCLNT020 (Holds APO part) DCLNT010 (Holds BI workbench).
    So a datasource was generated from the Planning area in DCLNT020 --> it was replicated in DCLNT010 --> data from Planning Area was extracted to BI cube using this.
    Now we transported this datasource to the Test environment which has only one client (TCLNT010). I maintained the Source to target mapping there such that DCLNT020 -- TCLNT010 and DCLNT010 -- TCLNT010.
    However the Transport fails and the error message is:
    Cannot replicate DataSource
    Errors occurred during post-handling RS_AFTER_IMPORT for ISFS L
    If I go to the Test system and try to generate the transported Datasource directly from the Planning area again, it says this DataSource already exists. However I cannot see this datasource in the system even after replicating and refreshing multiple times.
    Please provide your inputs as to what might be wrong and hat I need to do to solve this.
    TIA
    Amrita

    Hi   Amrita Goswami
    Based on the above post it seems to be your maintain two clients in Dev one is for creation and another is for testing and when it comes to test environment your maintain only one client and maintain the DS in one it wont give any impact..
    Based on the error
    > +Cannot replicate DataSource+
    > +Errors occurred during post-handling RS_AFTER_IMPORT for ISFS L+
    There could be two reasons
    1) Needs to replicate the data source once you have imported it to test environ ment and than ran the program "RSDS_DATASOURCE_ACTIVATE_ALL" by giving the name of the source and DS name if its BI 7.0
    If its 3.x then have to execute the program :"RS_TRANSTRU_ACTIVATE_ALL" By specifying the transfer structure name.
    2) RS_AFTER_IMPORT  in some cases its because of improper transport of the update rules.
    Solution would be recollect the transport and release the DS transport first and execute the ( 1)Activities and then transport the remaining._
    Hope its clear a little..!
    Thanks
    K M R
    ***Even if you have nothing, you can get anything.
                                                But your attitude & approach should be positive..!****
    >
    Amrita Goswami wrote:
    > Hi All,
    > I need help with APO-BI datasource generated from Planning Area.
    >
    > In the Dev environment we had two clients:
    >
    > DCLNT020 (Holds APO part) DCLNT010 (Holds BI workbench).
    >
    > So a datasource was generated from the Planning area in DCLNT020 --> it was replicated in DCLNT010 --> data from Planning Area was extracted to BI cube using this.
    >
    > Now we transported this datasource to the Test environment which has only one client (TCLNT010). I maintained the Source to target mapping there such that DCLNT020 -- TCLNT010 and DCLNT010 -- TCLNT010.
    >
    > However the Transport fails and the error message is:
    > Cannot replicate DataSource
    > Errors occurred during post-handling RS_AFTER_IMPORT for ISFS L
    >
    > If I go to the Test system and try to generate the transported Datasource directly from the Planning area again, it says this DataSource already exists. However I cannot see this datasource in the system even after replicating and refreshing multiple times.
    >
    > Please provide your inputs as to what might be wrong and hat I need to do to solve this.
    >
    > TIA
    > Amrita
    Edited by: K M R on Feb 6, 2009 12:03 PM
    Edited by: K M R on Feb 6, 2009 12:18 PM

  • APO DP - loading from InfoCube to planning area

    I am using APO DP V5.
    I have the following situation:
    1. I am extracting sales history data at the end of each day from a connected ECC system into an InfoCube, using delta logic, so in the Cube I just have the new sales history transactions (despatches) for the day
    2. I am then loading data from the Cube to my DP planning area via transaction /SAPAPO/TSCUBE
    I assume I must have the 'add data' flag set for each key figure in this transaction, to ensure I see the consolidated sales history in the planning area.
    Is this the 'best practice' approach for regular loading of sales history? I assume it's better to have a 'delta' approach for the Cube, for improved performance.
    Thanks, Bob Austin

    Hi,
            Good questions!
    1. What does the 'period' really refer to? Is it referring to the date of a particular key figure? Or the 'date of data being added to Cube'?
    A: Both are same
    The date is generally the date in your cube like the calendar day, month etc. This date is again based on a time stamp in the format DDMMYYYYHHMMSS. The calendar day is part of this field as DDMMYYYY. Also the system recognizes the changes by the same time stamp. So, if a customer changes the qty 05/15 at 3.30 pm, then the time stamp is 15052007153000. The calendar day in your cube or your key figure date is 15052007 and the delta is recognized by the changes in time stamp, between last load at the current system time. So , you are talking about the same time field.
    Check in your system if this is the same case you got. let me know if not.
    2. Suppose original dispatch qty = 100 (two weeks ago), and 'today' the customer returns a qty of 60; how does system handle this - I would want this posted to the original date of two weeks ago.
    A: The data from your ECC system is generally brought to an ODS first. The reason is we overwrite the data there if there is any data that has the same key. If your key for the ODS is Customer and division. Then you overwrite the customer qty for that division whenever the value changes. If you need it by time, lets say per month, include it in the key. The system over writes the value for that month only. For next month, it's a new record.
    In your case, if the qty. is 100 2 weeks back and now it's 60, if the time stamp is not in key, the system overwrites it to 60 and you have only 60 when you load it to your ODS and thereby to your PA as it overwrites. Delete the delta in your ODS and it shows the same 100 again. Then load it to PA. This is not a good option. The alternative is to include time stamp like calweek in your ODS key and load it over to cube. That way you have two records.
    I hope I answered all your questions.  Please feel free to ask more. I would love to answer that way I can also brush things up that were unused.
    Thanks.

  • APO DP - load from InfoCube to a 'fixed' cell in planning area

    I am using APO DP V5.
    Suppose I have a key figure which is fixed for a particular time bucket cell.
    If I try and load data into this cell from an InfoCube, what will be the effect? I assume that the cell will NOT be udpated.
    Regards,
    Bob Austin, Atos Origin

    Hi,
    Fixing information will be
    extracted when saving the data into an InfoCube and will be reloaded
    when using transaction /SAPAPO/TSCUBE to load the data back into the
    planning area. If you load the data back only on the level of the
    basis planning object structure, the data will be aggregated to the
    level of the aggregates but no fixing information will then be
    available on the level of an aggregate.
    and also /sapapo/tscube doesn't check fixing info in LC when data is written to LC. In case there is fixed data in iCube, that fixing info will be loaded into LC.
    You can also refer the note  1057308.
    Regards,
    Srikanth.

  • Loading data from APO-BW cube into DP planning area

    Hello,
    Please explain to me how to load data from an infocube within APO into the DP planning area.  So far, I generated the data mart, created the update rules for the DP planning area (cube) and replicated data sources.  The problem that I am having is that when I create the infopackage, there is no data targets on the data target tab.  How can I get the data targets to display?  All the other actions appear to be set up correctly (confirmed by BW experts).  Can data be loaded from an APO-BW cube directly to the DP planning area cube?
    Thanks

    Hi James,
    Loading of data from an InfoCube to a Planning Area is done using transaction /SAPAPO/TSCUBE (program /SAPAPO/RTSINPUT_CUBE). You just specify the InfoCube, target Planning Area, Selection condition, Infocube Key Figure to PA KF mapping, etc).
    Have you tried this transaction already?
    Hope this helps.

  • Datasource on APO Planning Area - Transportation Error

    Hi All,
                 I have created the Datasource on APO Planning Area. Datasource working fine check in RSA3 and also in BW side. when transporting the data source from APO Dev to APO QA i am getting following error and transport fails. Please suggest.
    Thanks
    Christopher
       Execution of programs after import (XPRA)
       Transport request   : AD1K909333
       System              : AQ3
       tp path             : tp
       Version and release: 372.04.10 700
       Post-import methods for change/transport request: AD1K909333
          on the application server: hllsap112
       Post-import method RSA2_DSOURCE_AFTER_IMPORT started for OSOA L, date and time: 20080725125524
       Execution of "applications after import" method for DataSource '9ADS_PP_APO'
       Import paramter of AI method: 'I_DSOURCE:' = '9ADS_PP_APO'
       Import paramter of AI method: 'I_OBJVERS:' = 'A'
       Import paramter of AI method: 'I_CLNT:' = ' '
       Import paramter of AI method: 'LV_CLNT:' = '100'
       DataSource '9ADS_PP_APO': No valid entry in table /SAPAPO/TSAREAEX
       Planning area for DataSource '9ADS_PP_APO' does not exist in target system
       Extract structure /1APO/EXT_STRU100002737 is not active
       The extract structure /1APO/EXT_STRU100002737 of the DataSource 9ADS_PP_APO is invalid
       Errors occurred during post-handling RSA2_DSOURCE_AFTER_IMPORT for OSOA L
       RSA2_DSOURCE_AFTER_IMPORT belongs to package RSUM
       The errors affect the following components:
          BC-BW (BW Service API)
       Post-import method RSA2_DSOURCE_AFTER_IMPORT completed for OSOA L, date and time: 20080725125532
       Post-import methods of change/transport request AD1K909333 completed
            Start of subsequent processing ... 20080725125524
            End of subsequent processing... 20080725125532
       Execute reports for change/transport request: AD1K909333
       Reports for change/transport request AD1K909333 have been executed
            Start of................ 20080725125532
            End of.................. 20080725125532
       Execution of programs after import (XPRA)
       End date and time : 20080725125532   Ended with return code:  ===> 8 <===

    Christopher,
    There seems to be no extract strucutre available for this data source in quality. This is creating the problem in quality. The extract strucutre which is created in your scenario would be available in temp folder and that will not be availbale for transport, so you need to have the datasource generated in quality and then you transport the active version to quality so that it will be available with the changes as that of development.
    Regards
    Vijay

  • APO DP - loading InfoCube data into a planning area

    I am using APO DP V5.
    In my planning object structure I have planning characteristics, some of which have 'navigation attribute' characteristics. For example, my planning characteristic 'product' has navigation attribute 'product group'.
    In the standard transaction to load data from an InfoCube to a planning area, can the characteristics used in this transaction be 'navigation attribute' characteristics? (or do they have to be plannng characteristics?)
    So, for example, could I load data at 'product group' level?
    Thanks,
    Bob Austin, Atos Origin

    am not sure if you can load data from cube to a navigation attribute (product group)
    you might need to maintain it in the Admin workbench and also assign values there. Am not sure if you will get the attribute characterisitics in yout TSCUBE loading screen

  • RE: Need help on cornering the APO BI issue relevant to Planning area

    HI Guys,
    Iam loading historical data from my infocube to APO planning area...
    MY Planning will be done in apo for weekly based..
    for thatmy client has configuired Fisc VArnt which is
    specific for 48 periods but not 52 periods...
    My clinet will be planning in week which will will be like it starts from 07.01.2010,
    14.01.2010,21.01.2010,31.01.2010..
    but for testing purpose we are taking thisfrom a flat file andloaded into the infocube..
    and created gen export datasource into scmapo system and loaded into cune from that cube iam feeding itto planning area..
    when I execute the transaction /n/sapapo/tscube tha data copied was successful,
    but the keyfigure amnt when I saw in the planning area Transaction /n.sapapo.sdp94which will be distributed across the week.
    lets say my keyfigure values is 100 in the infocube for the jan month and week 1
    and the value which i CAN SEE IN THE PLANNING AREA IS week1 25 week 2 25 week 3 25 and week 4 25
    which will be 100 total for amonth,
    but it should not be like that 100 should go into a particular week and should display 100 for that particular week..
    I have calmonth calday, fiscper(posting period) which we have maintained in ob29 as 48 periods
    when i derived calweek in the transformation iam getting 48 weeks but
    when i try to load it to planning area iAM GETTING AN ERROR LIKECOMBINATION IS INCONSISTENT..with the calmonth
    CODE WHICH i HAVE DERIVED CALWEEK FROM CALDAY:
    DATA: lv_year(4) TYPE c,
    lv_month(2) TYPE c,
    lv_day(2) TYPE c,
    lv_result(6) TYPE c,
    v_poper TYPE poper.
    lv_year = SOURCE_FIELDS-calday+0(4).
    lv_month = SOURCE_FIELDS-calday+4(2).
    lv_day = SOURCE_FIELDS-calday+6(2).
    SELECT SINGLE poper FROM t009b INTO v_poper WHERE bumon = lv_month
    AND butag = lv_day.
    IF sy-subrc = 0.
    CONCATENATE lv_year v_poper+1(2) INTO lv_result.
    RESULT = lv_result.
    CLEAR: lv_result.
    ENDIF.
    gURUS CAN ANY BODY THROW SOME LIGHT ON THIS.. iWILL BE HIGHLY OBLIGED
    when i load the data from infocube to planning arae using /SAPAPO/TSCCUBE, the copy was succeeful.. But the issue is the keyfigure value is dis aggregating..
    For ex If my sales hostory for week 1 001.2010 and for calmonth 01.2010 is 100, but it is disaggegating the values for whole month 4 weeks as 25,25,25,25 but it needs to b written as 100 for week 1.rather it should be aggregated on the highlevel as 100 for week 1.
    Do I need to check any Charecterstics combination but all are consistent....
    Even the periodicities in the planning area and infocube are consistent , since i am able to copy in to planning area..
    I dont have calweek in my infocube i have derived calweek with logic provided in my earlier thread, but as of now iam going with calyear and calmonth, fisper3 (postig period), since 48 posting periods are maintained in the 0b29 or t009b tables.
    whether I need to implement any oss notes for this, If I include calweek and calmonth and try to copy in to planning area Iam getting an error periodicities are not matching  SAP Note 1408753 - /sapapo/tscube wrong error message
    /SAPAPO/TSM 232
    Regards
    Balaram

    thnx for replying the thread..
    Where do I maintain this the PLANNING OBJECT STRUCTURE(STOR BKT Profile) and Planning Book-data View(Time Bkt Profile) with the time periodicities for Fisc Varnt and Posting period specificu2026
    Can you pls elaborate on this Iam new to APO BW implementation part, Sine this is a burning issue
    Actually what seeting I need to do there, I think infocube structures are good and copying data in to planning area..
    I have calmonth and Fiscper3 for 48 periods in my infocube
    also whether I need to maintain any calweek in the PLANNING OBJECT STRUCTURE(STOR BKT Profile) and Planning Book-data View(Time Bkt Profile)
    when I checked in planning book keyfigure overview it is maintained there for 48 periods
    For Planning Book-data View(Time Bkt Profile)  how can we achieve this...
    If you could throw some light more on this I have my APO Counter part I will ask him to change the master planning structure accordingly
    Regards
    Ram

Maybe you are looking for