Do we have to load data in BPS cube before entering values ?

Hello,
Can someone explain me the flow of data in BPS. Like I have created a copy of a regular cube and will be working on that making it into a transactional cube (as per help.sap). So do i need to load the same data in tran. cube from R/3 as load in regular cube. or do I just have to create the planning folders and layouts and let the user enter their planning values?
Do we have to send the data that users input back to R/3 ?
In short if someone can explain me the flow of data it would be great.
Thanks a lot in advance.

Hi,
If you are working on funds management then I think there's an extractor for it and clients usually plan and retract the  data back to SAP. Again, loading of data is not just for comparing plan and actuals the client might be reporting on the data.
BPS is a part of BI, the main area that is normally implemented first is the BW part which is for reporting and then go for BPS.
What I am not sure is these are very generic questions can you be more specific as to what's the problem you are facing. All the points mentioned by you will be handled/answered during a blue printing session.
thanks

Similar Messages

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Thanks,
    Gopi R

    Hi,
    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).
    Regards,
    Gregor

  • To load data from a cube in SCM(APO) system to a cube in BI system.

    Experts,
         Please let me know whether it is possible to load data from a cube in SCM(APO) system to a cube in BI system.If so explain the steps to perform.
    Thanks,
    Meera

    Hi,
    Think in this way,
    To load the data fro any source we need datasource Ok. You can genare Export data source for Cube in APO,  then use that datasource for BW extraction, try like this. I think it will work, in my case I'm directly loading data from APO to BW using the DataSource sthat are genaraed on Planning Area.
    Why you need to take data from APO cube?. Is there any condition for that?. If it is not mandatory, you can use the dame datasource and load the data to BW, if they have any conditions while loading the data from APO to APO cube, they you try check wherther it is possible in BW or not. If possible then you use DataSource and do the same calculation in BW directly.
    Thanks
    Reddy

  • Automatically trigger the event to load data from Planning cube to Standard Cube

    Hello,
    We have a below set up in our system..
    1. A Planning BEx query using which user makes certain entries and writes back data to the planning cube.
    2. An actual reporting cube which gets data from the planning cube above.
    Now, what we want to do is to automate the data load from Planning cube to Reporting cube.
    This involves 2 things..
    1. Change the setting " Change real time load behaviour " of the planing cube to Planning.
    2. Trigger the DTP which loads data from Planning cube to reporting cube.
    We want to automate the above two steps...
    I have tried few things to achieve the same..
    1. Created an event in SM64,
    2. In the Planning cube "Manage" Screen,  clicked on "Subsequent Processing" and provided the event details. ( Not sure if its the correct place to provide the event detail ).
    3. Wrote a ABAP program which changes the setting of the planning cube (  " Change real time load behaviour " to Loading )
    4. Created a process chain, where we have used the event as a start variant, used the ABAP program as a next step, DTP run as the last step..
    This i hoped, would trigger the event as soon as a new request comes and sits in planning cube, which in turn would trigger the process chain which would load the data from Planning cube to reporting cube..
    This is not working. I dont think the event is triggering, and even if does i am not sure whether it will start the process chain automatically. Any ideas please ?

    hi,
    try to do the transformation directly in the input cube by using CR of type exit, more details :
    http://help.sap.com/saphelp_nw70ehp2/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
    hope it helps.

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading Data from one Cube into another Cube

    Hi Guys,
    I am trying to load data from one cube A to another cube B. Cube A has data around 200,000 records. I generate export datasource on Cube A. Replicated the datasource and created InfoSource and activated it.
    I created update rules for Cube B selecting Source as Cube A. I do have a start routine to duplicate records in Cube A. Now when I schedule load,
    It stops at Processing Datapacket and says no data. Is there something wrong with the update routine or is there any other way to load form cube to cube in a simpler way?
    Thanks in advance

    This is the start routine to duplicate records in two currencies.
    DATA: datew   TYPE /bi0/oidateto,
          datew2  TYPE rsgeneral-chavl,
          fweek   TYPE rsgeneral-chavl,
          prodhier TYPE /bi0/oiprod_hier,
          market  TYPE /bic/oima_seg,
          segment TYPE /bic/oizsegment.
    DATA: BEGIN OF S_DATA_PACK OCCURS 0.
            INCLUDE STRUCTURE /BIC/CS8ZSDREV.
    DATA: END OF S_DATA_PACK.
    S_DATA_PACK[] = DATA_PACKAGE[].
      REFRESH DATA_PACKAGE.
      LOOP AT S_DATA_PACK.
        move-corresponding s_data_pack to DATA_PACKAGE.
        if DATA_PACKAGE-loc_currcy = 'EUR'.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalloc.
          DATA_PACKAGE-CURRENCY = 'EUR'.
          APPEND DATA_PACKAGE.
        else.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
        endif.
      ENDLOOP.
    This is to load Quantity field
    RESULT = COMM_STRUCTURE-BILL_QTY.
    This is to load Value field
    RESULT = COMM_STRUCTURE-NETVAL_INV.
    UNIT = COMM_STRUCTURE-currency.

  • How to load data to a cube from multiple infosources ?

    Hi friends,
    How to load data to a cube from multiple infosources ? could u please answer this question .
    thanks in advance......

    Hi ,
    say for example you need to load data to 1 cube from 3 info sources:
    1) You need to create 3 update rules for the Cube.
    2) Each time you create the update rules. Mention the name the name of the Info source. and create update rules correspondingly.
    Regards
    satish
    Message was edited by:
            satish murthy

  • Loaded data amount into cube and data monitor amount

    Hi,
    when I load data into the cube the inserted data amount in the administrator section shows 650000 data sets. The monitor of that request shows a lot of data packages. When I sum the data packages, the sum is about 700000 data sets.
    Where is the difference coming from?
    Thanks!

    Hi ,
       If it is a full load to the cube , all the records are updated in it since in a cube data can be overwritten.
       If it is a delta load and u want to see why the difference occurs between the records transferred and added in cube ,
       u can go to the manage tab in dso , go to the contents tab ,there click change log button at the below , check the number of entries in that table , the number of entries are the added records in cube since only these records are the new records other records with the same key are already present in the cube.

  • Dump while loading data from 1 cube to other cube

    Hi,
    We created a direct mapping between 2 cubes ( no start/end routines, formula exist in the mapping). Strange part is while I am loading data from one cube to other cube all data packages executing successfully only last data package giving dump with ITAB_DUPLICATE_KEY. This is in our QA system. But it is working/loading correctly in DEV.
    Please help. Thanks in advance for your support.
    Regards,
    Sai

    Hi Sai,
    Please refer to below blog for ITAB Duplicate Key error.
    ITAB_DUPLICATE_KEY Dump error during Master Data loading
    This can occur when a BW client has many different source systems, unused objects still remains, creating a duplicate entry in RSISOSMAP table.
    Regards
    Shabnam

  • Could not see loaded data in the Cube

    Hi Gurus,
    Pl. help me in this: I am not able to get teh data loaded into the cube. In PSA it is loaded right. after that I created Transformation on InfoCube and I got Proposal Generated. Then I created DTP and executed, and monitor shows green. but when I right click on the cube and look for data(also maintain master data on InfoObject), I see only the headers of the table but no data is displayed.
    I hope you have similar situations. Thanking you in advance.

    Check if the request is reportable, sometimes the load could be fine, but the request may not be reportable. Make sure in your start routine you don't have a routine which will filter or delete the data out from the cube before it gets update.
    thanks.
    Wond

  • Process Chain which loads data to a cube

    Hi All
    Please help me create a process chain in BI 7.0. I have to load the data in to a cube and before that i need to delete the existing request in the Cube. I think i should use " Delete overlapping requests from Infocube".  In the maintain variant for this process type what object types i need to add. Do i need to add both Execute infopackage and Data transfer process objects.
    Regards
    Naga

    Hi Ravi
    I am loading the data from PSA to Cube using DTP. Actually my data source is export datasource (8ZHYP- prefix 8). So accroding to your answer i should use DTP object type in the process type delete overlapping request. But when i create a variant for that process type it is getting Indexs > Create Index > Data Transfer Process > Request to delete data in the infocube.
    I just want to delete the data in the infocube before i load the data. So can i delete all the remaining processes.
    Regards
    Naga

  • Loading data to a cube from a DSO and a Flat file

    Hi All,
    I have an Infocube with fields
    product
    plant
    customer
    quantity
    country
    I am trying to load data into it , few fields from a DSO and others from a flat file
    e.g plant and country from a dso
    and product, customer and quantity from a flat file.
    I have created 2 transformations,
    one from the DSO -> cube (which works)
    another transfromation does not get activated -- it gives an error msg saying (no source fields are assigned to the rule)
    Is it possible to make a load of this sort? If so, then why is the transformation giving that error?
    Please do help!
    Thanks and Regards,
    Radhika

    Dear Friend,
    Not sure what have you done, it should be like this
      datasouce---DSO1--
    Cube1
    Flat2--DSO(optional)-Cube2
    then you should build a multiprovider on top of these cubes (cube1 and cube2 ) and then create Query.
    Please check if this is something you have done.
    Hope this helps.
    Thanks
    sukhi

  • Loading data to a cube

    Hi All,
    We have created one cube and loaded data successfully.
    But we have one dimension named periodType: the members of the period are:Annual, Quarterly and Monthly.
    Client asked to make Monthly as + and remaining ~.
    For that we have created one view in that we have added one column aggr_cons.
    I have defined it as +* for monthly and the rest as *~*. I’m using that column as the consolidation operator.
    But the data load is laoding only a few records.
    Actually client asked me to create an Attribute dimension but it is not possible in EIS.
    We are using EIS 7.1.2 and SQL server 2005.
    For PeriodType dimension we have written the query like
    select distinct PeriodType,aggr_cons
    from ClaimsData_2
    PeriodType is a column in table and it contains Annual,Quarterly,Monthly.
    Please let me know any ideas to do this.
    Only thing is i have to make Monthly as +* and remaining *~*.
    Thanks,
    prathap

    Hi Pratap,
    Ffirst do following changes in datasource (i.e. SQL Server, Oracle whatevere you are using): -
    1- Create a new table say 'Population' and add add two column say id and population like 100,200....etc. Define id as primary key.
    2- Now assuming that you have a SQL table called 'Product' so add column called 'Attribute' create it relationship with column 'id' of table 'Population' through foriegn key.
    Now do follolwing changes in OLAP metadata & metaoutline:-
    1- Suppose you have Product dimension and enable one of its column as attribute.
    2- Ok now open metaoutline and expand Product dimension in left panel. Now it will show attributes that you associated.
    3- Select an attribute and drag to right panel. It will create a attribute dimension automatically.
    Hope it answers you.
    Atul K,

  • Loading data from Info cube to PA

    Hi All,
    I am trying to load the CVC from Infocube to Planning Area in Development environment.
    I am getting error message for mismatch of periodicity of infocube and Planning Area.
    how to resolve this. Where do we define the periodicty of infocube ? I don't think so ?
    Please help me
    Thanks a lot
    Prabhat Sahay

    Prabhat,
    I understand that you already have the data in the infocube and want to load it into the planning area.
    However it is important to ensure that the data in the cube and the planning area have the same periodicity failing which we will run into such issues.
    So pls go back to the step of defining the infocube.
    In RSA1 if you double click on the infocube name which you are using as source, you can see the details in the right. In that, look for the dimensions section, under that look for the time section. If that contains a calweek or a calmonth and you are using a fiscal variant, this issue occurs.
    You have to drop your infocube data, redefine the cube to use FISCAL PERIOD and FISCAL VARNT instead of calweek or calmonth and then load the data back again. You need to ensure you populate the fields FISCAL PERIOD and FISCAL VARNT. You might want to do that with routines in your update rules or your infosource communication structure. REmember to activate the entire structure from cube to data source.
    Once you load the data into the cube with the new setting, and then try to copy the data into planning area, it will be successful.
    To answer your other question. Yes Time Characteristics is necessary for loading data between cube and planning area.
    Hope this helps.
    Thanks
    Mani Suresh

  • Loading data into Fact/Cube with surrogate keys from SCD2

    We have created 2 dimensions, CUSTOMER & PRODUCT with surrogate keys to reflect SCD Type 2.
    We now have the transactional data that we need to load.
    The data has a customer id that relates to the natural key of the customer dimension and a product id that relates to the natural key of the product dimension.
    Can anyone point us in the direction of some documentation that explains the steps necessary to populate our fact table with the appropriate surrgoate key?
    We assume that we need to have an lookup table between the current version of the customer and the incoming transaction data - but not sure how to go about this.
    Thanks in advance for your help.
    Laura

    Hi Laura
    There is another way to handling SCD and changing Facts. This is to use a different table for the history. Let me explain.
    The standard approach has these three steps:
    1. Determine if a change has occurred
    2. End Date the existing record
    3. Insert a new record into the same table with a new Start Date and dummy End Date, using a new surrogate key
    The modified approach also has three steps:
    1. Determine if a change has occurred
    2. Copy the existing record to a history table, setting the appropriate End Date en route
    3. Update the existing record with the changed information giving the record a new Start Date, but retaining the original surrogate key
    What else do you need to do?
    The current table can use the surrogate key as the primary key with the natural key being being a unique key. The history table has the surrogate key and the end date in the primary key, with a unique key on the natural key and the end date. For end user queries which in more than 90% of the time go against current data this method is much faster because only current records are in the main table and no filters are needed on dates. If a user wants to query history and current combined then a view which uses a union of the main and historical data can be used. One more thing to note is that if you adopt this approach for your dimension tables then they always keep the same surrogate key for the index. This means that if you follow a strict Kimball approach to make the primary key of the fact table be a composite key made up of the foreign keys from each dimension, you NEVER have to rework this primary key. It always points to the correct dimension, thereby eliminating the need for a surrogate key on the fact table!
    I am using this technique to great effect in my current contract and performance is excellent. The splitter at the end of the map splits the data into three sets. Set one is for an insert into the main table when there is no match on the natural key. Set two is when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the current row needs to be copied into history, setting the End Date to the system date en route. Set three is also when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the main record is simply updated with the Start Date being reset to the system date.
    By the way, I intend to put a white paper together on this approach if anyone is interested.
    Hope this helps
    Regards
    Michael

Maybe you are looking for

  • Mxmlc / compc: automatically link in all source files?

    In the standard configuration, compc and mxmlc only link in classes actually used by your project. We're using code generation to get ActionScript classes from our Java classes for AMF serialization. We deal with most of these generically, so we don'

  • [SOLVED]"The requested URL returned error: 407" while using pacman

    pacman worked fine until now. When I try to upgrade with pacman -Syu it returns what is written below even though internet works fine without problems(I'm posting from the same system using firefox) How can I get pacman working again? :: Synchronizin

  • White Balance forgets setting

    Hi, I have this wacky issue where I'll edit the white balance in a photo and the preview will update like it should. Then sometimes I come back to the photo, and I can see in the thumbnail preview thing that it's the way I remember doing it, and then

  • Layer Edit Menu Froze Photoshop

    Ok, something really weird happened to my photoshop. I was busy designing an advert for work, happily adding layer styles to text as per usual and when I was busy selecting a colour for the gradient overlay my colour picker menu froze. I can still se

  • Fingerprin​t reader not detected

    I have a HP pavilion dv6 1152tx laptop. I have upgraded the system form pre-installed vista to win7 and installed the driver for fingerprint reader that came with the upgrade disk but it says reader not detected. I also tried installing the driver pr