RSA1 Cube statusoverview instead of DSO

In the RSA1 administration menu there is an option to check the DSO Loads statusses. Is there also an option to directly check all Cubes involved?
Kind regards,
Thanks in advance.

Hi,
There are multiple threads on SDN for the same topic, some of them are:
[InfoCube vs DSO;
[diff between infocube and dso;
Please search before posting the questions.
Regards,
Arminder Singh

Similar Messages

  • What is the effect of INDEXING on Cube? and on DSO?

    Hi,
    I am trying to figure out the effect of indexing on a Cube and DSO.
    1. My findings so far indicates that Indexing is a database concept but so is partition but partitioning applies in the case of a Cube. Right?
    2.What is the effect of Indexing on a cube? How is it related to the "database concept that I keep reading?
    How is it implement in BI?
    3.What is the effect of Indexing on a DSO? How is it related to the "database concept that I keep reading?
    How is it implement in BI?
    Thanks

    Hi,
    thanks for the detailed information. I will appreciate some help on these follow ups on your posting:
    On your answer to 1:
    I read that there are 2 partitions, logical and physical. It is the physical partitioning which takes place at the database level, so can I assume that when you said u201CEven DSO can be partitionedu201D you meant the physical partition? And that logical partitioning is not possible in DSOs?
    On your answer to 2:
    So in the above you suggested that indexes apply not only to cubes but also to DSO with the statement that u201C..secondary indexes can be built on DSO u2026u201D
    If so, how come there is not feature to drop and rebuild indexes for data load to DSOs?
    I will appreciate a hint on the difference between u201Cindexesu201D and u201Csecondary indexesu201D. I am ok with what u201Cindexesu201D is, is the use of u201Csecondaryu201D that threw me off.//
    On your answer to 3:
    i.
    Based on the link, there does not appear to be a simply way to drop and rebuild indexes in DSOs. It appears an ABAPer needs to be involved. Isnu2019t there a simply way as in cubes, to click on buttons for the index implementation for DSOs?
    ii.
    It suggests one must choose between B-tree and Bit map indexes, right? Is this the function of the BI consultant or as DBA? If the BI consultant should be able to do this, where is it implemented in BI?
    Can you explain what you meant by u201Ccardinality of a column is lowu201D?
    iii.
    I know how in rsa1, a Cube can be partitioned on 0calmonth and 0fisper. If this is done at the database level, then can it be done also for DSO? If so, where in BI is it done for DSOs?
    You also noted that SAP DB does not support table partitioning but Oracle does.
    I know in my environment, the BASIS consultants are always talking about Oracle so I am assuming we have Oracle and not SAP DB.
    I never heard of SAP DB, is it a substitute for Oracle in SAP BI environments?
    Is it a DB that comes directly from SAP; if so, why do companies not keep the total SAP package but prefer to use Oracle?  Is it because the SAP DB does not support table partitioning?
    Thanks

  • New field added to cube, delta DTP from DSO to cube is failing

    Dear all,
    Scenerio in BI 7.0 is;
    data source -delta IPs-> DSO -
    delta DTP---> Cube.
    Data load using proces chain daily.
    We added new field to Cube --> transformation from DSO to cube is active, transport was successful.
    Now, delta from DSO to cube is failing.
    Error is: Dereferencing of the NULL reference,
    Error while extracting from source <DSO name>
    Inconsistent input parameter (parameter: Fieldname, value DATAPAKID)
    my conclusion, system is unable to load delta due to new field. And it wants us to initialize it again ( am i right ?)
    Do I have only one choice of deleting data from cube & perform init dtp again ? or any other way ?
    Thanks in advance!
    Regards,
    Akshay Harshe

    Hi Durgesh / Murli,
    Thanks for quick response.
    @ durgesh: we have mapped existing DSO field to a new field in cube. So yes in DTP I can see the field in filter. So I have to do re-init.
    @ Murli: everything is active.
    Actully there are further complications as the cube has many more sources, so wanted to avoid seletive deletion.
    Regards,
    Akshay

  • RSA1 cube role

    I need to give a user a new role so that they can just view the contents of a specific cube within RSA1.  Say the cube is called ztest what entries do I need to enter in the following within PFCG so they can just see this cube
    Cross-application Authorization Objects
    Basis: Administration
    Basis - Central Functions
    Business Information Warehouse
    SAP Business Information Warehouse - Reporting
    regards
    Dom

    HI Domnic,
                 Try to give authorization for TX -listcube.
    Thanks,
    Vijay.
    Edited by: vijay m on Feb 10, 2009 4:54 PM

  • RSA1 - Cube Data

    Is there any function module which can be used to get the details of the Cube maintained in RSA1 in APO. The cube gets the data populated from BW.

    Hi Padmanabhan,
    Please try these function modules
    /FRE/FU_BW_LAST_EXTRACTION_GET
    OBJ2_VARALV_EXTRACT_SAVE
    /SAPTRX/BWEX (function group)
    ALV_EXTRACT_LOAD
    REUSE_ALV_EXTRACT_LOAD
    REUSE_ALV_EXTRACT_SAVE_LEVELS
    Please let me know whether your issue is resolved
    Regards
    R. Senthil Mareeswaran.

  • Data is not coming correctly through delta from DSO to Cube

    HI All,
    When there is change in Org. Unit, Job or any characteristic of employee then the the delta of data source picks that change records.
    I have used DSO before Cube. Now in DSO active table have one record against the Employee and same Schedule hours as ECC have.
    But while loading to the cube, this data of schedule hours will get added to existing schedule hours  of Cube , which makes the schedule hours double for that Employee. For exmaple.
    DSO Data.
    Emp NO.     Date Org. Unit     Sch. Hrs
    1                    ABC                              16
    CUBE Data
    Emp NO.     Org. Unit     Sch. Hrs
    1                    ABC               16
    Now Employee Org. Unit get changed to XYZ.
    DSO Data.
    Emp NO.     Org. Unit     Sch. Hrs
    1                    XYZ               16
    CUBE Data
    Emp NO.     Org. Unit     Sch. Hrs
    1                    ABC               16
    1                    XYZ               16
    I need data as it is in the DSO.
    Please help, thanks in advance.
    Arvind

    Hi Arvind,
    DSO and Infocube works different for char values......you can define Key Fields in DSO...
    whereas in Infocube each Characteristic value work as Key Field.....that's the reason if there is any change in any characteristic value ...
    DSO Active data will not change if Key field has not change..... the newly updated Data field value will be available in DSO Active Table.....whereas in Infocube... it will create a new entry......
    In order to get the updated value each time you need to explore some other options .....like complete deletion of Info cube data and full load everyday.......
    Other options depends on design and data volume and other parameters.
    Hope this helps........
    Regards
    Mayank

  • Data model that takes part of data to DSO and part to cube

    Hello experts,
    I have a question about the data model I am currently working on. This model is about a reservation agency, where a customer contacts the agency through a sales representative and makes a reservation for a certain event. He later has the option of confirming or cancelling the reservation. We will be receiving data for future reservations every day, with the updated status of future and recently past reservations, sales representative in charge of the reservations, income that will be perceived from each reservation (if the reservation is confirmed) or that was perceived if the event already took place, and more.
    I have found the need to use a DSO to update the reservation status of incoming reservations, as well as other fields. I am loading from a flat file with 20 fields, and the destination is a cube, after the data gets cleansed in the DSO, where there are 2 key fields (reservation code and reservation date).  However, not all the remaining data from the flat file needs to get updated. Out of the other 18 fields, only around 9 may change on an update, while the others will never change (for example, the customer name, confirmation date and event name will not change, while the sales representative, reservation status and income perceived can change from one day to another).
    I was designing a model in which the DSO contained all 18 non-key fields as data fields, so that everything got updated and passed to the cube, but then I thought that perhaps since many fields will not change, it would be better if the DSO only updated those fields that could actually change, while the other fields "jumped" directly to the cube. After the DSO updates the changing fields, the others will already be in the cube. This way, the DSO would receive and have to process less amount of data, making the DSO tables smaller and the process more efficient. But I cannot find a proper way to do this from a single data source, since the data has to pass from the data source to the DSO and then through a transformation from the DSO to the cube.
    Is there a way to accomplish this, or should I just have all fields go through the DSO and from there to the cube?

    Hi, thanks for the suggestion. The reservation code, as of right now, is not master data, and it is unique almost always, except for rare occurences where the same customer makes several reservations, each one for a different date (that is why for the DSO I am using as key fields both reservation date and reservation code). Because of that and the fact that there are lots of reservations made each day, we had defined that object without master data (since the master data table would be huge in a short amount of time). Perhaps I should change it to become master data? Or what should we do?

  • Regarding Info cube and DSO

    Hi,
    Generally reporting will be done on Info Cube rather than DSO.
    Suppose If we assign the same data source to Info Cube and DSO then both contains the same data.
    Info cube have additive and aggregated functionality where DSO have the overwrite functionality .
    Are we using cube for this functionality only ?
    What about the Dimensions in Cube how they differ from data fields and key fields in DSO when we are developing same Bex Report on both ?
    Please advice me .
    Thanks in advance.
    Thanks & Regards,
    Ramnaresh.p

    It is hard to compare Cube and DSO.
    Both thier own usage.
    1. InfoCube is always additive, while DSO supports overwrite functionality.
    2. In InfoCube, combination of all the characteristic value is a Key in the Fact Table, while in ODS, you can specify your own Key Fields based on which you want to generate unique record in the DSO.
    3. DSO supports many delta modes like D, R, N, X, after image, before image, while cube does not support all the modes. You can not delete the record based on the key from the cube by just loading the data. While DSO automaitcally deletes the record from active table and generates the reverse entry for Cube.
    4. DSO is a flat structure and therefore, it is used to store information at detail level, while cube is used to store information at aggregated level.
    So both the structures are very much different from each other. One can replace other at some places, but both the objects have thier own functionality.
    - Danny

  • I am planning to create custom defined  DSO Object & Info cube

    Hi ,
                     i am planning to create custom defined  DSO Object & Info cube.what ratio i can calculate what is the keyfields & what are the data fields in DSO.How can i calculate.
                     2. how can i create  suitable dimensions, suitable characterstics  for dimensions.what ratio i can decide.
    Thanks,
    chandu.

    Hi Diego Garu,
                               Thanks for your fast response.i
    VBELN     VBAP     2LIS_11_VAITM                                              0DOC_NUMBER
    POSNR     VBAP     2LIS_11_VAITM                                                0S_ORD_ITEM
    KUNNR     VBAK     2LIS_11_VAHDR                                                 0SOLD_TO
    VBELN     VBRP     2LIS_13_VDITM                                                    0BILL_NUM
    FKDAT     VBRK     2LIS_13_VDHDR                                                 0BILL_DATE
    INCO1     VBRK     2LIS_13_VDHDR(INCO1FieldNot Available in Data Source)     0INCOTERMS
    ZTERM     VBRK     2LIS_13_VDHDR(Payment terms field Not Available in Data Source)                                                                                0UCPYTERMS
    NETWR     VBRP     2LIS_13_VDITM                                                           0NETVAL_INV.
                                           here data is coming from the multible tables.that why i am planning to create custom defined data source based on view. here how can i calucate dso is suitable or cube is suitable.
    suppose dso is suitable how can i decide which on is the data field and which one is the key field.
                                        how can i decide how many dimensions are needed here.and which chara are suitable for that dimensions.
    Thanks ,
    chandu.

  • Creation of cube and dso in sap bi 7.0

    This topic has ben discussed here often. Kindly search the forums before posting a new thread. Unfortunately I would have to lock this thread.
    Hi
    Can any body tell me the how to create the cube as well an dso in bi 7.0 version thougt i how to create cube and ods in sap bi 3.5 version but i want to update my knowledge more what are main diif b/w 3.5 and 7.0 i had ask this question further also but i will be very greate full if any give me own hand made Answer rather than the for diff b/w 3.5 and biw 7.0
    regards
    gurkiran
    Edited by: Dinesh Lalchand on Feb 12, 2009 12:53 PM
    Edited by: Dinesh Lalchand on Feb 12, 2009 12:55 PM

    Hi ,
    I am trying to migrate the Transaction data sources before moving into the DEV from Sandbox.
    So i have to delete the existing TR/UR's and trying to create the transformation, but while creating transformation fields mapping is not happend automatically. (For manul field mapping the transaction data is the  time taking process)
    We are using the spk 8 . Please let me know any one has faced the above problem . 
    Regards
    Jose

  • Creation of DSO for a standard cube

    Hi All,
    I am new in BI modeling area & looking for help from you experts. What I have to do is , I need to create a DSO for a standard Cube  0PS_CS01.
    In this I am doing is to  take that data source which is in BI content for this Cube  and from there I am extracting data fields and keyfields. Once  this DSO is ready I will  create update rules from this standard Data source to this newly created to DSO and after that I will create Update rules for Cube based on this DSO.
    Can you guys help me to find the best way to collect  keyfields & data fields based on the data source. Also suggest me if  it is the right way to design data model fot the same as I desribed above.
    All helpful replies will be appriciated with points.
    Thanks a lot in Advance.
    Sincerely,
    John

    1>double  click on cube(0PS_CS01) --copy all characteristics(these will be key fields in DSO).
    DSO only take 16 key fields so rest will go under data fields.most unique 16 fields will go under key fields.
    2.all the keyfigues will go under data fields of DSO.
    change settings and activate DSO.create transformations between (DSO and DS) and ( DSO and CUBE);
    Hope this helps.
    if you need further info just check these two 0FIAP_C03(cube) and 0FIAP_O03(DSO).compare the characteristics and key figure with the key fields and data fields respectively.you will get good idea.
    in the above cube the data flow is like DS-DSO-CUBE

  • Transformations not copied from cube flow to dso

    hi
    i have 3.x flow from ds to cube. i want use dso in the middle before load to cube.
    i migrated flow 3.x to 7.x?
    now i created dso with same cube structure?
    i am trying copy of cube transfromations ( which migrated 3.x to 7.x flow) to dso? most of the fields not mapping? we can do manual this is no issue, my doubt is why filed mapping is not done when i copy transformations?

    If you are mapping from a DSO to cube then you will not get overwrite option.
    In cube only data aggregation is possible.

  • Do you see anything wrong with this GL mapping in a cube?

    hi,
    Can you verify if something is wring with this mapping:
    In model 1:
    0fi_gl_4 --> 0figl_O02 --> 0figl_C02
    the standard cube shows the following key figures and I mapped them to the following fields from the datasource:
    DMSOL(in datasource) to 0Debit_LC (in DSO) to 0Debit (in cube)
    DMHAB(in datasource) to 0Credit_LC (in DSO) to 0Credit (in cube)
    In model 2:
    0fi_gl_6 --> 0figl_C01
    the standard cube shows the following key figures and I mapped them to the following fields from the datasource:
    UMSOL(in datasource) to 0Debit (in cube)
    UMHAB(in datasource) to 0Credit (in cube)
    1. Do you see anything wrong with this mapping? And what is the solution?
    2. What is the problem with the mapping if both cubes sit under the same multiprovider, since 0debit and 0credit are from differnt souce fields?
    thnx

    Hi,
    Thanks for the explanation.
    on #1,
    if my goal is to provide a sort of trial balance, credits and debits as two columns, will it be ok to assume that it should be Local Currency or could it be Document Currency, Local Currency 2 or Local Currency 3? If the latter, then how do I decide?
    on #2,
    i. why will there be difficulty creating a union between the two?
    ii. will there also be an inconsistency if I created 2 separate multiproviders for each model, and create a jump query from the 0fi gl_6 model to 0fi gl_c4 model. If not, why not in this case?
    iii. So, is there any advantage in placing a multiprovider on each of these models separately? i.e. instead of directly reporting off the cubes separately?
    3. I am wondering why in the case of gl_6 (the totals) help.sap recomended to go directly from the datasource to the cube i.e. without DSO in between.
    If without the DSO step is an advantage, what are the advantages?
    And why then do we use DSOs in most model instead? I will appreciate any reasons you may think of..
    Thnx.
    Edited by: AmandaBaah on Jul 28, 2009 11:07 AM
    Edited by: AmandaBaah on Jul 29, 2009 2:25 AM

  • Lookup DSO in Start Routine and End Routine to update date

    Hi Gurus,
    I need a Start Routine & help, while loading the data from DSO to Info Cube.
    In the DSO ( /BIC/AZSD_O0300 ) & in the InfoCube I have a common fields are
    0DOC_NUMBER ( DOCUMENT NUMBER), 0S_ORD_ITEM ( Sales Document Item Number ), 0SCHED_LINE ( Schedule Line ) and 0MATAV_DATE ( Material Available Date ).
    Data in DSO is as follows
    0DOCNUMBER_ 0SORD_ITEM_ 0SCHEDLINE_ 0MATAVDATE_
    1001 10 1 05/21/2011
    1001 20 2 05/26/2011
    1001 10 2 05/22/2011
    1001 20 1 05/18/2011
    1002 20 1 05/20/2011
    1002 10 1 05/24/2011
    1002 10 2 05/28/2011
    Data should load in the InfoCube as below.
    0DOCNUMBER    0S_ORD_ITEM     0SCHED_LINE    0MATAV_DATE
    1001          10    1   05/21/2011
    1001          20    2   05/18/2011
    1001          10    2   05/21/2011
    1001          20     1  05/18/2011
    1002         20     1  05/20/2011
    1002         10    1   05/24/2011
    1002         10     2   05/24/2011  
    When ever schedule line is 2 for the record we have to check both Document number and 0S_ORD_ITEM are same change the Material moving date to the date schedule line is 1 for the same document and same order item number.
    Your help is really appreciated.
    Thanks
    Ganesh Reddy.

    Thanks Parth and Raj. I have written code almost as you suggested just little changes.
    Start Routine
    TABLES:/BIC/AZSD_O0300.
    DATA:ITAB_DSO LIKE /BIC/AZSD_O0300.
    TYPES: BEGIN OF ITABTYPE,
    DOC_NUM TYPE /BI0/OIDOC_NUMBER,
    SAL_DOC_NUM TYPE /BI0/OIS_ORD_ITEM,
    SCHED_NUM TYPE /BI0/OISCHED_LINE,
    MAT_AV_DATE TYPE /BI0/OIMATAV_DATE,
    END OF ITABTYPE.
    DATA : ITAB TYPE STANDARD TABLE OF ITABTYPE
           WITH HEADER LINE
           WITH NON-UNIQUE DEFAULT KEY INITIAL SIZE 0.
    data: wa_itab type ITABTYPE.
    SELECT DOC_NUMBER S_ORD_ITEM SCHED_LINE MATAV_DATE
      FROM /BIC/AZSD_O0300
      INTO TABLE ITAB
      WHERE SCHED_LINE = '1'.
    Field Routine.
    if SOURCE_FIELDS-S_ORD_ITEM <> '1'.
          read table itab into wa_itab with key    doc_num =
          SOURCE_FIELDS-DOC_NUMBER
                                             SAL_DOC_NUM =
                                             SOURCE_FIELDS-S_ORD_ITEM.
          if sy-subrc = 0.
          RESULT = wa_itab-MAT_AV_DATE.
        endif.
      endif.
    I haven't check how the data is coming. But for now I am closing this issue.
    Thanks Again
    Ganesh Reddy.

  • To Compress cube

    HI Experts,
    I am begineer to SAP BI .When I executed a outbound job through process chain at background,but got struck in a step due to SQL statement error.The DBA team informed to compress this cube and  It has 17M+ rows in the F fact table and they may also be able to re-create the F fact table as partitioned and with the new index structure for ~P indexu2026 meaning with time dimension in the 1st column as opposed to the packet dimension.
    Cube has last 3 years data and need to do in slots .Can you tell me procedure to do it.

    There are 2 different solution you mentioned to improve the performance.
    For the compress, you can do it at
    1) RSA1 -> Cube -> Management
    2) Find the process type and add it into your process chain.
    Compress will change the request id to '0' and aggregate the records with same char.
    After you do it, the total number of records will decrease. It depends on the data in cube itself.
    For the Partition, you can find it at RSA1 -> Cube and find it from the menu: Extras -> DB performance -> Partition.
    It devide the data into groups through some particular char. value.
    And you have chance to do the partition in physical level which may not supported by all DB. Check some technical notes.

Maybe you are looking for