Loading data into two cubes

We want to set up a delta refresh from R/3 data that will pull data into two cubes.  One cube I want all the records to be loaded, in the second cube I want to filter the records that are loaded into that cube.  I can figure out how to load the data into two cubes, but can I place a filter on one of the loads.  (i.e. only load records of where type = 'X')
Thanks,
Chris

You can do that in the Update Rules to the second cube... In the Start Routine...
DELETE FROM DATA_PACKAGE
WHERE type EQ 'X'.
(Please, verify the right syntax)
Then with only one Infopackage, you can load both cube with different conditions.
Hope it helps.
Regards,
Luis

Similar Messages

  • Loading data into two files

    Hi
    I have written sql loader script for loading data into two tables.
    Script is working ..But output is not coming properly.
    I want to load data into first table which lines are having first char 'R'.
    In the second table I have to load data which are related from the first line. Incase
    first line data is not properly (means discarded) then related second table data
    will not be load.
    But I am getting both rows.Though the first table record is discarded. Please find below
    the output.
    Any other solution also ....Ok..external tables..Utl_file.
    LOAD DATA
    infile "inputFileForRmaReceiptAcknowledgement.dat"
    BADFILE 'inputFileForRmaReceiptAcknowledgement.bad'
    DISCARDFILE 'inputFileForRmaReceiptAcknowledgement.dsc'
    APPEND
    INTO TABLE XXGW_RMA_HEDR_RCPTACK_TAB
       WHEN (01)='R'
          (  LINE_TYPE              POSITION(1:1) "substr(:Line_Type, 1)",
             RMA_ORDER_NO           POSITION(2:16)  CHAR,
             ACKNOWLEDGEMENT_NO     POSITION(17:31) CHAR,
             ACKNOWLEDGEMENT_DATE   POSITION(32:45) 
    "to_date(substr(:acknowledgement_date,3),'YYMMDDHH24MISS')",
             DETAIL_LINE_COUNT      POSITION(46:51) INTEGER EXTERNAL,
             FLAG                   CHAR)
    INTO TABLE XXGW_RMA_RCPT_ACKLDGMNT_TAB
       WHEN (01) = 'D'
          (  LINE_TYPE              POSITION(1:1) "substr(:Line_Type, 1)",
             RMA_ORDER_NO           POSITION(2:16)    CHAR,
             RMA_ORDER_LINE         POSITION(17:19)   INTEGER EXTERNAL,
             SERIAL_NUMBER          POSITION(20:49)   CHAR,
             SKU                    POSITION(50:63)   CHAR,
             QUANTITY               POSITION(64:69)   INTEGER EXTERNAL,
             WAREHOUSE_CODE         POSITION(70:71)   CHAR,
             WAYBILL_NUMBER         POSITION(72:121)  CHAR,
             COURIER                POSITION(122:146) CHAR,
             RETURN_DEALER_FLAG     POSITION(147:156) CHAR)
    inputFileForRmaReceiptAcknowledgement.dat
    R12345678901    2345456789123200   21111228241113000002  --- discarded record
    D12345678901    00159123687402 45678925803   00000102name                                  
    D12345678901    00159143687402 45678925603   00000102name                                  
    T000004Regards
    Ar

    Pl post details of OS and database versions.
    Create a foreign key constraint between the detail table and the master table. If the row fails to load into the master table, then the detail table rows will fail the foreign key constraint and will not load.
    http://docs.oracle.com/cd/E11882_01/server.112/e25789/datainte.htm#CNCPT1649
    HTH
    Srini

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Unable to load data into Planning cube

    Hi,
    I am trying to load data into a planning cube using a DTP.
    But the system throws a message saying that, real time data loading cannot be performed to the planning cube.
    What would be the possible reason for the same?
    Thanks & Regards,
    Surjit P

    Hi Surjit,
    To load data into cube using DTP, it should be put in loading mode only.
    The real time cube behaviour is set to planning mode only when the data is input to the cube via the planning layouts / or through the Fileupload program.
    You can change the behaviour of the cube in RSA1 -> Right click on the cube -> Change Real-time load behaviour and select the first option(real time cube can be loaded with data; planning not allowed)
    Best Rgds
    Shyam
    Edited by: Syam K on Mar 14, 2008 7:57 AM
    Edited by: Syam K on Mar 14, 2008 7:59 AM

  • Loading data into Fact/Cube with surrogate keys from SCD2

    We have created 2 dimensions, CUSTOMER & PRODUCT with surrogate keys to reflect SCD Type 2.
    We now have the transactional data that we need to load.
    The data has a customer id that relates to the natural key of the customer dimension and a product id that relates to the natural key of the product dimension.
    Can anyone point us in the direction of some documentation that explains the steps necessary to populate our fact table with the appropriate surrgoate key?
    We assume that we need to have an lookup table between the current version of the customer and the incoming transaction data - but not sure how to go about this.
    Thanks in advance for your help.
    Laura

    Hi Laura
    There is another way to handling SCD and changing Facts. This is to use a different table for the history. Let me explain.
    The standard approach has these three steps:
    1. Determine if a change has occurred
    2. End Date the existing record
    3. Insert a new record into the same table with a new Start Date and dummy End Date, using a new surrogate key
    The modified approach also has three steps:
    1. Determine if a change has occurred
    2. Copy the existing record to a history table, setting the appropriate End Date en route
    3. Update the existing record with the changed information giving the record a new Start Date, but retaining the original surrogate key
    What else do you need to do?
    The current table can use the surrogate key as the primary key with the natural key being being a unique key. The history table has the surrogate key and the end date in the primary key, with a unique key on the natural key and the end date. For end user queries which in more than 90% of the time go against current data this method is much faster because only current records are in the main table and no filters are needed on dates. If a user wants to query history and current combined then a view which uses a union of the main and historical data can be used. One more thing to note is that if you adopt this approach for your dimension tables then they always keep the same surrogate key for the index. This means that if you follow a strict Kimball approach to make the primary key of the fact table be a composite key made up of the foreign keys from each dimension, you NEVER have to rework this primary key. It always points to the correct dimension, thereby eliminating the need for a surrogate key on the fact table!
    I am using this technique to great effect in my current contract and performance is excellent. The splitter at the end of the map splits the data into three sets. Set one is for an insert into the main table when there is no match on the natural key. Set two is when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the current row needs to be copied into history, setting the End Date to the system date en route. Set three is also when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the main record is simply updated with the Start Date being reset to the system date.
    By the way, I intend to put a white paper together on this approach if anyone is interested.
    Hope this helps
    Regards
    Michael

  • Error while loading data into the cube

    Hi,
    I loaded data on to PSA and when I am loading the data to the cube through DataTransferProcess, I get an error (red color).
    Through "Manage", I could see teh request in red. How do I get to knoe the exact error? Also what could be the possibel reason for this?
    Also can some one explain the Datatransfer process(not in process chain)?
    Regards,
    Sam

    Hi Sam
        after you load the data  through DTP(after click on execute button..) > just go to monitor screen.. in that press  the refresh button..> in that it self.. you can find the  logs..
       otherwise.. in the request  screen also..  beside of the request number... you can see the logs icon.. you can click on this..
    DTP  means..
    DTP-used for data transfer process from psa to data target..
    check thi link..for DTP:
    http://help.sap.com/saphelp_nw04s/helpdata/en/42/f98e07cc483255e10000000a1553f7/frameset.htm
    to load data in to the datatargets or infoproviders like DSO, cube....
    in your  case.. the problem may be.. check the date formats.. special cherectrs... and
    REGARDS
    @JAY

  • Unable to load data into Essbase cube using Essbase Studio

    Hi
    We are creating an essbase cube using Essbase Studio using flat files as data sources.
    We have taken different hierarchies into different flat files and created one fact file having dimension intersection along with data.
    We are able to create the cube and the hierarchy but not able to load any data.
    We are getting the following error
    Failed to deploy Essbase cube.
    Caused by: Unable to perform dataload from more than one flat file.
    Could anyone please help on this?

    Oh this was killing me, so I did this test in 11.1.1.3:
    1) Excel 2007 format -- no go, Essbase didn't see it
    2) Excel 2003 format, three sheets -- only the first sheet was read into an empty rule
    3) Excel 2003 format, one sheet -- the first sheet was read into an empty rule
    4) Excel 95 format, one sheet -- the first sheet was read into an empty rule
    The lesson?
    1) Excel 2007/2010 sheets don't work (no surprise there as the .xlsx format isn't supported).
    2) Excel 2003 and lower (hey, if you have Excel 4, I'll bet that works as well) work, but only the first sheet is recognized.
    Regards,
    Cameron Lackpour

  • Short Dump when loading Data into Cube UNCAUGHT_EXCEPTION(Pls Help)

    Dear Guru's
    i am trying to load data into a Cube from a ODS and the load gets dump...
    In Dump Analysis.....
    The exception 'ZCX_EDW_MAP_ROOT' was raised, but it was not caught anywhere         
    along   
    "UNCAUGHT_EXCEPTION" "ZCX_EDW_MAP_ROOT"                                       
    "ZCL_EDW_MAP_ROOT==============CP" or "ZCL_EDW_MAP_ROOT==============CM002"   
    "LOAD_MAPPING_INFOPROV"                                                                               
    So it would be great if some one can give me some suggestion to handle this issue
    Thanks in Adv
    Dev

    Dear Arun,
    i have checked the routine there exists some code
    Eg: data: vref_monitor type ref to zcl_edw_monitor,
    But to say fact i ddnot i understood it....
    Any suggestions dude...pls
    thanks in adv
    Dev

  • Error while loading data into cube

    hi BW gurus,
    when ever i am trying to load data into the cube from flat file after scheduling iam getting short dump in BW system. I checked it in st22 it is giving me a error as exception add_partition_failed. please help me to sort out this problem. If you know the error recovery please give me the answer in detail.
    I will assign points for good answers.

    This is what the note says:
    Symptom
    The process of loading transaction data fails because a new partition cannot be added to the F fact table.The loading process terminates with a short dump.
    Other terms
    RSDU_TABLE_ADD_PARTITION_ORA, RSDU_TABLE_ADD_PARTITION_FAILED, TABART_INCONSITENCY, TSORA, TAORA , CATALOG
    Reason and Prerequisites
    The possible causes are:
    SQL errors when creating the partition
    Inconsistencies in the Data Dictionary control tables TAORA and TSORA
    Solution
    BW 3.0A & BW 3.0B
    In the case of SQL errors:Analyze the SQL code in the system log or short dump and if possible, eliminate the cause. The cause is often a disk space problem or lock situations on the database catalog or, less frequently: the partitioning option in the ORACLE database is not installed.
    The most common cause of the problem is inconsistencies in the TAORA and TSORA tables. As of Support Package 14 for BW 3.0B/Support Package 8 for BW 3.1C, the TABART_INCONSITENCY exception is issued in this case. The reason is almost always missing entries in TSORA for the tablespaces of the DDIM, DFACT and DODS data classes.
    The TAORA table contains the assignment of data classes to data tablespaces and their attributes, for example:
    Tablespace Data class
    DDIM PSAPDIMD ........
    DFACT PSAPFACTD ........
    DODS PSAPODSD .......
    Foreach data tablespace, the TSORA table must contain an entry for the corresponding index tablespace, for example:
    TABSPACE INDSPACE
    PSAPDIMD PSAPDIMD
    PSAPFACTD PSAPFACTD
    PSAPODSD PSAPODSD
    In most cases, these entries are missing and have to be added. See also notes 502989 and 46272.

  • Getting an error while trying to load data into cube

    Hello,
    I am trying to load data into cube but it gives me following error:
    fiscal year variant not expected.
    Can someone tell what could be done on fiscal year
    Cheers
    Jim.

    Hi,
    May be you have not mapped anything to fiscal variant or not included fiscal variant in the cube but you have used fiscal period inside it.
    You should load fiscal variant if you wanna use fiscal period.
    Map fiscal variant to a constant K4 or from the source if present and then see the result.
    Generally it is K4 except for FI-CO cubes .
    In the transfer rules, select field '0fiscvarnt' 'Tp' click
    Enter K4 for Constant................Save and Activate
    Iam sure you'd fix the issue.
    Regards,
    Ray

  • Error encountered while loading data into an Essbase cube from oracle DB

    Hi All,
    I am trying to load data into an Essbase cube from Oracle db. Essbase is installed in an 64 bit HP-UX machine and oracle database on HP -UX 11.23 B server. While performing the above I am getting the below error messages:
    K/INFO - 1021013 - ODBC Layer Error: [08001] ==> [[DataDirect][ODBC Oracle Wire Protocol driver][Oracle]ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    HP-UX Error: 2: No such file or directory].
    OK/INFO - 1021014 - ODBC Layer Error: Native Error code [1034] .
    ERROR - 1021001 - Failed to Establish Connection With SQL Database Server. See log for more information.
    Can you please help me out in identifying the issue, as why it could have occured , because of network problem or is there anything related to databse?
    Regards,
    Priya

    The error is related to Oracle, I would check that it is all up and running
    ORA-01034: ORACLE not available
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading Data from one Cube into another Cube

    Hi Guys,
    I am trying to load data from one cube A to another cube B. Cube A has data around 200,000 records. I generate export datasource on Cube A. Replicated the datasource and created InfoSource and activated it.
    I created update rules for Cube B selecting Source as Cube A. I do have a start routine to duplicate records in Cube A. Now when I schedule load,
    It stops at Processing Datapacket and says no data. Is there something wrong with the update routine or is there any other way to load form cube to cube in a simpler way?
    Thanks in advance

    This is the start routine to duplicate records in two currencies.
    DATA: datew   TYPE /bi0/oidateto,
          datew2  TYPE rsgeneral-chavl,
          fweek   TYPE rsgeneral-chavl,
          prodhier TYPE /bi0/oiprod_hier,
          market  TYPE /bic/oima_seg,
          segment TYPE /bic/oizsegment.
    DATA: BEGIN OF S_DATA_PACK OCCURS 0.
            INCLUDE STRUCTURE /BIC/CS8ZSDREV.
    DATA: END OF S_DATA_PACK.
    S_DATA_PACK[] = DATA_PACKAGE[].
      REFRESH DATA_PACKAGE.
      LOOP AT S_DATA_PACK.
        move-corresponding s_data_pack to DATA_PACKAGE.
        if DATA_PACKAGE-loc_currcy = 'EUR'.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalloc.
          DATA_PACKAGE-CURRENCY = 'EUR'.
          APPEND DATA_PACKAGE.
        else.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
        endif.
      ENDLOOP.
    This is to load Quantity field
    RESULT = COMM_STRUCTURE-BILL_QTY.
    This is to load Value field
    RESULT = COMM_STRUCTURE-NETVAL_INV.
    UNIT = COMM_STRUCTURE-currency.

  • Optimize the data load process into BPC Cubes on BW

    Hello Gurus,
    We like to know how to optimize the data load process and our scenario for this is that we have ECC Classic Ledger,  and we are looking for the best way to load data into the BW Infocubes from an ECC source.
    To complement the question above, from what tables the data must be extracted and then parsed to BW so the consolidation it´s done ?  also, is there any other module that has to be considered from other modules like FI or EC-CS for this?
    Best Regards,
    Rodrigo

    Hi Rodrigo,
    Have you looked at the BW Business Content extractors available for the classic GL? If not, I suggest you take a look. BW business content provides all the business logic you will normally need to get data out of ECC and into BW for pretty much every ECC application component in existence: [http://help.sap.com/saphelp_nw70/helpdata/en/17/cdfb637ca5436fa07f1fdc0123aaf8/frameset.htm]
    Ethan

  • Loading data to a cube from a DSO and a Flat file

    Hi All,
    I have an Infocube with fields
    product
    plant
    customer
    quantity
    country
    I am trying to load data into it , few fields from a DSO and others from a flat file
    e.g plant and country from a dso
    and product, customer and quantity from a flat file.
    I have created 2 transformations,
    one from the DSO -> cube (which works)
    another transfromation does not get activated -- it gives an error msg saying (no source fields are assigned to the rule)
    Is it possible to make a load of this sort? If so, then why is the transformation giving that error?
    Please do help!
    Thanks and Regards,
    Radhika

    Dear Friend,
    Not sure what have you done, it should be like this
      datasouce---DSO1--
    Cube1
    Flat2--DSO(optional)-Cube2
    then you should build a multiprovider on top of these cubes (cube1 and cube2 ) and then create Query.
    Please check if this is something you have done.
    Hope this helps.
    Thanks
    sukhi

  • Automatically trigger the event to load data from Planning cube to Standard Cube

    Hello,
    We have a below set up in our system..
    1. A Planning BEx query using which user makes certain entries and writes back data to the planning cube.
    2. An actual reporting cube which gets data from the planning cube above.
    Now, what we want to do is to automate the data load from Planning cube to Reporting cube.
    This involves 2 things..
    1. Change the setting " Change real time load behaviour " of the planing cube to Planning.
    2. Trigger the DTP which loads data from Planning cube to reporting cube.
    We want to automate the above two steps...
    I have tried few things to achieve the same..
    1. Created an event in SM64,
    2. In the Planning cube "Manage" Screen,  clicked on "Subsequent Processing" and provided the event details. ( Not sure if its the correct place to provide the event detail ).
    3. Wrote a ABAP program which changes the setting of the planning cube (  " Change real time load behaviour " to Loading )
    4. Created a process chain, where we have used the event as a start variant, used the ABAP program as a next step, DTP run as the last step..
    This i hoped, would trigger the event as soon as a new request comes and sits in planning cube, which in turn would trigger the process chain which would load the data from Planning cube to reporting cube..
    This is not working. I dont think the event is triggering, and even if does i am not sure whether it will start the process chain automatically. Any ideas please ?

    hi,
    try to do the transformation directly in the input cube by using CR of type exit, more details :
    http://help.sap.com/saphelp_nw70ehp2/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
    hope it helps.

Maybe you are looking for

  • Partners not getting determined inspite of settings in place.

    Hi Gurus This problem is regarding partners not determining,. Cr/dr memo request is created with ref to invoice. there are partners X,Y,Z  at invoice item but are not getting copied to Cr/dr memo request item. copy control is maintained. can you help

  • Swipe features in Lion on older macbook?

    I just upgraded to Lion with my 13" macbook pro with the glass trackpad.  But I want to upgrade my macbook without the multi touch trackpad.   My concern is all the gestures that you use with Lion on the trackpad.   Currently with 10.6 I use expose b

  • Error in validation of Form 24Q

    Dear All We are faced with a problem in Form 24Q, We have the payment date for the salary as the first of the next month because of which in any quarter the third month validation is giving errors. with the error saying "Invalid date on which the tax

  • Modifying Base Permissions in SharePoint Online

    Hi,  I am new to SharePoint and trying to understand the security model of SharePoint 2013. As far as I understand, Base Permissions in SharePoint cannot be modified either via UI or via APIs. However, Permissions levels (combination of base permissi

  • List of calendar events??

    Before the iOS7 update you used to be able to look at a complete list of calendar events and scroll through them but now it appears you can only view the list at the bottom of each calendar date. Can someone tell me if there is a way to look at the c