Regarding Reloading of Data Into The Cube

Hi All,
I have a situation where I have to reload the data into the InfoCube. The InfoCube is already having Full Load and Deltas for a period of 1 month. Now if I do a full load will I get the data as of today and can I do deltas tomorrow onwards after doing an Init???
Should I have to delete the Data from Delta Queue in Source System (RSA7)???
Please clarify me.
Thanks in advance.
Regards
Jay

Hi,
  if you already have intialisations on the cube you can do a repair ful load. which will bring all the data and you can continue using the delta functionality from tomorrow.
hope this helps.
Ravi.

Similar Messages

  • Error while loading data into the cube

    Hi,
    I loaded data on to PSA and when I am loading the data to the cube through DataTransferProcess, I get an error (red color).
    Through "Manage", I could see teh request in red. How do I get to knoe the exact error? Also what could be the possibel reason for this?
    Also can some one explain the Datatransfer process(not in process chain)?
    Regards,
    Sam

    Hi Sam
        after you load the data  through DTP(after click on execute button..) > just go to monitor screen.. in that press  the refresh button..> in that it self.. you can find the  logs..
       otherwise.. in the request  screen also..  beside of the request number... you can see the logs icon.. you can click on this..
    DTP  means..
    DTP-used for data transfer process from psa to data target..
    check thi link..for DTP:
    http://help.sap.com/saphelp_nw04s/helpdata/en/42/f98e07cc483255e10000000a1553f7/frameset.htm
    to load data in to the datatargets or infoproviders like DSO, cube....
    in your  case.. the problem may be.. check the date formats.. special cherectrs... and
    REGARDS
    @JAY

  • Error while loading the data into the cube using DTP.

    No SID found for the value'UL' of characterstics 0BASE_UOM.
    Please give me the idea ,to solve the error.
    Thanks&Regards
    Syam Prasad Dasari

    https://forums.sdn.sap.com/click.jspa?searchID=23985990&messageID=6733764
    https://forums.sdn.sap.com/click.jspa?searchID=23985990&messageID=5062570

  • Data Load into the Cube based on Fiscal Year

    Hi All,
    I was told to load the data into the cube coming from 3 different datasources, but based on fiscal year. I was told to load the data for 2010, 2009 and so on.
    Any suggestions please...

    Hi Dear,
                Write the following code in start Routine of Update Rule:-
    In BI 3.x
                   Delete DATA_Package where calday lt '20090101' and calday gt '20091231'.
               to load data onlyfor 2009.
              same thing for 2010.
    In BI 7.0
               Start Routine of Transformation:-
              delete source_package where calday lt  '20090101' and calday gt '20091231'.
    Regards
    Obaid

  • How can i retrieved data into the infocube from archived files

    hi,
    i have archived cube data and i have to load data into the cube from archived files.
    so now i want to find archived files and how to load data into the cube.
    thanks

    Hi.....
    Reloading archived data should be an exception rather than the general case, since data should be
    archived only if it is not needed in the database anymore. When the archived data target is serving also as a
    datamart to populate other data targets, Its recommend that you load the data to a copy of the original
    (archived) data target, and combine the two resulting data targets with a MultiProvider.
    In order to reload the data to a data target, you have to use the export DataSource of the archived data
    target. Therefore, you create an update rule based on the respective InfoSource (technical name 8<data
    target name>). You then trigger the upload either by using ‘Update ODS data in data target’ or by
    replicating the DataSources of the MYSELF source system and subsequently scheduling an InfoPackage
    for the respective InfoSource
    If you want to read the data for reporting or
    control purposes, you have to write a report, which reads data from the archive files sequentially.
    Alternatively, you can also use the Archiving Information System (AS). This tool enables you to define an
    InfoStructure, and create reports based on these InfoStructures. The InfoStructures define an index for
    the archive file data. At the moment, the archiving process in the BW system does not fill the
    InfoStructures during the archiving session automatically. This has to be performed manually when
    needed.
    Another way of displaying data from the archive file is by using the ‘Extractor checker’ (TCODE RSA3).
    Enter the name of the export DataSource of the respective data target (name of the data target preceded
    by ‘8’), and choose the archive files that are to be read. The extractor checker reads the selected archive
    files sequentially. Selection conditions can be entered for filtering but have to be entered in in internal
    format
    It will remain same in the change log table.
    Check this link :
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b32837f2-0c01-0010-68a3-c45f8443f01d
    Hope this helps you...........
    Regards,
    Debjani............

  • How to reduce the run time of ABAP code (BADI) when it is reading huge data from BPC Cube and thus writing  back huge data to the cube.

    Hi All ,
    In Case of reading huge amount of record from BPC Cube  from BADI code , performing calculations and writing back huge amount of data into the cube , It takes lot of time . If there is any suggestion to read the data  from Cube  or writing data into the cube using some Parallel Processing  methods , Then Please suggest .
    Regards,
    SHUBHAM

    Hi Gersh ,
    If we have a specific server say 10.10.10.10 (abc.co.in) on which we are working, Then under RZ12 we make the following entry  as :
    LOGON GROUP          INSTANCE
    parallel_generators        abc.co.in_10         ( Lets assume : The instance number is 10 )
    Now in SM59 under ABAP Connections , I am giving the following technical settings:
    TARGET HOST          abc.co.in
    IP address                  10.10.10.10
    Instance number          10
    Now if we have a scenario of load balancing servers with following server details (with all servers on different instance numbers ) :
    10.10.10.11   
    10.10.10.13
    10.1010.10
    10.10.10.15
    In this case how can we make the RZ12 settings and SM59 settings such that we don't have to hardcode any IP Address.
    If the request is redirected to 10.10.10.11 and not to 10.10.10.10 , in that case how will the settings be.
    I have raised this question on the below thread :
    How to configure RZ12  and SM59 ABAP connection settings when we have work with Load Balancing servers rather than a specific server .
    Regards,
    SHUBHAM

  • Could not see the data in the cube

    Hi all,
    Iam trying to load data from one cube to another. Everything is fine but Iam not able to see the data in the new cube.
    I created 2 cubes--> Cube A new & cube B new which are copied from existing cubes cube A and Cube B. I need to load data from these new cubes cube Anew & cube B new to another cube (cube C new)  which is also a copy of existing cube C. I created update rules and activated them.I tried to load the data into the cube C new. I can see the load to be green but I cannot see the data in the cube C new. it shows 0 records in cube C new. there are records in cube A new & cube B new.
    Can anyone plz help me out.
    Thanks in advance.
    Thanks,
    Preethi.

    Hi Balu,
    I dont know what exactly the problem was but even after the load was successful i could not see any records in the new cube. The source cubes had 60,446 and 7595 records. but the target cube had 0 records.
    I ran the program SAP_FACTVIEWS_RECREATE as Bhanu suggested and deleted the data from the new cube(inspite of having 0 records I deleted them). I tried to load the data again which was successful and I could see the data in the new cube.
    Hope its clear.
    Thanks,
    Preethi.

  • Base unit of measure values  not loading into the cube.

    HI Xperts,
    i have created a generic data source and in that i took unit as base_uom and one quantity keyfigure and in my cube i have taken 0base_uom and a quantity key figure infoobject.
    in transformation i mapped the  base_uom(source) to 0base_uom (target)and soures quantity KF infoobject to targer keyfigure infoobject.
    after loading the data into my cube..i am getting the values for quantity keyfiure infoobject but for  unit (CS,IT) i am not getting any value in my cube.
    so in my report also the for a specific quantity unit value is showing ERR.which is wrong.it should show me as CS or IT(10CS or 15 IT etc).
    Can any one help me on this?
    Regards,
    satihs

    > in transformation i mapped the base_uom(source) to 0base_uom (target)and soures quantity KF infoobject to targer keyfigure infoobject.
    This is not how you map Key Figure and Currency in cube..
    First check in RSD5 whether you have Unit Infoobjet assigned to the key figure.
    in Type/Unit tab ..you have to assign 0BASE_UOM to your Key Figure.
    Then in transformation from DataSource to Cube  : connect both the Unit and Key ( in source) figure to target KF.
    In Rule Detail you have to select .
    Target Unit = 0BASE_UOM
    Unit  = From Source
    Source Unit = Your source field name for unit
    Let me know after doing all these whether are getting unit field in cube
    Regards
    Anindya

  • How to enhance the new field(mblnr) into the cube.

    Dear Experts,
                               My client wants GR document number and GR posting date in the drill down option in reporting, i have taken the field names of that. MBLNR and BUDAT, I only found BUDAT in the datasourse field level,please help me how to enchance MBLNR into the cube 0PUR_C01.
    Thank You
    Regards
    Srikanth

    SELF

  • How to write plan data into Transactional Cube from Visual Composer ?

    Dear experts,
    Visual Composer is very powerfull to build a 'nice-looking' web-report (I think is much better than Web Application Designer in term of freely position the chart/graph/table/selection parameter in anywhere we like. We can't have this flexibility in WAD) plus it also has strong integration to R/3 for posting the  transactions via RFC-Enabled Function Module (such as changing customer credit limit, approve PO, approve Sales Order, etc), so that the R/3 user won't even have to know the TrxCode.
    In addition to this, I just wondering if VC also capable to write plan data into transactional cube via Write-Enable Query. We can simply insert a BW Query into VC as planning layout (set the edit-mode to "editable") plus having some Wizards to make it more user-friendly and self-explanatory for the users.
    But I am not sure how to save the changes back to transactional cube from VC.
    Could anyone please kindly advise how to achieve this.
    "Seeing is believing" is always my learning method, so I would appreciate some "actioanable" explanation or hints.
    Many thanks,
    Sen

    Hi Sen,
    As far as I know, it's not possible to write data into transactional cubes directly from a table in Visual Composer through standar method, but it's possible to integrate a WAD with your input ready query into a URL element in VC (You can fill entry variables in VC and call WAD dinamically), so it's possible to create a planning application in VC calling dinamically WAD's.
    Another way to interact through VC is creating button objects that calls planning sequences (with BAPI calling RSPLS_PLSEQ_EXECUTE). Also can fill variables in VC and pass them to sequence.
    Regards,
    Enrique

  • Loading data into two cubes

    We want to set up a delta refresh from R/3 data that will pull data into two cubes.  One cube I want all the records to be loaded, in the second cube I want to filter the records that are loaded into that cube.  I can figure out how to load the data into two cubes, but can I place a filter on one of the loads.  (i.e. only load records of where type = 'X')
    Thanks,
    Chris

    You can do that in the Update Rules to the second cube... In the Start Routine...
    DELETE FROM DATA_PACKAGE
    WHERE type EQ 'X'.
    (Please, verify the right syntax)
    Then with only one Infopackage, you can load both cube with different conditions.
    Hope it helps.
    Regards,
    Luis

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Steps for loading data into the infocube in BI7, with dso in between

    Dear All,
    I am loading data into the infocube in BI7, with dso in between. By data flow looks like....
    Top to bottom:
    InfoCube (Customized)
    Transformation
    DSO (Customized)
    Transformation
    DataSource (Customized).
    The mapping and everything else looks fine and data is also seen in the cube on the FULL Load.
    But due to some minor error (i guess), i am unable to see the DELTA data in DSO, although it is loaded in DataSource through process chains.
    Kindly advise me, where did i miss.
    OR .. Step by step instructions for loading data into the infocube in BI7, with dso in between would be really helpful.
    Regards,

    Hi,
    my first impulse would be to check if the DSO is set to "direct update". In this case there is no Delta possible, because the Change log is not maintained.
    My second thought would be to check the DTP moving data between the DSO and the target cube. If this is set to full, you will not get a delta. It is only possible to create one DTP. So if you created one in FULL mode you can't switch to Delta. Just create the DTP in Delta mode.
    Hope this helps.
    Kind regards,
    Jürgen

  • Could not see loaded data in the Cube

    Hi Gurus,
    Pl. help me in this: I am not able to get teh data loaded into the cube. In PSA it is loaded right. after that I created Transformation on InfoCube and I got Proposal Generated. Then I created DTP and executed, and monitor shows green. but when I right click on the cube and look for data(also maintain master data on InfoObject), I see only the headers of the table but no data is displayed.
    I hope you have similar situations. Thanking you in advance.

    Check if the request is reportable, sometimes the load could be fine, but the request may not be reportable. Make sure in your start routine you don't have a routine which will filter or delete the data out from the cube before it gets update.
    thanks.
    Wond

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Thanks,
    Gopi R

    Hi,
    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).
    Regards,
    Gregor

Maybe you are looking for

  • Brand New MacBook Air Randomly Resetting (IMO WiFi Issue)

    Good evening, My problem is as follows: I purchased a brand new Macbook Air last week and since the get go have been having issues while transferring large amounts of data over WIFI. The issue is that it seems when there is no activity (I leave the f

  • ITunes is "Not Responding" to my iPod

    Since the 7.7 upgrade, I have been unable to access my non-touch iPod (it is the last generation before it became the Classic model). My iPod Touch syncs up and loads fine, and I am able to transfer things to that device without any issue. But it's t

  • In Service contract, to change the valid form date on service code level.

    Hi. I want to change the price valid form date on service code level ( i.e. in condition)  in Service contract. I know i can  do that with ME32 transaction but i have lot of Service contract with more then 10,000 line items Is there any way that i ca

  • Using functions in Resultset while getting database data.

    Well I'm trying to reset my resultset and it doesn't work. Frist I extract data from it. Then I want to start to get other data from hte beginning again. Is this not possible. Can I just get thew data once. I'm working with an access database and dev

  • Deploy form, question about database object or connection parameters

    When form to be deployed to the dev/test/production environment, should it be compiled with that specific database connection username, password and sid? Another words, when I compile the database objects created for the compile is embedded in the .f