Asset loading date

I have two legacy  assets, asset A, acquired in 2009, APC value: 10000, accumlated depreciation 3000,
asset B, acquired in this year, apc value:20000, accumulated depreciation:8000. the last depreciation run in legacy system will
be July.2011(july.31), after that it will move to SAP.
any difference between we use as91 to upload asset A and B, they are prior. year acqusition and current year acquistion?
what should be asset take over day we configured before using as91?
so what should be the capitalization date for asset A or asset B?

Hi,
1) What should be asset take over Date?
Example:
transfer date = December 31, 2010
=> last closed fiscal year = 2010
2) What should be the capitalization date for asset?
Capitalization date must be earlier than or same as transfer date
Example:
Capitalization date 2011-06-01 is later than the transfer date 2010-12-31 for legacy data transfer of assets.
This is not possible.
regards Bernhard

Similar Messages

  • Group and Individual Assets Legacy data load

    Hi, I work for a oil and gas company and I am having issues for the initial load for the group and individual assets into SAP .
    Any help in regards to this would be helpful.
    Below are the values that I would like to load into SAP for group ( grp 1 )  and individual assets ( Ind 1 and Ind 2 ) into SAP.  We would like the depreciation to be carried on the group asset level and allocated on the NBV of individual assets ( which I think seems not possible but the standard allocates the depreciation only on acquisition value ) .
    Asset              APC Value         Acc Depreciation      NBV
    grp 1                20,000.00           10,000.00                 10,000.00
    Ind  1               10,000.00            6,000.00                    4,000.00
    Ind  2               10,000.00            4,000.00                    6,000.00
    The problem is for intital load as well once the Ind 1 and Ind 2 are assigned to group asset grp 1 , it doesnt consider the acc depreciation at all , and the system allocations the accumulated depreciation based on apc values rather then NBV as of the load date.
    Also, if I plan to load only the NBV of the group and individual assets only, the clients wants the APC values as well for reporting , so not where I can get those from if I only load the NBV.
    Please let me know other approaches for initial load of group and individual assets that you guys might have worked on .
    Thanks
    bharani

    hi,
    Hope your requirement is to see the takeover values as on legacy asset takeover date, for those group assets mistakenly created with AS21...instead of AS81.
    Am i right...........
    Then you can do this by doing below.
    1) Go to OAAQ, for your CC, the last closed FY should be same as the FY, which is actually used during legacy asset upload.
    2) now go to AS82, for your wrongly created group assets and in takeover values tab, enter the values manually.
    Note: Assets created by AS01/AS21 can changed with AS92/AS82, then you will have takeover values tab in addition to normal AS02/AS22 transactions.
    once updation is finished, then again revert back the changes in OAAQ.
    This will solve your issue.
    Thanks,
    Srinu

  • Error while loading data after level 17 patch

    Hi all,
    A week ago we had patch level 17 implemented on our SAP NetWeaver BI 7.0 system. Since then, we cannot load any data to some cubes using update rules. We get "Error when generating the update program" (Message no. RSAU484), status red while loading data.
    I tried activating the update rules
    We get te same error when loading data from a DSO into a cube.
    The strange thing is that it doesn't happen for all the objects in the system. We have troubles loading deltas from 2LIS_12_VCITM (delivery items), 2LIS_13VAITM (sales order items) and from a BI content DSO in assets accountng.
    Please if you have any idea or if this happend to you, i appreciate any suggestion.
    This is not an implementation, these processes are already live for months, and we didtn't have any problems before the patch..
    thanks,
    Andreea

    We ran into same issue & another one during our BI 7.0 SPS15/SP17 upgrade.
    The solution:
    1. Apply 2 notes:
    [Note 1146851 - BI 7.0 (SP 18): Error RSAU 484 when generating update|http://service.sap.com/sap/support/notes/1146851]
    [Note 1152453 - Termination generating InfoCube write program|http://service.sap.com/sap/support/notes/1152453]
    2. Then run program RSDD_RSTMPLWIDTP_GEN_ALL
    to generate the new programs for InfoCubes from the template RSTMPLWIDTP and cleans up the staging table RSUPDINFO per note 1152453.

  • SQL LOADER DATA INSERT ISSUES

    I tried inserting data into an empty table in my database but got the following error:
    SQL*Loader: Release 11.1.0.6.0 - Production on Tue Jul 7 22:01:35 2009
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    SQL*Loader-350: Syntax error at line 7.
    Expecting "," or ")", found "VARCHAR2".
    countryname VARCHAR2(25),
    ^
    What is wrong what this please? Find below the contents of .CTL and .DAT files.
    Content of load_country.CTL
    LOAD DATA
    infile '/D:/Data/Load/load_country.dat'
    INTO TABLE ASSET.COUNTRY
    INSERT
    FIELDS TERMINATED BY ';'
    (countrycode,
    countryname VARCHAR2(25),
    population NUMBER(10))
    The content of the load_country.DAT file:
    BEN;Benin;8791832
    BFA;Burkina Faso;15746232
    CPV;Cape Verde;503000
    Thank you.
    Ofonime Essien.

    Hello,
    After solving the Varchar2 problem, I received this error for the NUMBER data type:
    SQL*Loader: Release 11.1.0.6.0 - Production on Wed Jul 8 11:47:16 2009
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    SQL*Loader-350: Syntax error at line 8.
    Expecting "," or ")", found "number".
    population number(10))
    ^
    I replaced the Number data type with Float but received this error:
    SQL*Loader: Release 11.1.0.6.0 - Production on Wed Jul 8 11:52:27 2009
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    SQL*Loader-350: Syntax error at line 8.
    Expecting valid column specification, "," or ")", found "(".
    population Float(10))
    ^
    Replacing Number data type with Integer, I received this error:
    SQL*Loader: Release 11.1.0.6.0 - Production on Wed Jul 8 12:17:11 2009
    Copyright (c) 1982, 2007, Oracle. All rights reserved.
    Control File: D:\Data\Load\tsl.CTL
    Data File: D:\Data\Load\c.dat
    Bad File: D:\Data\Load\c.bad
    Discard File: none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array: 64 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table C, loaded from every logical record.
    Insert option in effect for this table: INSERT
    Column Name Position Len Term Encl Datatype
    COUNTRYCODE FIRST * ; CHARACTER
    COUNTRYNAME NEXT 27 VARCHAR
    POPULATION NEXT 10 INTEGER
    SQL*Loader-941: Error during describe of table C
    ORA-04043: object C does not exist
    Thank you so much.
    Ofonime.
    Edited by: ofonime on Jul 8, 2009 7:09 PM

  • Asset:load

    Hi,
    for
    asset:load
    from database which table will be queried for retriving the data.
    Thanks in advance

    Hello,
    This tag is mainly used (and should only be used) to query Basic assets.
    After you have created basic asset say - "TestBasicAsset", tablename with same name is created.
    Hence, with asset:load tag, you can query an asset either with type+objectid attribute or name+type+value attribute. You have to provide this tablename - TestBasicAsset in type. Check the tag - asset:load
    Read about Basic and Flex assets from Developer's Guide - http://docs.oracle.com/cd/E29542_01/doc.1111/e29634/toc.htm
    Regards,
    Guddu

  • Error in Data Extraction for Asset Accouting Data Source

    Hi all,
    We have install the new flow of the Asset Accounting i.e 0FI_AA_20 data source. We are trying to load the data in BW. The problem is the volume of data is so high. Around 35 lacs records.
    I am getting below error while loading in the BW.
    "Error 7 when sending an IDoc"
    "No more storage space available for extending an i 3 "
    Please suggest..
    Regards,
    Macwan James.

    Hi Macwan James,
    Not only the requests from DS to which you are loading data. You can delete the requests from other DS--> manage screens also if they are no longer required.
    Or else use selection screen of Infopackage to load the data. Change the selections & try to load the data.
    Hope it helps!
    Regards,
    Pavan

  • Loading data one year at a time

    Hi,
    We have a situation where we need to load data one year at a time. I saw this done a few years ago but do not remember the details.
    What I am thinking is that we could initially run a full load with the following parameters:
    $$ANALYSIS_START: 1/1/2006
    $$ANALYSIS_START_WID: 1/1/2006
    $$INITIAL_EXTRACT_DATE: 1/1/2006
    $$ANALYSIS_END_WID: 1/1/2006
    And this should give us one year. What I am not sure about is how to load each subsequent year???
    Regards

    Is the issue a performance issue (ETLs running for too long)? The problem is that if you do Year by year..and you want to do a "incremental" load for each year, that would be even more of a load..since you are not allowing for BULK load (where the tables get truncated). Either you can truncate and do BULK or incrmental..which may be an even heavier load. I think you are assuming that this approach will somehow help you from a hardware limitation standpoint..do you know for sure that it will?
    If you really do want to do it, as I mentioned, you can edit the INITIAL and END parameters. It would help if you clarify the Hardware limitation...I think there are better ways to handle this than to do what you are doing.

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Thanks,
    Gopi R

    Hi,
    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).
    Regards,
    Gregor

  • Can we load data using .xls in user define format(without using default template)

    Hi All,
    I'm new bee to FDM. Part of HFM support i use FDM to load flatfile data. Just has a bit more knowledge than end user.
    Requirement is that i need to load data from MS excel to Planning application via FDM.
    Previously application is in Excel(Macro driven) and upstream(data) is also in Excel(multi tab).
    As of my knowledge data can be loaded from .csv file(Excel save as CSV) with single tab.
    Could you please let me know possibilities to load data from .xls(.xlsx) to FDM.
    Thanks in advance.

    If you want to load data using Excel, utilising FDM's out-of-the-box functionality you will have to use one of the templates supplied i.e. Excel Trial Balance or Excel Multi-load template.

  • Can we load data in chunks using data pump ?

    We are loading data using data pump. So I want to clear my understanding.
    Please correct me if I am wrong on my understandings -
    ODI will fetch all data from source (whether it is INIT or CDC ) in one go and unload into staging area.
    If it is true, will performance hamper in case very huge data (50 million records at source) at source as ODI tries to load entire data in one go. I believe it will give better performance if we load in chunks using data pump.
    Please confirm and correct.
    Also I would like to know how can we configure chunk load using data-pump.
    Thanks in Advance.
    Regards,
    Dinesh.

    You may consider usingLKM Oracle to Oracle (datapump)
    http://docs.oracle.com/cd/E28280_01/integrate.1111/e12644/oracle_db.htm#r15c1-t2
    In 11g ODI reads from source and write to target in parallel. This is the case where you specify select query in source command and insert/update query in the target command. At source side Odi reads records from source and add them to a data queue. At target side a parallel thread reads data from the data queue and writes to the target. So the overall performance would be the slower of the read or write process.
    Thanks,

  • Can not load data into the ods.

    When I was trying to load data into BW3.5 from R3 system, I am getting this message
    "For InfoSource ZSD_A507_011, there is no Hierarchies for master data in source system R3PCLNT300
    "  any idea.  There is no hierarchies for this.

    Hello Mat,
    What are you trying to load??
    If the data source is a hierarchy datasource??
    just check this thread
    There are no hierarchies for this InfoSource in source system
    Thanks
    Ajeet

  • Can't load data through smart view (ad hoc analysis)

    Hi,
    There is EPM application where I want to give ability to planners to load data through smart view (ad hoc analysis). In Shared Services there are four options in
    EssbaseCluster-1: Administrator, Create/Delete Application, Server Access, Provisioning Manager. Only Administrator can submit data in smart view (ad-hoc analysis). But I don't want to grant Essbase administrator to planners, I'm just interested to give them ability to load data through ad-hoc analysis. Please suggest!

    I take that you refreshed the Planning security, If not refresh the security of those users. Managing Security Filters
    Check in EAS whether those filters are created with "Write" permissions.
    Regards
    Celvin
    http://www.orahyplabs.com

  • Can we load data for all levels in ASO?

    Hi All,
    Im creating cube in ASO
    can i load data for all levels in ASO
    we can load data for all Levels In BSO but in ASO i need confirmation????
    and one more
    wat is the consider all levels option in ASO is used for ? wat is the purpose?
    Can any one help ,it would be appriciate.
    Thanks

    In an ASO cube you can only load to level zero
    The consider all levels if used for aggregation hints. It allows you to tell the aggregation optimizer to look at all levels when deciding if aggregation needs to be done on the dimension

  • Can we load data for all levels in ASO cube

    Hi All,
    Can we load data for all levels of members in ASO cube in 9.3.1.
    Regards

    Yes you can load data for all levels in an ASO cube in any version HOWEVER, none of the upper level data in any cube will be there when you look for it. You will get a warning message in the load because ASO cubes don't store data at upper levels. It is the same as loading data into dynamic calc members in BSO cube. It will do the load without compalints, but there will be no data there (At least you get the warning in ASO)

  • Not able to load data in tables with correct way

    Hi
    i made one trigger to load data in view and tables.
    CODE FOR TRIGGER>>>>
    Object Details Code Errors SQL
    CREATE OR REPLACE TRIGGER "WELL_GENERATOR_TRIGGER_1"
    INSTEAD OF INSERT ON VIEW_WELL_GENERATOR_FORM
    FOR EACH ROW
    DECLARE
    rowcnt number;
    BEGIN
    INSERT INTO facility (FAC_PK) VALUES (:NEW.FAC_PK);
    SELECT COUNT(*) INTO rowcnt FROM WELL WHERE WEL_PK = :NEW.WEL_PK;
    IF rowcnt = 0 THEN
    INSERT INTO WELL (WEL_PK,SITE,FAC_FK) VALUES(:NEW.WEL_PK,:NEW.SITE,:NEW.FAC_PK);
    ELSE
    UPDATE WELL SET WELL.SITE = :NEW.SITE,
    WELL.FAC_FK = :NEW.FAC_PK
    WHERE WELL.WEL_PK = :NEW.WEL_PK;
    END IF;
    SELECT COUNT(*) INTO rowcnt FROM WELL_STATUS WHERE WELL_STATUS.STA_PK = :NEW.STA_PK;
    IF rowcnt = 0 THEN
    INSERT INTO WELL_STATUS (WELL_TYPE, WELL_TYPE_DATE, OPER_STATUS, CLASS,WEL_FK)
    VALUES(:NEW.WELL_TYPE, :NEW.WELL_TYPE_DATE, :NEW.OPER_STATUS, :NEW.CLASS,:NEW.WEL_PK);
    ELSE
    UPDATE WELL_STATUS SET WELL_STATUS.WELL_TYPE = :NEW.WELL_TYPE,
    WELL_STATUS.WELL_TYPE_DATE = :NEW.WELL_TYPE_DATE,
    WELL_STATUS.OPER_STATUS = :NEW.OPER_STATUS,
    WELL_STATUS.CLASS = :NEW.CLASS,
    WELL_STATUS.WEL_FK = :NEW.WEL_PK
    WHERE STA_PK = :NEW.STA_PK;
    END IF;
    SELECT COUNT(*) INTO rowcnt FROM PERMIT WHERE PERMIT.PER_PK = :NEW.PER_PK;
    IF rowcnt = 0 THEN
    INSERT INTO PERMIT (AUT_STATUS,WEL_FK) VALUES (:NEW.AUT_STATUS,:NEW.WEL_PK);
    ELSE
    UPDATE PERMIT SET PERMIT.AUT_STATUS = :NEW.AUT_STATUS,
    PERMIT.WEL_FK = :NEW.WEL_PK
    WHERE PERMIT.PER_PK = :NEW.PER_PK;
    END IF;Now But still i am not getting result which i want.Like in WELL_STATUS table i could nt able to insert the value (WEL_FK) from (WEL_PK).And In PERMIT table column (WEL_FK) still i could nt able to insert the value (WEL_PK).
    Now WEL_PK value autogenerate from the SEQ same with STA_PK and PER_PK.But instead of taking value from seq STA_PK take the value from WEL_PK (dnt know why ? But in data its shows me Same value WEL_PK and STA_PK).Same thing With PER_PK.
    so where i am wrong ? i really need your help.
    workspace:PRACTISE
    UN:[email protected]
    PW:testing
    Application:     39289 - TESTTING
    Thanks
    Edited by: vijaya45 on Jul 13, 2009 9:44 PM
    Edited by: vijaya45 on Jul 14, 2009 12:48 AM

    Hello vijaya45,
    Not sure if this will help, but I noticed you currently have a WELL_GENERATOR_TRIGGER_1 and a WELL_GENERATOR_TRIGGER_2, and "1" is disabled but "2" is enabled. Which one is supposed to be running at this point?
    John

Maybe you are looking for

  • Displaying the number of files in a folder with Mountain Lion?

    I recently purchased a new iMac with Mountain Lion and I pretty much hate it. I have been a fan of just about EVERY update, so I am not resistant to change. I just think this is an inferior product. It is slow, clunky, and the update alerts remind me

  • I want to run simple procedure in TOAD

    Hi guys ,I have already compiled the procedure .When i run it it gives me another window.please help me out .I will give you the procedure and what i get after i execute This is the procedure CREATE OR REPLACE PROCEDURE find_sname (i_student_id IN NU

  • Bug signing in to Adobe DPS Forum?

    I think there is a problem when signing in to this Forum. There is a few days ago I am having trouble to sign in through DPS Forum (don't know others). To get in I have to type adobe.com and sign in from the home page. Then, I click "learning > adobe

  • IDvd 5 problem on Intel Macbook!

    I am having problems running iDVD 5 on Intel Macbooks. iDVD 5 loads up, the themes show up in the side tray and the music plays. However, the themed "images and drop zones" are not visible in the main window.... in its place, only a black background

  • How to create table with dynamic amount of columns which are nested columns

    M trying to fetch data and show in a javaFX table. My table displays the details of different items from ItemVO , where : public class ItemVO() String itemName; List<ItemTypeVO> type; and My ItemTypeVO has the following attributes : public class Item