Problem when upload legacy data

Hi all,
I've faced a problem when uploading asset legacy data (AS91) for customer. My customer have started depreciation on May 21, 2009, and I set transfer date is November 30, 2010. So SAP doesn't calculate right planned value for December, 2010. I think that problem related to the first month depreciation. It calculated full month for depreciation (instead of from May 21) and change the planned value for December 2010. So I want SAP just calculate 11 days depreciation for May 2009. How can I config to do this?
Thanks in advance!

Hi,
Check in the asset master Opening Start Date for depreciation.. If it is under open mode change that date and see asset explorer for depreciation values.
Thanks,
Padmaja

Similar Messages

  • How to avoid the split problem when uploading the data from csv file

    Dear Friends,
                  I have to upload data from the .csv file to my custom table , i have found a problem when uploading the data .
    i am using the code as below , please suggest me what i have to do in this regard
          SPLIT wa_raw_csv  AT ',' INTO
                    wa_empdata_csv-status
                     wa_empdata_csv-userid
                     wa_empdata_csv-username
                     wa_empdata_csv-Title
                     wa_empdata_csv-department.
    APPEND wa_empdata_csv TO  itab.
    in the flat file i can see for one of the record for the field Title  as
    Director, Finance - NAR............there by through my code the  wa_empdata_csv-Title is getting splited data as "Director, and  Department field is getting Finance - NAR" , i can see that even though  " Director, Finance - NAR"  is one word it is getting split .
    .......which is the problem iam facing.Please could any body let me know how in this case i should handle in my code that this word
    "Director,Finance - NAR"   wil not be split into two words.
    Thanks & Records
    Madhuri

    Hi Madhuri,
    Best way to avoid such problem is to use TAB delimited file instead of comma separated data. Generally TAB does not appear in data.
    If you are generating the file, use tab instead of comma.
    If you cannot modify the format of file and data length in file is fixed character, you will need to define the structure and then move data in fixed length structure.
    Regards,
    Mohaiyuddin

  • Please send detail steps for uploading legacy data

    Hi friends,
    please send detail steps for uploading legacy data
    Thanking u in advance,
    Diwa.

    HI U CAN USE LSMW TO UPLOAD LEGACY DATA
    LSMW is used for migrating data from a legacy system to SAP system, or from one SAP system to another.
    Apart from standard batch/direct input and recordings, BAPI and IDocs are available as additional import methods for processing the legacy data.
    The LSMW comprises the following main steps:
    Read data (legacy data in spreadsheet tables and/or sequential files).
    Convert data (from the source into the target format).
    Import data (to the database used by the R/3 application.
    But, before these steps, you need to perform following steps :
    Define source structure : structure of data in the source file.
    Define target structure : structure of SAP that receives data.
    Field mapping: Mapping between the source and target structure with conversions, if any.
    Specify file: location of the source file
    Of all the methods used for data migration like BDC, LSMW , Call Transaction which one is used most of the time?
    How is the decision made which method should be followed? What is the procedure followed for this analysis?
    All the 3 methods are used to migrate data. Selection of these methods depends on the scenario, amount of data need to transfer. LSMW is a ready  tool provided by SAP and you have to follow some 17 steps to migrate master data. While in BDCs Session method is the better choice because of some advantages over call transaction. But call transaction is also very useful to do immediate updation of small amout of data. (In call transaction developer has to handle errors).
    SO Bottom line is make choice of these methods based of real time requirements.
    These methods are chosen completely based on situation you are in. Direct input method is not available for all scenario, else, they are the simplest ones. In batch input method ,you need to do recording for the transaction concerned. Similarly, IDoc, and BAPI are there, and use of these need to be decided based on the requirement.
    Try to go through the some material on these four methods, and implement them.  You will then have a fair idea about when to use which.
    LSMW Steps For Data Migration
    How to develop a lsmw for data migration for va01 or xk01 transaction?
    You can create lsmw for data migration as follows (using session method):
    Example for xk01 (create vendor)
    Initially there will be 20 steps but after processing 1 step it will reduced to 14 for session method.
    1. TCode : LSMW.
    2. Enter Project name, sub project name and object name.
        Execute.
    3. Maintain object attributes.
        Execute
        select Batch Input recording
        goto->Recording overview
        create
        recording name.
        enter transaction code.
        start recording
        do recording as per ur choice.
        save + back.
        enter recording name in lsmw screen.
        save + back
    Now there will be 14 steps.
    2. MAINTAIN SOURCE STRUCTURES.
        Here you have  to enter the name of internal table.
        display change
        create
        save + back
    3. MAINTAIN SOURCE FIELDS.
        display change
        select structure
        source_fields->copy fields.
        a dialogue window will come .
        select -> from data file
        apply source fields
        enter No. of fields
        length of fields
        attach file
        save + back
    4. MAINTAIN STRUCTURE RELATIONS
        display change
        save + back
    5. MAINTAN FIELD MAPPING & CONVERSION RULE
        display change
        click on source field, select exact field from structue and enter
        repeat these steps for all fields.
        save+back
    6. MAINTAIN FIXED VALUES, TRANSACTION, USER DEFINED
        execute
        save + back
    7. SPECIFY FILES.
        display change
        click on legacy data
        attah flat file
        give description
        select tabulatore
        enter
        save + back
    8. ASSIGN FILE
        execute
        display  change
        save + back
    9. IMPORT DATA.
        execute
        display  change
        save + back
    10. DISPLAY IMPORTED DATA
          enter ok, it willl show records only.
          back
    11. CONVERT DATA
          execute
          display  change
          save + back
    12. DISPLAY CONVERTED DATA
          execute
          display  change
          save + back
    13. CREATE BATCH INPUT SESSION
          tick keep batch input folder
          F8
          back
    14. RUN BATCH INPUT SESSION.
          sm35 will come
          Object name will be shown here
          select object & process

  • Problem when uploading a large file in PI - weird SQL I/O errors

    Hi guys,
    I'm facing a very difficult problem when uploading a 35 MB with an FTPs adapter. I see in the logs that, after the translation to XML, it's going to 170 MB.
    I receive the following error in the CC Monitoring:
    Error: com.sap.aii.af.ra.ms.api.DeliveryException: Problem inserting 41827ca7-6b8c-4a87-198d-ad8a81fcb12b(OUTBOUND) into the database: com.sap.engine.services.dbpool.exceptions.BaseSQLException: Connection is invalid.
    When I look in the NWA Monitoring, I see the following details:
    SQL error occurred on connection affhb201:X11:SAPSR3DB: code=17,002, state="null", message="Io exception: Socket closed";
    SQL statement is "INSERT INTO "XI_AF_MSG" ("MSG_ID","DIRECTION","MSG_BYTES","TIMES_FAILED","SENT_RECV_TIME","STATUS","CONN_NAME","MSG_TYPE","REF_TO_MSG_ID","ADDRESS","TRANSPORT","CREDENTIAL","TRAN_HEADER","MSG_PROFILE","CONVERSATION_ID","SCHEDULE_TIME","PERSIST_UNTIL","FROM_P........
    I cannot check the Visual Admin Logs 'cause I don't have access to them yet.
    I'm pretty convinced that some swap memory, message size or whatever setting on the adapter engine or on the Java stack is preventing this. I do not get any message in CC Monitoring when uploading a smaller, 6 MB version of the same file.
    Can you please help me solve this or give me some interesting pointers?
    Never before did we experience something like this in the PI system. In addition, I didn't find any useful resource on the SDN and throughout the SAP notes for this.
    Let me know if you need more info about this.
    Best regards,
    George

    Hi George
    I am facing the same issue, where did you configure the message split in the Communication Channel?
    If I do the message split as you said, is it going to create several files or how does it work?
    Thanks in advanced
    Emmanuel

  • Upload legacy data into SAP

    Hi experts,
           I have to upload legacy data into SAP by using eCATT recording Transaction. For doing this I have completed  ‘Test script’ in tcode SECATT. After this what I have to do.. How can I attach my legacy file.. Can anyone help me. ..Its really urgent…
    Thanks in advance

    Hi,
    There are a series of weblogs in this area. Check them out, I think there are 8 in a series:
    /people/sapna.modi/blog/2006/04/10/ecatt--an-introduction-part-i
    cheers
    Aveek

  • Upload legacy data

    Hi experts,
           I have to upload legacy data into SAP by using eCATT recording Transaction. For doing this I have completed  ‘Test script’ in tcode SECATT. After this what I have to do.. How can I attach my legacy file.. Can anyone help me. ..Its really urgent…

    hi silviya,
    check these threads...
    /people/sapna.modi/blog/2006/04/10/ecatt-scripts-creation--sapgui-mode-part-iii
    Steps in ECATT
    uploading data from external files with multiple entries in ecatt
    ecatt upload
    Error in executing eCATT GUI Script
    hope these help,
    do reward if it helps,
    priya.

  • Encounter Problem when upload Data for T-Code CJ92

    Dear All,
    I want to upload Standard WBS element through transaction code CJ92.  I am encountering a problem when I want to upload more than 13 Standard WBS for a project using recording (SHDB). 
    I have also tried to upload data through BAPI u201CBAPI_BUS2054_CREATE_MULTIu201D but this BAPI didnu2019t work for standard WBS.
    Please provide me a solution to upload Standard WBS (CJ92).

    Thanks Shyam,
    But this BAPI "BAPI_PROJECT_MAINTAIN" works only for operative WBS and I want to upload Standard WBS.  The T-Code to create Standard WBS is CJ92. The screen upload can't be possible due to screen limitations.
    Please provide me any solution to upload Standard WBS.
    regards,
    Saif

  • Issue when uploading Sales data from DSO to Cube.

    Dear All,
    I have an issue when I am uploading Sales data from DSO to Cube.  I am using BI 7.0 and I have uploaded all sales document level data to my DSO.  Then I use a transformation rule to calculate the Sales Value when I am doing DTP to cube.  Cube has customer wise aggregation data.
    In DSO I have NetPrice(KF) and Delivered_QTY(KF).  I do a simple multiplication routine in the transformation from DSO to Cube.
    RESULT =   SOURCE_FIELDS-NET_PRICE * SOURCE_FIELDS-DLV_QTY .
    At the moment I use Active Table(With out Archive) on the DSO to get the data since this is my first load.
    Issue is that the figure (Sales Value) in the cube is incorrect. I am getting a very large values which is impossible.
    Can someone please help me.
    Shanka

    Hi,
    are you sure that cube has customer wise aggregated data..since it will always aggregate the values of the key figures for the same set of characters...
    did you checked the values for the key figures as well and if even they are inflated or is the problem with this key figure only.
    During the data load the records may aggregates first for the same values of characters and then multiplication will be happening....if that is the case then you may have to multiply the values before storing it in the data package and then let it aggregate....can be achived through start routine...
    but first verify if other key figures are also having the same issue.
    Thanks
    Ajeet

  • Problem while uploading master data through two files

    Hi all,
    When i try to upload master data of CUSTOMER with two files , Attribute and Text.
    Text gets uploaded perfectly.
    for Attribute, the data upload is done and psa i can see the exact data. But when i check the CUSTOMER object, i am not able to find the actually Attributs data. Instead those colmns are filed with 'E'.
    Can anybdy please explain me wheret the problem is.
    Thanks..

    Hi Sun,
    First check the data in PSA whether the records are fine or any discrepancy in them. Also check the properties of the COSTUMER object into which you are loading. The properties "type" or the "size" of the object may not be matching with the same in the file.
    Check them and try to load once again.
    Hope this helps u...
    Regards,
    Koundinya Karanam.
    Edited by: koundinya karanam on Jan 15, 2008 12:05 AM

  • Problem to upload the data into internal table record length more than 6000

    Hi all
            I stuck with this problem from past 3 days. I have to upload the data from excel sheet. iam making it tab delimited and trying to upload from gui_upload. but in the structure of file, we have, one field of 4000 characters, and other fields of 255 characters.
    how can i upload this into internal table . From excel sheet or from tab delimeted or any other format? or any special function module is there?  while iam doing this its truncating the datat 255 characters and not uploading the other fields also...
    can any one of you help me out. ASAP
    thnks in advance

    from one of the forum iam just pasting code which it is used in lsmw, try the same logic in ur code hope it can work.
    you have to create multiple lines with do...enddo loop., like this:
    (assuming excel_long_text-text is 924 characters long, 7 lines X 132 char)
    __GLOBAL_DATA__
    data: offset type i,
    text_132(132) type c.
    __BEGIN_OF_RECORD__ Before Using Conversion Rules
    Rule : Default Settings Modified
    Code: /sapdmc/ltxtl = init_/sapdmc/ltxtl.
    CLEAR offset.
    DO 7 TIMES.
    text_132 = excel_long_text-text+offset(132).
    offset = offset + 132.
    __END_OF_RECORD__ After Using Conversion Rules
    Rule : Default Settings Modified
    Code: transfer_record.
    ENDDO.
    also check this
    COMMIT_TEXT
    To load long text into SAP
    READ_TEXT
    To load long text into SAP

  • Problem when recording the data using BDC for Tcode CJ02.

    Dear Experts,
    When i am trying to record the data for TCODE : CJ02 i need to enter the project  Definition and enter the WBS element it takes me to the screen then i should select the WBS element and attach a file for that selected WBS element . The option for me to attach the attachment of file  will be available on the application area(Services for the Object).
    Now the problem when i try to do recording in SHDB this option like create attachement is not visible in the recodring . Kindly suggest me what can i do such that i attach the file for the particular project def and WBS element.
    Either suggest any function module or other procedure .......
    Regards,
    Sana.

    Hi,
      in BDC each and every action is recording. If your press enter in same screen that also recorded once aging may be this is your case repeating field values will appear. we can solve the problem for repeat fields like below.
    suppose in your excel having repeated field X1 X2 X3 the X2 contains repeated field X3 means delete the X3 field.
    Now In your itab having X1 and X2 fields. While in the LOOP the ITAB pass the X2 field to repeated the fields.
    LOOP at ITAB to WA.
    CLEAR bdcdata_wa.
    bdcdata_wa-fnam = 'BDC_CURSOR'.
    bdcdata_wa-fval = 'RM08M-EBELN'.
    APPEND bdcdata_wa TO bdcdata_tab.
    CLEAR bdcdata_wa.
    bdcdata_wa-fnam = 'INVFO-BLDAT'.
    bdcdata_wa-fval = wa-X2." 1st time pass the X2 fields
    APPEND bdcdata_wa TO bdcdata_tab.
    CLEAR bdcdata_wa.
    bdcdata_wa-fnam = 'INVFO-BLDAT'.
    bdcdata_wa-fval = wa-X2." pass the same value to repeated field
    APPEND bdcdata_wa TO bdcdata_tab.
    Endloop.
    Hope you can understand.
    Regards,
    Dhina..

  • A strange problem when uploading a file in Struts

    When i upload a file, the other elements in the form of the jsp cannot be got by the ActionForm, instead, the getters show that they are "null"s. But when i upload no file, everthing is ok.
    Moreover, the problem happens when i update an article which can have an picture with it, but when i add a new article, the problem does not appear, instead, the file can be uploaded and other element can also be got.
    i think i might be a problem of the volume of the request , because i put more information to update than to add.
    Hope someboby can help me with the problem.

    When i upload a file, the other elements in the form of the jsp cannot be got by the ActionForm, instead, the getters show that they are "null"s. But when i upload no file, everthing is ok.
    Moreover, the problem happens when i update an article which can have an picture with it, but when i add a new article, the problem does not appear, instead, the file can be uploaded and other element can also be got.
    i think i might be a problem of the volume of the request , because i put more information to update than to add.
    Hope someboby can help me with the problem.

  • Problem in Uploading Excel Data ! - Urgent

    Dear Experts,
    I am uploading excel data using FM 'TEXT_CONVERT_XLS_TO_SAP' , its directly convert data and store in internal table same as in excel sheet(row , column wise).
    But the problem is , one of the column having description more then 500 characters. so while uploding excel to itab its truncating the text and only uploding 255 characters.
    Is there any other way , so that i can upload more then 500 characters long text. Also i have tried 'ALSM_EXCEL_TO_INTERNAL_TABLE' FM , it was also not working.
    Please Help me , its Urgent.
    Points will be rewared.
    Thanks & Regards,

    Hi,
    Please use FM 'GUI_UPLOAD'.
    types: begin of ttab,
           rec(1000) type c,
           end of ttab.
    types: begin of tdat,
           fld1(10) type c,
           fld2(10) type c,
           fld3(10) type c,
           end of tdat.
    data: itab type table of ttab with header line.
    data: idat type table of tdat with header line.
    data: file_str type string.
    parameters: p_file type localfile.
    at selection-screen on value-request for p_file.
      call function 'KD_GET_FILENAME_ON_F4'
           exporting
                static    = 'X'
           changing
                file_name = p_file.
    start-of-selection.
      file_str = p_file.
      call function 'GUI_UPLOAD'
           exporting
                filename                = file_str
                filetype                = 'ASC'
           tables
                data_tab                = itab
           exceptions
                file_open_error         = 1
                file_read_error         = 2
                no_batch                = 3
                gui_refuse_filetransfer = 4
                invalid_type            = 5
                no_authority            = 6
                unknown_error           = 7
                bad_data_format         = 8
                header_not_allowed      = 9
                separator_not_allowed   = 10
                header_too_long         = 11
                unknown_dp_error        = 12
                access_denied           = 13
                dp_out_of_memory        = 14
                disk_full               = 15
                dp_timeout              = 16
                others                  = 17.
    Best regards,
    Prashant

  • Problems with uploading master data in BI 2004S

    Hi all,
    I wanna upload master data in FI from R/3 System to BI System. I created its <b>transformation</b>. Then, I <b>migrated</b> it WITH EXPORT option. After that, I created <b>Data Transfer Process</b> (DTP), I didn't change any thing, just "Save and Active". Then I created <b>InfoPackage</b> for it => choose Schedule tab and press Start. Open DTP and choose Execute tab -> press Execute.
    1. In the case of successful process (the log showed no  errors), How can we check that data is transferred successfully?
    2. <b><u>Case 2:</u></b> I did all above steps for a data source. But when I opened InfoPackage -> it said <b>"This InfoPackage has been already in Process Chain"</b>... I still continued to execute its DTP but I failed. The log said that the data was duplicated.
    What does it mean? And what should I do in this case?
    3. <b><u>Case 3</u></b>: I have a data souce Account Type with <0ACCT_TYPE_TEXT> (icon dialog). I migrated it but failed. Then, after I refreshed, this datasource moved to the group that had been migrated before (in the end of the tree) but <0ACC_TYPE_TEXT> disappeared.
    So, I cannot create DTP or InfoPackage for it to transfer. In this case, Do I have to re-create a DataSource? What should I do?
    Thanks in advance.

    case 1:
    Check there might be duplicate record. as it is a text datasource language may be maintained. U can check in monitor what record is showing error.
    case 2:
    If u migrate the data source then it is a BI7.0 version.
    If u want to recover 3.x datasource then u need to use RSDS .
    Then again u can create Data Transaformation and DTP and migrate once again.
    then load the data.
    Normally if u r working on BI7.0 do all the stuffs(DT, DTP) using 3.X version then do the migration of datasource else if u migrate first and do the transformation then u will have to manual mapping and u may missed and do a lot of mistakes.
    Thanks,
    Debasish

  • Having problem while uploading the data.

    Hi All,
    i am facing some problem when trying to upload csv file in the database.
    ERROR:
    "SQL*Loader-704: Internal error: ulconnect: OCIServerAttach [0]
    ORA-12560: TNS:protocol adapter error".
    have anybody aware of this kind of problem or having knowledge regarding this problem.database is working fine.
    -rajnish

    Hi,
    About this error, you need to enable tracing and get the exact error from the listener trace files:
    12560, 00000, "TNS:protocol adapter error"
    // *Cause: A generic protocol adapter error occurred.
    // *Action: Check addresses used for proper protocol specification. Before
    // reporting this error, look at the error stack and check for lower level
    // transport errors.For further details, turn on tracing and reexecute the
    // operation. Turn off tracing when the operation is complete.Regards,
    Yoann.

Maybe you are looking for

  • Hierarchy leaves have different InfoObject name than the value

    Hi, I am working on an hierarchy and loaded data to BW. I noticed that when I expanded the tree, the right side of the leaves saids 0COSTCENTER but the left hand side of the leaves has the department values. Would this be a prolem for reporting or an

  • Data miner problem,help!! Thanks

    why the data in data miner are not Chinese,but other strange characters?? My oracle Database's nls_Lang is AMERICAN_AMERICA.US7ASCII. When I pull the Chinese data from Oracle Database to Oracle data miner, the data turn to strange characters . How co

  • ITunes 8.2.1 quits on startup

    After upgrading to 8.2.1 iTunes quits during the last moments of startup - every time. This is what I've tried, none of it was successful: - tried starting up with or without iPhone attached - created new, empty iTunes library - created new user acco

  • How can i activate full text search in DMS

    Dear Expert,                   please anyone can guide me how to activate text search in DMS ? Regards Bhuwan

  • Cannot import OD archive on standalone server to promote to OD master

    According to page 169 of th OD server manual, I should be able to restore an OD archive on a standalone server - which will then promote it to OD master. The "..." in the Archive/Restore window is shaded and will not allow me to choose a file to rest