Hierarchy  data loading  in sap  bw 7.3 for 0 costcenter-dtp error

Hi ,
I am loading data from ECC with data source 0COSTCENTER_0101_HIER to BW.For this i have created transformations between info object HIERARCHIES 0COSTCENTER and data source and DTP as well.I have loaded data upto PSA with info package  no issues,while loading data from PSA to hierarchy i got error in DTP like "Node [00000006, 00000010 ]: Leaf ' ' already exists as child of node 00000005 ".
Can any one faced similar issue in BW7.3 ,please share your thoughts
Thanks,
Sri.

Hi Srivardhan,
I too got the same error. This error probably rises if you are entering hierarchy with some characteristics in your infoObject whose data type is numc (character string with only digits).
Instead try using a infoobject with Char as data type. or simply change the infoobject in which u are implementing your hierarchy to char data type.
And give suitable sufficient length.
Regards.

Similar Messages

  • Data load into SAP ECC from Non SAP system

    Hi Experts,
    I am very new to BODS and I have want to load historical data from non SAP source system  into SAP R/3 tables like VBAK,VBAP using BODS, Can you please provide steps/documents or guidelines on how to achieve this.
    Regards,
    Monil

    Hi
    In order to load into SAP you have the following options
    1. Use IDocs. There are several standard IDocs in ECC for specific objects (MATMAS for materials, DEBMAS for customers, etc., ) You can generate and send IDocs as messages to the SAP Target using BODS.
    2. Use LSMW programs to load into SAP Target. These programs will require input files generated in specific layouts generated using BODS.
    3. Direct Input - The direct input method is to write ABAP programs targetting on specific tables. This approach is very complex and hence a lot of thought process needs to be applied.
    The OSS Notes supplied in previous messages are all excellent guidance to steer you in the right direction on the choice of load, etc.,
    However, the data load into SAP needs to be object specific. So targetting merely the sales tables will not help as the sales document data held in VBAK and VBAP tables you mentioned are related to Articles. These tables will hold sales document data for already created articles. So if you want to specifically target these tables, then you may need to prepare an LSMW program for the purpose.
    To answer your question on whether it is possible to load objects like Materials, customers, vendors etc using BODS, it is yes you can.
    Below is a standard list of IDocs that you can use for this purpose to load into SAP ECC system from a non SAP system.
    Customer Master - DEBMAS
    Article Master - ARTMAS
    Material Master - MATMAS
    Vendor Master - CREMAS
    Purchase Info Records (PIR) - INFREC
    The list is endless.........
    In order to achieve this, you will need to get the functional design consultants to provide ETL mapping for the legacy data to IDoc target schema and fields (better to ahve sa tech table names and fields too). You should then prepare the data after putting it through the standard check table validations for each object along with any business specific conversion rules and validations applied. Having prepared this data, you can either generate flat file output for load into SAP using LSMW programs or generate IDoc messages to the target SAPsystem.
    If you are going to post IDocs directly into SAP target using BODS, you will need to create a partner profile for BODS to send IDocs and define the IDocs you need as inbound IDocs. There are few more setings like RFC connectivity, authorizations etc, in order for BODS to successfully send IDocs into the SAP Target.
    Do let me know if you need more info on any specific queries or issues you may encounter.
    kind regards
    Raghu

  • EDT Cat. 15 Data loaded to SAP but failed when updating

    Hi,
    I have an structure to load data into BP using KCLJ with Cat 15.
    Here it is.
    AKTYP
    TYPE
    PARTNER
    ROLE1
    KUNNR
    KUNNR_EXT
    BU_GROUP
    FIBUKRS
    CHIND_ADDR
    NAME_CO
    NAME_ORG1
    NAME_ORG2
    STREET
    STR_SUPPL1
    STR_SUPPL2
    STR_SUPPL3
    LOCATION
    HOUSE_NUM1
    POST_CODE1
    CITY1
    COUNTRY
    REGION
    LANGU
    CHIND_TEL
    TEL_NUMBER
    TEL_EXTENS
    We had the data loaded to SAP with ROLE 000000 and TR0100 but somehow when we tried to update the address thru ROLE TR0100, we had a dump and failed to update. However, the update is good when we use ROLE 000000 to update.
    Anyone would have a clue of this?
    Thanks

    I am new to EDT also. I know Cat 15 which can load BP. SAP has more categories which can load more using KCLJ. Here's my experience.
    Under Tcode SIMGH, find IMG structure External Data Transfer for SAP Banking. Under that structure, you can find Display Required and Optional Entry Fields for SEM Banking. When you enter 15 in the Category box, you'll see a whole list of fields about BP. You can use these fields to create your own structure.
    After you created your structure, use Define Sender Structure under External Data Transfer for SAP Banking to define your structure. After that, it is done. You can try use KCLJ to load your BP.
    If you still have other issues, mostly will be the configuration.
    Enjoy.

  • Mass data load into SAP R/3 - with XI?

    Hi guys!
    I have an issue - mass data migration into SAP R/3. Is XI a good solution? It will be about 60GB of data. Or is there a better way of this data load?
    Thanx a lot!
    Olian

    hi,
    SAP doesn't recomment using XI for mass data migration
    and 60 Gb is certainly too much
    use LSMW for that purpose
    Regards,
    michal

  • Full and delta Data loads in SAP BI 7.0

    Hi,
    Kindly explain me the difference between full and delta data loads in cube and DSO.
    When do we select change log or active table(with or without archive) with full/delta in DTP for DSO.
    Please explain the differences in data load for above different combinations.
    With thanks and regards,
    Mukul

    Please search the forums. This topic has already been discussed a lot of times.

  • Modify the Data Load wizard to remove the check for unique columns

    Hi,
    I'm trying to make a Data Loading Wizard into my application, which basically works nice except for one thing:
    I want to remove the check for unique columns, and insert ALL data which the users wants to upload.
    (Like the Data load utility under development).
    The reason is, that the data I want to upload is an export from my bank statements (in XLS / CSV format)
    And these exports don't have any primary key.
    Using the Apex wizard created Data Upload pages, I will loose data in case I had two identical payments in a day.
    Is there a way to do this, while keeping the pages created by the wizard to make Data Loading pages?
    apex 4.2.1

    I would suggest to load the data into a view and process the records into the real table(s) with instead-of-triggers. When you add a "1=2" where clause to the view, there never is any data and everything will be loaded.
    Otherwise you have to have a "real" PK value, like a timestamp or a statement_nr+line_nr.

  • Data load from SAP ECC to SAP BI

    Hello Gurus and Experts
                     I am new to BI and I m almost struggling to find a step by step process in loading data from SAP ECC to SAP BI through generic extractors from VBAK. The Problem is I can create an Infopackage for full update and loaded the records.(Lets say Upto this step 100 records are moved to PSA)
    Then I create a sale order in ECC which is a new record that has to be transfered as delta record.
    Now my questions are:
    Am I correct if I moved the initial records into PSA with Full Mode InfoPackage?
    Should I create another infopackage for delta records with Delta Mode or to change the update mode of the infopackage created earlier for Full Update mode?
    Should I create a new infopackage every time to load the delta records?
    Does separate DTP has to be created for Full and Delta updates?
    Please dont be frustrated if my questions are pointless or childish
    I would be glad if someone demonstrate the solutions in steps
    Thanks in Advance
    Karthik

    Hi Ram,
           I have followed all the steps mentioned by you, yet something is lagging in my end.
    Let me lists the steps done by me and please correct me if i am wrong anywhere.
    1. RSO2 in ECC..
    2. Created DS with short, med and long names in SD application component for VBAK with all the fields unchecked in hide selection except document number, time of document, document date, net
    value of the document and document currency.
    3. mentioned a numeric pointer as Document number for delta and used new status for changed records.
    3. saved the datasource. checked the newly created ds using the tcode RSA3.
    4. Replicated the datasource under SD Application Component in BI and Activated it
    6. Performed Transformations
    7.  created an Info Package with FULL UPDATE and executed. I got 6898 records in my PSA
    6. Created DTP with DELTA. Saved and Executed it. Got 6898 records in my DSO
    7. Created Info Package with INIT WITHOUT DATA TRANSFER and scheduled it.
    8. Switched over to ECC and added a sale order
    9. Created an InfoPackage with DELTA UPDATE and Scheduled it.
    -----I didnt receive any records in my PSA
    10. performed DTP.
    ---- I didnt received any new records in my DSO as my delta run is not working..
    PS: I am using WO DSO
    Please correct me if i am wrong anywhere
    Thanks in advance
    Karthik

  • InfoPackage error for Hierarchy data loads after upgrading to BI 7.31

    Hi All,
    We have recently upgraed to BI 7.31 from BI 7.01. After the upgrade we are observing info package error while loading the Hierarchy for 0BPARTNER infoobject, below is the long text for the error message.
    Datasource name : 0BPARTNER_NAME_HIER
    Error in the hierarchy structure: see error log
    Message no. RH003
    Diagnosis
    The structure of the hierarchy contains errors. You can find a detailed description of the error in the log.
    Procedure
    Correct the error in the hierarchy structure as described in the log.
    Screenshot for the Infopackage Details tab,
    The error message suggests checking the error log, but there are no error logs to be reffered.
    Please let us know if anyone have faced similar issue and help us resolving the same. Also let us know where to find the error log in such scenario.
    We have already tried replicating the data source at BW side, but no success. We also checked the extractor at source system side and found it to be working alright.
    Thanks & Regards,
    Viraj

    Hi Viraj,
    Did you find a solution for your problem ?
    Even am facing the same issue .Process chain is running fine without throwing any errors but if i try to run manually it getting failed.
    when we check the data few of the nodes are not getting reflected which are maintained in source.
    Error Message : Error in the hierarchy structure.
    Source system : Extraction check performed and the record count is matching. We took the dump of data and where able to find the required data but in the maintain hierarchies it is not getting reflected
    Could you please help me in the solution.
    Thank you
    Regards,
    Madhav

  • Material Hierarchy data load

    Hi,
    I am loading data for material hierarchy master data but getting the following error message.
    0MATERIAL_LKLS_HIER is the hierarcy transfer structure i am using.
    I have not used any FM like RSAR_DATA_RECEIVED and i am not getting why the error is? Pl help me in resolving this.
    Function module not allowed: RSAR_DATA_RECEIVED
         Message no. B1252
    Diagnosis
         The function module RSAR_DATA_RECEIVED and the application object type
         which were determined are not valid for this IDoc.
    Procedure
         1.  Please check that the process code in the  inbound partner profile
             is correct.
         2.  If this is the case, you should check the ALE inbound methods ALE
             inbound methods for the process code and see whether the specified
             function module and application object type are correct.
    3.  If this is also the case, then the function module and the
        application object type are not permitted for the logical message
        type, message variant, message function and basis type that are
        contained in the IDoc control record. You should check whether the
        correct values have been assigned to these fields in the control
        record. If they do have the correct values, then the assignment to
        the function module and the application object type needs to be
        maintained.
    Full points will be assigned!!
    Thanks & Regards,
    ViJaya

    In the InfoPackage created to load your hierarchy, ensure that you select the correct hierarchy available in the "Hierarchy Selection" tab. All available huierachies for that particular IO that are available in R/3 will show up here. If its blank that would mean there are no hier. maintained in R/3.
    Select Flag for Activation. In the processing tab choose ALE Inbox & InfoObject.
    Hope this helps,
    Regards
    Anujit Ghosh

  • MASTER DATA LOADING  IN SAP BI7.0

    how to load master data attribute, text , hierarchies in sapBI7.0

    check these
    http://help.sap.com/saphelp_nw2004s/helpdata/en/e8/2cd937651f1a13e10000009b38f8cf/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/c8/e92637c2cbf357e10000009b38f936/content.htm

  • How to avoid duplicate data loading from SAP-r/3 to BI

    Hi !
           I have created one process chain that will load data into some ODS from R/3,where(in R/3)the datasources/tables r updated daily.
           I want to scheduled the system such that ,if on any day the source data is not updated (if the tables r as it is) then that data shuold not be loaded into ODS.
           Can any one suggest me such mechanism,so that I can always have unique data in my data targets.
           Pls ! Reply soon.
          Thank You !
           Pankaj K.

    Hello Pankaj,
    By setting the unique records option, you pretty much are letting the system know to not check the uniqueness of the records using the change log and the ODS active table log.
    Also, in order to avoid the problem where you are having dual requests which are getting activated at the same time. Please make sure you select the options "Set Quality Status to 'OK' Automatically" and "Activate Data Automatically" that way you would be having an option to delete a request as required without having to delete the whole data.
    This is all to avoid the issue where even the new request has to be deleted to delete the duplicate data.
    Untill and unless the timestamp field is available in the table on top of which you have created the datasource it would be difficult to check the delta load.
    Check the table used to make sure there is no timestamp field or any other numeric counter field which can be used for creating a delta queue for the datasource you are dealing with.
    Let me know if the information is helpful or if you need additional information regarding the same.
    Thanks
    Dharma.

  • Data load in SAP-BW

    Hi all,
    What is the strategy to load for example 500,000 entries in BW (material master, transactional data)?
    How to separate this entries in small packages and transfer it to BW in automatic?
    Is there some strategy for that?
    Is there some configuration for that?
    regards
    ashwin

    hi
    See OSS note 411464 (example concerning Info Structures from purchasing documents) to create smaller jobs in order to integrate a large amount of data.
    For example, if you wish to split your 500,000 entries in five intervals:
    - Create 5 variants in RMCENEAU for each interval
    - Create 5 jobs (SM36) that execute RMCENEAU for each variant
    - Schedule your jobs
    - You can then see the result in RSA3
    regards
    ashwin

  • Create Document from Data Load and Link to Transaction Record for Long Text

    Hi,
    I have a DBConnect Oracle datasource which contains a large text field.  I would like to build a process that will, as part of the load, create a text file from the content of this large field, upload the file into BW and create the document association with the transaction record.
    Is anyone aware of a HOW-TO to create the BW document entries and upload the files using ABAP?  I thought that I had seen a HOW-TO or instructions approx a year ago, but cannot locate them now.
    Thanks in advance,
    Mel W.

    Hi,
    I hope this is the how to document you were looking for:
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/8046aa90-0201-0010-5e99-962948c83331
    -Vikram

  • Data Loading issue - No local currency found for plant 0001 and sales org

    Hi Guru's,
    I am trying to load data to 0PLANT_ATTR and I am getting the following error message:
    1. No local currency found for plant 0001 and sales organization 0001
    2. Plant 1010 : No sales organization found; Local currency cannot be checked
    I have also maintained the global setting for currencies in RSA1 for that particular Source System.
    Please let me know how to resolve this issue.
    Thanks in advance.

    Hi,
    Refer the lnik and the note mentioned it.
    Re: Problem when storing master data 0PLANT_ATTR-Full (i.e. Full Update)
    With rgds,
    Anil Kumar Sharma .P

  • Hierarchy Data not saved after Hierarchy Load

    Hi Gurus,
    I am facing a problem with the hierarchy data load in SAP BPC NW 7.5 SP06. I have already used /CPMB/IMPORT_IOBJ_MASTER & /CPMB/IMPORT_IOBJ_HIER once to load the data from BI to BPC and it worked fine that time.
    Now I have made some changes in the Hierarchy at the BI side. When I run the package link list, the /CPMB/IMPORT_IOBJ_MASTER chain runs fine and the master data, description for Attribute and Hierarchies are loaded perfectly. But whn the /CPMB/IMPORT_IOBJ_HIER is run, the changes are not saved in the BPC Member sheet. I have manually make the changes in the member sheet and Process the Dimension.
    Can we process the Dimension explicitly after the Hierarchy load again?
    Does anyone have an idea about this?
    Any help is appreciated.
    Thanks,
    Abhishek

    Abhishek,
    Double check whether  /CPMB/IMPORT_IOBJ_MASTER is bringing all the ID's,Texts and Hierarchy Nodes. If required drop the data from dimesion member sheet and load it again.
    Double check the Hierarchy and level setup while running /CPMB/IMPORT_IOBJ_HIER.
    (when we are in SP05, it was not bringing the multiple Hierarchies, now we are in SP07 loading multiple hierarchies but not the TEXT fo the Id's some inconsistency)
    Thanks

Maybe you are looking for