Import ECC data to MDG hub

Hi,
I am using MDMGX tool to import the data(Cost center, Accounts, company, Profit center) from ECC to MDG hub. can I use the same tool MDMGX for importing Cost center group, cost center group hierachy data ? also for Profit centers as well.
Please help me if there is any other way for importing the above data to the MDG hub.

Hi Anusha,
You can use the report KBAS_MDGF_SET_DOWNLOAD to download the data from ECC to your CSV file for MDG objects.
Create a blank CR for holding this data of file that you will be uploading in MDG system by File Upload/Download.
Use the same file to upload in MDG. Complete the approval process.
Best Regards,
Kaustubh

Similar Messages

  • Import ECC Data Structures into PI

    Does anyone know if it is possible to import and ECC data structure for example BFKKZP into PI? Is there a way we can export this as XML and then import that XML into PI?
    We are using PI 7.1 and ECC 6.0.

    Hi,
    you could import the structure via a JDBC sender adapter, but this will not be the same as importing RFC and IDOC.
    You can get the structure of a database in xml via JDBC. You would then need to convert it to an xsd and import as external defintion.
    THis way is somewhat complicated. Instead you could manually copy and paste the structure in SE11 and create a schema out of it in XMLSpy (or other XML editor)
    Hope, this was helpful.
    Best regards,
    Markus

  • MDG-Hub: Tried and tested methods for ECC customizing synchronization

    MDG Experts, I am looking for guidance on synchronizing configuration and customizing data between ECC and an MDG-Hub. I reviewed documentation on (a) synchronizing via BC set import and (b) via TMS although I was looking for a little bit more detail and steps involved in setting up ongoing synchronization between ECC and an MDG-Hub.
    I am looking for experiences and suggestions on this topic. Thanks in advance.

    Hello Karthik
    I have used - BC sets to transport the ECC configuration to MDG. Another way is to transport the tables directly from ECC environment to MDG.
    Check this -
    http://wiki.scn.sap.com/wiki/display/Basis/Business+Configuration+Sets+(BC+Sets)+and+their+use?original_fqdn=wiki.sdn.sap.com
    Kiran

  • Data Services - Open hub read - No Open Hub DTP error

    Hi,
    I have created a open hub destination in BW (7.0 SP21), created transformation, DTP and process chain. In Data Services (3.2) I have created flow, job etc. and every thing seems to be as expected. However when I start the job in DS it terminates with an error:
    1136     5624     BIW-230334     13-11-2009 12:28:22     Process Chain <ZTEST2> of Open Hub destination <ZTEST> at SAP system <.......> no more contains Open Hub DTP. Please reimport the Open Hub destination to execute the appropriate Process Chain.
    I've tried re-activating everything in BW and also re-importing the OHD in BW, but nothing works. I've even tried to re-create everything from scratch. The process chain I've created do indeed contain an Open Hub DTP.
    One addtional question: I've activated the process chain, but do I also need to schedule it in BW (like you do with e.g. process chains for time points to be used in Broadcasting)? Note, I've tried both scheduling/no scheduling and does not work either way.
    Any ideas/suggestions?
    Thanks in advance,
    Jacob

    You must implement following SAP Notes in SAP systems in order to make Data Services Open Hub functionality work correctly:
    SAP Note  1270828 : (Open Hub Process chain is not imported into Data Services repository).
    SAP Note  1294774:  (Open Hub import failed - when you call the module RSB_API_OHS_DEST_SETPARAMS, th system tries to open a GUI)
    SAP Note 1330568 version 3: (This note has fixed "Process Chain <ZTEST2> of Open Hub destination <ZTEST> at SAP system <.......> no more contains Open Hub DTP. Please reimport the Open Hub destination to execute the appropriate Process Chain").
    SAP Note 1079698 - 70SP16 (Enable check box "Automatically Repeat Red Requests in Process Chain" in the DTP Execute)
    SAP Note 1346049:  (Error 028(OSOH): Request <n> cannot be locked.)
    SAP Note 1338465 #with 1293589 as pre-requisite: (The ABAP/4 Open SQL array insert results in duplicate database records.)
    SAP Note 1342955: (0 data records transferred)
    Edited by: Ajit Dash on Nov 13, 2009 10:37 PM
    Edited by: Ajit Dash on Nov 13, 2009 10:39 PM
    Edited by: Ajit Dash on Nov 13, 2009 10:39 PM

  • HCI error - unable to import the data because the row is duplicated in the master data

    Hi, i am working on HCI while i am mapping the data from ECC to HCI target table i am getting this error :Unable to import the data because the row is duplicated in the master data .Could you please help me how to solve this
    Message was edited by: Mariana Mihaylova

    Moved to Process Integration (PI) & SOA Middleware as per New SAP HANA Cloud Integration (HCI) content category (unfortunately, I can't edit the discussion and add the category as I'm not a moderator in this space). Perhaps Mariana Mihaylova can help.
    Susmitha Yerradhodi please read and follow The SCN Rules of Engagement - in particular, the more details you provide, the better the chances of getting your question answered. (e.g. "HCI" is not a helpful subject)

  • ECC Data extraction

    Hello All,
    I have a situation in my department where I need to extract the ECC data to third party database (MS SQL). Initially we were thinking of using SAP BW extractors for extraction through SAP PI to push the data out but this doesn't seem to be feasible. Is this a fair assessment of the situation?
    What other methods we can use for data extraction? We are evaluating options for ABAP programs, SAP Queries and BAPIs. Is there any other way we can extract this data set. The data set is required for sd,fi-ar/ap,mm,pp.
    We donot need data fro real time time basis, but on the same lines of a data warehouse extraction like SAP BW.
    Appreciate your help.
    Thanks

    Hi
    I recently figured out what to do and here is what we do :
    Not many would like to allow you to extract data directly from SAP R/3. But we have been successful in extracting data for our in-house implementation of BI at ORG level. You need to understand which tables of R/3 would give you what information.
    If that information could be gathered. Then extracting data on BODS is very simple. Though it is a bit tricky and it is not direct. But a we do it as a process of different stages.
    We use BODS 3.1
    Step 1 :
    You need to create 2 connections on SAP R/3 Datastore :
    1) conn 1 :
    Use the option generate and execute - ABAP Execution option
    data transfer mtd : direct download
    working directory on SAP Sever : here give local path on which BODS server is running e.g. ; d:\bods
    local directory : : d:\bods
    Generated ABAP directory : d:\bods.
    execute in bg : no
    Create data flow :
    SAPR/3 >>> (sap r/3 table>>>query transformation>>data transport (dat file))
    ||__query transformation >> target database (we have chosen oracle) . You could create a datastore for sql server.
    And Import that table table under this connection.
    Even for SAP R/3. Import that table immediately after you create the SAP R/3 datastore.
    We have used the option of replace file in data transport option in dat file
    So that each time you run the job the file gets replaced.
    Step 2 :
    Create a job for this work flow
    Run the job.
    It will create a dat file and a ABAP File on the specified D: Bods   folder.
    Step 3 :
    Send the ABAP File : to sap bw / sap team - requesting them to create this ABAP program. and make it available on your SAP R/3 System.
    (We found this is easy to do for anyone, all that you need is to copy paste and activiate and do the transport to the sap r/3 system and does not require any time to do any ABAP programming - as such)
    Note : For the data store creations of SAP R/3 : THE user name being used. Need to have access to all the required tables of SAP R/3. Otherwise data access will not happen. You can chek the data flow at each stage and it will say, not authorised. If you dont have access.
    Step 4 :
    Now create another SAP R/3 Connection conn2 ;
    This connection will use the option : execute preloaded (abab execution option)
    execute in bg : no
    data transfer method : shared directory
    working directory on sap server :  d:\bi_share
    (Create a shared folder on sap r/3 server : :\BI_SHARE (the user on data store should have complete access of read and write on this shared folder)
    application path to share directory :
    sapservername
    bi_share
    Step  5 :
    Now if you have the ABAP Prog already available on your R/3 system.
    Re-run the job finally using the new data store for execute - preloaded.
    which is pointing to the shared directory of sap server itself.
    And this job can be scheduled on production BODS server finally.
    ======================================================
    Wherever required, import the tables under the related data store.
    Although it might sound a lengthy process. Actually, it hardly takes a very little time, compared to the time involved in trying to ask someone on ABAP to code / or take help of BW or whatever.
    We have done things on DEV BODS tested the same.
    then we have moved everything into production.
    And till date : it has been successful for us.
    We have taken help of ABAP programers only when we explicity required some customised logical programming. Which was not available directly on SAP R/3 system. 
    Otherwise if we generally pull all the fields once we take any table from SAP R/3 so that i could be used any time later. Even if not required today.
    And SO with a little effort, even a person who is new to SAP R/3  or BW or BODS.... like me, is able to manage. So, it is not difficult i suppose.
    Good Luck.
    We have used oracle for reporting database. And sql server should also work the same way.
    We use sql server only for meta data information.
    like BO Repository or BODS repository
    Cheers
    Indu
    Edited by: Indumathy Narayanan on Jul 27, 2011 8:23 AM
    Edited by: Indumathy Narayanan on Jul 27, 2011 8:24 AM
    Edited by: Indumathy Narayanan on Jul 27, 2011 8:27 AM

  • Import ECC No Through DTW

    Hello Expert
    I want Import ECC No and CE division(Excise Informations)  of  the Party through DTW but no standatd Templetes are  available for ECCNo. How can i do it ?please suggest me ASAP.
    Thanks & Regards
    Bhupendra Shekhawat

    In DTW using maintain interface create a new excel template for tax information and then try to upload excise datas.
    But you need to map exact fields in table while entering datas before uploading.

  • Using SAP ECC as a MDM Hub

    I would like to know the pros and cons of using SAP ECC as a MDM hub instead of using standard MDM tool like SAP MDM.  Many folks in my company thinks that it is  possible by adding some  governance functionality & quality tools to standard SAP ECC. 
        Please  let me know if this is a good idea , if not what are the reasons.
    Thanks

    Hi,
    To answer in general terms, It is like going to a physician for surgery instead of going to a surgeon.
    There is no restriction of exploring ECC for master data & governance. But SAP's EIM solution products like SAP NW MDM are result of growing market requirements for speciality tools. Read more about customer success stories of using SAP NW MDM solutions at - http://www.sap.com/platform/netweaver/components/mdm/index.epx
    Few highlights -
    SAP MDM is designed to be a central hub of master data. MDM will maintain global attributes that can be sent to different target systems including your ECC. Then each of your target systems can maintain their local attributes in additon to the global attributes sent by MDM. MDM will maintain the global unique ID's thus providing a single version of truth for your master data across the organization. SAP NW MDM can be integrated with SAP BOBJ DS and a continuous cleansing/validating/removing duplicates can be achieved. Master data Address cleansing/D&B enrichment can be achieved with real time services present in BOBJ DS & MDM. Enterprise Portal/BPM has standard MDM API's providing data governance capability at its best.
    SAP ECC on other hand is a transactional system/operational system. The load on ECC of maintaining global master data will unncessarily affect the performance. Maintaining master data, removing duplicates, running data vaidations/assignments for cleansing data, then adding central data maintenance with data governance is like adding load on your transactional system. The outmost priority of ECC system is its performance and availability.
    Similar SDN thread - MDM Vs ECC
    Thanks,
    Shambhu.

  • Open Hub: How-to doc "How to Extract data with Open Hub to a Logical File"

    Hi all,
    We are using open hub to download transaction files from infocubes to application server, and would like to have filename which is dynamic based period and year, i.e. period and year of the transaction data to be downloaded. 
    I understand we could use logical file for this purpose.  However we are not sure how to have the period and year to be dynamically derived in filename.
    I have read in sdn a number of posted messages on a similar topic and many have suggested a 'How-to' paper titled "How to Extract data with Open Hub to a Logical Filename".  However i could not seem to be able to get document from the link given. 
    Just wonder if anyone has the correct or latest link to the document, or would appreciate if you could share the document with all in sdn if you have a copy.
    Many thanks and best regards,
    Victoria

    Hi,
    After creating open hub press F1 in Application server file name text box from the help window there u Click on Maintain 'Client independent file names and file paths'  then u will be taken to the Implementation guide screen > click on Cross client maintanance of file name > create a logical file path by clicking on new entiries > after creating logical file path now go to Logical file name definition there give your Logical file , name , physical file (ur file name followed by month or year what ever is applicable (press f1 for more info)) , data format (ASC) , application area (BW) and logical path (choose from F4 selection which u have created first), now goto Assignment of  physical path to logical path > give syntax group >physical path is the path u gave at logical file name definition.
    however we have created a logical path file name to identify the file by sys date but ur requirement seems to be of dynamic date of tranaction data...may u can achieve this by creating a variable. U can see the help from F1 that would be of much help to u. All the above steps i have explained will help u create a dynamic logical file.
    hope this helps u to some extent.
    Regards

  • Extract Data with OPEN HUB to a Logical Filename

    Hi Experts,
    Can anybody help me in sending the link for How to guide...Extract Data with OPEN HUB to a Logical Filename?
    Thanks in advance.
    BWUser

    Hi,
    check this links...
    http://searchcrm.techtarget.com/generic/0,295582,sid21_gci1224995,00.html
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/e698aa90-0201-0010-7982-b498e02af76b
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1570a990-0201-0010-1280-bcc9c10c99ee
    hope this may help you ..
    Regards,
    shikha

  • Not enough space on my new SSD drive to import my data from time machine backup, how can I import my latest backup minus some big files?

    I just got a new 256GB SSD drive for my mac, I want to import my data from time machine backup, but its larger than 256GB since it used to be on my old optical drive. How can I import my latest backup keeping out some big files on the external drive?

    Hello Salemr,
    When you restore from a Time Machine back up, you can tell it to not transfer folders like Desktop, Documents. Downloads, Movies, Music, Pictures and Public. Take a look at the article below for the steps to restore from your back up.  
    Move your data to a new Mac
    http://support.apple.com/en-us/ht5872
    Regards,
    -Norm G. 

  • Comments not imported from Data Dictionary of SQL Server. SDDM 3.3.0.747

    Hi,
    SDDM 3.3.0.747 32-bit on Windows 7 64-bit.
    Comments are not imported from Data Dictionary of SQL Server 2008. Connection through Microsoft JDBC Driver 4.0 for SQL Server or jTDS 1.2.7
    What I have tried? In SDDM DDL generation, Comments in DBRMS for SQL Server are generated with "EXEC sp_addextendedproperty 'MS_Description' , 'Test Comment' ..." so I added extended property named "MS_Description" into SQL Server database, both on table and column. None of them were imported from Data Dictionary into SSDM. I have tried both drivers stated above. Is it a bug or am I missing something?
    I've found similar question thread Re: Data dictionary import doesn't import column comments for SDDM 3.0.0.665, so I guess it is a bug when importing with JDBC drivers.
    MiGli
    Edited by: MiGli_1006342 on May 25, 2013 8:32 AM
    Edited by: MiGli_1006342 on May 25, 2013 9:02 AM

    Extended properties were not imported correctly from SQLServer databases at DM 3.3.0.747.
    Calls to sp_addextendedproperty and fn_listextendedproperty have been modified.
    I don't think it is a problem with JDBC drivers.
    A bug fix should be available in the next release of DM.

  • SQL modeler can not import from data dictionary

    It was very frustruted to see that the SQL Modeler hang in import from data dictionary of a database as part of revise engineering. I have to question myself if sql modeler is a serious tool and should I give up.

    I am not sure if Data Modeller is still in Beta./Production. First couple off initial versions of a new product are normally buggy.
    Regards
    Alternatively, If this product is still in Beta, then you can contact the development team or report the issue so that they can take care of this issue in next beta release.
    Edited by: skvaish1 on Mar 30, 2010 3:18 PM
    Edited by: skvaish1 on Mar 30, 2010 3:26 PM

  • Crystal Report Using ECC Data has Auth error in Infoview

    We've created a Crystal report that successfully accesses ECC data from a function module in our Dev environment when the report is executed from within Crystal Designer.  When we try to execute the same report against the same development ECC system from within InfoView we get the following error:
    ERROR: The Database Logon Information for this report is either incomplete or incorrect. 
    The connection information is the same as it is in designer. We are able to successfully run Crystal Reports over BW data within InfoView. Any ideas about what might be wrong?
    Doug

    Ingo,
    The consumers of the report do have ECC access for order and inventory management so licensing should not be an issue.
    We were able to resolve the issue. The problem was with our version of SAP GUI on the BO Dev server.
    We had to update our u201CServices.dllu201D and our u201CSAPLogonTree.xmlu201D file with an accurate version.  Once that was done, the report executed as expected in InfoView using LDAP.
    Thanks,
    Doug
    Edited by: Douglas Eberle on Jun 13, 2011 6:54 PM

  • Import XML data to textfields in a PDF file

    Can I use LiveCycle Designer 8.2 to auto populate textfields in a PDF-file with data from a XML-file when the user opens the PDF-file? If so, how will I do it? Is there any guide?
    It works fine in the Livecycle Designer preview but when I open the PDF-file all textfields are empty. It also works fine if I import data from a OLEDB-connection.
    Regards
    Mikael Forssell

    I'm looking for a solution for the same thing. I have a pdf form with fields bound to an xml file, but Acrobat will only import the data manually (Forms / Manage Form Data / Import Data) and Reader doesn't work at all.
    Tech Support says that when in Livecycle, you're working in an xml environment, so the 'Preview PDF' works. Acrobat and Reader aren't xml environments, so they won't display the data.
    What we need is a javascript programmer who can figure this out. Possibly using the remerge function. Anyone???
    Don.

Maybe you are looking for

  • Back up data on iMac

    while using Imac and OS X Mountain Lion, and wanting to back up data on Time Machine, I am led to the net.....then searching for airport and finally the Mac regretfully announces " No configured airport base station have been found". Does it mean the

  • Bridge not working in CS2 or by self

    All of the sudden when I click on browse in CS2 the bridge doesn't pop up. In fact It wont pop up on it's own as well. I have tried re install on both and nothing changes. Any ideas?

  • ABAP short dump error while creating an invoice

    Hi All, We are creating a delivery(Unit is EA) for the material for which the Unit Conversion is maintained as 1CS = 48 EA. So while creating a delivery If I maintain the delivered quantity is less than 24, I am not able to create an invoice.It's thr

  • Bundled Runtime Installer

    The instructions for setting up the bundled runtime installer, to install AIR and my application from a CD, seem straightforward enough. Although, I'm having trouble getting the installer to automatically install my application after the AIR runtime

  • "Could not complete the Blending Options Command because of a program error"

    I'm using CS5, PhotoShop 12.04 x64. I received this error message this morning when I tried to use the blending options. My software is in another state, so reloading is not an option right now. Can anyone suggest another way to address this problem?