Data creation in data manager vs. ERP transaction

Hi experts,
I have a general question on working with MDM: When I use MDM as central data management tool then I create and change the whole data within MDM Data Manager or maybe in MDM iViews in EP. Doing that way I will loose a lot of positive aspects I had doing this in ERP formerly even when using business content.
For example:
- no process logic exists (which I had in ERP dynpros in the code): no preconfigured assignments, validations, relationships, etc.
- I have to know all relevant tables where I have to do my entries
- no standard searches for typical problems
- all roles and authorizations have to be created on my own (except standard roles like Admin, Data Expert, etc.), there is no possibility to transfer authorizations from ERP to MDM (authorizations can be very complex)
- you need more (expert) know-how to maintain data (which is also a job of business departments)
- etc.
So a change to central master data management means a step backwards from a dynpro controlled input GUI to a database frontend with only rudimental features and loosing a lot of functionality that was build within a lot of years.
Is this right or did I miss something important? I think loosing all this functionality can be a cause for resistance against introduction of MDM central data management in companys.
Thanks for your answers. Helpful answers will be rewarded.
BR, bd

Hi BD,
CMDM is one of the later stages in using MDM effectively.Before moving to a CMDM scenario it is necessary that ground base is set up.
- One goes for a CMDM scenario once all the one time data(exisiting data) is already cleansed.So firstly you need to extract all the possible high risk duplicates existing data from the ERP sytem into MDM ,run the matching merging strategies on them and then maintain the consistent data in MDM.Once this step is completed only then it makes sense to go for the creation of new data centrally through MDM.
- Just like in ECC when a user creates a new material for instance.He has to go through all the views and enter in all the required and optional fields value.In the same way this same scenario can be replicated in MDM by creting a repository with all the fields that one needs to enter while creating a new material. and by using MDM Validations we can simulate the same requirement of mandatory as in ECC.
- If you go for a Business content repository for Material,Vendor,Product etc you will have all the ready to use Roles with authorizations so its no a much of rework.
- Standard searches amy not be available in MDM but one can always desig his/her own search based on the customer requirement and MDM searches are very easy to use and also very dynamic.
I do agree that SAP MDM may not be a fully grown tool at this point of time but with the SAP MDM 7.1 vs most of these drawbacks will be addressed. and besides although ECC can do most of the Master data related work it is not a dedicated system for that purpose and will therefore have an influence on the performance and time involved aspects.
Hope It Helped,
Kindly Reward Points if found useful
Thanks & Regards
Simona Pinto

Similar Messages

  • Extracting Data from SAP ERP using BODI/Data Services 4.0

    HI,
    I am trying to extract data from SAP ERP via SAP extractors using BODI/Data Services 4.0.
    I do not have my own ERP system so I am renting remote access from one of the many available on the internet.
    I am able to connect BODI to the ERP system and import the extractors metadata.
    The problem I am experiencing is that when I run job to extract the data I get the following error:
    Vendor-supplied function module <Z_AW_RFC_READ_EXTRACTOR> not found. Ensure that you can execute the function module in SAP via transaction /nSE37.
    How do I create the function? Or is the function a SAP standard function?
    SAP ERP system being used is: ECC6 EHP4
    User has SAP FULL and DEVELOPER authorizations.
    Any assistance would be appreciated.

    You might have better luck in the (somewhat misnamed) [Data Integration and Data Quality Management|Data Services and Data Quality; forum:
    This forum is dedicated to topics related to SAP BusinessObjects Data Services (Data Integrator, Data Quality Management, Text Data Processing), SAP BusinessObjects Information Steward (Metadata Management, Data Insight), SAP BusinessObjects Rapid Marts and SAP BusinessObjects Data Federator.
    (emphasis added)
    Regards,
    Sean

  • Governance and Data Management

    Hello,
    We need to know about governance and data management, from this week I am working as master data coordinator in Empresas Polar (Mass Consumer Products Industry), where the role of this function is to assure the creation and maintenance of the master data, as initial activities we are raising statistics to know the most important information of the master data in all SAP modules, and the idea is across own developments in R/3, Portal, Workflow and Tools that exist on the market to simplify Update of the information on the part of the user, we know that in SAP the information is created in different views or transactions, which makes very difficult the opportune update of the data.
    I am interested to know about MDM or any tool/application which eolud help us,
    I appreciate your help,
    Thanks

    Hi torres,
    The MDM system for master data management and single-source cross-media catalog publishing does more than just database management, asset management, or even database publishing.It uses an industry-standard SQL-based DBMS as a u201Ccontaineru201D for the repository database. But then it goes much further to implement an elegant and seamless combination of many different technologies,concepts, and minute details to create a comprehensive solution to an
    age-old problem.Many of the features incorporated into the MDM system are unique.
    ADVANTAGES
    End-to-end solution. The MDM system provide an end-to-end
    solution that automates the entire process of managing master data from start to finish, including bulk data import, centralized master data management, and published output to a variety of media.
    u2022 Database-driven system. MDM layers a thick shell of functionality on top of a powerful SQL-based DBMS so that the MDM system is fully scalable and the master data is fully accessible to other SQLbased applications and tools.
    u2022 Large capacity. The MDM system efficiently manages master data repositories containing up to millions of records.
    u2022 Superior performance. MDM breaks through SQL performance
    bottlenecks to deliver blazingly fast performance that is measured in milliseconds rather than seconds and is literally 100u20131000 times that of a SQL DBMS alone. No other system on the market today delivers comparable performance.
    u2022 Powerful search and retrieval. All of the MDM modules include
    powerful search and retrieval capabilities, so that an entire repository of thousands or millions of items can be easily searched and any item or group of items located in a matter of seconds.
    u2022 Cross-media publishing. The MDM system is the only one on the market today that includes tightly integrated functionality for both electronic and printed output, for Web, CD-ROM, and paper output from a single MDM repository.

  • Classes in Master Data Management

    Hello All,
    Why Classes are used in Master Data management.What is the Concept of Distribution Class for distributing Master
    Data. Kindly help me to understand the Concept.
    Regards,
    Hazam

    Cookoid
    You comment regarding 'changing the nature of authorisations' is an interesting one because as you have already pointed out and is consistent with the way the product(s) work; you cannot change any more than your SAP credentials allow you to change.
    All that said, quite rightly; the idea of being able to perform mass changes on ANYTHING in your SAP system of record should not be taken lightly and due diligence should be applied in determining who should have such tools and under what circumstances they should be used. Design and testing in QA or pre-production environments is an absolute must and this discipline should be applied with the same level of enthusiasm and commitment as you would any objects that you develop for productive use even if they are LSMW or regular BDC's.
    The product supports a controlled release approach if implemented in a particular way, using centralised license management and workflow controls around the scripts and data templates and workbooks. Scripts can be restricted to particular SAP instances, particular volumes of data, particular transactions, particular times of day/week/month for execution.
    Additionally you have two types of licenses, author/developer licenses and runner-only licenses. This avoids you landing up in a situation where inexperienced users can create their own automation objects, in such circumstances these users would typically only be given a runner license.
    Some customers don't choose to use this governance and compliance approach though and choose to buy only the authoring tools, accordingly they essentially leave the door open to the creation of potentially harmful objects.
    Again exhaustive testing is always recommended. WIth over 1,000 customers globally there are plenty of ASUG member references and you should actually be able to get referrals to existing customers who can tell you how they have introduced and manage controls around the use of the products.

  • Integration Oracle Utilities CC&B w/ non-Oracle bespoke Meter Data Management system

    Hello all,
    In your opinion, what's the best way to integrate a bespoke (non-Oracle) Meter Data Management system with Oracle CC&B 2.3.1 or 2.4?
    I came up with a few possible scenarios and would love to hear your opinion:
    1. Use CC&B Meter Read upload (MUP) batch process to upload reads from our bespoke Meter Data system.
    2. The bespoke Meter Data Management system adds Meter Reads to CC&B using a CC&B Inbound Service.
    3. At Billing time, CC&B invokes a Get Consumption Algorithm (invoked before the Bill Segment Creation Algorithm) which is responsible for requesting the reads from our bespoke Meter Data Management system.
    This is the approach used by the Oracle CC&B-Oracle MDM integration when it comes to Usage requests/transactions.
    4. Any other idea? Feel free to share.
    Thanks

    Hello all,
    In your opinion, what's the best way to integrate a bespoke (non-Oracle) Meter Data Management system with Oracle CC&B 2.3.1 or 2.4?
    I came up with a few possible scenarios and would love to hear your opinion:
    1. Use CC&B Meter Read upload (MUP) batch process to upload reads from our bespoke Meter Data system.
    2. The bespoke Meter Data Management system adds Meter Reads to CC&B using a CC&B Inbound Service.
    3. At Billing time, CC&B invokes a Get Consumption Algorithm (invoked before the Bill Segment Creation Algorithm) which is responsible for requesting the reads from our bespoke Meter Data Management system.
    This is the approach used by the Oracle CC&B-Oracle MDM integration when it comes to Usage requests/transactions.
    4. Any other idea? Feel free to share.
    Thanks

  • Preview transformation file in data manager package

    Dear BPC Experts,
    When we try to preview Tranformation file while running data manager package to import transaction data from BW, we are getting following error. We do not get this error if we use load from flat file package.
    We are on BPC 10 PS06, EPM Add-in SP14 patch3.
    Has anybody seen this issue before? We can paste the entire log if required.
    See the end of this message for details on invoking
    just-in-time (JIT) debugging instead of this dialog box.
    ************** Exception Text **************
    System.ArgumentException: Separator cannot be null and must contain only one char
    Parameter name: separator
       at FPMXLClient.DataManager.CsvParser.Parse(String data, String separator, Boolean hasHeader) in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\DataManager\CsvParser.cs:line 15
       at FPMXLClient.DataManager.UI.Forms.FilePreview.BuildDataArrayFromCsv(String data) in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 487
       at FPMXLClient.DataManager.UI.Forms.FilePreview.BuildDataArray(String data, Boolean formatted) in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 414
       at FPMXLClient.DataManager.UI.Forms.FilePreview.SpecialFilesProcessing() in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 406
       at FPMXLClient.DataManager.UI.Forms.FilePreview.DisplayData() in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 351
       at FPMXLClient.DataManager.UI.Forms.FilePreview.InitializePreview() in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 102
       at FPMXLClient.DataManager.UI.Forms.FilePreview.FilePreview_Load(Object sender, EventArgs e) in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\DataManagerUI\Forms\FilePreview.cs:line 740
       at System.Windows.Forms.Form.OnLoad(EventArgs e)
       at FPMXLClient.UILayer.Forms.BaseForm.OnLoad(EventArgs e) in d:\Olympus_100_REL_XLCLIENT\src\FPMXLClient\src\UILayer\UI\Forms\Base\BaseForm.cs:line 70
       at System.Windows.Forms.Form.OnCreateControl()
       at System.Windows.Forms.Control.CreateControl(Boolean fIgnoreVisible)
       at System.Windows.Forms.Control.CreateControl()
       at System.Windows.Forms.Control.WmShowWindow(Message& m)
       at System.Windows.Forms.Control.WndProc(Message& m)
       at System.Windows.Forms.ScrollableControl.WndProc(Message& m)
       at System.Windows.Forms.Form.WmShowWindow(Message& m)
       at System.Windows.Forms.Form.WndProc(Message& m)
       at System.Windows.Forms.Control.ControlNativeWindow.OnMessage(Message& m)
       at System.Windows.Forms.Control.ControlNativeWindow.WndProc(Message& m)
       at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
    Best Regards,
    Ashwin.

    Hi Raju,
    Thank you for your reply.
    It seems that it is SP related issue. When we downgrade our EPM Add-in to SP13 patch4, it did not throw any error.
    Best Regards,
    Ashwin.

  • Error in process chain when running package (data manager)

    Hello experts,
    I try to run the following how to : "move date between applications using Data Manager ..."
    Yon can find the pdf here : http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/b0480970-894f-2d10-f9a5-d4b1160be203
    I'm looking after someone who used it successfully, on my side after implementing all steps and running the package I get the message "IMMEDIATE RUN: Request to run the package on the server was successful The package is now running"
    When I look the log for the process chain, overall status of the process chain is yellow, and steps "modify dynamically" and "clear BPC Tables" are in error.
    I can't get any usefull error messages.
    If someone has an idea to have details about those errors, it will be appreciated.
    Points will be awarded, thanks in advance.
    Guillaume P.

    Hello Experts,
    I am also facing the same issue with PC failing at Modify_Dynamically step....
    This process type is working fine when used  in standard delivered process chain to load transaction data.
    I tried using this variant in my cutsom chain, but it fails.
    Please suggest.....
    Thx....D

  • Debugging LCDS Data Management from Flexbuilder3 on Tomcat does not work

    I have a sample with Data Management and when i use it on my tomcat it works fine. But i can´t debug it from flexbuilder 3. When i start it from the flexbuilder i get the connection and then i crashed with "No registered fault handler or token responder - throwing an error for destination:"
    It is the same tomcat i use to run standalone.
    is it a problem with the context.xml?
    [SWF] /flexine/dataService_wt2.swf - 1.422.965 bytes after decompression
    'cds-consumer-sandwich-null' consumer set destination to 'sandwich'.
    Configuration for destination='sandwich':
    New DataService for destination: sandwich
    DataService.fill() called for destination: sandwich with args: []
    '4A853118-F24B-5FE2-BF01-FCDFF6C93161' producer set destination to 'sandwich'.
    Creating a new independent data store for destination: sandwich
    Adding data service: sandwich to the data store: null initialized: false
    Finished validating destination: sandwich loadOnDemand/paged associations: [] sub-types:
    data store: null is initialized
    'my-rtmp' channel endpoint set to rtmp://localhost:2038
    'my-rtmp' channel settings are:
    'ds-producer-sandwich' producer sending message '52AF7949-7B16-7C8E-3B87-FCDFF6F8D3A2'
    'my-rtmp' channel got connect attempt status. (Object)#0
      code = "NetConnection.Connect.Success"
      description = "Connection succeeded."
      details = (null)
      DSMessagingVersion = 1
      id = "E35ECC7A-710F-4D9B-A507-330FDD83B283"
      level = "status"
      objectEncoding = 3
    'my-rtmp' channel is connected.
    'my-rtmp' channel sending message:
    (mx.messaging.messages::CommandMessage)
      body=(Object)#0
      clientId=(null)
      correlationId=""
      destination="sandwich"
      headers=(Object)#0
      messageId="CDC66303-F24B-1E87-4103-FCDFF8211298"
      operation="client ping"
      timeToLive=0
      timestamp=0
    'my-rtmp' channel sending message:
    (mx.data.messages.DataMessage)
      messageId = '52AF7949-7B16-7C8E-3B87-FCDFF6F8D3A2'
      operation = fill
      destination = sandwich
      identity = (null)
      body = []
      headers = {}
    'ds-producer-sandwich' producer connected.
    'ds-producer-sandwich' producer acknowledge of 'CDC66303-F24B-1E87-4103-FCDFF8211298'.
    'ds-producer-sandwich' producer acknowledge of '52AF7949-7B16-7C8E-3B87-FCDFF6F8D3A2'.
    'ds-producer-sandwich' producer fault for '52AF7949-7B16-7C8E-3B87-FCDFF6F8D3A2'.
    Dispatching fault event for destination: sandwich
    No registered fault handler or token responder - throwing an error for destination: sandwich
    [RPC Fault faultString="There was an unhandled failure on the server. javax/transaction/SystemException" faultCode="Server.Processing" faultDetail="null"]
         at mx.data::ConcreteDataService/http://www.adobe.com/2006/flex/mx/internal::dispatchFaultEvent()[C:\depot\flex\branches\enterprise_corfu_b2\frameworks\projects\data\src\mx\data\ConcreteDataService.as:2401]
         at DataListRequestResponder/fault()[C:\depot\flex\branches\enterprise_corfu_b2\frameworks\projects\data\src\mx\data\ConcreteDataService.as:6970]
         at mx.rpc::AsyncRequest/fault()[E:\dev\3.0.x\frameworks\projects\rpc\src\mx\rpc\AsyncRequest.as:103]
         at NetConnectionMessageResponder/statusHandler()[E:\dev\3.0.x\frameworks\projects\rpc\src\mx\messaging\channels\NetConnectionChannel.as:523]
         at mx.messaging::MessageResponder/status()[E:\dev\3.0.x\frameworks\projects\rpc\src\mx\messaging\MessageResponder.as:222]

    Here's an update for future victims doing google searches :-)
    The issue appears to be related to compiz. When I'm doing basic 2D desktop with metacity everything simply rocks: the machine suspends/resumes with no issues.
    When compiz is enabled, there is about 10 second delay after resume, during which the screen is black and you better not touch the keyboard! If I sit and wait for 8-12 seconds, the dark screen goes away GDM login dialog shows up. But if I touch the keyboard or the touchpad, it will either freeze or become exceptionally slow.

  • Default File Selection while Running Data Manager Package

    Hi BPC Experts,
    What i wanted to know is that is possible to make a file selection as "Default" for running a data manager package.Say for example i wanted to run a simple DMP "Load Transaction data from from BW InfoProvider UI" and the file that i always select is "ZFIGL010" for this is it possible to get  "ZFIGL010" file as pre-populated (or pre-selected)instead user going and choosing from long list of files available "Info Provider Selection".
    Thanks
    Rahul

    Rahul,
    As Ashish said you cant get rid of the info provider name selection as you are using the /CPMB/LOAD_INFOPROV_UI pkg. But if you use /CPMB/LOAD_INFOPROVIDER package you can follow the below steps to get rid of the selection.
    Comment the below prompt statement in the Advance script of the pkg which asks you to select the Info provider.Its the first statement of the script. You can comment by putting a ' before begining of the statement.
    PROMPT(TEXT,%InforProvide%,"Please enter InfoProvider to download data",)
    Add a INFO statement and pass the same INFO statement to the Task. Please refer to image below.
    Regards
    Nikhil

  • Scheduling a data Manager Package

    Hi,
    Please le t me know how to schedule the Data manager package from BPC Client.
    I tried th option in Data Manager Package "Schedule" but is not repeating the job. It just run once and stop.
    Anurodh

    Hello,
    Please check the schedule for the InfoProvider you are using. You can check this within transaction code RSA1. There may not be a schedule that was created to load data. Please verify that this is or not the case. Thanks.
    Regards,
    Jason

  • BPC 7.5 NW -- Data Manager Import Fails When Loading 40,000 Records?

    Hi Experts,
    Can't believe I'm posting this because the Data Manager component has always been one of the best parts of BPC.  But, since we got SP04 applied last week, every IMPORT process that I've run that has more than 40,000 records fails out.
    The result logs show that the CONVERT task completes just fine, but it doesn't really show the LOAD task.  ...Not exactly sure what's going on here.  So far I've also taken the following two steps to try for resolution:
    (1.)  Re-added the IMPORT package in Organize Package List from the Library to have a "fresh" version.  Didn't help.
    (2.)  In the "Modify Package" screens, there is a PACKAGESIZE parameter that is 40,0000 by default...  I've been able to locate that in BI using transaction RSA1 and have changed it to 10,000,000.  Saved it.  Tried it.  Didn't help either
    Has anyone seen this kind of behavior before?
    Thanks,
    Garrett

    Update -- This problem may now be resolved.
    I have been able to conduct test IMPORTs of 48,000, then 96,000 and then 1.7 million records.  All were fine.
    It turns out that that difference is that the text files were sorted by amount in the ones that failed.  They were sorted by GLAccount in column A for the ones that succeeded.
    Edit:  Yep, all files loaded normally when re-sorted by GLACCOUNT, etc. on the left-hand side.  Apparently, when you're doing a lot of records that might confuse the system or something
    Edited by: Garrett Tedeman on Nov 18, 2010 11:41 AM

  • POS Data Management

    Hi folks,
    first I have to say that I'm a newbie in SAP but I have to try out a scenario with a connection from SAP Exchange Infrastructure (XI) to SAP POS Data Management (I think included in BW 3.5). Task is to transfer POS Data, stored in a XML-file, into the transaction database in the PIPE (POS Inbound Processing Engine).
    My questions:
    - Is there a special POS workbench in SAP BW?
    - How can I write the data into the transaction database?
    - Is there more information in the web about SAP POS Data Management?
    Regards,
        Josef

    Hello,
    have you set the "COMMIT"-Flag in the structure to 'X' for a commit work per package or used the function module for BAPI commit work?
    Stefan

  • Passing dynamic data from UWL to transaction iview

    Hi Experts,
    Can some one please suggest how can we take data from UWL task to a transaction iview on click of button.
    I have a webdynpro java sceen which is coming as a task in UWL of a user. When user clicks on Button with complete event, i have to get data from the Webdynpro Sceen and set data to a transaction iview. This means i cant har code the data in Application parameter of transaction iview.
    Catches here are:
    1. Passing data from UWL to Transaction Iview.
    2. Data is dynamic i.e. is entered by user through web dynpro screen.
    Thanks in advance.
    Regards
    Pranav

    Hi Scott,
    Thanks for reply. But i am not understanding the solution provided by you. I think these are ABAP FM and are used user management. I am working in Netweaver environment CE7.1.1.
    Regards
    Pranav

  • What is master data management

    i want know clearly abut MDM in SAP.

    Hi naresh reddy  ,
    SAP NetWeaver MASTER DATA MANAGEMENT
    SAP NetWeaver Master Data Management (SAP NetWeaver MDM) is an enabling foundation for enterprise services and business process management -- providing a single version of the truth for customer, product, employee, supplier, or user-defined data objects.
    Working across heterogeneous systems at disparate locations, SAP NetWeaver Master Data Management ensures cross-system data consistency through interactive distribution. It integrates business processes across the extended value chain, delivering features and functions to enable:
    Master data consolidation -- Consolidate master data for companywide analysis and reporting. SAP NetWeaver Master Data Management consolidates and cleanses master data objects from disparate systems. After consolidation, it stores information from different systems in a centralized repository. Related rich content can be included to augment the data store.
    Synchronization and distribution of master data -- Enable consistent data maintenance and distribution to ensure permanent harmonization of master data. Using global attributes, you can ensure that all systems receive the same master data during distribution -- and enrich the distributed data objects with additional attribute values in the target systems. Distribution can be controlled, visible, and traceable, with active status management at each distribution step.
    Centralized management of master data -- Supports companywide quality standards by ensuring that central control of master data begins as soon as the data is created. Centrally created master data can subsequently be distributed to client systems as required using interactive distribution.
    Administration of master data -- Manage master data without custom code. A powerful interface supports administrative tasks such as data exception handling and assignment of role-based access to business processes and information. Data managers use the interface to configure data source merging, business rules, and distribution details to downstream applications.
    Management of internal content -- Collect and centralize all your content -- including parametric information and rich content such as images, paragraphs of text, PDF documents, and organizational intelligence about content -- in an enterprisewide repository.
    Catalog search -- Deploy intuitive interfaces that help you locate items internally, publish Web catalogs on e-commerce storefronts and in supplier enablement programs, and integrate easy-to-search catalogs into e-procurement solutions -- all from a centralized repository, and all at speeds surpassing normal SQL-based queries.
    Print catalog customization -- Disseminate product information directly from a centralized catalog repository to popular desktop publishing programs, and automatically generate fully formatted and populated page layouts.
    Multichannel syndication of product catalog content -- Publish restructured and reformatted extracts or incremental updates of your product catalog content -- and distribute them to trading partners in several delimited text and XML formats -- on an unscheduled or regular basis.
    Business process support -- Enable communication in a heterogeneous environment, and insert master data into other systems.
    Business analytics and reporting -- Leverage synchronized data for reliable analysis and accurate reporting.
    The following websites will clearly explain u wt is MDM. These websites contain PDF & PPT presentations:
    SAP Netweaver MDM Overview
    http://www.asug.com/DesktopModules/Bring2mind/DMX/Download.aspx?TabId=66&DMXModule=370&Command=Core_Download&EntryId=3431&PortalId=0
    MDm
    http://www.asug.com/DesktopModules/Bring2mind/DMX/Download.aspx?TabId=66&DMXModule=370&Command=Core_Download&EntryId=1666&PortalId=0
    SAP Netweaver MDM Overview
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b09b548d-7316-2a10-1fbb-894c838d8079
    SAP NETWEAVER MDM Leverage MDM in ERP Environments - An Evolutionary Approach -
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/4059f477-7316-2a10-5fa1-88417f98ca93
    Master Data Management architecture patterns
    http://www-128.ibm.com/developerworks/db2/library/techarticle/dm-0703sauter/
    MDM and Enterprise SOA
    http://www.saplounge.be/Files/media/pdf/Lagae---MDM-and-Enterprise-SOA2007.10.10.pdf
    Effective Hierarchy Management Using SAP NetWeaver MDM for Retail
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/70ee0c9e-29a8-2910-8d93-ad34ec8af09b
    MDM World
    http://mdm.sitacorp.com/
    MDM: Master Data for Global business
    http://www.sitacorp.com/mdm.html
    MDM Master Data Management Hub Architecture
    http://blogs.msdn.com/rogerwolterblog/archive/2007/01/02/mdm-master-data-management-hub-architecture.aspx
    Improve Efficiency and Data Governance with SAP NetWeaver MDM
    http://www.sapnetweavermagazine.com/archive/Volume_03_(2007)/Issue_02_(Spring)/v3i2a12.cfm?session=
    Data Modeling i MDM
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/5d4211fa-0301-0010-9fb1-ef1fd91719b6
    http://www.sap.info/public/INT/int/index/Category-28943c61b1e60d84b-int/0/articlesVersions-31279471c9758576df
    SRM-MDM Catalog
    http://help.sap.com/saphelp_srmmdm10/helpdata/en/44/ec6f42f6e341aae10000000a114a6b/frameset.htm
    cheers!
    gyanaraj
    ****Pls reward points if u find this helpful

  • Central Master Data Management

    Hi All,
    Can anyone tell me about central master data management ?
    In what way it is difference from consolidation and harmonization?
    I have gone through several documents but i did n't get any idea on that topic.
    If anyone have knowledge on CMDM, Please explain in their own words.
    Thanks
    Narendra

    Consolidation and Harmonization procedures are very different from CMDM (central master data management), however often times CMDM is not very useful without considering the other two as well. 
    Consolidation essentially means creating identity around your master data across your landscape.  More specifically you want to identify duplicate records and merge them into one record.  Harmonization is the process of pushing the new cleansed data back out to your partner systems.
    Central master data management is the process of creating and managing enterprise level attributes in one place.  One common mistake that people make is that central master data management means having one place to create your records and containing ALL data.  This is not correct, MDM focuses on enterprise level attributes.  In other words, which attributes are KPI's, shared across multiple processes in scope, shared across multiple systems, important for reporting, need ultimate quality, etc.  Once you have determined your data model then you begin developing the workflow (if needed) around the creation of master data.  This way you can easily keep your data clean and free of duplicates moving forward following a consolidation and harmonization process.

Maybe you are looking for

  • Errors in archive shipping in a standby environment 10.2.0.4

    Hi, I've problem with a standby configuration with real time apply. For month everything works fine, then one day the log shipping from primary to standby fails with error 3135 and sometimes with 1041 and 12571. There's no change in the configuration

  • About T500 battery issue

    I bought a T500 in Jan, 2009. So far my battery only has 53 cycles. But it only could last at most 20min. Sometimes, the gauge shows 60% full, but the next minute it shows less than 1% full and shut down automatically. Although I bought a 3-year warr

  • [SOLVED]can't compile from aur

    /usr/bin/ld: main.o: undefined reference to symbol 'XPending' /usr/bin/ld: note: 'XPending' is defined in DSO /usr/lib/libX11.so.6 so try adding it to the linker command line /usr/lib/libX11.so.6: could not read symbols: Invalid operation collect2: e

  • Elements 10 did not install properly but blocked my computer - need help, please!

    I've purchased an Elements 10 software, the requirements are met with my computer (My computer has 4G RAM), yet, after I had installed it and clicked on the icon for the first time, it blocked my computer (it became VERY slow for everything, like hal

  • Submit Statement with layout details

    HI Experts, I am using a Sumbit statement to the Standard report 'ME2L'. Also i have created a Layout for this standard report in the output. Please let me know how to trnasfer this layout name in the Sumbit statement. Thanks, Kothai