Syndication process - Tags in Data manager Selecting the map in Syndication

Hello Experts,
I am supporting a repository in MDM 7.1, and there is a part of the syndication process that i don´t know how does it work. This is the process what i mean:
The repository works with 3 different maps, the map is selected in  MDM Data Manager belong a 'syndicatable' tag. Two of the maps are selected inside Qalified tables.
What i need to know is where could i see the configuration for that automatic mapping (it is performed belong MDSS), could anybody help me?
Thanks in advance,
Mariano Pinkava.

Hi Mariano,
In addition, please go through page 223 (mds.ini) & 272 of Console Reference Guide.
Syndication File Location:
When MDSS completes a syndication to a port, it places the syndication file in the portu2019s Ready folder. The Ready folder is part of the following fixed directory structure, located beneath the MDM Serveru2019s distribution root directory:
root/DBMSinstance_DBMStype/RepositoryName/Outbound/ RemoteSystem/PortName/Ready
where:
u2022 root is the distribution root directory (set in mds.ini).
u2022DBMSinstance is the network identifier used to specify the DBMS instance name and DBMStype is the four-character identifier for the DBMS type (i.e. MSQL, ORCL, IDB2).
u2022RepositoryName, RemoteSystem, and PortName are the values entered in the Code property for each item in MDM Console.
Thanks,
Priti

Similar Messages

  • Hi. I am reinstalling Macbook Pro OS X from the "question mark icon". During process, there is a question "select the disk where to install Mac OS X" but there is no option/selection in the box. I really need your help guys. I need it badly. :'(

    Hi. I am reinstalling Macbook Pro OS X from the "question mark icon". During process, there is a question "select the disk where to install Mac OS X" but there is no option/selection in the box. I really need your help guys. I need it badly. :'(

    If you got the question mark and you do not see any partition to install OS X to, it is not good.
    Close the OS X installer by pressing the red button at the top left corner of the Install OS X window, and open Disk Utility. Then, look at the left sidebar and tell me if you see your hard drive. You should see something like this:
    Do not worry if you do not see Macintosh HD or if the option with an arrow shows a different name. This is normal.
    If you see it, choose it, go to the Erase tab and erase the drive. Finally, close Disk Utility and reinstall OS X. You will lose everything if you do not have a backup.
    If you do not see the drive, it is damaged and it has to be replaced. You can replace the hard drive of a MacBook Pro without voiding the warranty, and any 2'5" SATA hard drive is valid. If you are not an experienced user or you do not want to go inside the MacBook, take it to an Apple Store or reseller.

  • How is index data managed within the cluster?

    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    Peter

    user11279467 wrote:
    Hello
    We are in the process of evaluating Coherence.
    With regard to indexes, am I correct in thinking that the data for the index is stored in an (internally managed) cache on the same node as the attendant indexed values? I'm not interested in getting at the index data, but rather just understanding how the index data itself is treated/stored within the cluster.
    So in a 2 node cluster where an index is applied to a distributed cache that only has 2 entries, will there be an index cache on each node that has 1 {Binary,key} entry where the key points to the value that is stored on that same node? That is, I trust that the index data is spread out across the cluster in the case of a partitioned topology?
    Regards
    PeterHi Peter,
    up to 3.4.x indexes were supported only on partitioned topology, and yes, indexes on each node contain only data corresponding to entries which have their primary copy on that node (referred to as local entries from now on).
    The indexes are not stored in a cache but they are managed as part of the information Coherence maintains about the backing map (the map which contains local entries).
    The indexes are made of 2 parts:
    - forward index: this is a mapping from the cache key in an internal representation (backing map keys from now on) to the value extracted with the extractor used to create the index from the entry belonging to the cache key (extracted value from now on)
    - reverse index (aka. reverse map): this is a mapping from the extracted value to a set of backing map keys of entries from which the reverse index key was extracted with the extractor used to create the index
    The index can be sorted which means that the reverse index will be a SortedMap (so sorting order is the ordering of extracted values)
    Let's see an example. Assuming you have the following cache content within a particular storage-enabled node :
    A --> Object (getX=5 ,...)
    B --> Object (getA=3 ,...)
    C --> Object (getA=3 ,...)
    D --> Object (getA=4 ,...)
    The forward index in that node will contain:
    Binary(A) --> 5
    Binary(B) --> 3
    Binary(C) --> 3
    Binary(D) --> 4
    The reverse index will contain:
    3 --> { Binary(B), Binary(C) }
    4 --> { Binary(D) }
    5 --> { Binary(A) }
    Where Binary(...) means the internal (binary) representation of the ... object.
    If the index is not ordered, then the order of iteration on entries in the reverse index are not deterministic.
    If the index is ordered, then the iteration of entries in the reverse index will be as shown above. Also, you can cast the reverse index to SortedMap so you have the very useful headMap and tailMap and firstKey and lastKey methods.
    Hope this helps.
    Best regards,
    Robert

  • Firefox won't start properly. The process begins in task manager, but the program doesn't start.I have restarted the computor, and updated my plug-ins. Don't know what to do next. It will open in safe mode, but not regularly.

    When we click on the icon to start firefox even though the process opens in task manager, the program screen doesn't open

    If it does work in Safe-mode then disable all extensions and then try to find which is causing it by enabling one at a time until the problem reappears.
    * Use "Disable all add-ons" on the [[Safe mode]] start window to disable all extensions.
    * Close and restart Firefox after each change via "File > Exit" (Mac: "Firefox > Quit"; Linux: "File > Quit")
    * https://support.mozilla.com/kb/Troubleshooting+extensions+and+themes

  • Compare Dates and select the max date ?

    Hello,
    I am trying to write a script and will compare the dates in " eff_startdt" and give me the lastest date at the outcome.
    I have data that some service locations have more than one contract date and I need to get the latest dated conract dates to work on the data but I tried many things I am always getting errors. When I run the script below I get " missing expression" error. I dont' see anything missing.
    also somehow Max() is keep giving me errors when I do something like [  ON service_locs = vmeterid WHERE SERVICE_LOCS = SERVICE_LOCS AND EFF_STARTDT = MAX(EFF_STARTDT)  ]
    Can someone pls give me advice on this. Thanks
    SELECT DISTINCT Broker, customer_name, service_locs, fee_kwh, qtr_monthly, eff_startdt, eff_enddt
    FROM VMETER
    INNER JOIN BROKER_DATA
    ON service_locs = vmeterid WHERE SERVICE_LOCS = SERVICE_LOCS AND (SELECT MAX(EFF_STARTDT) FROM VMETER)
    -----------------------------------------------------------

    Hi,
    I will try to explain on my example. I have got a table:
    DESC SOLD_ITEMS;
    Name                                    Value NULL? Type
    COMPONENT                                          VARCHAR2(255)
    SUBCOMPONENT                                       VARCHAR2(255)
    YEAR                                               NUMBER(4)
    MONTH                                              NUMBER(2)
    DAY                                                NUMBER(2)
    DEFECTS                                            NUMBER(10)
    DESCRIPTION                                        VARCHAR2(200)
    SALE_DATE                                          DATE
    COMP_ID                                            NUMBERI have insert example data into my table:
    select component, subcomponent, sale_date,comp_id
      2  from sold_items;
    COMPONENT       SUBCOMPONENT    SALE_DAT    COMP_ID                            
    graph           bar             06/04/03          1                            
    graph           bar             06/04/01          2                            
    search          user search     06/04/02          3                            
    search          user search     06/04/01          4                            
    search          product search  06/03/20          5                            
    search          product search  06/03/16          6                            
    graph           bar             06/05/01          7                            
    graph           bar             06/05/02          8                            
    graph           bar             06/05/02          9
    As you can see there are a few components and subcomponents duplicated with different date and comp_id value.
    I want to get component and subcomponent combination with latest date.
    SELECT COMPONENT, SUBCOMPONENT, MAX(SALE_DATE)
      2  FROM SOLD_ITEMS
      3* GROUP BY COMPONENT, SUBCOMPONENT;
    Efect:
    COMPONENT       SUBCOMPONENT    MAX(SALE                                       
    graph           bar             06/05/02                                       
    search          user search     06/04/02                                       
    search          product search  06/03/20
    For your purpose I will do it using join and subquery. Maybe it will help you resolve your problem:
    SELECT COMPONENT, SUBCOMPONENT, SALE_DATE, RANK
      2  FROM (SELECT T1.COMPONENT, T1.SUBCOMPONENT, T2.SALE_DATE,
      3          ROW_NUMBER() OVER (PARTITION BY T1.COMPONENT, T2.SUBCOMPONENT ORDER BY T2.SALE_DATE DESC) AS ROW
      4          FROM SOLD_ITEMS T1, SOLD_ITEMS T2
      5          WHERE T1.COMP_ID = T2.COMP_ID)
      6* WHERE ROW = 1;
    I joined the same table but it act as two different tables inside subquery. It will group values (partition by statement) and order result descending using t2.sale_date column. As you can see columns are returned from both tables. If you would like to add some conditions, you can do it after WHERE ROW=1 code.
    Results:
    COMPONENT       SUBCOMPONENT    SALE_DAT       RANK                            
    graph           bar             06/05/02          1                            
    search          product search  06/03/20          1                            
    search          user search     06/04/02          1Hope this help you
    Peter D.

  • LiveCycle 2.6.1 Data Management with The ColdFusion 8.0 DataManagement Event Gateway Issue

    Hello all,
         I've recently been developing a project that involves sending out events from ColdFusion to LiveCycle 2.6.1 using the Data Management event gateway to Flex 4.0 clients (LiveCycle and ColdFusion are on different Instances, but the same server).  To begin with, I used ColdFusion assemblers, DAO's, and models and everything worked fine locally.  After deploying this setup to a beta site, I decided that this setup would be very troublesome in terms of configuring clustered instances across multiple servers.  I then decided to convert my assemblers, DAO's, and models to Java.  The conversion went well and the flex clients see the exact same data as they did with the ColdFusion adapter.
         Once I tried to send an update through from my ColdFusion application to a Flex client, I get an error stating that:
    "Unable to find the Flex adapter for destination My_Dest in the RMI registry on localhost:1099.The Flex adapter may not be running or the destination may be incorrect."
    After seeing this error, downloaded a Java-based RMI inspector to see what was going on.  To get a good idea of what was happening when the ColdFusion adapter was being used, I switched my data-management-config file back to the CF adapter.  I noticed that the RMI entry was as follows:
    localhost:1099/cfdataserviceadapter/My_Dest
    localhost:1099/cfassembler/my_cf_instance
    Once I gathered this data as the base, I converted back to the Java adapter in my data-management-config file, restarted the servers, and ran the RMI inspector again.  Only the "localhost:1099/cfassembler/my_cf_instance" was showing.  (This one shows because I have "Enable Remote Adobe LiveCycle Data Management Access" checked in my CF instance's CF Admin -> Flex Integration).  Since I don't need this checked anymore, I unchecked it and re-ran the RMI inspector.  As it should, the "localhost:1099/cfdataserviceadapter/My_Dest" went away.  Since no destination shows up, it means that the Flex adapter isn't registering my "my_Dest" destination with RMI.  Since it isn't registered, I can't see it when I try to send a message through the CF Data Management event gateway.
    Can anyone help me out here?  I certainly may be missing something when it comes to RMI (I don't work with Java very often).  Any advice would be greatly appreciated!
    Thank you,
    Dustin Blomquist

    Dustin,
    Without the ColdFusion based data management destination defined on the LCDS server, the destination will not show up in the RMI registry.  It is only the CF adapter code that does this.  The 'stock' LCDS adapter does not support invoking via RMI the way the CF version does.
    I would recommend you run the LCDS MessageBrokerServlet inside the ColdFusion web application.  This will give you two things:
    1. You will not have the overhead of RMI between CF and LCDS as they will share the same VM (better performance!).
    2. You will be able to use the CF Data Management Gateway to pass messages to Java-based destinations.  The APIs the gateway uses should work fine with either CF or Java based Data Management destinations.
    The CF/LCDS integration doesn't support what you are trying to do when you run two seperate instances.

  • Is there a way for GPS coordinates to override a 'places' tag when viewing photos on the map?

    I like to have 'places' tags for all of my photos in Elements Organizer so I can filter photos without the map view.  The problem is that some of my photos already have GPS information from my camera.  When I tag a photo that has EXIF geo data with a 'places' tag, it moves the photo on the map to the less specific lat/long of the tag and no longer uses the lat/long of the photo itself.  I would like to use the data EXIF data from the photo, if it exists, rather than the lat/long from the places tag.
    Alternately, it would be cool if Elements would auto-create places tags based on the lat/long from a photos EXIF data.  That would also solve my conflict between places tags and EXIF data from the photo. 
    Any ideas?  Am I out of luck?

    Why don't you just put a passcode lock on the phone so no-one else can use it when you're not around?
    You can look at the history in Safari, but that won't prevent someone from clearing the history and cookies after they use it.

  • Maximum IMD size that can be processed by Meter Data Management v2.x

    Does anyone know the maximum IMD size that MDM 2.x can process? I was trying to process an MV90 Interval IMD with 15 minute cuts for a period of 6 months (trying to load historical usage in one go). MDM is timing out and failing to process the IMD. The IMD has been created and is failing when I am trying to do Additional Processing (with just a dummy VEE group so no 'real' validation is happening).
    Edited by: user12984550 on Jun 4, 2012 1:15 PM

    You have to options:
    1. The router that is internal must have a static route to the ASA to reach the VPN networks and must have a distribute static so that other routers that form part of EIGRP know how to route to the VPN networks.
    2. You can configure on the ASA "set reverse-route" on the crypto map then configure EIGRP on the ASA and add redistribute static so that routes learned via VPN (considered static routes) can be pushed through EIGRP.

  • HT1414 How do I sync my device to a different device's library?  Do I have to restore to factory settings and start over? Or, is there a simpler process where i can just select the library/device that I want to sync with?

    Is there an easier way other than restoring to factory settings and starting over?

    Open itunes, connect ipod, select what you want, sync.
    No need to restore.  Just sync.  It will erase the current content and replace with the content of the new computer.

  • MPxIO, traffic manager, select the active path

    I have one lun with 2 paths, and I use MPxIO without loadBallance,.
    NOTE: I have v240 server with solaris 10 the latest that connected to the SAN.
    Two Question :
    1. How I can see which of the paths is the active path?
    2. Is the MPxIO can understand which path(from 2 paths to the lun) is the path througth the default controller, and make this path to e the active path?
    thanks

    For that array, Hitachi needs to provide a specific string to enter into the scsi_vhci.conf file, telling MPXIO how to deal with that array.
    It should be in the array Release Notes or Install Guide, along with any particular levels of microcode and array port settings needed. The scsi_vhci.conf entry will provide the string that a host scsi inquiry will return, along with the specific scsi options that the OS will need to talk to the array using multi pathing. Once those changes are done usually a boot -r will be required.
    Until that is done, Solaris will see both paths to the array as seperate and distinct arrays, so no multi pathing will be done.
    HTH
    Dom

  • MDM Data Manager step-by-step process with an example

    Hi All,
    I recently started reading MDM. I am clear with MDM-Console concepts. But I have problem with MDM Data Manager. I am confused with data entries in qualified tables, work flows(using microsoft studio), Matching mode- rules(High treshold, low threshold) etc.. I read material posted in forums- but still I am not clear.
    I need a step-by-step process for data entries with *examples*. A good example which covers all Data Manager concepts. PDF's with screen shots, videos any thing would help.
    Please help me......It would be of great help.
    Thanks in Advance.
    Suchir

    Hi Suchir ,
    MDM 5.5 has 4 MDM clients to perform 4 main functionality on the MDM master records.
    - Console- This is where all the Administrative work is performed
    - Data Manager- This is where the actual maintainence work is done on the master data like consolidation deduplication etc.
    - Import manager- Thsi is from where the source matser records are taken into MDM repoistory
    - Syndicator- This is from where teh consolidated master data is send back to the target systems.
    MDM Data Manager is the heart of MDM
    The activities that can be performed in the MDM Data manager are:
    - Data Consolidation
    - Data Validation
    - Data Deduplication
    - Data Assignment
    - Data Governanace
    etc......
    Qualified Table:
    - Qulaified tables are used to store relational records where one field value is determining the other field or fields value.
    - Qualified tables are maintained in Console and can be viewed in Data Manager
    - Qualified Table are viewed under a separate section in Data Manager in the right hand side.
    - Qualified table has Non qualifiers and Qualifiers.
    - Non qualifiers can be seen in the  Qualified table and in the linking Qualified field
    - Qualifiers can be seen separately under the Qualified section in the data manager
    Kindly refer the below link to know more on this:
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/00a15239-684e-2b10-b8ae-b936b7d1c1fe
    Workflows
    - Workflows are designed in MDM for Governanace
    - You can create the MDM workflow using the MS Visio 2003 stencils in the MDM Data manager in the record mode under the workflow table
    - The saved workflow in DM can then be executed either mnaully or automatically based on teh trigger actions
    - The trigger action for teh MDM wf are record Add,Update,Import,Manual
    Kindly refer the below link to know more on this:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/60559952-ff62-2910-49a5-b4fb8e94f167  (MDM Workflows Overview)
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/90990743-91c0-2a10-fd8f-fad371c7ee40  (Demo of MDM Workflows)
    Matching Mode
    - This mode is exclusively for Master data deduplication
    - We need to create Rules,Strategies,Transformation in MDM data manager in teh record mode
    - Then you need to run these strategies on the selected records
    - Based on your score and threshold set in the strategy MDM will identify duplicate records
    - This score will be seen in teh Matching mode
    - Colour coding is used along with scores to find out dupliactes
    - Green - 100%dupliacte Blue- 50?% Red- No duplicate
    Kindly refer the below link to know more on this:
    /people/michael.reil/blog/2006/05/18/mdm-matching-strategies-for-master-data-consolidation (matching merging in mdm)
    Hope It Helped
    Thanks & Regards
    Simona Pinto

  • Not able to see data in the qualifier table of the main tbl , Data Manager

    Hi,
    I have an issue of not able to see the data of two qualified table after populating them.
    It is in mdm-5.5 ps4.
    When populating data first time ,it shows up in those two table slots in the right side of the Data Manager.
    However subsequently it does not show up in those slots , only by right click on the table and selecting "View/edit", the window pops up where those data shows up.
    However unlike other qualified tables the data does not showup automatically for these two tables.
    Appreciate any suggestion or feedback on this.
    regards,
    -reo

    You may have checked the Filter Check Box next to the Qualified Lookup cell in Data Manager, when the current table is the Main Table.
    You use the Filter Checkbox to limit the qualified table records by the current search selections.
    Secondly, you have see if there are any Qualified Links to the main table record you are viewing.
    If not, create the Qualified links in Data manager, for the main table record and the Qualified Table Record.
    Once this is done, you will see the Display fields of the Qualified table for which the links exists for the given main table record.
    Message was edited by:
            Adhappan Thiagarajan

  • Sort all the Records in Data manager based on Update Date

    How to sort the reocrds in Data Manager based on Date or something ?
    And also I have three records for a same customer in ( for three different company codes ) and when I search for this customer with the Customer number then it is showing three records , but when I search with Update Date then no records are fetched ...what is the reason ?

    If you make a field as type "Time Stamp" in your data model using the MDM Console, make sure that the option Sort Index is set to Normal.  If this is the case, whenever a record is updated in the repository, the date will be updated.  If the sort index is normal, then MDM will allow you to sort the records based on the date and time by which they were updated.  You can do this by finding your time stamp field in the MDM data manager and clicking on it.  If there is an up and down arrow next to the name of your field, it means it can be sorted.  Also, to answer your other question, you can definitely search based on date.  Use the Free-Form Search on the left hand side of the data manager at the bottom of the screen.  Simply select a date, and it will show you all the records updated on that date.

  • How to view the key mapping when viewing Hierarchy in Data Manager

    Hi! I previously posted a question on doing value mapping using the hierarchy code rather than the name. I have got that working now using concept of remote key. When I first loaded the hierarchy with remote key and check in Data Manager, I can see the hierarchy structure with the key mappings. However, when I subsequently went back to the hierarchy to have a second look, the keys are gone. Does anyone know how I can view hierarchy in Data Manager with the key mappings?
    Cheers!
    SF

    Hi,
    Go to Record mode, and open Hierarchy table. Now to select click on any hierarchy node record and then right click and select "Edit Key Mappings". This will show you existing keys as well as allow you to manually modify keys. You can only Check Remote keys on hierarchy in Record mode and in Hierarchy Mode "Edit Key Mappings" option will be disabled.
    Please revert if you face any issue.
    Regards,
    Shiv

  • How to import the data to Data manager

    Hi All,
    I am new to MDM and need your help in handling the following issue
    In our setup we are sending the Customer master data from R/3->XI->MDM (R/3-> XI IDOCs,  XI->MDM XML files).
    with the help of extraction job in R/3 IDOCs r generated and XI has onverted these IDOCs to XML Files and created at MDM Server. Now my Question is how to import the Data from these XML files to Data Manager !!!!
    Note : Import Map settings are done already in the system.
    Can any one help me the steps requeired to get this data to Data manager .
    and am I need to do some activity in MDM to import the data to Data manager each and every time or is there any automatic way of doing this .
    Please guide me.
    Thanks in advance!

    Hi MDM User,
    Generally there are 2 cases while importing the data,
    1) Initial Load
    2) Delta Load
    Initial Load means importing the data into MDM at first time.This is usually done by manual importing
    without using import server.
    Delta Load means importing the data into MDM when we have changes to existing data or added new data
    to existing data.This is usually done by automatically using import server.
    To manually import the data,just create a remote system and port in console and login into import manager
    using remote system.Create ,save the map and execute the map.There is no other configurations needed.
    To automatically import the data without manual intervention,we go for import server.
    To let this happen we need to set some configuarations in mdis.ini file of import server.
    Give the repository name,username and password which are set in console for that repository and also
    give schedule time based on which import server picks the file from ready folder.
    Go to the import manager , change the map and import action depends on your requriment and save that
    map.
    and also you need to set port type as inbound and automatic and select the map which you saved earlier in
    MDM console.
    As soon as we start the import server, it picks the file from ready folder of MDM server based on
    configurations which we set in mdis.ini file and executes the map.
    Here i am giving complete path of ready folder.
    SAP MDM 5.5> Server>Distributions>bp1bocap080.bp.co_MSQL> BP3_PoC_Customer> Inbound>Siebel> SIEBEL_OB_CUS_SIEBELCUS01> Ready)
    This path has to be given in target directory of receiver communication channel of XI so that it puts XML
    file in the above path.
    For better understanding,go through below blogs.
    /people/balas.gorla/blog/2007/02/05/r3-xi-mdm-outbound-scenario
    /people/balas.gorla/blog/2006/09/27/mdm-xi-r3-integration
    Hope it helps
    Reward points,if found useful
    Thanks
    Narendra

Maybe you are looking for