ObjectGUID as matching attribute for reconciliation between AD and IdM

Hello together,
I want to use the AD attribute objectGUID as matching attribute between AD and IdM. It is one of the attribute which will never be changed (in case of name changes of a person or similiar changes).
Our IdM can read this attribute and save it (with the help of a java script). This works fine.
But if I want write back something to AD I do not know how the "To LDAP directory" path must be configured. I get always the error that the account can not be found in the AD. Maybe the attribute must be changed with a java script ...
Has anyone already used this attribute?
Thanks!
BR
Michael

Michael, I have done this many times in pretty much the same way you have indicated.  While DN is good for basic LDAP / AD operations, ObjectGUID is preferred for ModRDN operations or if you want to change the user's DN.  I have written about these operations in a blog, Using modRDN with SAP NW IDM.
Hope it helps!
Regards,
Matt

Similar Messages

  • Can any one send procedure for integration between XI and Geographic inform

    can any one send procedure for integration between XI and Geographic information System(GIS).
    Please clarify ASAP..
    Thanks

    Hi,
    Please see the below links
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/d53af8ca-0801-0010-08b0-dce9e02130a8
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/d53af8ca-0801-0010-08b0-dce9e02130a8
    GIS VS XI
    Regards
    Chilla..
    Points rewarded if it is usefull..<i></i>

  • Reference Library for Converting Between LabVIEW and XML Data (GXML)

    Please provide feedback, comments and questions on the Reference Library for Converting Between LabVIEW and XML Data (GXML) in this thread.
    The latest version of the NI GXML Library is availble in VIPM on the NI LabVIEW  Tools Network repository.

    Francesco, Thank you for the feedback.  With this component it was my intention to make a more "terse" version of the LabVIEW Flatten to XML VI that was also supported on RT and that gave the user more flexbility regarding the structure of the parsing type definition. I think you are right that the XML parser is not compliant to section 2.11 of the XML spec.  The parser does specifically looking for a #D#A and this appears to be an oversight on my part.  Please confirm for me, the specifcation is saying that the XML parser should be able to recognize three possibilities as an "end of line" character: #D#A, #D, or #A.  Am I reading this right?There are more efficient (and in some cases much more efficient) ways of sharing data between LabVIEW and LabVIEW: some examples are flattened binary strings and the datalog binary format.  XML is slower than these optons but the upside is that it is human readable.  Furthermore XML is inherently hierarchical which is convenient for complex data structures like clusters of arrays of clusters, etc.  If you don't care about human readability then you are correct XML doesn't make as much sense.I will return to the GXML source code and try to fix this in the near future but I would hope that instead of creating yet another custom VI from scratch that you could reuse what I have provided for you.  I included enough documentation in the source code so that users could make some modificiations themselves. The target application for this reference library was LabVIEW to LabVIEW communication.  As such I documented the schema on the dev zone document from a LabVIEW perspective.  It includes all the supported datatypes and all the supported data structures (cluters, arrays, multidimensional arrays, clusters of multidimensional arrays, etc.)  I do see some value in making a more conventional XML spec but the time investment required didn't really line up with my intended use case. Were there any other downsides to GXML that I have missed?  Best Regards, Jeff TippsSystems Engineer - Sound and VibrationMessage Edited by Jeff T. on 04-21-2010 10:09 AM

  • Data Reconciliation between BI and R/3 systems

    Hi Gurus,
    I want to know how to reconciliation the data between BI and R/3 systems.
    is there any easy mentods to reconcile the data.i have also gone through the HOW TO GUIDE document, that document dosen't help me out.
    Regards
    Sreenivas.Y

    1) Either we can reconcile using standard R3 tables eg VBAK , VBAP for sales .
    2) You can go for automatic reconcillation where you will have to create a reconcillation data source and extract it to BW using virtual cube (Virtual cube would extract real time data from R3).
    Include original cube and this virtual cube in a multiprovider .
    Make a report taking keyfigures from both the infoproviders and add a column for difference .If there is 0 difference it means BW data matches with R3 .

  • Master control attribute value discrepency between Master and child org.

    Hi Guru,
    We are facing a strange issue in Track install Base Attribute values. There are some 60K item having discrepency in the value of the this attribute between master and child.
    We are wondering how it happens as this attribute is master controlled one and we never changed this attribute control and kept as Master controlled. The master value is true
    If you guys experience this issue, please let us know the solution that you had applied to fix this issue.
    regards
    venkat

    Try executing the vo1 query before setting the bind variable for vo2, then execute the vo2 query (if that's what you want to do).

  • Data Inconsistency for 0EMPLOYEE between ECC and BI

    Hi,
    We do a full load to 0EMPLOYEE using 0EMPLOYEE_ATTR from ECC. There were records deleted for lot of employees (some action types) in ECC. This has caused data inconsistency for 0EMPLOYEE master data (time-dependent) between ECC and BI. In BI we have more records for these employees (Additional records have time-dependent range that were deleted from ECC but still exist in BI). These employee records are already being used in lot of InfoProviders. Is there an efficient way to fix this issue? One of the solution is to delete data from all InfoProviders and then delete from Master data for 0EMPLOYEE, but the deletion of employee records can happen quite often, so we don't want to take this route. Also, I tried to re-organize master data attributes for 0EMPLOYEE through process chain, but that didn't work either.
    Message was edited by:
            Ripel Desai

    Hi Ripel,
    I share your pain. This is one of the real pains of Time-Dependant master data in BW. I have been in your exact position and the only way round the issue for me was to clear out all the cubes that used 0EMPLOYEE and then delete and re-load 0EMPLOYEE data.
    I know this responce doesn't help you but at least you are not alone.
    Regards,
    Pete
    http://www.teklink.co.uk

  • Issue for integration between GL and AR AP

    Hi All,
    We have problem with integration real time between GL and AR AP follows:
    - now system automatic transfer realtime Data from AR or AP module to GL. but my customer don't want transfer automatic realtime because on AR or AP module may enter data incorrect at the end moth they chage and transfer to GL.
    pls let me now how to disable transfer automatic this function.
    Thanks for support

    Hai
    I think Hold documents function will very much suits to your rqquirement.Any way Hold documents will not have any impact in the financial transactions and will not even post in GL accounts.
    when ever u feel that the dodument is completley correct u can post the hold documents.
    Regards
    Raghavendra.M
    SAP-Practice

  • Data conversion for communication between Java and C/C++ program

    The real problem, i guess, is about data type conversion between Java and C programs. For instance, and int is supposed to be 4 bytes in Java, and also in C/C++. But, as far as I know, the size of and int in C depends on the processors architecture, so if it were a 64 bit arch., size of and int, double, etc. in C, would change. The real scenario would be a communication between a 32 bit machine (Java) and a 64 bit machine (C/C++).
    First of all, is this assumption correct?
    If so,
    how is it possible to deal with this problem in Java?
    Is there any way to know 'type length' automatically? or
    would it be necessary to modify the Java program to work with a C program in 32-bit or 64-bit arch?
    thx in advance

    cotton.m wrote:
    Yes you should develop the C part of the protocol first and then build from there.
    See http://forum.java.sun.com/thread.jspa?threadID=5243547 for a previous discussion of this topic.
    I understand what you mean, also the topic you referenced. I will explain the situation a bit more. The protocol is already defined: (16 bits message size, 16 message id, 32 bit time, etc.); it is also defined in C (structures and so on). So it would be easy to write my Java code using the specified protocol in bytes, and C types wouldn't be necessary to be considered. But, let's say that the C code can't be modified, and I thought (and don't know yet if its correct), that size of C types CAN changed depending on architecture. So if size changes for C types (code is specified in short, int, etc, and not in bit format), it all would be a mess, if I don't implement something in my Java code.
    cotton.m wrote:
    I don't believe that 32 or 64 bit enters the equation here at all.Maybe 32 o 64 doesn't mind, but sparc or PC could take a point here. Communication would be between a sun solaris sparc and windows x86 PC. But if type size in C/C++ does not change, I wouldn't really have any problem then...
    The situation is not ideal, I know, but is what I have; and I think I could have problems if I run the C program in other machines; so I would have to try to solve it whithin my Java Program.
    I hope it is clear now what I need.
    thx
    Edited by: MGasa on Jan 16, 2008 1:51 AM

  • Solution for connection between BODS and WCF web service ?

    Hi,
    we have a unique requirement of connecting BODS to WCF webservice. This web service acts as a layer for connection between proprietary databases (secured) and other external systems. Queries are to be sent to this web service in order to fetch data from proprietary databases.
    In order to connect to these web services, we should also provide Address, Binding and contract whose details are already present with us.
    Example of web service address:     net.tcp://op12345:10101/sd
    I heard Adapters are the best way of connecting to the above web service, could you please provide some guidance on connecting through adapters if it is the best solution.
    Could someone suggest the best possible approach(ASAP) to bring connection between BODS and web services and to execute the above requirement.
    Please let me know in case of any more information.
    Regards
    Surya

    Hello Nir,
    I tried to import the my VS2008 + .net 3.5 as Wsdl local file to the Xcelsius 2008 but there is message unable to load url.
    I tried to import your Wsdl developed by .net:
    http://www.webservicex.net/airport.asmx?WSDL
    and it works ok.
    Can you send me or print here the service class definition and the configuration file for  http://www.webservicex.net/airport.asmx?WSDL?
    Thanks in advance,
    Sergey Osmachkin

  • GUID for relation between BP and Sales Org?

    Hi,
    I need a GUID for the relation between a business partner and the sales org. I already have a GUID for the sales org, and one for the BP, but I also need one for the relation itself in addition to the field RelNr in BUT050.
    What table are these stored in?
    Regards
    Thomas

    Hello Manuel,
    in neither of these tables is a field for a GUID for the relation, there is only one GUID for the business partner set.
    Since my GUID does not really need to be globally unique, but rather locally, I will now use the FM BUPA_MWXT_RELKEY_TO_RELGUID (which basically concats both partner numbers and the partner relationship number).
    Thanks for your help, though.
    Thomas

  • Any soft on android for iMessages between android and ios

    I m using my ipad3 and now got a new samsung s5 so dm stuck on using imessages begoz I liked it very much
    help someone if hv any info

    Check the App Store for an app that allows texting between SMS and  your iPad. The app provider takes the SMS message and turns it into a iMessage.

  • Keyboard shortcut for toggling between belgian and russian keyboards ?

    I want to switch between "belgian" and "russian" keyboards by a keyboard shortcut while using a text editor (NeoOffice, Pages, ...). How can I do that: macro, applescript, ... ?
    Message was edited by: jjeener

    The standard shortcut is Apple/Command + space.  If that doesn't work you may need to disable it being used already by Spotlight.  You can set any shortcut you want in system prefs/keyboard/keyboard/shortcuts

  • How to syncronize data between resource and IdM?

    I created a form (assign to proxy administrator) for reconciliation, and it gets executed when I imported new user into IdM through full reconciliation. In this way I can assign email of user to waveset.email.
    But later when the email gets updated in resource directly, how can I propagate the new value to IdM (without activeSync). By running a full reconciliation again doesn't seem to re-run the proxy administrator's form. I always thought full reconciliation would do the trick, but only now I realize it does not work at all !!?
    What should I do then to synchronize the data?

    Recon was not really designed to automatically synchronize attributes. Typically, this is implemented via ActiveSync. With that said, there is an option in Recon to "Attribute-Level Reconciliation". This allows you to provide a list of resource attributes to watch and if there is a change, Recon calls a workflow, where you can execute any changes you would like.
    I believe this workflow is called "Audit Native Change To Account Attributes". Look at wfrecon.xml located in /sample.
    Edited by: R1100SLA on Dec 21, 2007 12:12 AM

  • Retain Pattern Match Attribute for Next Pattern Match Calculation

    When a MATCH_RECOGNIZE PATTERN match occurs, is there any way to retain the underlying event for the next pattern match?
    It's like I want to replay the underlying event (that triggered the match) into the next "batch" of events.

    Here is an approach
    1) Use a (local) cache to hold the "lastMatchedTime"
    2) Augment each input event with the "lastMatchedTime" by first (left outer join) joining with cache
    3) Run the MATCH_RECOGNIZE on the augmented stream
    4) Use a match to update the cache
    Here are the queries that I tried -
      <view id="S">
            <![CDATA[
                 RSTREAM (
                    SELECT
                           id,
                           obs,
                           epoch,
                           nvl(C.lastMatchedTime, 0) as lastMatchedTime
                    FROM
                           InputChannel[now] AS I LEFT OUTER JOIN lastMatchCache AS C
                    ON
                           C.keyValue = "key"      
            ]]>
          </view>
          <view id="v1">
            <![CDATA[                 
                      SELECT
                              T.id,
                              T.obsTotal,
                              T.elapsedTime,
                              T.firstts,
                              T.lastts
                      FROM S MATCH_RECOGNIZE (
                              PARTITION BY id
                              MEASURES
                                        id as id,
                                        sum(obs) as obsTotal,
                                        B.epoch - decode(B.lastMatchedTime, 0L, min(epoch), B.lastMatchedTime) as elapsedTime,
                                        decode(B.lastMatchedTime, 0L, min(epoch), B.lastMatchedTime) as firstts,
                                        B.epoch as lastts
                              PATTERN ( A+? B )
                              DEFINE
                                       B AS (epoch- decode(lastMatchedTime, 0L, min(epoch), lastMatchedTime)) > 20*60*1000
                      ) as T
            ]]>
          </view>
          <query id="qcache">
             <![CDATA[
                 SELECT
                       "key" as keyValue,
                       lastts as lastMatchedTime
                 FROM
                       v1
             ]]> 
          </query>
           <query id="q1">
             <![CDATA[
                 SELECT
                 FROM
                        v1
             ]]> 
          </query>
        Also, include in the config file
    <channel>
           <name>CacheChannel</name>
           <selector>qcache</selector>
       </channel> 
       <channel>
           <name>OutputChannel</name>
           <selector>q1</selector>
       </channel>The EPN that I used was
      <wlevs:event-type-repository>       
            <wlevs:event-type type-name="CacheEvent">
                <wlevs:class>com.bea.wlevs.event.example.helloworld.CacheEvent</wlevs:class>
            </wlevs:event-type>
            <wlevs:event-type type-name="InputEvent">
               <wlevs:properties>
                  <wlevs:property name="id" type="int"/>
                  <wlevs:property name="obs" type="int"/>
                  <wlevs:property name="epoch" type="bigint"/>
               </wlevs:properties>
            </wlevs:event-type>
            <wlevs:event-type type-name="OutputEvent">
               <wlevs:properties>
                  <wlevs:property name="id" type="int"/>
                  <wlevs:property name="obsTotal" type="int"/>
                  <wlevs:property name="elapsedTime" type="bigint"/>
                  <wlevs:property name="firstts" type="bigint"/>
                  <wlevs:property name="lastts" type="bigint"/>
               </wlevs:properties>
            </wlevs:event-type>
        </wlevs:event-type-repository>
        <wlevs:adapter id="InputAdapter" provider="csvgen">
             <wlevs:instance-property name="port" value="9061" />
             <wlevs:instance-property name="eventTypeName" value="InputEvent" />
             <wlevs:instance-property name="eventPropertyNames" value="id,obs,epoch" />
        </wlevs:adapter>     
        <wlevs:channel id="InputChannel" event-type="InputEvent" >
            <wlevs:listener ref="helloworldProcessor"/>
            <wlevs:source ref="InputAdapter"/>
        </wlevs:channel>
        <wlevs:caching-system id="local-caching"/>
        <wlevs:cache id="lastMatchCache" name="lastMatchCache" key-properties="keyValue" value-type="CacheEvent">
           <wlevs:caching-system ref="local-caching"/>
        </wlevs:cache>
        <!-- The default processor for OCEP 11.0.0.0 is CQL -->
        <wlevs:processor id="helloworldProcessor">
           <wlevs:listener ref="CacheChannel"/>
           <wlevs:cache-source ref="lastMatchCache" />      
        </wlevs:processor>
        <wlevs:channel id="CacheChannel" event-type="CacheEvent">
           <wlevs:listener ref="lastMatchCache"/>
        </wlevs:channel>
        <wlevs:channel id="OutputChannel" event-type="OutputEvent" advertise="true">
            <wlevs:listener>
                <bean class="com.bea.wlevs.example.helloworld.HelloWorldBean"/>
            </wlevs:listener>
            <wlevs:source ref="helloworldProcessor"/>
        </wlevs:channel>And the output that I get seems to match your requirement -
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=9, v1.elapsedTime=7228000, v1.firstts=1199311210000, v1.lastts=1199318438000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=17, v1.elapsedTime=30935000, v1.firstts=1199318438000, v1.lastts=1199349373000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=7, v1.elapsedTime=41436000, v1.firstts=1199349373000, v1.lastts=1199390809000
    eventType=OutputEvent object=q1  v1.id=4143, v1.obsTotal=13, v1.elapsedTime=5904000, v1.firstts=1199390809000, v1.lastts=1199396713000Edited by: Anand Srinivasan on Sep 30, 2010 12:49 AM
    Edited by: Anand Srinivasan on Sep 30, 2010 12:50 AM

  • Best practice for migrating between environments and versions?

    Hi to all,
    we've got a full suite of solutions custom developed in SAP BPC 7.0, SP 7. We'd like to understand if
    - there are best practice in order to copy this applications from an environment to another environment (another client)
    - there are best practice in case the client has got a newer version of SAP BPC (they should install the 7.5, while we're still stucked with the 7.0).
    Thank you very much
    Daniele

    Hi Daniele
    I am not entirely sure, what you are asking, Please could you provide additional information.
    Are you looking for best practice recommendations for Governance, for example: Change transports between DEV, QA and PRD in BPC 7.0?
    What is the best method? Server Manager backup and restore, etc  ?
    And
    Best Practice recommendations on how to upgrade to a different version of BPC, for example: Upgrading from BPC 7.0 to 7.5 or 10.0 ?
    Kind Regards
    Daniel

Maybe you are looking for

  • Media Encoder CC conflicting with Premiere Pro CC

    Why does Adobe Media Encoder pause while exporting a video when I try to edit another video in Premiere Pro? On the older version I was using before I upgraded, they would both run simultaneously. Also, is there a phone number that I can call to get

  • My i tunes wont work

    i click on it and this think flashes then it juss stops. the same thing is happening to my quicktime.Somebody tell me wat to Do!

  • Memory Consumption

    We are using Solaris 10 and running oracle 11.1.0.7 on it. My Sysadmin told me about 97% RAM usage on my machine by my oracle user (dbora). Now, i want to match OS process with oracle session fo find out that which session is consuming the top memory

  • Can I use old drives with Mac Pro?

    Can I transfer the IDE drives from my Gigabit Ethernet G4 into empty bays on a new Mac Pro? If not, what would be the best way to continue to use them? I currently have 3 internal drives in the G4, including one 120G Seagate Barracuda. PowerPC G4 Gig

  • Where do we assign ITEM CATEGORY USAGE to a material?

    Hi all,   I have understood the calculation of ITEM CATEGORY based upon ITEM CATEGORY USAGE and three values. Can someody tell me where do we link this ITEM CATEGORY USAGE and the Material? Where do we assign this ITEM CATEGORY USAGE to a Material? T