Map Publish Data Smart link

Is there a way to configure the smart link right after a Map Published Data to only allow the next activity to run if the output patches a pattern in the Map Published Data?

Hi
if not ... you can use the "Compare Values" Activity and subscribe the Published Data from the "Map Published Data" Activity there and trigger only if "comparison result equals true".
Regards,
Stefan
www.sc-orchestrator.eu ,
Blog sc-orchestrator.eu

Similar Messages

  • Date Smart Links not opening iCal

    Hi,
    I'm getting an error that says "No associated application could be found" when clicking on a Date Smart Link on the Mail.app.
    The Date Smart Link is actually the words "Next Thursday" with containing the link x-apple-data-detectors://0
    Any suggestions to fix this?
    Thanks!

    Hi CY,
    Yes, I have iCal and it is in my Applications folder. The Date Smart Link doesn't work even when iCal is running.

  • Custom Activity not returning Published Data when ActivityWarning() is thrown

    I'm using Visual Studio 2012 on Windows 8.1 to develop a custom
    ActivityMonitor class using the
    imperative approach with the Microsoft System Center 2012 R2 Orchestrator Integration Toolkit. The
    ActivityMonitor inherits from IActivity, has two inputs and three outputs declared in the
    Design method, and publishes three outputs in the Execute method.
    I want to use this ActivityMonitor such that the Orchestrator
    smart link only processes published data when a Warning is encountered. Therefore, after calling the
    IActivityResponse.Publish method three times (once for each output), I throw a new
    ActivityWarning exception.
    Unfortunately, after throwing the ActivityWarning exception, no published data is passed down the smart link. Since this published data is used by the next
    Activity, it is causing that Activity to fail.
    Why is my published data getting discarded, just because an ActivityWarning was thrown? How can I make this work?
    If this post was helpful, please click the little "Vote as Helpful" button :)
    Trevor Sullivan
    Trevor Sullivan's Tech Room
    Twitter Profile

    Hi Mithun Sharma
    Are you getting desired result, when you execute the RFC FM directly in SAP, with the same values, which you pass from .NET
    Regards
    Madhan Doraikannan

  • How To View An Item When Publish Date Is Set Greater Than Current Date

    Using Oracle Portal Version: 10.1.2.0.2 (Build: 139).
    The customer(content manager) uploads a document into the portal on 28-JUL-2006 07:08 AM.
    The publish date is change to 1-AUG-2006 00:01 AM.
    There is a portal user group needing to review the document prior to the publish date.
    I created an oracle procedure to display the links to the unpublished items.
    Unfortunately the links generate an item can not be found error.
    When the publish date is met the links function properly.
    How can provide a method to allow this user group to view the unpublished document without granting them the manage content privilege on the page?
    Thanks

    For anyone else struggling with this... In my case this was not working due to a bug in XI R2, XI 3.0, and XI 3.1. QueryService modifyDataProvider (tracking ADAPT01113234) has been fixed with XI Release 2 FixPack 4.5 and I've tested that it did work for me. No such fix is available for XI 3.0 or XI 3.1 as of this post but I was told that it would get a fix in the next SP (fingers crossed)

  • Insert data into linked table

    Problem when you try to insert data into linked table.
    I made the view with all linked table columns (field1 is PK, field2, field3......) plus new primary key field "field1_PK" for tabular form wizard primary key field definition. After submit data on tabular form appears:
    "Error in mru internal routine:ORA-2001:Error in MRU: row=1,ORA-00957: duplicate column name ORA-02063:preciding line from "SHEMA_NAME", insert into "field1","field1_PK","field3",........) values (:b1,:b2,:b3.........)
    Error Unable to process update.
    Is it bug too?

    Hi,
    Have you tried to do your updates using an INSTEAD OF TRIGGER on your view? An INSTEAD OF TRIGGER allows you to update a non-updateable view.
    The following book contains an excellent introduction to this topic.
    "Oracle PL/SQL for Dummies"
    written by Rosenblum and Dorsey
    published by Wiley Publishing Inc.
    The following book is also excellant reference that deals with INSTEAD OF TRIGGERS.
    "Oracle PL/SQL programming
    written by Steven Feuerstein
    published by O'Reilly
    Hope this helps.
    Cheers,
    Patrick Cimolini

  • HTML Code for Smart Link Edit and Logout items

    I am using an HTML template and would like to include the Smart Link Edit and Logout items in a TABLE. How do I do this?
    Is there a way to get the code for these Smart Links (ORACLE tags) and include them?

    According to Oracle Metalink "Note:368720.1 - When ILS is Enabled the Edit Smartlink is visible to Users that are not entitled to see it", that is the expected behavior whether you are using the "edit" substitution tags or the "edit" smartlink on a page when the page has item-level access enabled. I don't like it but I can understand why that is the behavior since an end user may have access to manage/edit just one item on a page or just one tabbed region on a page and they need a way to put the page into edit mode in order to edit their item(s). If a user has nothing to edit on a page, then they won't be able to do anything in edit mode anyway.
    On the employee portal I work on, there is the possibility that every page may have a different set of users/groups that need to edit items/regions on the page AND the edit smartlink is published on the portal page template so basically, all employees can see the Edit link but if they click on it, only those employees that have items/regions to edit/manage will be able to really do anything while in edit mode.
    However, on the customer portal I work on, there is only one group that manages content site-wide so I was able to publish the Edit smartlink on the portal page template and change its access to the one group that manages content so they're the only ones that can see the Edit link on every page.
    The substitution tags for Logout are #PORTAL.LOGOUT.URL# and #PORTAL.LOGOUT.LABEL# (I use those so I can control the html & css) or just #PORTAL.LOGOUT# (which does both the label and url). The list of all substitution tags is available in Oracle Application Server Portal Help.

  • How to map bapimtcs-data to bapimatmra ?

    Hello friends,
    In order to implement the user exit z_uexit_material_vtl_status mentioned in Note Nr. 623026, I need a function which can convert (or map) bapimtcs-data to bapimatmra. This is necessary beacuse the standart code includes lines which contain syntax errors ("LS_BAPIMARA" and "LS_BAPISTRUCT-DATA" are not mutually convertible in a Unicode program. error) The conversion (mapping) of
    bapimatmvk to bapimtcs is necessary, too. Here is the link leading to the problematic user exit. Any useful help would appreciated and rewarded.
    Regards;
    Özcan.

    Hi,
    This MIGHT work, I am doing something similar in one of my user exits:
    in your variable declarations put this:
    DEFINE move_casting.
      assign &1 to <unicode_x1> casting.
      assign &2 to <unicode_x2> casting.
      move <unicode_x1> to <unicode_x2>.
    END-OF-DEFINITION.
    And then at the point where you need to convert the data:
    move_casting ls_bapimtcs-data ls_bapimatmra
    I hope it helps!
    Jeroen

  • Error when Publish data to passive cache

    Hi,
    I am using active-passive push replication. When I add data to active cache, the data added to cache but get the below error during publish data to passive cache.
    011-03-31 22:38:06.014/291.473 Oracle Coherence GE 3.6.1.0 <Error> (thread=Proxy:ExtendTcpProxyService:TcpAcceptor, member=1): Failed to publish EntryOperation{siteName=site1, clusterName=Cluster1, cacheName=dist-contact-cache, operation=Insert, publishableEntry=PublishableEntry{key=Binary(length=12, value=0x154E094D65656E616B736869), value=Binary(length=88, value=0x12813A15A90F00004E094D65656E61
    6B736869014E07506C6F74203137024E074368656E6E6169034E0954616D696C4E616475044E06363030303432401A155B014E0524737263244E0E73697465312D436C757374657231), originalValue=Binary(length=0, value=0x)}} to Cache dist-contact-cache because of
    (Wrapped) java.io.StreamCorruptedException: invalid type: 78 Class:com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher
    2011-03-31 22:38:06.014/291.473 Oracle Coherence GE 3.6.1.0 <D5> (thread=Proxy:ExtendTcpProxyService:TcpAcceptor, member=1): An exception occurred while processing a InvocationRequest for Service=Proxy:ExtendTcpProxyService:TcpAcceptor: (Wrapped: Failed to publish a batch with the publisher [Active Publisher] on cache[dist-contact-cache]) java.lang.IllegalStateException: Attempted to publish to cache dist-contact-cache
    Here is my coherence cache config xml file
    !DOCTYPE cache-config SYSTEM "cache-config.dtd">
    <cache-config xmlns:sync="class:com.oracle.coherence.patterns.pushreplication.configuration.PushReplicationNamespaceContentHandler">
    <caching-schemes>
    <sync:provider pof-enabled="true">
    <sync:coherence-provider />
    </sync:provider>
    <caching-scheme-mapping>
    <cache-mapping>
    <cache-name>dist-contact-cache</cache-name>
    <scheme-name>distributed-scheme-with-publishing-cachestore</scheme-name>
    <sync:publisher>
         <sync:publisher-name>Active Publisher</sync:publisher-name>
         <sync:publisher-scheme>
         <sync:remote-cluster-publisher-scheme>
         <sync:remote-invocation-service-name>remote-site2</sync:remote-invocation-service-name>
         <sync:remote-publisher-scheme>
         <sync:local-cache-publisher-scheme>
         <sync:target-cache-name>dist-contact-cache</sync:target-cache-name>
         </sync:local-cache-publisher-scheme>
         </sync:remote-publisher-scheme>
         <sync:autostart>true</sync:autostart>
         </sync:remote-cluster-publisher-scheme>
         </sync:publisher-scheme>
    </sync:publisher>
    </cache-mapping>
    </caching-scheme-mapping>
    <proxy-scheme>
    <service-name>ExtendTcpProxyService</service-name>
    <thread-count>5</thread-count>
    <acceptor-config>
    <tcp-acceptor>
    <local-address>
    <address>localhost</address>
    <port>9099</port>
    </local-address>
    </tcp-acceptor>
    </acceptor-config>
    <autostart>true</autostart>
    </proxy-scheme>
    <remote-invocation-scheme>
    <service-name>remote-site2</service-name>
    <initiator-config>
    <tcp-initiator>
    <remote-addresses>
    <socket-address>
    <address>localhost</address>
    <port>5000</port>
    </socket-address>
    </remote-addresses>
    <connect-timeout>2s</connect-timeout>
    </tcp-initiator>
    <outgoing-message-handler>
    <request-timeout>5s</request-timeout>
    </outgoing-message-handler>
    </initiator-config>
    <serializer>
              <class-name>com.tangosol.io.pof.ConfigurablePofContext</class-name>
         </serializer>
    </remote-invocation-scheme>
    </caching-schemes>
    </cache-config>

    Please find below for active and passive server start command, log files.
    Thanks.
    Active server start command
    java -server -showversion -Xms128m -Xmx128m -Dtangosol.coherence.ttl=0 -Dtangosol.coherence.cacheconfig=active-cache-config.xml
    -Dtangosol.coherence.cluster="Cluster1" -Dtangosol.coherence.site="site1" -Dtangosol.coherence.clusteraddress="224.3.6.2"
    -Dtangosol.coherence.clusterport="3001" -Dtangosol.pof.config=pof-config.xml -cp "config;lib\custom-types.jar;lib\coherence.jar;lib/coherence-common-1.7.3.20019.jar;lib/coherence-pushreplicationpattern-3.0.3.20019.jar;lib/coherence-messagingpattern-2.7.3.20019.jar;" com.tangosol.net.DefaultCacheServer
    Passive server start command
    java -server -showversion -Xms128m -Xmx128m -Dtangosol.coherence.ttl=0 -Dtangosol.coherence.cacheconfig=passive-cache-config.xml
    -Dtangosol.coherence.cluster="Cluster2" -Dtangosol.coherence.site="site2" -Dtangosol.coherence.clusteraddress="224.3.6.3"
    -Dtangosol.coherence.clusterport="3003" -Dtangosol.pof.config=pof-config.xml -cp "config;lib\custom-types.jar;
    lib\coherence.jar;lib/coherence-common-1.7.3.20019.jar;lib/coherence-pushreplicationpattern-3.0.3.20019.jar;
    lib/coherence-messagingpattern-2.7.3.20019.jar;" com.tangosol.net.DefaultCacheServer
    Active Server log
    <Error> (thread=PublishingService:Thread-3, member=1): Failed to publish the range MessageTracker{MsgId{40-1} } of messages for subscription SubscriptionIdentifier{destinationIdentifier=Identifier{dist-contact-cache}, subscriberIdentifier=Identifier{Active Publisher}} with publisher com.oracle.coherence.patterns.pushreplication.publishers.RemoteClusterPublisher@423d4f Class:com.oracle.coherence.patterns.pushreplication.providers.coherence.CoherencePublishingService
    2011-04-01 20:23:20.172/65.972 Oracle Coherence GE 3.6.1.0 <Info> (thread=PublishingService:Thread-3, member=1): Publisher Exception was as follows Class:com.oracle.coherence.patterns.pushreplication.providers.coherence.CoherencePublishingService
    (Wrapped: Failed to publish a batch with the publisher [Active Publisher] on cache [dist-contact-cache]) java.lang.IllegalStateException: Attempted to publish t
    o cache dist-contact-cache
    at com.tangosol.util.Base.ensureRuntimeException(Base.java:293)
    at com.oracle.coherence.patterns.pushreplication.publishers.RemoteClusterPublisher$RemotePublishingAgent.run(RemoteClusterPublisher.java:348)
    at com.tangosol.coherence.component.net.extend.proxy.serviceProxy.InvocationServiceProxy.query(InvocationServiceProxy.CDB:6)
    at com.tangosol.coherence.component.net.extend.messageFactory.InvocationServiceFactory$InvocationRequest.onRun(InvocationServiceFactory.CDB:12)
    at com.tangosol.coherence.component.net.extend.message.Request.run(Request.CDB:4)
    at com.tangosol.coherence.component.net.extend.proxy.serviceProxy.InvocationServiceProxy.onMessage(InvocationServiceProxy.CDB:9)
    at com.tangosol.coherence.component.net.extend.Channel.execute(Channel.CDB:39)
    at com.tangosol.coherence.component.net.extend.Channel.receive(Channel.CDB:26)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Peer.onNotify(Peer.CDB:96)
    at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
    at java.lang.Thread.run(Thread.java:619)
    Caused by: java.lang.IllegalStateException: Attempted to publish to cache dist-contact-cache
    at com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher.publishBatch(AbstractCachePublisher.java:163)
    at com.oracle.coherence.patterns.pushreplication.publishers.RemoteClusterPublisher$RemotePublishingAgent.run(RemoteClusterPublisher.java:343)
    ... 9 more
    Caused by: (Wrapped) java.io.StreamCorruptedException: invalid type: 78
    at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:266)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.PartitionedService$ConverterKeyToBinary.convert(PartitionedService.CDB:16)
    at com.tangosol.util.ConverterCollections$ConverterInvocableMap.invoke(ConverterCollections.java:2156)
    at com.tangosol.util.ConverterCollections$ConverterNamedCache.invoke(ConverterCollections.java:2622)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ViewMap.invoke(PartitionedCache.CDB:11)
    at com.tangosol.coherence.component.util.SafeNamedCache.invoke(SafeNamedCache.CDB:1)
    at com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher.publishBatch(AbstractCachePublisher.java:142)
    ... 10 moreCaused by: java.io.StreamCorruptedException: invalid type: 78
    at com.tangosol.util.ExternalizableHelper.readObjectInternal(ExternalizableHelper.java:2266)
    at com.tangosol.util.ExternalizableHelper.readObject(ExternalizableHelper.java:2254)
    at com.tangosol.io.DefaultSerializer.deserialize(DefaultSerializer.java:74)
    at com.tangosol.util.ExternalizableHelper.deserializeInternal(ExternalizableHelper.java:2708)
    at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:262)
    ... 16 more
    Passive Server log
    2011-04-01 20:23:20.141/56.925 Oracle Coherence GE 3.6.1.0 <Error> (thread=Proxy:ExtendTcpProxyService:TcpAcceptor, member=1): Failed to publish EntryOperation{siteName=site1, clusterName=Cluster1, cacheName=dist-contact-cache, operation=Insert, publishableEntry=PublishableEntry{key=Binary(length=12, value=0x154E094D65656E616B736869), value=Binary(length=88, value=0x12813A15A90F00004E094D65656E616
    B736869014E07506C6F74203137024E074368656E6E6169034E0954616D696C4E616475044E06363030303432401A155B014E0524737263244E0E73697465312D436C757374657231), originalValue=Binary(length=0, value=0x)}} to Cache dist-contact-cache because of (Wrapped) java.io.StreamCorruptedException: invalid type: 78 Class:com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher2011-04-01 20:23:20.141/56.925 Oracle Coherence GE 3.6.1.0 <D5> (thread=Proxy:ExtendTcpProxyService:TcpAcceptor, member=1): An exception occurred while processing a InvocationRequest for Service=Proxy:ExtendTcpProxyService:TcpAcceptor: (Wrapped: Failed to publish a batch with the publisher [Active Publisher] on cache [
    dist-contact-cache]) java.lang.IllegalStateException: Attempted to publish to cache dist-contact-cache
    at com.tangosol.util.Base.ensureRuntimeException(Base.java:293)
    at com.oracle.coherence.patterns.pushreplication.publishers.RemoteClusterPublisher$RemotePublishingAgent.run(RemoteClusterPublisher.java:348)
    at com.tangosol.coherence.component.net.extend.proxy.serviceProxy.InvocationServiceProxy.query(InvocationServiceProxy.CDB:6)
    at com.tangosol.coherence.component.net.extend.messageFactory.InvocationServiceFactory$InvocationRequest.onRun(InvocationServiceFactory.CDB:12)
    at com.tangosol.coherence.component.net.extend.message.Request.run(Request.CDB:4)
    at com.tangosol.coherence.component.net.extend.proxy.serviceProxy.InvocationServiceProxy.onMessage(InvocationServiceProxy.CDB:9)
    at com.tangosol.coherence.component.net.extend.Channel.execute(Channel.CDB:39)
    at com.tangosol.coherence.component.net.extend.Channel.receive(Channel.CDB:26)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.Peer.onNotify(Peer.CDB:96)
    at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
    at java.lang.Thread.run(Thread.java:619)
    Caused by: java.lang.IllegalStateException: Attempted to publish to cache dist-contact-cache
    at com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher.publishBatch(AbstractCachePublisher.java:163)
    at com.oracle.coherence.patterns.pushreplication.publishers.RemoteClusterPublisher$RemotePublishingAgent.run(RemoteClusterPublisher.java:343)
    ... 9 more
    Caused by: (Wrapped) java.io.StreamCorruptedException: invalid type: 78
    at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:266)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.PartitionedService$ConverterKeyToBinary.convert(PartitionedService.CDB:16)
    at com.tangosol.util.ConverterCollections$ConverterInvocableMap.invoke(ConverterCollections.java:2156)
    at com.tangosol.util.ConverterCollections$ConverterNamedCache.invoke(ConverterCollections.java:2622)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.partitionedService.PartitionedCache$ViewMap.invoke(PartitionedCache.CDB:11)
    at com.tangosol.coherence.component.util.SafeNamedCache.invoke(SafeNamedCache.CDB:1)
    at com.oracle.coherence.patterns.pushreplication.publishers.cache.AbstractCachePublisher.publishBatch(AbstractCachePublisher.java:142)
    ... 10 more
    Caused by: java.io.StreamCorruptedException: invalid type: 78
    at com.tangosol.util.ExternalizableHelper.readObjectInternal(ExternalizableHelper.java:2266)
    at com.tangosol.util.ExternalizableHelper.readObject(ExternalizableHelper.java:2254)
    at com.tangosol.io.DefaultSerializer.deserialize(DefaultSerializer.java:74)
    at com.tangosol.util.ExternalizableHelper.deserializeInternal(ExternalizableHelper.java:2708)
    at com.tangosol.util.ExternalizableHelper.fromBinary(ExternalizableHelper.java:262)
    ... 16 more

  • Finding Published date of bid invitation

    Dear All,
    I Want to know which table/ Function module stores the 'published date' of a bid invitation. Table CRMD_ORDERADM_H stores only created date and changed date. We are using SRM 4.0
    Thanks.
    Regards
    Amit

    Hi Amit,
    Try using the following FM
    BBP_PD_BID_GETDETAIL
    see this link too
    Publish date of a bid invitation
    Hope this solves ur query
    Thanks,
    Pradeep.
    Edited by: Pradeep Purushottam on Sep 11, 2008 8:40 AM

  • Publish dates into iCal from Numbers

    I am trying to publish dates from a chart in Numbers to iCal or Outlook calendar. Is this something that can be done?

    Solved, I think! What you need to do is log on to your icloud account at www.icloud.com, go to the calendar screen, click the gearwheel and choose New Calendar. Create a calendar here (ie up on the iCloud) with the name you want for your imported calendar data. Then switch to iCal - you may have to quit and relaunch for the new iCloud calendar to appear (or it will probably do so automatically if you wait long enough) then import the data using File->Import->Import... When you get to the step where it asks which calendar you want to import the data into, choose the new iCloud calendar you've just created. Job done!
    To be complete in my answer, I don't think you can exactly do what you want to do, at least not at present. There doesn't seem to be any way to publish a calendar to iCloud after the first time you set up iCloud on your Mac when it exports any calendar data that is on the Mac at that time. I suppose that Apple feel that the way I have described works by exporting from one calendar and importing into another, so no need to provide a way of doing it directly.

  • How can I read published data from a datasocket server running in a different network?

    Hi all,
    I have been trying to solve this problem but without any success. I want to develop data acquisition Vi to run on a computer on a remote location. I want to use the datasocket technology to acquire and publish this data so that in my office (in a different network), I can read the published data and perform analysis with it. I can successfully use this approach for the two computers in the same network but not when they are in different networks. My real problem is how to specify the URL for the acquisition computer in one network while the analysis Vi runs on a different network.
    I have attached two Vis to illustrate my point (I simply want to be able to read the random numbers generated in the acquisition computer in a different network). The first Vi (RemoteDatasocketWrite.vi) will run on the remote computer with an IP address (192.168.0.110). My office computer on which RemoteDatasocketRead.vi runs has an IP address (192.168.0.11) and I can log in remotely into the remote computer using Remote Desktop Protocol with name (emelvin.001.eairlink.com) from my office computer.
     Is it possible to do what I am trying to do or is there a simpler way to solve this problem?
    I will really appreciate any help towards a solution to this problem.
    Thanks
    Attachments:
    RemoteDatasocketWrite.vi ‏9 KB
    RemoteDatasocketRead.vi ‏9 KB

    Are you getting an error? The way you have it set up, you can have an error and it will never be displayed. Put and error control on your front panel and see what it gives you. Also, shared variables in a project can be useful. Look at some examples for that.

  • Problem with context mapping and data flow in a FPM application

    Hi All,
    I am trying to develop an ESS application using FPM. For the same, the requirement is to see the history of an employee in the second view.
    The first view has got just the overview information and the second one has got the detail. So, the records or the fields are the same on both the views.
    As per the FPM guidelines, the Model is residing in the Fc component and the respective Vc components are using the model data accordingly.
    I am executing the model in the Fc component calling the executable method in the interfaceController of the first view and then trying to display the output data of the BAPI in the first view which provides the overview information.This is working fine.
    But when i am trying to map the same output node to the Table UI for the second view, the record size is coming zero and thus no information is available.
    For the above issue, I am again executing the RFC in the InterfaceController of the second view to populate the records, which is incorrect as it is already executed and the data is available for the first view.
    I request you to let me know the correct approach to Context mapping and data flow when using FPM-roadmap. Is their any standard method or approach available to deal with such requirements? Please let me know.
    Thanks in advance.
    Regards
    DK

    Hi Idhaya,
    I model node is available in Fc and the Fc interface controller is being used in the first Vc and the second Vc.
    So the idea is, as the executable method is generated in the Fc, so i have created a custom method to call the executable method in Fc, where the input parameter is getting passed and this custom method is finally getting called is the first Vc.
    So , now my first Vc is ready to call the custom method in Fc and execute the RFC. Once the RFC is executed, the nodes in the Fc should get populated which is the ideal case.
    And as the Fc is used as a component in the second Vc, the same node is available to the UI elements.
    But, when I check the record size for the output node, it is always zero, for the second Vc.
    Regards
    DK

  • How are attribute and text master data tables linked in SAP R/3?

    Hello,
    how are attribute and text master data tables linked in SAP R/3?
    Most tables with attribute master data like T001 for company codes,
    have a text master data table T001T (add "T" to table name).
    When looking at the content of table T001 via transaction se11,
    the text are automatically joined.
    But for some tables there is no "T"-table (e.g. table TVBUR for sales offices
    has no text table TVBURT), but in se11 you get texts. There is an address
    link in TVBUR, but the Name1, etc. are empty.
    a) Where are the text stored?
    b) How does the system know of the link?
    Hope someone can help!
    Best regards
    Thomas

    Hi Thomas
    The master and text table are not linked by name, of course, if you see the text table, it has the same key fields of master table, only it has the field key spras and the field for description.
    The link beetween the tables is done by foreign key: if you check the text table TVKBT u need to see how the foreign key for field VKBUR is done:
    -> Foreing key with table TVBUR
    -> Foreing key field type -> KEY FIELD FOR A TEXT TABLE
    ->Cardinality-> 1-:CN
    It's very important the attribute sets for Foreing key field type, if it's KEY FIELD FOR A TEXT TABLE, it'll mean the table is a text table: i.e. that mean the master table is a check table for the text table, where the foreign key type is for text table.
    U can find out the text table of master table by SE11: GoTo->Text Table
    U can fined some information in table DD08L.
    Max

  • Problem in mapping xml data with header details from IPM 11g to BPEL

    Hi,
    I want to map xml data as a supporting content from IPM application to BPEL.
    My xml is
    <?xml version="1.0" encoding="utf-8"?>
    <DocumentFile xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns="http://tempuri.org/Document.xsd">
    <Document DocumentType="Invoice">
    <DocumentImage>
    <Filename>\\10.205.0.209\Img\10883212.TIF</Filename>
    </DocumentImage>
    </DocumentFile>
    If I remove header details from root element <DocumentFile> i.e. modified xml is
    <DocumentFile>
    <Document DocumentType="Invoice">
    <DocumentImage>
    <Filename>\\10.205.0.209\Img\10883212.TIF</Filename>
    </DocumentImage>
    </DocumentFile>
    it works fine but i need to pass header details as well.
    Please suggest.
    Thanks,
    Priya

    Hi Naveen,
    In sxmb_moni the content transmitted to the adapter(RFC)is as follows
    <?xml version="1.0" encoding="UTF-8" ?>
    - <ns:ZRFID_EQUIP xmlns:ns="urn:sap-com:document:sap:rfc:functions">
    - <RECORDS>
    - <item>
      <FLOC>f1-01-01</FLOC>
      <RFID_NO>I006</RFID_NO>
      </item>
    - <item>
      <FLOC>f1-01-02</FLOC>
      <RFID_NO>I002</RFID_NO>
      </item>
    - <item>
      <FLOC>f1-01-03</FLOC>
      <RFID_NO>I003</RFID_NO>
      </item>
    - <item>
      <FLOC>f1-01-04</FLOC>
      <RFID_NO>I004</RFID_NO>
      </item>
    - <item>
      <FLOC>f1-01-05</FLOC>
      <RFID_NO>I005</RFID_NO>
      </item>
    - <item>
      <FLOC>f1-01-06</FLOC>
      <RFID_NO>I001</RFID_NO>
      </item>
      </RECORDS>
      </ns:ZRFID_EQUIP>
    At r/3 side the field floc and rfid_no gets mapped to floc which is of char30
    eg floc=f1-01-01I006
       rfid_no=

  • Sharepoint 2013 Reporting services not fetching data via linked server

    We have a functional SharePoint 2013 farm in production with WFE (Server 1 , windows server 2012) and WFE ( Server 2 , windows server 2012) and SharePoint databases including report server databases on  (Server 3, which SQL Server 2012 on windows
    server 2012)
    Kerberos Authentication is enabled and Windows Authentication happens. Claims to Windows Service seems to work correctly too. No double hop issue. The SharePoint Reporting Services seem to be configured correctly and Reports get generated as expected
    in the following cases.
     1. If data is being pulled from Server A (SQL Server 2008 R2 Server on Windows Server 2008), Report gets generated no issues.
     2. If data is being pulled from Server B (SQL Server 2008 R2 Server on Windows Server 2008), Report gets generated no issues.
    The problem:-
    We have a view on Server A, which is fetching data from Server B. This report does not get generated. Gives an NT/Authority logon failure. Server B is created as a linked Server on Server A. The setting to run under users security context is checked.
    If I execute Query Builder from Report Builder 3.0, the query on the view gets the data and data is displayed in the report for next 10-15 min but after that the connections gets lost or if the user logs out and logs back in, the connection is lost
    again.
    Has some one experienced a issue like this? Can a report not work on a view getting data from linked server? Or I have missed  some thing really important.
    Any help is highly appreciated?
    cheers!
    Rachna
    Rachna

    Hi,
    From your description, my understanding is that you got an error when you clicked System settings after configuring Reporting Services service application.
    What size did you use for the memory? Please make sure the server has enough memory to run SharePoint and all services.
    In addition, please open
    http://localhost:xxxxx/SecurityTokenServiceApplication/securitytoken.svc to check if there is something about this issue.
    Here is a similar post about this issue, please take a look at:
    http://whitepages.unlimitedviz.com/2012/07/setting-up-reporting-services-2012-with-sharepoint-2013/
    As this issue is about Reporting Services, I suggest you create a new thread on SSRS forum, more experts will assist you with SSRS.
    SSRS forum:
    https://social.technet.microsoft.com/Forums/office/en-US/home?forum=sqlreportingservices
    Best Regards,
    Wendy
    TechNet Community Support
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected]

Maybe you are looking for

  • Trial Balance Report not showing in months

    Hi All. A client is trying to run a trial balance, wanting it to display in months. He chooses monthly report, but it still gives just one total amount. The company he is running is a seconday company. He ran the report in the main company and it wor

  • Solaris 10 ldapclient setup question

    Hi all, I'm setting up ldap authentication for ssh logins on a Solaris 10 system. Up to now, I've been successful in finding my answers in: 1 this forum 2 http://www.sun.com/bigadmin/features/articles/nis_ldap_part2.jsp 3 http://web.singnet.com.sg/~g

  • IdleRemover notifying pools, interval: 450000

    Hi, I am using Jboss 3.2.1 and oracle 9.1 i. I have this error in the jboss error log. Does anyone know what's the problem? Thanks in advance. DEBUG [org.jboss.resource.connectionmanager.IdleRemover] run: IdleRemover notifying pools, interval: 450000

  • Infotype screen control

    Hi, Can anyone help me with the difference between the STD and RF screen control option for the fields of an infotype in the table T588M? I understand that both STD and RF option makes a field mandatory. Also, 'STD' means standard Setting and 'RF' me

  • Safari closing unexpectedly

    Hi all, I'm having a problem with my macbook pro shutting down safari randomly. its never on the same site. I dont really know where to start. I'm running osx 10.6.8. when it does this I get the message 'safari has quit unexpectadly' then the options