IPM and Process Injector

Hi All,
We have just started using IPM and are busy configuring the Process Injector.
We have created a application/process/template and can add documents to to the system from the filer and manually add those filed documents to a process. We want to configure the process injector to automatically add to the process from the filer.
Have followed the install/admin guide for configuring it, but it does not do anything.
When starting and stopping the process injector we get a logon error, but not sure what this logon is related to as we can connect to db from the service configuration?
Any help appreciated.
Details of Service Manager log for Process Injector:
Server started: Computer: APPECMTS01, Tool: Process Injector.
Logon failure: unknown user name or bad password.
Injector Worker Thread Disabled
UpdateActiveServer: Entered.
UpdateActiveServer: Entered.
UpdateActiveServer: Entered.
Thx

IPM documentation/downloads are here.
http://www.oracle.com/technology/software/products/content-management/index_ipm.html
Suggest that you read some/all of this and come back with specific questions
Hope that helps
Tim

Similar Messages

  • How to Integrate Oracle Imaging and Process Management(IPM) 7.7.10 with EBS

    Hi All,
    We are implementing the following modules from Oracle Content Management to one of our Client, the following are the versions of the tools:
    Configure and deploy Oracle Forms Recognition (OFR) 10GR3
    Configure the Oracle Adopter for integration of OFR with EBS Version 11.5.10.2
    Configure the Adopter for integration of OFR with Oracle Imaging and Process Management (IPM) 7.7.10
    I really like to have any technical design document on integrations front, especially from OFR -> EBS and also IPM -> EBS.
    Please feel reach me on my following contact details:
    [email protected]
    Phone: 1865-582-2369
    Mobile: 1865-202-6430
    Any pointers for the same is well appreciated.
    Regards
    Chandra

    Try replacing the .dll file with the one in your WINDOWS directory. I had that problem and I copied the .dll from C:/WINDOWS/system32 to overwrite the one located in the install directory.

  • How to develp a subreport in the same page of the main report and processed when is called

    SSRS 2012
    Hi guys,
    I am developing a report that includes a subreport.
    therefore The report is made up by 2 sections:
    1) Show a grid (several rows - may be 100 or more - and some columns). It represent high level info  (Each row represent a specific item)
    2) The user click on a Row/column (eg: Item1, to see details that must be shown as graph.
    Here the problem: Searching info about SSRS reporting for subreports, drilldown, nested report and so on, no ones accomplish the requirements for the following reasons:
    1) Subreports: NOT open in the same page. The user must <go back> to see detail for other items, such as Item 2....
    2) Nested and DrillDown: Process data as the Main report (they are hidden and displayed after user action). In my case this leads to performance problem, because: if the main report has 200 objects (it means 200 graphs processed and hidden at the same time
    for the main reports).
    3) DrillThrough and Subreport: The detail data are processed only when the user select the interested item in the main report. Here the problem: the report is shown externally.
    Please, it there any possibility to have a subreport (or anything else) to be executed in the same report as the main report and processed only when the user make action?
    Thanks for your help.

    Hi Fasttrack2,
    According to your description, there is a main report with more than 100 rows of summary information, when users click the item of the row, the report need to jump to detail information shows as graph. The problem you are facing is that you are not sure
    subreport, drilldown or drillthrough report should be used? You hope to set display detail information in the main report and keep high performance of report.
    In Reporting Services, each subreport instance is a separate query execution and a separate report processing task. Subreports are recommended when there are just a few subreport instances. We should not use subreports inside a group when there are many
    group instances, instead, consider using drillthrough reports. Drilldown reports process all data even when the data is first hidden. In order to improve the performance of the report, we can split drilldown reports into parameterized drillthrough reports,
    because Drillthrough reports do not run until a user clicks the drillthrough link in the main report.
    In this case, we can consider using drilldown or drillthrough report. In drilldown report, when we click a plus or minus button to expand or collapse a section of a report, the detail data will be displayed in place. By using drillthrough report, when we
    click the link in the main report, the detail data will be displayed, we can come back to the main report by click return button in the drillthrough report.
    Reference:
    Troubleshooting Reports: Report Performance
    Drillthrough, Drilldown, Subreports, and Nested Data Regions
    If you have any more questions, please feel free to ask.
    Best Regards,
    Wendy Fu
    If you have any feedback on our support, please click
    here.

  • Difference between connection, session and process

    Hi all,
    Can anyone please update me on the difference between connection,session and process.
    Thanks in advance,
    - Sri

    I got this useful note by googled in net. It describes session,connection,process gracefully.
    A connection is a physical circuit between you and the database.A connection
    might be one of many types -- most popular begin DEDICATED server and SHARED
    server. Zero, one or more sessions may be established over a given connection
    to the database as show above with sqlplus. A process will be used by a session
    to execute statements. Sometimes there is a one to one relationship between
    CONNECTION->SESSION->PROCESS (eg: a normal dedicated server connection).
    Sometimes there is a one to many from connection to sessions (eg: like
    autotrace, one connection, two sessions, one process). A process does not have
    to be dedicated to a specific connection or session however, for example when
    using shared server (MTS), your SESSION will grab a process from a pool of
    processes in order to execute a statement. When the call is over, that process
    is released back to the pool of processes.

  • Can i use labview to capture and process images from ip cameras connected to a switch.

    Hi,
    i have more then one ip cameras conneted in LAN by a switch. i have a PC conneted to same switch. is it possible to capture images from these ip cameras and process them(for example performing simple barcode scan or QR-code scan or OCR)?  Which IP cameras will be supported?

    There is functionally no difference whether the camera is connected directly to the system or via a switch. Both should work the same. The only real requirement that you'd need to care about is whether the switch can pass multicast UDP traffic correctly (this is used for network discovery of IP cameras). Generally all unmanaged switches (probably what you are using) will treat multicast traffic identical to broadcast traffic, but some managed switches/routers may need to be configured to pass multicast.
    Eric

  • Why do I have to Open and physically Save a provided .xlsx File in order to source it and process it in SSIS?

    It sounds CRAZY...but it's true and I proved it out over many tests...
    We receive a .xlsx File from an outside vendor. If we simply move the file over to our processing Path and then try to source it and process it in SSIS, we get this error...
    Exception from HRESULT: 0xC02020E8
    Error at Data Flow Task [Excel Source [1]]: SSIS Error Code DTS_E_OLEDBERROR. An OLE DB Error has occurred.
    Error code: 0x80004005
    However, if we open the file and Save It and/or Save As to our processing library path, we can source the .xlsx File and process it accordingly.
    Why is this? Is it because something funky is happening under the covers when the the vendor creates and saves the .xlsx File and then when we open and save it on our network it then cleans it up?
    Is there anyway that we can get around this? I hate for the processing clerk to have to go through what seems like a meaningless task....but I cannot think of any way around this if we cannot source the original file. And a C# Edit Script using Microsoft.Interpol.Excel
    is not possible because we are not allowed to put Microsoft Office on the Server where the SSIS Package will eventually reside.
    Any thoughts...suggestions...etc...would be GREATLY appreciated.
    Thanks for your review and am very hopeful (Holding my breath) for a reply and possible solution and answer to this strangeness.
    ITBobbyP85

    First of all...THANK YOU ARTHUR!
    You're probably sick and tired of seeing my posts out here on this.
    We are running Microsoft Office 32-bit which is why I downloaded the 32-bit driver of Microsoft Access Database Engine 2010 from...
    http://www.microsoft.com/en-us/download/details.aspx?id=13255
    Sooooo I think what you're saying is that it is a Microsoft Office 2010 .xlsx File but utilizing the 32-bit Microsoft Access Database Engine 2010 only allows me to source as Microsoft Excel 2007.
    Nowwww....this will run on a Server...soooo a couple of questions...
    1.) Are you saying I need to use the 64-bit Microsoft Access Database Engine 2010 in order to process this file without having to open and Save it? Can I do that if we're running Microsoft Office 32-bit? Will I be able to source this on the Server because
    we cannot install Microsoft Office 2010 on the Server where this SSIS Package will live and breathe.
    2.) How can I tell which Microsoft Access Database 2010 Engine is on the server where this will eventually reside and run? 32-bit or 64-bit?
    Welllll...maybe a 3rd question...
    3.) Is there anyway to get around this driver dependency? Can I create a script to bulkcopy it and it would simply take what's on the server.
    Now mind you...we DO NOT have Microsoft Office installed on our server where this will reside...a licensing issue...
    So maybe now this explains my struggles with this damn file(excuse my French and frustration)...and why I was using a C# Edit Script to Open and Save the File until I then deployed it to the server and the Microsoft.Office.Interop.Excel DLL was not available
    on the Server because we cannot install Microsoft Office on the Server.
    Sooooo it seems as though I'm back at square one.
    UGH!
    Any additional help you can suggest or even if we can continue this "offline" would be GREATLY GREATLY appreciated.
    Thanks!

  • 'AND' process in a process chain failure - need help

    For the first time and out of no where, two separate 'AND' processes in two separate process chains failed on two separate days. It displayed error the same messages in both isntances as below. It seems that when the first load completed and the second was still in progress, this 'AND' attempted excecute and failed.
    I had to manually resume the chain by setting the 'AND' status'to 'G' in RSPCPROCESSLOG table. I also plan to recreate those 'AND'.
    But did anyone here experience this problem and know the root cause?
    Thanks for any assistance!
    Julian
    This AND process is not waiting for event RSPROCESS, parameter      
    Job cancelled after system exception ERROR_MESSAGE

    Thanks, Shambhu!
    I am planning to create new ones. But my problem is that I hate to recreate all 'AND's I have in all our chains. There are a lot of them. But if I don't, I have no idea which is the next victim ...
    Julian

  • How to get the data from mysql database which is being accessed by a PHP application and process the data locally in adobe air application and finally commit the changes back in to mysql database through the PHP application.

    How to get the data from mysql database which is being accessed by a PHP application and process the data locally in adobe air application and finally commit the changes back in to mysql database through the PHP application.

    If the data is on a remote server (for example, PHP running on a web server, talking to a MySQL server) then you do this in an AIR application the same way you would do it with any Flex application (or ajax application, if you're building your AIR app in HTML/JS).
    That's a broad answer, but in fact there are lots of ways to communicate between Flex and PHP. The most common and best in most cases is to use AMFPHP (http://amfphp.org/) or the new ZEND AMF support in the Zend Framework.
    This page is a good starting point for learning about Flex and PHP communication:
    http://www.adobe.com/devnet/flex/flex_php.html
    Also, in Flash Builder 4 they've added a lot of remote-data-connection functionality, including a lot that's designed for PHP. Take a look at the Flash Builder 4 public beta for more on that: http://labs.adobe.com/technologies/flashbuilder4/

  • Process chain AND process

    Hi,
    I have a process chain that has two parallel processes of loading DSO and Infocube data. There are two infopackages on one side and two on the other side. The two processes are connected with and AND process at the end and after that there is a datastore activation process.
    When I executed the process chain, the infopackages on both sides started loading data. The two infopackages on the left side finished first while the second infopackage on the right side had not finished yet when the AND process errored. I get the following message:
    This AND process is not waiting for event RSPROCESS , parameter
    47VKUORHMXQKR2AN73AVAR0C9
    After the AND process I have a Datastore generation process and I want it to continue only when all infopackages have been loaded.
    Why does the process stop at AND even though all infopackages have not finished yet?
    thanks
    Michalis

    There was no error message in ST22 or SM37. There was very little info to go on. I just deleted the requests and run the process chain again and there's no error. However the process chain does fail at a later step.
    Update Datastore Object Data (Further update)
    I don't know what that further update is for. It was automatically selected in my Process chain when I selected the infopackage for the DSO.
    There's a MESSAGE_TYPE_X error.
    Information on where terminated
        Termination occurred in the ABAP program "SAPLRSM2" - in "RSSM_MON_WRITE_3".
        The main program was "RSBATCH1 ".
        In the source code you have the termination point in line 2659
        of the (Include) program "LRSM2U16".
        The program "SAPLRSM2" was started as a background job.
        Job Name....... "BI_PROCESS_ODSPROCESS"
        Job Initiator.. "CSPAPADM"
        Job Number..... 14103000
    Source Code Extract
    Line  SourceCde
    2629       CALL FUNCTION 'RSSM_GET_TIME'
    2630         EXPORTING
    2631           i_datum_utc  = rsreqdone-tdatum
    2632           i_uzeit_utc  = rsreqdone-tuzeit
    2633         IMPORTING
    2634           e_timestamps = l_s_status-t_timestamp.
    2635       IF rsreqdone-tstatus(3) = icon_green_light(3) OR
    2636          rsreqdone-tstatus(3) = icon_red_light(3).
    2637         CALL FUNCTION 'RSSM_GET_TIME'
    2638           IMPORTING
    2639             e_timestamps = l_s_status-ts_proc_ended.
    2640         l_not_yellow = 'X'.
    2641       ENDIF.
    2642       IF l_not_yellow = 'X'.
    2643         CALL FUNCTION 'RSSTATMAN_SET_STATUS'
    2644           EXPORTING
    2645             i_s_status    = l_s_status
    2646             i_with_commit = 'X'.
    2647         IF l_s_status-t_status(3) = icon_green_light(3) AND
    2648            l_s_status-dta_type   = 'ODSO' AND
    2649            t_data-aufrufer <= '70'.
    2650           CALL FUNCTION 'RSS2_CALLBACK_ODSADM_ODS_SUCC'
    2651             EXPORTING
    2652               i_ods = l_s_status-dta(30)
    2653               i_rnr = l_s_status-rnr.
    2654         ENDIF.
    2655       ENDIF.
    2656     ENDLOOP.
    2657     IF sy-subrc <> 0.
    2658       IF t_data-aufrufer <> '09'.
    2660       ENDIF.
    2661     ELSE.
    2662       IF l_not_yellow = 'X'.
    2663         CALL FUNCTION 'RSB1_OSNAME_TO_ICNAME'
    2664           EXPORTING
    2665             i_osource        = g_s_minfo-oltpsource
    2666 *           I_OSTYPE         = RSAOT_C_OSTYPE-TRAN
    2667           IMPORTING
    2668             e_icname         = l_icname
    2669           EXCEPTIONS
    2670             name_error       = 1
    2671             OTHERS           = 2.
    2672         IF sy-subrc = 0.
    2673           CALL FUNCTION 'RSAWB_MAINTAIN_DTA'
    2674             EXPORTING
    2675               i_method                   = rsatr_c_dta_get
    2676 *             I_TYPE                     =
    2677               i_objvers                  = rs_c_objvers-active
    2678               i_dta                      = l_icname
    Any idea what that means? Also do I need that further update step in my process chain?

  • Process Chain hangs on AND process

    Help Super gurus!
    I've tried several versions of a chain that loads multiple InfoPackages serially then multiple DTP in parallel using filtering from the PSA to a DSO. All load successfully. Then I have the DTPs all meet at an 'AND' process before processing should continue to a cube.
    The 'AND' process also shows green but it does not continue to the next process step. It just stops there, no more jobs running, no yellow, nothing. It thinks it's all done, everything after that is still blue.
    Why does it hang there and what can I do to get it to continue?
    Thanks,
    Al
    Points available.

    Hi,
    Not sure what is the root cause in your case.
    Try this. It may work. Normally this is done after a step turns red and no repeat is available.
    Method 1 (when it fails in a step/request)
    /people/siegfried.szameitat/blog/2006/02/26/restarting-processchains
    How is it possible to restart a process chain at a failed step/request?
    Sometimes, it doesn't help to just set a request to green status in order to run the process chain from that step on to the end.
    You need to set the failed request/step to green in the database as well as you need to raise the event that will force the process chain to run to the end from the next request/step on.
    Therefore you need to open the messages of a failed step by right clicking on it and selecting 'display messages'.
    In the opened popup click on the tab 'Chain'.
    In a parallel session goto transaction se16 for table rspcprocesslog and display the entries with the following selections:
    1. copy the variant from the popup to the variante of table rspcprocesslog
    2. copy the instance from the popup to the instance of table rspcprocesslog
    3. copy the start date from the popup to the batchdate of table rspcprocesslog
    Press F8 to display the entries of table rspcprocesslog.
    Now open another session and goto transaction se37. Enter RSPC_PROCESS_FINISH as the name of the function module and run the fm in test mode.
    Now copy the entries of table rspcprocesslog to the input parameters of the function module like described as follows:
    1. rspcprocesslog-log_id -> i_logid
    2. rspcprocesslog-type -> i_type
    3. rspcprocesslog-variante -> i_variant
    4. rspcprocesslog-instance -> i_instance
    5. enter 'G' for parameter i_state (sets the status to green).
    Now press F8 to run the fm.
    Now the actual process will be set to green and the following process in the chain will be started and the chain can run to the end.
    Of course you can also set the state of a specific step in the chain to any other possible value like 'R' = ended with errors, 'F' = finished, 'X' = cancelled ....
    Check out the value help on field rspcprocesslog-state in transaction se16 for the possible values.
    Thanks,
    JituK

  • Triggering Process chain based on execution of R3 job and Process Chain

    Hi All,
    I need your help in arriving at a solution for a reqirement.
    Scenario is., I need to trigger a process chain B based on successful execution of Process chain A and an R3 job. If both conditions(Completion of R3 job and Process Chain A) are met then only it has to trigger process Chain B.
    I know that we can use events to trigger a process chain using R3 job.But my case is differenet I need to consider the success message from both process chain and R3 job.Is there any way to solve it ?
    Please provide me with your valuable inputs.
    Thanks,
    Bindu

    Hi Hima,
    You can use the  'AND' variant for both, If both are success then trigger the process chain.
    Regards,
    Ravi Kanth
    Edited by: Ravi kanth on Apr 30, 2009 3:36 PM

  • Process chain Error: This AND process is not waiting for event RSPROCESS

    Hi All,
    I am facing an error in the process chain in PRD.
    Error message: This AND process is not waiting for event RSPROCESS.
    We had a process chain which had two sub chains which run parallel and below of this we had AND process type and below of the AND process we had 4 more jobs.
    Once the two subchains got successfully loaded, the and process should turn to green and further 4 jobs should start.
    It is a daily load , it worked fine from past years,but suddenly the AND processtype is getting failed.The thing is in the two subchains if one got completed, the AND process process is not waiting for 2nd subchain to get completed.The AND process is turning to RED (status: cancelled) . I tried to repeat the AND process once after above 2 subchains loaded, but it didnt worked.
    And i removed the existing AND process and created a new one and placed it in same place and activated and schedules again.but still it is getting failed with same error.
    Thanks in advance,
    Sai Chand.S

    Hi,
    If you did any transports related to that metachain we face similar kind of issues.
    not only the AND process , we need to remove all the process before executing the AND process and create it again.
    It helps you, you said your in production So you should take the proper approvals and do it.
    Regards,
    Yerrabelli.

  • How to solve delay in a program with data aquisition and processing

    Hello, I am a starter in Labview programing. I am working on a system which contains a roller, a piston and a A/D cart which is from Data Translation Inc (DT304). I am using labview to get speed data of the roller (which is voltage first then be converted into speed) from Analog Input Channels, and then using Analog Output Channels to control the friction force being appylied to the roller by using a piston.
    I am using a Waveform Chart to show the speed of roller. However, as I am adding more components to the program, I always get delay in the speed display and also the output control of the piston. Also, as I spent more time using the program eg. from morning to night, it became slower too.
    My question is, how does labview store data and how are the data been stored and released in the buffer.Is it because I am loading too much data in the buffer so that it became slower? And also how to solve these delays?
    I am using the "save to file" function to write data to a ".lvm" file, but I did some changes to the saving by using a button to enable and disable saving. So I don't think the problem is because I am always saving the data into a file.
    Thank you very much.
    Jessie

    I'm betting it's two things:
    1. Program Architecture
    and 2. Dynamically resizing arrays.
    With the program architecture... if you have one do-while loop that acquires data from teh DAQ, processes it, opens/indexes/writes/closes a file, then comes back around... as your arrays and files increase in size, the loop is going to take longer to execute.  A Producer-Consumer loop (a good tutorial on them can be found here) has one loop that acquires data and stuffs the data into a queue.  This will buffer the data while it's in transition between acquisition and processing.  A second loop in parallel takes data in from the queue, processes it, then comes back around.  The two loops operate independently of each other, so even if the consumer loop takes longer as the files or math gets more complex, the producer loop continues to run full speed.
    Second is the arrays.  Every time you append data into an array, LabVIEW has to make a copy of the data that's in the array.  If you append small amounts of data to an array over and over and over again, eventually LabVIEW is going to be copying very large amounts of data over and over and over again.  The producer-consumer architecture can alleviate this problem.

  • Dequeue from JMS Queue 10 at a time and process it.

    Hi,
    I have created a JMS queue in weblogic.
    Say i enqueue 1000 records in the queue at one go.
    Now, i want a bpel process to dequue from it and process it.
    However as per my requirement i want at a time 10 instances being active at a time i.e. 10 records from the queue are being processed at a time.
    Once any one of them finishes, the other will pick up the message.
    1 BPEL instance will process only 1 record from Q at a time
    Please let me know how to achieve this.

    You need to understand here that it is adapter which deques the message from queue and it will deliver it to BPEL for processing which is an async process. So adapter will keep dequeing messages and keep publishing it to BPEL irrespective of whether BPEL process has completed processing.
    You may need to tweak the adapter configuration here. First of all create the synchronous BPEL composite so that both adapter and BPEL work in same transaction -
    https://blogs.oracle.com/adapters/entry/creating_a_synchronous_bpel_composite_using_file_adapter
    You may control the number of threads for inbound JMS adapter using binding property "adapter.jms.receive.threads" in composite.xml
    Regards,
    Anuj

  • Ask the Expert: Overview of Cisco Prime Service Catalog and Process Orchestrator Solutions

    Welcome to this Cisco Support Community Ask the Expert conversation. This is an opportunity to learn and ask questions about the Cisco Prime Service Catalog and Process Orchestrator solutions.
    Cisco expert Jason Davis will discuss Cisco’s network management products offered under the Cisco Prime framework. If you have questions about Cisco Prime infrastructure or data center automation with our Cisco Prime Service Catalog and Process Orchestrator solutions, join us on the Cisco Support Community.
    Jason Davis is a distinguished services engineer in the Intelligent Infrastructure Practice team of Cisco Advanced Services. His role is to provide strategic and tactical consulting for hundreds of Advanced Services customers, lead service innovation, and assess new services and technologies. Jason's primary expertise areas are in network management systems, intelligent automation, virtualization, data center operations, software-defined networking, and network programmability.
    Based out of the Research Triangle Park (RTP) campus, Jason is also responsible for administering the Research Triangle Park Network Management Lab, Cisco's largest network management lab.
    Since joining Cisco in 1998, Jason has been a frequent speaker at Cisco's Networkers and CiscoLive conferences in the United States and Europe. In the past five years he has also been involved in the conference network setup and monitoring. He is a much sought-after resource by the field sales teams to assist with presales solutions and executive briefings. He has provided strategic and tactical network management consulting for several hundred customers.
    Jason is a subject matter expert with the following products and features:
    Cisco Prime LAN management solution
    Cisco Prime infrastructure
    CiscoSecure ACS
    Cisco Prime Network Registrar
    Cisco Process Orchestrator
    Cisco Prime Service Catalog
    Cisco IP SLA
    Embedded Event Manager
    SNMPv3
    onePK and OpenFlow
    Cisco UCS
    Device instrumentation
    VMware ESX, ESXi, and vCenter
    ITIL
    Jason received his bachelor of science degree in electrical engineering from the University of Miami (FL). He has been married for 20 years and has 4 children. His interests include providing audiovisual technical support for churches and conference venues, camping and biking with his family, remote-control helicopter piloting, paintball, and recreational shooting.
    Remember to use the rating system to let Jason know if you have received an adequate response.
    Because of the volume expected during this event, Jason might not be able to answer every question. Remember that you can continue the conversation in Data Center > Intelligent Automation under the subcommunity Cisco Prime Service Catalog shortly after the event. This event lasts through September 12, 2014. Visit this forum often to view responses to your questions and the questions of other Cisco Support Community members.

    Hello Jason,
    Thank you very much for welcoming me to your expert discussion :) I feel to be in the right place, at the right time. Thank you also for answering question beyond your scope here, much appreciated. The information received will help me to go further as such I have submitted a 5 start rating for your first reply.
    That sounds promising about the LMS part so yes, I stay tuned and wait patiently.
    Ok, now let’s revert to the actual topic discussed here. Cisco Prime Service Catalog and Process Orchestrator solutions I have briefly read up on this on CCO (where elseJ) and picked out the following quote
    ---- Quote from the Cisco Prime Service Catalog Data Sheet
     Today’s end users want self-service and easy access to IT tools and services.
    Simultaneously, organizations are seeking ways to extend their cloud management
    platforms beyond self-service delivery of virtual machines and infrastructure resources
    while increasing their use of cloud-based solutions to enhance business agility and effectiveness.
    Cisco Prime™ Service Catalog offers tremendous benefits to organizations that want to unify the ways in
    which all types of IT services are ordered and fulfilled, not just infrastructure requests
    ---- un quote ---
    I try to understand what (at high level of course) happens in the back ground when an order is raised and which vendor solution your product can interact with.
    As mentioned in the quoted text, this service catalogue goes beyond the standard infrastructure.
    Let’s say, a user wants to deploy a new email services, or in your example,  extends or create a new web-portal (i.e. for HR to view and manage holiday, staff absence and benefits).
    Your solution will need to interact somehow with the 3rd party vendor application that is capable building such portal I believe.
    Without disclosing to many information, I assume the portal is linked to backend VM,s that spin up requested resources (and more magic of course). Perhaps I am mixing this up with another cisco product where a user can go on the portal and spin up virtual Firewalls, virtual Routers can be provisioned in now time.
    Out if interest; Is this product also known as Mozart? (project code within Cisco?)
    I hope query is ok.
    Best wishes
    Markus

Maybe you are looking for