Automate BPM Process

Hi,
We are part of a testing team and we would like to automate the BPM process from start to finsih. Would PAPI / PAPI-WS help us for automation?
Some of the activities are of "Screenflow" type using "BPM Object" as "input". We are able to instantiate the process through PAPI. But we are getting the below exception while tryng to exceute an interactive activity. I saw in some threads that you can't execute an interactive activity that has a input method.
Is there a way to programmatically excute such tasks/activities? We would like to be able to automate the entire process for our testing.
Any suggestions would be greatly appreciated.
Thanks
Raj
Exception :
Could not perform the requested operationfuego.lang.ComponentExecutionException: The method 'input' from class 'fuego.xobject.runtime.components.XObjectInputImpl' could not be successfully executed.
     at fuego.component.InvocationImpl.doInvoke(InvocationImpl.java:141)
     at fuego.component.InvocationImpl.invoke(InvocationImpl.java:268)
Edited by: rajiblj on Apr 16, 2009 7:44 AM

Currently how are you getting file? The option is, in your InfoPackage, you can specify the source file from Application server. usually /usr/sap/<system name>/SYS/global etc.
Now you can decided one folder on the application server from where you will be uploading file to BW data targets. Put that path with file name into your Infopackage and schedule that InfoPackage through Process Chain.
If you are getting this file from 3rd party, you can ask them to upload this path to your SAP application server on teh specified folder.
You can also write a shell script in Unix or batch script in Windows to copy file from one folder to your desired folder on application server. It depends on the OS of your SAP system.
- Danny.

Similar Messages

  • Start BPM process automatically

    Hi ALL,
    I have created a BPM process. I can start that process automatically from nwa -> Process controll.
    But I want to run the process automatically by a certain time interval. How can I do it.
    Can anybody help me.
    Thanks in advance
    Chandan

    Hi Chandan,
    I am not 100% sure if I understood you correctly, but with BPMN there comes a model element named Intermediate Timer Event which can be configured in two different ways using:
    an offset (e.g. 1 hour after the timer has been created)
    an expression  (e.g. ProcessContext.aDateOneEntered + 1 day)
    Nevertheless this still requires you to start the process by yourself, but it 'waits' at the timer until the offset or the expression has been reached. Thereby you can achieve that an interval is taken into account upon execution of the process.
    More information about events can be found in the documentation:
    Modeling Events
    https://help.sap.com/saphelp_nwce711/helpdata/en/47/61470236013df6e10000000a42189b/frameset.htm
    Hope that helps,
    Martin

  • How to create automatic timer process in BPM?

    Can anyone tell me how to create an automatic timer process in BPM?
    Can I just create an manual process, and drag the timer event into the swam lane, can connect the timer to the service and then to the end event?
    Thank you

    Hi,
    You can make use of Global Automatic activity to kick off your process at a particular time. Place a Global Automatic activity in your process and right click on the Global Automatic activity and select Properties --> General --> Global Automatic Type: Automatic Schedule, Automatic Schedule: <Configure with your requirement>.
    Else you can make use of due transition/timer activity to get your task done.
    Bibhu

  • How to catch and handle soap foult in a BPM process?

    Hi,
    I have a BPM process which calls a webservice in one of its activities. The webservice throws an exception so the response will contain a soap fault element. I defined exception handlers to handle the error but none of the exception handlers are called. How can I catch and handle these types of error?
    The log contains this:
    A component failed while executing activity '/Process1#Default-1.0/Automatic' (BP-method Automatic) over instance '/Process1#Default-1.0/3/0'.
    Details:
    The method 'CIL_automatic' from class 'BEA_Systems__Inc__.Process1.Default_1_0.Instance' could not be successfully executed.
    Caused by: <?xml version="1.0" encoding="UTF-8" ?>
    <MyException faultString="" faultCode="{http://schemas.xmlsoap.org/soap/envelope/}Server"/>
    fuego.lang.ComponentExecutionException: The method 'CIL_automatic' from class 'BEA_Systems__Inc__.Process1.Default_1_0.Instance' could not be successfully executed.
         at fuego.component.ExecutionThreadContext.invokeMethod(ExecutionThreadContext.java:498)
         at fuego.component.ExecutionThreadContext.invokeMethod(ExecutionThreadContext.java:251)
         at fuego.fengine.FEEngineExecutionContext.invokeMethodAsCil(FEEngineExecutionContext.java:214)
         at fuego.server.execution.EngineExecutionContext.runCil(EngineExecutionContext.java:1068)
         at fuego.server.execution.TaskExecution.invoke(TaskExecution.java:389)
         at fuego.server.execution.TaskExecution.executeCIL(TaskExecution.java:481)
         at fuego.server.execution.TaskExecution.executeTask(TaskExecution.java:655)
         at fuego.server.execution.TaskExecution.executeTask(TaskExecution.java:616)
         at fuego.server.execution.TaskExecution.executeTask(TaskExecution.java:152)
         at fuego.server.execution.activities.XAutomatic.execute(XAutomatic.java:60)
         at fuego.metadata.Activity.execute(Activity.java:1022)
         at fuego.server.execution.ToDoItemAutomatic.execute(ToDoItemAutomatic.java:35)
         at fuego.server.execution.DefaultEngineExecution$AtomicExecutionTA.runTransaction(DefaultEngineExecution.java:290)
         at fuego.transaction.TransactionAction.startBaseTransaction(TransactionAction.java:462)
         at fuego.transaction.TransactionAction.startTransaction(TransactionAction.java:540)
         at fuego.transaction.TransactionAction.start(TransactionAction.java:213)
         at fuego.server.execution.DefaultEngineExecution.executeImmediate(DefaultEngineExecution.java:117)
         at fuego.server.execution.DefaultEngineExecution.executeAutomaticWork(DefaultEngineExecution.java:57)
         at fuego.server.execution.EngineExecution.executeAutomaticWork(EngineExecution.java:42)
         at fuego.server.execution.ToDoItem.executeAutomaticWork(ToDoItem.java:264)
         at fuego.server.execution.ToDoItem.run(ToDoItem.java:531)
         at fuego.component.ExecutionThread.processMessage(ExecutionThread.java:754)
         at fuego.component.ExecutionThread.processBatch(ExecutionThread.java:734)
         at fuego.component.ExecutionThread.doProcessBatch(ExecutionThread.java:140)
         at fuego.component.ExecutionThread.doProcessBatch(ExecutionThread.java:132)
         at fuego.fengine.ToDoQueueThread$PrincipalWrapper.processBatch(ToDoQueueThread.java:432)
         at fuego.component.ExecutionThread.work(ExecutionThread.java:818)
         at fuego.component.ExecutionThread.run(ExecutionThread.java:397)
    Caused by: fuego.soaptype.SoapExecutionException: <?xml version="1.0" encoding="UTF-8" ?>
    <MyException faultString="" faultCode="{http://schemas.xmlsoap.org/soap/envelope/}Server"/>
         at fuego.soaptype.SoapCall.processRemoteException(SoapCall.java:531)
         at fuego.soaptype.SoapCall.invoke(SoapCall.java:292)
         at fuego.soaptype.SoapObject.invoke(SoapObject.java:272)
         at fuego.lang.Invokeable.invokeImpl(Invokeable.java:220)
         at fuego.lang.Invokeable.invoke(Invokeable.java:161)
         at BEA_Systems__Inc__.Process1.Default_1_0.Instance.CIL_automatic(Instance.java:265)
         at BEA_Systems__Inc__.Process1.Default_1_0.Instance.CIL_automatic(Instance.java:281)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at fuego.component.ExecutionThreadContext.invokeMethod(ExecutionThreadContext.java:491)
         ... 27 more
    Thanks,
    fifty

    Hi fifty,
    Did you get a solution to the above problem you have mentioned? I have a similar issue i am trying to fix.
    I have a webservice call in a process activity and if the call does not work i get a soap fault and the fuego.lang.ComponentExecutionException . My process requires that i catch the exception infact any kind of exceptions that occur on that call and perform another activity in the process.
    I have defined an exception handler at the activity level for java.lang.Exception and java.lang.RunTimeException.
    i don't see anything in the catalog which would handle the SOAP fault OR the componentexception.

  • BPM process analytics

    Hi
    We are trying to report on BPM process analytics.
    Has anyone succesfully implemented the same.
    We have some basic questions on it
    1. where is the process context and generic data stored(is it in ECC backend or CE's database or file system)
    2. how to pull data from ce to BW (do the datasources get created automatically in source system?)
    3. Is there a way to pull data without using BI content datasources. (process context and generic)
    Please clarify.
    Regards
    Lavanya

    Hi Lavanya,
    to 1). The process context is stored on the CE server. Likewise, SAP delivers some predefined data sources that you can report on directly using Visual Composer or
    to 2) you can extract the information into your BW system. All the BI objects needed for that are part of BI Content 7.05 (it might be in an earlier version but I know it's in 7.05)
    to 3) Maybe, :-). Haven't tried it and I don't know what the underlying data model is but it may be possible to read the tables directly.
    Here's a good link to get started.
    http://help.sap.com/saphelp_nwce72/helpdata/en/60/794d3f1e5a4443b5f714b28f6f5fa1/content.htm
    HTH.
    O.

  • BPM Process deleted from ESR / ID but still visible in SXI_CACHE

    Hi,
    We are having some issues with BPM where even after deleteing the BPM processes from PI  ESR and its corresponsing ID reference the processes are still visible in SXI_CACHE.
    Can any one point to how we can get the CACHE alligned to wht is in the ESR and ID.
    Regards,
    Arunava

    where even after deleting the BPM processes from PI ESR and its corresponding ID reference the processes are still visible in SXI_CACHE.
    It gets refreshed automatically may be not immediately but daily , if you check the SXI_Cache after some time the deleted one will not appear.
    It will be visible in SXI_CACHE,please check the return value it should be 0 not 99.
    Try creating the IP with some other name , copy the existing one with new name and import in ID and try.
    Regards,
    Srinivas

  • How to start a BPM process from ERP system

    Hi,
    I created a process in Netweaver Developer Studio with the Process Composer in SAP CE 7.1 BPM, now I want to automatically start the process by saving a SAP document like purchase requisition or purchase order. Therefore I think about implementing a UserExit / Badi or something else.
    The following systems are available in my Landascape:
    - PI 7.1
    - ERP 2005
    - CE 7.1 EHP1 with ESR
    Does someone have an idea how to start the BPM Process or how to trigger start event?
    Greetings,
    Christian

    the process by saving a SAP document like purchase requisition or purchase order.
    Based on your business process, you will have to decide what data should be sent to other system. It could be in the form of Idoc trigger or an RFC calls. If there are non-SAP systems involved, you should be using PI 7.1 too. In that case, you can communicate using Abap proxy call too. For this you may either call an RFC internally in a report and schedule it.
    Regards,
    Prateek

  • A BPM process with out any human task

    Hi All,
    I am new bie to BPM. Can I create a BPM process without any human task?
    Like start event, service task (or any other activity) & end event.
    If it is possible, please let me know the proper steps to execute and test this process also.
    Thanks,

    Hi
    Yes, ofcourse you can have a Process without any User Task. Ex: Start Node -> Service Task -> End Node. Start Node can be 2 things say like DB Adapater and File Adapter. Read something from DB from some table and put into the Text File.
    But the thing is, BPM Process is more of Role Oriented. If you already saw, you define something called Swim Lanes and put each User Task in those Lanes and automatically users in that Role only can work on those Task.
    If you do not have any such things like User Task, you don't even need to create a BPM Process. Just create a simple BPEL with your custom Payload.
    Now, the most important question is how do you plan to trigger/start this process whether BPM or BPEL. Do you assign some users to the first Role and do users log into bpm/workspace and click the Link on the left side to start the process. ALSO since there is NO User Task, you will NOT have any Task Details Page and this means you MAY NOT even see the link on the left side in bpm/workspace.
    The only option I could see is have BPEL. This BPEL will have a WSDL automatically generated. And you can use this WSDL to instantiate the task.
    If your intention is to use this process/bpel in some other MAIN Process then it is totally possible.
    Thanks
    Ravi Jegga

  • Calling ADF page Custom Listener (Ex: ActionListener) from BPM Process Task (APPROVE/REJECT)

    Hi All,
    Jdeveloper version - 11.1.1.7
    I am very new to BPM / SOA development, but I have very good development skills on ADF.
    I am not using ADF BC, using EJB for business services and also using custom ADF pages for HumanTasks.
    Usecase:
    From the BPM Process task, when process submission (APPROVE / REJECT) , I need to invoke a Custom listener (Ex.Action Listener - a EJB call) in the ADF page.
    I am trying to use BPM APIs.
    Please clarify me how this will achieve using BPM APIs. I need detail guidelines to do it. Please make to understand this process.
    Provide some useful documentation or links to understand the following:
    1. Custom Human task pages
    2. BPM APIs - 11.1.1.7
    3. Call a BPM process task from ADF Listener and Call a ADF Listener from BPM Process task - Using BPM APIs.
    Please revert more clarifications needed.
    Thanks and Regards
    Mohanraj N

    Hi Joonas.
    Plese let me explain me better for your understanding
    A big summary for what I meant it's the following:
    1- In the procces you made, when you add the HT activity, you have to implement it, this means declare the input(s) parameters you want. This implementation create the .task file.
    2- Create an application, and projects as HT you have. Each poject are based on the .task file, and automatically create a Data Control (for each project based on a .task) with all you need.
    This w'll be an empty application, so you can customize it all you want. The task selected should have all the parameters previously defined. Those parameters can change if you want.
    2- Create a page(s) in the task flow for the task implementation. You can even split the the payload of the task in differents pages, create your custom pages and any logic you need.
    3- An important aspect is how to match these application with the HT implemented in the process. It's possible, it's a configuration en the Enterprise Manager.
    4- Deploy your application
    All these are explain in the book I mentioned
    Th book you can find it here:
    https://blogs.oracle.com/soacommunity/entry/oracle_soa_suite_11g_handbook_1
    Regards Dariel.
    PS: Please, let me know if you need more details.

  • SXMB_ADM_BPE - Automatic BPM Customizing

    Hi Friends,
    When i have done simple FILE TO IDOC, it worked fine. but when I am trying to implement FILE TO IDOC scenario using BPM, it was failing.
    in SXI_CACHE, process status code is 99 instead of 0.
    checked conditions, it was fine.
    but when i went to
    SXMB_ADM_BPE -> Automatic BPM Customizing
    all are not green.
    i think this is causing problem. ideally all should be GREEN.
    Can you please suggest how to solve this issue?
    Waiting for your valuable suggestions.
    Thanks
    Chris

    yep it compiled fine without errors. to compile i did F7 in the business process.
    colin, usually why automatic bpm customizing were not all green?
    but what i found is
    SXI_CACHE -> entered into workflow mode.
    then i did syntactlical check results one error but not able to find that error?
    is this error caused because automatic bpm customizing were not all green?
    or they are all not green because some error has occured in the business process?
    please advise.
    thanks
    sam

  • Oracle BPM Process Data mart

    I am required to create audit reports on BPM workflows.
    I am new to thid & need some guidance on configuring BPM Process Data mart. What are the pre-requisites for configuring it & what are the steps to do it.
    Also, need some inputs on BAM database. What is the frequency of data upload. Is it data update or insert in BAM.

    Hi,
    You might want to check out the Administration and Configuration Guides on http://download.oracle.com/docs/cd/E13154_01/bpm/docs65/index.html.
    I suspect you might find the BAM and Data Mart portions of this documentation a bit terse, so I've added the steps below that provides more detail. I wrote this for ALBPM 6.0, but believe it will still work for Oracle BPM 10g. It was created from an earlier ALBPM 5.7 document Support wrote called "ALBPM 5_7 Configuring and Troubleshooting the BAM and DataMart Updater.pdf.
    You can define how often you want the contents in both databases updated (actually inserted) and how long you want to persist the contents of the BAM database during the configuration.
    Here's the contents of the document:
    1. Introduction
    The use of BAM (Business Activity Monitoring) and Data Mart (or Warehouse) information is becoming more and more widespread in today’s BPM project implementations for the obvious benefits they bring to the management and tuning of processes.
    BAM is basically composed by a collection of measurements of current processes load and execution times. This gives us an idea of how the business is doing at this moment (in a pseudo real-time fashion).
    Data Mart, on the other hand, is a historical view of the processes load and execution times. And this gives us an idea of how the business has developed since the moment the projects have been put in place.
    In this document we are not going to describe exhaustively all configuration aspects of the BAM and Data Mart Updater, but rather we will quickly move from one configuration step to another paying more attention to subjects that have presented some difficulties in real-life projects.
    2. Creating the Service Endpoints
    The databases for BAM and for Data Mart first have to be defined in the External Resources section of the BPM Process Administrator.
    In this following example the service endpoint ‘BAMJ2EEWL’ is being defined. This definition is going to be used later as BAM storage. At this point nothing is created.
    Add an External Resource with the name ‘BAMJ2EEWL’ and, as we use Oracle, select the Oracle driver, then click <Next>:
    On the following screen, specify:
    ·     the hostname – here I have used ‘localhost’ as I am just setting this up to work on my laptop
    ·     the port for the Oracle service
    ·     the SID – here I have use Oracle Express so the SID is ‘XE’
    ·     the new user to create / use in Oracle for this database – here I have specified ‘BPMBAM’. This user, and its database, will be created later
    ·     the password for the user
    Scroll down to the bottom of the page and click <Save>.
    In addition to a standard JDBC connection that is going to be used by the Updater Service, a remote JDBC configuration needs to be added as the Engine runs in a WebLogic J2EE container. This Data Source is needed to grant the Engine access over BAM tables thru the J2EE Connection Pool instead of thru a dedicated JDBC. The following is an example of how to set this up.
    Add an External Resource with the name ‘BAMremote’ and select the Oracle driver, then click <Next>
    On the following screen, specify:
    ·     the Lookup Name that will be used subsequently in WebLogic - here I have given it the name ‘XAbamDS’
    Then click <Save>.
    In the next example the definition ‘DWHJ2EEWL’ is created to be used later as Data Mart storage. If you are not going to use a Data Mart storage you can skip this step.
    Add an External Resource with the name ‘DWHJ2EEWL’ and select the Oracle driver, then click <Next>:
    On the following screen, specify:
    ·     the hostname – here I have used ‘localhost’ as I am just setting this up to work on my laptop
    ·     the port for the Oracle service
    ·     the SID – here I have use Oracle Express so the SID is ‘XE’
    ·     the new user to create / use in Oracle for this database – here I have specified ‘BPMDWH’. This user, and its database, will be created later
    ·     the password for the user
    3. Configuring BAM Updater Service
    Once the service endpoint has been created the next step is to enable the BAM update, select the service endpoint to be used as BAM storage and configure update frequency and others. Here the “Updater Database Configuration” is the standard JDBC we configured earlier and the “Runtime Database Configuration” is the Remote JDBC as we are using the J2EE Engine.
    So, here’s the example of how to set up the BAM Updater service….
    Go into ‘Process Monitoring’ and select the ‘BAM’ tab and enter the relevant information (using the names created earlier – use the drop down list to select):
    Note that here, to allow me to quickly test BAM reporting, I have set the update frequency to 1 minute. This would not be the production setting.
    Once the data is input, click <Save>.
    We now have to create the schema and related tables. For this we will open the “Manage Database” page that has appeared at the bottom of the BAM screen (you may have to re-select that Tab) and select to create the database and the data structure. The user required to perform this operation is the DB system administrator:
    Text showing the successful creation of the database and data structures should appear.
    Once we are done with the schema creation, we can move to the Process Data Mart configuration screen to set up the Common Updater Service parameters. Notice that the service has not been started yet… We will get to that point later.
    4. Configuring Process Data Mart Updater Service
    In the case that Data Mart information is not going to be used, the “Enable Automatic Update” checkbox must be left off and the “Runtime Database Configuration” empty for this service. Additionally, the rest of this section can be skipped.
    In the case it is going to be used, the detail level, snapshot time and the time of update should be configured; in addition to enabling the updater and choosing the storage configuration. An example is shown below:
    Still in ‘Process Monitoring’, select the ‘Process Data Mart’ tab and enter the name created earlier (use the drop down list to select).
    Also, un-tick the Generate O3 Cubes (see later notes):
    Then click <Save>.
    Once those properties have been configured the database and the data structure have to be created. This is performed at the “Manage Database” page for which the link has appeared at the bottom of the page (as with BAM). Even when this page is identical to the one shown above (for the BAM configuration) it has been opened from the link in the “Process Data Mart” page and this makes it different.
    Text showing the successful creation of the database and data structures should appear.
    5. Configuring Common Updater Service Parameters
    In the “Process Data Mart” tab of the Process Monitoring section -along with the parameters that are specific to the Data Mart - we will find some parameters that are common to all services. These parameters are:
    • Log directory: location of the log file
    • Messages logged from Data Store Updater: severity level of the Updater logs
    • Language
    • Generate Performance Metrics: enables performance metrics generation
    • Generate Workload Metrics: enables workload metrics generation
    • Generate O3 Cubes: enables O3 Cubes generation
    In this document we are not going to describe in detail each parameter. But we will mention a few caveats:
    a. the Log directory must be specified in order for the logs to be generated
    b. the Messages logged from Data Store Updater, which indicates the level
    of the logs, should be DEBUG for troubleshooting and WARNING otherwise
    c. Performance and Workload Metrics need to be on for the typical BAM usage and, even when either metric might not be used on the initial project releases, it is recommended to leave them on in case they turn out to be useful in the future
    d. the Generation of O3 Cubes must be off if this service is not used, otherwise the Data Mart Updater service might not work properly .
    The only changes required on this screen was to de-select the ‘Generate O3 Cubes’ as shown in the last section.
    6. Set up the WebLogic configuration
    We need to set up the JDBC data source specified above, so go to Services / JDBC / Data Sources.
    Click on <Lock and Edit> and then <New> to add a New data source.
    Specify:
    ·     the Name – use the name you set up in the Process Administrator
    ·     the JNDI Name – again use the name you set up in the Process Administrator
    ·     the Database Type – Oracle
    ·     use the default Oracle Database Driver
    Then click <Next>
    On the next screen, click <Next>
    On the next screen specify:
    ·     the Database Name – this is the SID – for me that is XE
    ·     the Host Name – as I am running on my laptop, I’ve just specified ‘localhost’
    ·     the Database User Name and Password – this is the BAM database user specified in the Process Administrator
    Then click <Next>
    On the next screen, you can test the configuration to make sure you have got it right, then click <Next>
    On the next screen, select your server as the target server and click <Finish>:
    Finally, click <Activate Changes>.
    7. The Last Step: Starting Up and Shutting Down the Updater Service
    ALBPM distribution is different depending on the Operating System. In the case of the Updater Service:
    -     For Unix like Operating Systems the service is started or stopped with the albpmwarehouse.sh shell script. The command in this case is going to look like this:
    $ALBPM_HOME/bin$ ./albpmwarehouse.sh start
    -     For Windows Operating Systems the service is installed or uninstalled as a Windows Service with the albpmwarehouse.bat batch file. The command will look like:
    %ALBPM_HOME%\bin> albpmwarehouse.bat install
    After installing the service, it has to be started|stopped from the Microsoft Management Console. Note also that Windows will start automatically the installed service when the computer starts. In either case the location of the script is ALBPM_HOME/bin Where ALBPM_HOME is the ALBPM installation directory. An example will be:
    C:\bea\albpm6.0\j2eewl\bin\albpmwarehouse.bat
    8. Finally: Running BAM dashboards to show it is Working
    Now we have finally got the BAM service running, we can run dashboards from within Workspace and see the results:
    9. General BAM and Data Mart Caveats
    a. The basic difference between these two collections of measurements is that BAM keeps track of current processes load and execution times while Data Mart contains a historical view of those same measurements. This is why BAM information is collected frequently (every minute) and cleared out every several hours (or every day) and why Data Mart is updated infrequently (once a day) and grows indefinitely. Moreover, BAM measurements can be though of as a minute-by-minute sequence of Engine Events snapshots, while Data Mart measurements will be a daily sequence of Engine Events snapshots.
    b. BAM and Data Mart table schemas are very similar but they are not the same. Thus, it is important not to use a schema created with the Manage Database for BAM as Data Mart storage or vice-versa. If these schemas are exchanged by mistake, the updater service will run anyway but no data will be added to the tables and there will be errors in the log indicating that the schema is incorrect or that some tables could not be found.
    c. BAM and Data Mart Information and Services are independent from one another. Any of them can be configured and running without the other one. The information is extracted directly from the Engine Database (PPROCINSTEVENT table is the main source of info) for both of them.
    d. So far there has not been a mention of engines, projects or processes in any of the BAM or Data Mart configurations. This is because the metrics of all projects published under the current Process Administrator (or, more precisely, FDI Directory) are going to be collected.
    e. It is also important to note that only activities for which events are generated are going to be measured (and therefore, shown in the metrics). The project default is to generate events only for Interactive activities. This can be changed for any particular activity and for the whole process (where the activity setting, when specified, overrides the process setting). Unfortunately, there is no project setting for events generation so far; thus, remember to edit the level of event generation for every new process that is added to the project.
    f. BAM and Data Mart metrics are usually enriched with Business Variables. These variables are a special type of External Variables. An External Variable is a process variable with the scope of an Instance and whose value is stored on a separate column in the Engine Instances table. This allows the creation of views and filters based on this variable. A Business Variable, then, shares all the properties of an External Variable plus the fact that its value is collected in all BAM and Data Mart measurements (in some cases the value is shown as it is for a particular instance and in others the value is aggregated).
    The caveat here is that there is a maximum number of 256 Business Variables per FDI. Therefore, when publishing several projects into a single FDI directory it is recommendable to reuse business variables. This is achieved by mapping similar Business Variables of different projects with a unique real Variable (on the variable mapping performed at publish time).
    g. Configuring the Updater Service Log
    In section 5. Configuring Common Updater Service Parameters we have seen that there are two common Updater properties related to logging. These properties are “Log directory” and “Messages logged from Data Store Updater”, and they specify the location and level of these two files:
    - dwupdater.log: which is the log for the Data Mart updater service
    - bam-dwupdater.log: which is the log for the BAM updater service
    In addition to these two properties, there is a configuration file called ‘WarehouseService.conf’ that allows us to modify these other properties:
    - wrapper.console.loglevel: level for the updater service log
    - wrapper.logfile.loglevel: level for the updater service log
    - wrapper.java.additional.n: additional argument to the service JVM
    - wrapper.logfile.maxsize: maximum size of the updater service log files
    - wrapper.logfile.maxfiles: maximum number of updater service log files
    - wrapper.logfile: updater service log file name (the default value is dwupdater-service.log)
    9.1. Updater Service Log Configuration Caveats
    a. The first three parameters listed above have to be modified when increasing the log level to DEBUG (since the default is WARNING). The loglevel parameters have to be set to DEBUG and a java.additional.n (where n is a consecutive integer to the already used ones) has to be set to –ea to enable asserts, since without this option no DEBUG message is going to be generated.
    b. Of the other arguments, maxfiles might need to be increased to hold a few more days of data when the log level is set to DEBUG (with the default value up to two days are stored).
    c. The updater service has to be stopped, uninstalled, installed and then started for any of these changes to take effect.
    Hope this helps,
    Dan

  • Invoke Java code within BPM process

    Hi,
    I have a use case where I need to retrieve a property from database and set the value to a process variable. Is there a way I can achieve this?
    Also, is there a way to invoke a piece of java code from the automatic activity within a BPM process?
    Thanks,
    Suren.

    The best way to do this is to use the spring component. You will need your java code setup as a spring bean in the spring context file and then it can be exposed to the BPM process as an sca service.
    http://docs.oracle.com/cd/E21764_01/integration.1111/e10224/sca_spring.htm

  • BPM Process chain takes long time to process

    We have BI7, Netweaver 2004s on Oracle and SUN Solaris
    There is a process chain (BPM) which pulls data from the CRM system into BW. The scheduled time to run this chain is 0034 hrs. This chain should ideally complete before / around 0830 Hrs. <b>Now the problem is that every alternate day this chain behaves normally and gets completed well before 0830 hrs but every alternate day this chain fails…</b> there are almost 40 chains running daily. Some are event triggered (dependent with each other) or some run in parallel. In this, (BPM) process chain, usually there are 5 requests with 3 Delta and 2 full uploads (Master Data). The delta uploads finishes in 30 minutes without any issues with very few record transfers. The first full upload is from 0034 hrs to approximately 0130 hrs and the 2nd upload is from 0130 hrs to 0230 hrs. Now if the 1st upload gets delayed then the people who are initiating these chains, stop the 2nd full upload and continue it after all the process chains are completed. Now this entire BPM process chain sometimes takes 17 -18 hrs to complete!!!!!
    No other loads in CRM or BW when these process chains are running
    CRM has background jobs to push IDOCS to BW which run every 2 minutes which runs successfully
    Yesterday this chain got completed successfully (well within stipulated time) with over 33,00,000 records transferred but sometimes it has failed to transfer even 12,00,000 records!!
    Attaching a zip file, please refer the “21 to 26 Analysis screen shot.doc” from the zip file
    Within the zip file, attaching “Normal timings of daily process chains.xls” – the name explains it….
    Also within the zip file refer “BPM Infoprovider and data source screen shot.doc” please refer this file as the infopackage (page 2) which was used in the process chain is not displayed later on in page number 6 BUT CHAIN GOT SUCESSFULLY COMPLETED
    We have analyzed:--
    1)     The PSA data for BPM process chain for past few days
    2)     The info providers for BPM process chain for past few days
    3)     The ODS entries for BPM process chain for past few days
    4)     The point of failure of BPM process chain for past few days
    5)     The overall performance of all the process chains for past few days
    6)     The number of requests in BW for this process chain
    7)     The load on CRM system for past few days when this process chain ran on BW system
    As per our analysis, there are couple of things which can be fixed in the BW system:--
    1)     The partner agreement (transaction WE20) defined for the partner LS/BP3CLNT475 mentions both message types RSSEND and RSINFO: -- collect IDOCs and pack size = 1 Since the pack size = 1 will generate 1 TRFC call per IDOC, it should be changed to 10 so that less number of TRFCs will be generated thus less overhead for the BW server resulting in the increase in performance
    2)     In the definition of destination for the concerned RFC in BW (SM59), the “Technical Setting” tab says the “Load balancing” option = “No”. We are planning to make it “Yes”
    But we believe that though these changes will bring some increase in performance, this is not the root cause of the abnormal behavior of this chain as this chain runs successfully on every alternate day with approximately the same amount of load in it.
    I was not able to attach the many screen shots or the info which I had gathered during my analysis. Please advice how do I attach these files
    Best Regards,

    Hi,
    Normally  index  creation or deletion can take long time in case  your database statistics are not updated properly, so can check  stat  after your data loading is completed and index generation is done,  Do creation of database statistics.
    Then try to recheck ...
    Regards,
    Satya

  • Are 'read only' items causing my automatic row processing to fail?

    I have a form that uses automatic row processing to update my table.
    I recently had to add a check to lock down the fields using the read only condition where a record has been 'approved'. So, I turned off all of the fields and hid the update button.
    Now they want me to check to see if the user is a manager, and if so, allow them to see the update button, and make a single field on the form modifiable.
    I have the user identified as a manager, and can make the field accept the change (a yes/no combo box). Yet when they click the update button, the form just sits there. A branch is supposed to fire on hitting that button, taking them back to the previous report, but it doesn't happen. The source of the combo box in question uses a column on the table, and works fine when the form has all of the fields showing for modification.
    I had some validations that I finally set to never fire, thinking that they were failing and just not giving the error messages. I know a bad sql or pl/sql atatement in a validation will cause the form to just keep sitting there doing nothing, and thought it might help.
    Several of the various fields are combo boxes, several are text boxes.
    I'm stumped.
    Shall I keep the regular update button hidden whether it is a manager or not (because it has been 'approved'), and enable a new button that fires off a process to update the one field that the manager may want to change? Does making a field 'read only' always have a detrimental effect on the auto processes I've been counting on? Is that if fact what is happening now?
    Thanks in advance for any advice or comments on this.
    Clayton

    Hi, Scott.
    My version is 3.0.0.00.20
    I have received info to set up the application, though I have not activated it as yet. I'm reluctant to try to load up the various tables to make it work. It could be done.
    My workaround is to that the user (a manager) fires off to another form to show an individual record, and ask if he/she really wants to make this change. On submit, they will be taken back to the original locked down form, and I will update the single field in that process of returning them. This will actually be a plus, in that I expect to show them all of the 'duplicate' entries for that item, and show them that by 'activating' this particular item, they will 'deactivate' the one duplicate that is currently active. Or, if they are deactivating the sole item that is active, that there will in fact be no active items in the system. Convoluted as heck, but they swear this is how they've been doing business with their briefcases and file cabinets of Excel files. I'm surprise they even float at this point (these are ships we're talking about).
    This all may be due to the form originally having a next/previous/count process once arrived at my clicking on the edit button of the report of items. All of that has been hidden away, and though the fields exist of the form, they are not used. They only see the current item that they clicked on from the report.
    I read up a little on session state, as well as posts with the debug error I was getting. I'm still not on board with the whole issue of 'state'. Items stay as they are until they're changed, no? I suppose that sounds simplistic.
    One of the things that caught my eye was several combo box lists that on read only mode, showed [] square brackets. Which, in some cases is okay.
    What I'll try to do is set up a public application - for future use if nothing else. I'm going to be working on this for awhile. I have more than a bit of work to do to cleanse the data as well as get rid of the single sign on requirement (I'll assume that I should have to do this) in order to share the app.
    I know you're expecting me to have this up and running for you to look at, but I'm under the gun. And they actually like my workaround.
    I'll keep everybody informed, one way or another.
    Clayton

  • Dynamic Filename in BPM process (SOAP with attachm. and PayloadSwapBean)

    Hello together
    I have the following BPM process:
    1. IDoc=>WebServiceRequest
    2. WebServiceResponse (payload) => IDoc
    3. WebServiceResponse (attachment) => File
    XI receivs an IDoc an map it to an WebService. The Webservice is called by XI and we receive the WebServiceResponse including a PDF attachment.
    The challenge is to store the PDF attachment with a dynamic filename from the payload of WebServiceResponse.
    We use the PayloadSwapBean to change the payload to the PDF attachment. But then we are not able to access the required information on the original WebService-XML-Response via variable substitution.
    Is there a solution in the standard or have we to use a custom adapter module?
    Thx
    manuku

    Hi Jayasimha,
    We can do this by "Adapter Specific Message Properties" of ur comunication channels.
    1.If u want to keep the output filename same as input filename, no need to use the UDF. only the 'adapter specific parameters' in both sender n receiver file adapter will do that.
    In case if u want to get the filename inside our mapping we have to create a user defined function
    which will return the filename and map it to one of our XML tags. 2nd point gives solution 4 that:
    2. If u want to generate an output file taking some input from the payload,then u hav to use the UDF.There u hav to populate the name.
    Pretty much.... if you set an attribute from the sender side, for example, you can use a UDF and access the particular attribute sent and use it in the mapping. In another example, where no attributes are sent from the sender, you can still actually set a particular attribute, say a filename derived from the payload, using a UDF, and enable the receiver attribute to use it. That's where the UDFs come in - either to get or set particular adapter specific message attributes.
    This will be a very helpful blog which solves ur query:
    /people/michal.krawczyk2/blog/2005/11/10/xi-the-same-filename-from-a-sender-to-a-receiver-file-adapter--sp14
    Regards,
    Vinod.

Maybe you are looking for

  • Unable to clear customer through F-32

    We just went live today with our FM (BCS) Module. Everything is fine except when people is trying to clearing customers through F-32 We have the "No commitment itme in line 0001 xxxx xxxxxxxxxxx 2014" error when try to post the clearing document. We

  • ITunes library 11.0 (Windows 7) bug

    I updated my iTunes libray and I puchased a new song using version 11.0 iTunes library and in my songs Get Info my Apple ID and the Label Records was not in there.

  • Scorecard error "For input string" ....

    Hi has anyone had and solved the following error? When selecting "Last Run", one of my scorecards is returing an error "For Input string "7127912115" right after loading the Key data domain details. When selecting "Now" it works fine. However, when I

  • Urgent please send code for this one ( write code by using core java)

    hi sir my question i am taking one square box .. and i divide that square in again four parts .. now i am placing one object in 1st box now i want to move that object to 2 box or third box... like that only if my object is in 1st box now i want to ju

  • Contact sheet images are grayed out

    I have made an album of black and white images and want to print them as a series of contact sheets. The sheets look good when customized, but when I tell it to print, all but the first two of the previewed images are grayed out and not worth printin