ESS Job Definition and Testing

I have defined a ESS Job definition in Jdeveloper using the Using Oracle JDeveloper to Generate an Oracle Enterprise Scheduler Application - 11g Release 1 (11.1.1.6.1) and other supporting docs.
The roles attached in the Access control and already available in the jazn file.
I have edited the Resource Bundle and added the details for the new Job Definition.But unable to select the same in the Display Name Resource Key & Description Resource Key.
I have build the Ear file and deployed it in integrated weblogic server. But I'm unable to see the new job the in the Scheduled Process list.
Any help will be highly appreciated...
Thanks,

Thanks for the reply.
I had not added the Resource Bundle present in another project to this project Bundle list.
Once added, I was able to select.
I have another question.
In the Access Control section, I have added 2 roles, which are already present in the jazn-data.xml file.I select them, give the roles & when I save the job, I'm getting an error.
I have checked out the jazn-data.xml file present in my view. But the job is referring to jazn-data.xml file present in another location which is read only.
So I'm unable to save the job definition.
Also My main issue is The job I created & deployed in weblogic server is not visible in the ESS Testing UI.(http://localhost:7101/essUi/faces/EssUIShell)
P.S Pls let me know if I have to raise this as another question.I'll do the same.

Similar Messages

  • Unable to see Ess Job Definition in the ESS Test UI

    Hi,
    I am defining a new ESS Job. I have created the Jobdefinition.xml in the Jdev. I have defined the parameters in In the Application defined properties. In the system properties, I have added the Sys Effective Application & Sys_product. In the Access Control section, I have added 2 roles, which are already present in the jazn-data.xml file.I select them, give the roles with Read,write,delete & execute permissions.
    I have included the resource bundle, Display Name Resource key & Description Resource Key.
    But now when I save the job, I'm getting an error.
    It is referring to the Jazn file in some other location & not referring to the one which is present in my project. In my project, I have checked out the file and able to make changes.
    So I'm unable to save the job definition.
    Also My main issue is The job I created & deployed in weblogic server is not visible in the ESS Testing UI.(http://localhost:7101/essUi/faces/EssUIShell).
    Any help will be highly appreciated.
    Thanks,

    Hi,
    The issue is not occuring now.
    I could not identify the root cause but it looks like jazn-data.xml file was locked in another view. Though I have issued check out and it also showed check out, it was not actually available for writing.
    I unbranched and destroyed the transactions in my view and re-created the view, the issue did not occur then.
    Also I have created the job definition and now it is listing in the test UI but fails in the parameters page. I followed the steps mentioned in this post.
    http://myforums.oracle.com/jive3/thread.jspa?threadID=1289852&tstart=0
    But ended up in BUILD FAILURE for some reasons.
    A how-to series will be very helpful.
    Thanks,
    Sangheetha.

  • ESS job definition migration

    Hi All,
    I have created the ESS job in development environment.Now I have to migrate the definition to Test environment.In Oracle apps we have some thing like FNDLOAD to download the concurrent program definition.
    Could you please let me know,is there any similar kind of utility like FNDLOAD for ESS job.
    Regards,
    Venkat.

    You can use WLST to transfer metadata, for details on the commands refer to  Managing the Metadata Repository. For detailed example refer to
    How do Users View Enterprise Scheduler Service (ESS) Job Metadata? (Doc ID 1358281.1)
    Jani Rautiainen
    Fusion Applications Developer Relations
    https://blogs.oracle.com/fadevrel/

  • Background Job definition and relationship between jobs

    Hi All,
    I have 2 jobs J1 and J2. After finishing J1, J2 should start. While defining through SM36, if we use After Job in Start Condition, first time the runs are ok, but if we wish to run it periodic, the dependency is lost. Can someone help me out??
    RSPC option is not there.
    thanks,
    Partha Maity

    Hi,
    I have tried the option of Job definition through SM36, it works for the first time as i have said, in periodic running dependency is lost. Tried again , same output.
    Defining a single job, we have tried this option, and it is taking unusally longer time, thats why we want to have separate job,. and one single big job fails we have very little option of recovery.
    RSPC option is not there in APO 3.1
    By events, how this can be acheived can some one give detail workout.
    Thanks in anticipation.
    Partha Maity

  • SAP Scheduler. Multiple jobs for one job definition issue.

    Hi.
    I am facing one problem concerning scheduling jobs. I've created a job definition and deployed it on server. In onJob(...) method system queries to DB for a single result, when this result matches some conditions job is finished, otherwise it waits for a minute and then queries again. Then I tried to create several tasks for this job definition and set similar execution time for all tasks to check their simultaneous behavior. I was surprised to find out that only one job status is 'RUNNING' and for others status is 'STARTING'. But I expected to see all my jobs in status 'RUNNING'. At first I thought the problem was in MDB pool stuff(e.g. InitialSize of beans or something like that), supposing that there is only one working bean and huge queue  of messages. But nothing changed after setting different values for pool params.
    So, can anyone help me to make at least 2 simultaneous working  jobs for one job definition, if it possible, off cause. Or may be I am wrong somewhere. (By the way, I used Threads in MDB for waiting).
    Thanks in advance.

    May be I described my problem rather messy. The main question is it possible to create several jobs for one job definition and see them running at the same time (simultaneously)?
    Thanks.

  • How to migrate ESS job from one instance to another Instance ?

    Hi Experts,
    We have developed some custom ESS jobs in fusion applications. Now we have to migrate newly created ESS jobs to another environment.
    Is there any way like we use to extract ldt in EBS R12.
    Kindly suggest......!
    Thanks,
    Praveen

    Please refer to:
    How to Migrate ESS Job Definitions Using WLST (1941788.1)
    Migrating Ess Job Definitions From One Instance To Another Instance (1442541.1)
    Jani Rautiainen
    Fusion Applications Developer Relations
    https://blogs.oracle.com/fadevrel/

  • Remote and Testing Server Site Definitions

    I am confused about what information I should put in my
    remote site definition info and testing server FTC Host directory
    text boxes. My host placed 3 directories: database, log, and www on
    my server and Dreamweaver placed two directories, Connections and
    _mmServerScripts in the same directory as the above 3. I have
    placed all of my site files, including my home page and folders in
    the www directory and my Access database in the database directory.
    Dreamweaver placed the connection info in the Connections folder an
    d _mmServerScripts in the same directory.
    In order to construct my local site to match the remote site,
    I created a www folder on my local site and placed all of my HTML
    files and supporting folders in it.
    From various Adobe Tech Notes I have read, I was led to
    believe my FTP host directory
    Should be www/. However, when I do this and create a new web
    page on my local site and upload it to my remote site, Dreamweaver
    creates a new www directory inside of the one my host created on
    the remote site and uploads the file to it instead of placing it in
    the www directory my host created. Also, when I use F12 to test my
    page in my browser I get a 404 page not found because Dreamweaver
    placed a www in the page URL that shouldn’t be there and
    after I remove it I’m able to see my page in my browser. My
    URL prefix on my testing server site definition is:
    http://I.P.number.html.
    What am I not doing right?
    Should I use www/ as my FTP host directory or leave it blank?
    Also, what should my URL prefix be?
    Thanks for your help.

    What you need to do is set up so that in the remote server
    you point to the
    www level on the server, and don't replicate this locally.
    Paul Whitham
    Certified Dreamweaver MX2004 Professional
    Adobe Community Expert - Dreamweaver
    Valleybiz Internet Design
    www.valleybiz.net
    "Jim_Moore" <[email protected]> wrote in
    message
    news:[email protected]...
    >I am confused about what information I should put in my
    remote site
    >definition
    > info and testing server FTC Host directory text boxes.
    My host placed 3
    > directories: database, log, and www on my server and
    Dreamweaver placed
    > two
    > directories, Connections and _mmServerScripts in the
    same directory as the
    > above 3. I have placed all of my site files, including
    my home page and
    > folders
    > in the www directory and my Access database in the
    database directory.
    > Dreamweaver placed the connection info in the
    Connections folder an d
    > _mmServerScripts in the same directory.
    >
    > In order to construct my local site to match the remote
    site, I created a
    > www
    > folder on my local site and placed all of my HTML files
    and supporting
    > folders
    > in it.
    >
    > From various Adobe Tech Notes I have read, I was led to
    believe my FTP
    > host
    > directory
    > Should be www/. However, when I do this and create a new
    web page on my
    > local
    > site and upload it to my remote site, Dreamweaver
    creates a new www
    > directory
    > inside of the one my host created on the remote site and
    uploads the file
    > to it
    > instead of placing it in the www directory my host
    created. Also, when I
    > use
    > F12 to test my page in my browser I get a 404 page not
    found because
    > Dreamweaver placed a www in the page URL that shouldn?t
    be there and after
    > I
    > remove it I?m able to see my page in my browser. My URL
    prefix on my
    > testing
    > server site definition is:
    http://I.P.number.html.
    >
    > What am I not doing right?
    > Should I use www/ as my FTP host directory or leave it
    blank?
    > Also, what should my URL prefix be?
    >
    > Thanks for your help.
    >
    >

  • Problems with ESS in  "Career and Jobs" in option "Appraisal Document"

    hi
    My problem with ESS in "Career and Jobs" menu in the option "Appraisal Document" is not appear the appraisal documents or my my question is...
    What I can customize for appear the "Appraisal Documents"?
    My ESS shows the following screens:
    http://img295.imageshack.us/img295/3451/appraisalportal.jpg (there is also a problem with the flash player, I marked with red rectangle)
    I found looking on the internet that had to configure the Performance Management
    I entered in SPRO tcode, follow these steps
    http://img580.imageshack.us/img580/2598/appraisal1.jpg (Entered in "Define Templates in Performance Management")
    http://img215.imageshack.us/img215/6233/appraisal2.jpg
    http://img651.imageshack.us/img651/4243/appraisal3.jpg
    http://img340.imageshack.us/img340/622/appraisal4.jpg
    http://img52.imageshack.us/img52/1842/appraisal5.jpg (click in "End Configuration")
    http://img534.imageshack.us/img534/8746/appraisal6a.jpg (Error screen part 1)
    http://img4.imageshack.us/img4/1261/appraisal6b.jpg  (Error screen part 2)
    I hope the answers and posts, I wonder if my chosen way(configuring the Performance Managemente)  is correct
    thanks for posts
    Ivan

    thanks Chris for comments,
    Now I customize the "Appraisal Document"
    http://img18.imageshack.us/img18/4640/uiapraisal.jpg (Here I can see the evaluation that was done in the back-end)
    but, I clicked in the hiperlink of  "Appraisal Document Name",  I see this error:
    http://img80.imageshack.us/img80/8157/uiapraisal2.jpg (Error:<b>Page not found or not avaible</b>)
    even before I saw another error (<b>"There is no iView available for system "SAP_ERP_HumanResources": object "employee". For more information, contact your administrator."</b>) in others posts(questions in sap forums) mentioned, These included the following notes
    1468466 Performance Management Launchpad Configuration
    1463821 Performance Management Portal Configuration
    1416756 OBN Configuration in Performance Management
    1408243 Configuration for object-based navigation
    They recommended me in the back-end include roles in the resources:
    http://img401.imageshack.us/img401/8048/uiapraisala.jpg
    http://img689.imageshack.us/img689/9359/uiapraisalb.jpg
    The note number 1463821(especially), I recommend the following
    <a href="https://websmp230.sap-ag.de/sap(bD1lcyZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=1463821" target=nieves>note 1463821</a> (link note)
    http://img88.imageshack.us/img88/7484/uiapraisal1.jpg (LPD_CUST tcode)
    http://img266.imageshack.us/img266/7405/uiapraisal1b.jpg (Check in "Restricted Parameters" and "Active" - "EXT_HEAD")
    but in these notes appointing a role that I could not find on my system that is:
    pcd:portal_content/com.sap.pct/line_manager/com.sap.pct.erp.mss.bp_folder/com.sap.pct.erp.mss.14.bp_folder/com.sap.pct.erp.mss.14.pages/com.sap.pct.erp.mss.hcm/com.sap.pct.erp.mss.appraisal_document_wd_ui
    because, I was looking at my porta's role  (With reference "Appraisals Document")
    http://img97.imageshack.us/img97/5167/uiapraisal3a.jpg
    http://img413.imageshack.us/img413/5643/uiapraisal3b.jpg
    http://img534.imageshack.us/img534/4585/uiapraisal3c.jpg
    http://img203.imageshack.us/img203/4605/uiapraisal3d.jpg
    Finally, my main error is :<b>Page not found or not avaible</b> (I referred at the beginning of the post http://img80.imageshack.us/img80/8157/uiapraisal2.jpg)
    Why I see this error (Page not found or not avaible)?
    http://img80.imageshack.us/img80/8157/uiapraisal2.jpg (Error:Page not found or not avaible)
    I hope the helps, and thaks very much.
    Ivan

  • ESS job created not visible in the list of 'Scheduled Process'

    As a part of data migration , I am required to invoke an ODI scenario using a java program which in turn is invoked by an ESS job.I'm able to do the same using a standalone application created in my jdeveloper.For incorporating this into the Fusion application I created an ADF Model Project in the hcmEss Application and created a sample javaclass and jobdefinition in the project,similar to what i did in my standalone application.
    I have included my job definition in the MarHcmEss and included the MAR file in the EarHcmEss.Also I have given the jazn security for my job,similar to other visible jobs.Still the job that I created is not visible in the list of jobs in the 'Launch ESS Monitoring UI'
    All this i tested using a deployment of hcmEss and hcmTalent Applications into my standalone weblogic server.Is there anything that i might have missed during the process?.I am new to ESS.
    Thanks,
    Ajin

    I'm facing the same issue.
    Also I'm unable to populate the values from the Resource Bundle in the Localization section of Job Definition.
    Any help will be appreciated.
    Thanks,
    Sangita.

  • ESS Personnel File and other iViews

    Hi all,
    In our ESS portal, there's a link under "Personal Information", it's called "Pers. File". We are getting error for this page , and also I don't see the iView and page in portal content admin at all. Do you know where the iView is?
    Also, we are looking for some other links , such as "Previous Employer" , is this link supposed to be under "personal information"? how come I can not see this link anywhere at all? Am I missing some configuration step?

    Thanks for the reply.
    I had not added the Resource Bundle present in another project to this project Bundle list.
    Once added, I was able to select.
    I have another question.
    In the Access Control section, I have added 2 roles, which are already present in the jazn-data.xml file.I select them, give the roles & when I save the job, I'm getting an error.
    I have checked out the jazn-data.xml file present in my view. But the job is referring to jazn-data.xml file present in another location which is read only.
    So I'm unable to save the job definition.
    Also My main issue is The job I created & deployed in weblogic server is not visible in the ESS Testing UI.(http://localhost:7101/essUi/faces/EssUIShell)
    P.S Pls let me know if I have to raise this as another question.I'll do the same.

  • Regarding "Data does not match the job definition; job terminated"

    Dear Friends,
    The following job is with respect to an inbound interface that transfers data into SAP.
    The file mist.txt is picked from the /FI/in directory of the application server and is moved to the /FI/work directory of application server for processing. Once the program ends up without any error, the file is moved to /FI/archive directory.
    The below are the steps listed in job log, no spool is generated for this job and it ended up with an error "Data does not match the job definition; job terminated".Please see below for more info.
    1.Job Started
    2.Step 001 started (program Y_SAP_FI_POST, variant MIST, user ID K364646)
    3.File mist.txt copied from /data/sap/ARD/interface/FI/in/ to /data/sap/ARD/interface/FI/work/.
    4.File mist.txt deleted from /data/sap/ARD/interface/FI/in/.
    5.File mist.txt read from /data/sap/ARD/interface/FI/work/.
    6.PD-DKLY-Y_SAP_FI_POST: This job was started periodically or directly from SM36/SM37 (Message Class: BD, Message Number : 076)
    7.Job PD-DKLY-Y_SAP_FI_POST: Data does not match the job definition; job terminated (Message Class : BD, Message No. 078)
    8.Job cancelled after system exception
    ERROR_MESSAGE
    Could you please analyse and come up about under what circumstance the above error is reported.
    As well I heard that because of the customization issues in T.Code BMV0, the above error has raised.
    Also please note that we can define as well schedule jobs from the above transaction and the corresponding data is stored in the table TBICU
    My Trials
    1. Tested uplaoding an empty file
    2. Tested uploading with wrong data
    3. Tested uploading with improper data that has false file structue
    But failed to simulate the above scenario.
    Clarification Required
    Assume that I have defined a job using BMV0. Is that mandatory to use the same job in SM37/SM36 for scheduling?
    Is the above question valid?

    Hi dharmendra
    Good Day
    How are you
       I am facing the same problem which you have posted
    By any chance have you got the soultion for this
    If so please let me know the solution for this.
    Thanks in advance.
    Cheers
    Vallabhaneni

  • FinancialUtilService, downloadESSJobExecutionDetails method is not uploading ESS job output/log files to UCM.

    Hi,
    We are invoking financialUtilService web service using HTTP Proxy Client to upload data files and submit ESS and to get ESS job log files. We are able to successfully upload and submit ESS job. But downloadESSJobExecutionDetails is not uploading logs/out files to UCM.
              List<DocumentDetails> docDetails = financialUtilService.downloadESSJobExecutionDetails(requestId.toString(), "log");
            //  List<DocumentDetails> docDetails1 = financialUtilService.downloadExportOutput(requestId.toString());
              System.out.println("Ess Job output:" + docDetails);
              for(DocumentDetails documentDetails : docDetails){
                System.out.println("Account: "+documentDetails.getDocumentAccount().getValue());
                System.out.println("File Name: "+documentDetails.getFileName().getValue());
                System.out.println("Document Title: "+documentDetails.getDocumentTitle().getValue());
                System.out.println("DocumentName " + documentDetails.getDocumentName().getValue());
                System.out.println("ContentType " + documentDetails.getContentType().getValue());
    Below output is returned:
    Ess Job status:SUCCEEDED
    Ess Job output:[com.oracle.xmlns.apps.financials.commonmodules.shared.financialutilservice.DocumentDetails@5354a]
    Account: fin$/payables$/import$
    File Name: null
    Document Title: Uma Test Import
    DocumentName:  84037.zip
    ContentType zip

    Hey
    We have the same problem. On calling `downloadESSJobExecutionDetails`, a zipfile is returned, but it only contains my original upload file, not any logs. Although, if I call `downloadESSJobExecutionDetails` on a dependent child subprocess, the log is included.
    I also reported this to oracle support (SR 3-10267411981) - they referred me to known bug, to be fixed in v11: https://bug.oraclecorp.com/pls/bug/webbug_edit.edit_info_top?rptno=20356187 (not public accessible though )
    Is there any work around available?

  • How to create a LOV for a ESS Job Parameter

    Hi,
    I have an ESS job and for one of the parameter , I want to provide users to select from the static List of values.
    What is the easiest possible approach and how to achieve that.
    Thanks in Advance,
    Ankit

    Sorry I missed it somehow. So, assuming currently the job parameter is a text box. Here, is what you can do.
    1. Create a new Fusion application with a model project in it. Then Create a new view object with data based on static values type in the JDeveloper.
    2. Create a new Fusion Applications Customization Application based on the exploded EAR
    3. Open the page you want to change by customize option and then keep a note of the current binding value from the ess parameter inputText and then delete this field from UI.
    4. Shuttle the newly created vo in the AM used for this page. For this to be available you will need to add your project created in step 1 as a library to the Customization application.
    5. Now expand the data controls and drop the attribute from the VO as a LOV on your page and copy the value binding from the deleted field to this one.
    Please note that the above approach should work but I have not personally tried or tested it. Also you will need to go through the extensibility guide for setting up your environment, to understand the customization process, and deploying it to test. Extensibility guide can be found [http://docs.oracle.com/cd/E15586_01/fusionapps.1111/e16691/toc.htm]
    Thanks & Regards !
    Vik
    Fusion Apps Developer Relations
    http://blogs.oracle.com/fadevrel
    Please mark the response helpful or answered appropriately

  • Development and Test on same machine

    Hi,
    In my current setup I have a development application, a test application and a production application.
    The dev. and test is on the same machine, with different APP ID's. Dev is 102, and test is 110
    At first there was no problems with this. But now all of a sudden it starts to create problems for me. In APEX you can create a coloumn link on an item in a report. In the column link you can set link text, attributes, the page the link links to and you can set items that should be send to the new page.
    The problem that occurs with the link is that APEX uses the APPS name instead of the APP_ID, so that my link will be the same from the development app and the test app. F.x.:
    http://test.test.dk/pls/apex/f?p=APP
    instead of
    http://test.test.dk/pls/apex/f?p=110
    I think that APEX stores some cookie information on the machine that accesses the sites, and the problem hasnt occured untill these cookies ran out.
    I see 2 solutions to this problem:
    1. Install a new server with the development app
    2. Rename the development APP to something different.
    The problem with no 2. is that I will then have to (re) rename the app everytime I release the app to the test server.
    Let me hear some suggestions. How do other people do this?
    Edited by: bjarkekr on 2009-12-22 13:09

    I would recommend separating DEV and TEST into different database instances - they can still be on the same physical server and even share the web server (if you are using Apache/modplsql). All that's needed is a second DAD pointing to the second DB (e.g. your URLs would then look like http://test.test.dk/pls/apexdev/f?p=APP and http://test.test.dk/pls/apextest/f?p=APP if you define a DAD named apexdev to point to the DEV database and a second one named apextest to point to the TEST database).
    If you really have to stick with one DB, then I believe all you need to do is changing the application alias, not the application name itself, for your DEV application (and yes, you would need to change this setting once you migrate from DEV to TEST, but that shouldn't be really that big deal, it's just changing that one setting on the app definition page...). By default the alias is the same as the app name.
    Holger

  • Job Chaining and Quickcluster

    I always get
    Status: Failed - HOST [Macintosh.local] QuickTime file not found.
    after the first part of the job is successful.
    If I just submit with "This Computer" it works fine. Original file is ProRes 422, first job uses ProRes 422 to scale it down to 480x270. Second job compresses to h.264. I found some info on this board from 2008 saying that job chaining and quickclusters don't work together. Is that still how it is? that's really useless..
    I also found this from Jan 2009
    David M Brewer said:
    The reason the second rendering is failing is.......this has happen to me a few times until I figured it out.....make sure you set the dimension to the video in the h.264 settings, set to the same size as the Pro Rez dimensions.
    For the most part the dimensions are left blank for the second link, h.264. And don't use 100% of source. Put the physical numbers into the spaces. When you link one video to another, the second codec doesn't know the specs settings you made for the first video settings.
    Also make sure you (at least check) set the audio for the second video. I usually have the Pro Res do the audio conversion and just past-through to the second video settings. Again it can happen the the audio is disable in the h.264 settings. This has happen a few time for me........... Check and double check your settings!
    He doesn't mention anything about with or w/o Quickclusters, but I tried what he said and could not get it to work with quickclusters...
    Anyone got any new info on this?

    Studio X,
    Thanks for taking the time to run some tests and post your results.
    I'm finding the same results with converting ProRes422 to mp4, But...
    Other codecs are giving me very different results.
    I've run some random tests to try to get a grip on whats happening.
    First I was playing around with the # of instances. I've read here and on Barefeats that (at least for my model MacPro) the instances should be set to (# of processors/2), so I've been using 4 for quite a while now and thought I'd test it for myself.
    A single 5min ProRes422 1920x1080 29.97 file to h.264
    This Computer- 15:28
    2 Instances- 14:56
    3 Instances- 13:52
    4 Instances- 14:48
    5 Instances- 13:43
    6 Instances- 13:48
    7 Instances- 13:58
    In this case 5i was the fastest but not using a Quickcluster wasn't far off
    A single 2m30s ProRes422 1920x1080 29.97 file to h.264
    This Computer- 3:19
    2 Instances- 3:45
    3 Instances- 3:45
    4 Instances- 3:45
    5 Instances- 3:50
    6 Instances- 4:00
    7 Instances- 4:00
    Interesting...not using a Quickcluster is fastest
    A single 2m30s ProRes422 1920x1080 29.97 file Scaled down using original codec
    This Computer- 5:20
    4 Instances- 4:10
    5 Instances- 4:10
    7 Instances- 4:11
    A single 1m30s ProRes422 1920x1080 29.97 file to mpeg-2
    This Computer- 2:12
    5 Instances- 2:10
    When Quickclusters are faster, 4-5 instances does seem to be the sweet spot(again for my setup).
    In the mpeg-2 test, I should have used a longer clip to get a better test but it was getting late and I was just tring to get an idea of the codecs usage of my resources. I was also monitoring CPU usage with Activity Monitor in all tests.
    Now multiclip batches:
    I forgot to write down the length of the clips in this first test but it consisted of 8 ProRes 422 clips. 3 about 1m long and the rest between 13s and 30s
    8 ProRes 422 clips to mp4
    This Computer- 11:25
    4 Instances- 5:16
    Same results as Studio X
    Next tests with 5 clips(total 1m51s)
    5 ProRes 422 clips to h.264
    This Computer- 5:00
    4 Instances- 4:52
    5 ProRes 422 clips to mpeg-2
    This Computer- 2:55
    4 Instances- 3:01
    5 ProRes 422 clips to DV NTSC
    This Computer- 6:40
    4 Instances- 5:12
    5 ProRes 422 clips to Photo Jpeg
    This Computer- 2:44
    4 Instances- 2:46
    I re-ran the last test with 7 clips because of the time it took reassemble the segmented clips
    7 ProRes 422 clips to Photo Jpeg(total 3m14s)
    This Computer- 4:43
    4 Instances- 3:41
    One last test,
    A single ProRes 422 clip to Photo Jpeg(4:05;23)
    This Computer- 5:52
    4 Instances- 4:10
    Let me start off by saying it is clear that there are many factors that effect compression times such as # of clips, length of clips, and codecs, but here are some of the things I noted:
    1)Some codecs themselves seem to be "more aware" of the computers resources than others.
    When I compress to h.264 w/o a cluster it will use about 80-85% of all resources
    When I compress to h.264 with a cluster it will use about 90-95% of all resources
    When I compress to PhotoJpeg w/o a cluster it will use about 20-25% of all resources
    When I compress to PhotoJpeg with a cluster it will use about 80-85% of all resources
    2)The time it takes to reassemble clips can be quite long and could effect overall speed
    In the very last test, compressing a single file to photoJpeg using 4 instances took 4m10s. Watching Batch Monitor I noted that it took 2m0s to compress and 2m10s to reassemble.Wow...
    It would be interesting to see how the disassemble/reassemble of bigger and larger batches using clusters effect overall time. But that would take some time.
    I think the thing I will be taking with me from all of this is your workflow is your own. If you want to optimize it, you should inspect it, test it and adjust it where it needs adjusting. Now if anyone has the time and was to run similar tests with very different results I'd love to know about it...

Maybe you are looking for