Trying to submit a job for Compressor

I'm having trouble submitting my sequence once it's been imported into Compressor.
After performing the necessary steps then clicking on the submit button the window
displays the name and priority, but for the cluster it displays "no value". So I click on submit again, then a warning appears stating: "Unable to submit to queue"
Please restart or verify your Compressor installation is correct.
I've tried to restart which didn't work either.
I've also tried going into the QMaster under system preferences, but that seems to be fine. Please help me!

I have been experiencing the same issue with Compressor, unable to submit a job to be compressed.
This issue started recently, after the Pro App update.
I am using Compressor 3.5, there are no current updates available in "Software Update".
I have not tried reinstalling Final Cut Studio yet, I would like to avoid that if possible. I'm sure there is a software update fix coming soon... In the mean time, I am unable to use compressor.

Similar Messages

  • Submit background job for a report

    Hai Friends,
    I want to submit a report in background which should run on first monday of every month. otherthen SM36 and SM37. Because we cant give date and time like this.
    If any coding required Pl give me the code.
    regards,
    vamsykrishna.

    This is the Basic code to Achieve that.
    Call Function 'JOB_OPEN'
    SUBMIT RSBDCBTC
    via background Job <Job info>
    With QIE = <QID Returned by BDC_OPEN_Group FM>
    AND RETURN.
    Call Function 'JOB_CLOSE'
    2) Also go through this
    DATA: lv_job_name LIKE tbtco-jobname,
    lv_job_nr LIKE tbtco-jobcount,
    lv_job_released TYPE c,
    lv_job_start_sofort TYPE c,
    lv_print_parameters TYPE pri_params.
    lv_job_name = 'Z_BGPROGRAM'.
    CALL FUNCTION 'JOB_OPEN'
    EXPORTING
    jobname = lv_job_name
    IMPORTING
    jobcount = lv_job_nr
    EXCEPTIONS
    cant_create_job = 1
    invalid_job_data = 2
    jobname_missing = 3
    OTHERS = 4.
    IF syst-subrc = 0.
    " submit job with all the selection screen params...
    SUBMIT (lv_job_name)
    WITH applfile = applfile " these are the parameters of the background program
    WITH p_lines = p_lines
    USER syst-uname " you must use this
    VIA JOB lv_job_name NUMBER lv_job_nr AND RETURN.
    IF sy-subrc = 0.
    CALL FUNCTION 'JOB_CLOSE'
    EXPORTING
    jobcount = lv_job_nr
    jobname = lv_job_name
    strtimmed = 'X'
    IMPORTING
    job_was_released = lv_job_released
    EXCEPTIONS
    cant_start_immediate = 1
    invalid_startdate = 2
    jobname_missing = 3
    job_close_failed = 4
    job_nosteps = 5
    job_notex = 6
    lock_failed = 7
    OTHERS = 8.
    IF syst-subrc <> 0.
    MESSAGE i162(00) WITH
    'An error occured while closing the background job.'.
    STOP.
    ENDIF.
    ENDIF.
    ENDIF.
    SKIP 1.
    WRITE: / 'Background process', lv_job_name , 'called successfully' NO-GAP.
    WRITE: / 'You can check the job in transaction SM37'.

  • Would selling cool iPhone cases on eBay be a good ieda im 16 im trying to create a job for my self becasue times are hard

    selling cool iPhone cases a good ieda?

    While it sounds like a good idea, you will have an incredible amount of competition as anyone can go anywhere to buy an iPhone case. You would have to advertise your business and explain why someone would want to buy from you rather than Best Buy, RadioShack, or any other place where they can pick up a case in an afternoon and not have to wait for it to be shipped. Keep up the business spirit. Maybe look for something you can where you live. Offer a service of some type. Good luck with your search.

  • Getting an unusual error message in Compressor 4.1 when I try to submit a job

    I'm running Mavericks and have Compressor 4.1 on my Mac Pro along with FCP 10.1.  When I submit a job to compressor, I then add the Dolby Digital and Mpeg-2 6.2 Mbps/90.  When I hit Start Batch I get this error message
    /var/folders/k9/f6fyk4sj4f3_rj2wlrlwx9hr0000gn/T/46BDF064-B30F-4BF1-8D9C-D22DE91 8342B.export-archive
    I've tried to uninstall and re-install Compressor but to no avail.  What is this error message referring to and how do I rectify it?
    Thank you
    Drew

    Hi Drew if you haven't resolved this. TRy this to see if the issue is a TRANSIENT object access error instead for submitting directly to Compressor. Do this to isolate any content error defects before making you distribution (older 2 stage workflow)..
    in FCPX 10.1, make sure the projects Primary Storyline is completely rendered - select all and cntl+R (render selection or control+shift+R)In FCPX 10.1 watch the background tasks (+9) and wait for all the rendering to be completed.... (make sure no object errors)
    in FCPX 10.1 export a master File/Share/Master File (+E) as PRORES 4xx and save it ~/Movies/Drews_final_master.mov (prores)
    in Compressor.app V4.1,
    create a new BATCH ,
    import the MASTER object ~/Movies/Drews_final_master.mov)
    (and add your setting and submit it. (dont use FCPX SEND TO COMPRESSOR just yet! until you resolve this issue)
    This process will avoid the use of transient  file storage areas that seem to be used in the FCPX to COMPRESSOR.
    Post your results for others to see
    Warwick
    Hong Kong

  • Submit remote job to HDInsight cluster using its IP address.

    Hi there,
    I am using HDInsight and trying to submit jobs in it. I am able to submit jobs using the API provided by Azure. This works fine. Also, I am able to submit job in remote machine by opening the remote machine using VM. 
    I am now trying to submit the job to the HDInsight cluster from my machine using the IP address of the remote machine. I am not able to submit any job into it. It throws out some sort of error.
    Please help me on this.
    Regards,
    Athiram S

    Hi Sudhir,
    Thanks for looking into this.
    We can submit job in hadoop cluster using the IP address by the following method.
    1) Configure certain XML files like core-site.xml,Hdfs-site.xml, Yarn-site.xml in cluster machine(namenode) with the IP address of the machine. I also make the similar change in the configuration files in my machine under the location "..\\etc\\hadoopcluster_IPAddress".
    2)Now, execute the command pig --config "..\\etc\\hadoopcluster_IPAddress" in my machine(which is connected to namenode machine of the cluster through LAN). Now, the Map-reduce job gets executed in remote machine.
    I am trying a similar approach for submitting the job in HDInsight cluster. I use the Headnode IP address and modified the configuration files and used the same command as above. But, I am wondering why it not working.
    I am able to get the job successfully executed to my cluster machine and job submission in HDInsight cluster fails.
    Please help me on this issue.
    Regards,
    Athiram S

  • Submit background job in APEX -- count on all_jobs return shadow jobs.

    Hi, I am trying to submit a job in APEX. The setups are as below:
    On Submit - After Computation and Validations
    Run Process: Once Per Page Visit
    Process:
    DECLARE
    v_instance_cnt NUMBER;
    job_no NUMBER;
    BEGIN
    SELECT COUNT(0)
    INTO v_instance_cnt
    FROM user_jobs
    WHERE what LIKE 'pagl_refresh.master_refresh%';
    IF NVL(v_instance_cnt,0) = 0 THEN
    DBMS_JOB.SUBMIT(job_no, 'pagl_refresh.master_refresh('''||:G_BSYS_USER_NAME||''');');
    :P3_MESSAGE:= 'Job has submitted. Number is '||TO_CHAR(job_no);
    ELSE
    :P3_MESSAGE :='The refresh is in progress. Please wait ... ('||to_char(v_instance_cnt);
    END IF;
    END;
    Now, if I run the process, the :P3_MESSAGE message returns "'The refresh is in progress. Please wait ... (5)." . This is due to the count is 5 instead of expected 0.
    If I SELECT count(*) FROM dba_jobs WHERE lower(what) LIKE 'pagl_refresh.master_refresh%'; in SQLPLUS, it returns 0. Same result from all_jobs as well.
    My suspect is that it returns job counts include those that has been removed before. Yet, how APEX can see this? Does APEX use some special way to look into Job queue?
    Please help
    Thanks

    From the looks of it, the job is being submitted and run - although I would check the elapsed time to see if it's anywhere close to the 20-30 minutes you anticipate. Assuming not, I would suggest that the problem is in one of the following areas:
    1. The way in which you are passing in the arguments is not conforming to the expected input format or values and it's therefore not executing as expected.
    2. Your process implictly relies on the state of your apex application in some manner, which is not being reproduced within the procedure when the job is submitted.
    In the former case, I would check the procedure's specification against the page items types being passed in - you might have to explicitly convert some of your arguments into the appropriate type
    In the latter case, well... bearing in mind that we don't know what your procedure looks like and it's therefore kind of difficult to diagnose the problem, you'll possibly need to pass your session information into the procedure as additional parameters and re-create your session from within the code.

  • Error when scheduling job for Job Control LO

    Hi all,
    I tried to schedule the job for collective update after I set the job parameter in tcode LBWE for LO Data Extraction. But they said 'Error generating job' without any description why.
    Maybe you can help me about this since I still can't do the delta extraction for LO.
    Thanks and appreciate for your help.
    -Martin Darmawi-

    Hi!
    You have to pass two parameters for job control scheduling.In that you have traffic light for the respective parameters,defaulty those will appear in red.when you have specified parameters,both turns in to green.then you can say start.hope this will helps.....
    Best Regards
    Sunil.

  • Pre-processing job for Transactions

    Hi Archiving Forum,
    In SAP help it is mentioned that pre-processing job is optional job. I am trying to run Write job for few of the transactions where the status is completed without pre-processing job.
    But when i debugged the write program for - CRM_SERORD archiving object, the program internally checking the I1100 status for each Service ticket/ Order.
    Here, all Service tickets were marked completed status on 24/03/2009.
    Residency time for all transaction types is 1 day.
    I am running the Write job for Service Ticket and Service order on 24/03/2009. - but job is not able to write the data into Archive files.
    Should i wait for One day ( where it completes the Residency time) to run the Write job?
    If i wait for one day, then the write program is internally checking the status I100 (Archival status). Is this status is updated by default in the CRM_JEST table after completing Residency period ?
    Please need concrete answers.
    Thanks in Advance,
    Kishore

    Hi Kishore,
    Pre-processing job should be scheduled for archiving objects before runing write program because it will help in basic archivability check and makes the basement for write program.
    Without executing pre-processing program write program will fails to archive data. For example: CO_ORDER, MM_EKKO, MM_EBAN and so on... in R/3 system.
    Similarly in CRM system Pre-processing program will help
    In check, the following database changes for CRM_SERORD:
    1) The date of the archivability check is updated.
    2) For the archivable data, the status is changed to 'Archivable'. Then the
    data is no longer changeable.
    3) For the archivable data, the system sends a deletion BDoc type with
    message type 'A' (for archiving) via the CRM Middleware so that the other
    components involved in mySAP CRM solutions are informed about an archiving
    of the data in the CRM server. The system sends deletion BDoc types at this
    time because the change to an unchangeable status is the status change
    which is relevant from a business point of view.
    In CRM 3.0 version
    The check occurs in an archiving object-specific preprocessing program
    entered for the archiving object in transaction AOBJ.
    As of mySAP CRM5.0 it is controlled using transaction SARA check/delete
    In your case:
    Even after passing of 1 day of residence time service order or service ticket will not be archive since write program will not set the status for archive I1100. Preprocessing program calling functional module and setting the status as archivable (I1100) after all the pre-requisites are met and write program will archive those data.
    The status in ST / SO will not be automatically changed after meeting residence time.
    Please also read OSS note no 500551.
    -Thanks,
    Ajay
    Edited by: Ajay Kumar on Mar 25, 2009 2:17 AM

  • Unable to submit a job when the variable is a composite one

    Hi
    I was trying to submit a job from dbms_job.submit
    DBMS_JOB.SUBMIT(l_job,
    'gitm_goaml_job_proc('||''''||p_wrk_gidamlrp||''''||'),
    SYSDATE,
    'SYSDATE + (10/(24*60*60))');
    Here p_wrk_gidamlrp is a composite type of variable which i have created in a package spec.
    when i try to complie the code I receive an error that '|| have not been used properly' But if i replace p_wrk_gidamlrp with a simple data type variable things work fine.
    Can you please let me know the reason of this error?
    Thanks
    -vinod
    Edited by: 965358 on Oct 15, 2012 4:07 AM

    post details of your datatype etc. otherwise we don't know what you've got in there and how we can help.
    Also make sure you include your database version and use {noformat}{noformat} tags as described in the FAQ: {message:id=9360002}                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Job for transaction PBA6

    Hi,
    I'm trying to create a job for transaction PBA6 (program RPAPRT02) but once I get the list of all the activities I should complete I don't know how to select them all and execute de program SAPMSSY0.
    Any idea?
    Thanks

    Hi,
    Did you try to make a Parallel Processing referred to RFC?
    Maybe this link can help you.
    [http://help.sap.com/saphelp_nw04/helpdata/en/22/0425c6488911d189490000e829fbbd/content.htm]

  • TS1702 I have tried to submit a dissatisfaction report following your instructions from my invoice for the recent purchase of smart office 2.  I can get no further as a report code keeps appearing saying too many http redirects so I can not get any furthe

    I have tried to submit a dissatisfaction report following your instructions from my invoice for the recent purchase of smart office 2.  I can get no further as a report code keeps appearing saying too many http redirects so I can not get any further.
    I have also sent 3 emails to the company if Smart Office 2 with no response.
    How do I report my dissatisfaction which is growing with the frustration of not being able to contact any one!
    Zena Parton

    I have tried to submit a dissatisfaction report following your instructions from my invoice for the recent purchase of smart office 2.  I can get no further as a report code keeps appearing saying too many http redirects so I can not get any further.
    I have also sent 3 emails to the company if Smart Office 2 with no response.
    How do I report my dissatisfaction which is growing with the frustration of not being able to contact any one!
    Zena Parton

  • I am trying to get a login for eprint but I get this: Ajax submit failed: error = 403, Forbidden

    Can anyone help me.  I am trying to set up eprint for my chromebook and need to get an email address for my 6600
    This question was solved.
    View Solution.

    Hi,
    Please use other web browser OR clear cache, histories then try again.
    Regards.
    BH
    **Click the KUDOS thumb up on the left to say 'Thanks'**
    Make it easier for other people to find solutions by marking a Reply 'Accept as Solution' if it solves your problem.

  • OEM submit job for EJB

    Could I use OEM to submit a job and call an EJB session bean? If answer is yes , could you point me to the right direction , steps should I perform in OEM.
    Thanks
    Suhail

    This is the Basic code to Achieve that.
    Call Function 'JOB_OPEN'
    SUBMIT RSBDCBTC
    via background Job <Job info>
    With QIE = <QID Returned by BDC_OPEN_Group FM>
    AND RETURN.
    Call Function 'JOB_CLOSE'
    2) Also go through this
    DATA: lv_job_name LIKE tbtco-jobname,
    lv_job_nr LIKE tbtco-jobcount,
    lv_job_released TYPE c,
    lv_job_start_sofort TYPE c,
    lv_print_parameters TYPE pri_params.
    lv_job_name = 'Z_BGPROGRAM'.
    CALL FUNCTION 'JOB_OPEN'
    EXPORTING
    jobname = lv_job_name
    IMPORTING
    jobcount = lv_job_nr
    EXCEPTIONS
    cant_create_job = 1
    invalid_job_data = 2
    jobname_missing = 3
    OTHERS = 4.
    IF syst-subrc = 0.
    " submit job with all the selection screen params...
    SUBMIT (lv_job_name)
    WITH applfile = applfile " these are the parameters of the background program
    WITH p_lines = p_lines
    USER syst-uname " you must use this
    VIA JOB lv_job_name NUMBER lv_job_nr AND RETURN.
    IF sy-subrc = 0.
    CALL FUNCTION 'JOB_CLOSE'
    EXPORTING
    jobcount = lv_job_nr
    jobname = lv_job_name
    strtimmed = 'X'
    IMPORTING
    job_was_released = lv_job_released
    EXCEPTIONS
    cant_start_immediate = 1
    invalid_startdate = 2
    jobname_missing = 3
    job_close_failed = 4
    job_nosteps = 5
    job_notex = 6
    lock_failed = 7
    OTHERS = 8.
    IF syst-subrc <> 0.
    MESSAGE i162(00) WITH
    'An error occured while closing the background job.'.
    STOP.
    ENDIF.
    ENDIF.
    ENDIF.
    SKIP 1.
    WRITE: / 'Background process', lv_job_name , 'called successfully' NO-GAP.
    WRITE: / 'You can check the job in transaction SM37'.

  • How to share a job with Compressor 4.1?

    Can anyone explain how to set up Compressor on two or more computers to share a coding job? I was never successful, neither with the old versions nor the new one. I have connected two computers running Mavericks via Ethernet. They appear in the preferences list of Compressor as inactive and can be selected (with a tick). Starting a job produces no error. Only the little network monitor window shows some activity: "inactive" (in white or yellow), sometimes: "not found" (in red). The computer which sends the job waits endlessly.
    I deactivated the firewall, connected the computer with DHCP or fixed IP but no success. What else do I have to do?

    Hi Steffen, hats off to you for gathering this valuable information!  I'm going to title this post:
    Setup Distributed Node Processing for Distributed Segmented MULTIPASS Transcoding in Compressor.app V4.1 (2013 version)
    Summary:
    A quick look at those logs of yours.., Qmaster is having trouble accessing its cluster storage and probably your transcode source and target elements.
    This is a bit of a giveaway - looks like the part time helpers at Apple didn’t look at it hard enough
    msg="CSwampService::startupServer: servicecontroller:com.apple.stomp.transcoderx couldn't advertise server, error = error: DNSServiceRegister failed: CRendezvousPublisher::_publish"/>
    <mrk tms="412112937.953" tmt="01/22/2014 20:48:57.953" pid="1975" kind="begin" what="log-session"/>
    <mrk tms="412113195.964" tmt="01/22/2014 20:53:15.964" pid="1975" kind="begin" what="service-request" req-id="D6BAF26C-DD43-4F29-BD72-81BC9CF25753:1" msg="Processing."></mrk>
    <log tms="412113209.037" tmt="01/22/2014 20:53:29.037" pid="1975" msg="Shared storage mount failed: exception = CNFSSharedStorage::_subscribe: command [/sbin/mount 127.0.0.53:/Users/steffen/Library/Application Support/Compressor/Storage/21D262F0-BF7EC314/shared] failed, error = 61"/>
    Let’s look at this and then propose a clean method of establishing and consolidating your cluster.
    Simply the Bonjour service is having a hard time trying to find you and also qmaster been running ragged trying to mount your Cluster 21D262F0-BF7EC314 storage.
    Let's fix it.
    Basics for the above with Compressor v4.1 and Qmaster.
    much has been abstracted from the user to help easy implentation and use. This is good me thinks!
    avoid ticking every option that is available on each host , such facilities aggravate and add to the complexity of your workflow environment
    isolate the network subnets to use for all host access, file system paths, communication, diagnosis, monitoring  and finally data transfer (see later)
    use source objects that will develop segments that will be distributed for processing. A 3 minute clip generally won't segment to cause distribution.
    review any workflow gains by distributed transcoding: slow node hols up process and additional time to assemble qt segments. A cluster dedicated to an iMac or macpro can often be faster.  (Have several clusters defined and submit accordingly (long , most and short )!!)
    All elements/objects used in the source and any target folders SHOULD (not must) be mounted and accessible by each Qmaster node.  You can use sym links I recall. For reasons of efficiently and easy of diagnosis.   
    So.. I'd propose you try and test your setup as follows .
    Try this and start from beginning.  Do your best to perform these work instructions. Try also not to deviate if u can
    Simple Architecture Approach:
    Your main Macbookpro or main work mac (refered to by you as "master") shall be designated the qmasterd controller that services batch submissions AND that also provides transcode services.
    The other macs ("service or slave nodes) will ONLY perform transcoding services and will NOT accept batch submissions.  The slaves / service nodes  will not be able to send their jobs to your master controller for transcoding for example.
    Keep it simple! and please follow these steps.
    Step 1: Quiesce your clusters and Qmaster
    in Compressor.app v4.1 / Preferences / Shared Computers, stop / disable all hosts (both your macs) from automatic file sharing - tick it OFF (it causes issue u have).. More later
    In Compressor.app v4.1 / Preferences / My Computer, stop / disable all hosts (both your macs) stop allowing others to add batches to your host. Slide to OFF
    On all hosts, quit or force out compressor.app v4.1
    On all hosts (macs) use activity monitor.app or unix ps command to Force Quit ( or kill) any currently running qmasterd task and any compressord tasks if you can.
    On all hosts, purge | clean out | delete  the Qmaster and compressor structures. This is documented by several of us on this forum but fundamentally you want to preserve your settings and destination templates and delete the rest.  Do these sub-steps on all hosts where u intend to deploy compressor/Qmaster for your distributed transcode processing
    a. Navigate to /Users/shared/library/Application Support/ and delete the Compressor folder if it exists. By all means use the osx Finder to put it in the trash or just use the faithfully unix rm command to delete it immediately without serialisation : rm -rf /Users/Shared/Library/Application Support/Compressor
    b. Navigate to your home directory ~/Library/Application Support/Compressor and move or copy any personalised values to your desktop so we can reinstate them later. Copy these two folders if they exist.
    Settings
    Layouts
    And also copy /move any customised destination templates that u used. These are files ending in ".template"
    Now using the Finder or unix command delete your ~/Library/Application Support/Compressor folder and all it's objects such: rm -rf  ~/Library/Application Support/Compressor
    c. Dismount (+E or drag into trash) any shared file systems you have manually or automatically shared between your hosts. (Your two macs) . Turn off any auto mounts you may have setup in login items for Qmaster and your source and target libraries.
    d. After you have done Steps 1a - 1c on all your hosts ,
    then RESTART
    and log back into your hosts
    attempt to delete the trash on each.
         6. Check the activity monitor and confirm there are no compressord sub tasks running. Qmasterd might be running. That's ok
    Step 2: set up your dedicated network for your transcoding cluster .. Important!
    In this step you will isolate your cluster network to a dedicated subnet. BTW ,o DNS is needed unless you get exotic with many nodes!
    You will:
    use the macs Wifi network as your primary network for NON TRANSCODING work such as email, internet , iChat, iCloud and bonjour (.local) I'm assuming u have this in place
    use the Ethernet on your macs as your dedicated Qmaster cluster subnet.
    For this procedure will make an IP address range 1.1.1.x subnet and manually assign the IP addresses to each host. Ofcourse you can use a smart DHCP router if u have one or a simple switch and use OSX SERVER.app erver 10.9 ($HK190 , €19) on your MAcbookpro... The later for another time
    a). using system preferences/network on your controller mac("master"), configure the built in Ethernet to manual IP address of 1.1.1.1 Yes yes, dhcp would be nice if we also had a dns for this subnet to identify each host (machine name) however we don't. Leave the subnet default to 255.255.255.0, and router to 1.1.1.1 .. It's not escaping anywhere! ( very private LAN for your cluster! )
    b) repeat step 2a to set the other "slaves" service node only macs whose built in Ethernet to 1.1.1.2 , 1.1.1.3 and so on
    c) connect these hosts (macs) ethernets together in a dedicated hub / zoned switch or if only two macs, just use a cat5/cat6 Ethernet cable
    d) on each host (mac) using system preferences/network check for a Green light against the built in Ethernet
    e) on each host (mac)system preferences/network to make a new network configuration (so that you can fall back incase of error) :  using system preferences/network make a new network location on each mac
    - edit the Location listbox, edit and DUPLICATE the current location
    - select and over type the name and change it to "qmaster work" or some name u like save and close the dialogue
    - back in sys prefs / network select your new location "qmaster work" (from the location list box and "apply"
    - now click the gear wheel icon on lower left and Reorder the network interfaces so that the WIfi is top (first) followed by Builtin Ethernet .
    - click "apply"
    - confirm Ethernet is still green status
    Do this on each host (mac) .. The slave/service nodes
    f) on each host (mac) verify that you can address each mac over you new subnet. There's many ways to do it however do it simply via the /applications/utilities/Terminal.app.
    From mac #1 whose IP address is 1.1.1.1,
    Enter:
    traceroute 1.1.1.2 press return and one line should come back.
    ping 1.1.1.2 and a continuous lines appear with packets and time is ms. Watch 3-4 then use control+C to stop
    Do the same to the other nodes you may have such as 1.1.1.3 etc
    Repeat the above from the other hosts. For example from one of the service (slave) macs say 1.1.1.2, t
    Test the network path back to your main mac 1.1.1.1: using terminal.app from that slave,
    Enter:
    traceroute 1.1.1.1 press return and one line should come back.
    ping 1.1.1.1 and a continuous lines appear with packets and time is ms. Watch 3-4 lines then use control+c  to stop
    At this point you should have a solid network path between your hosts over Ethernet on the subnet 1.1.1.x
    Step 3: mount all filesystems over the Ethernet 1.1.1.x subnet that are to be used for transcoding source (input | read )  and target (output | to be written )
    Simplicity is important at this stage to make sure you know what being accessed.  This is one reason for disabling all the automatic compressor settings
    You will use the Finder's "Connect to Server" (+k) from each slave (server) node to access the source and target filesystems on your master mac for the transcoding.
    These can be saved as favourites in the "Connect to Server" dialogues
    Do this:
    A) locate the volumes / filesystems and folders on your mac master where your source objects is contained. Do the same for where the final distribution transcode is to be written with you user access.. "Steffen"
    B) On each slave mac, use the Finder's "connect to server" dialogue to MOUNT those folders as network volumes on your slave macs
    mount the Source folder. Finder / Go / Connect to Server  or +K
    enter "[email protected]//Users/steffen/movies/my-fcpx-masters. ( choose you source directory path) .
    Click connect & password and use the "+" sign to save as favourite
    - mount the target folder. Finder / Go / Connect to Server  +k
    - enter "[email protected]/users/movies/my-fcpx-transcodes. ( choose your target directory path) . Click connect n password and use the "+" sign to save as favourite
    Do these for all your slave macs.  Remember you are specifying file paths over the 1.1.1.x subnet
    Obviously make sure your slaves have read and write access. Yes and you could also mark these folders and Shared in Finder info then everyone can see them ... your choice
    C) verify your access: on each slave use the finder to create a ftest older in those recently mounted volumes. Delete the test folder.
    So now all your networks and workflow folders are all mounted and accessible by your slave hosts and it's user
    step 4: Set up Compressor v4.1 and Qmaster
    Care is initially required here NOT to click needlessly on options.
    Recall that you purged most of the compressor.app v4.1 state information on step 1. ?
    Compressor.app v4.1 will appear new when u start it. 
    on the master mac 1.1.1.1 , launch compressor.app v4.1
    open compressor v4.1 preferences (command+,)
    using compressor.app V4.1 preferences:
    My Computers tab: turn OFF . Don't allow others to process on this computer
    Share Computers tab: UNTICK (disable) automatically share files.
    Advanced tab: WHen Sharing My Computer listbox: select Ethernet as the preferred network interface 1.1.1.x). Don't use all interfaces : - trouble
    Do not specify additional instances yet! Less to troubleshoot
    On each slave mac 1.1.1.2 -1.1.1.x
    launch compressor.app v4.1
    open compressor v4.1 preferences (command+,)
    using compressor.app preferences:
    My Computers tab: turn ON (yes ON) to allow others to process their stuff on this slave computer
    Share Computers tab: UNTICK (disable) automatically share file
    Advanced tab: WHen Sharing My Computer listbox select Ethernet as the preferred network interface 1.1.1.x). Don't use all interfaces : - trouble
    Do not specify additional instances yet! Less to troubleshoot !
    On the master mac, 1.1.1.1
    using Compressor.app v4.1, select destinations and add a new CUSTOM destination and navigate the dialogue to your target folder you specified in step 3b ~/movies/my-fcpx-transcodes as an example.
    Use this custom destination on the batch later
    in Compressor.app V4.1 preferences/Shared Computers, click the plus "+" sign in bottom left cornet to create a new cluster called "unnamed ".
    - click in the name and change the name to "Steffenscluster" (I'm not connected to my network as I do this..)
    - tick on the slaves 1.1.1.2 to 1.1.1.x to add the to your new cluster .. assumethee are on the right in a list.
    Your cluster "Steffenscluster" is now active!
    Care care! One more thing to do. You SHOULD make the cluster storage for master 1.1.1.1 available to all your slaves. This is important for this test !!
    On the master 1.1.1.1, Use the finder to verify that you have these directories built from compressor.app v4.1
    /Users/Shared/Library/Application Support/Compressor
    and your own home directory: ~/Library/Application Support/Compressor
    Dig deeper for the /storage folder in each to see the hexadecimal named folder that represents this cluster "Steffencluster"!
    These should be manully mounted on each slave. Recall we DISABLED automatic file sharing.
    on each slave mac 1.1.1.2 - 1.1.1.x, mount the masters cluster storage file systems. Do this for verifying access of each cluster Slave
    on each slave mac, use the Finder's "connect to server" dialogue to MOUNT those folders as network volumes on your slave macs
    mount Qmaster cluster storage folder 
    Use Finder / Go / Connect to Server  or +k
    enter "[email protected]/Users/Shared/Library/Application Support/Compressor/Storage
    Click connect & password and use the "+" sign to save as favourite
    - mount users Qmaster cluster storage folder 
    Use Finder / Go / Connect to Server  or command+k
    enter "[email protected]/Users/steffen/Library/Application Support/Compressor/Storage
    Click connect & password and use the "+" sign to save as favourite
    This you may have 4 new network volumes (file systems) mounted on each mac slave over your dedicated 1.1.1.x subnet!
    Step5: submit a job
    On master mac 1.1.1.1 launch the new compressor.app v4.1 "Network Encoder Monitor " .. Use +E . new in COmpressor.app V4.1
    You should see all your nodes all cluster service points for your master and slaves.  Heres you see just one (this macbookair!)
    On each host (mac) Launch the activity monitor.app and filter compressord .. There they are on each host! 
    Nearly there.  
    On the mac that's the master 1.1.1.1 (controller )
    Submit a job:
    Use finder to move some footage etc that more than 15 mins for example into your source directory folder from step 3b (eg ~/movies/my-fcpx-masters.)
    In compressor.app v4.1 Add this +I to the batch
    Drag your custom destination on to the batch
    Set your encoding setting (apple devices best)
    Open the inspector in compressor.app and select "video" and make sure MULTIPASS is checked. Then change to frame controls at your leisure.  Better means slower
    Select the General tab and make sure JOB SEGMENTING is ticked!
    Now cross fingers and submit it (or +B)
    Step 6: Monitoring the Workflow Progress
    In compressor.app v4.1 Use the active tab to watch the progress
    Open the disclosure triangle and see the segments
    Unfortunately u can really see which node is processing. (No more share monitor .. btw for those who care.. thats buried now in /Applications/Compressor.app/Contents/PlugIns/Compressor/CompressorKit.bundle/Contents/Embedde dApps/Share Monitor.app/Contents/MacOS/Share Monitor
    Look at the network encoder monitor (cmd + E) to see the instances processing your work
    Lots of small and overdetailed steps here Steffen and its worth working through.
    Simply all these things need to be availble to get your cluster to work EVERYTIME.
    I might update this and post a more detailed disalogue/transcript on my blog and post it here.
    Epilogue:
    I for one rather like the new compressor.app V4.1. Qmaster is buried and works well when not teased or unintentionally fooled.
    I would like the ability to :
    specify the location of the qmaster cluster storage rather than have it in the root systems file system - I used to have it on my disk array
    compressor to be applescriptable
    Post your results for others to see.
    Warwick
    Hong Kong

  • Error while trying to submit Access request to GRC from IDM

    Hello
    We have SAP IDM 7.2 SP8 installed and done all the prerequisite for connecting to GRC AC 10 as in configuration document.
    We are trying to submit request to GRC using Standard GRC provisioning framework task ( AC Validation) but pass: Submit AC Request fails with error: "Pass stopped by script"
    Is there anything wrong with the script which put RoleData details since its getting aborted ?
    I tried providing Role name directly in Role data attribute inside the action task and got following error:
    Error
    putNextEntry failed
    storingcn=IDMUSR0023,ou=useraccessrequest,o=grc
    Exception from Add operation:javax.naming.NamingException: [LDAP: error code
    82 - (GRC User Access Request:82:Script execution failed)]; remaining name
    'cn=IDMUSR0023,ou=useraccessrequest,o=grc'
    I checked VDS Logs and there was one error :
    Additional message = msgcode=4;msgdescription=Mandatory field ITEM NAME  is empty in line no 1 ;msgtype=ERROR
    From where exactly ITEM NAME field value will be fetched and pass to GRC for request creation ?
    Regards
    Deepak Gupta

    Thanks Christopher
    I got my issue fixed, There was issue with my GRC Initial load job which couldn't enrich repository privileges and hence the issue was coming since script wasn't able to find GRC ROLE ID and Application ID attribute from privileges.
    Regards
    Deepak Gupta

Maybe you are looking for

  • Airdrop after iOS 8 upgrade

    HI, I have recently updated my iPad mini to iOS 8 and noticed a bad behavior when transfering photos to my iPhone 5s. the problem is that everytime I try to share a set of photos, only few (4 in my last attempt, when 6 photos were selected) are sent

  • Multiple display problem in OSX 10.9.2

    I just updated to 10.9.2 on my 15" Macbook Pro (Mid-2010) and after the restart my secondary display will not work. When I plug it in to my mini displayport my secondary display comes up black and my primary display comes back with the background and

  • Error when logging into account

    I'm getting an error when trying to log in to my account. I had 3 hard drives in my Mac and removed 2 of them and wiped/installed into another computer. I still have the hard drive with OSX on it installed, but when I try to log into my account it sa

  • Installing new software for ipod touch.

    Thanx to all who replyed to my situation. I have since been able to dwnload & install 4.0.2. software sucessfully to my ipod touch. Although it took at least 5 attempts I reinstalled itunes newest version (9.1.2) then clicked update,then unpluged my

  • Hp1200 v printing a page of PJL commands before the doucment is printed.

    Good afternoon, Firstly the details. Printer make \ model : Hp laserjet 1200 v Printer Driver             : Hp Laserjet 1200 series PCL5e Platform                   : Windows 2003 (sbs) SP2 Between printer and network there is a Castelle print server