Restart/resubmit job chain waiting of a file event

Hi Experts,
how can I restart/resubmit a job chain waiting of a file event?
In my case the variables OriginalPath and FinalPath are used in job chain parameters as an Default Expression (waitEvents...). If I restart the job chain I get the message "Could not evaluate default value for parameter...".
We are using SAP-CPS Build M28.20-37214.
Regards
Mathias

Hi,
You probably want to restart it because there was a failure somewhere in the chain. In this case it is easier to not let the chain end in error, but to let it go to status Console using the "Request Restart" postcondition.
Then when the step is restarted, or the chain is restarted from the beginning, it is restarted within the original chain and it reuses the already determined parameter values.
Regards,
Anton.

Similar Messages

  • Restart in Job chains and managing DST

    Hi,
    1) Given a 5 step job chain, if step 3 fails how is this handled in CPS? The following steps are chained but can the errored step be fixed and rerun to continue the job chain processing?
    2) How is day light saving handled for particular jobs. Is there anywhere this is captured in the detailed job request form solman?
    Regards,
    Esha

    Hi,
    1) Given a 5 step job chain, if step 3 fails how is this handled in CPS? The following steps are chained but can the errored step be fixed and rerun to continue the job chain processing?
       -- Yes, you can restart at any step of Job chain. (Not not at call level, remember one step might have multiple call inside it)
    2) How is day light saving handled for particular jobs. Is there anywhere this is captured in the detailed job request form solman?
    -- Use GMT timezone to submit jobs. There is nothing on redwood which can solve day light saving. It alwasy has to be at OS level. But redwood has all timezones availabe and you can submits job in any timezone.
    Regards,
    Sunil Rohit

  • Job Chain Restart Behaviour

    If a Job Chain with 3 steps fails in the second step - we would like the restart to resume execution with the second step, bypassing execution of the first step.   What are the correct settings to enable restarting the chain with the failed step? Thanks

    Hi Noel,
    For now, you could put an "on completed" "request restart" on the SendMessage step. If the SendMessage runs, there was an error so you would want to restart, but the restart options have some restrictions then because they don't take the step of the error but the SendMessage step as a reference now (you have the request restart on the send message step).
    Last but not least: if you already inform someone, you can also simply right click a step in the monitor and choose restart, this should restart the chain but from the step you selected. This is the same as submitting a new chain but specifying the step where it should start (on the same page in the submit wizard where you specify the queue).
    There also is an add-on available from Redwood called "active alerting" which allows you to do rule-based alerting (depending on source of the error send message to different teams) as well as escalation (if no reply within x minutes, send to next team) very easily.
    Regards,
    Anton.

  • Job Chaining and Quickcluster

    I always get
    Status: Failed - HOST [Macintosh.local] QuickTime file not found.
    after the first part of the job is successful.
    If I just submit with "This Computer" it works fine. Original file is ProRes 422, first job uses ProRes 422 to scale it down to 480x270. Second job compresses to h.264. I found some info on this board from 2008 saying that job chaining and quickclusters don't work together. Is that still how it is? that's really useless..
    I also found this from Jan 2009
    David M Brewer said:
    The reason the second rendering is failing is.......this has happen to me a few times until I figured it out.....make sure you set the dimension to the video in the h.264 settings, set to the same size as the Pro Rez dimensions.
    For the most part the dimensions are left blank for the second link, h.264. And don't use 100% of source. Put the physical numbers into the spaces. When you link one video to another, the second codec doesn't know the specs settings you made for the first video settings.
    Also make sure you (at least check) set the audio for the second video. I usually have the Pro Res do the audio conversion and just past-through to the second video settings. Again it can happen the the audio is disable in the h.264 settings. This has happen a few time for me........... Check and double check your settings!
    He doesn't mention anything about with or w/o Quickclusters, but I tried what he said and could not get it to work with quickclusters...
    Anyone got any new info on this?

    Studio X,
    Thanks for taking the time to run some tests and post your results.
    I'm finding the same results with converting ProRes422 to mp4, But...
    Other codecs are giving me very different results.
    I've run some random tests to try to get a grip on whats happening.
    First I was playing around with the # of instances. I've read here and on Barefeats that (at least for my model MacPro) the instances should be set to (# of processors/2), so I've been using 4 for quite a while now and thought I'd test it for myself.
    A single 5min ProRes422 1920x1080 29.97 file to h.264
    This Computer- 15:28
    2 Instances- 14:56
    3 Instances- 13:52
    4 Instances- 14:48
    5 Instances- 13:43
    6 Instances- 13:48
    7 Instances- 13:58
    In this case 5i was the fastest but not using a Quickcluster wasn't far off
    A single 2m30s ProRes422 1920x1080 29.97 file to h.264
    This Computer- 3:19
    2 Instances- 3:45
    3 Instances- 3:45
    4 Instances- 3:45
    5 Instances- 3:50
    6 Instances- 4:00
    7 Instances- 4:00
    Interesting...not using a Quickcluster is fastest
    A single 2m30s ProRes422 1920x1080 29.97 file Scaled down using original codec
    This Computer- 5:20
    4 Instances- 4:10
    5 Instances- 4:10
    7 Instances- 4:11
    A single 1m30s ProRes422 1920x1080 29.97 file to mpeg-2
    This Computer- 2:12
    5 Instances- 2:10
    When Quickclusters are faster, 4-5 instances does seem to be the sweet spot(again for my setup).
    In the mpeg-2 test, I should have used a longer clip to get a better test but it was getting late and I was just tring to get an idea of the codecs usage of my resources. I was also monitoring CPU usage with Activity Monitor in all tests.
    Now multiclip batches:
    I forgot to write down the length of the clips in this first test but it consisted of 8 ProRes 422 clips. 3 about 1m long and the rest between 13s and 30s
    8 ProRes 422 clips to mp4
    This Computer- 11:25
    4 Instances- 5:16
    Same results as Studio X
    Next tests with 5 clips(total 1m51s)
    5 ProRes 422 clips to h.264
    This Computer- 5:00
    4 Instances- 4:52
    5 ProRes 422 clips to mpeg-2
    This Computer- 2:55
    4 Instances- 3:01
    5 ProRes 422 clips to DV NTSC
    This Computer- 6:40
    4 Instances- 5:12
    5 ProRes 422 clips to Photo Jpeg
    This Computer- 2:44
    4 Instances- 2:46
    I re-ran the last test with 7 clips because of the time it took reassemble the segmented clips
    7 ProRes 422 clips to Photo Jpeg(total 3m14s)
    This Computer- 4:43
    4 Instances- 3:41
    One last test,
    A single ProRes 422 clip to Photo Jpeg(4:05;23)
    This Computer- 5:52
    4 Instances- 4:10
    Let me start off by saying it is clear that there are many factors that effect compression times such as # of clips, length of clips, and codecs, but here are some of the things I noted:
    1)Some codecs themselves seem to be "more aware" of the computers resources than others.
    When I compress to h.264 w/o a cluster it will use about 80-85% of all resources
    When I compress to h.264 with a cluster it will use about 90-95% of all resources
    When I compress to PhotoJpeg w/o a cluster it will use about 20-25% of all resources
    When I compress to PhotoJpeg with a cluster it will use about 80-85% of all resources
    2)The time it takes to reassemble clips can be quite long and could effect overall speed
    In the very last test, compressing a single file to photoJpeg using 4 instances took 4m10s. Watching Batch Monitor I noted that it took 2m0s to compress and 2m10s to reassemble.Wow...
    It would be interesting to see how the disassemble/reassemble of bigger and larger batches using clusters effect overall time. But that would take some time.
    I think the thing I will be taking with me from all of this is your workflow is your own. If you want to optimize it, you should inspect it, test it and adjust it where it needs adjusting. Now if anyone has the time and was to run similar tests with very different results I'd love to know about it...

  • How to Schedule a Job Chain to start automatically on SAP CPS.

    Hi,
    I did a job chain and i want to run automatically on sap cps Tuesday thru Saturday at 6:00 a.m., i make a calendar on sap cps with this specific options but the job chain doesn't start running.  I don't know if i need to do something more, so if someone can give a little help with this i will apreciate a lot.
    Thanks,
    Omar

    It finished ok but on the operator message i got the following message:
    Unable to resubmit this job.
    Details:
    com.redwood.scheduler.api.exception.TimeWindowExpectedOpenWindowException: CalculateNextClose should only be called on an open time window
    at com.redwood.scheduler.model.method.impl.TimeWindowMethodImpl.calculateNextCloseIntersectionInt(TimeWindowMethodImpl.java:388)
    at com.redwood.scheduler.model.method.impl.TimeWindowMethodImpl.calculateNextCloseIntersectInt(TimeWindowMethodImpl.java:249)
    at com.redwood.scheduler.model.TimeWindowImpl.calculateNextCloseIntersectInt(TimeWindowImpl.java:212)
    at com.redwood.scheduler.model.method.impl.SubmitFrameMethodImpl.calculateNextInt(SubmitFrameMethodImpl.java:178)
    at com.redwood.scheduler.model.SubmitFrameImpl.calculateNext(SubmitFrameImpl.java:176)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.resubmitSubmitFrameJob(JobStatusChangePrepareListener.java:763)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.resubmitJob(JobStatusChangePrepareListener.java:637)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.processJobToFinalState(JobStatusChangePrepareListener.java:520)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.modelModified(JobStatusChangePrepareListener.java:233)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.informListeners(LowLevelPersistenceImpl.java:728)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.writeDirtyObjectListRetry(LowLevelPersistenceImpl.java:207)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.access$000(LowLevelPersistenceImpl.java:38)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl$WriteDirtyObjectListUnitOfWork.execute(LowLevelPersistenceImpl.java:79)
    at com.redwood.scheduler.persistence.impl.PersistenceUnitOfWorkManager.execute(PersistenceUnitOfWorkManager.java:34)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.writeDirtyObjectList(LowLevelPersistenceImpl.java:102)
    at com.redwood.scheduler.cluster.persistence.ClusteredLowLevelPersistence.writeDirtyObjectList(ClusteredLowLevelPersistence.java:59)
    at com.redwood.scheduler.model.SchedulerSessionImpl.writeDirtyListLocal(SchedulerSessionImpl.java:648)
    at com.redwood.scheduler.model.SchedulerSessionImpl.persist(SchedulerSessionImpl.java:626)
    at com.redwood.scheduler.apiint.model.UnitOfWorkManager.perform(UnitOfWorkManager.java:32)
    at com.redwood.scheduler.apiint.model.UnitOfWorkManager.perform(UnitOfWorkManager.java:13)
    at com.redwood.scheduler.jobchainservice.JobChainService.childJobFinalStatus(JobChainService.java:223)
    at com.redwood.scheduler.core.processserver.ProcessServerRuntime.childJobFinalStatus(ProcessServerRuntime.java:836)
    at com.redwood.scheduler.core.processserver.ProcessServerRuntime.onMessage(ProcessServerRuntime.java:248)
    at com.redwood.scheduler.infrastructure.work.MessageEnabledWork.run(MessageEnabledWork.java:104)
    at com.redwood.scheduler.infrastructure.work.WorkerImpl.run(WorkerImpl.java:109)
    at java.lang.Thread.run(Thread.java:534)

  • OIA jobs.xml constrain execute of file import on only one of two app server

    I am trying to get the jobs.xml and scheduling-context.xml files to fire an import job on one of the two app servers. There are two application servers utilizing the same shared database. Files to import into OIA reside on one of the two application servers file systems only and we want to use the jobs.xml and scheduling-context.xml files on one of the app servers to setup the imports.
    I have these two beans enabled
    <ref bean="usersImportJob"/>
    <ref bean="usersImportTrigger"/>
    This is a snippet from the jobs.xml file
    <bean id="usersImportTrigger" class="org.springframework.scheduling.quartz.CronTriggerBean">
    <property name="jobDetail">
    <ref bean="usersImportJob"/>
    </property>
    <property name="cronExpression">
    <value>0 0/5 * * * ?</value>
    </property>
    </bean>
    <bean id="usersImportJob" class="org.springframework.scheduling.quartz.JobDetailBean">
    <property name="name">
    <value>Users Import</value>
    </property>
    <property name="description">
    <value>Users import Job</value>
    </property>
    <property name="jobClass">
    <value>com.vaau.rbacx.scheduling.manager.providers.quartz.jobs.IAMJob</value>
    </property>
    <property name="group">
    <value>SYSTEM</value>
    </property>
    <property name="durability">
    <value>true</value>
    </property>
    <property name="jobDataAsMap">
    <map>
    <!-- only single user name can be specified for jobOwnerName (optional)-->
    <entry key="jobOwnerName">
    <value>REPLACE_ME</value>
    </entry>
    <!-- multiple user names can be specified as
    comma delimited e.g user1,user2 (optional)-->
    <entry key="usersToNotify">
    <value>REPLACE_ME</value>
    </entry>
    <entry key="IAMActionName">
    <value>ACTION_IMPORT_USERS</value>
    </entry>
    <entry key="IAMServerName">
    <value>FILE_SERVER</value>
    </entry>
    <!-- Job chaining, i.e. specify the next job to run (optional) -->
    <entry key="NEXT_JOB">
    <value>rolesImportJob</value>
    </entry>
    </map>
    </property>
    </bean>
    QUESTION:_
    Is the IAMServerName value suppose to have the literal hostname of the application server (ex. "oia.bluebird.com") configured in place of the default "FILE_SERVER" text?
    Your insite is greatly appreciated

    Hi Meg,
    I have to say that I've only seen the jobs.xml be defined for data feeds only. It does not make calls out to app/DB servers OOTB.
    The alternative solution are the following:
    1. Use your IDM provisioning connector to the application then pull the data from the IDM product into OIA
    2. Extract the data from the app server using the ETL functionality/engine within the OIA product to extract the data into a data feed (then OIA will pick that up)
    3. Construct some extracting functionality (like a java class) instead of ETL
    Hope this has helped
    Regards,
    Daniel Redfern
    Technicalconfessions.com

  • Backing up Jobs, Chains and Programs in Oracle Job Scheduler

    What is the best way to back up Jobs, Chains and Programs created in the Oracle Job Scheduler via Enterprise Manager - and also the best way to get them from one database to another. I am creating quite a long chain which executes many programs in our test database and wish to back everything up along the way. I will also then need to migrate to the production database.
    Thanks for any advice,
    Susan

    Hi Susan,
    Unfortunately there are not too many options.
    To backup a job you can use dbms_scheduler.copy_job. I believe EM has a button called "create like" for jobs and programs but I am not sure about chains and this can be used to create backups as well.
    A more general purpose solution which should also cover chains is to do a schema-level export using expdp i.e. a dump of an entire schema.
    e.g.
    SQL> create directory dumpdir as '/tmp';
    SQL> grant all on directory dumpdir to public;
    # expdp scott/tiger DUMPFILE=scott_backup.dmp directory=dumpdir
    You can then import into a SQL text file e.g.
    # impdp scott/tiger DIRECTORY=dumpdir DUMPFILE=scott_backup SQLFILE=scott_backup.out
    or import into another database (and even another schema) e.g.
    # impdp scott/tiger DIRECTORY=dumpdir DUMPFILE=scott_backup
    Hope this helps,
    Ravi.

  • How to setup job chain with time dependant steps

    Hi
    I need to setup the following job chain with next job step waiting until specified time before starting
    Step 1 Job A Time 07:30 if complete go to job step 2
    Step 2 Job B start time 09:00 if complete go to job step 3
    Step 3 Job C start time 09:30 (end job)
    Have tried to use a pre-condition in step 2 and 3 but that just caused thoes steps to be skipped.
    can somebody point me in the right direction
    Thanks in advance
    Jon

    Hi Anton/Babu
    Thanks for your help, I managed to set this up by embedding multiple Job Chains with their own time window in a master job chain.
    As we continue our roll out not sure how scalable the solution will be as you have to create multiple job chains for essentially the same job with the only difference being a specific time window.
    I guess only time will tell.
    Thanks for you help again guys
    Jon

  • Job Chain in RAC Environment

    Hi,
    i've a job chain defined in a RAC environment.
    I need that all the steps of my chain are performed on the same instance.
    How can i do this?
    Now stepA run on the instance 1 and stepB run on the instance 2.
    Thanks.

    Again, If it is a file system write issue, use ACFS (11.2.0.x) to create a shared file system that all nodes can see. What are you using to delete/create these files? A shell script?
    1) create an ACFS file system of sufficient size to handle your data
    2) mkdir /some/acfs/location
    3) using sqlplus create a database directory " create directory foo as '/some/acfs/location' "
    4) put the file in this location
    5) use UTL_FILE to delete AND create the file (see fremove() - http://docs.oracle.com/cd/E11882_01/appdev.112/e10577/u_file.htm)

  • SAP CS 8.0 - question to parallel tasks in one job chain

    Dear all,
    I have a question about job chains in SAP CPS/Redwood
    We have a job chain like this:
    Job 1
         Job 1.1
                   Job 1.1.1
                   Job 1.1.2
         Job 1.2
                    Job 1.2.1
    The Jobs 1.1.1 and 1.1.2 should start  when the Job 1.1 is compled and don't have to wait until Job 1.2 is complete.
    How could I release it in Redwood in one job chain? Should I use a precondition in the job definition of 1.1 and 1.2?
    Thank you for your help.
    Best regards,
    Hans

    Assumptions :
    Job 1.1.1 and Job 1.1.2 are running in parallel.
    Job 1.1, job 1.2 are also running in parallel.
    Chain A -
    Step 1 - Job 1.1
    Step 2 - Job 1.1.1, Job 1.1.2 (these both will start as soon as Job 1.1 completes)
    Chain B -
    Step 1 - Job 1.2
    Step 2 - Job 1.2.1
    Chain C
    Step 1 - Job 1
    Step 2 - Chain A, Chain B (Chain A and Chain B will start as soon as Job 1 completes).
    Submit Chain C.
    Preconditions are usually  used to check the timewindow, for example , if you want to execute a step in a job chain, only if it is a Friday.
    thanks
    Nanda

  • Grafical overview of Job chains with dependencies by event

    Hi,
    we want to activate/monitor various process chains in APO and BW SAP systems which are dependent of each other.
    We intend to handle this with Events (wait events and raise events).
    For this reason we have created a job chain for each process chain. In the first and only step the job definition 'SAP_BW_ProcessChainRun' is started and activates the 'Technical Name of SAP BW Process Chain:
    Do we have a chance to get a graphical overview over all with event sceduling realized dependent job chains?
    Thanks
    Br
    Günter

    Hi,
    If there are dependencies between your BW Process Chains, you can use something like a master Job Chain definition in CPS to set up your dependencies between them (with failure or success outcomes, based on the step statuses). For chains that can run in parallel you can have multiple jobs (that can be other job chains) in a job chain step, and depedent jobs can be other steps that are triggered when a previous step completes / fails. Just check of course that thes jobs are your process chain jobs and not just event triggers.
    [link to SAP help for BW chains and CPS|http://help.sap.com/saphelp_nw04s/helpdata/de/bb/033b3c9f8c4b5ca34045688c6a9f7b/content.htm]

  • IBook G3 will not boot! Hangs at "Waiting for Apple Files"

    My iBook g3 will not boot past the "Waiting for Apple Files", Blue screen. I have tried everything from http://docs.info.apple.com/article.html?artnum=107210 to http://docs.info.apple.com/article.html?artnum=106214. I only have a Panther OSX 10.3 Update CD. Someone please help me!

    Mark:
    Welcome to Apple Discussions.
    I am not sure why you tried replacing the NetInfo database, or what effect that may have or have had on your startup. Here is where I would start:
    Reset PRAM and Safe Boot
    1. Shut down the computer.
    2. Locate the following keys on the keyboard: Command, Option, P, and R. You will need to hold these keys down simultaneously in step 4.
    3. Turn on the computer.
    4. Press and hold the Command-Option-P-R keys. You must press this key combination before the gray screen appears.
    5. Hold the keys down until the computer restarts and you hear the startup sound for the third time.
    6. Release the keys.
    7. Immediately hold down the "Shift" key until grey apple appears. This will make the computer start up in Safe Mode
    8. Log in and empty Trash if needed.
    9. Restart normally and log in.
    Please post back with update or further questions.
    Good luck.
    cornelius

  • Error while scheduling the Email Alert JOB chain

    Hi All,
    I have defined a job chain in CPS and when i am going to schedule it then it is giving me error message.We have taken the trial version.
    Please find the log attached below.
    11:18:31 PM:
    JCS-111004: Queue ETD.sapetd00_Queue has no ProcessServer with the required JobDefinitionType/Service/Resource for Job 932 (submitted from ETD.Z_MONI_BATCH_DP copy from 2009/12/30 18:22:23,113 Australia/Sydney) (submitted from Job Definition ETD.Z_MONI_BATCH_DP (Copy from 2009/12/30 18:22:23,113 Australia/Sydney)): Job Definition Type CSH/Service PlatformAgentService/"Empty"
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Job) to an object in an isolation group
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Chain Step) to an object in an isolation group
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Parent Job) to an object in an isolation group Show error details
    Thanks
    Rishi Abrol

    Hi
    Are you logged into the correct isolation group ?
    Ensure the process server is also assigned to the queue.
    Regards

  • Cheque printing error Restart number is not in the check file

    Dear Sap friends,
    I need to print the cheque using the standard program FBZ5. Since the client requires to change the description of vendor i am using FCH5 and manually creating the cheque for the payment document number.
    After the cheque no is attached to this document number i go to FBZ5 and am entering  this docuemnt number for printing the cheque.
    Initially i am getting the error
    You cannot print the manually created check AX002 AX002 500000
    Message no. FS675.
    This i have resovled by changing the message from error to warning.
    Later when i press enter i am getting the following error.
    Restart number is not in the check file or is not a From-number
    Message no. FS562
    How to i resolve this.
    Please help.
    Satish

    Exactly,
    If you really need such kind of facility, ask your ABAPER to develop a small report like below
    Selection criteria
    Document number - BELNR
    Conmpany code - BUKRS
    Fiscal year - GJAHR
    Output criteria
    Vendor name
    Amount
    Date
    Form has to develop for Check prinitng
    Thanks

  • Restart number is not in the check file or is not a From-number

    HI!
    I had  posted  payment Document  but i  got below error while  reprint  check from  FBZ5
    Restart number is not in the check file or is not a From-number
    "Message no. FS562"
    on that base, I had  check below  settinng 
    1) I  had check  Cheque no  register . It's  proper in series.
    2)Tcd:FBZP -> Pmt methods in country -> select your country
       and payment method 'C'. If 'Bank details' is checked on, then bank
       details of vendor master is mandatory for that payment.
    Awaiting for reply soon,
    thanking you,
    Rupang shah

    Hi,
    Commonly it will come for below situations:
    1. Client decided to use this check lots for manual payment & this checks already used, so this type of checks system not allowed to take reprint.
    2. You can only reprint checks created with F110, FBZ4, F-58 or FBZ5  (this checks have been printed by the system)
    Check the Payment method wise checks it is already issued / not PAYR - RZAWE
    system not allowed to reprint already exist in PAYR table checks.
    if you already use one check with payment method C now you want to use the same lot for payment method X then system not allow to print.
    Regards,
    Viswa

Maybe you are looking for

  • I need a new drive for my G4 Mac mini - where to start??

    Hello all, I have a approximately 4-5 year old 1.5GHz Power PC G4 Mac mini with 1 GB DDR SDRAM memory and an 80 GB hard drive. I was having issues with it running slow, not being able to access different files/applications, etc. so I ran Disk Warrior

  • Will not start with hd

    Insatalled a WD5000 hd in intel  21" imac 2006 worked fine for a couple of months then tried to switch on and got the Apple logo and spinning wheel. Tried to reset PRAM, starting from DVD and in Target mode - nothing. Will not respond to keyboard.  R

  • HT3924 using 2008 imac 24" as display for 2011 macbook pro

    can i use my  2008 iMac 24" as the display for my 2011 Mac book pro

  • Message no. RM109 The quantity at RP  is less than the quantity at next RP

    Hello everyone, I am getting an error in regards to a reporting point.  I tried going into MF4R to reset the reporting points and when I do I get the following message: The quantity at RP 0220 is less than the quantity at the next RP Message no. RM10

  • Best practice - applying Filters

    What's the best practice in terms of performance versus code reuse, when applying filters? Right now, I see 3 ways to specify a filter: 1. as a <mx:filter> or <s:filter> 2. in actionscript: var glowFilter:GlowFilter = new GlowFilter(); --OR-- 3. in S