Restart in Job chains and managing DST

Hi,
1) Given a 5 step job chain, if step 3 fails how is this handled in CPS? The following steps are chained but can the errored step be fixed and rerun to continue the job chain processing?
2) How is day light saving handled for particular jobs. Is there anywhere this is captured in the detailed job request form solman?
Regards,
Esha

Hi,
1) Given a 5 step job chain, if step 3 fails how is this handled in CPS? The following steps are chained but can the errored step be fixed and rerun to continue the job chain processing?
   -- Yes, you can restart at any step of Job chain. (Not not at call level, remember one step might have multiple call inside it)
2) How is day light saving handled for particular jobs. Is there anywhere this is captured in the detailed job request form solman?
-- Use GMT timezone to submit jobs. There is nothing on redwood which can solve day light saving. It alwasy has to be at OS level. But redwood has all timezones availabe and you can submits job in any timezone.
Regards,
Sunil Rohit

Similar Messages

  • Backing up Jobs, Chains and Programs in Oracle Job Scheduler

    What is the best way to back up Jobs, Chains and Programs created in the Oracle Job Scheduler via Enterprise Manager - and also the best way to get them from one database to another. I am creating quite a long chain which executes many programs in our test database and wish to back everything up along the way. I will also then need to migrate to the production database.
    Thanks for any advice,
    Susan

    Hi Susan,
    Unfortunately there are not too many options.
    To backup a job you can use dbms_scheduler.copy_job. I believe EM has a button called "create like" for jobs and programs but I am not sure about chains and this can be used to create backups as well.
    A more general purpose solution which should also cover chains is to do a schema-level export using expdp i.e. a dump of an entire schema.
    e.g.
    SQL> create directory dumpdir as '/tmp';
    SQL> grant all on directory dumpdir to public;
    # expdp scott/tiger DUMPFILE=scott_backup.dmp directory=dumpdir
    You can then import into a SQL text file e.g.
    # impdp scott/tiger DIRECTORY=dumpdir DUMPFILE=scott_backup SQLFILE=scott_backup.out
    or import into another database (and even another schema) e.g.
    # impdp scott/tiger DIRECTORY=dumpdir DUMPFILE=scott_backup
    Hope this helps,
    Ravi.

  • Restart/resubmit job chain waiting of a file event

    Hi Experts,
    how can I restart/resubmit a job chain waiting of a file event?
    In my case the variables OriginalPath and FinalPath are used in job chain parameters as an Default Expression (waitEvents...). If I restart the job chain I get the message "Could not evaluate default value for parameter...".
    We are using SAP-CPS Build M28.20-37214.
    Regards
    Mathias

    Hi,
    You probably want to restart it because there was a failure somewhere in the chain. In this case it is easier to not let the chain end in error, but to let it go to status Console using the "Request Restart" postcondition.
    Then when the step is restarted, or the chain is restarted from the beginning, it is restarted within the original chain and it reuses the already determined parameter values.
    Regards,
    Anton.

  • Job Chaining and Quickcluster

    I always get
    Status: Failed - HOST [Macintosh.local] QuickTime file not found.
    after the first part of the job is successful.
    If I just submit with "This Computer" it works fine. Original file is ProRes 422, first job uses ProRes 422 to scale it down to 480x270. Second job compresses to h.264. I found some info on this board from 2008 saying that job chaining and quickclusters don't work together. Is that still how it is? that's really useless..
    I also found this from Jan 2009
    David M Brewer said:
    The reason the second rendering is failing is.......this has happen to me a few times until I figured it out.....make sure you set the dimension to the video in the h.264 settings, set to the same size as the Pro Rez dimensions.
    For the most part the dimensions are left blank for the second link, h.264. And don't use 100% of source. Put the physical numbers into the spaces. When you link one video to another, the second codec doesn't know the specs settings you made for the first video settings.
    Also make sure you (at least check) set the audio for the second video. I usually have the Pro Res do the audio conversion and just past-through to the second video settings. Again it can happen the the audio is disable in the h.264 settings. This has happen a few time for me........... Check and double check your settings!
    He doesn't mention anything about with or w/o Quickclusters, but I tried what he said and could not get it to work with quickclusters...
    Anyone got any new info on this?

    Studio X,
    Thanks for taking the time to run some tests and post your results.
    I'm finding the same results with converting ProRes422 to mp4, But...
    Other codecs are giving me very different results.
    I've run some random tests to try to get a grip on whats happening.
    First I was playing around with the # of instances. I've read here and on Barefeats that (at least for my model MacPro) the instances should be set to (# of processors/2), so I've been using 4 for quite a while now and thought I'd test it for myself.
    A single 5min ProRes422 1920x1080 29.97 file to h.264
    This Computer- 15:28
    2 Instances- 14:56
    3 Instances- 13:52
    4 Instances- 14:48
    5 Instances- 13:43
    6 Instances- 13:48
    7 Instances- 13:58
    In this case 5i was the fastest but not using a Quickcluster wasn't far off
    A single 2m30s ProRes422 1920x1080 29.97 file to h.264
    This Computer- 3:19
    2 Instances- 3:45
    3 Instances- 3:45
    4 Instances- 3:45
    5 Instances- 3:50
    6 Instances- 4:00
    7 Instances- 4:00
    Interesting...not using a Quickcluster is fastest
    A single 2m30s ProRes422 1920x1080 29.97 file Scaled down using original codec
    This Computer- 5:20
    4 Instances- 4:10
    5 Instances- 4:10
    7 Instances- 4:11
    A single 1m30s ProRes422 1920x1080 29.97 file to mpeg-2
    This Computer- 2:12
    5 Instances- 2:10
    When Quickclusters are faster, 4-5 instances does seem to be the sweet spot(again for my setup).
    In the mpeg-2 test, I should have used a longer clip to get a better test but it was getting late and I was just tring to get an idea of the codecs usage of my resources. I was also monitoring CPU usage with Activity Monitor in all tests.
    Now multiclip batches:
    I forgot to write down the length of the clips in this first test but it consisted of 8 ProRes 422 clips. 3 about 1m long and the rest between 13s and 30s
    8 ProRes 422 clips to mp4
    This Computer- 11:25
    4 Instances- 5:16
    Same results as Studio X
    Next tests with 5 clips(total 1m51s)
    5 ProRes 422 clips to h.264
    This Computer- 5:00
    4 Instances- 4:52
    5 ProRes 422 clips to mpeg-2
    This Computer- 2:55
    4 Instances- 3:01
    5 ProRes 422 clips to DV NTSC
    This Computer- 6:40
    4 Instances- 5:12
    5 ProRes 422 clips to Photo Jpeg
    This Computer- 2:44
    4 Instances- 2:46
    I re-ran the last test with 7 clips because of the time it took reassemble the segmented clips
    7 ProRes 422 clips to Photo Jpeg(total 3m14s)
    This Computer- 4:43
    4 Instances- 3:41
    One last test,
    A single ProRes 422 clip to Photo Jpeg(4:05;23)
    This Computer- 5:52
    4 Instances- 4:10
    Let me start off by saying it is clear that there are many factors that effect compression times such as # of clips, length of clips, and codecs, but here are some of the things I noted:
    1)Some codecs themselves seem to be "more aware" of the computers resources than others.
    When I compress to h.264 w/o a cluster it will use about 80-85% of all resources
    When I compress to h.264 with a cluster it will use about 90-95% of all resources
    When I compress to PhotoJpeg w/o a cluster it will use about 20-25% of all resources
    When I compress to PhotoJpeg with a cluster it will use about 80-85% of all resources
    2)The time it takes to reassemble clips can be quite long and could effect overall speed
    In the very last test, compressing a single file to photoJpeg using 4 instances took 4m10s. Watching Batch Monitor I noted that it took 2m0s to compress and 2m10s to reassemble.Wow...
    It would be interesting to see how the disassemble/reassemble of bigger and larger batches using clusters effect overall time. But that would take some time.
    I think the thing I will be taking with me from all of this is your workflow is your own. If you want to optimize it, you should inspect it, test it and adjust it where it needs adjusting. Now if anyone has the time and was to run similar tests with very different results I'd love to know about it...

  • Automating start / restart of WebLogic Admin and managed servers on MS Windows

    Hi All,
    Please help me to do the automation of start and stop scrip of Oracle Weblogic 10.3.6 and Forms and Reports 11.1.2.2.0(Managed Servers) On Windows server 2008.
    Here are two options for automating start / stop of WebLogic Admin and managed servers on MS Windows. Please suggesst which one is standrad procedure.
    Use NodeManager (installed as a MS Windows Service)
    or 
    Create Windows Service(s) for the server(s) - no NodeManager
    Following Document : 1060058.1
    Please suggest me the best practice. Many thanks in advance.
    Kind Regards,
    Mohan

    Hi Borys,
    Thank you for quick reply...
    If we use the first option "Create Windows Service(s) for the server(s) - no NodeManager" then we will be facing issue Like "Admin Server or Managed Server may go into an unhealthy or failed state but the JVM is still running"
    In the document : 1060058.1 : Mentioned as Important Note: The second option has a limitations.
    Limitaion is below when we use first option : "Create Windows Service(s) for the server(s) - no NodeManager" -
    Windows Services Manager only monitors the JVM process. If the JVM process fails (shuts down), then Windows Services Manager will attempt to restart it. But there are some scenarios where the Admin Server or Managed Server may go into an unhealthy or failed state but the JVM is still running. The Windows Service will not know to restart the process in these cases. If a server is started using NodeManager, NodeManager would recognize such a state and restart the process accordingly. Furthermore, if you start up managed servers via a MS Windows Service, you cannot monitor and/ or stop the managed server via the Weblogic Admin Console and/or Fusion Middleware ( Enterprise Manager ) Control.
    I have implemented the second option "Use NodeManager starting admin and manged servers via WLST script" - Python script. Its working well without issue...No Limitation...
    Script Contains Below :
    print "================== Start scripts of Managed servers and Admin server ======================"
    nmConnect(userConfigFile='D:/myFiles/myuserconfigfile.secure',userKeyFile='D:/myFiles/myuserkeyfile.secure',host='00.0.00.00',port='5556',domainName='ClassicDomain',domainDir='D:\Oracle\Middleware\user_projects\domains\ClassicDomain',nmType='ssl')
    print "================== Adminserver Starting ================================================"
    nmStart('AdminServer')
    connect('weblogic','Password1','10.4.93.117:7001')
    print "================== MANAGED SERVER (FROMS) Starting ====================================="
    start('WLS_FORMS','Server')
    print "================== MANAGED SERVER (REPORTS) Starting ==================================="
    start('WLS_REPORTS','Server')
    exit()
    HOPE WE WILL MEET AGAIN AND AGAIN...
    Many thanks for your comments!
    Kind Regards,
    Mohan

  • How to Schedule a Job Chain to start automatically on SAP CPS.

    Hi,
    I did a job chain and i want to run automatically on sap cps Tuesday thru Saturday at 6:00 a.m., i make a calendar on sap cps with this specific options but the job chain doesn't start running.  I don't know if i need to do something more, so if someone can give a little help with this i will apreciate a lot.
    Thanks,
    Omar

    It finished ok but on the operator message i got the following message:
    Unable to resubmit this job.
    Details:
    com.redwood.scheduler.api.exception.TimeWindowExpectedOpenWindowException: CalculateNextClose should only be called on an open time window
    at com.redwood.scheduler.model.method.impl.TimeWindowMethodImpl.calculateNextCloseIntersectionInt(TimeWindowMethodImpl.java:388)
    at com.redwood.scheduler.model.method.impl.TimeWindowMethodImpl.calculateNextCloseIntersectInt(TimeWindowMethodImpl.java:249)
    at com.redwood.scheduler.model.TimeWindowImpl.calculateNextCloseIntersectInt(TimeWindowImpl.java:212)
    at com.redwood.scheduler.model.method.impl.SubmitFrameMethodImpl.calculateNextInt(SubmitFrameMethodImpl.java:178)
    at com.redwood.scheduler.model.SubmitFrameImpl.calculateNext(SubmitFrameImpl.java:176)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.resubmitSubmitFrameJob(JobStatusChangePrepareListener.java:763)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.resubmitJob(JobStatusChangePrepareListener.java:637)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.processJobToFinalState(JobStatusChangePrepareListener.java:520)
    at com.redwood.scheduler.model.listeners.JobStatusChangePrepareListener.modelModified(JobStatusChangePrepareListener.java:233)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.informListeners(LowLevelPersistenceImpl.java:728)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.writeDirtyObjectListRetry(LowLevelPersistenceImpl.java:207)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.access$000(LowLevelPersistenceImpl.java:38)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl$WriteDirtyObjectListUnitOfWork.execute(LowLevelPersistenceImpl.java:79)
    at com.redwood.scheduler.persistence.impl.PersistenceUnitOfWorkManager.execute(PersistenceUnitOfWorkManager.java:34)
    at com.redwood.scheduler.persistence.impl.LowLevelPersistenceImpl.writeDirtyObjectList(LowLevelPersistenceImpl.java:102)
    at com.redwood.scheduler.cluster.persistence.ClusteredLowLevelPersistence.writeDirtyObjectList(ClusteredLowLevelPersistence.java:59)
    at com.redwood.scheduler.model.SchedulerSessionImpl.writeDirtyListLocal(SchedulerSessionImpl.java:648)
    at com.redwood.scheduler.model.SchedulerSessionImpl.persist(SchedulerSessionImpl.java:626)
    at com.redwood.scheduler.apiint.model.UnitOfWorkManager.perform(UnitOfWorkManager.java:32)
    at com.redwood.scheduler.apiint.model.UnitOfWorkManager.perform(UnitOfWorkManager.java:13)
    at com.redwood.scheduler.jobchainservice.JobChainService.childJobFinalStatus(JobChainService.java:223)
    at com.redwood.scheduler.core.processserver.ProcessServerRuntime.childJobFinalStatus(ProcessServerRuntime.java:836)
    at com.redwood.scheduler.core.processserver.ProcessServerRuntime.onMessage(ProcessServerRuntime.java:248)
    at com.redwood.scheduler.infrastructure.work.MessageEnabledWork.run(MessageEnabledWork.java:104)
    at com.redwood.scheduler.infrastructure.work.WorkerImpl.run(WorkerImpl.java:109)
    at java.lang.Thread.run(Thread.java:534)

  • Weblogic server and managed servers sudden shutdown

    Hi,
    when trying to restart both the admin and managed osb servers. after some time the admin and managed servers are getting shut down. This is what the information i got from logs. Appreciate your help in this regard
    ####<Aug 18, 2011 12:51:44 PM EDT> <Info> <Health> <weblogic.GCMonitor> <<anonymous>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000011> <1313686304697> <BEA-310002> <38% of the total memory in the server is free>
    ####<Aug 18, 2011 12:52:06 PM EDT> <Error> <JDBC> <[ACTIVE] ExecuteThread: '7' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000109> <1313686326579> <BEA-001112> <Test "SELECT 1 FROM DUAL" set up for pool "GLTOdatasource" failed with exception: "oracle.jdbc.xa.OracleXAException".>
    ####<Aug 18, 2011 12:52:06 PM EDT> <Info> <JDBC> <[ACTIVE] ExecuteThread: '7' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000109> <1313686326580> <BEA-001128> <Connection for pool "GLTOdatasource" closed.>
    ####<Aug 18, 2011 12:52:44 PM EDT> <Info> <Health> <weblogic.GCMonitor> <<anonymous>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000011> <1313686364766> <BEA-310002> <61% of the total memory in the server is free>
    ####<Aug 18, 2011 12:56:45 PM EDT> <Info> <Health> <weblogic.GCMonitor> <<anonymous>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000011> <1313686605051> <BEA-310002> <44% of the total memory in the server is free>
    ####<Aug 18, 2011 12:58:18 PM EDT> <Info> <Security> <[ACTIVE] ExecuteThread: '6' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000111> <1313686698580> <BEA-090516> <The CredentialMapper provider has preexisting LDAP data.>
    ####<Aug 18, 2011 1:02:23 PM EDT> <Notice> <WebLogicServer> <Thread-1> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000117> <1313686943979> <BEA-000388> <JVM called WLS shutdown hook. The server will force shutdown now>
    ####<Aug 18, 2011 1:02:23 PM EDT> <Alert> <gifosb_admin> <Thread-1> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000117> <1313686943996> <BEA-000396> <Server shutdown has been requested by <WLS Kernel>>
    ####<Aug 18, 2011 1:02:24 PM EDT> <Notice> <WebLogicServer> <Thread-1> <<WLS Kernel>> <> <8beb4c4d69347bb8:6c6553af:131ddaf5db4:-7ff7-0000000000000117> <1313686944005> <BEA-000365> <Server state changed to FORCE_SUSPENDING>
    Thanks,
    Bharath

    <JVM called WLS shutdown hook. The server will force shutdown now>This is your JVM process which is getting killed and hence server is forced to shutdown. On which OS, servers are running? How are you starting the server? If you are on windows, DO NOT close the command prompt through which you started the server.
    Regards,
    Anuj

  • Job Chain Restart Behaviour

    If a Job Chain with 3 steps fails in the second step - we would like the restart to resume execution with the second step, bypassing execution of the first step.   What are the correct settings to enable restarting the chain with the failed step? Thanks

    Hi Noel,
    For now, you could put an "on completed" "request restart" on the SendMessage step. If the SendMessage runs, there was an error so you would want to restart, but the restart options have some restrictions then because they don't take the step of the error but the SendMessage step as a reference now (you have the request restart on the send message step).
    Last but not least: if you already inform someone, you can also simply right click a step in the monitor and choose restart, this should restart the chain but from the step you selected. This is the same as submitting a new chain but specifying the step where it should start (on the same page in the submit wizard where you specify the queue).
    There also is an add-on available from Redwood called "active alerting" which allows you to do rule-based alerting (depending on source of the error send message to different teams) as well as escalation (if no reply within x minutes, send to next team) very easily.
    Regards,
    Anton.

  • Scheduling a BI job chain in Redwood

    The problem I am having is we are trying to schedule a BI job chain via Redwood software and are not getting any response. Within Redwood, I have executed these jobs IMPORT_BW_CHAINS, IMPORT_BW_CHAIN_DEFINITION, IMPORT_BW_INFOPACKAGES using BI job chain  0fcsm_cm_10 which is defined in BI as a job chain. These jobs run to completion but nothing is moved into Redwood to schedule as you would see from a import of a CCMS job. When I run job RUN_BW_CHAIN using the same BI job chain ID I receive the below error.  Not sure what I’m missing or doing with the process to get to schedule the BI job chains with Redwood.
    ORA-06502: PL/SQL: numeric or value error
    ORA-06512: at "RSI.RSIEXEC", line 1638
    ORA-06512: at "RSI.RSIEXEC", line 1759
    ORA-06512: at "RSI.RSI_RUN_BW_CHAIN", line 21
    ORA-06512: at "RSI.RSI_RUN_BW_CHAIN", line 80
    ORA-06512: at line 1
    ORA-06512: at "SYS.DBMS_SYS_SQL", line 1200
    ORA-06512: at "SYS.DBMS_SQL", line 323
    ORA-06512: at "SYSJCS.DDL", line 1085
    ORA-06512: at "SYSJCS.DDL", line 1118
    ORA-06512: at "SYSJCS.DDL", line 1177
    ORA-06512: at line 3
    JCS-00215: in statement RSOJ_EXECUTE_JOB

    I am also seeing the same issue 
    anton the last information  you requested
    The following products are installed in the Cronacle repository:
    Product                                  Version    Status
    Cronacle for SAP solutions               7.0.3      Production 
    Cronacle Forecast Module                 7.0.3      Production 
    Cronacle Reports Module                  7.0.3      Production 
    Cronacle &module Module                  7.0.2      development
    Cronacle Mail Module                     7.0.3      Production 
    Cronacle Audit Module                    7.0.2 r2.2 Production 
    Cronacle Process Manager for Web         7.0.3      Production 
    Cronacle Module Installer                7.0.3      Production 
    Cronacle Repository                      7.0.3.34   Production 
    Cronacle Monitor Module                  7.0.3      Production

  • How to setup job chain with time dependant steps

    Hi
    I need to setup the following job chain with next job step waiting until specified time before starting
    Step 1 Job A Time 07:30 if complete go to job step 2
    Step 2 Job B start time 09:00 if complete go to job step 3
    Step 3 Job C start time 09:30 (end job)
    Have tried to use a pre-condition in step 2 and 3 but that just caused thoes steps to be skipped.
    can somebody point me in the right direction
    Thanks in advance
    Jon

    Hi Anton/Babu
    Thanks for your help, I managed to set this up by embedding multiple Job Chains with their own time window in a master job chain.
    As we continue our roll out not sure how scalable the solution will be as you have to create multiple job chains for essentially the same job with the only difference being a specific time window.
    I guess only time will tell.
    Thanks for you help again guys
    Jon

  • 6398 ERRORS Cache cluster is down, restart the cache cluster and Retry. Collection was modified; enumeration operation may not execute.

    Recently started getting these 6398 errors with SharePoint 2013 Single Farm and haven't been able to fix with any google results.  Everything seems to run fine.  They usually appear overnight.  Average 5-9 daily event logs.  See
    Errors below.
    Get-CacheHost
    always reports UP
    Have already tried: 
    -  Upgrading to AppFabric CU5
    -  Restart-Service AppFabricCachingService
    -  Clear Configuration Cache
    - habaneroconsulting Distributed Cache Bug
    - nobadthing Unexpected exception in feedcacheservice
    - mmman itgroove fixing the appfabric cache cluster
    - Microsoft unable to start appfabriccachingservice
    - dhasalprashantsharepoint lets-troubleshoot-sharepoint
    ...and many other readings.  (sorry I cant post links even though I verified my account 100 times)
    Errors:
    The Execute method of job definition Microsoft.Office.Server.UserProfiles.LMTRepopulationJob (ID 414cbbe9-cdb1-4f7a-beed-85fbfd8a10c7) threw an exception. More information is included below.
    Unexpected exception in FeedCacheService.IsRepopulationNeeded: Cache cluster is down, restart the cache cluster and Retry.
    The Execute method of job definition Microsoft.Office.Server.UserProfiles.LMTRepopulationJob (ID 414cbbe9-cdb1-4f7a-beed-85fbfd8a10c7) threw an exception. More information is included below.
    Unexpected exception in FeedCacheService.IsRepopulationNeeded: Connection to the server terminated,check if the cache host(s) is running .
    The Execute method of job definition Microsoft.SharePoint.Administration.SPProductVersionJobDefinition (ID 95aee52d-88a1-4355-b6b6-9d43d753414e) threw an exception. More information is included below.
    Collection was modified; enumeration operation may not execute.

    Hi,
    Please firstly go to Central Administration > Application Management > Manage Services on server, make sure the Distributed Cache service has been started on all servers.
    If you execute Get-CacheHost and Get-Cache in SharePoint 2013 Management Shell, is it return the expected information? If it returns the red error message, please refer to the article below to remove SPDistributedCacheServiceInstance and add SPDistributedCacheServiceInstance.:
    http://blogs.technet.com/b/saantil/archive/2013/03/31/distributed-cache-in-sharepoint-2013-quot-unexpected-exception-in-feedcacheservice-isrepopulationneeded-cache-cluster-is-down-restart-the-cache-cluster-and-retry-quot.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected]

  • Error while scheduling the Email Alert JOB chain

    Hi All,
    I have defined a job chain in CPS and when i am going to schedule it then it is giving me error message.We have taken the trial version.
    Please find the log attached below.
    11:18:31 PM:
    JCS-111004: Queue ETD.sapetd00_Queue has no ProcessServer with the required JobDefinitionType/Service/Resource for Job 932 (submitted from ETD.Z_MONI_BATCH_DP copy from 2009/12/30 18:22:23,113 Australia/Sydney) (submitted from Job Definition ETD.Z_MONI_BATCH_DP (Copy from 2009/12/30 18:22:23,113 Australia/Sydney)): Job Definition Type CSH/Service PlatformAgentService/"Empty"
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Job) to an object in an isolation group
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Chain Step) to an object in an isolation group
    JCS-102064: Job 934 (submitted from System_Mail_Send copy from 2009/12/29 17:54:16,608 Australia/Sydney) is global but refers (via Parent Job) to an object in an isolation group Show error details
    Thanks
    Rishi Abrol

    Hi
    Are you logged into the correct isolation group ?
    Ensure the process server is also assigned to the queue.
    Regards

  • Process Chain and CPS

    Hello all,
    We are currently having a problem with submitting process chains. When the chain runs it does not display any of steps in the process chain and where they are running from the job we submit (RSI_START_BW_CHAIN) So what we see is that the PC has been submited and completes in say 17seconds...however everything is still running on the system...for example our CIF PC runs for 3hrs but alls we are able to see is that the job executed from CPS with no problem. I've been digging around trying to find clear documenation about PC and CPS with little luck. Am I approaching the job submission incorrectly? We are on 7.0.3.

    Hello,
    You should be using the RSI_RUN_BW_CHAIN job to start process chains. Maybe you can try that first. For the rest things should be straight forward. Depending on the BW backend systems you might encounter some issues, with BW 7 the synchronization has changed and you would be better of using one of the latest 7.0.4 versions (SP6 has just been released).
    Regards Gerben

  • Process chain and Control M

    Hi Gurus
    We are designing process chain for our BW solution.
    We have identified the dependancies of various loads which includes flat file loads and loads from R/3.
    Now we would like to control ERP and BW jobs by Control M .
    I have the following questions:
    1. Which portion of BW process chain do we need to plug in to Control M. ? Do we need to connect the Meta chain to control M or different process chains to Control M?
    2.When we create Process chain for BW, what is the best practice? Is it good to create small chains and connect them via Meta chain? or we have to create small chains only?
    3. Can we include R/3 extraction Jobs and V3 Jobs in our BW process chain? so that they will trigger when we run the process chain?
    4. If we decide to run V 3 jobs after every half an hour on R3 side then how often we have to extract data to BW? and how to catch correct delta without missing single record?
    I would appreciate if you can help me with your knowledge.
    Thanks
    Kris

    Hi Kris,
    Control M is a third party scheduling tool provided by BMC software.
    1) You have option to either include a meta chain or individual chains in Control  M .
    2) It depends on how your data is extracted. If you have any dependecy jobs on sourcre system, then it is better to create small process chians and include them along with your dependecy jobs in Control M
    3)you can include R/3 dependency jobs and V3 jobs in Control M. Once those jobs are finished you can run BW process chains via Control M.
    4)You have several scheduling options availalbe in control M to take care of your BW and R/3 jobs.
    V3 jobs collect data from R/3 application tables and fill R/3 delta queue(RSA7). when you run BW process chain it will collect all the data available from R/3 delta queue.
    If both jobs are running at same time, then it is better to apply some wait time on BW jobs(in process chians).
    hope this helps

  • HOw to schedule ''landscape fetch' job in Solution Manager???

    HI,
    Can any one please help me in scheduling LAndscape fetch job in solution Manager?
    rEgards,
    fAisal

    Hi Venkatesh,
    Thanks for your reply.
    I checked by executing SLDCHECK. It opened the SLD page successfully. However, before that in the ABAP screen, it gave me error saying that:
    => Verify in the browser GUI that the SLD is in a healthy running state!
    Calling function LCR_LIST_BUSINESS_SYSTEMS
    Retrieving data from the SLD server...
    Function call returned exception code     3
    => Check whether the SLD is running!
    Summary: Connection to SLD does not work
    => Check SLD function and configurations
    Now checking access to the XI Profile
    Properties of RFC destination LCRSAPRFC
    The RFC LCRSAPRFC was of type 3 ABAP. It asked me to change it to T (TCP/IP). I did so.
    However, now I get this below error:
    Properties of RFC destination LCRSAPRFC
      RFC host:
      program id:
      gateway host:
      gateway service:
    Testing the RFC connection to the SLD java client...
    RFC ping returned exception with message:
    timeout during allocate / CPIC-CALL: 'ThSAPCMRCV' : cmRc=20 thRc=456#Timeout dur
    I even tested this RFC in SM59. It gave the same error.
    CAn you please help me?
    REgards,
    Faisal

Maybe you are looking for

  • How do I middle click and close tabs in safari? please help!

    I am used to google chrome and  adapted to use middle clicks to interact with tabs. My friend introduced me to safari and as i am using macbook pro I think using safari should be more compatible. The only thing that is holding me back is closing tabs

  • 500 Internal Server Error when including a JSP  withing a Report JSP

    Hello, Please help me. Thanks in advance. My setup is entirely 10g. (Database, Application Server - Infra, OC4J etc) When i am trying to include the content of a normal JSP within a Report JSP, i am getting an Internal Server Error. Am attaching the

  • Laserjet pro 400 mfp m425dn scan software

    I have a LaserJet Pro 400 MFP M425dn all-in-one device.  I have it connected via USB.  It works as a printer.  I cannot get any of the scan software to recognize that I even have a scanner. I have Windows 8.0 (junk). I have installed HP Scan and Capt

  • Cdemu - can't get it to work (can't connect do dbus, no cdemu module)

    Hi all! As the topic subject says, I can't get cdemu to work. I'm trying to use it to mount some .mdf images (wanna play Diablo 2 again, and I really don't wanna use the CDs for this). So here goes. First I tried mounting the .mdf image with cdemu, l

  • Trouble begining to use Reader 9

    I have just downloaded Adobe Reader 9 Version 9.1.2 and when I try to open a PDF I get an error messege "Before proceeding you must first launch Adobe Acrobat and accept the End User License Agreement". When I launch Adobe Acrobat I can't find the EU