Running Jobs in Parallel

Hi There,
I have two procedures A & B.
I want to run these two in parallel.
How can I achieve this ?
I want to run from Oracle.
Will DBMS_JOB work here?
Though two jobs are scheduled at the same time,
it will run one after the other based on the job_id
Thanks.

You can use DBMS_JOB to do that. Submit both procedures into job queue one after the other and when you COMMIT, they would be run at the next available oppourtunity.
it will run one after the other based on the job_idThat depends upon the parameter job_queue_processes configuration. What is this setting currently?

Similar Messages

  • Unix master shell script to run jobs in parallel with dependency

    I need a master shell script which would call the following shell scripts as per below order. Kindly note that if any shell script fails then the master shell script should abort.
    Level 1     PRODUCTS
         SP_ROST_DLY_STG_lD
         SP_ROST_DLY_ITG_lD
    After Level 1 completes run below two shell scripts in parallel:
    Level 2     SP_IDL_EDGE_CON_POSTN_STG_Ld ,     SP_IDL_EDGE_ACCT_POSTN_STG_LD
    After Level 2 completes run below shell scripts as per the seqeunce and in parallel:
    Level 3     SP_IDL_EDGE_ACCT_STG_LD ,     SP_IDL_EDGE_CONT_STG_LD
         SP_IDL_EDGE_ACCT_STG_VAL ,     SP_IDL_EDGE_CONT_STG_VAL
         SP_IDL_EDGE_ACCT_ITG_Ld ,     SP_IDL_EDGE_CON_ITG_Ld
    After Level 3 completes run below shell scripts as per the seqeunce and in parallel:
    Level 4     SP_IDL_EDGE_CONT_POSTN_ITG_LD ,     SP_IDL_EDGE_VISITS_LD
         SP_IDL_EDGE_ACCT_POSTN_ITG_LD ,     SP_IDL_EDGE_VISITS_DTL_LD

    can i avoid using set -e and use some other method so that if there is an error in previous job then the next job is not triggered.Sure, but it is ugly.
    The below is not the only way to do this...
    Wrap every place you start another script with:
    $ if ! MYSCRIPT; then
        echo Script MYSCRIPT failed. >&2
        exit 1
    fibut the wrapper for parallel scripts is even worse:
    $ MYSCRIPT &
    $ MYSCRIPT_JOBNO=$!
    $ MYSCRIPT2 &
    $ MYSCRIPT2_JOBNO=$!
    $ MYSCRIPT_STATUS=`wait ${MYSCRIPT_JOBNO}`
    $ if [ ${MYSCRIPT_STATUS} -ne 0 ]; then
        echo Script MYSCRIPT failed. >&2
        exit 1
    fi
    $ MYSCRIPT2_STATUS=`wait ${MYSCRIPT2_JOBNO}`
    $ if [ ${MYSCRIPT2_STATUS} -ne 0 ]; then
        echo Script MYSCRIPT2 failed. >&2
        exit 1
    fi
    ...Now, doesn't a simple 'set -e' look much better?

  • Release to accounting - Can we run jobs in parallel?

    Hi Guys
    While passing our Sales invoices from SD to Accounts receivables, our release to accounting job is taking very very long time. I was planning to split this job into multiple steps, each releasing accounting documents for a certain set of customers.
    In a normal scenario, the accounting documents are released simultaneously when the billing document in SD gets generated. However, at our client we follow a two step process, where the billing document gets generated but it is not passed to accounting immediately. We execute "Release to accounting" job to accomplish this. As this job is run open-ended , contrary to the billing job (which is split into multiple steps by RV60SBAT), it takes very long time to finish.
    Our FI Folks have raised concerns on splitting the release to accounting job into multiple steps, as it might cause problems on allocation of document numbers by the program. I don't see any issues as the normal scenario (simultaneously creating billing/accounting) does the same thing.
    Please let me know if you see any issues with this.
    Thanks.

    Hi,
    You can run parallel job that should not affect the performance of the system but you would have lot of document numbers which will not be used as system would buffer them. There is a option to remove the buffer but then the performance could get affected.
    Regards
    K.R

  • We are running 3 batches parallel(a.ksh,b.ksh,c.ksh parallel ) .if a.ksh will complete then d.ksh will start('d' start when 'a' will terminate successfully) and we have to handle error for all jobs ( if some job got aborted during runtime) ?

    we are running 3 batches parallel(a.ksh,b.ksh,c.ksh parallel ) .if a.ksh will complete then d.ksh will start('d' start when 'a' will terminate successfully) and we have to handle error for all jobs ( if some job got aborted during runtime) ?

    Moderator Action:
    You already asked this question, two days earlier.
    https://forums.oracle.com/thread/2585158
    Stay with your original post.   Deliberate multiple posting is the same as spamming the forums.
    This new post is locked.

  • Running SETUP jobs in parallel in Production R/3

    Hi All,
    I have identified the number ranges(for Billing) for which I have to run SETUP.
    Example,I have 5 parameters/Variables defined in 'oli9bw'....1 variable for 1 range in 'oli9bw' and say all are new runs.
    How can I run these variables/parameters in Parallel inorder to reduce the downtime of the Production system.??
    Do I have to have a ABAP code written to trigger these Variables (all at once)which inturn fills the setup tables.?
    or Is there a standard ABAP program where I can mention these Variables and then Kickoff the program?
    Note:I already took methods to have enough processes to run these 5 jobs in parallel.
    Any help is much appreciated.
    Thanks,
    Jb

    Abhijit,
    I have an alternate approach than running SETUP's for history data .since I am afraid that running SETUP's for 80 jobs may take long time in PRODUCTION and also to avoid any complications.
    Please let me know if this works
    For future data:
    1.clear qrfc and RSA7 by running delta twice for Billing DSO,so that Delta queue for 2LIS_13_VDITM is   empty before transporting new structure.
    2.send the new structure of 2LIS_13_VDITM to R/3 Production while R/3 Prod is down.
    3.Will check the delta queue for this data source and see if this DS is active with new structure.
    4.Turn on the Job control so that users can post new documents and then turn back on BW Process chain.
    In this way,we dont run the SETUP's for history data,but future forward we have new structure data coming in BW.
    In order to capture History:
    1.I thought of building generic DS based on a view on VBRK and VBRP with selction on Bill_date = 03/2008 to till date.
    2.I suppose view can handle 80 k billing documents
    3.then dump the data in to a new DSO (keys in DSO are Bill_doc and Bill_item ,data fields are just say 'xx' and 'yy.....xx and yy are the new fields that has been added to extract structure)')
    4. then load as full load to Billing ODS .(Bill_doc and Bill_item are the keys in Billing DSO )
    5.In this way ,all the data from Generic DS will over write the Billing DSO with new fields 'xx' and 'yy'
    By the way,can we have repair full from DSO to DSO ??So that the deltas won't get disturbed in Billing DSO.
    Please let me know if this works
    Thanks,
    Jb

  • Is it Possible to Run Reporting Agent job in Parallel?

    Hi Folks,
              I am sceduling a reporting agetn job and I see that it only takes one back Ground process and its a long running Job, so I was thinking about the possibilty of breaking this into parallel jobs..Is this possible if so please share with me how to do this...thanks in advance

    Farhan,
           I think I did not explain my question properly, I do not want to create 2 variants with different time selections and run it as 2 jobs, this way I cannot have a Super Set for using Cache, what I am want is use just one variant (Jan through Dec) and when I execute this in the background, i want this job to split into parallel processes so that it runs faster...Can someone explain if this is possible. if so how can i achieve this..thanks in Advance..

  • Running SQL agent jobs in parallel

    I have around 60 SQL agent jobs (each job has 1 step that loads data from ORACLE to SQL using attunity) that run SSIS packages that I need to run in parallel.  I did some testing and found that the server can handle 10 of them running at the same time.
     I know I need to loop through sysjobhistory and find the ones (all 60 jobs start with PS) that haven't ran today, currently not running and the first tow words of the name = 'PS'.  Any advice on how to get this accomplished would be appreciated.
    I have seen some articles that mention adding 10 data flows to a SSIS package and run them that way but I would like for SQL to handle it, so I can get the time.
    Thanks!

    That will definitely help identify the jobs but I need to figure the looping logic that will allow 10 jobs to run in parallel.
    try this.. not tested.. left comments where possible for you to understand..may need to tweak it a bit..
    declare @a intset @a=0
    declare @jobname nvarchar(200)
    -- checks if there are already 10 jobs running
    while (@a <10) and ((Select count(C.Name) as [JobName]
    from (Select max(Session_id) as Session_Id from msdb.dbo.syssessions) A
    INNER JOIN msdb.dbo.sysjobactivity B on A.Session_id=B.Session_ID
    INNER JOIN msdb.dbo.sysjobs C on B.job_id=C.Job_ID
    where B.stop_execution_date is null AND B.run_requested_date is not null and C.Name like 'PS%') < 10)
    begin
    --loops through to fetch one non-running job at a time and fetches upto 10 jobs
    Select top 1 @jobName = name
    from msdb.dbo.sysjobs X
    INNER JOIN msdb.dbo.sysjobactivity Z on Z.job_id=X.Job_ID
    where name like 'PS%' and
    --checks the job is currently not running
    name not in (
    Select C.Name as [JobName]
    from (Select max(Session_id) as Session_Id from msdb.dbo.syssessions) A
    INNER JOIN msdb.dbo.sysjobactivity B on A.Session_id=B.Session_ID
    INNER JOIN msdb.dbo.sysjobs C on B.job_id=C.Job_ID
    where B.stop_execution_date is null AND B.run_requested_date is not null)
    --makes sure the job already did not run today
    and cast(Z.run_requested_date as date)<>cast(getdate() as date)
    order by name
    Execute msdb.dbo.sp_start_job @job_Name=@jobName
    set @a=@a+1
    End
    Hope it Helps!!

  • Can not see running jobs from, package

    I have a problem with processing parallely runing jobs:
    I am creating another immediately runned jobs in a main job. Those two parallel jobs (2 loads from different databases) have to be finished before I run next operation (working out loaded data). Problem is that, after starting those 2 parallel jobs, I can not see them by select from ALL_SCHEDULER_RUNNING_JOBS that is executed immediately after I create those jobs in a package. If I take a look into ALL_SCHEDULER_RUNNING_JOBS from anonymous statement I can see all my running jobs. Let's sum it up:
    1/ Start of main job
    2/ Running 2 immediately created jobs (load data)
    3/ Checking in loop if jobs created in step 2 are still running
    3.1/ Jobs are running (ALL_SCHEDULER_RUNNING_JOBS check) - sleep for a
    while - It never happens - can't see any running jobs from select executed in
    package but can see them in
    anonymous statement
    3.2/ Jobs finished - start processing loaded data
    Can somebody help me with this task?
    Thanks alot!
    Jakub

    Hi,
    There is no reason a job should be visible from an anonymous block but not from inside a job. There are two things that may be happening here.
    - jobs scheduled to run immediately my not start running as soon as they are created/enabled, you may need to wait a bit before they start running (they will appear in all_scheduler_jobs immediately but maybe not all_scheduler_running_jobs immediately)
    - you may be running into privilege issues. Is the user that executes the anonymous block the same as the user that the job is running as (the job's schema) ? If not maybe the job user does not have privileges to see the job (you can grant alter on the job to the user to ensure this).
    Can you see the jobs in the all_scheduler_jobs view from within the job with status RUNNING ? If you can see jobs in all_scheduler_jobs as RUNNING but not in all_scheduler_running_jobs then this is a bug of some sort.
    Thanks,
    Ravi.

  • Run Jobs One After Another

    I know this is possible, I'm not sure how to go about it using Oracle's Scheduler. I simply want to create a special "queue" that I can submit jobs to which will run one after another. After they run, then then are gone. There many be no jobs at anyone time, and then at other times I may submit 10 jobs to the queue. Currently when I submit jobs they all run parrallel, this is not what I want. I guess I don't understand why Oracle calls it a Job Queue when all my jobs run in parrallel and not in a queue as the name implies.
    This is not say that I don't have other jobs which I would not want to send to this "queue". In other cases I would just want to submit the job and let it run right away and not wait on anything.
    I'm currently using the DBMS_JOB package, which I'm pretty sure won't let me do what I want.
    Any help would be greatly appreciated.
    Thanks,
    Joe

    Here is and example that creates Resource Manager objects and
    two job classes that will permit 1 and 3 jobs to
    run in parallel. Note this will work in 11R1 and 10.2.4
    Assumes user user scott/tiger exists.
    --  Remove existing scheduler objects, jobs and classes
    CONNECT /  as sysdba
    BEGIN
       DBMS_SCHEDULER.DROP_JOB_CLASS(
          JOB_CLASS_NAME => 'JOBQ_WIDTH_1_CLASS'
    END;
    BEGIN
       DBMS_SCHEDULER.DROP_JOB_CLASS(
          JOB_CLASS_NAME => 'JOBQ_WIDTH_3_CLASS'
    END;
    BEGIN
      FOR I IN 1..30 LOOP
      BEGIN
        SYS.DBMS_SCHEDULER.DROP_JOB(JOB_NAME => 'SCOTT.SERIAL_' || I, FORCE=>TRUE);
      EXCEPTION WHEN OTHERS THEN NULL;
      END;
    END LOOP;
    END;
    BEGIN
      FOR I IN 1..90 LOOP
      BEGIN
        SYS.DBMS_SCHEDULER.DROP_JOB(JOB_NAME => 'SCOTT.PAR3_' || I, FORCE=>TRUE);
      EXCEPTION WHEN OTHERS THEN NULL;
      END;
    END LOOP;
    END;
    ALTER SYSTEM SET RESOURCE_MANAGER_PLAN='';
    set echo on
    --  Remove previously defined  resource plan, resource plan
    --  directives, and resource consumer groups  for this test
    BEGIN
        -- Prepare the pending area
        DBMS_RESOURCE_MANAGER.CLEAR_PENDING_AREA;
        DBMS_RESOURCE_MANAGER.CREATE_PENDING_AREA;
        -- Delete resource plan, its resource plan directives, and
        -- any associated resource groups
        DBMS_RESOURCE_MANAGER.DELETE_PLAN_CASCADE(
           plan => 'JOBQS_PLAN'
        -- Submit the changes
        DBMS_RESOURCE_MANAGER.VALIDATE_PENDING_AREA;
        DBMS_RESOURCE_MANAGER.SUBMIT_PENDING_AREA;
    END;
    -- Create resource manager objects
    BEGIN
       DBMS_RESOURCE_MANAGER.CLEAR_PENDING_AREA();
       DBMS_RESOURCE_MANAGER.CREATE_PENDING_AREA();
       DBMS_RESOURCE_MANAGER.CREATE_CONSUMER_GROUP (
            CONSUMER_GROUP => 'JOBQ_WIDTH_1_GROUP',  
            COMMENT => 'Consumer group to force jobs to execute in serial'
       DBMS_RESOURCE_MANAGER.CREATE_CONSUMER_GROUP (
            CONSUMER_GROUP => 'JOBQ_WIDTH_3_GROUP',  
            COMMENT => 'Consumer group to allow jobs to run in parallel 3'
       DBMS_RESOURCE_MANAGER.SUBMIT_PENDING_AREA();
    END;
    BEGIN
       DBMS_RESOURCE_MANAGER.CLEAR_PENDING_AREA();
       DBMS_RESOURCE_MANAGER.CREATE_PENDING_AREA();
       DBMS_RESOURCE_MANAGER.CREATE_PLAN
          ('JOBQS_PLAN', 'PLAN TO SUPPORT JOBQS');
       DBMS_RESOURCE_MANAGER.CREATE_PLAN_DIRECTIVE(
          PLAN => 'JOBQS_PLAN',
          GROUP_OR_SUBPLAN => 'JOBQ_WIDTH_1_GROUP',
          COMMENT=>'Associates with job class JOBQ_WIDTH_1_CLASS',
          CPU_P1 => 25,
          PARALLEL_DEGREE_LIMIT_P1 => 1,
          ACTIVE_SESS_POOL_P1 =>1
       DBMS_RESOURCE_MANAGER.CREATE_PLAN_DIRECTIVE(
          PLAN => 'JOBQS_PLAN',
          GROUP_OR_SUBPLAN => 'JOBQ_WIDTH_3_GROUP',
          COMMENT=>'Associates with job class JOBQ_WIDTH_3_CLASS',
          CPU_P1 => 75,
          PARALLEL_DEGREE_LIMIT_P1 => 2,
          ACTIVE_SESS_POOL_P1 =>3
       DBMS_RESOURCE_MANAGER.CREATE_PLAN_DIRECTIVE(
          PLAN => 'JOBQS_PLAN',
          GROUP_OR_SUBPLAN => 'OTHER_GROUPS',
          CPU_P1 => 0,
          CPU_P2 => 100,
          COMMENT=>'Required'
       DBMS_RESOURCE_MANAGER.VALIDATE_PENDING_AREA();
       DBMS_RESOURCE_MANAGER.SUBMIT_PENDING_AREA();
    END;
    -- Create scheduler classes
    BEGIN
       DBMS_SCHEDULER.CREATE_JOB_CLASS(
          JOB_CLASS_NAME => 'JOBQ_WIDTH_1_CLASS',
          LOGGING_LEVEL => DBMS_SCHEDULER.LOGGING_FULL,
          LOG_HISTORY => 5,
          RESOURCE_CONSUMER_GROUP => 'JOBQ_WIDTH_1_GROUP',
          COMMENTS => 'Force jobs to run in serial, through plan JOBQ_PLAN and group JOBQ_WIDTH_1_GROUP'
    END;
    BEGIN
       DBMS_SCHEDULER.CREATE_JOB_CLASS(
          JOB_CLASS_NAME => 'JOBQ_WIDTH_3_CLASS',
          LOGGING_LEVEL => DBMS_SCHEDULER.LOGGING_FULL,
          LOG_HISTORY => 5,
          RESOURCE_CONSUMER_GROUP => 'JOBQ_WIDTH_3_GROUP',
          COMMENTS => 'Allow up to 3 jobs in parallel, enforced by plan JOBQ_PLAN and group JOBQ_WIDTH_3_GROUP'
    END;
    connect / as sysdba
    grant execute on jobq_width_1_class to scott;
    grant execute on jobq_width_3_class to scott;
    begin
      dbms_resource_manager_privs.grant_switch_consumer_group('scott','JOBQ_WIDTH_1_GROUP',false);
      dbms_resource_manager_privs.grant_switch_consumer_group('scott','JOBQ_WIDTH_3_GROUP',false);
    end;
    alter system set resource_manager_plan='';
    connect scott/tiger
    begin
      for i in 1..30 loop
      sys.dbms_scheduler.create_job(job_name => 'scott.serial_' || i,
          job_type=>'plsql_block', job_action=>'dbms_lock.sleep(2);',
          job_class=>'jobq_width_1_class',
          enabled=>true);
      end loop;
    end;
    begin
      for i in 1..90 loop
      sys.dbms_scheduler.create_job(job_name => 'scott.par3_' || i,
          job_type=>'plsql_block', job_action=>'dbms_lock.sleep(2);',
          job_class=>'jobq_width_3_class',
          enabled=>true);
      end loop;
    end;
    /-- The par3_90 job should finish about the same time as the serial_30 job since
    -- we run 3 jobs in paralled and serial respectively.
    -- check user_scheduler_job_run_details for after the test completed to verify.

  • Running Jobs in Parallely

    Hi All,
    We have to run 300 Jobs in parallel(All jobs start same time). Please advise the best approach to do this. Do I have to use multiple instances in Sql server ? to run all jobs at same time.
    Thanks in Advance.....

    Hi BHVS,
    Once a job is assigned to a thread, there is a limit on the number of threads used per agent subsystem. A Knowledge Base article for SQL Server 2000 that describes the thread count per subsystem in the registry can be found at
    support.microsoft.com/kb/306457.
    In SQL Server 2005 and later, these values are stored in the syssubsystem table in the msdb database. SQL Server Agent queries these values by executing the sp_enum_sqlagent_subsystems stored procedure. Once a job is assigned to a thread, the subsystem being
    used by each job step is important because of the maximum number of threads allowed per subsystem. If the maximum number of threads per subsystem has been reached, the thread running the current job will pause for five seconds, then try again. This behavior
    should be the same for both SQL Server 2000 and later versions.
    To further improve the situation, we could stagger the start times of the jobs by a minute or so. We could set things up such that 50 jobs would start at minute zero, 50 more at minute one, 50 more at minute two, and so on. Then you would only be attempting
    to start 50 jobs every minute instead of 300 jobs simultaneously every six minutes.
    References:
    https://technet.microsoft.com/en-us/magazine/2006.10.sqlqa.aspx
    http://blogs.msdn.com/b/karthick_pk/archive/2011/02/01/sql-agent-maxworkerthreads-and-agent-subsystem.aspx
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • How to stop a running job in 10g Scheduler?

    The following is a duplicate post. I posted the following to the general database forum before seeing that otn has a new scheduler forum:
    I am not able to find in the Admin Guide a method to stop a currently running instance of a job in the 10g scheduler.
    In 9i, I run the following script calling DBMS_JOB.broken and DBMS_JOB.remove to shut down currently running jobs:
    DECLARE
    jobid NUMBER;
    CURSOR c1
    IS
    SELECT job
    FROM dba_jobs
    WHERE priv_user = 'ME';
    BEGIN
    OPEN c1;
    LOOP
    FETCH c1
    INTO jobid;
    EXIT WHEN c1%NOTFOUND;
    DBMS_JOB.broken (jobid, TRUE);
    COMMIT;
    DBMS_JOB.remove (jobid);
    COMMIT;
    END LOOP;
    CLOSE c1;
    END;
    How may I create similar code to shut down currently running jobs using DBMS_SCHEDULER in 10g? According to the Admin Guide, disabling jobs with the force option will still allow the job to finish.
    How can I terminate a running job in 10g?

    You can stop a currently running job using the STOP_JOB api.
    STOP_JOB Procedure
    This procedure stops currently running jobs or all jobs in a job class. Any instance of the job will be stopped. After stopping the job, the state of a one-time job will be set to SUCCEEDED whereas the state of a repeating job will be set to SCHEDULED or COMPLETED depending on whether the next run of the job is scheduled.
    Syntax
    DBMS_SCHEDULER.STOP_JOB (
    job_name IN VARCHAR2
    force IN BOOLEAN DEFAULT FALSE);
    Parameters
    Table 83-44 STOP_JOB Procedure Parameters
    Parameter Description
    job_name
    The name of the job or job class. Can be a comma-delimited list. For a job class, the SYS schema should be specified.
    If the name of a job class is specified, the jobs that belong to that job class are stopped. The job class is not affected by this call.
    force
    If force is set to FALSE, the Scheduler tries to gracefully stop the job using an interrupt mechanism. This method gives control back to the slave process, which can update the status of the job in the job queue to stopped. If this fails, an error is returned.
    If force is set to TRUE, the Scheduler will immediately terminate the job slave. Oracle recommends that STOP_JOB with force set to TRUE be used only after a STOP_JOB with force set to FALSE has failed.
    Use of the force option requires the MANAGE SCHEDULER system privilege.
    Setting force to TRUE is not supported for jobs of type executable.
    Usage Notes
    STOP_JOB without the force option requires that you be the owner of the job or have ALTER privileges on that job. You can also stop a job if you have the CREATE ANY JOB or MANAGE SCHEDULER privilege.
    STOP_JOB with the force option requires that have the MANAGE SCHEDULER privilege.

  • A problem about the "Max running jobs". Thanks a million!

    Hello everyone,
    The "Max running jobs" was set to 15.
    (1). When I dragged two assets into FCSvr at a time, in the "Search all jobs" window, I saw the first one being uploaded then transcoded, and later the second one being uploaded then transcoded.
    (2). When I dragged them into FCSvr one by one, I saw them being uploaded, then the first one being transcoded, and the second one later.
    Thus with (2), there is no need to wait or queue between the uploading of the second one and the transcoding of the first one.
    However, I intend to drag the assets at the same time, besides, I wanna attain the aim in (2) that no queue to wait. Could someone tell me how to do it?
    With best regards,
    Steven Lee
    PINZ Media

    If I understand, you used for upload menu "Upload file..."
    For uploading and conversation in same time many files you can setup production action (response) for some folder on your device (XSAN, RAID, HDD)
    This action can create production for folder, and create assets for media in folder. The add action "SCAN" in response submenu. For transcoding many assets in same time check radio button "Background Analyze" in response "SCAN".
    You can read more about that in manual.

  • Adobe Acrobat XI pro version, Windows 7, running on iMac parallels, converting pdf to a pdf with reduced size is not possible, error: error in converting the file! What to do? Its a bit annoying not to be able to store pdf files in reduced size, any idea?

    Adobe Acrobat XI pro version, Windows 7, running on iMac parallels, converting pdf to a pdf with reduced size is not possible, error: error in converting the file! What to do? Its a bit annoying not to be able to store pdf files in reduced size, any idea?? Thanks, Jörg

    Hi Jorg ,
    Are you trying to reduce the file size with the "Reduced size PDF" in the save as other option.
    Give it a try if you haven't done it prior.
    Open that PDF>File>Save as Other>Reduced size PDF.
    If possible ,please share the snapshot of the error message with us so that we can have a look in order to assist you further.
    Regards
    Sukrit Dhingra

  • Is InfoPackage Group is the only way to run the InfoPackages Parallel???

    Hello BW Experts,
    If we want to divide a huge full upload InfoPackage with certain selection ranges in each InfoPackage, do we have to run them one by one according to our selcetions or can we run them parallel to each other??
    If we run these IPs parallel, is InfoPackage group is the only way or is there any other option and if yes, which one is better??
    plz help..!!
    Thanks & regards,
    Sapster.
    (assure points)

    Hi,
    Please do not assure us of your points...points are integral to SDN and everyone knows about them.
    In your previous posts, you have been advised that InfoPackage groups are obsolete and that you should proceed with process chains. Have you read those replies?
    Step-by-Step procedure to create an InfoPackage Group & a Scenario!!
    InfoPackage is one option. The better, more widely used and recommended option is using Process Chains.
    Hope this helps...

  • Re:How to determine the long running jobs in a patch

    Hi ,
    How to determine the long running jobs in a patch .
    Regards

    Hi,
    Check the below MY ORACLE SUPPORT note:
    Note.252422.1 .... Check Completed Long Running Jobs In Oracle Apps.
    Best regards,
    Rafi

Maybe you are looking for

  • HT4436 one apple ID for 2 people?

    My husband and i have a mac and 2 ipod shuffles. We both use one Apple ID to use iTunes to purchase music since we both like the same music. We want to get iPhones, but can we both access that iTunes acct on the phones while having separate AppleIDs

  • How to run a java program in command prompt

    hi i want to run a java program from in command prompt from another directory i want to run a file named First in C:\Program Files\Java\jdk1.6.0_07\bin so when i give the command java "C:\Program Files\Java\jdk1.6.0_07\bin\First" it doesnt works it s

  • High Pass Dialog Textbook Exercise

    I purchased the Power, Speed and Automatitation textbook to learn Photoshop Scripting. I am working through the exercises in the last chapeter Building Custom Dialogs and have stumbled on the last exercise building a High Pass Dialog. When I test the

  • Could not find version number data in TF_BUILD_BUILDNUMBER

    Hi, I'm using the default build process template TfvcTemplate.12.xaml that came with TFS2013. I'm trying to call ApplyVersionToAssemblies.ps1 (from the Community TFS Build Extensions) in the pre-build script to version assemblies. In lieu, I got this

  • How do i format the memory card on n78

    how do i format the memory card on my n78? do we another application to do it? can i do it with my nokia? the music player keeps on searching for files non stop and i wanna know if formatting the memory card solves the problem Solved! Go to Solution.