OWB workflow question

I'm an OWB-newbie (using PARIS) and have a question about OWB workflows.
I want to create several mapping workflows.
All the mapping workflows should send a mail with the outcome of the workflows (success, error, warning) to an administrator.
I don't want to always code the same mail activity with the same parameter values.
So my idea is to write one subprocess which handles the mail-routine with some input parameters (calling workflow-name, process state (success, warning, error)).
In the subprocess I want to test the input parameter "process-state". If the value is 0 (success) then a mail with the workflow name as subject should be generated incuding an static text e.g. "mapping successful". If the value is 1 then "mapping warning" and so on.
What OWB activity has to be choosen for testing the input parameter "process-state" like an if-then-else construct or a case-construct?
How can the value of the input parameter "workflow-name" be included in the mail-body?
Regards
Martin

To my experience, using the OWF for these things is not very practical. I would suggest that you would make your own PL/SQL procedude that does this, and then call that procedure from within the actual mapping (post-mapping), using the values you have there.
This might not be what you want, however. I simply find the email capabilities of OWF too limited for actual work.

Similar Messages

  • Start OWB workflow from APEX

    Hi together,
    My challenge is to start an OWB workflow from an apex procedure.
    Procedure:
    declare
    numb number;
    audit_id number;
    begin
    numb := rtruser_blueprint.exec_api.execute('LOC_A014_OWF_MGR', 'PROCESS', 'LOAD_CALM_DELIVERY', ',', 'DELIVERY_ID_IN=6', audit_id);
    end;
    My problem is the following error:
    ORA-06565: cannot execute SET ROLE from within stored procedure
    This error occures at the following row in the execute-Procedure in the package exec_api (standard owb execution package):
    dbms_session.SET_ROLE('OWB_D_rtr10201_blueprint');
    dbms_session.SET_ROLE('OWB_O_rtr10201_blueprint');
    If I run this script directly with sqlplus, the procedure works fine. As a result of this, I assume, the problem has to do with the grants which the apex execution user (APEX_PUBLIC_USER) doesn't have in contrast to the blueprint user (which is schema owner)
    More informations about the environment setup:
    The apex workspace is bound to the database - user: BLUEPRINT.
    BLUEPRINT has the grant to execute the procedure 'EXECUTE' in the package 'EXEC_API'. The 'EXEC_API' package belongs to the user 'RTRUSER_BLUEPRINT'.
    My concrete question is: Which rights / grants does which user need to execute the procedure directly from APEX?
    Thanks for your help!
    Mirco

    What account were you using when you ran the 'execute_api' procedure from SqlPlus ? Also, all SQL and PL/SQL from APEX is executed as the application's parsing schema id and not the schema used by APEX to connect to the database.

  • Calling owb workflow job created in oem using pl/sql

    is it possible using pl/sql to call owb workflow job which is created thru oem
    can you pl be kind enough to share a sample.
    thanks

    Hi,
    The same question was just answered by another member a couple a days ago want to run owb mappings thru job schedule :
    sqlplus_exec_template runtime_owner location_name PROCESS process_name "," ","
    sqlplus_exec_template is in "D:\oracle\oraWB\owb\rtp\sql"
    Nikolai Rochnik

  • RED Workflow questions with Mac Pro (including third party plugins)

    Hello all,
    I’ve been searching many forums for the better part of a day trying to get some workflow questions sorted. I’m experiencing (very) slow export times, and mediocre playback for a machine that should be screaming fast.
    Here is what I’m working with:
    2014 Mac Pro
    -2.7 GHz 12-core intel xeon E5
    -64GB Ram
    -Dual AMD FirePro D700 6GB
    -1TB Flash Storage
    Editing all footage off 96TB Raid 6 mini-sas server (getting about 1100mbs read/write rate according to AJA system test) which is faster than any Thunderbolt/TB2 drive array I have.
    Media I work with is footage from the RED Epic (normally 5K) as well as DSLR footage from the 5d.
    Software:
    -PrPro CC 2014 (8.1)
    -Magic Bullet Looks 2.5.2
    My question(s) pertains to RED post-pro workflow in combination with third party plug-ins and the different approaches to make it more efficient.
    Right now, majority of the clients need a 1080p HD master, and they are generally anywhere from 2-8 minutes (usually). So my sequence settings are as follows:
    Video:
    Editing Mode: RED Cinema
    Size: 1920 x 1080
    Audio: 48Hz
    Video Previews
    Preview File Format: I-Frame Only MPEG
    Codec: MPEG I-Frame
    1920x1080
    Maximum Bit Depth unchecked
    Maximum Render Quality unchecked
    Composite in Linear Color checked
    Export Settings
    H.264
    1920x1080
    VBR 1 pass
    Target Bitrate 12mbs
    Max bitrate 12mbs
    Maximum render quality/depth/previews unchecked
    Issues I have:
    -Playback is fine at 1/2 or even full, but once effects (especially magic bullet looks) start to go on the clips, it’s very choppy and has difficult playback at 1/4
    -Export times (especially with magic bullet looks) will take the better part of 1-4 hours for a video that is 3-6 minutes long. This doesn’t seem like it should be the case for a maxed out MacPro
    So my questions are:
    Do these seem like the right sequence/export settings for mastering at 1080p? If not, what would you suggest?
    Would using offline editing help at all?
    Do you place your effects on adjustment layers?
    Is there anyway to improve export settings when using an array of filters?
    Have you stopped using third party plugins for their inefficiency in unreliability and switched to more integrated applications like SpeedGrade?
    Is there any other tweaks that you would suggest for RED workflow with PrPro?
    Should I consider switching to FCPX or (besides the iMovie-likeness) does it carry problems of its own?

    Hi This Is Ironclad,
    thisisironclad wrote:
    Hello all,
    I’ve been searching many forums for the better part of a day trying to get some workflow questions sorted. I’m experiencing (very) slow export times, and mediocre playback for a machine that should be screaming fast.
    The biggest issue is that most people have is that updating OS X causes certain folders to be set to Read Only. See this blog post: Premiere Pro CC, CC 2014, or 2014.1 freezing on startup or crashing while working (Mac OS X 10.9, and later).
    thisisironclad wrote:
    Hello all,
    I’ve been searching many forums for the better part of a day trying to get some workflow questions sorted. I’m experiencing (very) slow export times, and mediocre playback for a machine that should be screaming fast.
    Here is what I’m working with:
    2014 Mac Pro
    -2.7 GHz 12-core intel xeon E5
    -64GB Ram
    -Dual AMD FirePro D700 6GB
    -1TB Flash Storage
    It's a nice base system. How about an additional speedy disk for media cache files. You also did not mention which version of OS X you are running.
    thisisironclad wrote:
    Software:
    -Magic Bullet Looks 2.5.2
    The Red Giant website does not indicate that this software is yet updated to work with Premiere Pro CC 2014.1 (8.1). Proceed with caution here.
    thisisironclad wrote:
    Issues I have:
    -Playback is fine at 1/2 or even full, but once effects (especially magic bullet looks) start to go on the clips, it’s very choppy and has difficult playback at 1/4
    I would not use this plug-in until you get the OK from the manufacturer.
    thisisironclad wrote:
    -Export times (especially with magic bullet looks) will take the better part of 1-4 hours for a video that is 3-6 minutes long. This doesn’t seem like it should be the case for a maxed out MacPro
    Again, I suspect your plug-in.
    Keep in mind that exports are largely CPU based but you can make sure that GPU acceleration is enabled for AME at the bottom of the Queue panel.
    thisisironclad wrote:
    So my questions are:
    Do these seem like the right sequence/export settings for mastering at 1080p? If not, what would you suggest?
    It's OK.
    thisisironclad wrote:
    Would using offline editing help at all?
    No need when you should be able to edit natively. Relinking might also be an issue.
    thisisironclad wrote:
    Do you place your effects on adjustment layers?
    That's one way you can do it with the benefit of being more organized.
    thisisironclad wrote:
    Have you stopped using third party plugins for their inefficiency in unreliability and switched to more integrated applications like SpeedGrade?
    I do. Of course, that's a preference.
    thisisironclad wrote:
    Is there any other tweaks that you would suggest for RED workflow with PrPro?
    Try the following:
    Sign out from Creative Cloud, restart Premiere Pro, then sign in
    Update any GPU drivers
    Trash preferences
    Ensure Adobe preference files are set to read/write(Hopefully you checked this out already)
    Delete media cache
    Remove plug-ins
    If you have AMD GPUs, make sure CUDA is not installed
    Repair permissions
    Disconnect any third party hardware
    If you have a CUDA GPU, ensure that the Mercury Playback Engine is set to CUDA, not OpenCLYou have AMD GPUs.
    Disable App Nap
    Reboot
    thisisironclad wrote:
    Should I consider switching to FCPX or (besides the iMovie-likeness) does it carry problems of its own?
    I really shouldn't answer that question.
    Hope this helps.
    Thanks,
    Kevin

  • OWB Workflow Help

    Im conducting a proof of concept using OWB Workflow version 2.4. To give some background, the objective is to run numerous dependant SQL scripts that in effect hand craft the ETL process, at a pre-scheduled time, that provide logging and re-starting functionality.
    With the distinct lack of publications and documentations Id appreciate any feedback anyone may have on the following:
    1) Is it possible to restart a process at a step midway through the flow. For example the 3rd SQL Activity step out of 10 may have failed and instead of starting over, its a requirement to re-start the step that failed.
    2) I've read that Oracle are rumoured to be de-supporting the workflow functionality, if so then what is their proposed method of scheduling and controlling ETL jobs.
    3) Can a log file generated by the SQL Plus activity be attached to the failure or success email activity?
    Any input would be appreciated, also any pointers to OWB publications or documentation, other than the Oracle manuals of course.
    Thanks in advance
    James

    We have been working towards something similar, and the approach we have settled on is as follows:
    1) We have added our own runtime-audit tables which track the most recent process flow run date, final status, and the completion status of each mapping. So, a new run can see if a) this run is a restart, in which case we set our max date for drawing data over to be the run date of the last run, and b) what mappings ran succesfully. This way, a function called needs_to_run(mapping_name varchar2) returns a true/false answer to see if a given mapping needs to be executed on this run. The logic is pretty simple
    if run_type is full_run or (run_type is restart and nvl(last_run_status,'FAILURE') of this mapping is 'FAILURE')
    then
    return true;
    else
    return false;
    2) We do not use mapping objects in the process flows. We will only use SQL*Plus objects that do something like this:
    variable exec_return_code;
    declare
    retVal varchar2(30) := 'OK';
    begin
    if needs_to_run('ThisMapping') then
    ThisMapping.main(retVal);
    end if;
    :exec_return_code := retVal;
    exception
    when others then
    :exec_return_code := 'FAILURE';
    end;
    exit :exec_return_code
    So that each mapping will check first to see if it needs to execute before running thus skipping over fully completed mappings in the case of a restart. As to emailing log information, we currently do that in our post-mapping procedures in the event of a failure rather than logging to file and then attaching to email. We also have a last step that emails a summary of our audit info for each mapping run - but again this is all done inside pl/sql rather than spooling to file.
    Anyway, just some ideas to toss your way. And no, we haven't yet got it all completed and working yet, but it is coming along nicely and none of our individual proof-of-concept tests have provided us with any show-stoppers thus far.
    Cheers,
    Mike

  • Yet Another Workflow Question

    Ok I too, like many others here, am new to the Mac (thanks to Apple's I'm a Mac, I'm a PC ads that my wife couldn't get enough of). I have done some searching around and I see that there are quite a few iMovie workflow questions out there. I have not quite found what I am looking for however, so I thought I would make my first post tonight. So here it goes...
    I have 3 different ways I capture video:
    1. Canon Vixia HF10 (HD)
    2. Canon Powershot (SD)
    3. Blackberry Storm (SD...I know it isn't a good phone)
    I record everything to SD cards. I am wanting to know the best way to store my raw video for editing at any time. Do I copy the AVCHD file structure (for the Vixia) and .avi files (for the other non HD) to my hdd, or do I just import into iMovie '09 and let it reside there, or both? I noticed that iMovie had an archival option (which appears to just copy the AVCHD structure to my hdd), which is why I ask. I want to always keep my raw video in case I decide to go back later and create a new video.
    After I have the raw video archived, I would like to know the best way to use iMovie. Depending on where I end up storing the raw video, should I keep the imported video in iMovie once I am finished with a project, and then reimport it at a later date if need be? Or, do I leave it in iMovie as events? I guess this all rely depends on the first question...where do I store the raw video for archival purposes...
    Finally, when exporting my iMovie project, should I store that in more of a, pardon the Windows reference, "My Videos" folder with a original size, web optimized size, and ipod optimized size? Thus, keeping the actual exported version of the project separate from the raw video?
    I hope I have asked the right questions here. I appreciate any and all help I can get!
    Ron

    Welcome Ron to the  iMovie boards..
    very interesting : 'switchers' care sooo much for 'storage strategies' ..
    the by Apple intended workflow/concept for iApps is:
    any 'photocam' related material (still or movin') comes-in via iPhoto, and is stored in an iP Library (=you can tell iP to create 2/many Libs, if you prefer to organize manually....)
    any 'camcorder' related material HAS to be imported by iM - why? because, iM has some internal routines to make such material editable (codecs, thumnails, stuff....). the same material as 'file by Finder' does not import.. in most cases!
    storage..
    iP stores in its Library (local/internal HDD and/or ext. HDD)
    iM stores in Events (local/internal HDD and/or ext. HDD)
    to make Projects/Albums accessible to any iApp, you should keep your fingers off that structure.
    Erasing Events 'kills' projects.
    allthough, once 'shared to media browser' there's a 'copy' of your project WITHIN the project file. (= the socalled Media Browser is no single Folder somewhere hidden in the system)
    there's this Spacesaver feature to erase any Event content which is not in use in any project to keep Events lean.
    use the Archive feature from within iM to keep things easy and convenient.. if you miss a single file of the SDcard file-structure, the whole card's content is kaputt ..
    summary:
    • use iApps as intended.
    • use iP for cameras, it stores 'raws' (the avi too)
    • use iM for camcorders, use Archive to store raws..
    • purchase a dozend of HDDs to store your material..

  • Configure the Database for FTP in an OWB workflow

    I want to use the ftp-operator in an OWB-Workflow. I want to use a script, that is managed in the owb itself.
    I configure the ftp operator in this way:
    COMMAND: /usr/bin/ftp (i can start the command at the command line)
    PARAMETER_LIST: ?"${Task.Input}"?
    SUCCESS_THRESHOLD: 0
    SCRIPT: [my with the ftp Commands]
    After the execution of the workflow i get the error message: "/usr/bin/ftp -n: not found"
    in Runtime Audit Browser
    Does anybody know what i have to do to get the ftp operator running?
    The operating system of the database server is HP UX.
    Thanks in advance
    Alexander

    Hi,
    why do you want to use ftp? As against going directly to the source database? (So much faster and easier). I haven't used ftp since the last century.
    Regards,
    Donna

  • Sharpening export workflow question

    I have a sharpening workflow question. Say I have pictures from a portrait session I just finished. I have to send 10 pictures the client ordered to a print lab and I also will make some small facebook sized pictures and upload them to my business facebook page. The level of sharpening needed for large prints (I upload to print lab as RGB JPEGS) and sharpening needed for the very small sRGB facebook-sized pictures is different. In Lightroom I have the option to set the sharpening on export and have a bunch of presets that alter the export size, color space, sharpening, etc(WHCC print lab, facebook, Client CD, etc). I don't see how to do that in Aperture. I see they have the option if you have a printer, but not on normal export.
    For those of you that have to export batches of pictures in multiple different sizes (with different levels of sharpening), what is your workflow? I could use some photoshop droplets/actions after Aperture export but I was hoping there was a way to avoid the extra step. Am I overlooking an export feature? The BorderFX plug-in looks like the only other option.
    Thank you in advance for time and help!
    Scott

    Frank Scallo Jr wrote:
    The thing is guys - Once a file is sized down it WILL lose sharpening - what we are doing is sharpening the full size RAW file or rather what the full size output would be like. Once we export a version sized down it will lose some of the 'bite'. LR has sharpening options on 'output' which is not only smart but a necessity. Adobe realizes that output for screen needs another sharpen. Apple either doesn't know or didn't bother. It makes ANY output for screen less than best.
    Bear in mind that there seem to be two separate issues going on here - sharpening adjustments not being applied on export, and resizing.
    As far as resizing is concerned, Aperture appears to use something roughly equivalent to Photoshop's Bicubic Sharper setting. Because of this I've never had much problem with Aperture's exports when used for the web, but obviously everyone's taste for sharpening differs which is why an option for output sharpening would be good.
    Sharpening adjustments not being applied on export is a separate issue and should be reported via the feedback form ASAP by everyone who is experiencing the bug.
    Now printing is another animal - I wouldn't print directly from RAW in aperture either if I'm printing small. Again, LR beats Aperture here as well since they include output sharpening for print.
    Aperture has had output sharpening for printing since 2.0 came out (unless it in was 1.5). In A3 you need to turn on 'More Options' and scroll down, I can't remember where it is in A2. I don't know how effective it is as I print via a lab, but it's there and it's been there for a long time...
    Ian

  • Owb / Workflow install

    I am in the process of installing OWB and workflow (to be able to use/deploy process flows).
    My understanding is that I need to install Oracle HTTP server in a separate Oracle
    Home, and Oracle Workflow Server into the Oracle Home of the database I'm
    deploying to.
    I have also installed OWB into a separate Oracle Home .
    However my database Oracle Home is patched to 10.2.0.3 and I havent got a companion cd for 10.2.0.3 with Oracle Workflow Server - dont think it exists?
    Question is; can Oracle Workflow Server be installed in a Oracle Home that is different from the Oracle Home that I'm deploying to?
    Regards
    Ottar.

    Jayan,
    WF server is a part of (licensd with) Oracle DB, and located on Companion CD.
    For installing WF read more details for example in this thread
    Create Process Flow
    Regards,
    Oleg

  • OWB - Workflow- Cannot run program "/usr/bin/perl": CreateProcess error=2,

    Hi All,
    I am using OWB 11gR2 and I am getting "perl" location error after start workflow for any dimension or fact.
    *Cannot run program "/usr/bin/perl": CreateProcess error=2, The system cannot find the file specified
    The log is below. I am using Windows 2008 R2 OS and 11gR2 DB. The perl path is set in the environment variables and I can run the perl.exe from command line.
    Why OWB is looking for the Unix base path "/usr/bin/perl": is a question for me.
    If anybody knows the issue please guide me.
    Thanks
    OUBIWF_D_DEVICE_ACTIVITY_TYPE                                        2012-07-04 16:57:37.0     4
    Parameters                                             
    OUBIWF_D_DEVICE_ACTIVITY_TYPE:FILEMGR_PRE                              1          2012-07-04 16:57:38.0     2
    Parameters                                             
    RPE-02229: The following debug information was collected:                                             
    RPE-02230: Debug Information: os.name = Windows Server 2008 R2                                             
    RPE-02230: Debug Information: os.version = 6.1                                             
    RPE-02230: Debug Information: os.arch = amd64                                             
    RPE-02230: Debug Information: user.name = INT-A-MDM-01$                                             
    RPE-02230: Debug Information: user.dir = C:\product\11.2.0\dbhome_3\owb\bin\win32                                             
    RPE-02230: Debug Information: argv.arg0 = /usr/bin/perl                                             
    RPE-02230: Debug Information: argv.arg1 = /spl/intbase/DB/tools/filemgr/splfilemgr.plx                                             
    RPE-02230: Debug Information: argv.arg2 = -l                                             
    RPE-02230: Debug Information: argv.arg3 = /spl/BIDevelopment/bi2400db                                             
    RPE-02230: Debug Information: argv.arg4 = -n                                             
    RPE-02230: Debug Information: argv.arg5 = D_DEVICE_ACTIVITY_TYPE_EXT.DAT                                             
    RPE-02230: Debug Information: argv.arg6 = -p                                             
    RPE-02230: Debug Information: argv.arg7 = 1                                             
    RPE-02230: Debug Information: argv.arg8 = -s                                             
    RPE-02230: Debug Information: argv.arg9 = S                                             
    RPE-02230: Debug Information: argv.stdin =                                             
    RPE-02230: Debug Information: USERPROFILE = C:\Windows\system32\config\systemprofile                                             
    RPE-02230: Debug Information: JAVAPATH = C:\product\11.2.0\dbhome_3\jdk6                                             
    RPE-02230: Debug Information: NODEID = 1                                             
    RPE-02230: Debug Information: IAS_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: SystemDrive = C:                                             
    RPE-02230: Debug Information: RTHOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: Path = C:\product\11.2.0\dbhome_3\bin;C:\product\11.2.0\dbhome_3\jdk6\jre\bin\client;C:\product\11.2.0\dbhome_3\jdk6\jre\bin;C:\product\11.2.0\dbhome_3\owb\bin\admin;C:\Windows\system32                                             
    RPE-02230: Debug Information: CONNECT_SPEC = 10.34.32.136:1521:MDM                                             
    RPE-02230: Debug Information: PROCESSOR_REVISION = 1a05                                             
    RPE-02230: Debug Information: OCM_ORACLE_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: USERDOMAIN = WORKGROUP                                             
    RPE-02230: Debug Information: ALLUSERSPROFILE = C:\ProgramData                                             
    RPE-02230: Debug Information: TMP = C:\Windows\TEMP                                             
    RPE-02230: Debug Information: CommonProgramFiles = C:\Program Files\Common Files                                             
    RPE-02230: Debug Information: PROMPT = $P$G                                             
    RPE-02230: Debug Information: PROCESSOR_LEVEL = 6                                             
    RPE-02230: Debug Information: JVM_OPTIONS = -Xmx768M                                             
    RPE-02230: Debug Information: LOCALAPPDATA = C:\Windows\system32\config\systemprofile\AppData\Local                                             
    RPE-02230: Debug Information: COMPUTERNAME = INT-A-MDM-01                                             
    RPE-02230: Debug Information: SystemRoot = C:\Windows                                             
    RPE-02230: Debug Information: ORA_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: USERNAME = INT-A-MDM-01$                                             
    RPE-02230: Debug Information: APPDATA = C:\Windows\system32\config\systemprofile\AppData\Roaming                                             
    RPE-02230: Debug Information: ProgramData = C:\ProgramData                                             
    RPE-02230: Debug Information: PATHEXT = .COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC                                             
    RPE-02230: Debug Information: ORACLE_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: RTUSER = OWBSYS                                             
    RPE-02230: Debug Information: ProgramFiles(x86) = C:\Program Files (x86)                                             
    RPE-02230: Debug Information: TEMP = C:\Windows\TEMP                                             
    RPE-02230: Debug Information: OEM_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: ProgramFiles = C:\Program Files                                             
    RPE-02230: Debug Information: JDK_HOME = C:\product\11.2.0\dbhome_3\jdk6                                             
    RPE-02230: Debug Information: OCM_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: ProgramW6432 = C:\Program Files                                             
    RPE-02230: Debug Information: PROCESSOR_IDENTIFIER = Intel64 Family 6 Model 26 Stepping 5, GenuineIntel                                             
    RPE-02230: Debug Information: STARTUP_TYPE = -manual                                             
    RPE-02230: Debug Information: PROCESSOR_ARCHITECTURE = AMD64                                             
    RPE-02230: Debug Information: OS = Windows_NT                                             
    RPE-02230: Debug Information: FP_NO_HOST_CHECK = NO                                             
    RPE-02230: Debug Information: CommonProgramW6432 = C:\Program Files\Common Files                                             
    RPE-02230: Debug Information: ORACLE_SID = mdm                                             
    RPE-02230: Debug Information: DISCOVERER_ALLOW_DB_CONNECT_STRING = YES                                             
    RPE-02230: Debug Information: OWB_HOME = C:\product\11.2.0\dbhome_3                                             
    RPE-02230: Debug Information: NUMBER_OF_PROCESSORS = 2                                             
    RPE-02230: Debug Information: PUBLIC = C:\Users\Public                                             
    RPE-02230: Debug Information: PSModulePath = C:\Windows\system32\WindowsPowerShell\v1.0\Modules\                                             
    RPE-02230: Debug Information: CommonProgramFiles(x86) = C:\Program Files (x86)\Common Files                                             
    RPE-02230: Debug Information: ComSpec = C:\Windows\system32\cmd.exe                                             
    RPE-02235: The operating system has reported that it cannot execute the requested command. This may be because the command or the parameters are invalid, or that the command is shell command. Any result code return is specific to the operating system. Debug information has been produced.                                             
    Cannot run program "/usr/bin/perl": CreateProcess error=2, The system cannot find the file specified                                             
    OUBIWF_D_DEVICE_ACTIVITY_TYPE:EMAIL_FILEMGR                              1          2012-07-04 16:57:40.0Parameters                                             
    ORA-24247: network access denied by access control list (ACL)
    ORA-06512: at "SYS.UTL_TCP", line 17
    ORA-06512: at "SYS.UTL_TCP", line 267
    ORA-06512: at "SYS.UTL_SMTP", line 161
    ORA-06512: at "SYS.UTL_SMTP", line 197
    ORA-06512: at line 8

    Hi There
    Am experiencing the same issue. Did you manage to resolve it, is so can you advice what the resolution is. Thanks
    Regards,
    Kayode

  • OWB WORKFLOW

    Hi All,
    I have OWB 11g r1(with database 11g r1) installed on my pc,win xp
    I have created many mappings.
    Now I want to create a process flow.
    I searched in GOOGLE
    It seems that the workflow is not installed.
    1) Is workflow not installed by default with OWB11g as the 11g database ?
    2) the default user owf_mgr is NOT present in the database .
    3) I checked the ORACLE_HOME/owb folder ,but the wf folder doesn't exist......
    I also have OWB11gR2 installed in my pc.
    this does have the wf folder in oracle_home/owb folder.
    so i copied wf folder from this home to 11gR1 home .....when I install using the wfinstall.bat ....it points the 11gR2 folder..........
    what options do i Have ?
    please suggest.
    Thanks
    s

    In oracle 11g workflow is no longer provided with the database and is not on the 11g companion. It is included as part of warehouse builder 11g and located at %ORACLE_HOME%\owb\wf\ in the OWB installation oracle home.
    For oracle database 10g you would need to install workflow server component into the 10g database home via the 10g companion cd, not use the workflow included with OWB 11g. The two "should" be the same 2.6.4 final release version of workflow (from appearances of looking at the files they "seem" the same), but if you want to be running a supported configuration, use the 10g companion to install workflow into a 10g database and use the OWB 11g workflow to install on 11g database.
    Edited by: smithrh2 on Feb 14, 2012 1:53 PM
    Edited by: smithrh2 on Feb 14, 2012 1:54 PM

  • OWB / Workflow integration problem

    We have a OWB 11g installation running on Windows 2003 64-bit Server.
    We have 2 OWB databases, one design repository and one runtime repository.
    Workflow is installed and configured on the runtime database.
    Oracle Workflow Location is set to the runtime repository, and checks out ok.
    OWF_MGR user is also created in the OWB Design Center, although Oracle documentation is very unclear on this; only stating that
    +"6. Create a Workflow Proxy User.+
    +When the Workflow instance is remote from the database hosting the Warehouse+
    +Builder repository, you need to create a proxy-user.+
    +Within the database hosting the repository, use SQL Plus to create a user and grant+
    +it the OWB_USER role as a default. This enables the remote OWF instance to+
    +connect to the services provided by the Control Center."+
    We're able to deploy Process Flows from the design repository to the runtime rep, but when we try to execute them through workflow, the warehouse builder browser reports the following 3 errors:
    R+PE-02075: Oracle Workflow failed to process the execution request for Activity LOAD_FP_INDEX:START1. This may be because dependent objects have not yet been deployed.+
    ORA-20002: Failed to call "PFFPVSL_WB_RTI_WORKFLOW_UTIL".initialize. Please check that "EXECUTE ANY PROCEDURE" privilege is set and that the OWB Runtime is available. ORA-20009: User OWF_MGR does not have access to the workspace with name:OWBRTREP and owner:OWBRTREP. OWF_MGR must be registered to access the workspace by the administrator or owner of the workspace
    WB_RT_WORKFLOW_UTIL_11G.Execute_Start(PFFPVSL, WB_IK_20090213_135803_19093, 366, RUN) Wf_Engine_Util.Function_Call(WB_RT_WORKFLOW_UTIL_11G.EXECUTE_START, PFFPVSL, WB_IK_20090213_135803_19093, 366, RUN)
    We've done the EXECUTE ANY PROCEDURE on the OWF_MGR user. The OWF service_doctor.sql script reports that "There are errors in one or more PL/SQL packages and functions", but fails to say anything else - everything is loaded and seeded correctly.
    But what we notice when comparing with our working 10g installation, is that the runtime repository target user (OWBRTREP) is totally empty of tables, views, procedures etc on the new runtime database, whereas it has about 250 tables and lots of other objects on the 10g installation. This user is created through the OWB 11g design center, so why is there such a difference? Has something gone wrong when creating the runtime workspace?
    (It was created with the Repository Assistant, which we had some trouble with on the 64-bit server, but after some work it now seems to run ok, exiting with success message after creating workspace user.)
    Regards,
    -Haakon-

    Haakon,
    Lets assume you have two databases with OWB Repository A installed on database A and OWB repository B installed on database B.
    Repository A - your Design repository
    Repository B - your Runtime repository (this is your target schema / database)
    I hope you have control centre service running under Repository B and you have installed OWF schema (usually OWF_MGR) under repository B.
    I hope in your repository A you have created a new configuration and new control centre to point to repository B and that configuration is active.
    Now logon to repository B using repository owner login and add OWF_MGR and other target databases as your target user under SECURITY/USERS.
    Create similar users under repository A.
    It works for us in OWB 10g but i dont know about 11G might be something related to WORKSPACE concept.
    Hope this helps.
    Thanks,
    Sam.

  • OWB Workflow Queue Listener

    Hi ,
    We are trying to schedule the OWB 3i mappings using OEM 2.2 and Workflow 2.6.2.We have a queue listener 'workflowqlsnr.bat' on the client side,but we need to have the Workflow Queue Listenter on the Server side.Please let me know what is the procedure in order to deploy the listener.sh in the server(Unix).
    I have the following file installed in my owb3i\owb\bin\solaris directory on Windows 2000 -- workflowqlsnr.sh,workflowqlsnr1.sh.
    Environment:
    OWB 3i Client ( Windows 2000).
    Server OS: Sun Solaris
    Database : 8.1.7.3
    OEM 2.2
    OWF 2.6.2 Server/Client.
    Thanks for the help.
    Sharad.

    I am facing a problem in DB Link creation.
    Backend: Oracle 8i Server on my machine
    DW Software: Oracle warehouse builder 3i ( client , repository asistant.....)
    Operating system: Windows NT 4 SERVICE PACK 6
    I wants to use the scott database( default database given by oracle ) as my input source.
    How can I create the DB LINK ( for scott database) ?
    How can I create the DB LINK ( for any other database) ?
    Should I need to add anything in Setting of"ODBC DATASOURCE ADMINISTRATION"
    ==================
    Settings done:
    ==================
    DB Link Name :prashant
    Host name
    Host name: my machine's ip address
    port number: 1521
    oracle sid: prashant ( my oracle sid)
    user name:scott
    password:tiger
    ==================
    Gives error:
    ==================
    Testing...
    Failed.
    Cwm Error Message: Failed in the WBAPIFactory.createDBLink()
    Cwm Error: SQL Exception
    Class Name: oracle.wh.ui.integrator.common.RepositoryUtils
    Method Name: createDBLink(String, String, String, String)
    Method Name: -1
    Repository Error Message: java.sql.SQLException: ORA-02082: a loopback database link must have a connection qualifier
    ==================
    ==================
    How should I proceed further.
    I am expecting URGENT FEEDBACK.
    Reply me on : [email protected] or this OTN Network
    From
    Prashant

  • Workflow Questions - How to execute the steps in a workflow automatically?

    I am using MDM 5.5 SP6.
    My Question is how to execute the steps in the workflow automatically. That means NO user interaction.
    I have a simple workflow, Start- Process -- Syndicate -  Stop.
    When a new record added /updated/inserted this workflow will start automatically.  The workflow inbox, the workflow status is u2018Receivedu2019 Step is u2018Processu2019.  Now I have to right click on the record then u2018Next Stepu2019 then click on u2018Syndicateu2019 then workflow completes automatically.
    How to execute the Syndication step automatically?.  I donu2019t want to have any user interaction to syndicate these records.
    Is this possible?
    The reason for this is I want to do some field validation and based on the field validation some assignments then then I want  to syndicate these records without user interaction..

    Hi,
    If you want to do some validations and then based on result assignments on some of the fields before syndicating, then you have to use workflow. Only syndication server will not fulfill this requirement.
    you may try by making validation's automatic property as "Error". This way it will not allow to add any record not fulfilling validation. You can also try using calculated fields instead of assignment.
    I am not sure if you can make your workflow fully automatic, because atleast one process step is must for any workflow, and process step means user action.
    Regards,
    Shiv

  • Workflow Question - Recently moved from PC to Mac

    So, let me say that I'm a die hard Windows user -- in fact, in my "day job" I own a software development company and we're a "Microsoft shop". But, my side business is Photography and I'm making the switch to the Mac & Aperture. So, what I'm hoping to get from this post is a "yes, you can do that" or a "no, you can't do that" with Aperture. I'm happy to figure out how -- I just need someone to tell me if it's possible.
    Here's my current workflow (coming from the PC - which required several applications to accomplish these steps):
    1) Import images and rename "IMG_1234.CR2" to "1234.CR2" -- essentially dropping the "IMG_" from the name. Is this possible w/ Aperture?
    2) Do all of the processing on the images (No questions here yet -- got that pretty well figured out in Aperture).
    3) Create a black & white version of each image. Currently doing this by making a duplicate version in Aperture. But, I would like it to be named "1234-BW.CR2" rather than just have "version 2" tacked on the end. Is this possible in Aperture?
    4) "Develop" the pics into jpg's. Again, the final image names should be "1234.jpg" and "1234-BW.jpg" respectively. Again, the question has to do with renaming the original & duplicate version of the image.
    5) Choose all of the color versions of the photos and then selectively pick some of the b/w versions and export them to a "Web" size along w/ Thumbnail versions that will be posted on my web site. (Again, this should be easy if I can name things like I want them to be named in question 3 & 4.)
    So, right now I'm the biggest thing that I'm having a problem with in naming the images -- how flexible is Aperture with this? I've tried to customize the naming but can't see to figure out how to simply remove the "IMG_" when importing and then how to tack on "-BW" when creating a duplicate version.
    So far I really like Aperture -- it will ultimately save me from having to use 5 different applications on the PC and thus greatly improve the efficiency of my workflow. There are some things I've noticed that would be nice -- for example, on the PC I used a program called Bibble Pro to "develop" my Canon RAW files. Bibble is a little smoother (read faster) when making some adjustments. For example, when I'm in full screen mode and adjust the exposure, contrast or sharpening Aperture is not as smooth and "real-time" as Bibble. (And, the Mac I have is the top of the line iMac made today w/ 2GB of RAM. So, I would assume that processor speed is not an issue.)
    Anyway, if you have any insight into this post please let me know. Thanks in advance for any help you can provide... If you want to see my web site to understand what I'm doing the URL is http://www.level3photography.com/proofs.aspx
    Thanks again!
    TK Herman
    [email protected]
    iMac   Mac OS X (10.4.8)  

    ...the Mac I have is the
    top of the line iMac made today w/ 2GB of RAM. So, I
    would assume that processor speed is not an issue...
    Many folks from other disciplines think that applications should, well, just run on any computer. Not so with graphics apps. Photoshop, for instance, slows if not provided a second physical hard drive for scratch disk; Aperture craves GPU hardware; both gobble RAM; etc. Existing PSCS2 Photoshop performance improves up to 8 GB RAM, and we have not yet empirically determined the max amount of RAM Aperture will benefit from but it certainly is more than 2 GB.
    iMacs are consumer machines, not pro graphics boxes, even though many folks successfully use them as such. And, "top of the line iMac made today" would include not the 7300 GT graphics but the 7600 GT graphics card and 3 GB of RAM rather than 2 GB.
    All that said, your iMac should run Aperture well if you keep the box well maintained. Just don't be surprised if you do see some hardware imposed limitations.
    -Allen Wicks

Maybe you are looking for