Workflow and server HTTP

To run process flow from OWB I have to install Oracle Workflow. Can I install component Workflow without HTTP server???

Standalone workflow is a desupported product - what are your reasons for installing this ?
HTH
Srini

Similar Messages

  • Down Time for  Installing APEX 4.2.1 and oracle HTTP server?

    Hi, all:
    I plan to install APEX 4.2.1 and oracle HTTP server on 10.2.0.4 database.
    Is there any down time required during installation?
    I mean during installation, do I need to block users from accessing the database?
    Can they do their normal data entry / inquiry?
    thanks,
    John
    ps. There will be server backup before installation.

    Hi,
    Documentation says
    >
    3.1 Recommended Pre-installation Tasks
    Before installing Oracle Application Express, Oracle recommends that you complete the following steps:
    Review and satisfy all Oracle Application Express installation requirements. See "Oracle Application Express Installation Requirements".
    Shut down with normal or immediate priority the Oracle Database instances where you plan to install Oracle Application Express. On Oracle Real Application Clusters (Oracle RAC) systems, shut down all instances on each node.
    An alternative to shutting down the database is to prevent all users from accessing Oracle Application Express when upgrading your installation from a previous release of Oracle Application Express. Oracle only recommends this option in high availability production environments where planned outages are not available. For all other scenarios, the database should be shut down.
    To disable access to Oracle Application Express when the existing installation is using the Application Express Listener, shut down the appropriate application server where the Application Express listener is deployed.
    To disable user access to Oracle Application Express when the existing installation is using the Oracle HTTP Server with mod_plsql, you should either shut down the Web server or disable the Application Express Database Access Descriptor of the Web server.
    For an existing installation using the embedded PL/SQL gateway, you should disable the Oracle XML DB Protocol Server by setting the HTTP port to 0. This can be accomplished by starting SQL*Plus, connecting as SYS to the database where Oracle Application Express is installed, and running:
    EXEC DBMS_XDB.SETHTTPPORT(0);
    Once you have prevented access from Oracle Application Express users, you should log into SQL*Plus as SYS, connecting to the database where Oracle Application Express is installed, and query V$SESSION to ensure there are no long running sessions which would interfere with the upgrade process.
    Back up the Oracle Database installation.
    Oracle recommends that you create a backup of the current Oracle Database installation before you install Oracle Application Express. You can use Oracle Database Recovery Manager, which is included in the Oracle Database installation, to perform the backup.
    See Also:
    Oracle Database Backup and Recovery User's Guide
    Start the Oracle Database instance that contains the target database.
    After backing up the system, you must start the Oracle instance that contains the target Oracle database. Do not start other processes such as a Web Listener. However, if you are performing a remote installation, make sure the database listener for the remote database has started.
    >
    Regards,
    Jari
    My Blog: http://dbswh.webhop.net/htmldb/f?p=BLOG:HOME:0
    Twitter: http://www.twitter.com/jariolai

  • ISQLPlus and Oracle HTTP Server

    After I reinstalled Oracle 9i on Windows XP, I now get a a FastCGI server error whenever I attempt to launch Oracle HTTP Server: "Redefinition of previously defined FastCGI Server". How do I correct this? I uninstalled the database prior to reinstalling. Question 2: I also installed Oracle 9i on Windows 98 and Oracle HTTP Server appears NOT to be part of the install so I don't know how to make iSQLPlus work. Any ideas on either of these? I would prefer to work on Windows XP.

    In answer to your first question, it may be answered in this SQL*Plus FAQ entry:
    http://otn.oracle.com/support/tech/sql_plus/htdocs/runtime.html#isql_duplicate_entries
    In answer to your second question, I don't think iSQL*Plus is part of the Windows 98 install. I don't think you even have a database, just the Oracle Client.
    Alison

  • Sharepoint foundation and server 2013

    hi all
    i have scom 2012 sp1 on windows server 2008 r2. sql reporting services (mssql 2012 sp1) installed and configured on this server.
    OM, OMDW, scom`s reportserver db - on other server 2008 r2 (mssql 2012 sp1).
    i try to install management pack sharepoint foundation and server 2013 (use
    http://www.enduria.eu/the-sharepoint-2013-management-pack-and-my-experience-installing-it/).
    in scom console task status - success, but server (sharepoint 2013) in unidentified machines.
    in log on sp2013 server i get error:
    Cannot identify which SharePoint farm this server is associated with. Check the management pack guide for troubleshooting information.
    in log on scom management server i get same error every 10 minutes:
    Failed to deploy reporting component to the SQL Server Reporting Services server. The operation will be retried.
    Exception 'DeploymentException': Failed to deploy reports for management pack with version dependent id 'edf9e0b9-65aa-df29-6729-d16f0005e820'. Failed to deploy linked report 'Microsoft.SharePoint.Server_Performance_Report'. Failed to convert management pack
    element reference '$MPElement[Name="Microsoft.SharePoint.Foundation.2013.Responsetime"]$' to guid. Check if MP element referenced exists in the MP. An object of class ManagementPackElement with ID 75668869-f88c-31f3-d081-409da1f06f0f was not found.
    One or more workflows were affected by this. 
    Workflow name: Microsoft.SystemCenter.DataWarehouse.Deployment.Report
    Instance name: d2781f95-d488-4e35-8226-a9e1d7127149
    Instance ID: {96379E66-14B2-B413-F73F-1F39806AF714}
    Management group: scom-service_group
    what could be the problem?
    thanks

    Same issue...running SCOM 2012 RTM with UR3 on Windows 2008 R2 and SharePoint 2013.  I got the discovery to work after editing the .config file.  All appeared to be working with the exception of the Configuration Databases and Content Databases
    which are currently in a Not monitored state (this could be by design?). 
    I believe this is a bug with the SharePoint 2013 Management Pack.  I tried importing the SharePoint 2010 MP to validate my SCOM environment and Report Server and did not get any errors and the reports imported fine.  Has anyone openened
    a case with Microsoft or know if this is a known issue?  Thanks!
    Error:
    Data Warehouse failed to deploy reports for a management pack to SQL Reporting Services Server. Failed to deploy reporting component to the SQL Server Reporting Services server. The operation will be retried.
    Exception 'DeploymentException': Failed to deploy reports for management pack with version dependent id 'edf9e0b9-65aa-df29-6729-d16f0005e820'. Failed to deploy linked report 'Microsoft.SharePoint.Server_Performance_Report'. Failed to convert management
    pack element reference '$MPElement[Name="Microsoft.SharePoint.Foundation.2013.Responsetime"]$' to guid. Check if MP element referenced exists in the MP. An object of class ManagementPackElement with ID 75668869-f88c-31f3-d081-409da1f06f0f was not
    found.
    One or more workflows were affected by this.
    Workflow name: Microsoft.SystemCenter.DataWarehouse.Deployment.Report
    Instance name: 9afddb34-5f35-40fa-8564-1de5d33773d3
    Instance ID: {4BF8AF22-E5E4-1FC6-3951-F731DCA65317}
    Only two SharePoint 2013 reports are showing up in SCOM:
    SharePoint Foundation Performance
    SharePoint Server Performance
    Missing the following reports (according to the product guide):
    Server Alert
    This report renders all alerts raised by specified server(s).
    Service Alert
    This report renders all alerts raised by specified service(s).
    Server Event
    This report renders all events raised on specified server(s).
    Service Event
    This report renders all events raised on specified service(s).
    Top Server Events
    This report renders top 20 events raised on specified server(s).
    Top Service Events
    This report renders top 20 events raised on specified service(s).
    Top Alerts
    This Most Common Alert Report helps to identify high volume alerts, the volume a distinct alert contributes to the total number of
    alerts and the resolution times. This report helps in tuning the alerts.
    Server Performance
    This report renders performance data for specified server(s).
    Entity State
    This report renders entity state for specified SharePoint object(s) over time.

  • What are the fundamental procurement workflows and tasks?

    A discussion has begun in the Knowledge Management branch of SAP about the [Help documentation for SAP Sourcing|http://help.sap.com/saphelp_sourcing_wave7_p/helpdata/en/39/8ee60a3d74482e9e0111c46dc1e23d/frameset.htm]. Many feel that Help could better serve the needs of SAP Sourcing users if it were organized and written along the lines of the regular workflows and tasks performed by procurement professionals. Some questions arise from this discussion that would best be answered by the community of procurement professionals. That's you. Care to take a shot?
    What do you think about the idea of reorganizing Help according to procurement workflows?
    What exactly are the fundamental workflows and tasks you execute frequently toward the accomplishment of procurement goals?
    Take a look at the Table of Contents/navigation bar toward the left side of the SAP Sourcing Help page. Currently, its top-level topics are Getting Started, Workbench, Spend and Compliance, etc. If you could rename those top-level topics with the names of fundamental procurement workflows, what would you name them?
    Going down a level, what are the fundamental tasks you must execute within any of the fundamental procurement workflows mentioned above?
    Any input you'd give on these questions will help us improve how Help is organized and written as we go forward. It will be greatly appreciated.
    Stephen Muratore
    Information Developer
    SAP Sourcing

    Does this article contain anything useful?
    (19852)

  • How to view and restore sharepoint 2013 designer workflows and how to redeploy with newer version to environments

    how to view and restore sharepoint 2013 designer workflows and how to redeploy with newer version to environments
    MCTS Sharepoint 2010, MCAD dotnet, MCPDEA, SharePoint Lead

    Hi,
    In SharePoint Designer 2010, we could not save the workflow as a template directly except the reusable workflow.
    However, in SharePoint Designer 2013, we can just save all the types workflow as a template, then you can import the workflow into the new environment.
    http://blogs.msdn.com/b/workflows_for_product_catalogs/archive/2012/11/02/deploying-a-workflow-on-a-different-server.aspx
    In SharePoint Designer 2013, every time we publish the workflow, we would get a newer version workflow, and the old workflow version would be overwritten.
    So, when you deploy the workflow in the environment, the workflow would the newer version.
    Thanks,
    Jason
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Jason Guo
    TechNet Community Support

  • CRM Workflow and ERP Workflow integration

    Hi,
    I need to get workflow to work across several modules. One example is that I have some actions that needs to be completed in CRM, and when that has finished I need to trigger a workflow in ERP to continue the processing. How can I send information from the CRM workflow to the ERP workflow, and status message back to CRM? I have been looking at Wf-XML options, but that seems to be across http communication only.. these systems can be on the same server. Can anyone give me some tips on how this can be done? Help is greatly appreciated!
    Regards,
    Thomas

    Hello,
    My suggestion is to trigger the events with a RFC of a Function module.
    Program the Function module in the ERP system and enable it as RCF function. In this module program the trigger for the prefered object.
    In the CRM workflow call a background method of the corresponding object which triggers the Function Module in the ERP system.
    Another solution could be that the action which was taken in CRM leads to a replication in the ERP which already triggers a standard event.
    Hopefully above guides you to a solution.
    Kind regards,
    Peter

  • Connect to Server HTTPs through Proxy Server Problem

    I am trying to connect through a company proxy via port 8080 to a https connection on a server so that it can mount on a desktop. Using 10.4.11 on Mac Pro
    The connection works http through the company network and works http and https through my home network. The company require we use secure connections or something they will supply which will be out of date and clunky to use. I am using FullSwitch as a workflow so need the volume to mount.
    Basically I do the following:
    Go
    Connect to Server
    type address https://??????????????????/???????
    I then get a Proxy Auth where I put my domain, user name and password
    A certificate is then returned and I press continue
    It then attempts to pass through the proxy tunnel to connect but just keeps going and going......
    It should come up at this point with the server login but it doesn't
    We are told by the IT team which only supports PCs that it is then that the Mac tries to create 2000+ connections but doesn't just creates lots of network traffic.
    In between us and the proxy we sit behind a firewall but are told that there are no rules to stop https etc. I can connect to the site through Safari via https but as I need to upload files I need to be able to mount via https as a volume
    Are there any suggestions as to what the issue may be or questions that I can try and ask IT to gather more info?
    Many thanks

    Many thanks for your reply. We can use ftp but can not tunnel sftp through the proxy.
    As I use fullswitch for automation I would like to keep the simplicity of mounting the server as files can be uploaded easily. Currently using http no problems but am being forced to use https from IT Security and of course being in a large corporation with only 35 Macs they say they don't support Macs and hey you have to do it our way or else. Driving me mad.
    Sorry for going off on one but it is so frustrating as the supplier we are sending to wants the files the way we currently send them as they have their system workflow but my company wants us to use their sftp system that is clunky for in and out going file connections forcing suppliers to go to our sftp servers and not us supplying to them.....................
    Right then I currently use Transmit with Type 7 on the proxy type for FTPing files using FTP Protocol not SFTP and have no problems going to the sites we have asked to get specific access to (again takes 2-14 days to get all the firewall rules changed to allow access to new ftp sites we are given, very efficient working practice)
    So as you can see I really want to get the https working as it is the best option for workflow if IT security want us to secure.
    Many thanks, again sorry for my ranting

  • What is the difference between jsp :include and server side include

    what is the difference between jsp :include and server side include(request dispatcher include method)????
    i understand that both request dispatcher include method and jsp:include take dynamic data,so when would one use request dispatcher include and when jsp:include.
    Is the usage interchangeable?i believe jsp include is used only for jsp/html but include directive can be used to include servlets ,jsp and html....correct me if i m wrong and
    do suggest if u hav ny other diff in this context...

    The difference really is: in what format do you want your inclusions? If your environment has many Java developers and only a few designers that focus mainly on, say, Flash, that might push you more towards the server-side include() directive. Or, if you have a large set of pages that receive dynamic content that is displayed in a consistent fashion (such as a workflow header area on a page).
    If, on the other hand, you have more web designers, there may be a greater desire to deal in markup rather than Java code. Java developers themselves might prefer to view markup (JSP) that more resembles the eventual output than something occuring in Java code.
    Finally, there are considerations of tiering. While it is totally possible to (and I have previously) implement 'view classes' that render markup or generate layout templates, JSP's offer, IMO, a subtle, psychological advantage. By forcing a developer to work in a different format, markup versus Java source, the separation on view from controller and model becomes a bit easier. It is still possible to make mistakes, but if a developer at some point notices, "Wait, I'm in a JSP, should I be importing a java.sql class?", then the choice to use JSP includes has paid off in spades.
    - Saish

  • Problems with Safari and Firefox (HTTP?)

    Problems with Safari and Firefox (HTTP?)
    On a laptop G4, 10.5.8 and Safari 5.02 I experience the following:
    On the account of my oldest daughter everything works fine, i.e. wireless internet works and no problems with mail or safari.
    On the same laptop, on the account of my other daughter, the wireless is OK, she can mail etc. But safari nor firefox works. It says: can’t find server (whatever site) and in the activity window it looks if safari tries to open files (in the safari preferences-folder) in stead of http. Same applies to Firefox, so maybe it has more to do with HTTP in general?
    What goes wrong? What to do? I tried the following on the host terminal (tips from Apple)
    defaults write com.apple.safari WebKitDNSPrefetchingEnabled -boolean false
    and
    defaults delete com.apple.safari WebKitDNSPrefetchingEnabled
    but that did not help,
    Nanne

    I'm still wondering why it happens now at this moment in time...
    PC does seem to be a bit odd & inconsistent, the few times I've tested with it, at least so far as we content filtering goes; and if I remember rightly, you're not the first to report previously ok settings suddenly preventing some or all internet until pc is switched off altogether.
    It may work when re-enabled

  • How to make site root-relative links work in DW and Server both?

    See details on buggy DW image link behavior, below. My question is:
    1) how to make site root-relative links work in DW and Server both? Or…
    2) how to reliably automate the change of several hundred legacy root-relative links of the form
    /images/image.jpg  to document-relative?
    That is, to
    ../images/image.jpg or
    ../../images/image.jpg or
    ../../../images/image.jpg etc…depending on where the directory is.
    The old format (/images/image.jpg ) used to work fine in my previous DW 8 configuration but appear grey in DW after “upgrading” to DW cs5.5 mac. (they look fine on the server, but it’s hard to edit image-heavy pages locally when they are all grey).
    I tried changing the files to how DW creates root relative links now:
    /public_html/images/image.jpg, which is a very easy, attractive root flow since there’s a one-to-one mapping. These look great in DW but are broken on the server!
    I looked at the “advanced” site setup, and it looked like it might be possible to nuke the /public_html/ part of my server info…but it also looked like there was the potential for doing damage changing these settings, which are automatically generated from our server connection settings, which seem to work.
    The “links relative to document/ site root” toggle…does that change how DW interprets existing links, or just change the default when you are adding a link?  I have made 80% of the file links document relative…before wondering if root-relative isn’t better?
    It sure seems less ambiguous for all those images if theres a way to make root relative work for DW design view, DW link check, and server.
    Summary of buggy behavior: (see test with images here)
    "old style" site root link
          /images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: broken (grey w/ broken icon)
          Link check in DW: "external link" (i.e., uncheckable, + file could appear orphaned)
          Browser: good
          Ease of switching: n/a (existing format)
    "new style" site root relative link
          /public_html/images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: good
          Link check in DW: good
          Browser: broken
          Ease of switching: easy
    Document relative link
          ../../images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: good
          Link check in DW: good
          Browser: good
          Ease of switching: hard (how to automate?)
    Absolute link
          http://www.oasisdesign.net/images/img_book/WScover120x150_NEW.jpg
          Design veiw in DW: broken (grey w/ broken icon)
          Link check in DW: external (i.e., uncheckable, + file could appear orphaned)
          Browser: good
          Ease of switching: n/a...not a real option
    Thanks!
    Similar discussion on "/"

    Hello again Jon!
    Thanks for jumping on this.
    All clear and understood about where publc_html is etc.
    No contemplation of nuking the actual public_html directory on the server, just the "/public_html" text string at the start of the DW-generated links.
    "/public_html" is automatically added to the front of the link in DW if I create the link with any of the GUI tools, if I have "site root relative" selected. And ""/public_html" ends up in the code, and gets uploaded that way to the server, where it (obviously) doesn't work.
    Doesn't sound like it is supposed to work this way. Also, what seems to be the usual root relative format (/images/image.jpg) shows as a broken link in the GUI and an external link in the DW link check. All this togther makes me thinkI have some obscure setting incorrect?
    The setting that caught my eye is manage sites/ site setup/ advance settings/ local info/ web url,  which is automatically set to http://www.oasisdesign.net/public_html/
    it gives an option to change it but it makes every effort to make this NOT look like something users should mess with:
    Having gone through the more careful thought process during this post, I'm ready to do the experiment of changing the remote server web URL (why is it wrong by default?)...think I'll eat dinner first so there's 45 min to avert disaster if anyone knows this to be a bad idea!
    Art
    PS--don't  have a local testing server...don't think this will solve the GUI broken link/ link shows as external problems.
    Is there an easy, automated way to change links sitewide from document to root relative?

  • Every now and then (every 15 minutes-half an hour or so), my internet fails. It says in the Diagnostics that the internet and server have failed. Happens on my sister's mac as well. Is this our modem?

    Every now and then (every 15 minutes-half an hour or so), my internet fails. It says in the Diagnostics that the internet and server have failed. Happens on my sister's mac as well. Is this our modem? Our PCs don't do this so it has something to do with the macs.

    Hi Patrick,
    I'm starting to think it's either an AirPort card issue or something to do with the Mac OS. I have stopped using wireless as much as possible in my home (have wired up several new Ethernet outlets), and it sure is nice having Internet run at the speed it's supposed to all the time, and with no drops. Might be something to consider.
    In the meantime:
    Open System Preferences > Network > select AirPort > click on Advanced > TCP/IP > Configure IPv6 > Off > OK > Apply > Turn AirPort Off > Turn AirPort On
    Open System Preferences > Network > select Ethernet > click on Advanced > TCP/IP > Configure IPv6 > Off > OK > Apply
    (FOR LION: If you can’t see an option to turn ipv6 off, set it to link-local only.)
    IPv6 troubleshooting http://support.apple.com/kb/TS3802
    And here are some internet troubleshooting ideas:
    http://docs.info.apple.com/article.html?path=Mac/10.6/en/11395.html
    http://support.apple.com/kb/HT1714

  • Tapeless workflows and Sandy Bridge or other PC's: KISS or LOVE?

    Tapeless workflows and Sandy Bridge or other PC's: KISS or LOVE?
    Life used to be so simple when shooting video on a tape based camera. You shot your material, captured it for editing and stored your precious original footage on tape in a safe and dry place. Sure, it took time to capture, but the big advantage was that if you had a computer or drive failure, you would still have the original tape so everything could be recreated.
    Now with tapeless workflows we have the significant advantage of much faster import of the original footage. Connect the flash card or disk drive to the computer over USB and copy the data to a HDD on the computer, ready for editing. The data on the flash card or disk drive can then be erased, so you can reuse it for more shots. But, like Johan Cruyff has said repeatedly, every advantage has its drawback. In this case it simply means that you no longer have the original material to fall back on, in case of computer or drive failures. That is a very unpleasant and insecure feeling.
    The easy anwser to that problem is backups. Backup of the original media, backup of projects and backup of exports. This often means a bundle of externals for backup or NAS configurations. One thing is clear, it requires discipline to make regular backups and it costs time, as well as a number of disks. Four as a minimum: 1 for media, 1 for exports and at least 2 for projects. Note: This is excluding a backup drive for OS & programs.
    There are different backup strategies in use. Some say backup daily and use one disk for monday, one for tuesday, and so on.  Others say one disk for the first backup, the second for the second backup, then the first again for an incremental backup, etc. and once weekly a complete backup on a third disk. Whatever you choose, be aware that shelf live of a disk is far less than tape. There are horror stories everywhere about ball-bearings getting stuck after some time and without original tapes, you better be safe than sorry, so don't skimp on backups.
    What is the relevancy of all this? I thought this was about Sandy Bridge and other PC's.
    It is and let me try to explain.
    Card based cameras are for the most part DSLR and AVCHD type cameras, and we all know how much muscle is required to edit that in a convenient way. Adobe suggests in the system requirements to use raid configurations for HD editing and practice has shown that raid arrays do give a significant performance boost and improve responsiveness, making for a nicer editing experience. The larger the project and the longer the time-line, the more a raid array will help maintain the responsiveness.
    One thing you would not do is using a raid0 for projects, media and exports, even if you have backups. The simple reason is that the chance of disk failure multiplies by the number of disks in the raid0. Two disks double the chance of disk failure, three disks triple the chance, four disks quadruples the chance, etc.
    Remember: Disaster always strikes when it is most inconvenient.
    Imagine you have been working all day on a project, you decide to call it a day and to make your daily backup, but then the raid fails, before you made your backup. Gone is all of today's work. Then take into consideration the time and effort it takes to restore your backups to the state it was in yesterday. That does not make you happy.
    Another thing to avoid is using a software or mobo based parity raid, for the simple reason that it is slooowww and puts a burden on the CPU, that you want to use for editing, not house keeping.
    For temporary or easily recreated files, like the page-file, media cache, media cache database and preview files, it is very much advised to use a raid0. It makes everything a lot snappier and if disaster strikes, so what? These are easily recreated in a short time.
    This was a general overview of what is required with tapeless workflows. Now let's get down to what this means in terms of system design.
    Two approaches or train of thoughts
    KISS: Keep it stupidly simple or LOVE: Laughing over video editing
    The first one, the most economic one, is to use a system with 3 or 4 disks internally and 4 or more backup disks.
    A typical disk setup can look like this:
    This is a perfectly sensible approach if one does not have large or complex projects, long time-lines and is willing to take the risk of occasionally losing a whole days work, between backups. Many hobbyists and consumers fall in this category.
    The KISS approach keeps it stupidly simple. The drawback is that there is no logical way to add more disks or storage. The discipline, diligence and effort required for regular backups make it far from a laughing matter. In fact it can quickly become a bore. Add to that the fact that the disk setup is simple but not very fast, so less suited for situations where lots of clips are involved, multi-cam is a regularly recurring situation or lots of video tracks are involved.
    A number of video editors want more from their system than the occasional platonic KISS, they want to really LOVE their system, which lead to the other train of thought.
    This is more costly than the KISS approach, but you all know a fiancée or wife is more costly and dear than the occasional kiss on the cheek by an old friend.
    Let's start with a typical disk setup. It may look like this:
    Two striking differences in comparison to the KISS approach:
    1. Much easier disk organization and more disks and thus more space.
    2. It requires a hardware raid controller, causing a higher investment cost. It is like an engagement ring. You don't get LOVE for free, one of the guiding principles of the oldest trade in the world.
    These are easy statements to make, but what are the benefits or advantages, that you would fall in LOVE with such a system, and what are the drawbacks? Think back to Johan Cruyff's adage.
    The only drawback is cost. The advantages are multiple, easier organization, more speed, more storage, snappier editing, no jerkiness, lesser requirements for regular backups and - this is the major benefit - hardly a chance of losing a day's work in case of a drive failure. Keep in mind that a parity raid keeps all your data intact in case of a drive failure, so lessens the need for up-to-date backups.
    We all know, we get what we pay for: "If you pay peanuts, you get monkeys. OTOH, if you pay money to monkeys, you get rich monkeys". But in this case you get what you pay for, a much better editing experience with a much easier workflow.
    Using a parity raid (be it raid 3/5/6/30/50/60) you get security, ease of mind that you are protected against losing precious media, that you need not worry about the last time you made a backup, that the editing you did today may be lost and you save valuable time editing and a lot of aggravation because of a much more responsive system.
    How does this all relate to Sandy Bridge and other PC's?
    First of all, the price difference between a Sandy Bridge / P67 platform and an i7-950+ / X58 platform is very small. Of course the new architecture is slightly more expensive than the older one, but the differences are small, almost not worth talking about.
    So what are the differences? Look below:
    The first thing to keep in mind is that the Sandy Bridge is the successor of the i7-8xx CPU and as such it is much more evolutionary than revolutionary. The CPU power has increased significantly over the i7-8xx due to new architecture and a smaller production process (32 nm), but in essence all the capabilities have remained unchanged. Same memory, same PCI-e lanes, same version, same L3 cache and no support for dedicated raid controllers.
    It is great that the processor performs much better than the older i7-8xx CPU's, almost achieving the level of the i7-9xx range of processors, but is still limited:
    The Sandy Bridge is unsuitable for anything more than a KISS system.
    Why? Because it lacks the required PCI-e lanes to accomodate more than a 16 x PCI-e nVidia card with CUDA support to enable hardware MPE acceleration and the integrated graphics are not supported by CS5.
    You may wonder if that is a bad thing. The plain and simple anser is NO. It is a great processor, it delivers great value for money, is a solid performer, but it has its limitations. Intel had a reason to position this CPU as a mid-level CPU, because that is what it is, a mid-level performer in comparison to what is to come.
    The term mid-level performer may seem strange when compared to the old generation of i7-9xx CPU's, because they perform almost equally well, but keep in mind that there is a generation difference between them.
    So what about the i7-9xx and X58 platform?
    It still is going strong. About the same performance as a Sandy Bridge, with only the much more expensive hexa-cores clearly in the lead, both performance and price wise. The quad cores deliver about the same value for money.  The main difference however is the platform that allows a dedicated raid controller to be installed, thus making it the platform of choice for those who want to go from a passing KISS to true LOVE.
    And what lies ahead?
    Sandy Bridge E on the Waimea platform (X68). Now that is revolutionary. More than double almost everything a processor can offer: double the cores, double the PCI-e lanes, triple the memory, more than double the L3 cache, increase the PCI-e support from 2.0 to 3.0, etc...
    This is why Intel calls this a high-end CPU / platform.
    So what now?
    If you prefer a KISS approach, choose either a Sandy Bridge/P67 or an i7-950+/X58 platform.
    If you wonder whether in the future you may need multi-cam more frequently, edit more complex projects and longer timelines or even progress to RED, look at KISS/LOVE solutions, meaning the i7-950+/X58.
    If you can't have downtime, time pressure is high, delivery dates to clients are critical or you edit highly complex projects, lots of multi-cam situations or lengthy time-lines, choose a LOVE solution, an i7-950+/X58 platform.
    If you have the time to wait till Q4/2011, Sandy Bridge E/Waimea looks to be worth the wait.
    Hope this gives you some more insight into recent and future developments and helps you make wise investment decisions.

    I'm upgrading from an AMD 3800+, cutting with Vegas 7 Pro. Usually shoot DSLR or HDV, sometimes P2, EX or RED. I have ridiculously cheap access to Macs, FCP/FCS, all kinds of software.
    I've been agonizing over this for the last month, was originally hoping the UD7 mobo was the solution, read the read about the NF200/PCIe issue a few days ago, http://www.dvinfo.net/forum/non-linear-editing-pc/489424-i7-980x-now-wait-sandybridge-2.ht ml- and still decided to go for a 2600k. 
    My preference is to treat my video footage the same way as my digital imagery: I make (at least) duplicate back ups of everything before reformatting the cards, never delete the back ups, and only worry about the day-to-day stuff at night. Unless I'm rendering or involved in other long processes, in which case I'll back up the work in process the next day. If I am under a really really tight deadline I might back up as I go.
    Yes, a RAID might make it easier, but I'm paranoid enough to prefer a slower, safer backup. You can always duplicate, and usually improve upon, a days work, but you can never get back original footage you lost. I have only ever had one hard drive die on me (a few enclosures crapped out, though)- it took a couple of (mostly unattended) hours to rectify. As a matter of act, I've had far more loss/damage from tapes than from hard drives.
    I ordered the UD7, 2 F4s and 4 F3Rs, understanding I will probably want to upgrade to SBE when it comes out, or maybe next year. The 2600k/mobo/RAM will likely hold its value better than a 950/X58, likely because of the marketplace as much as merit.
    The UD7 / RAID card issue is in it's early days, there may be a solution/mitigation. Probably not. But if I really really need a RAID card, then I probably really really need a 980, NAS, etc etc.
    But Harm still rocks!

  • Difference SAP Business Workflow and ccBPM

    Hi,
    well both mentioned are for connecting Applications. In my point of view there is only a difference in...
    SAP Business Workflow:
    connects Applications via Tasks across SAP-Systems
    ccBPM:
    with help of PI it should connect Applications via Adapter and is only message-driven!
    am I right?!
    Can someone please explain how to implement Business Workflow in SAP-Systems?! I am only aware of ccBPM!
    br Jens

    Hi Carsten
    how to implement Business Workflow in SAP-Systems
    <b>Business Workflow - Tutorials</b>
    http://help.sap.com/saphelp_nw04/helpdata/en/fb/135962457311d189440000e829fbbd/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/04/926f8546f311d189470000e829fbbd/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/a6/68063a5a636405e10000000a11402f/frameset.htm
    http://help.sap.com/saphelp_nw04s/helpdata/en/ae/b82cc8e10611d2a62f0060087a79ea/frameset.htm
    This url below(first Url) is a very good document to integrate workflow into portal. The other documents are for customization.
    https://media.sdn.sap.com/html/submitted_docs/Best_Practices/EP/documentation/EP/N10_BB_ConfigGuide_EN_US.doc
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/fa21a890-0201-0010-708b-d0cfc117e7cd
    http://help.sap.com/saphelp_nw2004s/helpdata/en/39/a1bb5c4c0d4ab4a417e87ef35f1efa/frameset.htm
    business workplace
    /thread/369036 [original link is broken]
    /thread/183229 [original link is broken]
    Difference between Guided Procedures and Business Workflow?
    CAF / GP  Versus Business Workflow
    Thanks !!

  • Internet not working - wifi connected but internet and server fail, ethernet cable doesn't work either

    All of a sudden the internet on our Macbook Pro does not work.  We connect through an Airport Extreme and other mac works fine.  It shows that wifi is fully connected but when i go to diagnostics, i see the internet and server are failing.  I tried connecting an ethernet cable and it doesn't work that way either.  Have restarted and reset modem and router, nothing is working… help!

    Hi ehamzic1,
    Welcome to the Apple Support Communities!
    I understand how frustrating it can be when your computer cannot connect to the internet. In your situation, you will need to isolate to find what is causing the issues connecting. I would recommend reading over and working through the attached article. The article should help you troubleshoot and resolve this situation.
    Solutions for connecting to the Internet, setting up a small network, and troubleshooting
    http://support.apple.com/kb/HT1714
    Have a great day,
    Joe

Maybe you are looking for

  • A G4 running/booting off a lacie with TIGER 10.4.7

    i have amongst many G5's , a G3 pismo and a G4 400MHZ desktop that i wish to have boot off of and run as if Tiger was their OS. The pismo cant run tiger itself i think but could it boot off of TIGER if that was on an external 7200 firewire HD? could

  • VArious problems with intel graphic card

    There is an old regarding part of this subject with no improvement. I have a 943/940GML graphic card. Normally the frame rate should synchronize with vfresh but it does not seems to works well: glxgears is not smooth at all unless you move the mouse

  • Mail wants to create new account after crash

    so, i ran out of batteries and my laptop just died. power attached, restart, my firefox session is there with all windows intact, but when i start up mail, it gives me the "welcome to mail" create account wizard. HELP! of course i know all my mail is

  • Help Needed - Accidentally Reset Lumia 610 - Lost ...

    Hi. I need help rescuing some family videos from my maid's Lumia 610. I've accidentally reset the phone, so all the data seems to have disappreared. Can someone help me, with get the videos back, or maybe, I could send it to Nokia to help retrieve th

  • Shared Disk solution for Transaction Logs and JMS?

              In a machine failure, two key resources for a clean recovery are the Transaction Logs           and the JMS Store.           The WebLogic Server Administration Guide discusses moving the Transaction Logs to           a new machine before st