Cloud service auto scale delay

Hi team, 
We have a cloud service that have 10 worker roles,  to let the cloud service can auto scale up/down, we configure the autoscale settings that if the average CPU reaches 70% and keeps last 30 minutes, new worker role instance will be created.
We also configure a email notification monitor that if the worker role average CPU reach 75% and keeps last 30 minute it will send a notification to our administrator, we receivce the notification at 2015/1/26 0:55 and 1 hour later the instances ware scaled
up, is that means  the autoscale have been delayed about 1 hour , is there some best practise that avoid this kind of problem and let the instances can be scaled up immeditatly
thanks in advance.

Hi,
Please have a look at this article:
http://azure.microsoft.com/en-us/documentation/articles/cloud-services-how-to-scale/, here is a snippet.
All instances are included when calculating the average percentage of CPU usage and the average is based on use over the previous hour. Depending on the number of instances that your application is using, it can take longer than the specified wait time
for the scale action to occur if the wait time is set very low. The minimum time between scaling actions is five minutes. Scaling actions cannot occur if any of the instances are in a transitioning state.
Hope this helps.
Best Regards,
Jambor
We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
Click
HERE to participate the survey.

Similar Messages

  • Changing cloud services auto scale weekends

    hi
    i want to change the default weekends in the azure auto scale ( our weekend is from Friday to Saturday ).
    can it be done using the portal or in any other way ?

    hi,
    >>( our weekend is from Friday to Saturday )?
    Base on my experience, if your weekend has two days, you could set weekend by selected the prior time zone .
    About customize the weekend time, it may be difficult on azure portal.
    Another way, I suggest you could try to auto scale could service by customize time using the Windows Azure Monitoring Services Management Library . I recommend you could refer to this code sample about how to use this library (http://blogs.msdn.com/b/cie/archive/2014/02/20/how-to-use-windows-azure-monitoring-services-management-library-to-create-an-autoscale-rule.aspx
    ). And you also see this document (http://msdn.microsoft.com/en-us/library/hh680945(v=pandp.50).aspx ).
    Hope this helps.
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Azure Cloud Service Scaling - do I have to configure a Load Balancer?

    I'm a little bit confused by how scaling in Azure works. I'm using a Cloud Service and have 2 web roles running a PHP application. I can RDP on both machines and both applications run great on each machine. Also I don't have any problems calling the staging
    URL.
    But I can't figure out if I configure scaling so that 2 machines run always, if I have to configure a load balancer somehow. Or is this already done for me?
    In Azure VM's I had to create a load-balanced set endpoint for an endpoint, but what about cloud services?
    And how is this done in the XML configuration file for my service? What if I don't do it?

    Hi,
    Scaling is affected by core usage. Larger role instances or Virtual Machines use more cores. You can only scale an application within the limit of cores for your subscription. For example, if your subscription has a limit of twenty cores
    and you run an application with two medium sized Virtual Machines (a total of four cores), you can only scale up other cloud service deployments in your subscription by sixteen cores. All Virtual Machines in an availability set that are used in scaling an
    application must be the same size.
    Windows Azure supports load balance for cloud services and standard websites, we just need to set instance count to more than 1 to enable load balance. For virtual machines, it needs to set up manually.
    Please refer this link for Load Balance a Virtual Machine:
    http://www.windowsazure.com/en-us/manage/windows/common-tasks/how-to-load-balance-virtual-machines/
     for more information.
    Auto scale lets you set scaling limits and scheduling goals to ensure you are always getting optimal performance
    Please refer this link for Scaling on Cloud Services:
    http://azure.microsoft.com/en-us/services/cloud-services/
    Also, Please refer this link for Scaling an Application :
    http://azure.microsoft.com/en-us/documentation/articles/cloud-services-how-to-scale/
    XML configuration : Azure (Load-balanced) Endpoints can only be used for TCP/UDP based services. please check
    https://techlib.barracuda.com/display/BNGv54/How+to+Configure+a+High+Availability+Cluster+in+Azure/printable for the detailed information
    Hope this helps.
    Regards,
    Shirisha Paderu.

  • A question regarding using a JDBC class to connect with cloud service database

    Hello,
    I am currently working on a small scale cloud service report where the company I chose is obviously Oracle. My question is regarding the cloud service in the following way. I was doing my report with the free trial until it just came to me that
    why not to do a small one class program with my netBeans or Eclipse that uses JDBC but I am not sure what username, password and the url to use in the connection to retrieve, manipulate and store values. Can somebody help me please if this is possible or not?
    edit: Anyone please? I have a deadline in 15.8. and could create something great until then if I get the anwser in few days

    To correct my question, I already have the oracle account and I created the cloud service trial account with database and java section.

  • Need advice for starting a Managed Cloud Service for Small Businesses

    I hope this is in the right forum.  I have done a lot of research and searching but havent found anything that specifically answers, in total, what I am wanting to accomplish.  I live in a small town and want to start a Managed Cloud Services for
    small to small-medium business in my area (2-30 users for each business).  I want to market this to have businesses replace their in-house server(s) to virtual ones I would host in a local Data Center with my own equipment that I would maintain.  I
    am just starting off so I don't have any clients I do this for currently, but I get asked about this frequently.  I want to run a 2012 R2 Domain Controller and a Hyper-V 2012 R2 server.  The virtual servers I will host are going to be for AD, RDS,
    FTP, and files.  Software examples that people are going to be using these virtual servers for are Quickbooks, Sage Accounting, Remote Desktop or RemoteApp, custom CRM or small database software, Office 2013, etc.  No Exchange currently but will
    probably configure something for that in the future (maybe run 1-3 virtually for now if someone asks, but will only do it if the user base is fairly small ~under 10 users).  I only have 1 static IP to work with over a 100Mbps connection up and down.
    For hardware, I am figuring something along the lines of this:
    (1) 1U, single CPU w/2-4 cores, 8GB, 2x73GB SAS 10k RAID 1, Dual PSU, running Windows Server 2012 R2
    Domain Controller
    (1) 2U, 2x 8-core Xeon ~2.6Ghz, 80GB RAM, 8x600GB SAS 15k in Raid 10 for Storage (VHDX files, etc), RAID 1 small Basic drives (or USB stick) for OS, Dual PSU, Quad GB Nic which I can use for load balancing/teaming, Hyper-V
    2012 R2
    Hyper-V Virtual Server
    (1) GB Unmanaged Network Switch & (1) Cisco 5510 Firewall
    Most of my questions are about the best way to configure this.  I am planning on managing my Hyper-V from the physical Domain Controller server.  Each virtual server will have RDS & (possibly) AD services on a single server.
    1) I want to replicate the physical Domain Controller.  Should I get another server or just virtualize the replica in Hyper-V?  I understand that if the Hyper-V goes down, so does my DC replica.
    2) Should I use my Domain Controller to manage ALL users on each virtual Server, by creating separate Organizational Units for each business?
    3) Should I setup my domain controller with Hyper-V management and then each Virtual Server I setup be a separate domain (Ex. mydomain.local, business1.local, business2.local, etc)?  Each one has no connection to any
    other, completely seperate.  Or should I do subdomains (business1.mydomain.local).
    4) What I have read is that Subdomains are a pain to manage with user rights, etc.  I want to keep each server complete separated from one another over a network connection, I suppose the VLan through Hyper-V options
    do this?  I dont want wondering users to stumble upon another businesses files (I know they would probably be prompted with a login for that business/domain).
    5) For each virtual server, I want to create and have an HTTP subdomain point to that server from my domain name. (Ex: business1.mydomain.com, business2.mydomain.com, etc.)  I want them to be able to have access to
    only their RemoteApps or be able to type that address in their Remote Desktop program as the host name.  This would be for viewing the RemoteApp login page and RemoteApps for that business over HTTP/S through a browser.
    6) If I do not have separate DC's in each virtual and my main DC manages each one, is their a way to connect up each companies RemoteApps using a single site that only shows what they are assigned to based upon their login?
    (Ex. http://login.mydomain.com which then shows that user what they are assigned on their own virtual server)
    7) Since each business will use the same ports for RemoteApp (443) & RDC (3389 unless I change it), how would I setup the subdomains to point to their correct server and not overlap for mess with any of the other servers
    since its all over 1 static WAN IP for all servers.  Thats why I figured setting up IIS subdomains would solve this.
    8) For backups or Hyper-V replication, is it better to have software that backs up the ENTIRE Hyper-V server (Acronis Advanced Backup for Hyper-V) as well as replication or just backups?  Or should I do separate file
    backup on each virtual with a replica?  Can a replica be a slower server since its just a backup? (Ex. 1x 8 core, 80GB, 8x600GB 10k SAS)
    9) For the servers that will be using FTP, can I again rely on the subdomains to determine which server to connect to on port 21 without changing each FTP servers ports?  I just want each business/person to type in
    the subdomain for their business and it connect up to their assigned FTP directory over port 21.
    10) If the physical DC manages DNS for all Virtual servers, can I forward sub domain requests to the proper virtual server so they connect to the correct RemoteApp screen etc.  Again all I have is 1 IP.
    I hope all of these questions make sense.  I just want every business to be independent of each other on the Hyper-V, each on their own virtual server, all without changing default ports on each server, each server running RDS, (possibly) AD, (a few) FTP,
    and all over a common single WAN IP.  Hoping subdomains (possibly managed through IIS on the physical DC) will redirect users to their appropriate virtual server.

    If you really want to run your own multi-tenant service provider cloud, Microsoft has defined the whole setup needed.
    They call it Infrastructure as a Service Product Line Architecture.  You can find the full documentation here -
    http://blogs.technet.com/b/yuridiogenes/archive/2014/04/17/infrastructure-as-a-service-product-line-architecture.aspx
    There are several different ways of configuring and installing it.  Here is a document I authored that provides step-by-step instructions for deploying into a Cisco UCS and EMC VSPEX environment -
    http://www.cisco.com/c/dam/en/us/td/docs/unified_computing/ucs/UCS_CVDs/ucs_mspc_fasttrack40_phase1.pdf
    This document contains the basic infrastructure required to manage a private cloud.  I will soon be publishing a document to add the Windows Azure Pack components onto the above configuration.  That is what would more easily provide a multi-tenant
    experience with a Azure look and feel.  It is not Azure, but the Azure pack is a series of applications, some of which came from Azure, the provides Azure-like capabilities only in a service provider type of environment.
    Whether you use my document or not (which has actually corrected errors found in the Microsoft documentation), you should take a look at it to see what it takes to put something like this up, if you are really serious about it.  It is not a small undertaking. 
    It requires a lot of moving pieces to be coordinated.  Yes, my document is designed to scale to a large environment, but you need the components that are there.  No need re-inventing the wheel.  Microsoft's documentation is based on a lot of
    real hands on experience of their consulting organization that has been doing this for customers for years.  This one is also know as Fast Track 4.  I've done 2 (2008 R2) and 3 (2012), also and it just keeps getting more complicated based on customer
    demands and expectations.
    Good luck!
    . : | : . : | : . tim

  • What are Azure limitations for Websockets in Cloud Services (web and worker role)?

    A WebSocket Server should be built on Azure platform with OnPrem connections and have questions regarding limitations for Websockets in Azure Cloud Services - web and worker roles.
    Websockets can be configured for Web Sites and limitations are understood, but Azure Websites is not an option. 
    Nevertheless it is planned to run a web service (without UI - no web site) as a Cloud service which has secure websocket (WSS) connections to OnPrem machines. Websocket protocol is enabled for IIS8 on Cloud services web and worker roles. Azure Service Bus Relay
    is not an option.
    Questions:
    1) Are Websockets supported for Azure Cloud services web and worker roles? we assume yes
    2) What are potential limitations from Azure side to support concurrent Websocket connections? We are aware that CPU, memory etc are limitations, but are there additional limitations from MS Azure side? 
     

    Hi,
    As I know, azure cloud service web and worker role support Websockets, users can connect to the role via the special endpoint, if we use Azure cloud service, I think we can monitor the metrics such as CPU, memory, etc... and scale our cloud service via these
    metrics to keep the websockets working, refer to
    http://azure.microsoft.com/en-us/documentation/articles/cloud-services-how-to-scale/ for more information about how to scale a cloud service.
    Regards

  • Async / Await in Cloud Services Worker Roles?

    Good evening,
    I was wondering if and how one would be able to utilize async / await for Azure Cloud Services' Worker Roles? Basically I have background jobs that will/shall run on worker roles, however many of those are performing up/downloads from Azure Storage blobs
    and ideally I'd like to await those calls.
    So is async/await natively supported for CS Worker Roles (using Azure SDK 2.5) and if so, how? And if not.. is it planned at all?
    Thanks,
    -Jörg

    Hi,
    I would suggest you have a look at this thread:
    http://stackoverflow.com/questions/15991287/async-await-in-azure-worker-role-causing-the-role-to-recycle
    A simple solution is to just do this:
    public override void Run()
    RunAsync().Wait();
    public async Task RunAsync()
    while (true)
    await Task.Delay(60000);
    Hope this helps.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Error when trying to deploy ADF application in Java Cloud Service - SaaS Extension

    Hello guys,
    I'm trying to deploy a simple ADF application in "Oracle Java Cloud Service - SaaS Extension" and i'm still having the error below
    The job turns to Failed at "Deploy Application" step:
    Did someone already got this error ?
    Thanks at advance
    Sid
    2014-12-10 11:06:20 PST: Starting action "Deploy Application"
    2014-12-10 11:06:20 PST: Deploy Application started
    2014-12-10 11:06:28 PST: weblogic.application.ModuleException: weblogic.application.ModuleException:
      at weblogic.servlet.internal.WebAppModule.startContexts(WebAppModule.java:1531)
      at weblogic.servlet.internal.WebAppModule.start(WebAppModule.java:488)
      at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
      at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
      at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
      at weblogic.application.internal.flow.ScopedModuleDriver.start(ScopedModuleDriver.java:200)
      at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:247)
      at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
      at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
      at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
      at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:27)
      at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:671)
      at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
      at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:212)
      at weblogic.application.internal.SingleModuleDeployment.activate(SingleModuleDeployment.java:44)
      at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:161)
      at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:80)
      at weblogic.deploy.internal.targetserver.operations.AbstractOperation.activate(AbstractOperation.java:573)
      at weblogic.deploy.internal.targetserver.operations.ActivateOperation.activateDeployment(ActivateOperation.java:150)
      at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doCommit(ActivateOperation.java:116)
      at weblogic.deploy.internal.targetserver.operations.AbstractOperation.commit(AbstractOperation.java:327)
      at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentCommit(DeploymentManager.java:844)
      at weblogic.deploy.internal.targetserver.DeploymentManager.activateDeploymentList(DeploymentManager.java:1253)
      at weblogic.deploy.internal.targetserver.DeploymentManager.handleCommit(DeploymentManager.java:440)
      at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.commit(DeploymentServiceDispatcher.java:163)
      at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doCommitCallback(DeploymentReceiverCallbackDeliverer.java:195)
      at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.access$100(DeploymentReceiverCallbackDeliverer.java:13)
      at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer$2.run(DeploymentReceiverCallbackDeliverer.java:68)
      at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:545)
      at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
      at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)
    Caused by: java.security.AccessControlException: access denied ("java.net.NetPermission" "specifyStreamHandler")
      at java.security.AccessControlContext.checkPermission(AccessControlContext.java:372)
      at java.security.AccessController.checkPermission(AccessController.java:559)
      at java.lang.SecurityManager.checkPermission(SecurityManager.java:549)
      at java.net.URL.checkSpecifyHandler(URL.java:649)
      at java.net.URL.<init>(URL.java:373)
      at weblogic.application.io.MergedDescriptorFinder.getSource(MergedDescriptorFinder.java:46)
      at weblogic.application.io.DescriptorFinder.getSource(DescriptorFinder.java:44)
      at weblogic.utils.classloaders.MultiClassFinder.getSource(MultiClassFinder.java:67)
      at weblogic.application.utils.CompositeWebAppFinder.getSource(CompositeWebAppFinder.java:71)
      at weblogic.servlet.internal.War$ResourceFinder.getSource(War.java:1213)
      at weblogic.servlet.internal.War$ResourceFinder.getSource(War.java:1203)
      at weblogic.servlet.internal.War.getResourceAsSource(War.java:512)
      at weblogic.servlet.internal.WebAppServletContext.getResourceAsSource(WebAppServletContext.java:3436)
      at weblogic.servlet.internal.WebAppServletContext.getResourceAsSource(WebAppServletContext.java:3427)
      at weblogic.servlet.internal.WebAppServletContext.getResourceAsStream(WebAppServletContext.java:872)
      at com.sun.faces.config.ConfigureListener$WebXmlProcessor.scanForFacesServlet(ConfigureListener.java:805)
      at com.sun.faces.config.ConfigureListener$WebXmlProcessor.<init>(ConfigureListener.java:768)
      at com.sun.faces.config.ConfigureListener.contextInitialized(ConfigureListener.java:178)
      at weblogic.servlet.internal.EventsManager$FireContextListenerAction.run(EventsManager.java:481)
      at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
      at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
      at weblogic.servlet.internal.EventsManager.notifyContextCreatedEvent(EventsManager.java:181)
      at weblogic.servlet.internal.WebAppServletContext.preloadResources(WebAppServletContext.java:1871)
      at weblogic.servlet.internal.WebAppServletContext.start(WebAppServletContext.java:3173)
      at weblogic.servlet.internal.WebAppModule.startContexts(WebAppModule.java:1529)
    2014-12-10 11:06:28 PST: WL action state: failed
    2014-12-10 11:06:28 PST: Action FAILED with WL action state: failed
    2014-12-10 11:06:28 PST: Check the server log of your Java cloud service for more info about the failure.
    2014-12-10 11:06:28 PST: Application deployment failed.
    2014-12-10 11:06:28 PST: "Deploy Application" complete: status FAILED

    The Application does nothing, it's just a simple page (the button does nothing too)  the aim is to deploy a jsf page with ADF forms.
    The deployment log is on the first message.
    I'm using Jdeveloper and i note the Jdev inserts some servlets into the web.xml file ! is it possible that the probleme is related to this ? (below the web.xml file)
    Sid
    <?xml version = '1.0' encoding = 'windows-1252'?>
    <web-app xmlns="http://java.sun.com/xml/ns/javaee" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
             xsi:schemaLocation="http://java.sun.com/xml/ns/javaee http://java.sun.com/xml/ns/javaee/web-app_2_5.xsd"
             version="2.5">
      <servlet>
        <servlet-name>Faces Servlet</servlet-name>
        <servlet-class>javax.faces.webapp.FacesServlet</servlet-class>
        <load-on-startup>1</load-on-startup>
      </servlet>
      <servlet>
        <servlet-name>resources</servlet-name>
        <servlet-class>org.apache.myfaces.trinidad.webapp.ResourceServlet</servlet-class>
      </servlet>
      <servlet>
        <servlet-name>BIGRAPHSERVLET</servlet-name>
        <servlet-class>oracle.adf.view.faces.bi.webapp.GraphServlet</servlet-class>
      </servlet>
      <servlet>
        <servlet-name>BIGAUGESERVLET</servlet-name>
        <servlet-class>oracle.adf.view.faces.bi.webapp.GaugeServlet</servlet-class>
      </servlet>
      <servlet>
        <servlet-name>MapProxyServlet</servlet-name>
        <servlet-class>oracle.adf.view.faces.bi.webapp.MapProxyServlet</servlet-class>
      </servlet>
      <servlet-mapping>
        <servlet-name>Faces Servlet</servlet-name>
        <url-pattern>/faces/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>resources</servlet-name>
        <url-pattern>/adf/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>resources</servlet-name>
        <url-pattern>/afr/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>BIGRAPHSERVLET</servlet-name>
        <url-pattern>/servlet/GraphServlet/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>BIGAUGESERVLET</servlet-name>
        <url-pattern>/servlet/GaugeServlet/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>MapProxyServlet</servlet-name>
        <url-pattern>/servlet/mapproxy/*</url-pattern>
      </servlet-mapping>
      <servlet-mapping>
        <servlet-name>resources</servlet-name>
        <url-pattern>/bi/*</url-pattern>
      </servlet-mapping>
      <filter>
        <filter-name>trinidad</filter-name>
        <filter-class>org.apache.myfaces.trinidad.webapp.TrinidadFilter</filter-class>
      </filter>
      <filter>
        <filter-name>ServletADFFilter</filter-name>
        <filter-class>oracle.adf.share.http.ServletADFFilter</filter-class>
      </filter>
      <filter-mapping>
        <filter-name>trinidad</filter-name>
        <servlet-name>Faces Servlet</servlet-name>
        <dispatcher>FORWARD</dispatcher>
        <dispatcher>REQUEST</dispatcher>
        <dispatcher>ERROR</dispatcher>
      </filter-mapping>
      <filter-mapping>
        <filter-name>ServletADFFilter</filter-name>
        <servlet-name>Faces Servlet</servlet-name>
        <dispatcher>FORWARD</dispatcher>
        <dispatcher>REQUEST</dispatcher>
      </filter-mapping>
      <context-param>
        <param-name>javax.faces.STATE_SAVING_METHOD</param-name>
        <param-value>client</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.PARTIAL_STATE_SAVING</param-name>
        <param-value>false</param-value>
      </context-param>
      <context-param>
        <description>If this parameter is true, there will be an automatic check of the modification date of your JSPs, and saved state will be discarded when JSP's change. It will also automatically check if your skinning css files have changed without you having to restart the server. This makes development easier, but adds overhead. For this reason this parameter should be set to false when your application is deployed.</description>
        <param-name>org.apache.myfaces.trinidad.CHECK_FILE_MODIFICATION</param-name>
        <param-value>false</param-value>
      </context-param>
      <context-param>
        <description>Whether the 'Generated by...' comment at the bottom of ADF Faces HTML pages should contain version number information.</description>
        <param-name>oracle.adf.view.rich.versionString.HIDDEN</param-name>
        <param-value>true</param-value>
      </context-param>
      <context-param>
        <description>Security precaution to prevent clickjacking: bust frames if the ancestor window domain(protocol, host, and port) and the frame domain are different. Another options for this parameter are always and never.</description>
        <param-name>org.apache.myfaces.trinidad.security.FRAME_BUSTING</param-name>
        <param-value>differentOrigin</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.VALIDATE_EMPTY_FIELDS</param-name>
        <param-value>true</param-value>
      </context-param>
      <context-param>
        <param-name>oracle.adf.view.rich.geometry.DEFAULT_DIMENSIONS</param-name>
        <param-value>auto</param-value>
      </context-param>
      <context-param>
        <param-name>oracle.adf.view.rich.SYNCROWS</param-name>
        <param-value>enable</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.FACELETS_SKIP_COMMENTS</param-name>
        <param-value>true</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.FACELETS_DECORATORS</param-name>
        <param-value>oracle.adfinternal.view.faces.facelets.rich.AdfTagDecorator</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.FACELETS_RESOURCE_RESOLVER</param-name>
        <param-value>oracle.adfinternal.view.faces.facelets.rich.AdfFaceletsResourceResolver</param-value>
      </context-param>
      <context-param>
        <param-name>javax.faces.FACELETS_VIEW_MAPPINGS</param-name>
        <param-value>*.jsf;*.xhtml</param-value>
      </context-param>
      <mime-mapping>
        <extension>swf</extension>
        <mime-type>application/x-shockwave-flash</mime-type>
      </mime-mapping>
      <mime-mapping>
        <extension>amf</extension>
        <mime-type>application/x-amf</mime-type>
      </mime-mapping>
      <listener>
        <listener-class>oracle.adf.mbean.share.config.ADFConfigLifeCycleCallBack</listener-class>
      </listener>
      <login-config />
    </web-app>

  • Add Operational Insights to Cloud Service during publish

    Hello,
    I have several Azure Services that I am already utilizing Application Insights on. What I would like to do is have Operational Insights added during the standard "Publish to Windows Azure" that I use from Visual Studio thereby removing the need
    to manually remote into the machine and then install Microsoft Monitoring Agent and configuring my account. Can you point me to some resources demonstrating or explaining how to accomplish this?
    A couple of notes:
    My group has dozens of services and workers so manual install/configure is not an option for us.
    All of our machines currently also have Application Insights setup and running so MMA already has a tab for Application Insights that displays if you open MMA however, there is no tab available in MMA for Operational Insights. The only way that I can add
    this tab is to uninstall and then re-install using the installer found on the Operational Insights site however, doing so removes Application Insights.
    Thanks!

    We are working on a VM Extension to push the agent into IaaS VMs from Azure Portal as a first class experience.
    The community has also come up with a DSC resource already anyway!
    http://www.powershellmagazine.com/2014/11/26/dsc-resource-module-for-microsoft-monitoring-agent-install-and-configuration-for-azure-operational-insights/
    For Cloud Services, we don't need/use the agent but support them already by pulling their logs from WAD - if the roles are configured to write to WAD (the usual way) we'll pick Windows Events and IIS logs (at the moment) - more data sources ideas pls see
    this category on the feedback forum
    http://feedback.azure.com/forums/267889-azure-operational-insights/category/88086-log-management-and-log-collection-policy
    We think a lot of the other scenarios/IP's don't make sense for PaaS:
    - System Updates: since PaaS roles are auto-updated
    - AntiMalware: we are investigating how to support Anti Malware scenario using logs - read the comment from Richard on this thread for queries using Windows Events you can do today for that scenario
    http://feedback.azure.com/forums/267889-azure-operational-insights/suggestions/6519202-support-other-antivirus-products-in-malware-assess#comments
    - Capacity is currently only focused on HyperV (private cloud) but we'll be able to support 'performance' type information thru WAD again
    http://feedback.azure.com/forums/267889-azure-operational-insights/suggestions/6662146-open-up-the-capacity-management-pack-for-other-sys so no agent needed either...
    - SQL Assessment: you don't install SQL in Cloud Services roles; you do in IaaS VMs.
    and so on...
    The agent for AppInsights and OpInsights at the moment are not compatible - see this
    http://feedback.azure.com/forums/267889-azure-operational-insights/suggestions/6660334-it-s-good-if-we-have-single-mma-file-for-appinsigh but again, you use Appinsights agent in your Web Roles, and you use OpInsights agent in your VMs, and for the
    web roles you get the info into OpInsights from WAD. that is the most common use case.

  • Azure Cloud Service internal server error (500) while scaling up

    I experienced multiple times that scaling up a Cloud Service from 1 (small) instance to 2 (small) instances causes a few minutes of downtime with request returning a an error 500. Is this how it's supposed to work?

    Thanks for the reply!
    Every initialization code is in OnStart() and in case of .cmd scripts, in <Startup> <Task>s inside the service definition... So I suppose until that is run to completion the Azure load balancer shouldn't dispatch any requests to the new instance.
    I tested today everything again, this time with Windows Server 2012 R2 (the Cloud Service didn't use R2 before). The issue is the same:
    Scaling up the service to two instance.
    Second instance is being deployed and starts.
    Requests going to the first instance work (i.e. since clients stick to their instance users who visited the site before the scale-up will continue to be serve through the first instance, hence not noticing anything). Requests going to the second instance
    don't work (internal server error), but the main issue is that requests start to go to the second instance before it is fully initialized!
    (I.e. Azure Management Portal says the role is starting, but still, I can manage to get my request to it, thus I suppose the load balancer does dispatch requests while the role is not fully started yet.)
    Failed request tracing doesn't show anything related (it started to trace after the role got full initialized, but requests reached the instance before this). I can clearly see from the event logs that the ARR installation (run as a startup task) ended after
    the first requests could reach it.
    To summarize, when scaling up the following happens with the second, initializing instance:
    The instance's state is correctly shown on the Management Portal and it's busy, starting.
    The instance still gets requests before it gets fully started.
    After the instance is fully started everything works normally.
    Is this normal?

  • BO XI 3.1 SP3 Auto Scale, how it works...

    Hi,
    I thought that the new feature: "auto scale" in graphs (SP3), helps us to create X-Y graphs with the same scale...
    But I don't know how I have to use...
    Could anybody give me a example?
    Thanks

    Hi,
    What kind of SSO authentication are you trying to set up? (AD, LDAP,...)
    I think it's AD regarding your command line.
    But be aware that in SSO, you don't need to configure the command line to run the client.
    Have a look at the following guide.
    [Configuring Manual Kerberos Authentication and-or SSO in Distributed Environments with XI 3.1 SP3.pdf|https://bosap-support.wdf.sap.corp/sap/support/sapnotes/public/services/attachment.htm?iv_key=002007204200000183782010&iv_version=0005&alt=2BCE4CB10DF674B172F4F3F7B32A284F49333135358877720E883731B332AF34CACD2AB52C0A2C8DCACA09084EF4CB494E4E0F2ECE8E2F89772908C9CE70CD2DF77675F7F2D1750C09514BCECFCFCE4C8DCF4BCC4DB5F575F4F4F3F57771F571F6F70B01B25D83D4120B0A722092A599504EB16D715E3E00&iv_guid=DF838310BFAAE8F1B486001A64C54696]
    Regarding accessing CMC with SSO, it's not recomended at all as if you break this access, than you can't connect anymore to the CMC and modify settings.
    Regards,
    Philippe
    Edited by: Philippe Tavares on Feb 15, 2011 4:11 PM

  • Adobe Cloud Services

    How comes Adobe Cloud services is not available to people in Africa?

    There is a special sub forum for iOs Reader:
    http://forums.adobe.com/community/adobe_reader_forums/ios?view=discussions

  • Unable to connect to VM's in new cloud service via express route

    We have changed our express route setup, initially we had an express route via London, but we have added a second one via Amsterdam and removed the one via London. All existing and new vm's in the different vnet's have connection to our local datacenter,
    but as soon as we create vm's in a new cloud service the published routes don't seem to be picked up and the machine are only reachable in their local vnet on azure.
    Does anyone have an idea where to look, it looks like the route publishing does not seem to work correctly, but it is strange that new vm's in existing cloud service do work correctly. BGP peering and vnet have been provided access via the expressroute and
    all have status provisioned.

    Hi Syed,
    When I try to connect to a new vm via rdp or try to do a tracert to the machine (with firewall turned off on the vm) I don't get a response (traffic is routed via the express-route correctly). If I do a tracert to an ip on the on premise network
    from the vm in question the trace is directed to internet instead of to the on premise network via the express route.
    the new cloud services were created in the same region as the working cloud services and the vm's are also in the same vnet/subnet as the working vm's. If I delete a vm (keeping the disks) from a new cloud service and redeployed it in an existing cloud service
    I can reach it again via the internal ip.
    We have checked the route publishing and the correct routes are published to the express route/vnet.
    When I check the provisioning of the vnet's via get-azurededicatedcircuitlink all the vnet's in question are listed as provisioned.
    I'll try to remove the bpgrouting for the original expressroute this evening to see if that helps.
    kind regards
    Xander

  • Creative Cloud Service "Not avilable in your country"

    Dear Administrator,
    We have urgent business needs for using Adobe Creative Cloud services. Our company resides in Türkiye(TURKEY +90 Area Code).
    We would like to join the mebership system for minimum 50 users/per year business plan on Creative Cloud but referral warns "Adobe does not have services to/in your country.
    So, my first question is;
    Will it be a live datacenter servicing business clients in TURKEY soon ?
    Or in a way of  solution - Shall i try to participate services in the name of an international company which we are part of.(Because we even cannot manage to enter payment site ( restiricted zone from tr-TR) But we can from England.)
    Secondly;
    If we subscribe services using method mentioned above , will it be a problem after payment and participate in services where our business groups head office in England and using services from TURKEY (Telecoms Network) and deploying applications (CS6 or Adobe pro XI etc) on internal network .
    Is there a possibility that happens to occur  restrictions on use of applications or activation problems maybe ?
    Urgent help and assistance needed!

    Adobe Resellers http://adobedealreg.force.com/PartnerSearch?lang=en
    -http://www.adobe.com/content/dam/Adobe/en/products/creativecloud/cc/pdfs/cc-availability-m atrix.pdf

  • We have concerns with Adobe Acrobat DC - document cloud services and how it relates to HIPAA.  Has anyone else worked through these issues?

    We have questions relating to the Document Cloud services offered with Acrobat DC.  We are considering disabling the cloud feature to stay in compliance with HIPAA.  Has anyone else worked through these issues?

    Hi mikem82897618,
    You can refer the following link to know what can be done with Adobe Document Cloud Services when working in Acrobat DC:
    Store Files Online, Share & Access From Anywhere | Adobe Acrobat.com
    Also, visit this link to know more on how Adobe Document Cloud services are compliant with HIPAA security standards:
    E-signatures vs digital signatures | eSign services from Adobe
    Please specify more details on the same as why you would need to disable the Cloud feature.
    What and where exactly are you facing the trouble?
    Let me know.
    Regards,
    Anubha

Maybe you are looking for

  • Report viewer issue

    I have an SSRS report that is being displayed in a custom asp .net report application through the ReportViewer control (VS2010). The SSRS report renders fine on the Report Server and through Report Manager, but in the Report Application, the column w

  • Nokia 6700 supported by Nokia Multimedia Transfer

    Hello All, Please note that Nokia 6700 device is supported by nokia multimedia Transfer for mac http://europe.nokia.com/support/download-software/nokia-multimedia-transfer/download Download Nokia Multimedia Transfer Connect the device and select PC s

  • Session Conflict

    Hello, I build a game in asp.net with vb. It is a simple survey. There is questions and 4 answer for each question. You can pass or 50% for once. Some times the game conflicts. When a user plays the game in the middle of questions it jumps to an othe

  • Price guide for selling my MacBook Pro mid-2009 17"

    Hey everyone, Does anyone have information on how much I could sell my Mac Book Pro 17" mid-2009 3.06 GHz Intel Core 2 Duo with 16Gb Ram? It has a brand new logic board and battery. Or if you know where I could find out this information? Just trying

  • New to daisy chaining my external hard drives need some guidence on this

    I have 2; 3.5 inch 7200rpm 1TB hard drives, & bought some external enclosures with firewire ports for them. what would the steps be to set them up? some things I cannot seem to get a clear answer for even after Google searches is; say I have drive A