Azure Storage Maximum Threshold for alert

Hi Azure team,
Azure Storage:
What is the maximum threshold after which azure will send notification when maximum capacity is going to be reached?
Thanks,

You can create a rule based upon capacity. The units has to be in bytes. So if you wanted to get an Alert when the storage account is at 80% of 1 TB. you would set the value to 858993459. Note not to forget to enable the storage account for Capacity Alerts.
Johnny Coleman [MSFT] Any code posted to this Forum is [As-Is] With No Warranties

Similar Messages

  • Encode shared key signature in Azure Storage Queue

    I'd like to use Azure storage REST api for queue, and in order to add the authentication infomation to the http header, I have to get the shared key.
    From http://msdn.microsoft.com/en-us/library/azure/dd179428.aspx, I saw the string to sign like this:
    I found it was very strange, I prefer it's C# code, and some place "\n" missed and so is "+", as I made up all stuffs, the authentication still failed, I do not know why, here is what I did:
    using (var httpClient = new HttpClient())
    var verb = "PUT";
    var contentEncoding = "";
    var contentLanguage = "";
    var contentLenght = "0";
    var contentMd5 = "";
    var contentType = "text/plain; charset=utf-8";
    var date = DateTime.UtcNow.ToString("R", CultureInfo.InvariantCulture);
    var ifModifiedSince = "";
    var ifMatch = "";
    var ifNoneMatch = "";
    var ifUnmodifiedSince = "";
    var range = "";
    var canonicalizedHeaders = "";
    var canonicalizedResource = "";
    var stringToSign = verb + "\n" + contentEncoding + "\n" + contentLanguage + "\n" + contentLenght + "\n" +
    contentMd5 + "\n" + contentType + "\n" + date + "\n" + ifModifiedSince + "\n" +
    ifMatch + "\n" + ifNoneMatch + "\n" + ifUnmodifiedSince + "\n" + range + "\n" +
    canonicalizedHeaders + "\n" + canonicalizedResource;
    httpClient.BaseAddress = new Uri(string.Format("https://{0}.queue.core.windows.net", accountName));
    httpClient.DefaultRequestHeaders.Clear();
    httpClient.DefaultRequestHeaders.Add("Authorization", string.Format("SharedKey {0}:{1}", accountName, Hash(accountKey, stringToSign)));
    httpClient.DefaultRequestHeaders.Add("x-ms-date", date);
    var result = httpClient.PutAsync("myqueue", new StringContent(""));
    result.Wait();
    static string Hash(string message, string secret)
    var messageBytes = Encoding.UTF8.GetBytes(message);
    var secrectBytes = Encoding.UTF8.GetBytes(secret);
    using (var hash = new HMACSHA256(secrectBytes))
    var hashMessage = hash.ComputeHash(messageBytes);
    return Convert.ToBase64String(hashMessage);
    Anyone can help me?

    Hi,
    Please try this code, I tested it on my side, it works fine.
    String accountName = "***";
    String accountKey = "***+**";
    String signature = "";
    String requestMethod = "POST";
    String urlPath = String.Format("{0}/messages", "inputtext");
    String storageServiceVersion = "2012-02-12";
    String dateInRfc1123Format = DateTime.UtcNow.ToString("R", CultureInfo.InvariantCulture);
    String messageText = String.Format(
    "<QueueMessage><MessageText>{0}</MessageText></QueueMessage>", "test2");
    UTF8Encoding utf8Encoding = new UTF8Encoding();
    Byte[] messageContent = utf8Encoding.GetBytes(messageText);
    Int32 messageLength = messageContent.Length;
    String canonicalizedHeaders = String.Format(
    "x-ms-date:{0}\nx-ms-version:{1}",
    dateInRfc1123Format,
    storageServiceVersion);
    String canonicalizedResource = String.Format("/{0}/{1}", accountName, urlPath);
    String stringToSign = String.Format(
    "{0}\n\n\n{1}\n\n\n\n\n\n\n\n\n{2}\n{3}",
    requestMethod,
    messageLength,
    canonicalizedHeaders,
    canonicalizedResource);
    using (HMACSHA256 hmacSha256 = new HMACSHA256(Convert.FromBase64String(accountKey)))
    Byte[] dataToHmac = System.Text.Encoding.UTF8.GetBytes(stringToSign);
    signature = Convert.ToBase64String(hmacSha256.ComputeHash(dataToHmac));
    String authorizationHeader = String.Format(
    CultureInfo.InvariantCulture,
    "{0} {1}:{2}",
    "SharedKey",
    accountName,
    signature
    Uri uri = new Uri(string.Format("https://{0}.queue.core.windows.net/", accountName) + urlPath);
    HttpWebRequest request = (HttpWebRequest)WebRequest.Create(uri);
    request.Method = requestMethod;
    request.Headers.Add("x-ms-date", dateInRfc1123Format);
    request.Headers.Add("x-ms-version", storageServiceVersion);
    request.Headers.Add("Authorization", authorizationHeader);
    request.ContentLength = messageLength;
    using (Stream requestStream = request.GetRequestStream())
    requestStream.Write(messageContent, 0, messageLength);
    using (HttpWebResponse response = (HttpWebResponse)request.GetResponse())
    String requestId = response.Headers["x-ms-request-id"];
    Also, you could see this page:http://convective.wordpress.com/2010/08/18/examples-of-the-windows-azure-storage-services-rest-api/
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to Delete Directory from Azure Storage Account?

    Hi All,
    Currently i am working on Microsoft Windows Azure (SaaS - Application).
    I am facing a problem with the AzureStorage.
    I want to delete a whole directory from the Azure Storage Account.
    For Example : I want to delete following Directory(14).
    https://azurestorage.blob.core.windows.net/blobname/Images/14
    Any help will be appriciated. 
    Thanks in Advance.
    Rakesh T. Gupta, Web Engineer, Ahmedabad, India

    Hi Rakesh,
    We have not heard you in days. Have you managed to write the code to delete a blob directory?
    If not, I'd like to share my code for your reference:
        public void RemoveBlobDirectory()
            string containerName = "blobname";
            string directoryName = "blobname/Images/14/";
            var storageAccount = CloudStorageAccount.Parse("UseDevelopmentStorage=true");
            var blobStorage = storageAccount.CreateCloudBlobClient();
            // Ensure the container is exist.
            var blobContainer = blobStorage.GetContainerReference(containerName);
            blobContainer.CreateIfNotExist();
            foreach (IListBlobItem item in blobStorage.ListBlobsWithPrefix(directoryName))
                if (item.GetType() == typeof(CloudBlob) || item.GetType().BaseType == typeof(CloudBlob))
                    ((CloudBlob)item).DeleteIfExists();
    Please note that the ListBlobsWithPrefix method will scan through all blob entities in order to search the result. So it will potentially have low performance when you have large number of blobs in you storage account.
    Thanks,
    Wengchao Zeng
    Please mark the replies as answers if they help or unmark if not.
    If you have any feedback about my replies, please contact
    [email protected]
    Microsoft One Code Framework

  • How to set same domain name for Azure Storage and Hosted Service

    I have a web application running on azure and using azure storage with blob. My application allows to edit html files that are in the azure storage to save with another name and publish them later. I am using javascript to edit the html content that
    I display in an Iframe but as the domain of my web application and the html that I try to edit are not the same, I got and this error "Unsafe JavaScript
    attempt to access frame with URL "URL1" from frame with URL "URL2". Domains, protocols and ports must match".
    I have been doing some research about it and the only way to make it work is to have the web application and the html that I want to access using javascript under the same domain. 
    So my question is: is it possible to have the same domain name in azure for the hosted service and the storage.
    Like this:
    hosted service: example.com
    storage: example.com
    By the way I already customize the domain names so they looks like this:
    hosted service <mydomainname>.com
    storage <blob.mydomainname>.com
    Actually I have my application running in another hosting and I have no problem there since I have a folder where I am storing the files that I am editing so they are in the same domain as the application. I have been thinking in to do the same with Azure,
    at least to have a folder where I can store the html file meanwhile I am editing it but I am not sure how much space I have in the hosted service to store temporary files.
    let me know if you have a good tip or idea about how to solve this issue.

    Hi Rodrigo,
    Though both Azure Blob and Azure applications support custom domain, one domain could have only one DNS record (in this case is CNAME record) at one time. For Steve's case, he has 3 domains, blog.smarx.com, files.blog.smarx.com and cdn.blog.smarx.com.
    > I would like to find a way to storage my html page that I need to edit under the same domain.
    For this case, a workaround will be adding a http handler in your Azure application to serve file requests. That means we do not use the actual blob url to access blob content but send the request to a http handler then the http handler gets the content
    from blob storage and returns it.
    Please check
    Accessing blobs in private container without Shared Access Secret key for a code sample.
    Thanks.
    Wengchao Zeng
    Please mark the replies as answers if they help or unmark if not.
    If you have any feedback about my replies, please contact
    [email protected]
    Microsoft One Code Framework

  • How to connect a web upload form(for video uploads) to Azure Storage / Media services?

    Hi I have a form on my Joomla site (see image under for viewers to upload and send me their videos, images and info.
    Question: 
    1.Instead of receiving this form content via email,  I just need a
    HTML and/or PHP code with tags code so I can send the content viewers give me (videos with up to 500MB) directly into my existing Azure blob.
    Seems I can also direct the form to a any database... They gave me this code sample but it was for AWS
    https://www.chronoengine.com/faqs/63-cfv4/cfv4-working-with-form-data/5196-can-i-save-files-to-aws.html
    Therefore they suggested me to reach out to Azure devs.
    I don't code so please show me the exact code HTML and/or PHP code to connect my "pre-built forms" with Azure Storage/Media services.
    kd9000

    You may refer the following links which may help you in building forms to direct to azure table/blob storage:
    http://www.joshholmes.com/blog/2010/04/15/creating-a-simple-php-blog-in-azure/
    http://stackoverflow.com/questions/27464839/use-form-post-php-to-upload-a-file-to-azure-storage-and-return-the-link
    Regards,
    Manu

  • UCCX 7 - Change alert threshold for disk space

    We record every call into our UCCX 7 platform which makes my backups quite large. This causes an alert every night when the backups run from CRSAdmin stating that the server is running low on disk space and needs at least 40 GB free for B&R to run.
    Is there a way to adjust this threshold?
    Thanks

    Pablo, thanks for the reply! I checked the event viewer on the UCCX servers and they are clean. This is not a Windows alert. This alert is coming from CRSAdministrator and is routed to the email address I have configured in the email subsystem (Appadmin / Subsystems / email.)
    The server is not actually getting low on disk space when this alert is generated, it is simply going below the 40 GB free threshold that is built somewhere in the UCCX application. This is what the actual email looks like:
    From CRSAdministrator@AUS-SVR-ICD-01
    WARNING: CRS server AUS-SVR-ICD-01(172.21.X.X) is running low in free disk space
    **DO-NOT-REPLY-TO-THIS-EMAIL**
    CRS required that you have at least 43377659904 bytes available in drive C:, otherwise B&R will fail in operating. Please free up some disk space if required.
    -Cisco Customer Response Application Administration.
    So maybe the solution is remove the email address I have configured in the email subsystem for alerting and set up a syslog server instead for alarms?
    Brett

  • Alerting Metric thresholds for Notification rules

    Hi Guys,
    Can i edit the Response Time (msec) metric under the Listener Availability Notification rule to alert only if the response time is breaching the alerting threshold for 5 minutes,as in for 5 consecutive occurences when this metric is pooled ?
    For example we can execute dbms_server_alert.set_threshold(.....
    consecutive_occurrences=>5
    for a metric like redo_generated_sec
    Can we do this for Response Time (msec) metric under the Listener Availability Notification rule ?
    I am also not seeing this metric listed under v$METRICNAME
    Any ideas...
    Regards,
    Swanand

    I am not sure if this is related, but you can added Metrics for the Service Test associated with the Beacon via the Monitoring Setting.
    Example for homepage:
    Select All Targets > EM Website > Metric and Policy Settings >Tests >Monitoring Settings of homepage
    Then Add/edit Metric Thresholds and Collection settings accordingly.

  • Learning resource for Azure Storage

    Dive Deep into Networking Storage and Disaster Recovery Scenarios
    http://www.microsoftvirtualacademy.com/training-courses/dive-deep-into-networking-storage-and-disaster-recovery-scenarios?WT.mc_id=13361-ITP-ming-man-chan-mvp-content
    chanmm

    There are some other great resources available.
    1)  Azure storage documentation: It's just awesome. you will get all latest bits here.
         http://azure.microsoft.com/en-us/documentation/services/storage/
    2)  Checkout MVA course for azure storage
    http://www.microsoftvirtualacademy.com/training-courses/windows-azure-storage-design-and-implementation-jump-start
    http://www.microsoftvirtualacademy.com/training-courses/windows-azure-storage-design-and-implementation-jump-start
    There are some cool examples also there on Azure Storage from the following link.
    https://azurestoragesamples.codeplex.com/
    Jalpesh Vadgama,  http://www.dotnetjalps.com

  • Azure account maximum capacity varies.

    I see that Azure initially provided 100TB per account. After 8 June 2012 they have increased the account capacity to 200TB for accounts created after the specified
    date. <o:p></o:p>
    Also the capacity storage schema in the msdn website indicates that only capacity used by the account is stored using statistics.
    <o:p>There is API to query for used account capacity.</o:p>
    If as a user I create azure accounts over the years, how do I query for the total capacity of my accounts if I hold multiple accounts whose total
    capacity may be different depending on when I created it. I need this info to determine how much capacity is left in each of the accounts I hold.
    So how do I query the maximum capacity of my accounts from Azure ? 
    Is there an API for query the maximum capacity of an Azure account ?

    Hi,
    Base on my experience, it may have not the method directly get the how much capacity is left in the accounts. But we could use the compromised approach. We could get the blob storage and table storage usage capacity.
    Please see this blog (http://blogs.msdn.com/b/windowsazurestorage/archive/2011/08/03/windows-azure-storage-metrics-using-metrics-to-track-storage-usage.aspx?Redirected=true
    ) and this docs (http://msdn.microsoft.com/library/azure/hh343258.aspx). And I suggest you could refer to Gaurav Mantri's reply (http://stackoverflow.com/a/17058225
    ) and this powershell code sample (http://www.amido.co.uk/richard-slater/windows-azure-storage-capacity-metrics-with-powershell/ &
    http://michaelwasham.com/2011/09/28/windows-azure-storage-analytics-with-powershell/).
    Hope it helps.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How do I improve performance while doing pull, push and delete from Azure Storage Queue

           
    Hi,
    I am working on a distributed application with Azure Storage Queue for message queuing. queue will be used by multiple clients across the clock and thus it is expected that it would be heavily loaded most on the time in usage. business case is typical as in
    it pulls message from queue, process the message then deletes the message from queue. this module also sends back a notification to user indicating process is complete. functions/modules work fine as in they meet the logical requirement. pretty typical queue
    scenario.
    Now, coming to the problem statement. since it is envisaged that the queue would be heavily loaded most of the time, I am pushing towards to speed up processing of the overall message lifetime. the faster I can clear messages, the better overall experience
    it would be for everyone, system and users.
    To improve on performance I did multiple cycles for performance profiling and then improving on the identified "HOT" path/function.
    It all came down to a point where only the Azure Queue pull and delete are the only two most time consuming calls outside. I can further improve on pull, which i did by batch pulling 32 message at a time (which is the max message count i can pull from Azure
    queue at once at the time of writing this question.), this returned me a favor as in by reducing processing time to a big margin. all good till this as well.
    i am processing these messages in parallel so as to improve on overall performance.
    pseudo code:
    //AzureQueue Class is encapsulating calls to Azure Storage Queue.
    //assume nothing fancy inside, vanila calls to queue for pull/push/delete
    var batchMessages = AzureQueue.Pull(32); Parallel.ForEach(batchMessages, bMessage =>
    //DoSomething does some background processing;
    try{DoSomething(bMessage);}
    catch()
    //Log exception
    AzureQueue.Delete(bMessage);
    With this change now, profiling results show that up-to 90% of time is only taken by the Azure Message delete calls. As it is good to delete message as soon as processing is done, i remove it just after "DoSomething" is finished.
    what i need now is suggestions on how to further improve performance of this function when 90% of the time is being eaten up by the Azure Queue Delete call itself? is there a better faster way to perform delete/bulk delete etc?
    with the implementation mentioned here, i get speed of close to 25 messages/sec. Right now Azure queue delete calls are choking application performance. so is there any hope to push it further.
    Does it also makes difference in performance which queue delete call am making? as of now queue has overloaded method for deleting message, one which except message object and another which accepts message identifier and pop receipt. i am using the later
    one here with message identifier nad pop receipt to delete message from queue.
    Let me know if you need any additional information or any clarification in question.
    Inputs/suggestions are welcome.
    Many thanks.

    The first thing that came to mind was to use a parallel delete at the same time you run the work in DoSomething.  If DoSomething fails, add the message back into the queue.  This won't work for every application, and work that was in the queue
    near the head could be pushed back to the tail, so you'd have to think about how that may effect your workload.
    Or, make a threadpool queued delete after the work was successful.  Fire and forget.  However, if you're loading the processing at 25/sec, and 90% of time sits on the delete, you'd quickly accumulate delete calls for the threadpool until you'd
    never catch up.  At 70-80% duty cycle this may work, but the closer you get to always being busy could make this dangerous.
    I wonder if calling the delete REST API yourself may offer any improvements.  If you find the delete sets up a TCP connection each time, this may be all you need.  Try to keep the connection open, or see if the REST API can delete more at a time
    than the SDK API can.
    Or, if you have the funds, just have more VM instances doing the work in parallel, so the first machine handles 25/sec, the second at 25/sec also - and you just live with the slow delete.  If that's still not good enough, add more instances.
    Darin R.

  • Azure TechNet Guru Announced for June 2014

    The Results are in! And the winners of the TechNet Guru Competition June 2014 have been posted on the
    Wiki Ninjas Blog.
    Below is a summary, heavily trimmed to fit the size restrictions of forum posting.
     BizTalk Technical Guru - June 2014  
    Steef-Jan Wiggers
    BizTalk Server: Custom Archiving
    TGN: "This one was my favorite this month. Archiving is a topic that is brought up often. Well done explaining it simply and how to do it according to best practice"
    Sandro Pereira: "Love the topic, well explain and with everything you need, my favorite."
    Mandi Ohlinger: "Another great addition to the Wiki. "
    boatseller
    BizTalk: Reducing and Consolidating WCF Serialization Schema Types
    TGN: "Very good, keeping the code clean, and only referencing what you need and consolidate it is important!"
    Mandi Ohlinger: "Great solution to somewhat-annoying behavior. Nice addition to the Wiki!"
    Sandro Pereira: "Great article."
    Murugesan Mari Chettiar
    How to Implement Concurrent FIFO Solution in BizTalk Server
    Ed Price: "Incredibly thorough in your explanations! Great formatting. Good job!"
    TGN: "First in, first out. Great article Murugesan!"
    Sandro Pereira: "Good additional to the TechNet Wiki, good work."
     Forefront Identity Manager Technical Guru - June 2014  
    Remi Vandemir
    Custom Reports in FIM2010R2
    AM: "Great step-by-step guide for generating custom reports. Thanks for taking the time to put this together."
    PG: "Nice article, in an area that is less known!"
    Søren Granfeldt: "Very comprehensive."
    Ed Price: "Great job on the intro, and a lot of images really help clarify all the steps!"
    GO: "Thank you "
    Eihab Isaac
    FIM 2010 R2: Review pending export changes to Active Directory using XSLT
    Ed Price: "Great introduction, great steps, and great job on the image and code formatting!"
    GO: "An introduction, a sample code, images, a TOC and a conclusion. Nothing here to preserve the GOLD medal!"
    PG: "Nice article!"
    Søren Granfeldt: "Nice and precise"
    Scott Eastin
    A Practical Alternative to the PeopleSoft
    AM: "Thank you for sharing. Great (and probably superior) alternative for those using PeopleSoft as import-only data source."
    GO: "Amazing article, love it so much"
    PG: "Would like to see more elaborated details in this article."
    Søren Granfeldt: "A little more technical stuff would be nice"
    Ed Price: "Some good community collaboration in removing blog-like personalization. This is a great topic with some good holistic thinking!"
     Microsoft Azure Technical Guru - June 2014  
    Mr X
    Configuration of WATM (Windows Azure Traffic Manager) for Web Portals hosted
    on Azure VMs
    JH: "Two simple words: Love it! The detailed explanation on how Traffic Manager works is awesome."
    Ed Price: "Wow! Incredibly well written, with beautiful diagrams and a great use of images and tables! Great topic!"
    GO: "This is a great article! Thanks Mr.X"
    Mr X
    How to use Windows Azure as Traffic Manager for Web portals
    hosted in multiple on-premise datacenters
    JH: "Very detailed! Great explanation at the beginning followed by a good step-by-step guide."
    Ed Price: "A much needed article! Great job on the formatting and images!"
    GO: "Thanks again, MR.X"
    Mr X
    How to connect Orchestrator to Windows Azure
    GO: "I really enjoyed reading this article, clever and well written. Lovely done!"
    JH: "Great article! I especially love the amount of pictures provided in the article."
    Ed Price: "Good procedural article! Great use of images!"
     Microsoft Visio Technical Guru - June 2014  
    Mr X
    How to open Visio files without Visio
    AH: "This Article is pretty basic and lacks details. Visio Viewer doesn't just open in IE but also in Outlook and File explorer. The writer should include the link to http://blogs.office.com/2012/11/28/download-the-free-microsoft-visio-viewer/
    this blog which has lot more details "
    Ed Price: "Good. I think the SEO on the title will drive more awareness of the Visio Viewer."
    GO: "Thanks you Mr.X! Again a great article!"
     Miscellaneous Technical Guru - June 2014  
    Ed Price - MSFT
    Yammer: Announcements Feature
    TGN: "Wow, not only is this a good way on how to write annoncments on Yammer, but in generel. Really, really great write-up Ed! T"
    GO: "Tord says on the comment section: "Very nice article, Ed. I really enjoyed reading it and you had a great set of tips. Thanks for sharing!".. I only can respond AMEN! Thanks Ed!"
    Margriet Bruggeman: "Good discussion of announcements feature."
    Anthony Caragol
    Backing Up and Restoring Lync 2013 Contacts
    Margriet Bruggeman: "Short & Sweet"
    GO: "Great article, but I'm missing, examples, images, definitions etc for a huge section like "backup and restore""
    TGN: "Very good, Lync has eaten up the market and is a key product in most companies, articles like this is very valuable. Great work Anthony!"
     SharePoint 2010 / 2013 Technical Guru - June 2014  
    Geetanjali Arora
    SharePoint Online : Working with People Search and User Profiles
    Benoît Jester: "A very good article, a must-read for those interested by SharePoint Online and the use of search and user profile API."
    Jinchun Chen: "Excellent. Just a tip, if you would like to improve the performance, please use the Search Service to search user profiles"
    Craig Lussier: "Good walkthrough and code example for getting started with People Search!"
    Margriet Bruggeman: "Good starter for working with search and profiles"
    Jaydeep Mungalpara
    Creating Bookmarks in Wiki Pages - SharePoint Rich Text Editor Extension
    Margriet Bruggeman: "Really cool! In the past, I was actually looking for this and its a nice implementation of this functionality. This article gets my vote!"
    Craig Lussier: "Great solution for extending out of the box functionality. I like the synergy between the TechNet Wiki and TechNet Gallery!"
    GO: "Simple but powerfull. We should all take an example about how this article has been written. This article has a TOC, headings and even a code! Well done!"
    Jinchun Chen: "Nice. "
    Benoît Jester: "A simple button which can save a lot of time!"
    Dan Christian
    PowerShell to copy or update list items across SharePoint sites and farms
    GO: "The best artice for June! Thanks Dan, you deserve the GOLD medal!"
    Benoît Jester: "A good article with useful scripts, as they can be used fior many scenarios (data refresh, migration tests, ...)"
    Jinchun Chen: "Good and low-cost solution. To be automatic, we can use EventHandle instead. "
    Craig Lussier: "Nice PowerShell script solution and explanation of the scenario. Consider using functions with parameters for easier reuse so input parameters are not hard coded."
    Margriet Bruggeman: "This script can be useful, although typically migration scenarios are more complex than this. Having said that, I probably end up using this script some time in in the future"
     Small Basic Technical Guru - June 2014  
    litdev
    Small Basic: Sprite Arrays
    Ed Price: "An important topic that's well described with fantastic examples! Great article!"
    Michiel Van Hoorn: "Great starter for Sprite Fundamentals and how to handle them. Briljant start point for greating you 2D shooter"
    Jibba Jabba
    Small Basic - Monthly Challenge Statistics
    Ed Price: "Jibba Jabba brings us astonishing insights and data about LitDev's Small Basic Monthly Challenges!"
    RZ: "This is very nicely done and showed all the statistics visually"
    Nonki Takahashi
    Small Basic: Challenge of the Month
    RZ: "This is very nicely done and organized all challenges of the month in one place"
    Ed Price: "Although this is very basic, it's incredibly helpful to get all these in one list and to access all the great challenges!"
    Michiel Van Hoorn: "Good explainer on  fundamental structures."
     SQL BI and Power BI Technical Guru - June 2014  
    Anil Maharjan
    Using Power Query to tell your story form your Facebook Data
    Jinchun Chen: "Interesting. I liked this best"
    PT: "Plenty to like here" 
    Ed Price: "Great! I love to see Power Query articles like this! Great formatting and use of images!"
    Tim Pacl
    SSRS Expressions: Part 1 - Program Flow
    PT: "A very comprehensive article about program flow expressions. Nice job. I'm sure many will benefit from this article. Just a little feedback about some terminology that could be more clear: The entire statement that
    is typically used to set a property value for an object in an SSRS report is an "expression". Each of the three programming constructs you've mentioned (e.g. IIF, SWITCH & CHOOSE) are "functions" and not expressions or statements."
    Jinchun Chen: "Perfect! Good article for SSRS newbie." 
    Ed Price: "The table and images help bring it more value. Great job!"
    Anil Maharjan
    How to Schedule and Automate backups of all the SSAS catalogs within the
    Server Instance
    PT: "This is a very useful article about automating multiple Analysis Services database backups using an SSIS package and the SQL Server Agent. Nice job."
    Jinchun Chen: "Good." 
    Ed Price: "Good use of images. Could be improved with better code formatting. Good job!"
     SQL Server General and Database Engine Technical Guru - June 2014  
    Shanky
    SQL Server: What does Column Compressed Page Count Value Signify
    in DMV Sys.dm_db_index_physical_stats ?
    DB: "Interesting and detailed"
    DRC: "• This is a good article and provides details of each and every step and the output with explanation. Very well formed and great information. • We can modify the create table query with “DEFAULT VALUES". CREATE TABLE [dbo].[INDEXCOMPRESSION](
    [C1] [int] IDENTITY(1,1) NOT NULL, [C2] [char](50) NULL DEFAULT 'DEFAULT TEST DATA' ) ON [PRIMARY]"
    GO: "Very informative and well formed article as Said says.. Thanks for that great ressource. "
    Durval Ramos
    How to get row counts for all Tables
    GO: "As usual Durva has one of the best articles about SQL Server General and Database Engine articles! Thanks, buddy!" "
    Jinchun Chen: "Another great tip!"
    PT: "Nice tip" 
    Ed Price: "Good topic, formatting, and use of images. This would be far better if the examples didn't require the black bars in the images. So it would be better to scrub the data before taking the screenshots. Still a good article. Thank
    you!"
     System Center Technical Guru - June 2014  
    Prajwal Desai
    Deploying SCCM 2012 R2 Clients Using Group Policy
    Ed Price: "Great depth on this article! Valuable topic. Good use of images."
    Mr X
    How to introduce monitoring and automatic recovery of IIS application
    pools using Orchestrator
    MA: "Good job Mr X, However I would like to see this runbook integrated as a recovery task with Operations Manager IISapppools Monitors in order to maintain a standard way of notifications and availability reporting."
    Ed Price: "Good formatting on the images, and great scenario!"
    Prajwal Desai
    How to deploy lync 2010 using SCCM 2012 R2
    Ed Price: "Great job documenting the entire process!!!"
     Transact-SQL Technical Guru - June 2014  
    Saeid Hasani
    T-SQL: How to Generate Random Passwords
    JS: "I loved the article, well structured, to the point. Not missing any caveats that might occur, really good in the end. I would suggest changing the function to accept a whitelist / blacklist as well as a length of
    the password to be created. This would be the cherry on the pie :-)"
    Samuel Lester: "Very nice writeup for a real world problem!"
    Richard Mueller: "Clever and apparently well researched. I liked the detailed step by step explanations."
    Jinchun Chen: "Excellent!"
    Manoj Pandey: "A good and handy utility TSQL that I can use and levarage if I have to use similar feature in future."
    Hasham Niaz
    T-SQL : Average Interval Length
    Richard Mueller: "A good article, but I need more explanation of the concepts."
    Manoj Pandey: "A handy TSQL script that I can use and levarage if I have to use similar feature in future."
    Visakh16
    T-SQL: Retrieve Connectionstring Details from SSIS Package
    Manoj Pandey: "Good shortcut by using TSQL with XML to read metadata information from SSIS XML file."
    Samuel Lester: "Handy trick, thanks for posting!"
    Richard Mueller: "Good code, but more explanation needed. Could use a See Also section."
     Visual Basic Technical Guru - June 2014  
    The Thinker
    Better to Ask for forgiveness then permission
    Richard Mueller: "Good use of images and code. The humorous title might be better in a blog."
    MR: "Great topic!"
    GO: "Well, to be honnest, many people worked on that article, but still, the owner "the thinker" should receive the credits! muchos gracias "The Thinker" for the Most Revised Article"
     Visual C# Technical Guru - June 2014  
    Jaliya Udagedara
    Entity Framework Code First - Defining Foreign Keys using Data Annotations
    and Fluent API
    Ed Price: "Wow. Good descriptions, great code snippets, and great job highlighting sections on your images!"
    GO: "Thank you."
     Wiki and Portals Technical Guru - June 2014  
    XAML guy
    History and Technology Behind the TechNet Wiki Ninja Belt Calculator
    Ed Price: "It's amazing to see all the details of what this tool does. Great job on the descriptions and formatting the images and text!"
    Richard Mueller: "Great documentation. Good links to explain everything."
    GO: "I love your articles XAML guy! Always clear and always a pleasure to read! Thanks for you help and commitment for this tool."
    Durval Ramos
    HTML5 Portal
    Ed Price: "This is great to see this HTML5 resource!"
    Richard Mueller: "A great contribution to our collection of portals"
    GO: "The HTML5 Portal is A-W-E-S-O-M-E !"
    João Sousa
    ASP.NET Portal
    Ed Price: "Good job on this portal! The Return to top links are helpful!"
    Richard Mueller: "More should be done to distinguish this portal from
    here."
    GO: "Thanks Joao!"
     Windows Phone and Windows Store Apps Technical Guru - June 2014  
    Dave Smits
    Theming of your application
    Peter Laker: "Another great article from the mighty Dave. Very useful. Not sure if MS like us want us to work around the accents so much though ;)"
    Ed Price: "Very useful topic and great formatting on the code! Could benefit from more explanation on the code toward the bottom and a See Also section. Great article!"
    saramgsilva
    Creating Windows Phone and Window 8.1 applications using Microsoft App Studio
    Peter Laker: "A great introduction! Nice walkthrough, and plenty to look at!"  
    Ed Price: "This is good. I love the narrative and use of images! Good conclusion!"
    Carmelo La Monica
    Create Universal Application with Windows Phone App Studio (en-US).
    Peter Laker: "Sensational article. A real attention grabber and written very clearly."  
    Ed Price: "Fantastic job on the narrative and images. Some amazing articles this month!"
     Windows Presentation Foundation (WPF) Technical Guru - June 2014  
    Magnus (MM8)
    WPF: How To Tab Between Items In a ListBox
    KJ: "This article seemed very useful to me. The kind of thing that I might need and here's the answer."
    GO: "Thanks for that great article!."
    Ed Price: "Another amazing article from Magnus! Great job on the topic choice (very needed scenario), formatting, code, explanations, and See Also section. Fantastic article!"
    Sugumar Pannerselvam
    Lets forget about limitations and temprorary fix... Think about 4.5 features
    KJ: "Wish there were code samples and more flushed out scenarios"
    GO: "Why second place? the layout and way to explain didn't convince me. Doesn't mean that the article is bad. The article is awesome; but it's missing something."
    Ed Price: "Short and sweet. Could benefit from adding in some code snippet examples and images. Good topic choice."
     Windows Server Technical Guru - June 2014  
    Mr X
    DHCP on Windows Servers – Why are the expired IP addresses not getting re-assigned?
    JM: "This is an excellent article, thanks for your contribution."
    Richard Mueller: "Important information with good explanation. Needs a See Also section."
    Philippe Levesque: "Good article ! I like how it's explained versus Windows Server. An image with the DHCP's process could be a good addition for reference. (DHCP OFFER, DHCP ACK, etc..)"
    Mr X
    How to force a DHCP database cleanup for expired leases in a specific scope
    GO: "I'm actually thinking that nobody can defaut you Mr.X"
    Philippe Levesque: "Good article ! I would add that changing the lease time to be shorted could help too."
    JM: "A very good article, however you might consider adding this content as a section in your article about expired IP addresses in DHCP"
    Richard Mueller: "More good information. Should be linked to the other DHCP article."
    GL: "This is OK but a better solution for a highly utilized DHCP scope would be to shorten the lease time and/or configure a superscope."
    Hicham KADIRI
    Windows Server Core 2012 R2 - Initial configuration
    GL: "This is good required information. I would really like to see information added about how to add a server role. You might consider providing PowerShell alternatives to the netsh and other commands."  
    JM: "This is a great to-the-point article on how to configure a Core install of Windows Server, nice work."
    Richard Mueller: "A great collection of useful tools. Some could use images, more detail, or examples. The example sections could be added to the Table of Contents."
    GO: "Well, our new french MVP! Well written Hicham! Do not forget to pray attention for the layout! It's capital for readers and judges!"
    Philippe Levesque: "I like the article, a good resumé of the command you need to do to configure a server."
    Don't forget the full version, with runners up is available
    here.
    More about the TechNet Guru Awards:
    TechNet Guru Competitions
    How it works
    #PEJL
    Got any nice code? If you invest time in coding an elegant, novel or impressive answer on MSDN forums, why not copy it over to the one and only
    TechNet Wiki, for future generations to benefit from! You'll never get archived again!
    If you are a member of any user groups, please make sure you list them in the
    Microsoft User Groups Portal. Microsoft are trying to help promote your groups, and collating them here is the first step.

    Congrats to Mr X! Also, the runner ups did a great job!
     Microsoft Azure Technical Guru - June 2014  
    Mr X
    Configuration of WATM (Windows Azure Traffic Manager) for Web Portals hosted
    on Azure VMs
    JH: "Two simple words: Love it! The detailed explanation on how Traffic Manager works is awesome."
    Ed Price: "Wow! Incredibly well written, with beautiful diagrams and a great use of images and tables! Great topic!"
    GO: "This is a great article! Thanks Mr.X"
    Mr X
    How to use Windows Azure as Traffic Manager for Web portals
    hosted in multiple on-premise datacenters
    JH: "Very detailed! Great explanation at the beginning followed by a good step-by-step guide."
    Ed Price: "A much needed article! Great job on the formatting and images!"
    GO: "Thanks again, MR.X"
    Mr X
    How to connect Orchestrator to Windows Azure
    GO: "I really enjoyed reading this article, clever and well written. Lovely done!"
    JH: "Great article! I especially love the amount of pictures provided in the article."
    Ed Price: "Good procedural article! Great use of images!"
    Also worth a mention were the other entries this month:
    Configure a custom domain name for a web site - Windows Azure (Godaddy)  by
    Saad Mahmood
    Ed Price: "Great details!"
    GO: "Well written.. Thanks for your help and article for the community"
    JH: "A very relevant topic. Would love to see more articles related to other hosting providers."
    Azure storage block blob upload from Android by
    Isham Mohamed
    JH: "Good explanation of block blobs. Unfortunately the biggest part is a code snippet provided by Microsoft. It would be great to have a real-world implementation a reader can use directly."
    Ed Price: "Good topic and commenting in the code! But a lot of this article is from other sources." 
    Ed Price, Azure & Power BI Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • In SOLUTION MANAGER, the thresholds for log attrib are not taken in account

    I’m using a SAP SOLUTION MANAGER to monitor any satellites systems.
    In this case, I have a problem with the monitoring of BDoc messages for a satellite.
    I’m declaring some threshold for BDocs in error status and anothers thresholds for BDocs in waiting for response in SOLMAN.
    Unfortunately, the thresholds are not taken in account.
    Here, you should find the threshold in the setup monitoring:
          green to yellow: 10
          yellow to green:   11
          red to yellow:       9
          yellow to green:   8
    When I have only 2 errors in the satellite system and 1 entrie in Waiting status ,I have 2 red alerts in SOLMAN. It is not correct because my thresholds in SOLMAN are greater than this values.
    unfortunately thresholds are only considered for performance attributes
    (see legend in the monitoring setup). For all other MTE types in CCMS
    the tresholds are not considered. The BDoc MTEs that I have used
    in my example is a log attribute.
    Please, someone could you help me to solve this problem or indicate me how to correct this problem with another tips?
    best regards
    cesario

    Sorry, but what you say is not a good idea, since it would mean changing a standard object. But what I posted is ok, using a redefinition of the method GET_DQUERY_VALUEHELPS works perfectly well for the need.
    Regards,
    Xavier

  • Can't install windows azure storage sdk on windows phone 8

    From Visual studio 2013 ultimate, I opened Nuget package manager console and typed following command for my windows phone 8 solution:
    install-package windowsazure.storage
    And I got the following error:
    install-package : Could not install package 'Microsoft.WindowsAzure.ConfigurationManager 1.8.0.0'. You are trying to install this package into a project that targets 'WindowsPhone,Version=v8.0', but the package does not contain any assembly references or
    content files that
    are compatible with that framework. For more information, contact the package author.
    At line:1 char:1
    I get the same problem using UI version of package manager as well.
    Did something change with the newer version of windows azure storage?

    The package you are trying to install does not include the Windows Phone library. Please use the following command instead:
    Install-Package WindowsAzure.Storage-Preview -Pre
    Thanks for this. It worked for me. Can I just add, for those that don't know how to run this command. You go to tools > Library package manager > package manager console then paste in the command and hit enter.
    Thanks again!

  • Cannot install Windows Azure Storage Emulator - 3.0 Error: No available SQL Instance was found

    When trying to install Windows Azure SDK for .NET (VS 2013) - 2.3 from Web Platform Installer 4.6, the install fails because  Windows Azure Storage Emulator - 3.0 (Dependency) does not install successfully.  
    Possibly relevant lines from the install log are:
    CAQuietExec:  Entering CAQuietExec in C:\WINDOWS\Installer\MSI1223.tmp, version 3.6.3303.0
    CAQuietExec:  "C:\Program Files (x86)\Microsoft SDKs\Windows Azure\Storage Emulator\WAStorageEmulator.exe" init -forcecreate -autodetect
    CAQuietExec:  Windows Azure Storage Emulator 3.0.0.0 command line tool
    CAQuietExec:  Error: No available SQL Instance was found.
    CAQuietExec:  Error 0xfffffff6: Command line returned an error.
    CAQuietExec:  Error 0xfffffff6: CAQuietExec Failed
    CustomAction RunInitialize returned actual error code 1603 (note this may not be 100% accurate if translation happened inside sandbox)
    Action ended 11:50:13: InstallFinalize. Return value 3.
    Action ended 11:50:13: INSTALL. Return value 3.
    In terms of SQL Instance, SQL LocalDB is installed and working properly.  SQL Server 2012 is also installed and working properly.

    Hi,
    It is a SDK version issue. I suggest you could remove all azure sdk form your PC and use WPI to install the latest version again.
    If you have any questions, please let me know.
    Regards,
    Will 
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to specify the storage access key for a ResourceFile?

    The azure batch tutorial shows how to put program file into a public container in a storage account and let azure batch to download them to TVMs and run.
    In real world scenario, if I don't want to use a public container or a shared access signature and want the azure batch to use a access key to access the container where my task program file is located, is it possible? How to do it?

    I see that you are conversant with the issues here but for other readers let me provide a quick review:
    The properties of a task (ICloudTask/CloudTask) include a collection of ResourceFile instances. ResourceFile instances
    map blobs in Azure Storage to local files in the Container/VM/Guest-OS.  Azure Batch copies the files from storage into the VM before the task runs and it uses the SAS (and other data) in the ResourceFile to do so.
    The ICloudTask/CloudTask.FilesToStage collection exposes the object model's mechanism for customizable file staging.
     The collection accepts instances of IFileStagingProvider which ultimately are invoked to create/augment the ResourceFile collection on the task.
    A default implementation is provided: FileToStage.
    An instance of FileToStage maps a file local to the client library to a file ultimately in the VM (indirecting through
    blob storage/SAS).  When instances of FileToStage are added to the CloudTask.FilesToStage the following occurs on Commit()/AddTask:
    A container is created in the given storage account.  The name is constructed to avoid collisions.
    The container is given a restricted SharedAccessBlobPolicy.
    All of the local files specified are uploaded to that container
    An SAS for each blob is created
    (24hr expiry)
    and a ResourceFile is constructed for each FileToStage
    The ResourceFile for each FileToStage is added to the CloudTask.ResourceFiles collection.
    FileToStage and the FilesToStage collection are intended to assist the customers that either want a shortcut around the issues of blob containers and SAS or want to control the file staging process via a custom implementation of IFileStagingProvider.
    When using the default implementation FileToStage to stage local files, care should be taken to monitor the number of containers created and the storage cost implications.
    Your concerns about SAS based methods are not directly addressed by the default implementation.  I would only note that SAS values can be re-used across tasks and jobs so the existing implementation can be used to get local data into storage and usable
    SAS values.  However, you already have these sorts of features implemented it seems and as you point out, there is the problem of SAS expiry. 
    daryl

Maybe you are looking for