URGENT: Azure Storage Table Outage

UPDATE: Problem appears to have been fixed, are affected accounts eligible for a refund due to the downtime?
Hi. I'm having trouble querying an Azure Storage table that is essential to my business operations. The queries seem to simply not be going through, and when using the Azure Storage Explorer (a 3rd party problem); I encounter the same issue. For some reason;
Azure does not seem to be responding to my storage requests. I also cannot open a support ticket with Microsoft; as our small business does not have a support agreement with Microsoft. The storage account name is chapteradviser, and it is not even possible
to query the table service for a list of tables (it keeps timing out). This seems to be a problem at the Azure datacenter, not my end. I am also an IT consultant for the company and do not have the binding authority to enter into a support agreement with Azure.
Thanks for any assistance,
- Brian Bosak
- Consulting with ChapterAdviser Llc.

Yup I see it too.. it looks like tables is really slow :(
Adding curl tracing if someone from MS is looking ..  (i redacted out the account name..)
$ curl -v --trace-time -X GET https://xxxxxxxxxx.table.core.windows.net
17:14:22.296000 * Adding handle: conn: 0x1e67e80
17:14:22.296000 * Adding handle: send: 0
17:14:22.296000 * Adding handle: recv: 0
17:14:22.296000 * Curl_addHandleToPipeline: length: 1
17:14:22.312000 * - Conn 0 (0x1e67e80) send_pipe: 1, recv_pipe: 0
17:14:22.312000 * About to connect() to xxxxxxxxxx.table.core.windows.net port 443 (#0)
17:14:22.312000 *   Trying 23.99.32.80...
17:14:25.375000 * Connected to xxxxxxxxxx.table.core.windows.net (23.99.32.80) port 443 (#0)
17:14:25.640000 * successfully set certificate verify locations:
17:14:25.656000 *   CAfile: C:\Program Files (x86)\Git\bin\curl-ca-bundle.crt
  CApath: none
17:14:25.656000 * SSLv3, TLS handshake, Client hello (1):
17:14:30.859000 * SSLv3, TLS handshake, Server hello (2):
17:14:30.875000 * SSLv3, TLS handshake, CERT (11):
17:14:30.890000 * SSLv3, TLS handshake, Server finished (14):
17:14:30.921000 * SSLv3, TLS handshake, Client key exchange (16):
17:14:30.921000 * SSLv3, TLS change cipher, Client hello (1):
17:14:30.937000 * SSLv3, TLS handshake, Finished (20):
17:14:41.937000 * SSLv3, TLS change cipher, Client hello (1):
17:14:41.953000 * SSLv3, TLS handshake, Finished (20):
17:14:41.953000 * SSL connection using AES128-SHA
17:14:41.968000 * Server certificate:
17:14:41.984000 *        subject: CN=*.table.core.windows.net
17:14:42.000000 *        start date: 2014-02-20 12:59:18 GMT
17:14:42.000000 *        expire date: 2016-02-20 12:59:18 GMT
17:14:42.031000 *        subjectAltName: xxxxxxxxxx.table.core.windows.net matched
17:14:42.046000 *        issuer: DC=com; DC=microsoft; DC=corp; DC=redmond; CN=MSIT Machine Auth CA 2
17:14:42.062000 *        SSL certificate verify ok.
17:14:42.078000 > GET / HTTP/1.1
17:14:42.078000 > User-Agent: curl/7.30.0
17:14:42.078000 > Host: xxxxxxxxxx.table.core.windows.net
17:14:42.078000 > Accept: */*
17:14:42.078000 >
17:15:35.078000 < HTTP/1.1 400 The requested URI does not represent any resource on the server.
17:15:35.093000 < Content-Length: 360
17:15:35.093000 < Content-Type: application/xml
17:15:35.093000 * Server Microsoft-HTTPAPI/2.0 is not blacklisted
17:15:35.109000 < Server: Microsoft-HTTPAPI/2.0
17:15:35.109000 < x-ms-request-id: f2e0b20e-5888-43ce-bbf0-68589e7ad972
17:15:35.109000 < Date: Sat, 09 Aug 2014 00:15:30 GMT
17:15:35.125000 <
<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<error xmlns="http://schemas.microsoft.com/ado/2007/08/dataservices/metadata">
  <code>InvalidUri</code>
  <message xml:lang="en-US">The requested URI does not represent any resource on the server.
RequestId:f2e0b20e-5888-43ce-bbf0-68589e7ad972
Time:2014-08-09T00:15:31.4896331Z</message>
</error>17:15:35.125000 * Connection #0 to host xxxxxxxxxx.table.core.windows.net left intact

Similar Messages

  • Azure Storage Table Skip Query?

    Currently azure storage table api support limited query feature, i was wondering what other queries they plan to support in future.
    I am mostly interested in skip as i am trying to show audit log of my application though table which has paging, sorting and filtering capacity, for jumping through pages i need to have skip support otherwise i need to use some workaround .
    If azure team is planning to provide this it will be very helpful.
    Thanks.

    Hi,
    Skip is not support at currently, I would suggest you vote the feedback below.
    #http://feedback.azure.com/forums/217298-storage/suggestions/1097257-support-skip-in-azure-table
    #http://feedback.azure.com/forums/217298-storage/suggestions/2914587-please-give-us-count-takelast-int-i-skip-int
    All of the feedback in feedback forums will be monitored and reviewed by the Microsoft engineering teams responsible for building Azure.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Nodejs query azure storage table by timestamp

    I can query the table with partition key by using the query:
    var query=new azure.TableQuery().where('PartitionKey ne ?','test')
    But I tried to query with the timestamp but failed.
    var query=new azure.TableQuery().where('Timestamp ge ?',azure.TableUtilities.entityGenerator.DateTime(new Date(Date.UTC(2014,11,11,00,00,00))));
    I also tried 
    var query=new azure.TableQuery().where('Timestamp ge datetime?','2014-11-11T00:00:00Z')
    But it still didn't work
    in the azure sdk says that I should add datetime keyword before UTC time format. And in nodejs I think the DateTime is the keyword for the datetime. But the query is wrong. Can anyone help me with that?
    Thanks

    Please see my comment on your question on Stack Overflow:
    http://stackoverflow.com/questions/28198972/nodejs-query-azure-storage-table-according-to-timestamp.
    Essentially, can you try the following code for specifying where clause:
    where("Timestamp ge datetime?", '2014-11-11T00:00:00Z')
    I ran this code and didn't get any error.
    Hope this helps.

  • Export Table Data to JSON via Azure Storage SDK

    Hi,
    I'm looking to export the contents of a table in JSON format and then save that JSON to one or more files.
    I see that we can have the table return JSON using: tableClient.DefaultRequestOptions.PayloadFormat = TablePayloadFormat.Json;
    I also see how to deserialize the JSON into an array of objects using something like this to get an array of CustomerEntity:
    IQueryable<CustomerEntity> query = from customer in table.CreateQuery<CustomerEntity>()
    where string.Compare(customer.PartitionKey, "I") >= 0 &&
    string.Compare(customer.PartitionKey, "X") <= 0 &&
    customer.Rating >= 2
    select customer;
    CustomerEntity[] customers = query.ToArray();
    But what if I don't want the results as CustomerEntity objects, I just want the raw JSON?
    The
    CloudTable.CreateQuery method requires a type that inherits from ITableEntity...
    I guess I could switch from using the Azure Storage SDK client to an HTTP client and query via OData, but I'd prefer a solution within the Azure Storage SDK...
    Thanks,
    Aron

    Thanks Will,
    Here is a more complete code snippet. As you can see, I have the payload set to JSON.
    const string customersTableName = "Customers";
    string connectionString = string.Format("DefaultEndpointsProtocol=https;AccountName={0};AccountKey={1}", accountName, accountKey);
    CloudStorageAccount storageAccount = CloudStorageAccount.Parse(connectionString);
    CloudTableClient tableClient = storageAccount.CreateCloudTableClient();
    // Values supported can be AtomPub, Json, JsonFullMetadata or JsonNoMetadata with Json being the default value
    tableClient.DefaultRequestOptions.PayloadFormat = TablePayloadFormat.Json;
    // Create the Customers table
    CloudTable table = tableClient.GetTableReference(customersTableName);
    table.CreateIfNotExists();
    // Insert a couple of customers into the Customers table
    foreach (CustomerEntity customer in CustomerEntity.GetCustomersToInsert())
    table.Execute(TableOperation.Insert(customer, echoContent: false));
    // The response have a payload format of JSON no metadata and the
    // client library will map the properties returned back to the CustomerEntity object
    IQueryable<CustomerEntity> query = from customer in table.CreateQuery<CustomerEntity>()
    where string.Compare(customer.PartitionKey, "I") >= 0 &&
    string.Compare(customer.PartitionKey, "X") <= 0 &&
    customer.Rating >= 2
    select customer;
    CustomerEntity[] customers = query.ToArray();
    However, the way the query is set up it automatically casts the results as CustomerEntity. That's the challenge - How to get the JSON payload before it gets cast to CustomerEntity...
    Thanks,
    Aron

  • Is there any way to automate purging Application Logging in azure storage account tables

    is there any way to automate purging Application Logging in azure storage account tables
    Rohit Pasrija

    hi Rohit,
    If you want to delete old data automatically, I think you need develop this feature by yourself. You could code the logic methods on your project, and set a timer to execute the methods to delete old data. Please refer to this thread (http://stackoverflow.com/questions/13602629/is-it-possible-to-acces-a-table-storage-in-azure-from-an-azure-web-site
    ), you could operate table storage data using azure sdk or REST API.
    Please try it.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to read azure storage data using JAVA with REST API

    Hi,
    We have a cloud service in our lab. We have enabled diagnostics
    in cloud services. So WADPerformanceCounterTable was created in storage account. Now , We want to read the WADPerformanceTable using
    JAVA with REST API. Is there any way to collect these details. please give me any
    sample code to connect azure storage using table service REST API.
    Thanks & Regards
    Rathidevi

    Hi,
    Please have a look at this article:
    https://convective.wordpress.com/2010/08/18/examples-of-the-windows-azure-storage-services-rest-api/, it demonstrate how to use table service Rest API, it also give us the code, hope this helps. Of course, the
    MSDN article could also help us coding.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to connect a web upload form(for video uploads) to Azure Storage / Media services?

    Hi I have a form on my Joomla site (see image under for viewers to upload and send me their videos, images and info.
    Question: 
    1.Instead of receiving this form content via email,  I just need a
    HTML and/or PHP code with tags code so I can send the content viewers give me (videos with up to 500MB) directly into my existing Azure blob.
    Seems I can also direct the form to a any database... They gave me this code sample but it was for AWS
    https://www.chronoengine.com/faqs/63-cfv4/cfv4-working-with-form-data/5196-can-i-save-files-to-aws.html
    Therefore they suggested me to reach out to Azure devs.
    I don't code so please show me the exact code HTML and/or PHP code to connect my "pre-built forms" with Azure Storage/Media services.
    kd9000

    You may refer the following links which may help you in building forms to direct to azure table/blob storage:
    http://www.joshholmes.com/blog/2010/04/15/creating-a-simple-php-blog-in-azure/
    http://stackoverflow.com/questions/27464839/use-form-post-php-to-upload-a-file-to-azure-storage-and-return-the-link
    Regards,
    Manu

  • Is "sp_purge_data" available only to datawarehouses or can it be used with both a normal database and Azure storage as well ?

    Is "sp_purge_data" available only to datawarehouses or can it be used with both a normal database  and Azure storage as well ?

    Thank you for the reply Qiuyun , the article was really helpful!
    I do have couple of other questions for you :
    How do we execute our SQL queries on Windows Azure tables and create horizontal partitions ? (I know that we have our SQL Server management Studio to execute normal queries on a SQL database , do we have a similar platform for Azure or do we have to get a local
    copy of the database to execute our queries and the  publish everything back to Azure)? I am looking to partition data on one of our databases and would like to know if it can be done in Azure directly or if we have to bring a local copy down ,write the
    partition function and
    and partition scheme or create a partition key and a row key - do the needful and publish it back to Azure?
    Also, how do I create a partition key and row key in Windows Azure?
    I am in the process of designing data archiving strategy for my team and would like to know more about the questions I just mentioned.
    Hoping to hear back from you soon.
    Thanks in advance for all the help!
    -Lalitha.

  • Latest Azure Storage SDK 4.2.1 fails under dev emulator when called at CloudTableClient.CreateIfNotExists()

    I am using Azure SDK 2.2 and Azure Storage SDK 4.2.1 and storage config. 2.0 in my cloud solution.  I am trying to initialize the storage as follows.
    CloudStorageAccount storageAccount = CloudStorageAccount.Parse(connectionKey);
                    // Create the table client.
                    CloudTableClient tableClient = storageAccount.CreateCloudTableClient();
                    // Create the table if it doesn't exist.
                    table = tableClient.GetTableReference(tableName);
                    table.CreateIfNotExists();
    Table.CreateIfNotExists() throws an exception in my dev emulator. I have been seeing this issue for quite sometime. It looks like MSFT has not fixed this issue at all. The same code works fine when I connect to the real Azure storage in the cloud. It just
    does not work under emulator. How can I debug anything. If I use Storage SDK 2.1, it works fine in dev emulator. After 3.0, nothing seems to work. Am I only one or anybody else have the same issue. Is there any workaround for this version or should I stick
    to the old version. Any help is appreciated.
    here is the detail exception.
    ex {"The remote server returned an error: (400) Bad Request."}
    System.Exception {Microsoft.WindowsAzure.Storage.StorageException}
    + [Microsoft.WindowsAzure.Storage.StorageException]
    {"The remote server returned an error: (400) Bad Request."}
    Microsoft.WindowsAzure.Storage.StorageException
    _className
    null string
    + _data
    {System.Collections.ListDictionaryInternal}
    System.Collections.IDictionary {System.Collections.ListDictionaryInternal}
    _dynamicMethods
    null object
    _exceptionMethod
    null System.Reflection.MethodBase
    _exceptionMethodString
    null string
    _helpURL
    null string
    _HResult
    -2146233088 int
    + _innerException
    {"The remote server returned an error: (400) Bad Request."}
    System.Exception {System.Net.WebException}
    + _ipForWatsonBuckets
    8791483017288 System.UIntPtr
    _message
    "The remote server returned an error: (400) Bad Request."
    string
    _remoteStackIndex
    0 int
    _remoteStackTraceString
    null string
    + _safeSerializationManager
    {System.Runtime.Serialization.SafeSerializationManager}
    System.Runtime.Serialization.SafeSerializationManager
    _source
    null string
    + _stackTrace
    {sbyte[192]} object {sbyte[]}
    _stackTraceString
    null string
    _watsonBuckets
    null object
    _xcode
    -532462766 int
    + _xptrs
    0 System.IntPtr
    + Data
    {System.Collections.ListDictionaryInternal}
    System.Collections.IDictionary {System.Collections.ListDictionaryInternal}
    HelpLink
    null string
    HResult
    -2146233088 int
    + InnerException
    {"The remote server returned an error: (400) Bad Request."}
    System.Exception {System.Net.WebException}
    + IPForWatsonBuckets
    8791483017288 System.UIntPtr
    IsTransient
    false bool
    Message
    "The remote server returned an error: (400) Bad Request."
    string
    RemoteStackTrace
    null string
    Source
    "Microsoft.WindowsAzure.Storage" string
    StackTrace
    "   at Microsoft.WindowsAzure.Storage.Core.Executor.Executor.ExecuteSync[T](RESTCommand`1 cmd, IRetryPolicy policy, OperationContext operationContext)\r\n   at Microsoft.WindowsAzure.Storage.Table.TableOperation.Execute(CloudTableClient client,
    CloudTable table, TableRequestOptions requestOptions, OperationContext operationContext)\r\n   at Microsoft.WindowsAzure.Storage.Table.CloudTable.Exists(Boolean primaryOnly, TableRequestOptions requestOptions, OperationContext operationContext)\r\n  
    at Microsoft.WindowsAzure.Storage.Table.CloudTable.CreateIfNotExists(TableRequestOptions requestOptions, OperationContext operationContext)\r\n   at PSI.Applications.RemoteManagement.TraceLogManagement.AzureTableStore.InitializeStorageTable(String tableName,
    String connectionKey) in c:\\AzureTraceCloudService\\AzureTableStoreLib\\AzureTableStore.cs:line 27"
    string
    + TargetSite
    {T ExecuteSync[T](Microsoft.WindowsAzure.Storage.Core.Executor.RESTCommand`1[T], Microsoft.WindowsAzure.Storage.RetryPolicies.IRetryPolicy, Microsoft.WindowsAzure.Storage.OperationContext)}
    System.Reflection.MethodBase {System.Reflection.RuntimeMethodInfo}
    WatsonBuckets
    null object
    + Static members
    Thanks,
    Ravi

    Hi Ravi,
    This issue maybe caused by storage client library is still not compatible with storage emulator and that's why your code is failing. I would recommend downgrading your storage client library to the previous version and your code should work just fine,
    the best option I think is to upgrade the SDK to the latest version, refer to
    http://azure.microsoft.com/blog/2014/08/04/announcing-release-of-visual-studio-2013-update-3-and-azure-sdk-2-4/ for more details.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Azure Storage Backup and Restore

    Hi,
    We have Azure Tables deployed in our Azure Storage Account. Is there a Backup & Recovery strategy we can use for Production Business Continuity plan.
    Thanks.
    Regards,
    Subhash Konduru
    Please remember to mark the replies as answers if they help and unmark them if they provide no help.

    Hi Subhash,
    Thanks for posting here.
    I suggest you to check this link for BCP on storage backup and recovery.
    https://technet.microsoft.com/en-us/library/dn621063.aspx
    http://blogs.technet.com/b/uspartner_ts2team/archive/2015/02/17/new-solution-accelerator-azure-business-continuity-and-disaster-recovery.aspx
    Hope this helps you.
    Girish Prajwal

  • Windows azure storage DevelopmentStorageAccount CloudTable.CreateIfNotExists() always get 400 http error

    environment:
    windows 8.1,  Visual Studio 2013,  Windows Azure SDK2.2 ,  Microsoft.WindowsAzure.Storage V3.0.0.0,  Windows Azure Storage Emulator 2.2
    Code snippet:
    CloudStorageAccount account = CloudStorageAccount.DevelopmentStorageAccount;
    CloudTableClient tableClient = account.CreateCloudTableClient();
    CloudTable table = tableClient.GetTableReference(tableName);
    table.CreateIfNotExists();  // this line cause error
    Error Message:
    System.Net.HttpWebRequest.GetResponse() +6594148
    Microsoft.WindowsAzure.Storage.Core.Executor.Executor.ExecuteSync(RESTCommand`1 cmd, IRetryPolicy policy, OperationContext operationContext) +1948
    [StorageException: remote server return error: (400) bad request]Microsoft.WindowsAzure.Storage.Core.Executor.Executor.ExecuteSync(RESTCommand`1 cmd, IRetryPolicy policy, OperationContext operationContext) +7389
    Microsoft.WindowsAzure.Storage.Table.TableOperation.Execute(CloudTableClient client, CloudTable table, TableRequestOptions requestOptions, OperationContext operationContext) +206
    Microsoft.WindowsAzure.Storage.Table.CloudTable.Exists(Boolean primaryOnly, TableRequestOptions requestOptions, OperationContext operationContext) +394
    I have search many threads on internet.  And I cannot get any solution.  Who can help? Thanks a lot.

    Hi,
    I didn’t find any issue in the code that you provided, I also try it and in my PC it created a new table in my storage, my system is windows 8, Visual Studio 2012, Windows Azure SDK2.2, Microsfot.WindowsAzure.Storage V2.0.0.0. From my experience, there is
    something wrong in you Windows Azure SDK, I recommended you to reinstall, and then try the code again. If you have any further issue, please feel free to let me know.
    Best Regards
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Microsoft.WindowsAzure.Storage.Table.CloudTable.ExecuteBatchAsync() truncates message

    When I call this method with large EntityProperty (around 17Kb of text), it truncates the string.
    I know that there is a limitation of 64Kb for a column and 1Mb for 1 entire row when it comes to Azure Table.
    Any insights?

    Hi,
    It sounds very strange, if you save the string value to other entity property, will it be truncated? If you use
    Azure storage explorer to save that string value, will it be truncated?
    If the string is not very confidential, could you share here? If you try to use the latest version of the azure storage client library, does this issue will be fixed? Please give us further information for a better support.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to set web job like a ssis which can take xmil file from azure storage to up date azure sql

    Hi,
    I have xml files under azure storage, i would like to set up web job which should be scheduled and
    load xml to update/insert/delete azure database.
    I have done SSIS locally and but cant use in azure.
    According to several advices and forum pages,I need to create azure web job.
    Please advice link or advice.
    Thanks.
    Superman

    Hi,
    To get started with Azure webjobs you could refer the following link:
    http://azure.microsoft.com/en-in/documentation/articles/websites-dotnet-webjobs-sdk-get-started/
    The article shows how to create a multi-tier ASP.NET MVC application that uses the WebJobs SDK to work with Azure blobs in an Azure Website.
    Meanwhile, we will research regarding your specific requirement and revert with our finds.
    Regards,
    Malar

  • How do I improve performance while doing pull, push and delete from Azure Storage Queue

           
    Hi,
    I am working on a distributed application with Azure Storage Queue for message queuing. queue will be used by multiple clients across the clock and thus it is expected that it would be heavily loaded most on the time in usage. business case is typical as in
    it pulls message from queue, process the message then deletes the message from queue. this module also sends back a notification to user indicating process is complete. functions/modules work fine as in they meet the logical requirement. pretty typical queue
    scenario.
    Now, coming to the problem statement. since it is envisaged that the queue would be heavily loaded most of the time, I am pushing towards to speed up processing of the overall message lifetime. the faster I can clear messages, the better overall experience
    it would be for everyone, system and users.
    To improve on performance I did multiple cycles for performance profiling and then improving on the identified "HOT" path/function.
    It all came down to a point where only the Azure Queue pull and delete are the only two most time consuming calls outside. I can further improve on pull, which i did by batch pulling 32 message at a time (which is the max message count i can pull from Azure
    queue at once at the time of writing this question.), this returned me a favor as in by reducing processing time to a big margin. all good till this as well.
    i am processing these messages in parallel so as to improve on overall performance.
    pseudo code:
    //AzureQueue Class is encapsulating calls to Azure Storage Queue.
    //assume nothing fancy inside, vanila calls to queue for pull/push/delete
    var batchMessages = AzureQueue.Pull(32); Parallel.ForEach(batchMessages, bMessage =>
    //DoSomething does some background processing;
    try{DoSomething(bMessage);}
    catch()
    //Log exception
    AzureQueue.Delete(bMessage);
    With this change now, profiling results show that up-to 90% of time is only taken by the Azure Message delete calls. As it is good to delete message as soon as processing is done, i remove it just after "DoSomething" is finished.
    what i need now is suggestions on how to further improve performance of this function when 90% of the time is being eaten up by the Azure Queue Delete call itself? is there a better faster way to perform delete/bulk delete etc?
    with the implementation mentioned here, i get speed of close to 25 messages/sec. Right now Azure queue delete calls are choking application performance. so is there any hope to push it further.
    Does it also makes difference in performance which queue delete call am making? as of now queue has overloaded method for deleting message, one which except message object and another which accepts message identifier and pop receipt. i am using the later
    one here with message identifier nad pop receipt to delete message from queue.
    Let me know if you need any additional information or any clarification in question.
    Inputs/suggestions are welcome.
    Many thanks.

    The first thing that came to mind was to use a parallel delete at the same time you run the work in DoSomething.  If DoSomething fails, add the message back into the queue.  This won't work for every application, and work that was in the queue
    near the head could be pushed back to the tail, so you'd have to think about how that may effect your workload.
    Or, make a threadpool queued delete after the work was successful.  Fire and forget.  However, if you're loading the processing at 25/sec, and 90% of time sits on the delete, you'd quickly accumulate delete calls for the threadpool until you'd
    never catch up.  At 70-80% duty cycle this may work, but the closer you get to always being busy could make this dangerous.
    I wonder if calling the delete REST API yourself may offer any improvements.  If you find the delete sets up a TCP connection each time, this may be all you need.  Try to keep the connection open, or see if the REST API can delete more at a time
    than the SDK API can.
    Or, if you have the funds, just have more VM instances doing the work in parallel, so the first machine handles 25/sec, the second at 25/sec also - and you just live with the slow delete.  If that's still not good enough, add more instances.
    Darin R.

  • Can't install windows azure storage sdk on windows phone 8

    From Visual studio 2013 ultimate, I opened Nuget package manager console and typed following command for my windows phone 8 solution:
    install-package windowsazure.storage
    And I got the following error:
    install-package : Could not install package 'Microsoft.WindowsAzure.ConfigurationManager 1.8.0.0'. You are trying to install this package into a project that targets 'WindowsPhone,Version=v8.0', but the package does not contain any assembly references or
    content files that
    are compatible with that framework. For more information, contact the package author.
    At line:1 char:1
    I get the same problem using UI version of package manager as well.
    Did something change with the newer version of windows azure storage?

    The package you are trying to install does not include the Windows Phone library. Please use the following command instead:
    Install-Package WindowsAzure.Storage-Preview -Pre
    Thanks for this. It worked for me. Can I just add, for those that don't know how to run this command. You go to tools > Library package manager > package manager console then paste in the command and hit enter.
    Thanks again!

Maybe you are looking for