Storage

How can we improve Azure Storage?

You've used all your votes and won't be able to post a new idea, but you can still search and comment on existing ideas.

There are two ways to get more votes:

  • When an admin closes an idea you've voted on, you'll get your votes back from that idea.
  • You can remove your votes from an open idea you support.
  • To see ideas you have already voted on, select the "My feedback" filter and select "My open ideas".
(thinking…)

Enter your idea and we'll search to see if someone has already suggested it.

If a similar idea already exists, you can support and comment on it.

If it doesn't exist, you can post your idea so others can support it.

Enter your idea and we'll search to see if someone has already suggested it.

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. make it possible to use SSL on blob storage using custom domains

    Currently you can use SSL but you have to user the standard URL. You can create a CNAME to your storage account but most browsers complain that the traffic was rerouted and is possibly an attack. There should be a way to install a domain certificate to your containers.

    4,468 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    135 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    You can now use the Azure CDN to access blobs with custom domains over HTTPS. See the following article for instructions on how to do so: https://docs.microsoft.com/en-us/azure/storage/storage-https-custom-domain-cdn. Having talked to a number of customers, we concluded that this solution addresses many scenarios where the need for HTTPS access to blobs with custom domains exists.

    Native Azure Storage support for using SSL to access blobs at custom domains is still on our backlog. We would love to hear about your scenarios where using the Azure CDN is not an acceptable solution, either by posting on this thread or sending us an email at azurestoragefeedback@microsoft.com.

  2. Static website hosting in Azure blob storage

    This would require supporting default file to render and somekind of redirect from root to a particular container.

    2,205 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    71 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for your feedback. We are currently in public preview of static website hosting for Azure Storage to enable this scenario. Check out the blog post here for more details: https://azure.microsoft.com/en-us/blog/azure-storage-static-web-hosting-public-preview. The feature set includes support for default documents and custom error documents for HTTP status code 404.

    For any further questions, or to discuss your specific scenario, send us an email at azurestoragefeedback@microsoft.com.

  3. Provide Time to live feature for Blobs

    If I need to provide a user (or external system) some data (blob) which might be outcome of some processing (or other) and it has some expiration time I'd like to just put a new blob and set TTL property with TimeSpan (or set absolute DateTime). When the period is over my blob is deleted. So I don't have to pay for it and don't need to spin up some service for doing it myself.

    1,266 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    36 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for your feedback. We are currently in public preview of blob storage lifecycle management. The feature set offers a rich, rule-based policy engine which you can use to transition your data to the best access tier and to expire data at the end of its lifecycle. See our post on the Azure Blog to get started: https://azure.microsoft.com/en-us/blog/azure-blob-storage-lifecycle-management-public-preview/preview/.

    For any further questions, or to discuss your specific scenario, send us an email at DLMFeedback@microsoft.com.

  4. Rename blobs without needing to copy them

    Copying blobs in order to rename them is a heavy operation, especially when the blob is big, or when you need to change many files, such as when changing a directory name.

    747 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    20 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Our apologies for not updating this ask earlier. Renaming Blobs is on our backlog, but is unlikely to be released in the coming year. Today, you can use the “Copy Blob” API as a workaround.

    Using the Azure Files service you can address Azure Storage like a network share using the SMB2.1 protocol.  This enables usage of normal Windows API’s to rename files and directories.  You can get started with the Files service by visiting https://docs.microsoft.com/en-us/azure/storage/storage-dotnet-how-to-use-files.

    For any further questions, or to discuss your specific scenario, send us an email at azurestoragefeedback@microsoft.com.

  5. Allow user-based access to Blob Containers (for support employees)

    For auditing purposes and to prevent data corruption, we want to give our support employees a user-centric, read-only access to Blob Containers in order to be able to investigate possible data corruptions (caused by bugs in systems).

    This is not possible now because the security architecture of Blob Service does not even know the concept of users or roles.

    SAS is not secure enough mechanism because it gives access to anyone by just sharing a link + you can't track who's actually using it.

    310 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    10 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for you feedback. Currently we are in public preview of Azure Active Directory authentication for storage. This feature set allows you to use Azure’s role-based access control framework to grant specific permissions to users, groups and applications down to the scope of an individual blob container or queue. You can see the public preview announcement here: https://azure.microsoft.com/en-us/blog/announcing-the-preview-of-aad-authentication-for-storage/

    For any further questions, or to discuss your specific scenario, send us an email at azurestoragefeedback@microsoft.com.

  6. Allow setting Archive tier at the account and container levels

    Currently the archive tier can be set only at the blob level. There are plenty of uses cases for having entire storage accounts or containers for archival where setting the tier for each blob is tedious and non-value-adding.

    135 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    8 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for your feedback. We are currently in public preview of blob storage lifecycle management. The feature set offers a rich, rule-based policy engine which you can use to transition your data to the best access tier and to expire data at the end of its lifecycle. This framework can be used to match and tier blobs based on prefix to enable batch archiving of an account, containers, or even virtual directories. Having talked to a number of customers, we concluded that this solution addresses many scenarios where the need for account and container level archiving exists. See our post on the Azure Blog to get started: https://azure.microsoft.com/en-us/blog/azure-blob-storage-lifecycle-management-public-preview/preview/.

    For any further questions, or to discuss your specific scenario, send us an email at DLMFeedback@microsoft.com.

  7. Azure Policy For Preventing Public Blob Containers

    There are no ARM REST API properties for Blob containers which means we can't create an Azure Policy for not allowing people to ever create public blob containers.

    This is how every AWS breach has occurred due to accidentally setting storage to public.

    123 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    6 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  8. Allow adding headers to static website hosting in blob storage

    The static website hosting is fantastic, to make this feature even more awesome please allow adding of custom headers like CSP, HSTS etc secure the static website.

    91 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  9. Auto defragment append blobs

    Append blobs are great, however, when a such a blob is generated through a long series of small increments, the read performance of the resulting blob is very poor: from 10x to 20x slower than reading a regular page blob.

    The performance problem goes away if the app rewrite the append-blob in large 4MB chunks. However, this process is complicated to setup, and collides with any 'always-on' property of the app.

    As append blobs are append-only, it would be much better if Azure was taking care of defragmenting the append blobs on its own; possible through a dedicated API operation…

    55 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  10. Make Blob access policies Azure Resource Manager managed

    In order to apply Azure Policies the item you want to control has to be ARM managed. Apparently Blob Access Policies are not ARM managed (i.e. resource that can be managed by calling https://management.azure.com/[ResourceId]). Because of this the Azure Policies product group can't make a policy alias to audit/deny the use of public containers/blobs, see: https://github.com/Azure/azure-policy/issues/131.

    This is a big issue because the use of anonymous access to storage containing enterprise data is a huge risk.

    44 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  11. hotlink prevention

    Preventing others to consume my Azure resources via HTTP_REFERER header validation.
    This is a common scenario for blogs, websites etc.

    Scott Hanselman has even written a blog post about it: http://www.hanselman.com/blog/BlockingImageHotlinkingLeechingAndEvilSploggersWithIISUrlRewrite.aspx

    22 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  12. Support GZIP & DEFLATE Content-Encoding for List Blobs Responses

    For collections with large numbers of blobs, the XML response payload easily approaches 100s of MBs. In our testing, each block of 5000 objects resulted in a ~2.8MB response. This is without any other flags set (i.e. no snapshots, no metadata, no uncommitted blobs). The response XML is highly compressible. In our testing we saw a 93% size reduction. (2.8MB -> 210KB). This would be a huge improvement in transfer performance, and a large cost savings in egress bandwidth situations.

    20 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    3 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  13. We would like to audit/log delete operations on Azure blob storage containers. Currently we can only do this at a Storage Account layer.

    We are simply looking for more granularity with our storage logging in Azure. If someone were to view/delete our blob containers, we would like to see these operations logged and have the ability to alert on them.

    17 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  14. Enable Authorization to the Azure Static Websites using Azure Active Directory

    Right now, there is development done to use Blob storage for a static website (known as Azure Static Websites). The feature is in preview.

    However, the downside of this is that the website is public. There is no built-in to restrict the access to the website.

    In my use case, I would have used Azure Static Websites to store the static documentation produced by DocFx (https://dotnet.github.io/docfx/). This documentation includes the classes as well including technical documentation and wikis.

    I would be great if I could use my Azure Active Directory to control who can actually see the pages…

    16 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  15. Server side delete, rename of virtual folder

    Currently if I have millions of documents inside a virtual folder, there is not way to delete them without actually iterating through everything, which is extremely time consuming (weeks to process deletions)

    For example, if you have multiple containers, which in turn have multiple virtual folders with millions of blobs, moving, renaming or deleting the "folder" is impossible without some painfully slow iterating through everything.

    thanks!

    15 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  16. Support If-Modified-Since precondition in blob enumeration APIs

    (Unless I've missed it somewhere) It would be real nice if last modification time filtered blob enumerations could be applied server-side.

    12 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  17. more detailed error messages

    We've run into an issue where data we tried to store wouldn't fit in the blob file. The error message was (416) The page range specified is invalid.

    At first it wasn't clear whether we were trying to start saving beyond the page file or data didn't fit or something else was culprit. Eventually we figured it out.

    It would be nice if the error message would provide more details: page size, size we tried to save, location we tried to save within the file, etc. This would've showed us the issue right away. Instead we had to spend a…

    12 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for your feedback. We are currently working on providing this functionality and will provide updates when they become available. See the following article for the latest: https://docs.microsoft.com/en-us/rest/api/storageservices/status-and-error-codes2. Note that for REST API version 2017-07-29 and later, failed API operations also return the storage error code string in a response header. For any further questions, or to discuss your specific scenario, send us an email at azurestoragefeedback@microsoft.com.

  18. Add capability for bulk deletion of blobs

    Right now, deletion of a container is the only way a number of blobs can be deleted together. The drawback, however, is that the container name cannot be used again for some time. This causes inconvenience for a number of use cases.

    Proposal: There should at least be way for emptying the container in one API call

    11 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  19. Support "Last Accessed" timestamp in blob metadata

    With the new "Cool Storage" feature announced recently, it would be useful if we could tell when our "Hot" blobs were last accessed, so we could push them to Cool storage where appropriate. We could achieve this with a proxy service and SAS signatures, but having this automatically updated in the blob metadata itself would be much simpler and remove the requirement for any compute resource being required for simply accessing a blob.

    10 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
  20. Support chunked transfer encoding on Blob PUT

    Currently the PUT API for block blobs requires a Content-Length header. This essentially makes the API unusable for clients that use async pipelining to start sending the blob bytes while they are being received or generated, and therefore before the total content length is known. Rather, it requires the caller to buffer the entire blob locally, determine the length, and then call the PUT API with that length. This is inefficient. When using an async client SDK, like the Java Storage SDK v10, knowing the entire content length ahead of time should not be necessary. See https://github.com/Azure/azure-storage-java/issues/336, for example.

    10 votes
    Vote
    Sign in
    (thinking…)
    Password icon
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Blobs  ·  Flag idea as inappropriate…  ·  Admin →
← Previous 1 3 4
  • Don't see your idea?

Feedback and Knowledge Base