Storage

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. User assigned identity in storage account (ARM template for custom key SSE)

    We are trying to define ARM template for storage accounts using custom key for SSE. Such definition is required for Azure Blueprints.

    Currently the only way to enable custom key for SSE is 3 step process: 1-Create SystemAssigned identity in storage account, 2-Update Keyvault access policies for that identity, 3-Update storage encryption settings.

    If we can get User (customer) assigned identity into storage account for accessing Keyvault, then we can pre-prepare / isolate step 1 and 2. Then we can have ARM template definition with custom key for SSE defined for a new storage account as a single step (3).

    6 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  2. Immutable storage cannot be unlocked

    Function: Immutable storage

    Issue: Customer can easily lock his resources in storage accounts, but he cannot unlock them by himself when the lock is no longer needed.

    Ref document: https://docs.microsoft.com/en-us/azure/storage/blobs/storage-blob-immutable-storage#faq

    Recommend: Provide an "unlock" button to customer.

    6 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  3. Automate domain replacement in Azure Storage Explorer link generator

    Azure Storage Explorer comes with a default URL generator for all files - http(s)://<storage account>.blob.core.windows.net/<container>/<file name>.

    The application also allows you to bind a custom URL to the service that takes the place of the "<storage account>.blob.core.windows.net/" section of the URL, making it http(s)://<custom domain>/<container>/<file name>.

    It would be super helpful to be able to be able to update the settings to indicate a custom domain is being used so that the "Copy URL" function replaced the root domain for you.

    Currently the advice received from MS is to manually replace the indicated URL segment manually, every time Storage Explorer…

    6 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  4. Firewall rule that can accept connections from VNET on the other AAD tenant

    Currently Storage firewall have a limitation, that source VNET must be in the same AAD tenant.
    https://docs.microsoft.com/en-us/azure/storage/common/storage-network-security?toc=%2fazure%2fvirtual-network%2ftoc.json#required-permissions

    It would be nice if cross tenant connection is enabled. It is quite useful to protect storage data by VNET basis and at the same time can accept connection from other partner companies who want to work on the specific storage account.

    5 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  5. Improving the indexed partitioning criteria for handling storage accounts with same naming conventions

    As described in detail here: https://azure.microsoft.com/en-us/documentation/articles/storage-performance-checklist/#subheading47 , Azure handles indexed partitioning with an index with aphabetical criteria, so storage accounts starting with A-H will go into one partition, H’-R will go into a different partition and last R’-Z will go into another different partition too.
    We discovered that this could affect High Availability of VMs if you take care of putting them into Availability Set and use different storage accounts for their virtual disks, but same naming convention for accounts. In this case, if a fail happens on a given partition for example during an internal maintenance task or storage…

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  6. 4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    under review  ·  0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  7. Search for container or files based on a date or date range

    I would like to search for blobs (container or file) by single date or date range using either Microsoft Azure Storage Explorer on my desktop or portal. This will allow me to delete or review items without having to know the name of the container or the file.

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  8. Increase storage account create calls per subscription limit

    “RDFE throttles the storage account creation; the default throttling limit is 20 per subscription per hour "

    Most of our automation creates storage account on the fly and they fails with following reasons

    "The maximum number storage account creations calls allowed for a subscription in an hour has been exceeded. Please try again after some time."

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    under review  ·  0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  9. Filter and sort files/blobs easily

    Enable filtering of files/blobs by their type rather than only their name, and enable ordering list per date of modification or file size.
    Right now it's a bit cumbersome to use when we're debugging.

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  10. Market Share

    Either charge for Data Transactions or total storage usage.... you're already charging for Active Time, this might deter Larger companies.
    People might start thinking that they are double charged and it gets out of hand.

    Microsoft might want to make money but i think it should focus more on the market share first.

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  11. Allow Azure SignalR service CORS rules to be applied via ARM template

    There is currently no support to configure CORS settings of an Azure SignalR Service using ARM templates.

    (See template reference: https://docs.microsoft.com/en-us/azure/templates/microsoft.signalrservice/2018-10-01/signalr)

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  12. Accept-Ranges: bytes header

    Blob storage REST API supports returning partial responses when a Range header is included in the HTTP request

    Blob storage however doesn't provide 'Accept-Ranges: bytes' in the response headers for GET nor HEAD requests, which would indicate to the client/browser that the server supports these types of requests.

    Documented here:
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Accept-Ranges

    Http spec here:
    https://tools.ietf.org/html/rfc7233#section-2.3

    Please add 'Accept-Ranges: bytes' to the headers of Blob Storage REST API responses

    4 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  13. Forecast or Show Approximate Storage Size for Selected Items for Azure Backup before the actual/scheduled Backup

    Hi Azure Microsoft Team

    Highly appreciate if you can calculate or forecast the selected items to Backup in the Schedule Backup Wizard for Azure Backup so that we can determine the file size storage are being backup rather than after the backup job success.

    It will be another awesome feature to be added.

    Thank you

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  14. Ability to set Proxy Server

    Allow the ability to set the proxy server for requests on windows azure storage

    I will say yes you can set the global but this isn't always ideal in all environments.

    3 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    under review  ·  0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  15. Storage tiering / Replication


    • Offer capability like SSD/Sata/Sas high perf, Low cost with automatic tiering, offer deduplication capability mean in the same Data center store 10 times the same file for the same price, the capability to use DPM for Backup , SCOM capability ...
      Offer live replication in active /active Mode with GSLB like front end capability ..

    But Azure is a very good product.

    2 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    under review  ·  0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  16. Storage operations

    On Azure storage I would like to (from the new portal):
    - list content of tables
    - run queries against tables
    - peek/pop/put queues and show approximate queue size
    - browse blob hierarchies
    - download and upload blobs

    This functionality would alone be a reason to move from old portal to new portal.

    2 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  17. Give size limitations on imports and exports. Is there any size limitation especially dealing with drives.

    Please update the documentation with import/export sizes and increase the level of detail with application in limitations.

    2 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  18. Allow Logic App Connectors to dynamically pull a Storage Account Key rather than sticking with a statically configured key value

    Currently, Connectors to Storage Accounts are configured with a Storage Account key during setup. Implementing Microsoft's recommended key rotation breaks the Connector as it's not able to dynamically pull the key value.

    There is a workaround via PowerShell and adding an Azure Automation step to grab the key value and update the Connector prior to running any Logic App steps that require this Connector. While functional, this adds more access and complexity than should be required.

    Data Factory v2 has a connector that dynamically pulls values from a Key Vault (screenshot attached). It would be great if the Storage Account…

    2 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  19. Create an Azure Instance type with NVDIMM memories for high speed Storage Cache

    An NVDIMM (https://en.wikipedia.org/wiki/NVDIMM) is a type of memory used by Storage vendors as local cache, to accelerate High Performance Computing, Deep Learning, Deep Analytics, Simulations, and other storage-and-compute-intensive applications.

    Currently no cloud vendor supports instances with NVDIMMs, and Azure could break new ground here. We already have GPU's, FPGA's and other specialized hardware on our instances.

    NVDIMM would enable certain ISVs who cater for the HPC, Deep Learning, Simulation, and Deep Analytics crowd, and enable them to provide innovative solutions that today only exist in Private Cloud on on-premises applications.

    1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  20. 1 vote
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  General  ·  Flag idea as inappropriate…  ·  Admin →
  • Don't see your idea?

Feedback and Knowledge Base