SQL Data Warehouse

Do you have an idea or suggestion based on your experience with SQL Data Warehouse? We would love to hear it! Please take a few minutes to submit your idea or vote up an idea submitted by another SQL Data Warehouse customer. All of the feedback you share in these forums will be monitored and reviewed by the SQL Data Warehouse engineering team. By suggesting or voting for ideas here, you will also be one of the first to know when we begin work on your feature requests and when we release the feature.

Remember that this site is only for feature suggestions and ideas! If you have technical questions please try our documentation, MSDN forum, or StackOverflow. If you need support, please open a support ticket with us.

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. sp_send_db_mail needs to be supported to send mails from stored procedure

    sp_send_db_mail needs to be supported to send mails from stored procedures created on Azure.

    We should have the abilities of sp_send_dbmail which was available with databases

    7 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Management  ·  Flag idea as inappropriate…  ·  Admin →
  2. Support for MapR on Polybase

    MAPR are encountering a stumbling block in the lack of support for MapR on Polybase.

    4 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  3. scheduled Auto Scale Up/Down for Azure SQL DWs on Azure Portal

    CAn you please provide the Auto scale up and Down functionality along with Alerts mechanism in Azure portal .

    11 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    3 comments  ·  Flag idea as inappropriate…  ·  Admin →
  4. Persist Data dictionary to a separate database for DBA

    Persist Data dictionary to a separate database by allowing the customer to persist the DMV, which can be used by DBA for fixing performance issues

    3 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Management  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for voting for this feature folks. We’d like additional information before commenting on this topic. Please provide your scenario in the comments below. We currently have Query Data Store on our roadmap.

  5. Polybase: PARTITIONED BY functionality when creating external tables.

    I have a large number of files which are read with Hive using a partitioning scheme. PARTITIONED BY functionality, which is so commonly used in HIVE is missing from polybase.

    64 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    8 comments  ·  Polybase  ·  Flag idea as inappropriate…  ·  Admin →
  6. Allow install of Linux ODBC to be automated via puppet

    Allow install of Linux ODBC to be automated via puppet. The current Microsoft SQL Server ODBC packages can't be installed via puppet as they require EULA acknowledgment. If we adopted the ADW technology, we would inevitably have to install the ODBC software on several grids of computers that host our various ETL software. Puppet installation allows the implementers of new systems to get the software automatically installed without the DBA team's intervention.

    2 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    2 comments  ·  Management  ·  Flag idea as inappropriate…  ·  Admin →
  7. Space quota at schema level

    Need to be able to specify a space quota for a schema and abort current work that causes the quota to be exceeded.

    Forces user groups to be good stewards and manage their storage usage or be prevented from additional work.

    1 vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Management  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for providing for this feedback folks. We’d like additional information before commenting on this topic. Please provide your scenario in the comments below specifically which quota configuration at the schema level.

  8. sp_help

    Support sp_help

    9 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    4 comments  ·  T-SQL  ·  Flag idea as inappropriate…  ·  Admin →
  9. Provide a way to expose the actual DWU usage (not the same as the scale setting).

    You can see what the database is scaled to i.e. DWU 200, but how do you know how much is actually being used over time. The portal display a graph of both the DWU limit and the DWU used but there is no way to programmatically monitor how much is being used.

    31 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    2 comments  ·  Management  ·  Flag idea as inappropriate…  ·  Admin →

    We are actively improving our monitoring experience. Currently we have ‘DWU Used’ in the portal which is a blend between CPU and IO to indicate data warehouse utilization. We also have future improvements on our road map such as Query Data Store and integrating with Azure Monitor for near real time troubleshooting in the Azure portal. If anyone has any other feedback, please elaborate on your scenario on this thread. Thank you for your continued support!

  10. PolyBase:allow encoding in file format and polybase will take care of encoding.

    PolyBase: Allow encoding in file format and polybase will take care of encoding.

    8 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Polybase  ·  Flag idea as inappropriate…  ·  Admin →
  11. Once a proc is executed and if we try to stop it, its not working. Sometimes it takes 6-8 hours to get it cancelled

    Once a proc is executed and if we try to stop it, its not working. Sometimes it takes 6-8 hours to get it cancelled

    7 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Reliability  ·  Flag idea as inappropriate…  ·  Admin →
  12. 'Replication' solution from sql server to sql dataware house?

    Why not implement a Sql server 'Replication' solution from sql server (on premise / Azure) to sql dataware house (on Azure)?
    It would be useful if we have a processing DB in sql server and a accumulating datawarehouse in sql datawarehouse, with not too many structural changes.

    30 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    2 comments  ·  Integration  ·  Flag idea as inappropriate…  ·  Admin →

    Thank you for voting on this feature. Please describe your scenario in the thread below. We currently have many migration solutions from SQL Server to SQL Data Warehouse such as Azure Data Factory and Redgate through the Quick Load Azure portal experience (link below). If this is a continuous replication process, we’d love to understand your hybrid scenario so please comment below.

    https://azure.microsoft.com/blog/enhanced-loading-monitoring-and-troubleshooting-experience-for-azure-sql-data-warehouse/

  13. Performance with respect to dynamic SQL

    In our solutions we load data from csv-files into SQL DW internal tables using polybase/external tables.

    Per roll-out we face the challenge that the schema (columns) in the csv may differ for the same table. Therefore we implemented a logic to lookup the column name and data type from a dictionary table and we create the external and internal table schemes dynamically. As we have round about 500 tables with up to 20 columns per table automating this process is the only way to go.

    The issue we face is that compared to an on premise SQL Server the dynamic…

    16 votes
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)

    We’ll send you updates on this idea

    2 comments  ·  Performance  ·  Flag idea as inappropriate…  ·  Admin →
  • Don't see your idea?

Feedback and Knowledge Base