Azure Search

Azure Search is a search-as-a-service solution that allows developers to incorporate a sophisticated search experience into web and mobile applications without having to worry about the complexities of full-text search and without having to deploy, maintain or manage any infrastructure

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. 355 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    24 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    This functionality is now available to all customers as a part of Cognitive Search feature of Azure Search. See:
    - Introduction to Cognitive Search: https://docs.microsoft.com/azure/search/cognitive-search-concept-intro
    - Process and extract information from images: https://docs.microsoft.com/azure/search/cognitive-search-concept-image-scenarios

    Please try it out and give us your feedback.
    Thanks,
    Your Azure Search team

  2. Blob Indexing w File Crackers

    We would love to have blob containers indexed. One option may be to use iFilter support.

    205 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    completed  ·  11 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  3. Azure Search and SQL Server integration

    Integrating SQL Server (Azure and on-prem) with Azure Search. A search in table/column that is Azure Search enabled would fall into a Azure Search query in the background, not visual for the developer.

    Only Azure Search enabled columns and tables would be handled by Azure Search and it would not be visible for the database consumer.

    Only a Azure Search Package that is installed on a on-prem machine that would enable the DBA to define which tables or columns that should be Azure Search enabled. Database triggers for SELECT, INSERT, UPDATE and DELETE would translate into Azure Search REST based…

    174 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  4. 106 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  5. 80 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    5 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  6. DocumentDB indexer should be able to index MongoDb collections

    If I create a MongoDB collection backed by DocumentDB, I should be able to use the DocumentDB indexer to ingest it into Azure Search.

    69 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    5 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  7. Support user-provided query in SQL indexer

    Let's say I run an indexer to read a really large table, with a billion records, out of which I'm interested in only a few thousands. This is a database I do not own, and I have only read permissions to it.

    In this case it would be better to have a simple filter on which data to keep in the indexer throw away the irrelevant data.

    For eg: "Index this table where TeamName equals MyTeam"

    Even though the indexer runs on a large set of data, my index would only contain a smaller subset of data.

    Smaller the index,…

    66 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    5 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  8. Datasource can contain multiple DocumentDB collections

    Because each DocumentDB collection can be only 10GB, we had to partition our data over multiple collections (app01, app02, etc..).

    When creating a indexer Data Source, the container field can only take one collection name. We can get around the problem by creating multiple indexer and multiple data source for each documentDB collection, but it would be much easier if:


    1. DataSource's container property can accept a list of names.

    or


    1. The container property can accept regex expression. So it would be something like "container" : { "regex": "[0-9]"}

    42 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    DocumentDB now supports partitioned collections with potentially infinite scale (see https://azure.microsoft.com/en-us/documentation/articles/documentdb-partition-data/), and Azure Search now has support for those collections as well. This obviates the need for “old-style” explicitly partitioned collections. If you haven’t done so already, please consider switching to the new DocumentDB collections.

    Thank you for using Azure Search!
    Your Azure Search team

  9. Auto-indexing of DocDB

    Lately, I've been working a lot with Cloudant, which is a distributed CouchDB platform with integrated Lucene search. I absolutely love the fact that you can define some sort of triggers that, when a document is inserted or updated, update the index to reflect that operation. Likewise, when a document is deleted, the corresponding indexes are deleted from Lucene.

    This is particularly convenient as you don't have to manually sync your DB and your search index, everything happens automatically under the cover. It would be great to have that on top of Azure Storage or the upcoming DocDB (even though…

    40 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  10. Blob Indexer should continue indexing documents after encountering an error

    This is related to...

    https://feedback.azure.com/forums/263029-azure-search/suggestions/11628435-blob-indexer-should-be-able-to-skip-unsupported-co

    ...whilst a change has been added to easily configure the indexer to exclude unsupported document formats, this is inadequate. Bugs in the indexer and corrupt documents are going to ensure that there will always be a percentage of documents that can't be processed by the indexer. This shouldn't stop the indexer from processing data it can index.

    When the indexer encounters an error processing a document it should log the error then proceed with indexing of the rest of the documents in storage.

    A run shouldn't be halted until the indexer has attempted to index…

    34 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    3 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  11. Extract document structure from JSON blobs

    Currently, blob indexer extracts entire document into a single 'content' field. Instead, we want the blobs containing JSON documents to be interpreted as documents with multiple fields.

    30 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  12. 25 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  13. 24 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    3 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  14. 21 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  15. SQL Azure indexer support for Collection(Edm.String)

    Currently migrating from FTS to Azure search. While in preview we setup our indexes with collections for some one to many data types (phone numbers, addresses, etc...) It would be nice to realize these collections with an indexer.. As far as the structure of the column we don't really have an existing structure in our system so we'd just build it based on what made the most sense for the way you process data.. I realize fundamentally that type of structure in SQL is unnatural so there's probably not something that just naturally fits. Possibly XML?

    Few thoughts
    -JSON (since…

    19 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  16. 18 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    5 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  17. Support for Geography SQL Server column

    Actually, the indexer returns an error if the view or the table contain a geography column, because indexer doesn't support it

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  18. Allow custom defined Field Mapping Function

    Currently when defining the indexers, one can specify the field mappings from source to target fields. Only a limited set of functions such as Extract token at position etc are supported. It would be desirable to support Func syntax such that each row is passed onto the func which then decide the target field value based on developer defined logic in func.

    Eg: we have a cosmo db source and a azure table storage data source. The partition key for cosmo db collection is combination of tenantId:studentId for efficient read/writes across partitions based on our data. The azure table storage…

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    We recently introduced a capability of Azure Search that we call Cognitive Search.

    Cognitive Search allows you to apply enrichment and transformation steps to your data that we call Cognitive Skills. You can select from predefined skills or apply your own custom skill.

    The best way to transform your input data when you want to define your own function would be through a custom skill that you run as part of your enrichment pipeline.

    You can read how to do this at https://docs.microsoft.com/en-us/azure/search/cognitive-search-custom-skill-interface

    Thanks!
    -Luis Cabrera

  19. Allow SkipContent to be set directly on the blob indexer

    The azure blob indexer currently checks for a metadata property AzureSearch_SkipContent to skip the processing of the content. It would be easier to just tell the indexer directly to skip processing content for all blobs.

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    This functionality is now available in REST API, using 2015-02-28-Preview version.

    Excerpt from the docs:
    Using indexer parameters to control document extraction

    If you want to skip content extraction for all blobs, you can do this using indexer configuration object, instead of having to add custom metadata to each blob individually. To do this, set SkipContent configuration property to true in the parameters object:

    PUT https://[service name].search.windows.net/indexers/?api-version=2015-02-28-Preview
    Content-Type: application/json
    api-key: [admin key]

    {
    … other parts of indexer definition
    “parameters” : { “configuration” : { “SkipContent” : true } }
    }

    Thanks!
    Your Azure Search team

  20. Support shared access signature for blob datasource credentials

    Currently, blob datasource requires a full connection string. Sometime, specifying a SAS container URL is preferable.

    13 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
← Previous 1
  • Don't see your idea?

Feedback and Knowledge Base