Azure Search

Azure Search is a search-as-a-service solution that allows developers to incorporate a sophisticated search experience into web and mobile applications without having to worry about the complexities of full-text search and without having to deploy, maintain or manage any infrastructure

  • Hot ideas
  • Top ideas
  • New ideas
  • My feedback
  1. asciifolding support

    I have some content that contains diacritics like Caffè, which azure search would not be able to pick up if user searches Caffe. Elastic search has asciifolding support.

    Is it possible to implement or configure somewhere? http://www.elasticsearch.org/guide/en/elasticsearch/guide/current/asciifolding-token-filter.html

    17 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    5 comments  ·  Indexing  ·  Flag idea as inappropriate…  ·  Admin →
  2. Create a Sitecore Azure Search Provider

    Sitecore comes with/supports Lucene and SOLR sitecore search providers.
    Coveo have built/supports their own Sitecore search provider.
    Could Microsoft build/support a Sitecore Azure Search provider?

    Currently are a few open source implementations, e.g. https://github.com/jscott1277/SitecoreAzureSearchProvider

    In a recent press release
    http://www.sitecore.net/about/press-and-media/press-releases/2016/07/sitecore-partners-microsoft-transform-digital-experience-at-speed-microsoft-azure.aspx
    "the two companies plan to expand their long-standing strategic alliance to enrich Sitecore’s Microsoft Azure cloud offerings over the next two years"
    Might this be included as part of that?

    16 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Indexing  ·  Flag idea as inappropriate…  ·  Admin →
  3. Support for Geography SQL Server column

    Actually, the indexer returns an error if the view or the table contain a geography column, because indexer doesn't support it

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  4. Allow custom defined Field Mapping Function

    Currently when defining the indexers, one can specify the field mappings from source to target fields. Only a limited set of functions such as Extract token at position etc are supported. It would be desirable to support Func syntax such that each row is passed onto the func which then decide the target field value based on developer defined logic in func.

    Eg: we have a cosmo db source and a azure table storage data source. The partition key for cosmo db collection is combination of tenantId:studentId for efficient read/writes across partitions based on our data. The azure table storage…

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    We recently introduced a capability of Azure Search that we call Cognitive Search.

    Cognitive Search allows you to apply enrichment and transformation steps to your data that we call Cognitive Skills. You can select from predefined skills or apply your own custom skill.

    The best way to transform your input data when you want to define your own function would be through a custom skill that you run as part of your enrichment pipeline.

    You can read how to do this at https://docs.microsoft.com/en-us/azure/search/cognitive-search-custom-skill-interface

    Thanks!
    -Luis Cabrera

  5. Allow SkipContent to be set directly on the blob indexer

    The azure blob indexer currently checks for a metadata property AzureSearch_SkipContent to skip the processing of the content. It would be easier to just tell the indexer directly to skip processing content for all blobs.

    15 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →

    This functionality is now available in REST API, using 2015-02-28-Preview version.

    Excerpt from the docs:
    Using indexer parameters to control document extraction

    If you want to skip content extraction for all blobs, you can do this using indexer configuration object, instead of having to add custom metadata to each blob individually. To do this, set SkipContent configuration property to true in the parameters object:

    PUT https://[service name].search.windows.net/indexers/?api-version=2015-02-28-Preview
    Content-Type: application/json
    api-key: [admin key]

    {
    … other parts of indexer definition
    “parameters” : { “configuration” : { “SkipContent” : true } }
    }

    Thanks!
    Your Azure Search team

  6. Support shared access signature for blob datasource credentials

    Currently, blob datasource requires a full connection string. Sometime, specifying a SAS container URL is preferable.

    13 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  7. Permit bulk loading of documents from SQL server

    We have about 2.5 million documents, currently stored in a table in a SQL database on Azure.

    We would like to insert these documents into an Azure search service, hosted in the same datacentre.

    Currently the only way to do this is by reading each row and inserting using the REST API.

    Something like the request feature "Azure Search and SQL Server integration" would help in this case.

    Although that request is asking for the data to be stored in SQL, in our case we want the data to be moved from SQL.

    An alternative is for the REST API…

    12 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  8. Expose search quality metrics like CTR (click-through rate) in the Search Traffic Analytics dashboard

    That would require collecting information about user clicks. The dashboard could also show the number of unique users over time and MRR (mean reciprocal rank).

    12 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Flag idea as inappropriate…  ·  Admin →
  9. 11 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Scoring  ·  Flag idea as inappropriate…  ·  Admin →
  10. 10 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Portal  ·  Flag idea as inappropriate…  ·  Admin →
  11. Able to do simple search on portal

    To be able to view index and count of documents is what we have today. We are not able to see the actual documents also.
    Can we have a simple search bar for the respective index to search on ?
    This way we don't have to either write our calls to api, or use fiddler etc tools. To have this capability handy on portal itself, will be benevolent in terms of testing, verifying data upload etc.

    10 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    2 comments  ·  Portal  ·  Flag idea as inappropriate…  ·  Admin →
  12. Ability to index more than 1M documents in Basic

    I have a basic account for $40\month, and have 1.1M documents. But I'm not able to add an additional partition. My only options are a) limit my documents to 1M (what's the point?) or b) upgrade to Standard which supports 15M documents (overkill) for $250/month?!? Of course, I could sign up two basic accounts for less, but then have to write specific code to query two search accounts. Please add the ability to add more partitions to the Basic Tier.

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Pricing and Quotas  ·  Flag idea as inappropriate…  ·  Admin →
  13. Make it so SearchMode=All matches if stop words are present even when querying across multiple analyzers

    Not all of the documents in our corpus are localized so we must search the default language (English) and the user's locale with the provided user query. Queries that contain a stop word for either default or user language analyzer causes no results to be returned.

    Example:
    The documents contain 2 searchable fields (plus an ID field).
    Title_EN -> English Analyzer
    Title_ES -> Spanish Analyzer

    A document exists with the following values:
    Title_EN -> "The best document"
    Title_ES -> "Un documento excelente"

    When a user inputs a query, we search on Title_EN and Title_ES. If the user provides the query…

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Query - Search  ·  Flag idea as inappropriate…  ·  Admin →
  14. wildcard sql query for node.js mobile services

    I want to be able to query using wildcard like so, [foo like '%?%', 'bar']
    I can't do this in node js on azure mobile services.

    Please add this. I can't add search feature on my app without this.

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  Crawlers  ·  Flag idea as inappropriate…  ·  Admin →
  15. configurable (or unlimited) hit highlights

    Per https://docs.microsoft.com/en-us/rest/api/searchservice/search-documents "Azure Search returns only up to 5 highlights per field. This limit is not configurable" >> this should be changed. Either by offering a configurable threshold, by pagination or any other way. We'd need to be able to access all hit results and not just a partial subset

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    completed  ·  1 comment  ·  Query - Search  ·  Flag idea as inappropriate…  ·  Admin →
  16. Get documents from an array of IDs

    an equivalent feature to that of the Ids Query in Elasticsearch (https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl-ids-query.html) to retrieve all documents whose Ids appear in a provided array of Ids.

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Flag idea as inappropriate…  ·  Admin →
  17. Enable gzip compression

    I'm adding fairly large documents via the rest API in fairly large batches. Being able to compress the JSON content seems like it would have a large impact on my ability to upload 10's of millions of documents more efficiently. Gzip for the response would be nice as well as there are times I wan't to get a large resultset in my response.

    9 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    1 comment  ·  REST API  ·  Flag idea as inappropriate…  ·  Admin →
  18. Most common searches by Analytics

    PowerBI can parse the Analytics data to obtain the "Most common search queries".
    It'd be awesome if the REST API could provide this information too.

    7 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  REST API  ·  Flag idea as inappropriate…  ·  Admin →
  19. Allow specifying query time analyzer

    Issueing a query with searchMode=all against multiple fields that use different analyzers may lead to unexpected results.

    Let’s take an index with two fields where one is analyzed with English Lucene analyzer, and the other with standard (default) analyzer.

    {
    "fields": [
    {"name": "docId", "type": "Edm.String", "key": true, "searchable": false},
    {"name": "field1", "type": "Edm.String", "analyzer": "en.lucene" },
    {"name": "field2", "type": "Edm.String" }
    ]
    }

    Let’s add these two documents:

    {
    "value": [
    {
    "docId": "1",
    "field1": "Waiting for a bus",
    "field2": "Exploring cosmos"
    },
    {
    "docId": "2",
    "field1": "Run to the hills",
    "field2": "run for your life"
    }
    ]
    }

    8 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Query - Search  ·  Flag idea as inappropriate…  ·  Admin →
  20. Support for parsing html/xml string content

    Many documents to be indexed already contain HTML or XML content.

    An example for us would be something like the following

    Article
    - string: Title
    - string/html: Summary
    - string/html: Body
    - string: author
    - datetime: created/updated date

    etc.

    It would be very helpful if the indexer could ignore HTML/XML/CSS/JavaScript content in the string. The current suggested alternative is to provide a surrogate field with the html tags stripped out of it and use that for the index.

    The challenge is then with hit highlighting, where we'd like to display the matching content. We'd rather not have to then find…

    7 votes
    Vote
    Sign in
    (thinking…)
    Sign in with: Microsoft
    Signed in as (Sign out)
    You have left! (?) (thinking…)
    0 comments  ·  Indexing  ·  Flag idea as inappropriate…  ·  Admin →
  • Don't see your idea?

Feedback and Knowledge Base