Skip to main content

A Demo For Understand Search Relevance In ElasticSearch Using Kibana

In this article, we will understand Relevance Search in ElasticSearch using Kibana.

In Elasticsearch relevance search means the output of the search can be done in 2 different ways like 'Precision', 'Recall'.

The 'Precision' result means the most accurate results from the ElasticSearch. This means it fetches exact matches of the search value. So in this search output will be more accurate, but the total output results count might be very very less.
  • Here all green color docs are more accurate documents for the search.
  • Here all red color docs are either less accurate documents or not matching documents for the search.
  • Here we can observe search results contain only accurate documents which means it is a 'Precision' result.
The 'Recall' results mean more in the count of the output results because it fetches partially matched data along with accurate data. Since it is fetching high count results, sometimes it skips the few accurate documents as well.
  • Here we can observe that the output search result contains both accurate and partially(not so accurate) documents.
  • Here we can observe a few accurate results left out from the search results.
  • So this search result is called the 'Recall' result.

Run ElasticSearch And Kibana Docker Containers:

Create a network first that can help to connect our services like 'Elasticsearch' & 'Kibana' under it.
Command to create a network:
docker network create your_network_name_any_name

Let's pull and create the Elasticsearch docker container.
Command To Create Elasticsearch Docker Container:

docker run -d --name your_container_name_any_name --net network_name_just_created -p 9200:9200 -e "discovery.type=single-node" elasticsearch:7.16.3
  • [ -d ] run the docker command in detach mode, which means runs as a background service.
  • [ --name your_container_name_any_name] define the name to the docker container.
  • [ --net network_name_just_created] specify the network name under which our service should run.
  • [ -p 9200:9200] right-hand side port number(fixed port number) is the default port number for the 'ElasticSearch', the left-hand side port number is exposing port number we can define any of our custom port numbers.
  • [-e "discovery.type=singlenode"] setting the environment variable to run our Elastic search on a single node. it can be changed for production applications.
  • [ elasticsearch:7.16.3] name of service and its version.
Let's pull and create the Kibana docker container.
Command To Create Kibana Docker Container:

docker run -d --name your_container_name_any_name --net network_name_just_created -p 5601:5601 kibana:7.16.3
  • [ -d ] run the docker command in detach mode, which means runs as a background service.
  • [ --name your_container_name_any_name] define the name to the docker container.
  • [ --net network_name_just_created] specify the network name under which our service should run.
  • [ -p 5601:5601] right-hand side port number(fixed port number) is the default port number for the 'Kibana', the left-hand side port number is exposing port number we can define any of our custom port numbers.
  • [ kibana:7.16.3] name of service and its version.

Upload Some Test Data Into Our ElasticSearch:

(1)Let's upload some test data into our elastic search. So from "https://perso.telecom-paristech.fr/eagan/class/igr204/datasets", this website downloads the 'Films' CSV file(any sample CSV file of your choice).

(2)Now open the 'Kibana' tool 'http://localhost:5601/app/home' and then click on the 'Upload A File' option.
(3) Now select the 'Film' CSV file and then click on the 'Import' Button.
(4) Create an index name for our test data.

Query To Retrieve Information About Document In An Index:

The syntax for Query to retrieve information about the document in an index.
GET Name_of_your_Index/_search

The sample query to retrieve information about the document in an index
GET film_info/_search
  • 'GET' - HTTP verb
  • 'film_info' - the name of the index.
  • '_search' - keyword of Elasticsearch.

Identify A Most Significant Text:

This section is totally an optional section, where I'm going to explain a query to get significant keywords from our Elastic store so that using those significant keywords we can frame a keyword to search against Elastic store so that we can understand the Relevance search results.

A sample query to get the Significant keywords.
GET film_info/_search
{
 "query": {
   "match": {
   "Subject": "Drama"
   }
  },
 "aggs": {
  "popular_kewords": {
    "significant_text": {
       "field": "Title"
     }
   }
  }
 }
  • GET - HTTP verbs
  • 'film_info' - the name of the index.
  • 'query' - Elasticsearch keyword.
  • 'match' - Elasticsearch keyword, inside match object, define 'Property Name' and its 'Value' for search against the documents.
  • 'aggs' - Elasticsearch keyword 'aggregation', the 'aggregation' summarizes our data as metrics, statics, or other analytics.
  • 'popular_keywords' - name of the 'aggregation'.
  • 'significant_text' - Elasticsearch keyword, inside of its define 'field' whose value must be property name of the document.

Now frame a sentence by using the above words to search against ElasticSearch to understand the Relevance search.

Recall Search Results:

Recall search result means it will bring the result if at least one word matched that means it will bring accurate and partial accurate data as well.

The sample query returns 'Recall Search Results'.
GET film_info/_search
{
   "query":
   {
      "match": {
          "Title":
             {
              "query": "until the last spring"
             }
      }
   }
}
  • GET - HTTP verb
  • film_info - the name of the index
  • _search - Elasticsearch keyword
  • query - Elasticsearch keyword
  • match - Elasticsearch keyword used to match data inside of it. Inside 'match' create an object with 'Property Name' of the document on which we want to search, inside of it add 'query' to it we have assigned our search keyword(eg: 'until the last spring').

Precision Search Result:

Precision search result, try to match every word in the search text against ElasticSearch data, here the position of the words don't matter. So, in this case, we will receive very less count results since every word must match.

The sample query returns 'Precision Search Result'.
GET film_info/_search
{
   "query":
     {
      "match":
      {
        "Title":
         {
          "query": "until the last spring",
         "operator": "and"
         }
       }
      }
}
  • 'operator' - Elasticsearch keyword and value is 'and' which means that need to match every word in the 'query' against the documents of the Elasticsearch.

Combine Precision & Recall Results:

To get more reasonable results we can combine 'Precision' & 'Recall' search results.

The sample query
GET film_info/_search
{
    "query":
   {
   "match":
  {
   "Title":
   {
   "query": "until the last spring",
   "minimum_should_match": 2
   }
   }
   }
}
  • 'minimum_should_match' - ElasticSearch keyword specifies minimum words of search keyword need to be matched against the documents.

Support Me!
Buy Me A Coffee PayPal Me

Video Session:

Wrapping Up:

Hopefully, I think this article delivered some useful information on Relevance Search in ElasticSearch. using I love to have your feedback, suggestions, and better techniques in the comment section below.

Follow Me:

Comments

Popular posts from this blog

.NET6 Web API CRUD Operation With Entity Framework Core

In this article, we are going to do a small demo on AspNetCore 6 Web API CRUD operations. What Is Web API: Web API is a framework for building HTTP services that can be accessed from any client like browser, mobile devices, desktop apps. In simple terminology API(Application Programming Interface) means an interface module that contains a programming function that can be requested via HTTP calls to save or fetch the data for their respective clients. Some of the key characteristics of API: Supports HTTP verbs like 'GET', 'POST', 'PUT', 'DELETE', etc. Supports default responses like 'XML' and 'JSON'. Also can define custom responses. Supports self-hosting or individual hosting, so that all different kinds of apps can consume it. Authentication and Authorization are easy to implement. The ideal platform to build REST full services. Create A .NET6 Web API Application: Let's create a .Net6 Web API sample application to accomplish our

Usage Of CancellationToken In Asp.Net Core Applications

When To Use CancellationToken?: In a web application request abortion or orphan, requests are quite common. On users disconnected by network interruption or navigating between multiple pages before proper response or closing of the browser, tabs make the request aborted or orphan. An orphan request can't deliver a response to the client, but it will execute all steps(like database calls, HTTP calls, etc) at the server. Complete execution of an orphan request at the server might not be a problem generally if at all requests need to work on time taking a job at the server in those cases might be nice to terminate the execution immediately. So CancellationToken can be used to terminate a request execution at the server immediately once the request is aborted or orphan. Here we are going to see some sample code snippets about implementing a CancellationToken for Entity FrameworkCore, Dapper ORM, and HttpClient calls in Asp.NetCore MVC application. Note: The sample codes I will show in

A Small Guide On NestJS Queues

NestJS Application Queues helps to deal with application scaling and performance challenges. When To Use Queues?: API request that mostly involves in time taking operations like CPU bound operation, doing them synchronously which will result in thread blocking. So to avoid these issues, it is an appropriate way to make the CPU-bound operation separate background job.  In nestjs one of the best solutions for these kinds of tasks is to implement the Queues. For queueing mechanism in the nestjs application most recommended library is '@nestjs/bull'(Bull is nodejs queue library). The 'Bull' depends on Redis cache for data storage like a job. So in this queueing technique, we will create services like 'Producer' and 'Consumer'. The 'Producer' is used to push our jobs into the Redis stores. The consumer will read those jobs(eg: CPU Bound Operations) and process them. So by using this queues technique user requests processed very fastly because actually

Blazor WebAssembly Custom Authentication From Scratch

In this article, we are going to explore and implement custom authentication from the scratch. In this sample, we will use JWT authentication for user authentication. Main Building Blocks Of Blazor WebAssembly Authentication: The core concepts of blazor webassembly authentication are: AuthenticationStateProvider Service AuthorizeView Component Task<AuthenticationState> Cascading Property CascadingAuthenticationState Component AuthorizeRouteView Component AuthenticationStateProvider Service - this provider holds the authentication information about the login user. The 'GetAuthenticationStateAsync()' method in the Authentication state provider returns user AuthenticationState. The 'NotifyAuthenticationStateChaged()' to notify the latest user information within the components which using this AuthenticationStateProvider. AuthorizeView Component - displays different content depending on the user authorization state. This component uses the AuthenticationStateProvider

How Response Caching Works In Asp.Net Core

What Is Response Caching?: Response Caching means storing of response output and using stored response until it's under it's the expiration time. Response Caching approach cuts down some requests to the server and also reduces some workload on the server. Response Caching Headers: Response Caching carried out by the few Http based headers information between client and server. Main Response Caching Headers are like below Cache-Control Pragma Vary Cache-Control Header: Cache-Control header is the main header type for the response caching. Cache-Control will be decorated with the following directives. public - this directive indicates any cache may store the response. private - this directive allows to store response with respect to a single user and can't be stored with shared cache stores. max-age - this directive represents a time to hold a response in the cache. no-cache - this directive represents no storing of response and always fetch the fr

Part-1 Angular JWT Authentication Using HTTP Only Cookie[Angular V13]

In this article, we are going to implement a sample angular application authentication using HTTP only cookie that contains a JWT token. HTTP Only JWT Cookie: In a SPA(Single Page Application) Authentication JWT token either can be stored in browser 'LocalStorage' or in 'Cookie'. Storing JWT token inside of the cookie then the cookie should be HTTP Only. The HTTP-Only cookie nature is that it will be only accessible by the server application. Client apps like javascript-based apps can't access the HTTP-Only cookie. So if we use authentication with HTTP only JWT cookie then we no need to implement custom logic like adding authorization header or storing token data, etc at our client application. Because once the user authenticated cookie will be automatically sent to the server by the browser on every API call. Authentication API: To implement JWT cookie authentication we need to set up an API. For that, I had created a mock authentication API(Using the NestJS Se

Unit Testing Asp.NetCore Web API Using xUnit[.NET6]

In this article, we are going to write test cases to an Asp.NetCore Web API(.NET6) application using the xUnit. xUnit For .NET: The xUnit for .Net is a free, open-source, community-focused unit testing tool for .NET applications. By default .Net also provides a xUnit project template to implement test cases. Unit test cases build upon the 'AAA' formula that means 'Arrange', 'Act' and 'Assert' Arrange - Declaring variables, objects, instantiating mocks, etc. Act - Calling or invoking the method that needs to be tested. Assert - The assert ensures that code behaves as expected means yielding expected output. Create An API And Unit Test Projects: Let's create a .Net6 Web API and xUnit sample applications to accomplish our demo. We can use either Visual Studio 2022 or Visual Studio Code(using .NET CLI commands) to create any.Net6 application. For this demo, I'm using the 'Visual Studio Code'(using the .NET CLI command) editor. Create a fo

Different HttpClient Techniques To Consume API Calls In Minimal API[.NET6]

In this article, we are going to implement different HttpClient techniques to consume API calls in minimal API. The different HttpClient techniques that we are going to explore are like: Register HttpClient Object Explicitly In DI(Dependency Injection Service) Named Client Type Client HttpRequestMessage Object Create A .NET6 Minimal API Project: Let's create a .Net6 Minimal API sample project to accomplish our demo. We can use either Visual Studio 2022 or Visual Studio Code(using .NET CLI commands) to create any.Net6 application. For this demo, I'm using the 'Visual Studio Code'(using the .NET CLI command) editor. CLI command For Minimal API Project dotnet new webapi -minimal -o Your_Project_Name Create A Third Party API Response Model: Here I'm going to use a free third-party rest API that is "https://jsonplaceholder.typicode.com/posts". So to receive the response let's create a response model like 'Post.cs'. Program.cs:(Add Post.cs c

.Net5 Web API Managing Files Using Azure Blob Storage

In this article, we are going to understand the different file operations like uploading, reading, downloading, and deleting in .Net5 Web API application using Azure Blob Storage. Azure Blob Storage: Azure blob storage is Microsoft cloud storage. Blob storage can store a massive amount of file data as unstructured data. The unstructured data means not belong to any specific type, which means text or binary data. So something like images or pdf or videos to store in the cloud, then the most recommended is to use the blob store. The key component to creating azure blob storage resource: Storage Account:- A Storage account gives a unique namespace in Azure for all the data we will save. Every object that we store in Azure Storage has an address. The address is nothing but the unique name of our Storage Account name. The combination of the account name and the Azure Storage blob endpoint forms the base address for each object in our Storage account. For example, if our Storage Account is n

.Net5 Web API Redis Cache Using StackExchange.Redis.Extensions.AspNetCore Library

In this article, we are going to explore the integration of Redis cache in .Net5 Web API application using the 'StackExchange.Redis.Exntensions' library. Note:- Microsoft has introduced an 'IDistributedCache' interface in dotnet core which supports different cache stores like In-Memory, Redis, NCache, etc. It is simple and easy to work with  'IDistributedCache', for the Redis store with limited features but if we want more features of the Redis store we can choose to use 'StackExchange.Redis.Extensions'.  Click here for Redis Cache Integration Using IDistributedCache Interface . Overview On StackExchange.Redis.Extnesions Library: The 'StackExchange.Redis.Extension' library extended from the main library 'StackExchange.Redis'. Some of the key features of this library like: Default serialization and deserialization. Easy to save and fetch complex objects. Search key. Multiple Database Access Setup Redis Docker Instance: For this sampl