A Novel Multi-Modal Image Retrieval System by Researchers from Gwangju Institute of Science and Technology

Researchers from Korea develop a new image retrieval system using deep learning algorithms

GWANGJU, South Korea, Nov. 9, 2022 /PRNewswire/ — With the amount of information on the internet increasing by the minute, retrieving data from it is like trying to find a needle in a haystack. Content-based image retrieval (CBIR) systems are capable of retrieving desired images based on the user’s input from an extensive database. These systems are used in e-commerce, face recognition, medical applications, and computer vision. There are two ways in which CBIR systems work—text-based and image-based. One of the ways in which CBIR gets a boost is by using deep learning (DL) algorithms. DL algorithms enable the use of multi-modal feature extraction, meaning that both image and text features can be used to retrieve the desired image. Even though scientists have tried to develop multi-modal feature extraction, it remains an open problem.

To this end, researchers from Gwangju Institute of Science and Technology have developed DenseBert4Ret, an image retrieval system using DL algorithms. The study, led by Prof. Moongu Jeon and Ph.D. student Zafran Khan, was made available online on September 14, 2022 and published in Volume 612 of Information Sciences. “In our day-to-day lives, we often scour the internet to look for things such as clothes, research papers, news article, etc. When these queries come into our mind, they can be in the form of both images and textual descriptions. Moreover, at times we may wish to amend our visual perceptions through textual descriptions. Thus, retrieval systems should also accept queries as both texts and images,” says Prof. Jeon, explaining the team’s motivation behind the study.

The proposed model had both image and text as the input query. For extracting the image features from the input, the team used a deep neural network model known as DenseNet-121. This architecture allowed for the maximum flow of information from the input to the output layer and needed tuning of very few parameters during training. DenseNet-121 was combined with the bidirectional encoder representation from transformer (BERT) architecture for extracting semantic and contextual features from the text input. The combination of these two architectures reduced training time and computational requirements and formed the proposed model, DenseBert4Ret.

The team then used Fashion200k, MIT-states, and FashionIQ, three real-world datasets, to train and compare the proposed system’s performance against the state-of-the-art systems. They found that DenseBert4Ret showed no loss during image feature extraction and outperformed the state-of-the-art models. The proposed model successfully catered for multi-modalities that were given as the input with the multi-layer perceptron and triple loss function helping to learn the joint features.

“Our model can be used anywhere where there is an online inventory and images need to be retrieved. Additionally, the user can make changes to the query image and retrieve the amended image from the inventory,” concludes Prof. Jeon.

Here’s hoping to see the DenseBert4Ret system in application in our everyday-use search engines soon!

Reference

Title of original paper: DenseBert4Ret: Deep bi-modal for image retrieval

Journal: Information Sciences

DOI: https://doi.org/10.1016/j.ins.2022.08.119

Affiliations:      

  1. School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology   (GIST)
  2. National University of Science and Technology (NUST), Pakistan
  3. Department of Computer Engineering, Dankook University

*Corresponding author’s email: mgjeon@gist.ac.kr

About the Gwangju Institute of Science and Technology (GIST)

Website: http://www.gist.ac.kr/

Contact:

Chang Sung Kang
82 62 715 6253
348316@email4pr.com

SOURCE Gwangju Institute of Science and Technology (GIST)