Content-Based Information Retrieval (CBIR) Research Papers (original) (raw)
Image features play a vital role in image retrieval. This chapter presents the use of Zernike moment features for retrieving the binary and gray level images from established image databases. To retrieve a set of similar category of... more
Image features play a vital role in image retrieval. This chapter presents the use of Zernike moment features for retrieving the binary and gray level images from established image databases. To retrieve a set of similar category of images from an image database, up to 25 Zernike moment features from order zero to order 8 were utilized and exp erimented in this chapter. A total of 1400 binary images from MPEG-7 dataset and 1440 images from a COIL-20 dataset were used to evaluate the capability of Zernike moments features for image retrieval. The exp erimental results show that Zernike moments implementation is suitable for image retrieval due to rotation invariance and fast computation.
On-line keyword searching from documents in Chinese tends to use inverted indexing as the main technique, which has its difficulties. Suffix Array is widely used for processing text in Western languages. However, it fails to get widely... more
On-line keyword searching from documents in Chinese tends to use inverted indexing as the main technique, which has its difficulties. Suffix Array is widely used for processing text in Western languages. However, it fails to get widely used in Chinese processing because of the speciality of Chinese. Suffix Array is a powerful tool. However it costs too much space. That is the major bottleneck of suffix Array. A data
structure called Keyword-driven Suffix Array is proposed in this paper for on-line keyword searching from documents in Chinese, based on observation of on-line search pattern and traits of Chinese. Space efficiency is improved a lot using this data structure. When the document database is large enough, space efficiency is improved by about 5/6 using this data structure without sacrificing its time efficiency.
Images from the minute it was invented, has had an immense impact on the world we live in. The extracting the required images from the World Wide Web (WWW) is very difficult because web contains a huge number of images. To solve this... more
Images from the minute it was invented, has had an immense impact on the world we live in. The extracting the required images from the World Wide Web (WWW) is very difficult because web contains a huge number of images. To solve this problem we need a system that can retrieve the required images needed by the user. Image Crawler is a web based tool that collects and indexes group of web images available on the internet. This tool collects the keyword or phrase from the user to retrieve the images from the web. Then these collected keyword is applied to the different general search tools like Google, Yahoo, Bind etc,. The collected web page information is stored in the temporary file till 200KB file size from the server. Then this file content will be scanned and extract the image URL's and it is compared the URL which is present in the database to avoid the duplicate downloads. The extracted URL's images are downloaded and finally stores unique image and corresponding metadata like filename, url, size etc. in the database. In this paper we present the designing of an Image crawler tool. We build a search tool which is flexible, general-purpose image search framework and explore a directed result aggregating and removing of duplicates to achieve top results compared to other existing search tools. Finally this resulted images are used in the Content Based Image Retrieval (CBIR) system for extracting the relevant images need by the client using the content of the images rather than the text based information.
Abstract:-A fuzzy approach to color quantization and its application in content-based image indexing and retrieval is presented in this paper. There are some drawbacks associated with traditional color histogram such as color... more
Abstract:-A fuzzy approach to color quantization and its application in content-based image indexing and retrieval is presented in this paper. There are some drawbacks associated with traditional color histogram such as color quantization. The goal in this paper is to develop an effective method to overcome several limitations, which are related to traditional color histogram. In the proposed method, fuzzy logic is used to quantize color space and apply it in color histogram construction.
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications in texture analysis are... more
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications in texture analysis are discussed in the context of Content Based Image Retrieval (CBIR). The theoretical extension of GLCM to n-dimensional gray scale images are also discussed. The results indicate that trace features outperform Haralick features when applied to CBIR.
—The accuracy and efficiency are two main issues in the context of image retrieval system during searching, indexing and retrieving images from huge databases. In this paper, multi-level color and texture feature extraction has been... more
—The accuracy and efficiency are two main issues in the context of image retrieval system during searching, indexing and retrieving images from huge databases. In this paper, multi-level color and texture feature extraction has been performed for Content Based Image Retrieval (CBIR) system. The proposed method implements color moment descriptor over Local Binary Patterns (LBP) texture feature. The CBIR system has been designed for Amsterdam Library of texture (ALOT) images. The proposed method has been compared with single color moment descriptor, LBP texture feature and then combined color and texture separately. The experimental results were generated by using MATLAB which shows that accuracy and efficiency of proposed method are considerably higher in terms of overall precision, recall and its retrieval time.
In this paper, we focus on cross-modal (visual and textual) e-commerce search within the fashion domain. Particularly, we investigate two tasks: 1) given a query image, we retrieve textual descriptions that correspond to the visual... more
In this paper, we focus on cross-modal (visual and textual) e-commerce search within the fashion domain. Particularly, we investigate two tasks: 1) given a query image, we retrieve textual descriptions that correspond to the visual attributes in the query; and 2) given a textual query that may express an interest in specific visual product characteristics, we retrieve relevant images that exhibit the required visual attributes. To this end, we introduce a new dataset that consists of 53,689 images coupled with textual descriptions. The images contain fashion garments that display a great variety of visual attributes, such as different shapes, colors and textures in natural language. Unlike previous datasets, the text provides a rough and noisy description of the item in the image. We extensively analyze this dataset in the context of cross-modal e-commerce search. We investigate two state-of-the-art latent variable models to bridge between textual and visual data: bilingual latent Dirichlet allocation and canonical correlation analysis. We use state-of-the-art visual and textual features and report promising results.
The purpose of this report is to describe research and solution to the problem of designing a Content Based Image Retrieval, CBIR system. It outlines the problem, the proposed solution, the final solution and the accomplishments achieved.... more
The purpose of this report is to describe research and solution to the problem of designing a Content Based Image Retrieval, CBIR system. It outlines the problem, the proposed solution, the final solution and the accomplishments achieved. Due to the enormous increase in image database sizes, as well as its vast deployment in various applications, the need for CBIR development arose. Firstly, this report outlines a description of the primitive features of an image; texture, colour, and shape. These features are extracted and used as the basis for a similarity check between images. The algorithms used to calculate the similarity between extracted features, are then explained. Our final result was a MatLab built software application, with an image database, that utilized texture and colour features of the images in the database as the basis of comparison and retrieval. The structure of the final software application is illustrated. Furthermore, the results of its performance are illustrated by a detailed example.
Deep learning and image processing are two areas of great interest to academics and industry professionals alike. The areas of application of these two disciplines range widely, encompassing fields such as medicine, robotics, and security... more
Deep learning and image processing are two areas of great interest to academics and industry professionals alike. The areas of application of these two disciplines range widely, encompassing fields such as medicine, robotics, and security and surveillance. The aim of this book, 'Deep Learning for Image Processing Applications', is to offer concepts from these two areas in the same platform, and the book brings together the shared ideas of professionals from academia and research about problems and solutions relating to the multifaceted aspects of the two disciplines. The first chapter provides an introduction to deep learning, and serves as the basis for much of what follows in the subsequent chapters, which cover subjects including: the application of deep neural networks for image classification; hand gesture recognition in robotics; deep learning techniques for image retrieval; disease detection using deep learning techniques; and the comparative analysis of deep data and big data. The book will be of interest to all those whose work involves the use of deep learning and image processing techniques
Scope of the book: This book focusses on the technical concepts of deep learning and its associated branch Neural Networks for the various dimensions of image processing applications. The proposed volume intends to bring together... more
Scope of the book:
This book focusses on the technical concepts of deep learning and its associated branch Neural Networks for the various dimensions of image processing applications. The proposed volume intends to bring together researchers to report the latest results or progress in the development of the above-mentioned areas. Since there is a deficit of books on this specific subject matter, the editors aim to provide a common platform for researchers working in this area to exhibit their novel findings.
Topics of Interest:
This book solicits contributions, which include the fundamentals in the field of Deep Artificial Neural Networks and Image Processing supported by case studies and practical examples. Each chapter is expected to be self-contained and to cover an in-depth analysis of real life applications of neural networks to image analysis.
This article summarizes the research on computational methods to detect academic plagiarism by systematically reviewing 239 research papers published between 2013 and 2018. To structure the presentation of the research contributions, we... more
This article summarizes the research on computational methods to detect academic plagiarism by systematically reviewing 239 research papers published between 2013 and 2018. To structure the presentation of the research contributions, we propose novel technically oriented typologies for plagiarism prevention and detection efforts, the forms of academic plagiarism, and computational plagiarism detection methods. We show that academic plagiarism detection is a highly active research field. Over the period we review, the field has seen major advances regarding the automated detection of strongly obfuscated and thus hard-to-identify forms of academic plagiarism. These improvements mainly originate from better semantic text analysis methods , the investigation of non-textual content features, and the application of machine learning. We identify a research gap in the lack of methodologically thorough performance evaluations of plagiarism detection systems. Concluding from our analysis, we see the integration of heterogeneous analysis methods for tex-tual and non-textual content features using machine learning as the most promising area for future research contributions to improve the detection of academic plagiarism further.
With the increasingly growing size of digital image collections, known image search is gaining more and more importance. Especially when the objects in such collections do not possess appropriate metadata (e.g., tags, annotations),... more
With the increasingly growing size of digital image collections, known image search is gaining more and more importance. Especially when the objects in such collections do not possess appropriate metadata (e.g., tags, annotations), content-based image retrieval (CBIR) is a promising approach. However, the application of CBIR to known item search usually suffers from the unavailability of query images that are good enough to express the user's information need. In order to improve this situation, we propose the QbS system which provides an approach to content-based search in large image collections based on user-drawn sketches. By exploiting novel devices for human-computer interaction like interactive paper, tablet PCs, or graphic tablets, users are able to draw a sketch that reflects their information need and start a content-based search using this sketch. The QbS system provides query support and offers several invariances that allow the user-generated sketch to slightly deviate from the searched image in terms of rotation, translation, relative size, and/or unknown objects in the background. To illustrate the benefits of the approach, we show search results from the evaluation of QbS on the basis of the MIRFLICKR collection with 25'000 objects.
Существует широкий круг задач, где требуется анализ, аудио-визуальных моделей реальности. В частности, для многих военных и гражданских приложений, необходимо наличие поиска нечетких дубликатов видео. Для мирного применения, — это... more
Существует широкий круг задач, где требуется анализ, аудио-визуальных моделей реальности. В частности, для многих военных и гражданских приложений, необходимо наличие поиска нечетких дубликатов видео. Для мирного применения, — это группировка подсказок в выдаче поисковиков, поиск неправомерных публикаций видеофайлов, сокращение объема хранимой и передаваемой информации. Для военного применения можно тоже перечислить несколько вариантов.
This article summarizes the research on computational methods to detect academic plagiarism by systematically reviewing 239 research papers published between 2013 and 2018. To structure the presentation of the research contributions, we... more
This article summarizes the research on computational methods to detect academic plagiarism by systematically reviewing 239 research papers published between 2013 and 2018. To structure the presentation of the research contributions, we propose novel technically oriented typologies for plagiarism prevention and detection efforts, the forms of academic plagiarism, and computational plagiarism detection methods. We show that academic plagiarism detection is a highly active research field. Over the period we review, the field has seen major advances regarding the automated detection of strongly obfuscated and thus hard-to-identify forms of academic plagiarism. These improvements mainly originate from better semantic text analysis methods, the investigation of non-textual content features, and the application of machine learning. We identify a research gap in the lack of methodologically thorough performance evaluations of plagiarism detection systems. Concluding from our analysis, we s...
Growing importance of global English in professional contexts has been the rise of ESP teaching at all levels. ESP is oriented towards mastering skills for professional communication. According to Osama Khalifa (2012), ESP teaching... more
Growing importance of global English in professional contexts has been the rise of ESP teaching at all levels. ESP is oriented towards mastering skills for professional communication. According to Osama Khalifa (2012), ESP teaching produces many problems such as problems for learners, problems for teachers, problems of methodology and materials, and problems of assessment and testing can be solved which require better teachers and better training. They need to understand the nature of the ESP teaching, they have to be able to observe and organize the learner's progress and to diagnose his problems. They have to be familiar with the widest possible range of alternative teaching techniques, learning requirements and the learner's needs. ESP teacher need to exercise professionalism based on training and experience. The structure and content of language curricula, when modern languages were first introduced into universities, were much influenced by the need to achieve academic respectability (Radford, 1985). Language curricula is therefore dominated by the systematic study of grammar, the regular carrying out of translation exercises into and out of language, the close study of set literary and science texts, the broad study of philological studies of the language (Maddock, 1994). The core of content based instruction is the lofty ideal that language instruction cannot be devoid of the context in which it is presented, and moreover, that content rather than language structures should be the driving force in curriculum development (Erickson & Schulz, 1981). According to freiermuth(2001), Content based teaching generally adhere to the idea that content is superior to grammaticality of language structures, but that form and content need to be integrated into language instruction.
Content based image retrieval (CBIR) has become one of the most active research areas in the past few years. Many indexing techniques are based on global feature distributions. However, these global distributions have limited... more
Content based image retrieval (CBIR) has become one of the most active research areas in the past few years. Many indexing techniques are based on global feature distributions. However, these global distributions have limited discriminating power because they are unable to capture local image information. In this paper, we propose a content-based image retrieval method which combines color and texture features. To improve the discriminating power of color indexing techniques, we encode a minimal amount of spatial information in the color index. As its color features, an image is divided horizontally into three equal non-overlapping regions. From each region in the image, we extract the first three moments of the color distribution, from each color channel and store them in the index i.e., for a HSV color space, we store 27 floating point numbers per image. As its texture feature, Gabor texture descriptors are adopted. We assign weights to each feature respectively and calculate the similarity with combined features of color and texture using Canberra distance as similarity measure. Experimental results show that the proposed method has higher retrieval accuracy than other conventional methods combining color moments and texture features based on global features approach.
content based Image retrieval system
Resumen. El enorme crecimiento de imágenes almacenadas en internet crea nuevas e insospechadas oportunidades para el avance la ciencia y la tecnología. Sin embargo, para hacer un uso óptimo de esa información se requiere progresar... more
Resumen. El enorme crecimiento de imágenes almacenadas en internet crea nuevas e insospechadas oportunidades para el avance la ciencia y la tecnología. Sin embargo, para hacer un uso óptimo de esa información se requiere progresar sustancialmente en la forma en que las maquinas actuales procesan las imágenes. Es por ello que se hacen esfuerzos en ese sentido, siendo uno de ellos estandarizar el proceso formándose así el Moving Picture Expert Group. En este artículo se presentan y analizan algunos trabajos representativos del estado del arte en esta área de oportunidad. Palabras clave: MPEG-7, descriptores, color, textura, forma, CBIR. Abstract. The enormous growth of images it stored in Internet creates new and unexpected opportunities for the advancement of science and technology. However, to make optimal use of this information requires substantial progress in the way that the current machines process images. So there are efforts in this direction, one of which is to standardize the process thus forming the Moving Picture Expert Group. In this paper we present and analyze some representative works of the state of the art in this opportunity area.
Most news organizations provide immediate access to topical news broadcasts through RSS streams or podcasts. Until recently, applications have not permitted a user to perform content based search within a longer spoken broadcast to find... more
Most news organizations provide immediate access to topical news broadcasts through RSS streams or podcasts. Until recently, applications have not permitted a user to perform content based search within a longer spoken broadcast to find the segment that might interest them. Recent progress in both automatic speech recognition (ASR) and natural language processing (NLP) has produced robust tools that allow us to now provide users with quicker and more focused access to relevant segments of news broadcast videos. Our public online demonstrator of the Voxalead application currently indexes daily broadcast news content from 50 sources in English, French, Chinese, Arabic, Spanish, Dutch, Italian and Russian.
Data available on the web is growing at an exponential rate, creating Knowledge or extracting information is of paramount importance. Information Retrieval (IR) plays a crucial role in Knowledge management as it helps us to find the... more
Data available on the web is growing at an exponential rate, creating Knowledge or extracting information is of paramount importance. Information Retrieval (IR) plays a crucial role in Knowledge management as it helps us to find the relevant information from the existing data. This paper compares the performance of keyword-based retrieval and other architectural styles in information retrieval system with ontology-based retrieval on documents in regional language
Direttore ALBERTO PETRUCCIANI Comitato di direzione PAOLA CASTELLUCCI (coordinamento redazionale), GIOVANNI PAOLONI, MARINA RAFFAELI, FRANCESCA SANTONI Segreteria di redazione SAMANTA SEGATORI Hanno collaborato a questo volume: Conto... more
Direttore ALBERTO PETRUCCIANI Comitato di direzione PAOLA CASTELLUCCI (coordinamento redazionale), GIOVANNI PAOLONI, MARINA RAFFAELI, FRANCESCA SANTONI Segreteria di redazione SAMANTA SEGATORI Hanno collaborato a questo volume: Conto corrente postale 12707501 Abbonamento annuo 2014 ISTITUZIONI -INSTITUTIONS La quota per le istituzioni è comprensiva dell'accesso on-line alla rivista. Indirizzo IP e richieste di informazioni sulla procedura di attivazione dovranno essere inoltrati a periodici@olschki.it Subscription rates for institutions includes on-line access to the journal. The IP address and requests for information on the activation procedure should be sent to periodici@olschki.it Italia € 108,00 Foreign € 126,00 (solo on-lineon-line only € 89,00) PRIVATI -INDIVIDUALS (solo cartaceoprint version only) Italia: € 84,00 Foreign € 108,00
Color histograms are one of the earliest and best known image features used in Content-Based Image Retrieval (CBIR). There is a wealth of scientific work on this topic. However, different papers vary in the specific ways of determining... more
Color histograms are one of the earliest and best known image features used in Content-Based Image Retrieval (CBIR). There is a wealth of scientific work on this topic. However, different papers vary in the specific ways of determining histograms and distance between them. In this paper authors attempt to classify various types of histograms used in literature and compare them using contemporary datasets and metrics for evaluation. Histograms are compared based on their retrieval performance as well as resource usage.
The purpose of this report is to describe research and solution to the problem of designing a Content Based Image Retrieval, CBIR system. It outlines the problem, the proposed solution, the final solution and the accomplishments achieved.... more
The purpose of this report is to describe research and solution to the problem of designing a Content Based Image Retrieval, CBIR system. It outlines the problem, the proposed solution, the final solution and the accomplishments achieved. Due to the enormous increase in image database sizes, as well as its vast deployment in various applications, the need for CBIR development arose. Firstly, this report outlines a description of the primitive features of an image; texture, colour, and shape. These features are extracted and used as the basis for a similarity check between images. The algorithms used to calculate the similarity between extracted features, are then explained. Our final result was a MatLab built software application, with an image database, that utilized texture and colour features of the images in the database as the basis of comparison and retrieval. The structure of the final software application is illustrated. Furthermore, the results of its performance are illustrated by a detailed example.
Keywords: Content Based Image Retrieval(CBIR), Similarity, Features and Image Database.
The development of content based retrieval mechanisms is a very active research area. Present studies are mainly focused on automating the information extraction and indexing processes. Usually for the development and evaluation of... more
The development of content based retrieval mechanisms is a very active
research area. Present studies are mainly focused on automating the information
extraction and indexing processes. Usually for the development and evaluation of such
mechanisms there is always a need for a ground-truth database. In this paper we present a
software tool named qp that is able to semi-automatically produce a collection of random
3D vessels, with morphological characteristics similar to those found in ancient Greek
pottery, a ceramic group exhibited worldwide with great impact to scholars as well as
general public. A 3D vessel collection has been produced by qp and can be used as a test
bed dataset for the development of shape-based 3D descriptors applicable to pottery.
Additionally, qp can be considered as a 3D vessel modelling software tool which can be
used by people not related to computer graphics technology and particularly to 3D
modelling.
This Master project aimed at studying, analyzing and developing relevance feedback (RF) techniques to enhance similarity queries that employ the content-based image retrieval (CBIR) approach. The motivation to develop this project came... more
This Master project aimed at studying, analyzing and developing relevance feedback (RF) techniques to enhance similarity queries that employ the content-based image retrieval (CBIR) approach. The motivation to develop this project came from the iRIS (internet Retrieval of Images System), which is a Web server prototype to process similarity queries. The iRIS can be integrated to a PACS (Picture and Archiving and Communication System) adding the functionality of retrieval images comparing their inherent alikeliness. The main reservation about using CBIR techniques is the semantic gap, because the general use of low level features to describe the images. The low level features, such as color, texture and shape, mostly cannot bridge the gap between what the users expect/want to what they get, generating disappointment and refusal of employing the system. However, if the user is allowed to interact with the system, classifying the query results and using such information on refinement steps, the queries can be reprocessed and the results tend to comply with the users’ expectation. This is just the core of the relevance feedback techniques. Looking at this scenario, this project developed two relevance feedback (RF) techniques: the RF Projection and the RF Multiple Point Projection. The improvements on the similarity queries were expressive going to up 29% with only one interaction, and to 42% on the fifth interaction, when compared to the original query. Experiments performed with users, have shown us that in average they run 3 iterations before get satisfactory results. By the results given by the experiment, one can claim that RF is a powerful approach to improve the use of CBIR systems and enhance similarity queries.
The rapid growth of the worldwide web poses unprecedented scaling challenges for general-purpose crawlers and search engines. In this paper we describe a new hypertext information management system called a Focused Crawler. The goal of a... more
The rapid growth of the worldwide web poses unprecedented scaling challenges for general-purpose crawlers and search engines. In this paper we describe a new hypertext information management system called a Focused Crawler. The goal of a focused crawler is to selectively seek out pages that are relevant to a pre-defined set of topics. The topics are specified not using keywords, but using exemplary documents. Rather than collecting and indexing all accessible hypertext documents to be able to answer all possible ad-hoc queries, a focused crawler analyzes its crawl boundary to find the links that are likely to be most relevant for the crawl. It avoids irrelevant regions of the web. This leads to significant savings in hardware and network resources, and helps keep the crawl more up-to-date. We describe a prototype implementation that is comprised of three programs integrated via a re-lational database: a crawler that stores metadata for documents and links in relational tables and is guided by various attributes in these tables, a hypertext classifier that updates the metadata with topic information from a large taxonomy (such as Yahoo!), and a rating system that updates metadata fields signifying the value of a page as a access point for a large number of relevant pages. The two " mining " modules guide the crawler away from unnecessary exploration and focus its efforts on web regions of interest. We report on extensive focused-crawling experiments using several topics at different levels of speci-ficity. Focused crawling acquires relevant pages steadily while standard crawling quickly loses its way, even though they are started from the same root set. Focused crawling is robust against large perturbations in the starting set of URLs. It discovers largely overlapping sets of resources in spite of these perturbations. It is also capable of exploring out and discovering valuable resources that are dozens of links away from the start set, while carefully pruning the millions of pages that may lie within this same radius. Our anecdotes suggest that focused crawling is very effective for building high-quality collections of web documents on specific topics, using modest desktop hardware.
This paper presents an alternative approach for Content Based Image Retrieval (CBIR) using Scale Invariant Feature Transform (SIFT) algorithm for binary and gray scale images. The motivation to use SIFT algorithm for CBIR is due to the... more
This paper presents an alternative approach for Content Based Image Retrieval (CBIR) using Scale Invariant Feature Transform (SIFT) algorithm for binary and gray scale images. The motivation to use SIFT algorithm for CBIR is due to the fact that SIFT is invariant to scale, rotation and translation as well as partially invariant to affine distortion and illumination changes. Inspired by these facts, this paper investigates the fundamental properties of SIFT for robust CBIR by using MPEG-7, COIL-20 and ZuBuD image databases. Our approach uses detected keypoints and its descriptors to match between the query images and images from the database. Our experimental results show that the proposed CBIR using SIFT algorithm producing excellent retrieval result for images with many corners as compared to retrieving image with less corners.
In this paper we describe an efficient implementation of an IEEE 754 single precision floating point multiplier targeted for Xilinx Virtex-5 FPGA. VHDL is used to implement a technology-independent pipelined design. The multiplier... more
In this paper we describe an efficient implementation of an IEEE 754 single precision floating point multiplier targeted for Xilinx Virtex-5 FPGA. VHDL is used to implement a technology-independent pipelined design. The multiplier implementation handles the overflow and underflow cases. Rounding is not implemented to give more precision when using the multiplier in a Multiply and Accumulate (MAC) unit. With latency of three clock cycles the design achieves 301 MFLOPs. The multiplier was verified against Xilinx floating point multiplier core.
CBIR is a very important domain, especially in the last decade due to the increased need for image retrieval from the multimedia database. In general, we extract low level (color, texture, and shape) or high-level features (when we... more
CBIR is a very important domain, especially in the
last decade due to the increased need for image retrieval
from the multimedia database. In general, we extract low
level (color, texture, and shape) or high-level features (when
we include machine learning techniques) from the images. In
our work, we proposed a new CBIR system using Local
Neighbor Pattern (LNP) with supervised machine learning
techniques. We evaluated the performance of this system by
comparing the system with Local Tetra Pattern (LTrP) using
Corel 1k, Vistex and TDF face databases. We used three
types of the database (i.e color, texture and face databases) to
improve the effectiveness of our system. Performance
analysis shows that LNP gives better performance regarding
the average recall than LBP, LDP, and LTrP. To increase
the accuracy of this system we used the LNP method with
machine learning techniques and performance analysis
shows that local pattern with machine learning techniques
improves the average accuracy from 36.23% to 85.60% when
we use LNP with cubic SVM on DB1 (Corel1K), and from
82.51% to 99.50 % when we use LNP with fine KNN on DB2
(Vistex DB), and from 56.63% to 95% when we use LNP with
ensemble subspace discriminant on DB3 (face DB).
Image content on the Web is increasing exponentially. As a result, there is a need for image retrieval systems. Historically, there have been two methodologies, text-based and content-based. In the text-based approach, query systems... more
Image content on the Web is increasing exponentially. As a result, there is a need for image retrieval systems. Historically, there have been two methodologies, text-based and content-based. In the text-based approach, query systems retrieve images that have been manually annotated using key words. This approach can be problematic: it is labor-intensive and maybe biased according to the subjectivity of the observer. Content based image retrieval (CBIR) searches and retrieves digital images in large databases by analysis of derived-image features. CBIR systems typically use the characteristics of color, texture, shape and their combination for definition of features. Similarity measures that originated in the preceding text-based era are commonly used. However, CBIR struggles with bridging the semantic gap, defined as the division between high-level complexity of CBIR and human perception and the low-level implementation features and techniques. In this paper, CBIR is reviewed in a broad context. Newer approaches is feature generation and similarity measures are detailed with representative studies addressing their efficacy. Color-texture moments, columns-of-interest, harmonysymmetry-geometry, SIFT (Scale Invariant Feature Transform), and SURF (Speeded Up Robust Features) are presented as alternative feature generation modalities. Graph
matching, Earth Mover’s Distance, and relevance feedback are discussed with the realm of similarity. We conclude that while CBIR is evolving and continues to slowly close the semantic gap, addressing the complexity of human perception remains a challenge.
—Finding accurate positions of mobile devices based on visual information involves searching for query-matching images in a very large dataset, typically containing millions of images. Although the main problem is designing a reliable... more
—Finding accurate positions of mobile devices based on visual information involves searching for query-matching images in a very large dataset, typically containing millions of images. Although the main problem is designing a reliable image retrieval engine, accurate localization also depends on a good fusion algorithm between the GPS data (geo-tags) of each query-matching image and the query image. This paper proposes a new method for reliable estimation of the actual query camera position (geo-tag) by applying structure from motion (SFM) with bundle adjustment for sparse 3D camera position reconstruction, and a linear rigid transformation between two different 3D Cartesian coordinate systems. The experimental results on more than 170 query images show the proposed algorithm returns accurate results for a high percentage of the samples. The error range of the estimated query geo-tag is compared with other related research and indicates an average error less than 5 meters that improves on some of the published works.
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications in texture analysis are... more
Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications in texture analysis are discussed in the context of Content Based Image Retrieval (CBIR). The theoretical extension of GLCM to n-dimensional gray scale images are also discussed. The results indicate that trace features outperform Haralick features when applied to CBIR.
Image retrieval is a topic where scientific interest is currently high. The important steps associated with image retrieval system are the extraction of discriminative features and a feasible similarity metric for retrieving the database... more
Image retrieval is a topic where scientific interest is currently high. The important steps associated with image retrieval system are the extraction of discriminative features and a feasible similarity metric for retrieving the database images that are similar in content with the search image. Gabor filtering is a widely adopted technique for feature extraction from the texture images. The recently proposed sparsity promoting l1-norm minimization technique finds the sparsest solution of an under-determined system of linear equations. In the present paper, the l1-norm minimization technique as a similarity metric is used in image retrieval. It is demonstrated through simulation results that the l1-norm minimization technique provides a promising alternative to existing similarity metrics. In particular, the cases where the l1-norm minimization technique works better than the Euclidean distance metric are singled out.
This paper provides a mining approach to the research area of relevance feedback (RF) in content-based image retrieval (CBIR). Relevance feedback is a powerful technique in CBIR systems, in order to improve the performance of CBIR... more
This paper provides a mining approach to the research area of relevance feedback (RF) in content-based image retrieval (CBIR). Relevance feedback is a powerful technique in CBIR systems, in order to improve the performance of CBIR effectively. The drawbacks in CBIR are the features of the query image and the semantic gap between low-level features and high level concepts. Especially, Mining Image data is the one of the essential features in this present scenario since image data plays vital role in every aspect of the system such as business for marketing, hospital for surgery, engineering for construction, Web for publication and so on. In this paper, we are proposed an adaptive approach for relevance feedback in CBIR using mining techniques. Where in the processes of feedback we are using a new technique called Image retrieval based on optimum clusters is proposed for improving user interaction with image retrieval systems by fully exploiting the similarity information. The index is created by describing the images according to their color characteristics, with compact feature vectors, that represent typical color distributions.
In this article we present an advanced version of Dual-PECCS, a cognitively-inspired knowledge representation and reasoning system aimed at extending the capabilities of artificial systems in conceptual categorization tasks. It combines... more
In this article we present an advanced version of Dual-PECCS, a cognitively-inspired knowledge representation and reasoning system aimed at extending the capabilities of artificial systems in conceptual categorization tasks. It combines different sorts of common-sense cat-egorization (prototypical and exemplars-based categorization) with standard monotonic cate-gorization procedures. These different types of inferential procedures are reconciled according to the tenets coming from the dual process theory of reasoning. On the other hand, from a representational perspective, the system relies on the hypothesis of conceptual structures represented as heterogeneous proxytypes. Dual-PECCS has been experimentally assessed in a task of conceptual categorization where a target concept illustrated by a simple common-sense linguistic description had to be identified by resorting to a mix of categorization strategies, and its output has been compared to human responses. The obtained results suggest that our approach can be beneficial to improve the representational and reasoning conceptual capabilities of standard cognitive artificial systems, and –in addition– that it may be plausibly applied to different general computational models of cognition. The current version of the system , in fact, extends our previous work, in that Dual-PECCS is now integrated and tested into two cognitive architectures, ACT-R and CLARION, implementing different assumptions on the underlying invariant structures governing human cognition. Such integration allowed us to extend our previous evaluation.
Abstract. Metadata vocabularies provide various semantic relations between concepts. For content-based recommender systems, these relations enable a wide range of concepts to be recommended. However, not all semantically related concepts... more
Abstract. Metadata vocabularies provide various semantic relations between concepts. For content-based recommender systems, these relations enable a wide range of concepts to be recommended. However, not all semantically related concepts are interesting for end users. In this paper, we identified a number of semantic relations, which are within one vocabulary (eg a concept has a broader/narrower concept) and across multiple vocabularies (eg an artist is associated to an art style). Our goal is to investigate which semantic relations are ...
Social media have become a major source of health information for lay people. It has the power to influence the public's adoption of health policies and to determine the response to the current COVID-19 pandemic. The aim of this paper is... more
Social media have become a major source of health information for lay people. It has the power to influence the public's adoption of health policies and to determine the response to the current COVID-19 pandemic. The aim of this paper is to enhance understanding of personality characteristics of users who spread information about controversial COVID-19 medical treatments on Twitter.
The paper focuses on the algorithms of the event detection in content-based video retrieval. Video has a complex structure and can express the same idea in different ways. This makes the task of searching for video more complicated. Video... more
The paper focuses on the algorithms of the event detection in content-based video retrieval. Video has a complex structure and can express the same idea in different ways. This makes the task of searching for video more complicated. Video titles and text descriptions cannot give the whole information about objects and events in the video. This creates a need for content-based video retrieval. There is a semantic gap between low-level video features, that can be extracted, and the users’ perception. The task of event detection is reduced to the task of video segmentation. Complex content-based video retrieval can be regarded as the bridge between traditional retrieval and semantic-based video retrieval. The properties of video as a time series are described. Introduced the concept of anomalies in the video. A method for event detection based on comparing moving averages with windows of different sizes is proposed. According to the classification given at the beginning of this article, our method refers to statistical methods. It differs from other methods with low computational complexity and simplicity. The video stream processing language is proposed for function-based description of video handling algorithms. So, our method is formulated in the form of a declarative description on an interpreted programming language. Unfortunately, most of the existing video processing methods use exclusively imperative approach, which often makes understanding more difficult. Examples of the use of this language are given. Its grammar is described too. As shown by experiments, the implementation of the proposed video events retrieval method, unlike their counterparts, can work for video streams too with a real-time and potentially infinite frame sequences. Such advantages within low computational requirements make implementation of the method helpful in aviation and space technology. The algorithm has some disadvantages due to necessity parameter selection for particular task classes. The theorem on near-duplicates of video is formulated at the end of the article. It asserts the near-duplicate videos express the same sequence of phenomena.
The problem of object recognition is of immense practical importance and potential, and the last decade has witnessed a number of breakthroughs in the state of the art. Most of the past object recognition work focuses on textured objects... more
The problem of object recognition is of immense practical importance and potential, and the last decade has witnessed a number of breakthroughs in the state of the art. Most of the past object recognition work focuses on textured objects and local appearance descriptors extracted around salient points in an image. These methods fail in the matching of smooth, untextured objects for which salient point detection does not produce robust results. The recently proposed bag of boundaries (BoB) method is the first to directly address this problem. Since the texture of smooth objects is largely uninformative, BoB focuses on describing and matching objects based on their post-segmentation boundaries. Herein we address three major weaknesses of this work. The first of these is the uniform treatment of all boundary segments. Instead, we describe a method for detecting the locations and scales of salient boundary segments. Secondly, while the BoB method uses an image based elementary descriptor (HoGs + occupancy matrix), we propose a more compact descriptor based on the local profile of boundary normals' directions. Lastly, we conduct a far more systematic evaluation, both of the bag of boundaries method and the method proposed here. Using a large public database, we demonstrate that our method exhibits greater robustness while at the same time achieving a major computational saving -- object representation is extracted from an image in only 6% of the time needed to extract a bag of boundaries, and the storage requirement is similarly reduced to less than 8%.
the World Wide Web (WWW) is very difficult because web contains a huge number of images. To solve this problem we need a system that can retrieve the required images needed by the user. Image Crawler is a web based tool that collects and... more
the World Wide Web (WWW) is very difficult because web contains a huge number of images. To solve this problem we need a system
that can retrieve the required images needed by the user. Image Crawler is a web based tool that collects and indexes group of web
images available on the internet. This tool collects the keyword or phrase from the user to retrieve the images from the web. Then
these collected keyword is applied to the different general search tools like Google, Yahoo, Bind etc,. The collected web page
information is stored in the temporary file till 200KB file size from the server. Then this file content will be scanned and extract the
image URL’s and it is compared the URL which is present in the database to avoid the duplicate downloads. The extracted URL’s
images are downloaded and finally stores unique image and corresponding metadata like filename, url, size etc. in the database. In
this paper we present the designing of an Image crawler tool. We build a search tool which is flexible, general-purpose image search
framework and explore a directed result aggregating and removing of duplicates to achieve top results compared to other existing
search tools. Finally this resulted images are used in the Content Based Image Retrieval (CBIR) system for extracting the relevant
images need by the client using the content of the images rather than the text based information.
The resource usage in Content-Based Image Retrieval is a frequently neglected issue. This paper describes a novel compact feature vector based on image color histograms in the HSL color space. The images are represented using only 10... more
The resource usage in Content-Based Image Retrieval is a frequently neglected issue. This paper describes a novel compact feature vector based on image color histograms in the HSL color space. The images are represented using only 10 bytes per image. It is shown that, in the context of Query-by-Example (QbE) usage scenarios, the method described achieves retrieval performance close to the state of the art image retrieval methods that use considerably more memory. It is also shown that the described method outperforms other methods with similar memory usage.
Content-Based Image Retrieval (CBIR) locates, retrieves and displays images alike to one given as a query, using a set of features. It demands accessible data in medical archives and from medical equipment, to infer meaning after some... more
Content-Based Image Retrieval (CBIR) locates, retrieves and displays images alike to one given as a query, using a set of features. It demands accessible data in medical archives and from medical equipment, to infer meaning after some processing. A problem similar in some sense to the target image can aid clinicians. CBIR complements text-based retrieval and improves evidence-based diagnosis, administration, teaching, and research in healthcare. It facilitates visual/automatic diagnosis and decision-making in real-time remote consultation/screening, store-and-forward tests, home care assistance and overall patient surveillance. Metrics help comparing visual data and improve diagnostic. Specially designed architectures can benefit from the application scenario. CBIR use calls for file storage standardization, querying procedures, efficient image transmission, realistic databases, global availability, access simplicity, and Internet-based structures. This chapter recommends important and complex aspects required to handle visual content in healthcare.
Recent advances in text detection allow for finding text regions in natural scenes rather accurately. Global features in content based image retrieval, however, typically do not cover such a high level information. While haracteristics of... more
Recent advances in text detection allow for finding text regions in natural scenes rather accurately. Global features in content based image retrieval, however, typically do not cover such a high level information. While haracteristics of text regions may be reflected by texture or color properties, the respective pixels are not treated in a different way. In this contribution we investigate the impact of text detection on content based image retrieval using global features. Detected text regions are preprocessed to allow for different treatment by feature extraction algorithms, and we show that for certain domains this leads to a much higher precision in content based retrieval.
Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or... more
Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or sequences. The framework is based on simple image processing filters that compete with unprocessed greyscale input to yield a single matching score between individuals. It is shown how the discrepancy between illumination conditions between novel input and the training data set can be estimated and used to weigh the contribution of two competing representations. We describe an extensive empirical evaluation of the proposed method on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our algorithm consistently demonstrated a dramatic performance improvement over traditional filtering approaches. We demonstrate a reduction of 50-75% in recognition error rates, the best performing method-filter combination correctly recognizing 96% of the individuals.
Globally, educational systems are under great pressure to adopt innovative methodologies and to integrate new Information and Communication Technologies in teaching and learning process, to prepare students with the knowledge and skills... more
Globally, educational systems are under great pressure to adopt innovative methodologies and to integrate new Information and Communication Technologies in teaching and learning process, to prepare students with the knowledge and skills they need in the 21st century. Apparently, teaching profession is evolving from an emphasis on teacher -centred, lecture -based instructions to student -centred interactive learning environments. The present study aimed to find out the effect of CBI on the achievement of learning Zoology among teacher trainees. The sample consisted of 80 student teachers with 40 student teachers in the control group and experimental group respectively. The data were collected and analyzed with 't' and 'F' test. The findings revealed that there was a significant difference between the control group and experimental group in their mean score values.
This paper presents a novel compact shape descriptor designed specifically for content-based retrieval of complete or nearly complete 3D vessel replicas. The descriptor consists of two vectors that carry morphological features of the... more
This paper presents a novel compact shape descriptor designed specifically for content-based retrieval
of complete or nearly complete 3D vessel replicas. The descriptor consists of two vectors that carry
morphological features of the vessel’s main body and appendages. The extraction of the descriptor is
based on a pose normalisation preprocessing phase which is designed for axially symmetric objects. In
order to evaluate the efficiency of the descriptor, we created a calibrated ground-truth database of 1012
3D digitised and manually modelled vessels and performed multiple query-by-example experiments.
We present the performance of our descriptor in relation to the performance of the MPEG-7 3D shape
spectrum descriptor. Additionally, a web-based 3D content-based retrieval prototype system has been
developed based on open source technologies.