https://elcvia.cvc.uab.cat/issue/feed ELCVIA Electronic Letters on Computer Vision and Image Analysis 2022-05-19T11:28:25+02:00 Electronic Letters on Computer Vision and Image Analysis elcvia@cvc.uab.cat Open Journal Systems Electronic Journal on Computer Vision and Image Analysis https://elcvia.cvc.uab.cat/article/view/1436 Pre-trained CNNs as Feature-Extraction Modules for Image Captioning 2022-03-08T11:25:08+01:00 Muhammad Abdelhadie Al-Malla abdelhadie.almalla@gmail.com Assef Jafar assef.jafar@hiast.edu.sy Nada Ghneim nada.ghneim@hiast.edu.sy <p>In this work, we present a thorough experimental study about feature extraction using Convolutional Neural<br>Networks (CNNs) for the task of image captioning in the context of deep learning. We perform a set of 72<br>experiments on 12 image classification CNNs pre-trained on the ImageNet [29] dataset. The features are<br>extracted from the last layer after removing the fully connected layer and fed into the captioning model. We use<br>a unified captioning model with a fixed vocabulary size across all the experiments to study the effect of changing<br>the CNN feature extractor on image captioning quality. The scores are calculated using the standard metrics in<br>image captioning. We find a strong relationship between the model structure and the image captioning dataset<br>and prove that VGG models give the least quality for image captioning feature extraction among the tested<br>CNNs. In the end, we recommend a set of pre-trained CNNs for each of the image captioning evaluation metrics<br>we want to optimise, and show the connection between our results and previous works. To our knowledge, this<br>work is the most comprehensive comparison between feature extractors for image captioning.</p> 2022-05-10T00:00:00+02:00 Copyright (c) 2022 Muhammad Abdelhadie Al-Malla, Muhammad Abdelhadie Al-Malla, Assef Jafar, Nada Ghneim https://elcvia.cvc.uab.cat/article/view/1453 Retinal Blood Vessels Segmentation using Fréchet PDF and MSMO Method 2022-03-04T10:05:11+01:00 Sushil Kumar saroj sushil.mnnit10@gmail.com Rakesh Kumar rkiitr@gmail.com Nagendra Pratap Singh nps@nith.ac.in <p>Blood vessels of retina contain information about many severe diseases like glaucoma, hypertension, obesity, diabetes etc. Health professionals use this information to detect and diagnose these diseases. Therefore, it is necessary to segment retinal blood vessels. Quality of retinal image directly affects the accuracy of segmentation. Therefore, quality of image must be as good as possible. Many researchers have proposed various methods to segment retinal blood vessels. Most of the researchers have focused only on segmentation process and paid less attention on pre processing of image even though pre processing plays vital role in segmentation. The proposed method introduces a novel method called multi-scale switching morphological (MSMO) for pre processing and Fréchet match filter for retinal vessel segmentation. We have experimentally tested and verified the proposed method on DRIVE, STARE and HRF data sets. Obtained outcome demonstrate that performance of the proposed method has improved substantially. The cause of improved performance is the better pre processing and segmentation methods.</p> 2022-04-28T00:00:00+02:00 Copyright (c) 2022 sushil kumar saroj, Rakesh Kumar, Nagendra Pratap Singh https://elcvia.cvc.uab.cat/article/view/1482 Object Detection and Statistical Analysis of Microscopy Image Sequences 2022-04-21T14:55:49+02:00 Juliana Gambini mgambini@itba.edu.ar Sasha Hurovitz sashahuro@gmail.com Debora Chan mchan@frba.utn.edu.a Rodrigo Ramele rramele@itba.edu.ar <p>Confocal microscope images are wide useful in medical diagnosis and research. The automatic interpretation of this type of images is very important but it is a challenging endeavor in image processing area, since these images are heavily contaminated with noise, have low contrast and low resolution. <br>This work deals with the problem of analyzing the penetration velocity of a chemotherapy drug in an ocular tumor called retinoblastoma. The primary retinoblastoma cells cultures are exposed to topotecan drug and the penetration evolution is documented by producing sequences of microscopy images. It is possible to quantify the penetration rate of topotecan drug because it produces fluorescence emission by laser excitation which is captured by the camera.<br>In order to estimate the topotecan penetration time in the whole retinoblastoma cell culture, a procedure based on an active contour detection algorithm, a neural network classifier and a statistical model and its validation, is proposed.<br>This new inference model allows to estimate the penetration time. <br>Results show that the penetration mean time strongly depends on tumorsphere size and on chemotherapeutic treatment that the patient has previously received.</p> 2022-04-28T00:00:00+02:00 Copyright (c) 2022 Juliana Gambini, Sasha Hurovitz, Debora Chan, Rodrigo Ramele https://elcvia.cvc.uab.cat/article/view/1517 Material Classification with a Transfer Learning based Deep Model on an imbalanced Dataset using an epochal Deming-Cycle-Methodology 2022-04-26T14:18:42+02:00 Marco Klaiber marco.klaiber@studmail.htw-aalen.de <p>This work demonstrates that a transfer learning-based deep learning model can perform unambiguous classification based on microscopic images of material surfaces with a high degree of accuracy. A transfer learning-enhanced deep learning model was successfully used in combination with an innovative approach for eliminating noisy data based on automatic selection using pixel sum values, which was refined over different epochs to develop and evaluate an effective model for classifying microscopy images. The deep learning model evaluated achieved 91.54% accuracy with the dataset used and set new standards with the method applied. In addition, care was taken to achieve a balance between accuracy and robustness with respect to the model. Based on this scientific report, a means of identifying microscopy images could evolve to support material identification, suggesting a potential application in the domain of materials science and engineering.&nbsp;</p> 2022-06-14T00:00:00+02:00 Copyright (c) 2022 Marco Klaiber https://elcvia.cvc.uab.cat/article/view/1392 A neural network with competitive layers for character recognition 2022-05-19T11:28:25+02:00 Alexander Goltsev agoltsev@adg.kiev.ua Vladimir Gritsenko vig@irtc.org.ua <p class="AbstractBodytext">A structure and functioning mechanisms of a neural network with competitive layers are described. The network is intended to solve the character recognition task. The network consists of several competitive layers of neurons. Each layer is a neural network consisting of a number of neurons represented as a layer. The number of neural layers is equal to the number of recognized classes. All neural layers have one-to-one correspondence with one another and with the input raster. The neurons of every layer have mutual lateral learning connections, which weights are modified during the learning process. There is a competitive (inhibitory) relationship between all neural layers. This competitive interaction is realized by means of a “winner-take-all” (WTA) procedure which aim is to select the layer with the highest level of neural activity.</p><p class="AbstractBodytext">Validation of the network has been done in experiments on recognition of handwritten digits of the MNIST database. The experiments have demonstrated that its error rate is few less than 2%, which is not a high result, but it is compensated by rather fast data processing and a very simple structure and functioning mechanisms.</p><p class="AbstractBodytext"> </p> 2022-06-28T00:00:00+02:00 Copyright (c) 2022 Alexander Goltsev, Vladimir Gritsenko https://elcvia.cvc.uab.cat/article/view/1361 Feature selection based on discriminative power under uncertainty for computer vision applications 2021-11-28T20:55:18+01:00 Marwa Chakroun chakroun.marwa37@gmail.com Sonda Ammar Bouhamed sonda.ammar@gmail.com Imene Khanfir Kallel imen.khanfir.kallel@gmail.com Basel Solaiman basel.solaiman@imt-atlantique.fr Houda Derbel hd.derbel@gmail.com Feature selection is a prolific research field, which has been widely studied in the last decades and has been successfully applied to numerous computer vision systems. It mainly aims to reduce the dimensionality and thus the system complexity. Features have not the same importance within the different classes. Some of them perform for class representation while others perform for class separation. In this paper, a new feature selection method based on discriminative power is proposed to select the relevant features under an uncertain framework, where the uncertainty is expressed through a possibility distribution. In an uncertain context, our method shows its ability to select features that can represent and discriminate between classes. 2022-06-28T00:00:00+02:00 Copyright (c) 2022 Marwa Chakroun, Sonda Ammar Bouhamed, Imene Khanfir Kallel, Basel Solaiman, Houda Derbel https://elcvia.cvc.uab.cat/article/view/1433 Attention-based CNN-ConvLSTM for Handwritten Arabic Word Extraction 2022-05-19T10:02:51+02:00 takwa Ben Aicha takwa.ben.aichaa@gmail.com Afef Kacem Echi ff.kacem@gmail.com <p>Word extraction is one of the most critical steps in handwritten recognition systems. It is challenging for many reasons, such as the variability of handwritten writing styles, touching and overlapping characters, skewness problems, diacritics, ascenders, and descenders' presence. In this work, we propose a deep-learning-based approach for handwritten Arabic word extraction. We used an Attention-based CNN-ConvLSTM (Convolutional Long Short-term Memory) followed by a CTC (Connectionist Temporal Classification) function. Firstly, the text-line input image's essential features are extracted using Attention-based Convolutional Neural Networks (CNN). The extracted features and the text line's transcription are then passed to a ConvLSTM to learn a mapping between them. Finally, we used a CTC to learn the alignment between text-line images and their transcription automatically. We tested the proposed model on a complex dataset known as KFUPM Handwritten Arabic Text (KHATT \cite{khatt}). It consists of complex patterns of handwritten Arabic text-lines. The experimental results show an apparent efficiency of the used combination, where we ended up with an extraction success rate of 91.7\%.</p> 2022-06-28T00:00:00+02:00 Copyright (c) 2022 takwa Ben Aicha, Afef Kacem Echi