Similar Journals
![]() |
Journal of Imaging
Number of Followers: 3 ![]() ISSN (Online) 2313-433X Published by MDPI ![]() |
- J. Imaging, Vol. 8, Pages 118: Discriminative Shape Feature Pooling in
Deep Neural Networks
Authors: Gang Hu, Chahna Dixit, Guanqiu Qi
First page: 118
Abstract: Although deep learning approaches are able to generate generic image features from massive labeled data, discriminative handcrafted features still have advantages in providing explicit domain knowledge and reflecting intuitive visual understanding. Much of the existing research focuses on integrating both handcrafted features and deep networks to leverage the benefits. However, the issues of parameter quality have not been effectively solved in existing applications of handcrafted features in deep networks. In this research, we propose a method that enriches deep network features by utilizing the injected discriminative shape features (generic edge tokens and curve partitioning points) to adjust the network’s internal parameter update process. Thus, the modified neural networks are trained under the guidance of specific domain knowledge, and they are able to generate image representations that incorporate the benefits from both handcrafted and deep learned features. The comparative experiments were performed on several benchmark datasets. The experimental results confirmed our method works well on both large and small training datasets. Additionally, compared with existing models using either handcrafted features or deep network representations, our method not only improves the corresponding performance, but also reduces the computational costs.
Citation: Journal of Imaging
PubDate: 2022-04-20
DOI: 10.3390/jimaging8050118
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 119: X-ray Digital Radiography and Computed
Tomography
Authors: Maria Pia Morigi, Fauzia Albertin
First page: 119
Abstract: In recent years, X-ray imaging has rapidly grown and spread beyond the medical field; today, it plays a key role in diverse research areas [...]
Citation: Journal of Imaging
PubDate: 2022-04-21
DOI: 10.3390/jimaging8050119
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 120: Time Synchronization of Multimodal
Physiological Signals through Alignment of Common Signal Types and Its
Technical Considerations in Digital Health
Authors: Ran Xiao, Cheng Ding, Xiao Hu
First page: 120
Abstract: Background: Despite advancements in digital health, it remains challenging to obtain precise time synchronization of multimodal physiological signals collected through different devices. Existing algorithms mainly rely on specific physiological features that restrict the use cases to certain signal types. The present study aims to complement previous algorithms and solve a niche time alignment problem when a common signal type is available across different devices. Methods: We proposed a simple time alignment approach based on the direct cross-correlation of temporal amplitudes, making it agnostic and thus generalizable to different signal types. The approach was tested on a public electrocardiographic (ECG) dataset to simulate the synchronization of signals collected from an ECG watch and an ECG patch. The algorithm was evaluated considering key practical factors, including sample durations, signal quality index (SQI), resilience to noise, and varying sampling rates. Results: The proposed approach requires a short sample duration (30 s) to operate, and demonstrates stable performance across varying sampling rates and resilience to common noise. The lowest synchronization delay achieved by the algorithm is 0.13 s with the integration of SQI thresholding. Conclusions: Our findings help improve the time alignment of multimodal signals in digital health and advance healthcare toward precise remote monitoring and disease prevention.
Citation: Journal of Imaging
PubDate: 2022-04-21
DOI: 10.3390/jimaging8050120
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 121: Weakly Supervised Polyp Segmentation in
Colonoscopy Images Using Deep Neural Networks
Authors: Siwei Chen, Gregor Urban, Pierre Baldi
First page: 121
Abstract: Colorectal cancer (CRC) is a leading cause of mortality worldwide, and preventive screening modalities such as colonoscopy have been shown to noticeably decrease CRC incidence and mortality. Improving colonoscopy quality remains a challenging task due to limiting factors including the training levels of colonoscopists and the variability in polyp sizes, morphologies, and locations. Deep learning methods have led to state-of-the-art systems for the identification of polyps in colonoscopy videos. In this study, we show that deep learning can also be applied to the segmentation of polyps in real time, and the underlying models can be trained using mostly weakly labeled data, in the form of bounding box annotations that do not contain precise contour information. A novel dataset, Polyp-Box-Seg of 4070 colonoscopy images with polyps from over 2000 patients, is collected, and a subset of 1300 images is manually annotated with segmentation masks. A series of models is trained to evaluate various strategies that utilize bounding box annotations for segmentation tasks. A model trained on the 1300 polyp images with segmentation masks achieves a dice coefficient of 81.52%, which improves significantly to 85.53% when using a weakly supervised strategy leveraging bounding box images. The Polyp-Box-Seg dataset, together with a real-time video demonstration of the segmentation system, are publicly available.
Citation: Journal of Imaging
PubDate: 2022-04-22
DOI: 10.3390/jimaging8050121
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 122: Surreptitious Adversarial Examples through
Functioning QR Code
Authors: Aran Chindaudom, Prarinya Siritanawan, Karin Sumongkayothin, Kazunori Kotani
First page: 122
Abstract: The continuous advances in the technology of Convolutional Neural Network (CNN) and Deep Learning have been applied to facilitate various tasks of human life. However, security risks of the users’ information and privacy have been increasing rapidly due to the models’ vulnerabilities. We have developed a novel method of adversarial attack that can conceal its intent from human intuition through the use of a modified QR code. The modified QR code can be consistently scanned with a reader while retaining adversarial efficacy against image classification models. The QR adversarial patch was created and embedded into an input image to generate adversarial examples, which were trained against CNN image classification models. Experiments were performed to investigate the trade-off in different patch shapes and find the patch’s optimal balance of scannability and adversarial efficacy. Furthermore, we have investigated whether particular classes of images are more resistant or vulnerable to the adversarial QR attack, and we also investigated the generality of the adversarial attack across different image classification models.
Citation: Journal of Imaging
PubDate: 2022-04-22
DOI: 10.3390/jimaging8050122
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 123: Microwave Imaging for Early Breast Cancer
Detection: Current State, Challenges, and Future Directions
Authors: Nour AlSawaftah, Salma El-Abed, Salam Dhou, Amer Zakaria
First page: 123
Abstract: Breast cancer is the most commonly diagnosed cancer type and is the leading cause of cancer-related death among females worldwide. Breast screening and early detection are currently the most successful approaches for the management and treatment of this disease. Several imaging modalities are currently utilized for detecting breast cancer, of which microwave imaging (MWI) is gaining quite a lot of attention as a promising diagnostic tool for early breast cancer detection. MWI is a noninvasive, relatively inexpensive, fast, convenient, and safe screening tool. The purpose of this paper is to provide an up-to-date survey of the principles, developments, and current research status of MWI for breast cancer detection. This paper is structured into two sections; the first is an overview of current MWI techniques used for detecting breast cancer, followed by an explanation of the working principle behind MWI and its various types, namely, microwave tomography and radar-based imaging. In the second section, a review of the initial experiments along with more recent studies on the use of MWI for breast cancer detection is presented. Furthermore, the paper summarizes the challenges facing MWI as a breast cancer detection tool and provides future research directions. On the whole, MWI has proven its potential as a screening tool for breast cancer detection, both as a standalone or complementary technique. However, there are a few challenges that need to be addressed to unlock the full potential of this imaging modality and translate it to clinical settings.
Citation: Journal of Imaging
PubDate: 2022-04-23
DOI: 10.3390/jimaging8050123
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 124: Extraction and Calculation of Roadway Area
from Satellite Images Using Improved Deep Learning Model and
Post-Processing
Authors: Varun Yerram, Hiroyuki Takeshita, Yuji Iwahori, Yoshitsugu Hayashi, M. K. Bhuyan, Shinji Fukui, Boonserm Kijsirikul, Aili Wang
First page: 124
Abstract: Roadway area calculation is a novel problem in remote sensing and urban planning. This paper models this problem as a two-step problem, roadway extraction, and area calculation. Roadway extraction from satellite images is a problem that has been tackled many times before. This paper proposes a method using pixel resolution to calculate the area of the roads covered in satellite images. The proposed approach uses novel U-net and Resnet architectures called U-net++ and ResNeXt. The state-of-the-art model is combined with the proposed efficient post-processing approach to improve the overlap with ground truth labels. The performance of the proposed road extraction algorithm is evaluated on the Massachusetts dataset and it is shown that the proposed approach outperforms the existing solutions which use models from the U-net family.
Citation: Journal of Imaging
PubDate: 2022-04-25
DOI: 10.3390/jimaging8050124
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 125: Colored Point Cloud Completion for a Head
Using Adversarial Rendered Image Loss
Authors: Yuki Ishida, Yoshitsugu Manabe, Noriko Yata
First page: 125
Abstract: Recent advances in depth measurement and its utilization have made point cloud processing more critical. Additionally, the human head is essential for communication, and its three-dimensional data are expected to be utilized in this regard. However, a single RGB-Depth (RGBD) camera is prone to occlusion and depth measurement failure for dark hair colors such as black hair. Recently, point cloud completion, where an entire point cloud is estimated and generated from a partial point cloud, has been studied, but only the shape is learned, rather than the completion of colored point clouds. Thus, this paper proposes a machine learning-based completion method for colored point clouds with XYZ location information and the International Commission on Illumination (CIE) LAB (L*a*b*) color information. The proposed method uses the color difference between point clouds based on the Chamfer Distance (CD) or Earth Mover’s Distance (EMD) of point cloud shape evaluation as a color loss. In addition, an adversarial loss to L*a*b*-Depth images rendered from the output point cloud can improve the visual quality. The experiments examined networks trained using a colored point cloud dataset created by combining two 3D datasets: hairstyles and faces. Experimental results show that using the adversarial loss with the colored point cloud renderer in the proposed method improves the image domain’s evaluation.
Citation: Journal of Imaging
PubDate: 2022-04-26
DOI: 10.3390/jimaging8050125
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 126: Airborne Hyperspectral Imagery for Band
Selection Using Moth–Flame Metaheuristic Optimization
Authors: Raju Anand, Sathishkumar Samiaappan, Shanmugham Veni, Ethan Worch, Meilun Zhou
First page: 126
Abstract: In this research, we study a new metaheuristic algorithm called Moth–Flame Optimization (MFO) for hyperspectral band selection. With the hundreds of highly correlated narrow spectral bands, the number of training samples required to train a statistical classifier is high. Thus, the problem is to select a subset of bands without compromising the classification accuracy. One of the ways to solve this problem is to model an objective function that measures class separability and utilize it to arrive at a subset of bands. In this research, we studied MFO to select optimal spectral bands for classification. MFO is inspired by the behavior of moths with respect to flames, which is the navigation method of moths in nature called transverse orientation. In MFO, a moth navigates the search space through a process called transverse orientation by keeping a constant angle with the Moon, which is a compelling strategy for traveling long distances in a straight line, considering that the Moon’s distance from the moth is considerably long. Our research tested MFO on three benchmark hyperspectral datasets—Indian Pines, University of Pavia, and Salinas. MFO produced an Overall Accuracy (OA) of 88.98%, 94.85%, and 97.17%, respectively, on the three datasets. Our experimental results indicate that MFO produces better OA and Kappa when compared to state-of-the-art band selection algorithms such as particle swarm optimization, grey wolf, cuckoo search, and genetic algorithms. The analysis results prove that the proposed approach effectively addresses the spectral band selection problem and provides a high classification accuracy.
Citation: Journal of Imaging
PubDate: 2022-04-27
DOI: 10.3390/jimaging8050126
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 127: A Review of Watershed Implementations for
Segmentation of Volumetric Images
Authors: Anton Kornilov, Ilia Safonov, Ivan Yakimchuk
First page: 127
Abstract: Watershed is a widely used image segmentation algorithm. Most researchers understand just an idea of this method: a grayscale image is considered as topographic relief, which is flooded from initial basins. However, frequently they are not aware of the options of the algorithm and the peculiarities of its realizations. There are many watershed implementations in software packages and products. Even if these packages are based on the identical algorithm–watershed, by flooding their outcomes, processing speed, and consumed memory, vary greatly. In particular, the difference among various implementations is noticeable for huge volumetric images; for instance, tomographic 3D images, for which low performance and high memory requirements of watershed might be bottlenecks. In our review, we discuss the peculiarities of algorithms with and without waterline generation, the impact of connectivity type and relief quantization level on the result, approaches for parallelization, as well as other method options. We present detailed benchmarking of seven open-source and three commercial software implementations of marker-controlled watershed for semantic or instance segmentation. We compare those software packages for one synthetic and two natural volumetric images. The aim of the review is to provide information and advice for practitioners to select the appropriate version of watershed for their problem solving. In addition, we forecast future directions of software development for 3D image segmentation by watershed.
Citation: Journal of Imaging
PubDate: 2022-04-26
DOI: 10.3390/jimaging8050127
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 128: Elimination of Defects in Mammograms Caused
by a Malfunction of the Device Matrix
Authors: Dmitrii Tumakov, Zufar Kayumov, Alisher Zhumaniezov, Dmitry Chikrin, Diaz Galimyanov
First page: 128
Abstract: Today, the processing and analysis of mammograms is quite an important field of medical image processing. Small defects in images can lead to false conclusions. This is especially true when the distortion occurs due to minor malfunctions in the equipment. In the present work, an algorithm for eliminating a defect is proposed, which includes a change in intensity on a mammogram and deteriorations in the contrast of individual areas. The algorithm consists of three stages. The first is the defect identification stage. The second involves improvement and equalization of the contrasts of different parts of the image outside the defect. The third involves restoration of the defect area via a combination of interpolation and an artificial neural network. The mammogram obtained as a result of applying the algorithm shows significantly better image quality and does not contain distortions caused by changes in brightness of the pixels. The resulting images are evaluated using Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE) and Naturalness Image Quality Evaluator (NIQE) metrics. In total, 98 radiomics features are extracted from the original and obtained images, and conclusions are drawn about the minimum changes in features between the original image and the image obtained by the proposed algorithm.
Citation: Journal of Imaging
PubDate: 2022-05-02
DOI: 10.3390/jimaging8050128
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 129: Image Classification in JPEG Compression
Domain for Malaria Infection Detection
Authors: Yuhang Dong, W. David Pan
First page: 129
Abstract: Digital images are usually stored in compressed format. However, image classification typically takes decompressed images as inputs rather than compressed images. Therefore, performing image classification directly in the compression domain will eliminate the need for decompression, thus increasing efficiency and decreasing costs. However, there has been very sparse work on image classification in the compression domain. In this paper, we studied the feasibility of classifying images in their JPEG compression domain. We analyzed the underlying mechanisms of JPEG as an example and conducted classification on data from different stages during the compression. The images we used were malaria-infected red blood cells and normal cells. The training data include multiple combinations of DCT coefficients, DC values in both decimal and binary forms, the “scan” segment in both binary and decimal form, and the variable length of the entire bitstream. The result shows that LSTM can successfully classify the image in its compressed form, with accuracies around 80%. If using only coded DC values, we can achieve accuracies higher than 90%. This indicates that images from different classes can still be well separated in their JPEG compressed format. Our simulations demonstrate that the proposed compression domain-processing method can reduce the input data, and eliminate the image decompression step, thereby achieving significant savings on memory and computation time.
Citation: Journal of Imaging
PubDate: 2022-05-03
DOI: 10.3390/jimaging8050129
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 130: Weakly Supervised Tumor Detection in PET
Using Class Response for Treatment Outcome Prediction
Authors: Amine Amyar, Romain Modzelewski, Pierre Vera, Vincent Morard, Su Ruan
First page: 130
Abstract: It is proven that radiomic characteristics extracted from the tumor region are predictive. The first step in radiomic analysis is the segmentation of the lesion. However, this task is time consuming and requires a highly trained physician. This process could be automated using computer-aided detection (CAD) tools. Current state-of-the-art methods are trained in a supervised learning setting, which requires a lot of data that are usually not available in the medical imaging field. The challenge is to train one model to segment different types of tumors with only a weak segmentation ground truth. In this work, we propose a prediction framework including a 3D tumor segmentation in positron emission tomography (PET) images, based on a weakly supervised deep learning method, and an outcome prediction based on a 3D-CNN classifier applied to the segmented tumor regions. The key step is to locate the tumor in 3D. We propose to (1) calculate two maximum intensity projection (MIP) images from 3D PET images in two directions, (2) classify the MIP images into different types of cancers, (3) generate the class activation maps through a multitask learning approach with a weak prior knowledge, and (4) segment the 3D tumor region from the two 2D activation maps with a proposed new loss function for the multitask. The proposed approach achieves state-of-the-art prediction results with a small data set and with a weak segmentation ground truth. Our model was tested and validated for treatment response and survival in lung and esophageal cancers on 195 patients, with an area under the receiver operating characteristic curve (AUC) of 67% and 59%, respectively, and a dice coefficient of 73% and 0.77% for tumor segmentation.
Citation: Journal of Imaging
PubDate: 2022-05-09
DOI: 10.3390/jimaging8050130
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 131: BI-RADS BERT and Using Section Segmentation
to Understand Radiology Reports
Authors: Grey Kuling, Belinda Curpen, Anne L. Martel
First page: 131
Abstract: Radiology reports are one of the main forms of communication between radiologists and other clinicians, and contain important information for patient care. In order to use this information for research and automated patient care programs, it is necessary to convert the raw text into structured data suitable for analysis. State-of-the-art natural language processing (NLP) domain-specific contextual word embeddings have been shown to achieve impressive accuracy for these tasks in medicine, but have yet to be utilized for section structure segmentation. In this work, we pre-trained a contextual embedding BERT model using breast radiology reports and developed a classifier that incorporated the embedding with auxiliary global textual features in order to perform section segmentation. This model achieved 98% accuracy in segregating free-text reports, sentence by sentence, into sections of information outlined in the Breast Imaging Reporting and Data System (BI-RADS) lexicon, which is a significant improvement over the classic BERT model without auxiliary information. We then evaluated whether using section segmentation improved the downstream extraction of clinically relevant information such as modality/procedure, previous cancer, menopausal status, purpose of exam, breast density, and breast MRI background parenchymal enhancement. Using the BERT model pre-trained on breast radiology reports, combined with section segmentation, resulted in an overall accuracy of 95.9% in the field extraction tasks. This is a 17% improvement, compared to an overall accuracy of 78.9% for field extraction with models using classic BERT embeddings and not using section segmentation. Our work shows the strength of using BERT in the analysis of radiology reports and the advantages of section segmentation by identifying the key features of patient factors recorded in breast radiology reports.
Citation: Journal of Imaging
PubDate: 2022-05-09
DOI: 10.3390/jimaging8050131
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 132: Are Social Networks Watermarking Us or Are
We (Unawarely) Watermarking Ourself'
Authors: Flavio Bertini, Rajesh Sharma, Danilo Montesi
First page: 132
Abstract: In the last decade, Social Networks (SNs) have deeply changed many aspects of society, and one of the most widespread behaviours is the sharing of pictures. However, malicious users often exploit shared pictures to create fake profiles, leading to the growth of cybercrime. Thus, keeping in mind this scenario, authorship attribution and verification through image watermarking techniques are becoming more and more important. In this paper, we firstly investigate how thirteen of the most popular SNs treat uploaded pictures in order to identify a possible implementation of image watermarking techniques by respective SNs. Second, we test the robustness of several image watermarking algorithms on these thirteen SNs. Finally, we verify whether a method based on the Photo-Response Non-Uniformity (PRNU) technique, which is usually used in digital forensic or image forgery detection activities, can be successfully used as a watermarking approach for authorship attribution and verification of pictures on SNs. The proposed method is sufficiently robust, in spite of the fact that pictures are often downgraded during the process of uploading to the SNs. Moreover, in comparison to conventional watermarking methods the proposed method can successfully pass through different SNs, solving related problems such as profile linking and fake profile detection. The results of our analysis on a real dataset of 8400 pictures show that the proposed method is more effective than other watermarking techniques and can help to address serious questions about privacy and security on SNs. Moreover, the proposed method paves the way for the definition of multi-factor online authentication mechanisms based on robust digital features.
Citation: Journal of Imaging
PubDate: 2022-05-10
DOI: 10.3390/jimaging8050132
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 133: Integration of Deep Learning and Active
Shape Models for More Accurate Prostate Segmentation in 3D MR Images
Authors: Massimo Salvi, Bruno De Santi, Bianca Pop, Martino Bosco, Valentina Giannini, Daniele Regge, Filippo Molinari, Kristen M. Meiburger
First page: 133
Abstract: Magnetic resonance imaging (MRI) has a growing role in the clinical workup of prostate cancer. However, manual three-dimensional (3D) segmentation of the prostate is a laborious and time-consuming task. In this scenario, the use of automated algorithms for prostate segmentation allows us to bypass the huge workload of physicians. In this work, we propose a fully automated hybrid approach for prostate gland segmentation in MR images using an initial segmentation of prostate volumes using a custom-made 3D deep network (VNet-T2), followed by refinement using an Active Shape Model (ASM). While the deep network focuses on three-dimensional spatial coherence of the shape, the ASM relies on local image information and this joint effort allows for improved segmentation of the organ contours. Our method is developed and tested on a dataset composed of T2-weighted (T2w) MRI prostatic volumes of 60 male patients. In the test set, the proposed method shows excellent segmentation performance, achieving a mean dice score and Hausdorff distance of 0.851 and 7.55 mm, respectively. In the future, this algorithm could serve as an enabling technology for the development of computer-aided systems for prostate cancer characterization in MR imaging.
Citation: Journal of Imaging
PubDate: 2022-05-11
DOI: 10.3390/jimaging8050133
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 134: LightBot: A Multi-Light Position Robotic
Acquisition System for Adaptive Capturing of Cultural Heritage Surfaces
Authors: Ramamoorthy Luxman, Yuly Emilia Castro, Hermine Chatoux, Marvin Nurit, Amalia Siatou, Gaëtan Le Goïc, Laura Brambilla, Christian Degrigny, Franck Marzani, Alamin Mansouri
First page: 134
Abstract: Multi-light acquisitions and modeling are well-studied techniques for characterizing surface geometry, widely used in the cultural heritage field. Current systems that are used to perform this kind of acquisition are mainly free-form or dome-based. Both of them have constraints in terms of reproducibility, limitations on the size of objects being acquired, speed, and portability. This paper presents a novel robotic arm-based system design, which we call LightBot, as well as its applications in reflectance transformation imaging (RTI) in particular. The proposed model alleviates some of the limitations observed in the case of free-form or dome-based systems. It allows the automation and reproducibility of one or a series of acquisitions adapting to a given surface in two-dimensional space.
Citation: Journal of Imaging
PubDate: 2022-05-12
DOI: 10.3390/jimaging8050134
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 135: On Acquisition Parameters and Processing
Techniques for Interparticle Contact Detection in Granular Packings Using
Synchrotron Computed Tomography
Authors: Fernando Alvarez-Borges, Sharif Ahmed, Robert C. Atwood
First page: 135
Abstract: X-ray computed tomography (XCT) is regularly employed in geomechanics to non-destructively measure the solid and pore fractions of soil and rock from reconstructed 3D images. With the increasing availability of high-resolution XCT imaging systems, researchers now seek to measure microfabric parameters such as the number and area of interparticle contacts, which can then be used to inform soil behaviour modelling techniques. However, recent research has evidenced that conventional image processing methods consistently overestimate the number and area of interparticle contacts, mainly due to acquisition-driven image artefacts. The present study seeks to address this issue by systematically assessing the role of XCT acquisition parameters in the accurate detection of interparticle contacts. To this end, synchrotron XCT has been applied to a hexagonal close-packed arrangement of glass pellets with and without a prescribed separation between lattice layers. Different values for the number of projections, exposure time, and rotation range have been evaluated. Conventional global grey value thresholding and novel U-Net segmentation methods have been assessed, followed by local refinements at the presumptive contacts, as per recently proposed contact detection routines. The effect of the different acquisition set-ups and segmentation techniques on contact detection performance is presented and discussed, and optimised workflows are proposed.
Citation: Journal of Imaging
PubDate: 2022-05-12
DOI: 10.3390/jimaging8050135
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 136: Data Extraction of Circular-Shaped and
Grid-like Chart Images
Authors: Filip Bajić, Josip Job
First page: 136
Abstract: Chart data extraction is a crucial research field in recovering information from chart images. With the recent rise in image processing and computer vision algorithms, researchers presented various approaches to tackle this problem. Nevertheless, most of them use different datasets, often not publicly available to the research community. Therefore, the main focus of this research was to create a chart data extraction algorithm for circular-shaped and grid-like chart types, which will accelerate research in this field and allow uniform result comparison. A large-scale dataset is provided containing 120,000 chart images organized into 20 categories, with corresponding ground truth for each image. Through the undertaken extensive research and to the best of our knowledge, no other author reports the chart data extraction of the sunburst diagrams, heatmaps, and waffle charts. In this research, a new, fully automatic low-level algorithm is also presented that uses a raster image as input and generates an object-oriented structure of the chart of that image. The main novelty of the proposed approach is in chart processing on binary images instead of commonly used pixel counting techniques. The experiments were performed with a synthetic dataset and with real-world chart images. The obtained results demonstrate two things: First, a low-level bottom-up approach can be shared among different chart types. Second, the proposed algorithm achieves superior results on a synthetic dataset. The achieved average data extraction accuracy on the synthetic dataset can be considered state-of-the-art within multiple error rate groups.
Citation: Journal of Imaging
PubDate: 2022-05-12
DOI: 10.3390/jimaging8050136
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 137: Development of a Visualisation Approach for
Analysing Incipient and Clinically Unrecorded Enamel Fissure Caries Using
Laser-Induced Contrast Imaging, MicroRaman Spectroscopy and Biomimetic
Composites: A Pilot Study
Authors: Pavel Seredin, Dmitry Goloshchapov, Vladimir Kashkarov, Anna Emelyanova, Nikita Buylov, Yuri Ippolitov, Tatiana Prutskij
First page: 137
Abstract: This pilot study presents a practical approach to detecting and visualising the initial forms of caries that are not clinically registered. The use of a laser-induced contrast visualisation (LICV) technique was shown to provide detection of the originating caries based on the separation of emissions from sound tissue, areas with destroyed tissue and regions of bacterial invasion. Adding microRaman spectroscopy to the measuring system enables reliable detection of the transformation of the organic–mineral component in the dental tissue and the spread of bacterial microflora in the affected region. Further laboratory and clinical studies of the comprehensive use of LICV and microRaman spectroscopy enable data extension on the application of this approach for accurate determination of the boundaries in the changed dental tissue as a result of initial caries. The obtained data has the potential to develop an effective preventive medical diagnostic approach and as a result, further personalised medical treatment can be specified.
Citation: Journal of Imaging
PubDate: 2022-05-13
DOI: 10.3390/jimaging8050137
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 138: A Generic Framework for Depth
Reconstruction Enhancement
Authors: Hendrik Sommerhoff, Andreas Kolb
First page: 138
Abstract: We propose a generic depth-refinement scheme based on GeoNet, a recent deep-learning approach for predicting depth and normals from a single color image, and extend it to be applied to any depth reconstruction task such as super resolution, denoising and deblurring, as long as the task includes a depth output. Our approach utilizes a tight coupling of the inherent geometric relationship between depth and normal maps to guide a neural network. In contrast to GeoNet, we do not utilize the original input information to the backbone reconstruction task, which leads to a generic application of our network structure. Our approach first learns a high-quality normal map from the depth image generated by the backbone method and then uses this normal map to refine the initial depth image jointly with the learned normal map. This is motivated by the fact that it is hard for neural networks to learn direct mapping between depth and normal maps without explicit geometric constraints. We show the efficiency of our method on the exemplary inverse depth-image reconstruction tasks of denoising, super resolution and removal of motion blur.
Citation: Journal of Imaging
PubDate: 2022-05-16
DOI: 10.3390/jimaging8050138
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 139: Intraretinal Layer Segmentation Using
Cascaded Compressed U-Nets
Authors: Sunil Kumar Yadav, Rahele Kafieh, Hanna Gwendolyn Zimmermann, Josef Kauer-Bonin, Kouros Nouri-Mahdavi, Vahid Mohammadzadeh, Lynn Shi, Ella Maria Kadas, Friedemann Paul, Seyedamirhosein Motamedi, Alexander Ulrich Brandt
First page: 139
Abstract: Reliable biomarkers quantifying neurodegeneration and neuroinflammation in central nervous system disorders such as Multiple Sclerosis, Alzheimer’s dementia or Parkinson’s disease are an unmet clinical need. Intraretinal layer thicknesses on macular optical coherence tomography (OCT) images are promising noninvasive biomarkers querying neuroretinal structures with near cellular resolution. However, changes are typically subtle, while tissue gradients can be weak, making intraretinal segmentation a challenging task. A robust and efficient method that requires no or minimal manual correction is an unmet need to foster reliable and reproducible research as well as clinical application. Here, we propose and validate a cascaded two-stage network for intraretinal layer segmentation, with both networks being compressed versions of U-Net (CCU-INSEG). The first network is responsible for retinal tissue segmentation from OCT B-scans. The second network segments eight intraretinal layers with high fidelity. At the post-processing stage, we introduce Laplacian-based outlier detection with layer surface hole filling by adaptive non-linear interpolation. Additionally, we propose a weighted version of focal loss to minimize the foreground–background pixel imbalance in the training data. We train our method using 17,458 B-scans from patients with autoimmune optic neuropathies, i.e., multiple sclerosis, and healthy controls. Voxel-wise comparison against manual segmentation produces a mean absolute error of 2.3 μm, outperforming current state-of-the-art methods on the same data set. Voxel-wise comparison against external glaucoma data leads to a mean absolute error of 2.6 μm when using the same gold standard segmentation approach, and 3.7 μm mean absolute error in an externally segmented data set. In scans from patients with severe optic atrophy, 3.5% of B-scan segmentation results were rejected by an experienced grader, whereas this was the case in 41.4% of B-scans segmented with a graph-based reference method. The validation results suggest that the proposed method can robustly segment macular scans from eyes with even severe neuroretinal changes.
Citation: Journal of Imaging
PubDate: 2022-05-17
DOI: 10.3390/jimaging8050139
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 140: Comparison of Ultrasound Image Classifier
Deep Learning Algorithms for Shrapnel Detection
Authors: Emily N. Boice, Sofia I. Hernandez-Torres, Eric J. Snider
First page: 140
Abstract: Ultrasound imaging is essential in emergency medicine and combat casualty care, oftentimes used as a critical triage tool. However, identifying injuries, such as shrapnel embedded in tissue or a pneumothorax, can be challenging without extensive ultrasonography training, which may not be available in prolonged field care or emergency medicine scenarios. Artificial intelligence can simplify this by automating image interpretation but only if it can be deployed for use in real time. We previously developed a deep learning neural network model specifically designed to identify shrapnel in ultrasound images, termed ShrapML. Here, we expand on that work to further optimize the model and compare its performance to that of conventional models trained on the ImageNet database, such as ResNet50. Through Bayesian optimization, the model’s parameters were further refined, resulting in an F1 score of 0.98. We compared the proposed model to four conventional models: DarkNet-19, GoogleNet, MobileNetv2, and SqueezeNet which were down-selected based on speed and testing accuracy. Although MobileNetv2 achieved a higher accuracy than ShrapML, there was a tradeoff between accuracy and speed, with ShrapML being 10× faster than MobileNetv2. In conclusion, real-time deployment of algorithms such as ShrapML can reduce the cognitive load for medical providers in high-stress emergency or miliary medicine scenarios.
Citation: Journal of Imaging
PubDate: 2022-05-20
DOI: 10.3390/jimaging8050140
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 141: Image Augmentation Techniques for Mammogram
Analysis
Authors: Parita Oza, Paawan Sharma, Samir Patel, Festus Adedoyin, Alessandro Bruno
First page: 141
Abstract: Research in the medical imaging field using deep learning approaches has become progressively contingent. Scientific findings reveal that supervised deep learning methods’ performance heavily depends on training set size, which expert radiologists must manually annotate. The latter is quite a tiring and time-consuming task. Therefore, most of the freely accessible biomedical image datasets are small-sized. Furthermore, it is challenging to have big-sized medical image datasets due to privacy and legal issues. Consequently, not a small number of supervised deep learning models are prone to overfitting and cannot produce generalized output. One of the most popular methods to mitigate the issue above goes under the name of data augmentation. This technique helps increase training set size by utilizing various transformations and has been publicized to improve the model performance when tested on new data. This article surveyed different data augmentation techniques employed on mammogram images. The article aims to provide insights into basic and deep learning-based augmentation techniques.
Citation: Journal of Imaging
PubDate: 2022-05-20
DOI: 10.3390/jimaging8050141
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 142: upU-Net Approaches for Background Emission
Removal in Fluorescence Microscopy
Authors: Alessandro Benfenati
First page: 142
Abstract: The physical process underlying microscopy imaging suffers from several issues: some of them include the blurring effect due to the Point Spread Function, the presence of Gaussian or Poisson noise, or even a mixture of these two types of perturbation. Among them, auto–fluorescence presents other artifacts in the registered image, and such fluorescence may be an important obstacle in correctly recognizing objects and organisms in the image. For example, particle tracking may suffer from the presence of this kind of perturbation. The objective of this work is to employ Deep Learning techniques, in the form of U-Nets like architectures, for background emission removal. Such fluorescence is modeled by Perlin noise, which reveals to be a suitable candidate for simulating such a phenomenon. The proposed architecture succeeds in removing the fluorescence, and at the same time, it acts as a denoiser for both Gaussian and Poisson noise. The performance of this approach is furthermore assessed on actual microscopy images and by employing the restored images for particle recognition.
Citation: Journal of Imaging
PubDate: 2022-05-20
DOI: 10.3390/jimaging8050142
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 143: Digital Hebrew Paleography: Script Types
and Modes
Authors: Ahmad Droby, Irina Rabaev, Daria Vasyutinsky Shapira, Berat Kurar Barakat, Jihad El-Sana
First page: 143
Abstract: Paleography is the study of ancient and medieval handwriting. It is essential for understanding, authenticating, and dating historical texts. Across many archives and libraries, many handwritten manuscripts are yet to be classified. Human experts can process a limited number of manuscripts; therefore, there is a need for an automatic tool for script type classification. In this study, we utilize a deep-learning methodology to classify medieval Hebrew manuscripts into 14 classes based on their script style and mode. Hebrew paleography recognizes six regional styles and three graphical modes of scripts. We experiment with several input image representations and network architectures to determine the appropriate ones and explore several approaches for script classification. We obtained the highest accuracy using hierarchical classification approach. At the first level, the regional style of the script is classified. Then, the patch is passed to the corresponding model at the second level to determine the graphical mode. In addition, we explore the use of soft labels to define a value we call squareness value that indicates the squareness/cursiveness of the script. We show how the graphical mode labels can be redefined using the squareness value. This redefinition increases the classification accuracy significantly. Finally, we show that the automatic classification is on-par with a human expert paleographer.
Citation: Journal of Imaging
PubDate: 2022-05-21
DOI: 10.3390/jimaging8050143
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 144: Generation of Ince–Gaussian Beams
Using Azocarbazole Polymer CGH
Authors: Sumit Kumar Singh, Honoka Haginaka, Boaz Jessie Jackin, Kenji Kinashi, Naoto Tsutsumi, Wataru Sakai
First page: 144
Abstract: Ince–Gaussian beams, defined as a solution to a wave equation in elliptical coordinates, have shown great advantages in applications such as optical communication, optical trapping and optical computation. However, to ingress these applications, a compact and scalable method for generating these beams is required. Here, we present a simple method that satisfies the above requirement, and is capable of generating arbitrary Ince–Gaussian beams and their superposed states through a computer-generated hologram of size 1mm2, fabricated on an azocarbazole polymer film. Other structural beams that can be derived from the Ince–Gaussian beam were also successfully generated by changing the elliptical parameters of the Ince–Gaussian beam. The orthogonality relations between different Ince–Gaussian modes were investigated in order to verify applicability in an optical communication regime. The complete python source code for computing the Ince–Gaussian beams and their holograms are also provided.
Citation: Journal of Imaging
PubDate: 2022-05-21
DOI: 10.3390/jimaging8050144
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 145: What Is Significant in Modern Augmented
Reality: A Systematic Analysis of Existing Reviews
Authors: Athanasios Nikolaidis
First page: 145
Abstract: Augmented reality (AR) is a field of technology that has evolved drastically during the last decades, due to its vast range of applications in everyday life. The aim of this paper is to provide researchers with an overview of what has been surveyed since 2010 in terms of AR application areas as well as in terms of its technical aspects, and to discuss the extent to which both application areas and technical aspects have been covered, as well as to examine whether one can extract useful evidence of what aspects have not been covered adequately and whether it is possible to define common taxonomy criteria for performing AR reviews in the future. To this end, a search with inclusion and exclusion criteria has been performed in the Scopus database, producing a representative set of 47 reviews, covering the years from 2010 onwards. A proper taxonomy of the results is introduced, and the findings reveal, among others, the lack of AR application reviews covering all suggested criteria.
Citation: Journal of Imaging
PubDate: 2022-05-21
DOI: 10.3390/jimaging8050145
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 146: Dental MRI of Oral Soft-Tissue
Tumors—Optimized Use of Black Bone MRI Sequences and a 15-Channel
Mandibular Coil
Authors: Adib Al-Haj Husain, Esra Sekerci, Daphne Schönegg, Fabienne A. Bosshard, Bernd Stadlinger, Sebastian Winklhofer, Marco Piccirelli, Silvio Valdec
First page: 146
Abstract: Soft-tissue lesions in the oral cavity, one of the most common sites for tumors and tumor-like lesions, can be challenging to diagnose and treat due to the wide spectrum from benign indolent to invasive malignant lesions. We report an abnormally large, rapidly growing hyperplastic lesion originating from the buccal mucosa in a 28-year-old male patient. Clinical examination revealed a well-circumscribed, smooth-surfaced, pinkish nodular lesion measuring 2.3 × 2 cm, which suggested the differential diagnosis of irritation fibroma, pyogenic granuloma, oral lipoma, and other benign or malignant neoplasms such as hemangioma, non-Hodgkin’s lymphoma, or metastases to the oral cavity. Dental MRI using a 15-channel mandibular coil was performed to improve perioperative radiological and surgical management, avoiding adverse intraoperative events and misdiagnosis of vascular malformations, especially hemangiomas. Black bone MRI protocols such as STIR (short-tau inversion recovery) and DESS (double-echo steady-state) were used for high-resolution radiation-free imaging. Radiologic findings supported the suspected diagnosis of an irritation fibroma and ruled out any further head and neck lesions, therefore complete surgical resection was performed. Histology confirmed the tentative diagnosis. This article evaluates the use of this novel technique for MR diagnosis in the perioperative management of soft-tissue tumors in oral and maxillofacial surgery.
Citation: Journal of Imaging
PubDate: 2022-05-22
DOI: 10.3390/jimaging8050146
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 147: Scanning X-ray Fluorescence Data Analysis
for the Identification of Byzantine Icons’ Materials, Techniques,
and State of Preservation: A Case Study
Authors: Theofanis Gerodimos, Anastasios Asvestas, Georgios P. Mastrotheodoros, Giannis Chantas, Ioannis Liougos, Aristidis Likas, Dimitrios F. Anagnostopoulos
First page: 147
Abstract: X-ray fluorescence (XRF) spectrometry has proven to be a core, non-destructive, analytical technique in cultural heritage studies mainly because of its non-invasive character and ability to rapidly reveal the elemental composition of the analyzed artifacts. Being able to penetrate deeper into matter than the visible light, X-rays allow further analysis that may eventually lead to the extraction of information that pertains to the substrate(s) of an artifact. The recently developed scanning macroscopic X-ray fluorescence method (MA-XRF) allows for the extraction of elemental distribution images. The present work aimed at comparing two different analysis methods for interpreting the large number of XRF spectra collected in the framework of MA-XRF analysis. The measured spectra were analyzed in two ways: a merely spectroscopic approach and an exploratory data analysis approach. The potentialities of the applied methods are showcased on a notable 18th-century Greek religious panel painting. The spectroscopic approach separately analyses each one of the measured spectra and leads to the construction of single-element spatial distribution images (element maps). The statistical data analysis approach leads to the grouping of all spectra into distinct clusters with common features, while afterward dimensionality reduction algorithms help reduce thousands of channels of XRF spectra in an easily perceived dataset of two-dimensional images. The two analytical approaches allow extracting detailed information about the pigments used and paint layer stratigraphy (i.e., painting technique) as well as restoration interventions/state of preservation.
Citation: Journal of Imaging
PubDate: 2022-05-23
DOI: 10.3390/jimaging8050147
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 148: Three-Dimensional Finger Vein Recognition:
A Novel Mirror-Based Imaging Device
Authors: Christof Kauba, Martin Drahanský, Marie Nováková, Andreas Uhl, Štěpán Rydlo
First page: 148
Abstract: Finger vein recognition has evolved into a major biometric trait in recent years. Despite various improvements in recognition accuracy and usability, finger vein recognition is still far from being perfect as it suffers from low-contrast images and other imaging artefacts. Three-dimensional or multi-perspective finger vein recognition technology provides a way to tackle some of the current problems, especially finger misplacement and rotations. In this work we present a novel multi-perspective finger vein capturing device that is based on mirrors, in contrast to most of the existing devices, which are usually based on multiple cameras. This new device only uses a single camera, a single illumination module and several mirrors to capture the finger at different rotational angles. To derive the need for this new device, we at first summarise the state of the art in multi-perspective finger vein recognition and identify the potential problems and shortcomings of the current devices.
Citation: Journal of Imaging
PubDate: 2022-05-23
DOI: 10.3390/jimaging8050148
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 149: Deep Neural Network for Cardiac Magnetic
Resonance Image Segmentation
Authors: David Chen, Huzefa Bhopalwala, Nakeya Dewaswala, Shivaram P. Arunachalam, Moein Enayati, Nasibeh Zanjirani Farahani, Kalyan Pasupathy, Sravani Lokineni, J. Martijn Bos, Peter A. Noseworthy, Reza Arsanjani, Bradley J. Erickson, Jeffrey B. Geske, Michael J. Ackerman, Philip A. Araoz, Adelaide M. Arruda-Olson
First page: 149
Abstract: The analysis and interpretation of cardiac magnetic resonance (CMR) images are often time-consuming. The automated segmentation of cardiac structures can reduce the time required for image analysis. Spatial similarities between different CMR image types were leveraged to jointly segment multiple sequences using a segmentation model termed a multi-image type UNet (MI-UNet). This model was developed from 72 exams (46% female, mean age 63 ± 11 years) performed on patients with hypertrophic cardiomyopathy. The MI-UNet for steady-state free precession (SSFP) images achieved a superior Dice similarity coefficient (DSC) of 0.92 ± 0.06 compared to 0.87 ± 0.08 for a single-image type UNet (p < 0.001). The MI-UNet for late gadolinium enhancement (LGE) images also had a superior DSC of 0.86 ± 0.11 compared to 0.78 ± 0.11 for a single-image type UNet (p = 0.001). The difference across image types was most evident for the left ventricular myocardium in SSFP images and for both the left ventricular cavity and the left ventricular myocardium in LGE images. For the right ventricle, there were no differences in DCS when comparing the MI-UNet with single-image type UNets. The joint segmentation of multiple image types increases segmentation accuracy for CMR images of the left ventricle compared to single-image models. In clinical practice, the MI-UNet model may expedite the analysis and interpretation of CMR images of multiple types.
Citation: Journal of Imaging
PubDate: 2022-05-23
DOI: 10.3390/jimaging8050149
Issue No: Vol. 8, No. 5 (2022)
- J. Imaging, Vol. 8, Pages 82: Three-Dimensional Pharyngeal Airway Space
Changes Following Isolated Mandibular Advancement Surgery in 120 Patients:
A 1-Year Follow-up Study
Authors: Sohaib Shujaat, Eman Shaheen, Marryam Riaz, Constantinus Politis, Reinhilde Jacobs
First page: 82
Abstract: Lack of evidence exists related to the three-dimensional (3D) pharyngeal airway space (PAS) changes at follow-up after isolated bilateral sagittal split osteotomy (BSSO) advancement surgery. The present study assessed the 3D PAS changes following isolated mandibular advancement at a follow-up period of 1 year. A total of 120 patients (40 males, 80 females, mean age: 26.0 ± 12.2) who underwent BSSO advancement surgery were recruited. Cone-beam computed tomography (CBCT) scans were acquired preoperatively (T0), immediately following surgery (T1), and at 1 year of follow-up (T2). The volume, surface area, and minimal cross-sectional area (mCSA) of the airway were assessed. The total airway showed a 38% increase in volume and 13% increase in surface area from T0 to T1, where the oropharyngeal region showed the maximum immediate change. At T1–T2 follow-up, both volumetric and surface area showed a relapse of less than 7% for all sub-regions. The mCSA showed a significant increase of 71% from T0 to T1 (p < 0.0001), whereas a non-significant relapse was observed at T1–T2 (p = 0.1252). The PAS remained stable at a follow-up period of 1 year. In conclusion, BSSO advancement surgery could be regarded as a stable procedure for widening of the PAS with maintenance of positive space at follow-up.
Citation: Journal of Imaging
PubDate: 2022-03-22
DOI: 10.3390/jimaging8040082
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 83: Generative Adversarial Networks in Brain
Imaging: A Narrative Review
Authors: Maria Elena Laino, Pierandrea Cancian, Letterio Salvatore Politi, Matteo Giovanni Della Porta, Luca Saba, Victor Savevski
First page: 83
Abstract: Artificial intelligence (AI) is expected to have a major effect on radiology as it demonstrated remarkable progress in many clinical tasks, mostly regarding the detection, segmentation, classification, monitoring, and prediction of diseases. Generative Adversarial Networks have been proposed as one of the most exciting applications of deep learning in radiology. GANs are a new approach to deep learning that leverages adversarial learning to tackle a wide array of computer vision challenges. Brain radiology was one of the first fields where GANs found their application. In neuroradiology, indeed, GANs open unexplored scenarios, allowing new processes such as image-to-image and cross-modality synthesis, image reconstruction, image segmentation, image synthesis, data augmentation, disease progression models, and brain decoding. In this narrative review, we will provide an introduction to GANs in brain imaging, discussing the clinical potential of GANs, future clinical applications, as well as pitfalls that radiologists should be aware of.
Citation: Journal of Imaging
PubDate: 2022-03-23
DOI: 10.3390/jimaging8040083
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 84: Addressing Motion Blurs in Brain MRI Scans
Using Conditional Adversarial Networks and Simulated Curvilinear Motions
Authors: Shangjin Li, Yijun Zhao
First page: 84
Abstract: In-scanner head motion often leads to degradation in MRI scans and is a major source of error in diagnosing brain abnormalities. Researchers have explored various approaches, including blind and nonblind deconvolutions, to correct the motion artifacts in MRI scans. Inspired by the recent success of deep learning models in medical image analysis, we investigate the efficacy of employing generative adversarial networks (GANs) to address motion blurs in brain MRI scans. We cast the problem as a blind deconvolution task where a neural network is trained to guess a blurring kernel that produced the observed corruption. Specifically, our study explores a new approach under the sparse coding paradigm where every ground truth corrupting kernel is assumed to be a “combination” of a relatively small universe of “basis” kernels. This assumption is based on the intuition that, on small distance scales, patients’ moves follow simple curves and that complex motions can be obtained by combining a number of simple ones. We show that, with a suitably dense basis, a neural network can effectively guess the degrading kernel and reverse some of the damage in the motion-affected real-world scans. To this end, we generated 10,000 continuous and curvilinear kernels in random positions and directions that are likely to uniformly populate the space of corrupting kernels in real-world scans. We further generated a large dataset of 225,000 pairs of sharp and blurred MR images to facilitate training effective deep learning models. Our experimental results demonstrate the viability of the proposed approach evaluated using synthetic and real-world MRI scans. Our study further suggests there is merit in exploring separate models for the sagittal, axial, and coronal planes.
Citation: Journal of Imaging
PubDate: 2022-03-23
DOI: 10.3390/jimaging8040084
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 85: Exploring Metrics to Establish an Optimal
Model for Image Aesthetic Assessment and Analysis
Authors: Ying Dai
First page: 85
Abstract: To establish an optimal model for photo aesthetic assessment, in this paper, an internal metric called the disentanglement-measure (D-measure) is introduced, which reflects the disentanglement degree of the final layer FC (full connection) nodes of convolutional neural network (CNN). By combining the F-measure with the D-measure to obtain an FD measure, an algorithm of determining the optimal model from many photo score prediction models generated by CNN-based repetitively self-revised learning (RSRL) is proposed. Furthermore, the aesthetics features of the model regarding the first fixation perspective (FFP) and the assessment interest region (AIR) are defined by means of the feature maps so as to analyze the consistency with human aesthetics. The experimental results show that the proposed method is helpful in improving the efficiency of determining the optimal model. Moreover, extracting the FFP and AIR of the models to the image is useful in understanding the internal properties of these models related to the human aesthetics and validating the external performances of the aesthetic assessment.
Citation: Journal of Imaging
PubDate: 2022-03-23
DOI: 10.3390/jimaging8040085
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 86: Improving Scene Text Recognition for Indian
Languages with Transfer Learning and Font Diversity
Authors: Sanjana Gunna, Rohit Saluja, Cheerakkuzhi Veluthemana Jawahar
First page: 86
Abstract: Reading Indian scene texts is complex due to the use of regional vocabulary, multiple fonts/scripts, and text size. This work investigates the significant differences in Indian and Latin Scene Text Recognition (STR) systems. Recent STR works rely on synthetic generators that involve diverse fonts to ensure robust reading solutions. We present utilizing additional non-Unicode fonts with generally employed Unicode fonts to cover font diversity in such synthesizers for Indian languages. We also perform experiments on transfer learning among six different Indian languages. Our transfer learning experiments on synthetic images with common backgrounds provide an exciting insight that Indian scripts can benefit from each other than from the extensive English datasets. Our evaluations for the real settings help us achieve significant improvements over previous methods on four Indian languages from standard datasets like IIIT-ILST, MLT-17, and the new dataset (we release) containing 440 scene images with 500 Gujarati and 2535 Tamil words. Further enriching the synthetic dataset with non-Unicode fonts and multiple augmentations helps us achieve a remarkable Word Recognition Rate gain of over 33% on the IIIT-ILST Hindi dataset. We also present the results of lexicon-based transcription approaches for all six languages.
Citation: Journal of Imaging
PubDate: 2022-03-23
DOI: 10.3390/jimaging8040086
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 87: Phase Retardation Analysis in a Rotated
Plane-Parallel Plate for Phase-Shifting Digital Holography
Authors: Igor Shevkunov, Nikolay V. Petrov
First page: 87
Abstract: In this paper, we detail a phase-shift implementation in a rotated plane-parallel plate (PPP). Considering the phase-shifting digital holography application, we provide a more precise phase-shift estimation based on PPP thickness, rotation, and mutual inclination of reference and object wavefronts. We show that phase retardation uncertainty implemented by the rotated PPP in a simple configuration is less than the uncertainty of a traditionally used piezoelectric translator. Physical experiments on a phase test target verify the high quality of phase reconstruction.
Citation: Journal of Imaging
PubDate: 2022-03-24
DOI: 10.3390/jimaging8040087
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 88: YOLOv4-Based CNN Model versus Nested
Contours Algorithm in the Suspicious Lesion Detection on the Mammography
Image: A Direct Comparison in the Real Clinical Settings
Authors: Alexey Kolchev, Dmitry Pasynkov, Ivan Egoshin, Ivan Kliouchkin, Olga Pasynkova, Dmitrii Tumakov
First page: 88
Abstract: Background: We directly compared the mammography image processing results obtained with the help of the YOLOv4 convolutional neural network (CNN) model versus those obtained with the help of the NCA-based nested contours algorithm model. Method: We used 1080 images to train the YOLOv4, plus 100 images with proven breast cancer (BC) and 100 images with proven absence of BC to test both models. Results: the rates of true-positive, false-positive and false-negative outcomes were 60, 10 and 40, respectively, for YOLOv4, and 93, 63 and 7, respectively, for NCA. The sensitivities for the YOLOv4 and the NCA were comparable to each other for star-like lesions, masses with unclear borders, round- or oval-shaped masses with clear borders and partly visualized masses. On the contrary, the NCA was superior to the YOLOv4 in the case of asymmetric density and of changes invisible on the dense parenchyma background. Radiologists changed their earlier decisions in six cases per 100 for NCA. YOLOv4 outputs did not influence the radiologists’ decisions. Conclusions: in our set, NCA clinically significantly surpasses YOLOv4.
Citation: Journal of Imaging
PubDate: 2022-03-24
DOI: 10.3390/jimaging8040088
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 89: Union-Retire for Connected Components
Analysis on FPGA
Authors: Donald G. Bailey, Michael J. Klaiber
First page: 89
Abstract: The Union-Retire CCA (UR-CCA) algorithm started a new paradigm for connected components analysis. Instead of using directed tree structures, UR-CCA focuses on connectivity. This algorithmic change leads to a reduction in required memory, with no end-of-row processing overhead. In this paper we describe a hardware architecture based on UR-CCA and its realisation on an FPGA. The memory bandwidth and pipelining challenges of hardware UR-CCA are analysed and resolved. It is shown that up to 36% of memory resources can be saved using the proposed architecture. This translates directly to a smaller device for an FPGA implementation.
Citation: Journal of Imaging
PubDate: 2022-03-24
DOI: 10.3390/jimaging8040089
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 90: Object Categorization Capability of
Psychological Potential Field in Perceptual Assessment Using Line-Drawing
Images
Authors: Naoyuki Awano, Yuki Hayashi
First page: 90
Abstract: Affective/cognitive engineering investigations typically require the quantitative assessment of object perception. Recent research has suggested that certain perceptions of object categorization can be derived from human eye fixation and that color images and line drawings induce similar neural activities. Line drawings contain less information than color images; therefore, line drawings are expected to simplify the investigations of object perception. The psychological potential field (PPF), which is a psychological feature, is an image feature of line drawings. On the basis of the PPF, the possibility that the general human perception of object categorization can be assessed from the similarity to fixation maps (FMs) generated from human eye fixations has been reported. However, this may be due to chance because image features other than the PPF have not been compared with FMs. This study examines the potential and effectiveness of the PPF by comparing its performance with that of other image features in terms of the similarity to FMs. The results show that the PPF shows the ideal performance for assessing the perception of object categorization. In particular, the PPF effectively distinguishes between animal and nonanimal targets; however, real-time assessment is difficult.
Citation: Journal of Imaging
PubDate: 2022-03-26
DOI: 10.3390/jimaging8040090
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 91: Augmented Reality Games and Presence: A
Systematic Review
Authors: Anabela Marto, Alexandrino Gonçalves
First page: 91
Abstract: The sense of presence in augmented reality (AR) has been studied by multiple researchers through diverse applications and strategies. In addition to the valuable information provided to the scientific community, new questions keep being raised. These approaches vary from following the standards from virtual reality to ascertaining the presence of users’ experiences and new proposals for evaluating presence that specifically target AR environments. It is undeniable that the idea of evaluating presence across AR may be overwhelming due to the different scenarios that may be possible, whether this regards technological devices—from immersive AR headsets to the small screens of smartphones—or the amount of virtual information that is being added to the real scenario. Taking into account the recent literature that has addressed the sense of presence in AR as a true challenge given the diversity of ways that AR can be experienced, this study proposes a specific scope to address presence and other related forms of dimensions such as immersion, engagement, embodiment, or telepresence, when AR is used in games. This systematic review was conducted following the PRISMA methodology, carefully analysing all studies that reported visual games that include AR activities and somehow included presence data—or related dimensions that may be referred to as immersion-related feelings, analysis or results. This study clarifies what dimensions of presence are being considered and evaluated in AR games, how presence-related variables have been evaluated, and what the major research findings are. For a better understanding of these approaches, this study takes note of what devices are being used for the AR experience when immersion-related feelings are one of the behaviours that are considered in their evaluations, and discusses to what extent these feelings in AR games affect the player’s other behaviours.
Citation: Journal of Imaging
PubDate: 2022-03-29
DOI: 10.3390/jimaging8040091
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 92: A New Preclinical Decision Support System
Based on PET Radiomics: A Preliminary Study on the Evaluation of an
Innovative 64Cu-Labeled Chelator in Mouse Models
Authors: Viviana Benfante, Alessandro Stefano, Albert Comelli, Paolo Giaccone, Francesco Paolo Cammarata, Selene Richiusa, Fabrizio Scopelliti, Marco Pometti, Milene Ficarra, Sebastiano Cosentino, Marcello Lunardon, Francesca Mastrotto, Alberto Andrighetto, Antonino Tuttolomondo, Rosalba Parenti, Massimo Ippolito, Giorgio Russo
First page: 92
Abstract: The 64Cu-labeled chelator was analyzed in vivo by positron emission tomography (PET) imaging to evaluate its biodistribution in a murine model at different acquisition times. For this purpose, nine 6-week-old female Balb/C nude strain mice underwent micro-PET imaging at three different time points after 64Cu-labeled chelator injection. Specifically, the mice were divided into group 1 (acquisition 1 h after [64Cu] chelator administration, n = 3 mice), group 2 (acquisition 4 h after [64Cu]chelator administration, n = 3 mice), and group 3 (acquisition 24 h after [64Cu] chelator administration, n = 3 mice). Successively, all PET studies were segmented by means of registration with a standard template space (3D whole-body Digimouse atlas), and 108 radiomics features were extracted from seven organs (namely, heart, bladder, stomach, liver, spleen, kidney, and lung) to investigate possible changes over time in [64Cu]chelator biodistribution. The one-way analysis of variance and post hoc Tukey Honestly Significant Difference test revealed that, while heart, stomach, spleen, kidney, and lung districts showed a very low percentage of radiomics features with significant variations (p-value < 0.05) among the three groups of mice, a large number of features (greater than 60% and 50%, respectively) that varied significantly between groups were observed in bladder and liver, indicating a different in vivo uptake of the 64Cu-labeled chelator over time. The proposed methodology may improve the method of calculating the [64Cu]chelator biodistribution and open the way towards a decision support system in the field of new radiopharmaceuticals used in preclinical imaging trials.
Citation: Journal of Imaging
PubDate: 2022-03-30
DOI: 10.3390/jimaging8040092
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 93: Unified Probabilistic Deep Continual
Learning through Generative Replay and Open Set Recognition
Authors: Martin Mundt, Iuliia Pliushch, Sagnik Majumder, Yongwon Hong, Visvanathan Ramesh
First page: 93
Abstract: Modern deep neural networks are well known to be brittle in the face of unknown data instances and recognition of the latter remains a challenge. Although it is inevitable for continual-learning systems to encounter such unseen concepts, the corresponding literature appears to nonetheless focus primarily on alleviating catastrophic interference with learned representations. In this work, we introduce a probabilistic approach that connects these perspectives based on variational inference in a single deep autoencoder model. Specifically, we propose to bound the approximate posterior by fitting regions of high density on the basis of correctly classified data points. These bounds are shown to serve a dual purpose: unseen unknown out-of-distribution data can be distinguished from already trained known tasks towards robust application. Simultaneously, to retain already acquired knowledge, a generative replay process can be narrowed to strictly in-distribution samples, in order to significantly alleviate catastrophic interference.
Citation: Journal of Imaging
PubDate: 2022-03-31
DOI: 10.3390/jimaging8040093
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 94: Imaging PPG for In Vivo Human Tissue
Perfusion Assessment during Surgery
Authors: Marco Lai, Stefan D. van der Stel, Harald C. Groen, Mark van Gastel, Koert F. D. Kuhlmann, Theo J. M. Ruers, Benno H. W. Hendriks
First page: 94
Abstract: Surgical excision is the golden standard for treatment of intestinal tumors. In this surgical procedure, inadequate perfusion of the anastomosis can lead to postoperative complications, such as anastomotic leakages. Imaging photoplethysmography (iPPG) can potentially provide objective and real-time feedback of the perfusion status of tissues. This feasibility study aims to evaluate an iPPG acquisition system during intestinal surgeries to detect the perfusion levels of the microvasculature tissue bed in different perfusion conditions. This feasibility study assesses three patients that underwent resection of a portion of the small intestine. Data was acquired from fully perfused, non-perfused and anastomosis parts of the intestine during different phases of the surgical procedure. Strategies for limiting motion and noise during acquisition were implemented. iPPG perfusion maps were successfully extracted from the intestine microvasculature, demonstrating that iPPG can be successfully used for detecting perturbations and perfusion changes in intestinal tissues during surgery. This study provides proof of concept for iPPG to detect changes in organ perfusion levels.
Citation: Journal of Imaging
PubDate: 2022-03-31
DOI: 10.3390/jimaging8040094
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 95: Spiky: An ImageJ Plugin for Data Analysis of
Functional Cardiac and Cardiomyocyte Studies
Authors: Côme Pasqualin, François Gannier, Angèle Yu, David Benoist, Ian Findlay, Romain Bordy, Pierre Bredeloux, Véronique Maupoil
First page: 95
Abstract: Introduction and objective: Nowadays, investigations of heart physiology and pathophysiology rely more and more upon image analysis, whether for the detection and characterization of events in single cells or for the mapping of events and their characteristics across an entire tissue. These investigations require extensive skills in image analysis and/or expensive software, and their reproducibility may be a concern. Our objective was to build a robust, reliable and open-source software tool to quantify excitation–contraction related experimental data at multiple scales, from single isolated cells to the whole heart. Methods and results: A free and open-source ImageJ plugin, Spiky, was developed to detect and analyze peaks in experimental data streams. It allows rapid and easy analysis of action potentials, intracellular calcium transient and contraction data from cardiac research experiments. As shown in the provided examples, both classical bi-dimensional data (XT signals) and video data obtained from confocal microscopy and optical mapping experiments (XYT signals) can be analyzed. Spiky was written in ImageJ Macro Language and JAVA, and works under Windows, Mac and Linux operating systems. Conclusion: Spiky provides a complete working interface to process and analyze cardiac physiology research data.
Citation: Journal of Imaging
PubDate: 2022-04-01
DOI: 10.3390/jimaging8040095
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 96: Convolutional Neural Networks for the
Identification of African Lions from Individual Vocalizations
Authors: Martino Trapanotto, Loris Nanni, Sheryl Brahnam, Xiang Guo
First page: 96
Abstract: The classification of vocal individuality for passive acoustic monitoring (PAM) and census of animals is becoming an increasingly popular area of research. Nearly all studies in this field of inquiry have relied on classic audio representations and classifiers, such as Support Vector Machines (SVMs) trained on spectrograms or Mel-Frequency Cepstral Coefficients (MFCCs). In contrast, most current bioacoustic species classification exploits the power of deep learners and more cutting-edge audio representations. A significant reason for avoiding deep learning in vocal identity classification is the tiny sample size in the collections of labeled individual vocalizations. As is well known, deep learners require large datasets to avoid overfitting. One way to handle small datasets with deep learning methods is to use transfer learning. In this work, we evaluate the performance of three pretrained CNNs (VGG16, ResNet50, and AlexNet) on a small, publicly available lion roar dataset containing approximately 150 samples taken from five male lions. Each of these networks is retrained on eight representations of the samples: MFCCs, spectrogram, and Mel spectrogram, along with several new ones, such as VGGish and stockwell, and those based on the recently proposed LM spectrogram. The performance of these networks, both individually and in ensembles, is analyzed and corroborated using the Equal Error Rate and shown to surpass previous classification attempts on this dataset; the best single network achieved over 95% accuracy and the best ensembles over 98% accuracy. The contributions this study makes to the field of individual vocal classification include demonstrating that it is valuable and possible, with caution, to use transfer learning with single pretrained CNNs on the small datasets available for this problem domain. We also make a contribution to bioacoustics generally by offering a comparison of the performance of many state-of-the-art audio representations, including for the first time the LM spectrogram and stockwell representations. All source code for this study is available on GitHub.
Citation: Journal of Imaging
PubDate: 2022-04-01
DOI: 10.3390/jimaging8040096
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 97: Machine Learning for Early Parkinson’s
Disease Identification within SWEDD Group Using Clinical and DaTSCAN SPECT
Imaging Features
Authors: Hajer Khachnaoui, Nawres Khlifa, Rostom Mabrouk
First page: 97
Abstract: Early Parkinson’s Disease (PD) diagnosis is a critical challenge in the treatment process. Meeting this challenge allows appropriate planning for patients. However, Scan Without Evidence of Dopaminergic Deficit (SWEDD) is a heterogeneous group of PD patients and Healthy Controls (HC) in clinical and imaging features. The application of diagnostic tools based on Machine Learning (ML) comes into play here as they are capable of distinguishing between HC subjects and PD patients within an SWEDD group. In the present study, three ML algorithms were used to separate PD patients from HC within an SWEDD group. Data of 548 subjects were firstly analyzed by Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) techniques. Using the best reduction technique result, we built the following clustering models: Density-Based Spatial (DBSCAN), K-means and Hierarchical Clustering. According to our findings, LDA performs better than PCA; therefore, LDA was used as input for the clustering models. The different models’ performances were assessed by comparing the clustering algorithms outcomes with the ground truth after a follow-up. Hierarchical Clustering surpassed DBSCAN and K-means algorithms by 64%, 78.13% and 38.89% in terms of accuracy, sensitivity and specificity. The proposed method demonstrated the suitability of ML models to distinguish PD patients from HC subjects within an SWEDD group.
Citation: Journal of Imaging
PubDate: 2022-04-02
DOI: 10.3390/jimaging8040097
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 98: A Comparative Review on Applications of
Different Sensors for Sign Language Recognition
Authors: Muhammad Saad Amin, Syed Tahir Hussain Rizvi, Md. Murad Hossain
First page: 98
Abstract: Sign language recognition is challenging due to the lack of communication between normal and affected people. Many social and physiological impacts are created due to speaking or hearing disability. A lot of different dimensional techniques have been proposed previously to overcome this gap. A sensor-based smart glove for sign language recognition (SLR) proved helpful to generate data based on various hand movements related to specific signs. A detailed comparative review of all types of available techniques and sensors used for sign language recognition was presented in this article. The focus of this paper was to explore emerging trends and strategies for sign language recognition and to point out deficiencies in existing systems. This paper will act as a guide for other researchers to understand all materials and techniques like flex resistive sensor-based, vision sensor-based, or hybrid system-based technologies used for sign language until now.
Citation: Journal of Imaging
PubDate: 2022-04-02
DOI: 10.3390/jimaging8040098
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 99: Cardiac Magnetic Resonance Imaging in Immune
Check-Point Inhibitor Myocarditis: A Systematic Review
Authors: Luca Arcari, Giacomo Tini, Giovanni Camastra, Federica Ciolina, Domenico De Santis, Domitilla Russo, Damiano Caruso, Massimiliano Danti, Luca Cacciotti
First page: 99
Abstract: Immune checkpoint inhibitors (ICIs) are a family of anticancer drugs in which the immune response elicited against the tumor may involve other organs, including the heart. Cardiac magnetic resonance (CMR) imaging is increasingly used in the diagnostic work-up of myocardial inflammation; recently, several studies investigated the use of CMR in patients with ICI-myocarditis (ICI-M). The aim of the present systematic review is to summarize the available evidence on CMR findings in ICI-M. We searched electronic databases for relevant publications; after screening, six studies were selected, including 166 patients from five cohorts, and further 86 patients from a sub-analysis that were targeted for a tissue mapping assessment. CMR revealed mostly preserved left ventricular ejection fraction; edema prevalence ranged from 9% to 60%; late gadolinium enhancement (LGE) prevalence ranged from 23% to 83%. T1 and T2 mapping assessment were performed in 108 and 104 patients, respectively. When available, the comparison of CMR with endomyocardial biopsy revealed partial agreement between techniques and was higher for native T1 mapping amongst imaging biomarkers. The prognostic assessment was inconsistently assessed; CMR variables independently associated with the outcome included decreasing LVEF and increasing native T1. In conclusion, CMR findings in ICI-M include myocardial dysfunction, edema and fibrosis, though less evident than in more classic forms of myocarditis; native T1 mapping retained the higher concordance with EMB and significant prognostic value.
Citation: Journal of Imaging
PubDate: 2022-04-05
DOI: 10.3390/jimaging8040099
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 100: Transverse Analysis of Maxilla and Mandible
in Adults with Normal Occlusion: A Cone Beam Computed Tomography Study
Authors: Kyung Jin Lee, Hyeran Helen Jeon, Normand Boucher, Chun-Hsi Chung
First page: 100
Abstract: Objectives: To study the transverse widths of maxilla and mandible and their relationship with the inclination of first molars. Materials and Methods: Fifty-six untreated adults (12 males, 44 females) with normal occlusion were included. On each Cone Beam Computed Tomography (CBCT) image of the subject, inter-buccal and inter-lingual bone widths were measured at the levels of hard palate, alveolar crest and furcation of the first molars, and maxillomandibular width differentials were calculated. In addition, the buccolingual inclination of each first molar was measured and its correlation with the maxillomandibular width differential was tested. Results: At the furcation level of the first molar, the maxillary inter-buccal bone width was more than the mandibular inter-buccal bone width by 1.1 ± 4.5 mm for males and 1.6 ± 2.9 mm for females; the mandibular inter-lingual bone width was more than the maxillary inter-lingual bone width by 1.3 ± 3.6 mm for males and 0.3 ± 3.2 mm for females. For females, there was a negative correlation between the maxillomandibular inter-lingual bone differential and maxillary first molar buccal inclination (p < 0.05), and a positive correlation between the maxillomandibular inter-lingual bone differential and mandibular first molar lingual inclination (p < 0.05). Conclusions: This is a randomized clinical study on transverse analysis of maxilla and mandible in adults with normal occlusion using CBCTs. On average: (1) At the furcation level of the first molars, the maxillary inter-buccal bone width was slightly wider than mandibular inter-buccal bone width; whereas the mandibular inter-lingual bone width was slightly wider than maxillary inter-lingual bone width; (2) A statistically significant correlation existed between the maxillomandibular transverse skeletal differentials and molar inclinations.
Citation: Journal of Imaging
PubDate: 2022-04-05
DOI: 10.3390/jimaging8040100
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 101: Machine-Learning-Based Real-Time
Multi-Camera Vehicle Tracking and Travel-Time Estimation
Authors: Xiaohui Huang, Pan He, Anand Rangarajan, Sanjay Ranka
First page: 101
Abstract: Travel-time estimation of traffic flow is an important problem with critical implications for traffic congestion analysis. We developed techniques for using intersection videos to identify vehicle trajectories across multiple cameras and analyze corridor travel time. Our approach consists of (1) multi-object single-camera tracking, (2) vehicle re-identification among different cameras, (3) multi-object multi-camera tracking, and (4) travel-time estimation. We evaluated the proposed framework on real intersections in Florida with pan and fisheye cameras. The experimental results demonstrate the viability and effectiveness of our method.
Citation: Journal of Imaging
PubDate: 2022-04-06
DOI: 10.3390/jimaging8040101
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 102: Novel Hypertrophic Cardiomyopathy Diagnosis
Index Using Deep Features and Local Directional Pattern Techniques
Authors: Anjan Gudigar, U. Raghavendra, Jyothi Samanth, Chinmay Dharmik, Mokshagna Rohit Gangavarapu, Krishnananda Nayak, Edward J. Ciaccio, Ru-San Tan, Filippo Molinari, U. Rajendra Acharya
First page: 102
Abstract: Hypertrophic cardiomyopathy (HCM) is a genetic disorder that exhibits a wide spectrum of clinical presentations, including sudden death. Early diagnosis and intervention may avert the latter. Left ventricular hypertrophy on heart imaging is an important diagnostic criterion for HCM, and the most common imaging modality is heart ultrasound (US). The US is operator-dependent, and its interpretation is subject to human error and variability. We proposed an automated computer-aided diagnostic tool to discriminate HCM from healthy subjects on US images. We used a local directional pattern and the ResNet-50 pretrained network to classify heart US images acquired from 62 known HCM patients and 101 healthy subjects. Deep features were ranked using Student’s t-test, and the most significant feature (SigFea) was identified. An integrated index derived from the simulation was defined as 100·log10(SigFea/2) in each subject, and a diagnostic threshold value was empirically calculated as the mean of the minimum and maximum integrated indices among HCM and healthy subjects, respectively. An integrated index above a threshold of 0.5 separated HCM from healthy subjects with 100% accuracy in our test dataset.
Citation: Journal of Imaging
PubDate: 2022-04-06
DOI: 10.3390/jimaging8040102
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 103: A Hybrid Method for 3D Reconstruction of MR
Images
Authors: Loubna Lechelek, Sebastien Horna, Rita Zrour, Mathieu Naudin, Carole Guillevin
First page: 103
Abstract: Three-dimensional surface reconstruction is a well-known task in medical imaging. In procedures for intervention or radiation treatment planning, the generated models should be accurate and reflect the natural appearance. Traditional methods for this task, such as Marching Cubes, use smoothing post processing to reduce staircase artifacts from mesh generation and exhibit the natural look. However, smoothing algorithms often reduce the quality and degrade the accuracy. Other methods, such as MPU implicits, based on adaptive implicit functions, inherently produce smooth 3D models. However, the integration in the implicit functions of both smoothness and accuracy of the shape approximation may impact the precision of the reconstruction. Having these limitations in mind, we propose a hybrid method for 3D reconstruction of MR images. This method is based on a parallel Marching Cubes algorithm called Flying Edges (FE) and Multi-level Partition of Unity (MPU) implicits. We aim to combine the robustness of the Marching Cubes algorithm with the smooth implicit curve tracking enabled by the use of implicit models in order to provide higher geometry precision. Towards this end, the regions that closely fit to the segmentation data, and thus regions that are not impacted by reconstruction issues, are first extracted from both methods. These regions are then merged and used to reconstruct the final model. Experimental studies were performed on a number of MRI datasets, providing images and error statistics generated from our results. The results obtained show that our method reduces the geometric errors of the reconstructed surfaces when compared to the MPU and FE approaches, producing a more accurate 3D reconstruction.
Citation: Journal of Imaging
PubDate: 2022-04-07
DOI: 10.3390/jimaging8040103
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 104: Explainable Multimedia Feature Fusion for
Medical Applications
Authors: Stefan Wagenpfeil, Paul Mc Kevitt, Abbas Cheddad, Matthias Hemmje
First page: 104
Abstract: Due to the exponential growth of medical information in the form of, e.g., text, images, Electrocardiograms (ECGs), X-ray, multimedia, etc., the management of a patient’s data has become a huge challenge. Particularly, the extraction of features from various different formats and their representation in a homogeneous way are areas of particular interest in medical applications. Multimedia Information Retrieval (MMIR) frameworks, like the Generic Multimedia Analysis Framework (GMAF), can contribute to solving this problem, when adapted to special requirements and modalities of medical applications. In this paper, we demonstrate how typical multimedia processing techniques can be extended and adapted to medical applications and how these applications benefit from employing a Multimedia Feature Graph (MMFG) and specialized, highly efficient indexing structures in the form of Graph Codes. These Graph Codes are transformed to feature relevant Graph Codes by employing a modified Term Frequency Inverse Document Frequency (TFIDF) algorithm, which further supports value ranges and Boolean operations required in the medical context. On this basis, various metrics for the calculation of similarity, recommendations, and automated inferencing and reasoning can be applied supporting the field of diagnostics. Finally, the presentation of these new facilities in the form of explainability is introduced and demonstrated. Thus, in this paper, we show how Graph Codes contribute new querying options for diagnosis and how Explainable Graph Codes can help to quickly understand medical multimedia formats.
Citation: Journal of Imaging
PubDate: 2022-04-08
DOI: 10.3390/jimaging8040104
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 105: Face Attribute Estimation Using Multi-Task
Convolutional Neural Network
Authors: Hiroya Kawai, Koichi Ito, Takafumi Aoki
First page: 105
Abstract: Face attribute estimation can be used for improving the accuracy of face recognition, customer analysis in marketing, image retrieval, video surveillance, and criminal investigation. The major methods for face attribute estimation are based on Convolutional Neural Networks (CNNs) that solve face attribute estimation as a multiple two-class classification problem. Although one feature extractor should be used for each attribute to explore the accuracy of attribute estimation, in most cases, one feature extractor is shared to estimate all face attributes for the parameter efficiency. This paper proposes a face attribute estimation method using Merged Multi-CNN (MM-CNN) to automatically optimize CNN structures for solving multiple binary classification problems to improve parameter efficiency and accuracy in face attribute estimation. We also propose a parameter reduction method called Convolutionalization for Parameter Reduction (CPR), which removes all fully connected layers from MM-CNNs. Through a set of experiments using the CelebA and LFW-a datasets, we demonstrate that MM-CNN with CPR exhibits higher efficiency of face attribute estimation in terms of estimation accuracy and the number of weight parameters than conventional methods.
Citation: Journal of Imaging
PubDate: 2022-04-10
DOI: 10.3390/jimaging8040105
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 106: Adaptive Real-Time Object Detection for
Autonomous Driving Systems
Authors: Maryam Hemmati, Morteza Biglari-Abhari, Smail Niar
First page: 106
Abstract: Accurate and reliable detection is one of the main tasks of Autonomous Driving Systems (ADS). While detecting the obstacles on the road during various environmental circumstances add to the reliability of ADS, it results in more intensive computations and more complicated systems. The stringent real-time requirements of ADS, resource constraints, and energy efficiency considerations add to the design complications. This work presents an adaptive system that detects pedestrians and vehicles in different lighting conditions on the road. We take a hardware-software co-design approach on Zynq UltraScale+ MPSoC and develop a dynamically reconfigurable ADS that employs hardware accelerators for pedestrian and vehicle detection and adapts its detection method to the environment lighting conditions. The results show that the system maintains real-time performance and achieves adaptability with minimal resource overhead.
Citation: Journal of Imaging
PubDate: 2022-04-11
DOI: 10.3390/jimaging8040106
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 107: Example-Based Multispectral Photometric
Stereo for Multi-Colored Surfaces
Authors: Daisuke Miyazaki, Kazuya Uegomori
First page: 107
Abstract: A photometric stereo needs three images taken under three different light directions lit one by one, while a color photometric stereo needs only one image taken under three different lights lit at the same time with different light directions and different colors. As a result, a color photometric stereo can obtain the surface normal of a dynamically moving object from a single image. However, the conventional color photometric stereo cannot estimate a multicolored object due to the colored illumination. This paper uses an example-based photometric stereo to solve the problem of the color photometric stereo. The example-based photometric stereo searches the surface normal from the database of the images of known shapes. Color photometric stereos suffer from mathematical difficulty, and they add many assumptions and constraints; however, the example-based photometric stereo is free from such mathematical problems. The process of our method is pixelwise; thus, the estimated surface normal is not oversmoothed, unlike existing methods that use smoothness constraints. To demonstrate the effectiveness of this study, a measurement device that can realize the multispectral photometric stereo method with sixteen colors is employed instead of the classic color photometric stereo method with three colors.
Citation: Journal of Imaging
PubDate: 2022-04-11
DOI: 10.3390/jimaging8040107
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 108: Multi-Stage Platform for (Semi-)Automatic
Planning in Reconstructive Orthopedic Surgery
Authors: Florian Kordon, Andreas Maier, Benedict Swartman, Maxim Privalov, Jan Siad El-Barbari, Holger Kunze
First page: 108
Abstract: Intricate lesions of the musculoskeletal system require reconstructive orthopedic surgery to restore the correct biomechanics. Careful pre-operative planning of the surgical steps on 2D image data is an essential tool to increase the precision and safety of these operations. However, the plan’s effectiveness in the intra-operative workflow is challenged by unpredictable patient and device positioning and complex registration protocols. Here, we develop and analyze a multi-stage algorithm that combines deep learning-based anatomical feature detection and geometric post-processing to enable accurate pre- and intra-operative surgery planning on 2D X-ray images. The algorithm allows granular control over each element of the planning geometry, enabling real-time adjustments directly in the operating room (OR). In the method evaluation of three ligament reconstruction tasks effect on the knee joint, we found high spatial precision in drilling point localization (ε<2.9mm) and low angulation errors for k-wire instrumentation (ε<0.75∘) on 38 diagnostic radiographs. Comparable precision was demonstrated in 15 complex intra-operative trauma cases suffering from strong implant overlap and multi-anatomy exposure. Furthermore, we found that the diverse feature detection tasks can be efficiently solved with a multi-task network topology, improving precision over the single-task case. Our platform will help overcome the limitations of current clinical practice and foster surgical plan generation and adjustment directly in the OR, ultimately motivating the development of novel 2D planning guidelines.
Citation: Journal of Imaging
PubDate: 2022-04-12
DOI: 10.3390/jimaging8040108
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 109: Tracking Highly Similar Rat Instances under
Heavy Occlusions: An Unsupervised Deep Generative Pipeline
Authors: Anna Gelencsér-Horváth, László Kopácsi, Viktor Varga, Dávid Keller, Árpád Dobolyi, Kristóf Karacs, András Lőrincz
First page: 109
Abstract: Identity tracking and instance segmentation are crucial in several areas of biological research. Behavior analysis of individuals in groups of similar animals is a task that emerges frequently in agriculture or pharmaceutical studies, among others. Automated annotation of many hours of surveillance videos can facilitate a large number of biological studies/experiments, which otherwise would not be feasible. Solutions based on machine learning generally perform well in tracking and instance segmentation; however, in the case of identical, unmarked instances (e.g., white rats or mice), even state-of-the-art approaches can frequently fail. We propose a pipeline of deep generative models for identity tracking and instance segmentation of highly similar instances, which, in contrast to most region-based approaches, exploits edge information and consequently helps to resolve ambiguity in heavily occluded cases. Our method is trained by synthetic data generation techniques, not requiring prior human annotation. We show that our approach greatly outperforms other state-of-the-art unsupervised methods in identity tracking and instance segmentation of unmarked rats in real-world laboratory video recordings.
Citation: Journal of Imaging
PubDate: 2022-04-13
DOI: 10.3390/jimaging8040109
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 110: Salient Object Detection by LTP Texture
Characterization on Opposing Color Pairs under SLICO Superpixel Constraint
Authors: Didier Ndayikengurukiye, Max Mignotte
First page: 110
Abstract: The effortless detection of salient objects by humans has been the subject of research in several fields, including computer vision, as it has many applications. However, salient object detection remains a challenge for many computer models dealing with color and textured images. Most of them process color and texture separately and therefore implicitly consider them as independent features which is not the case in reality. Herein, we propose a novel and efficient strategy, through a simple model, almost without internal parameters, which generates a robust saliency map for a natural image. This strategy consists of integrating color information into local textural patterns to characterize a color micro-texture. It is the simple, yet powerful LTP (Local Ternary Patterns) texture descriptor applied to opposing color pairs of a color space that allows us to achieve this end. Each color micro-texture is represented by a vector whose components are from a superpixel obtained by the SLICO (Simple Linear Iterative Clustering with zero parameter) algorithm, which is simple, fast and exhibits state-of-the-art boundary adherence. The degree of dissimilarity between each pair of color micro-textures is computed by the FastMap method, a fast version of MDS (Multi-dimensional Scaling) that considers the color micro-textures’ non-linearity while preserving their distances. These degrees of dissimilarity give us an intermediate saliency map for each RGB (Red–Green–Blue), HSL (Hue–Saturation–Luminance), LUV (L for luminance, U and V represent chromaticity values) and CMY (Cyan–Magenta–Yellow) color space. The final saliency map is their combination to take advantage of the strength of each of them. The MAE (Mean Absolute Error), MSE (Mean Squared Error) and Fβ measures of our saliency maps, on the five most used datasets show that our model outperformed several state-of-the-art models. Being simple and efficient, our model could be combined with classic models using color contrast for a better performance.
Citation: Journal of Imaging
PubDate: 2022-04-13
DOI: 10.3390/jimaging8040110
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 111: Reliability of OMERACT Scoring System in
Ultra-High Frequency Ultrasonography of Minor Salivary Glands: Inter-Rater
Agreement Study
Authors: Rossana Izzetti, Giovanni Fulvio, Marco Nisi, Stefano Gennai, Filippo Graziani
First page: 111
Abstract: Minor salivary gland ultra-high frequency ultrasonography (UHFUS) has recently been introduced for the evaluation of patients with suspected primary Sjögren’s Syndrome (pSS). At present, ultrasonographic assessment of major salivary glands is performed using the Outcome Measures in Rheumatology (OMERACT) scoring system. Previous reports have explored the possibility of applying the OMERACT scoring system to minor salivary glands UHFUS, with promising results. The aim of this study was to test the inter-reader concordance in the assignment of the OMERACT score to minor salivary gland UHFUS. The study was conducted on 170 minor salivary glands UHFUS scans of patients with suspected pSS. Three independent readers performed UHFUS image evaluation. Intraclass correlation coefficient (ICC) was employed to assess inter-reader reliability. Bland and Altman analysis was employed to test the agreement with a gold standard examiner. ICC values > 0.9 were found for scores 0 and 1, while score 2 and score 3 presented ICCs of 0.873 and 0.785, respectively. The measurements performed by the three examiners were in agreement with the gold standard examiner. According to these results, UHFUS interpretation showed good inter-observer reliability, suggesting that OMERACT score can be effectively used for the evaluation of glandular alterations, even for minor salivary glands.
Citation: Journal of Imaging
PubDate: 2022-04-15
DOI: 10.3390/jimaging8040111
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 112: Spectral Photon-Counting Computed
Tomography: A Review on Technical Principles and Clinical Applications
Authors: Mario Tortora, Laura Gemini, Imma D’Iglio, Lorenzo Ugga, Gaia Spadarella, Renato Cuocolo
First page: 112
Abstract: Photon-counting computed tomography (CT) is a technology that has attracted increasing interest in recent years since, thanks to new-generation detectors, it holds the promise to radically change the clinical use of CT imaging. Photon-counting detectors overcome the major limitations of conventional CT detectors by providing very high spatial resolution without electronic noise, providing a higher contrast-to-noise ratio, and optimizing spectral images. Additionally, photon-counting CT can lead to reduced radiation exposure, reconstruction of higher spatial resolution images, reduction of image artifacts, optimization of the use of contrast agents, and create new opportunities for quantitative imaging. The aim of this review is to briefly explain the technical principles of photon-counting CT and, more extensively, the potential clinical applications of this technology.
Citation: Journal of Imaging
PubDate: 2022-04-15
DOI: 10.3390/jimaging8040112
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 113: Fuzzy Information Discrimination Measures
and Their Application to Low Dimensional Embedding Construction in the
UMAP Algorithm
Authors: Liliya A. Demidova, Artyom V. Gorchakov
First page: 113
Abstract: Dimensionality reduction techniques are often used by researchers in order to make high dimensional data easier to interpret visually, as data visualization is only possible in low dimensional spaces. Recent research in nonlinear dimensionality reduction introduced many effective algorithms, including t-distributed stochastic neighbor embedding (t-SNE), uniform manifold approximation and projection (UMAP), dimensionality reduction technique based on triplet constraints (TriMAP), and pairwise controlled manifold approximation (PaCMAP), aimed to preserve both the local and global structure of high dimensional data while reducing the dimensionality. The UMAP algorithm has found its application in bioinformatics, genetics, genomics, and has been widely used to improve the accuracy of other machine learning algorithms. In this research, we compare the performance of different fuzzy information discrimination measures used as loss functions in the UMAP algorithm while constructing low dimensional embeddings. In order to achieve this, we derive the gradients of the considered losses analytically and employ the Adam algorithm during the loss function optimization process. From the conducted experimental studies we conclude that the use of either the logarithmic fuzzy cross entropy loss without reduced repulsion or the symmetric logarithmic fuzzy cross entropy loss with sufficiently large neighbor count leads to better global structure preservation of the original multidimensional data when compared to the loss function used in the original UMAP algorithm implementation.
Citation: Journal of Imaging
PubDate: 2022-04-15
DOI: 10.3390/jimaging8040113
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 114: Resources and Power Efficient FPGA
Accelerators for Real-Time Image Classification
Authors: Angelos Kyriakos, Elissaios-Alexios Papatheofanous, Charalampos Bezaitis, Dionysios Reisis
First page: 114
Abstract: A plethora of image and video-related applications involve complex processes that impose the need for hardware accelerators to achieve real-time performance. Among these, notable applications include the Machine Learning (ML) tasks using Convolutional Neural Networks (CNNs) that detect objects in image frames. Aiming at contributing to the CNN accelerator solutions, the current paper focuses on the design of Field-Programmable Gate Arrays (FPGAs) for CNNs of limited feature space to improve performance, power consumption and resource utilization. The proposed design approach targets the designs that can utilize the logic and memory resources of a single FPGA device and benefit mainly the edge, mobile and on-board satellite (OBC) computing; especially their image-processing- related applications. This work exploits the proposed approach to develop an FPGA accelerator for vessel detection on a Xilinx Virtex 7 XC7VX485T FPGA device (Advanced Micro Devices, Inc, Santa Clara, CA, USA). The resulting architecture operates on RGB images of size 80×80 or sliding windows; it is trained for the “Ships in Satellite Imagery” and by achieving frequency 270 MHz, completing the inference in 0.687 ms and consuming 5 watts, it validates the approach.
Citation: Journal of Imaging
PubDate: 2022-04-15
DOI: 10.3390/jimaging8040114
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 115: Human Tracking in Top-View Fisheye Images:
Analysis of Familiar Similarity Measures via HOG and against Various Color
Spaces
Authors: Hicham Talaoubrid, Marina Vert, Khizar Hayat, Baptiste Magnier
First page: 115
Abstract: The purpose of this paper is to find the best way to track human subjects in fisheye images by considering the most common similarity measures in the function of various color spaces as well as the HOG. To this end, we have relied on videos taken by a fisheye camera wherein multiple human subjects were recorded walking simultaneously, in random directions. Using an existing deep-learning method for the detection of persons in fisheye images, bounding boxes are extracted each containing information related to a single person. Consequently, each bounding box can be described by color features, usually color histograms; with the HOG relying on object shapes and contours. These descriptors do not inform the same features and they need to be evaluated in the context of tracking in top-view fisheye images. With this in perspective, a distance is computed to compare similarities between the detected bounding boxes of two consecutive frames. To do so, we are proposing a rate function (S) in order to compare and evaluate together the six different color spaces and six distances, and with the HOG. This function links inter-distance (i.e., the distance between the images of the same person throughout the frames of the video) with intra-distance (i.e., the distance between images of different people throughout the frames). It enables ascertaining a given feature descriptor (color or HOG) mapped to a corresponding similarity function and hence deciding the most reliable one to compute the similarity or the difference between two segmented persons. All these comparisons lead to some interesting results, as explained in the later part of the article.
Citation: Journal of Imaging
PubDate: 2022-04-16
DOI: 10.3390/jimaging8040115
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 116: A Comparison of Dense and Sparse Optical
Flow Techniques for Low-Resolution Aerial Thermal Imagery
Authors: Tran Xuan Bach Nguyen, Kent Rosser, Javaan Chahl
First page: 116
Abstract: It is necessary to establish the relative performance of established optical flow approaches in airborne scenarios with thermal cameras. This study investigated the performance of a dense optical flow algorithm on 14 bit radiometric images of the ground. While sparse techniques that rely on feature matching techniques perform very well with airborne thermal data in high-contrast thermal conditions, these techniques suffer in low-contrast scenes, where there are fewer detectable and distinct features in the image. On the other hand, some dense optical flow algorithms are highly amenable to parallel processing approaches compared to those that rely on tracking and feature detection. A Long-Wave Infrared (LWIR) micro-sensor and a PX4Flow optical sensor were mounted looking downwards on a drone. We compared the optical flow signals of a representative dense optical flow technique, the Image Interpolation Algorithm (I2A), to the Lucas–Kanade (LK) algorithm in OpenCV and the visible light optical flow results from the PX4Flow in both X and Y displacements. The I2A to LK was found to be generally comparable in performance and better in cold-soaked environments while suffering from the aperture problem in some scenes.
Citation: Journal of Imaging
PubDate: 2022-04-16
DOI: 10.3390/jimaging8040116
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 117: HISFCOS: Half-Inverted Stage Block for
Efficient Object Detection Based on Deep Learning
Authors: Beomyeon Hwang, Sanghun Lee, Seunghyun Lee
First page: 117
Abstract: Recent advances in object detection play a key role in various industrial applications. However, a fully convolutional one-stage detector (FCOS), a conventional object detection method, has low detection accuracy given the calculation cost. Thus, in this study, we propose a half-inverted stage FCOS (HISFCOS) with improved detection accuracy at a computational cost comparable to FCOS based on the proposed half inverted stage (HIS) block. First, FCOS has low detection accuracy owing to low-level information loss. Therefore, an HIS block that minimizes feature loss by extracting spatial and channel information in parallel is proposed. Second, detection accuracy was improved by reconstructing the feature pyramid on the basis of the proposed block and improving the low-level information. Lastly, the improved detection head structure reduced the computational cost and amount compared to the conventional method. Through experiments, the proposed method defined the optimal HISFCOS parameters and evaluated several datasets for fair comparison. The HISFCOS was trained and evaluated using the PASCAL VOC and MSCOCO2017 datasets. Additionally, the average precision (AP) was used as an evaluation index to quantitatively evaluate detection performance. As a result of the experiment, the parameters were increased by 0.5 M compared to the conventional method, but the detection accuracy was improved by 3.0 AP and 1.5 AP in the PASCAL VOC and MSCOCO datasets, respectively. in addition, an ablation study was conducted, and the results for the proposed block and detection head were analyzed.
Citation: Journal of Imaging
PubDate: 2022-04-17
DOI: 10.3390/jimaging8040117
Issue No: Vol. 8, No. 4 (2022)
- J. Imaging, Vol. 8, Pages 50: Considerations on Baseline Generation for
Imaging AI Studies Illustrated on the CT-Based Prediction of Empyema and
Outcome Assessment
Authors: Raphael Sexauer, Bram Stieltjes, Jens Bremerich, Tugba Akinci D’Antonoli, Noemi Schmidt
First page: 50
Abstract: For AI-based classification tasks in computed tomography (CT), a reference standard for evaluating the clinical diagnostic accuracy of individual classes is essential. To enable the implementation of an AI tool in clinical practice, the raw data should be drawn from clinical routine data using state-of-the-art scanners, evaluated in a blinded manner and verified with a reference test. Three hundred and thirty-five consecutive CTs, performed between 1 January 2016 and 1 January 2021 with reported pleural effusion and pathology reports from thoracocentesis or biopsy within 7 days of the CT were retrospectively included. Two radiologists (4 and 10 PGY) blindly assessed the chest CTs for pleural CT features. If needed, consensus was achieved using an experienced radiologist’s opinion (29 PGY). In addition, diagnoses were extracted from written radiological reports. We analyzed these findings for a possible correlation with the following patient outcomes: mortality and median hospital stay. For AI prediction, we used an approach consisting of nnU-Net segmentation, PyRadiomics features and a random forest model. Specificity and sensitivity for CT-based detection of empyema (n = 81 of n = 335 patients) were 90.94 (95%-CI: 86.55–94.05) and 72.84 (95%-CI: 61.63–81.85%) in all effusions, with moderate to almost perfect interrater agreement for all pleural findings associated with empyema (Cohen’s kappa = 0.41–0.82). Highest accuracies were found for pleural enhancement or thickening with 87.02% and 81.49%, respectively. For empyema prediction, AI achieved a specificity and sensitivity of 74.41% (95% CI: 68.50–79.57) and 77.78% (95% CI: 66.91–85.96), respectively. Empyema was associated with a longer hospital stay (median = 20 versus 14 days), and findings consistent with pleural carcinomatosis impacted mortality.
Citation: Journal of Imaging
PubDate: 2022-02-22
DOI: 10.3390/jimaging8030050
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 51: Diagnosis of Vertical Root Fractures in
Endodontically Treated Teeth by Cone-Beam Computed Tomography
Authors: Fumi Mizuhashi, Yuko Watarai, Ichiro Ogura
First page: 51
Abstract: The purpose of this study was to investigate the characteristics and the detection ability of vertical root fractures in endodontically treated teeth by intraoral radiography and cone-beam computed tomography (CBCT). CBCT images of 50 patients with root fractures in endodontically treated teeth were reviewed, and 36 vertical root fractures were taken in this study. The cause of fracture, core construction, kind of teeth, and fracture direction (bucco-lingual and mesio-distal fractures) were investigated. Detection ability of vertical root fractures by intraoral radiography and CBCT was also examined. Statistical analyses concerning the characteristics were performed by χ2 test, and the detection ability was analyzed by cross-tabulation. All of the fractured teeth were nontraumatized teeth. The vertical root fracture occurrence was not differed by core construction. The vertical root fracture number was largest at the premolar teeth (p = 0.005), and the number of the bucco-lingual fracture was larger than the mesio-distal fracture (p = 0.046). Vertical root fractures were detectable using CBCT, while undetectable by intraoral radiography (p < 0.001). Vertical root fractures occurred easily in premolar teeth with bucco-lingual direction, and CBCT is an adequate radiographic method to diagnose vertical root fracture.
Citation: Journal of Imaging
PubDate: 2022-02-23
DOI: 10.3390/jimaging8030051
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 52: Qualitative Comparison of Image Stitching
Algorithms for Multi-Camera Systems in Laparoscopy
Authors: Sylvain Guy, Jean-Loup Haberbusch, Emmanuel Promayon, Stéphane Mancini, Sandrine Voros
First page: 52
Abstract: Multi-camera systems were recently introduced into laparoscopy to increase the narrow field of view of the surgeon. The video streams are stitched together to create a panorama that is easier for the surgeon to comprehend. Multi-camera prototypes for laparoscopy use quite basic algorithms and have only been evaluated on simple laparoscopic scenarios. The more recent state-of-the-art algorithms, mainly designed for the smartphone industry, have not yet been evaluated in laparoscopic conditions. We developed a simulated environment to generate a dataset of multi-view images displaying a wide range of laparoscopic situations, which is adaptable to any multi-camera system. We evaluated classical and state-of-the-art image stitching techniques used in non-medical applications on this dataset, including one unsupervised deep learning approach. We show that classical techniques that use global homography fail to provide a clinically satisfactory rendering and that even the most recent techniques, despite providing high quality panorama images in non-medical situations, may suffer from poor alignment or severe distortions in simulated laparoscopic scenarios. We highlight the main advantages and flaws of each algorithm within a laparoscopic context, identify the main remaining challenges that are specific to laparoscopy, and propose methods to improve these approaches. We provide public access to the simulated environment and dataset.
Citation: Journal of Imaging
PubDate: 2022-02-23
DOI: 10.3390/jimaging8030052
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 53: A Survey of 6D Object Detection Based on 3D
Models for Industrial Applications
Authors: Felix Gorschlüter, Pavel Rojtberg, Thomas Pöllabauer
First page: 53
Abstract: Six-dimensional object detection of rigid objects is a problem especially relevant for quality control and robotic manipulation in industrial contexts. This work is a survey of the state of the art of 6D object detection with these use cases in mind, specifically focusing on algorithms trained only with 3D models or renderings thereof. Our first contribution is a listing of requirements typically encountered in industrial applications. The second contribution is a collection of quantitative evaluation results for several different 6D object detection methods trained with synthetic data and the comparison and analysis thereof. We identify the top methods for individual requirements that industrial applications have for object detectors, but find that a lack of comparable data prevents large-scale comparison over multiple aspects.
Citation: Journal of Imaging
PubDate: 2022-02-24
DOI: 10.3390/jimaging8030053
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 54: Monochrome Camera Conversion: Effect on
Sensitivity for Multispectral Imaging (Ultraviolet, Visible, and Infrared)
Authors: Jonathan Crowther
First page: 54
Abstract: Conversion of standard cameras to enable them to capture images in the ultraviolet (UV) and infrared (IR) spectral regions has applications ranging from purely artistic to science and research. Taking the modification of the camera a step further and removing the color filter array (CFA) results in the formation of a monochrome camera. The spectral sensitivities of a range of cameras with different sensors which were converted to monochrome were measured and compared with standard multispectral camera conversions, with an emphasis on their behavior from the UV through to the IR regions.
Citation: Journal of Imaging
PubDate: 2022-02-25
DOI: 10.3390/jimaging8030054
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 55: Kidney Tumor Semantic Segmentation Using
Deep Learning: A Survey of State-of-the-Art
Authors: Abubaker Abdelrahman, Serestina Viriri
First page: 55
Abstract: Cure rates for kidney cancer vary according to stage and grade; hence, accurate diagnostic procedures for early detection and diagnosis are crucial. Some difficulties with manual segmentation have necessitated the use of deep learning models to assist clinicians in effectively recognizing and segmenting tumors. Deep learning (DL), particularly convolutional neural networks, has produced outstanding success in classifying and segmenting images. Simultaneously, researchers in the field of medical image segmentation employ DL approaches to solve problems such as tumor segmentation, cell segmentation, and organ segmentation. Segmentation of tumors semantically is critical in radiation and therapeutic practice. This article discusses current advances in kidney tumor segmentation systems based on DL. We discuss the various types of medical images and segmentation techniques and the assessment criteria for segmentation outcomes in kidney tumor segmentation, highlighting their building blocks and various strategies.
Citation: Journal of Imaging
PubDate: 2022-02-25
DOI: 10.3390/jimaging8030055
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 56: Principal Component Analysis versus
Subject’s Residual Profile Analysis for Neuroinflammation
Investigation in Parkinson Patients: A PET Brain Imaging Study
Authors: Rostom Mabrouk
First page: 56
Abstract: Dysfunction of neurons in the central nervous system is the primary pathological feature of Parkinson’s disease (PD). Despite different triggering, emerging evidence indicates that neuroinflammation revealed through microglia activation is critical for PD. Moreover, recent investigations sought a potential relationship between Lrrk2 genetic mutation and microglia activation. In this paper, neuroinflammation in sporadic PD, Lrrk2-PD and unaffected Lrrk2 mutation carriers were investigated. The principal component analysis (PCA) and the subject’s residual profile (SRP) techniques were performed on multiple groups and regions of interest in 22 brain-regions. The 11C-PBR28 binding profiles were compared in four genotypes depending on groups, i.e., HC, sPD, Lrrk2-PD and UC, using the PCA and SPR scores. The genotype effect was found as a principal feature of group-dependent 11C-PBR28 binding, and preliminary evidence of a MAB-Lrrk2 mutation interaction in manifest Parkinson’s and subjects at risk was found.
Citation: Journal of Imaging
PubDate: 2022-02-25
DOI: 10.3390/jimaging8030056
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 57: PRNU-Based Video Source Attribution: Which
Frames Are You Using'
Authors: Pasquale Ferrara, Massimo Iuliani, Alessandro Piva
First page: 57
Abstract: Photo Response Non-Uniformity (PRNU) is reputed the most successful trace to identify the source of a digital video. However, its effectiveness is mainly limited by compression and the effect of recently introduced electronic image stabilization on several devices. In the last decade, several approaches were proposed to overcome both these issues, mainly by selecting those video frames which are considered more informative. However, the two problems were always treated separately, and the combined effect of compression and digital stabilization was never considered. This separated analysis makes it hard to understand if achieved conclusions still stand for digitally stabilized videos and if those choices represent a general optimum strategy to perform video source attribution. In this paper, we explore whether an optimum strategy exists in selecting frames based on their type and their positions within the groups of pictures. We, therefore, systematically analyze the PRNU contribute provided by all frames belonging to either digitally stabilized or not stabilized videos. Results on the VISION dataset come up with some insights into optimizing video source attribution in different use cases.
Citation: Journal of Imaging
PubDate: 2022-02-25
DOI: 10.3390/jimaging8030057
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 58: Scanning Hyperspectral Imaging for In Situ
Biogeochemical Analysis of Lake Sediment Cores: Review of Recent
Developments
Authors: Paul D. Zander, Giulia Wienhues, Martin Grosjean
First page: 58
Abstract: Hyperspectral imaging (HSI) in situ core scanning has emerged as a valuable and novel tool for rapid and non-destructive biogeochemical analysis of lake sediment cores. Variations in sediment composition can be assessed directly from fresh sediment surfaces at ultra-high-resolution (40–300 μm measurement resolution) based on spectral profiles of light reflected from sediments in visible, near infrared, and short-wave infrared wavelengths (400–2500 nm). Here, we review recent methodological developments in this new and growing field of research, as well as applications of this technique for paleoclimate and paleoenvironmental studies. Hyperspectral imaging of sediment cores has been demonstrated to effectively track variations in sedimentary pigments, organic matter, grain size, minerogenic components, and other sedimentary features. These biogeochemical variables record information about past climatic conditions, paleoproductivity, past hypolimnetic anoxia, aeolian input, volcanic eruptions, earthquake and flood frequencies, and other variables of environmental relevance. HSI has been applied to study seasonal and inter-annual environmental variability as recorded in individual varves (annually laminated sediments) or to study sedimentary records covering long glacial–interglacial time-scales (>10,000 years).
Citation: Journal of Imaging
PubDate: 2022-02-25
DOI: 10.3390/jimaging8030058
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 59: Glossiness Index of Objects in Halftone
Color Images Based on Structure and Appearance Distortion
Authors: Donghui Li, Midori Tanaka, Takahiko Horiuchi
First page: 59
Abstract: This paper proposes an objective glossiness index for objects in halftone color images. In the proposed index, we consider the characteristics of the human visual system (HVS) and associate the image’s structure distortion and statistical information. According to the difference in the number of strategies adopted by the HVS in judging the difference between images, it is divided into single and multi-strategy modeling. In this study, we advocate multiple strategies to determine glossy or non-glossy quality. We assumed that HVS used different visual mechanisms to evaluate glossy and non-glossy objects. For non-glossy images, the image structure dominated, so the HVS tried to use structural information to judge distortion (a strategy based on structural distortion detection). For glossy images, the glossy appearance dominated; thus, the HVS tried to search for the glossiness difference (an appearance-based strategy). Herein, we present an index for glossiness assessment that attempts to explicitly model the structural dissimilarity and appearance distortion. We used the contrast sensitivity function to account for the mechanism of halftone images when viewed by the human eye. We estimated the structure distortion for the first strategy by using local luminance and contrast masking; meanwhile, local statistics changing in the spatial frequency components for skewness and standard deviation were used to estimate the appearance distortion for the second strategy. Experimental results showed that these two mixed-distortion measurement strategies performed well in consistency with the subjective ratings of glossiness in halftone color images.
Citation: Journal of Imaging
PubDate: 2022-02-27
DOI: 10.3390/jimaging8030059
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 60: Hierarchical Fusion Using Subsets of
Multi-Features for Historical Arabic Manuscript Dating
Authors: Kalthoum Adam, Somaya Al-Maadeed, Younes Akbari
First page: 60
Abstract: Automatic dating tools for historical documents can greatly assist paleographers and save them time and effort. This paper describes a novel method for estimating the date of historical Arabic documents that employs hierarchical fusions of multiple features. A set of traditional features and features extracted by a residual network (ResNet) are fused in a hierarchical approach using joint sparse representation. To address noise during the fusion process, a new approach based on subsets of multiple features is being considered. Following that, supervised and unsupervised classifiers are used for classification. We show that using hierarchical fusion based on subsets of multiple features in the KERTAS dataset can produce promising results and significantly improve the results.
Citation: Journal of Imaging
PubDate: 2022-03-01
DOI: 10.3390/jimaging8030060
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 61: Iterative Multiple Bounding-Box Refinements
for Visual Tracking
Authors: Giorgio Cruciata, Liliana Lo Presti, Marco La Cascia
First page: 61
Abstract: Single-object visual tracking aims at locating a target in each video frame by predicting the bounding box of the object. Recent approaches have adopted iterative procedures to gradually refine the bounding box and locate the target in the image. In such approaches, the deep model takes as input the image patch corresponding to the currently estimated target bounding box, and provides as output the probability associated with each of the possible bounding box refinements, generally defined as a discrete set of linear transformations of the bounding box center and size. At each iteration, only one transformation is applied, and supervised training of the model may introduce an inherent ambiguity by giving importance priority to some transformations over the others. This paper proposes a novel formulation of the problem of selecting the bounding box refinement. It introduces the concept of non-conflicting transformations and allows applying multiple refinements to the target bounding box at each iteration without introducing ambiguities during learning of the model parameters. Empirical results demonstrate that the proposed approach improves the iterative single refinement in terms of accuracy and precision of the tracking results.
Citation: Journal of Imaging
PubDate: 2022-03-03
DOI: 10.3390/jimaging8030061
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 62: A Real-Time Method for Time-to-Collision
Estimation from Aerial Images
Authors: Daniel Tøttrup, Stinus Lykke Skovgaard, Jonas le Fevre Sejersen, Rui Pimentel de Figueiredo
First page: 62
Abstract: When large vessels such as container ships are approaching their destination port, they are required by law to have a maritime pilot on board responsible for safely navigating the vessel to its desired location. The maritime pilot has extensive knowledge of the local area and how currents and tides affect the vessel’s navigation. In this work, we present a novel end-to-end solution for estimating time-to-collision time-to-collision (TTC) between moving objects (i.e., vessels), using real-time image streams from aerial drones in dynamic maritime environments. Our method relies on deep features, which are learned using realistic simulation data, for reliable and robust object detection, segmentation, and tracking. Furthermore, our method uses rotated bounding box representations, which are computed by taking advantage of pixel-level object segmentation for enhanced TTC estimation accuracy. We present collision estimates in an intuitive manner, as collision arrows that gradually change its color to red to indicate an imminent collision. A set of experiments in a realistic shipyard simulation environment demonstrate that our method can accurately, robustly, and quickly predict TTC between dynamic objects seen from a top-view, with a mean error and a standard deviation of 0.358 and 0.114 s, respectively, in a worst case scenario.
Citation: Journal of Imaging
PubDate: 2022-03-03
DOI: 10.3390/jimaging8030062
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 63: An Exploration of Pathologies of Multilevel
Principal Components Analysis in Statistical Models of Shape
Authors: Damian J. J. Farnell
First page: 63
Abstract: 3D facial surface imaging is a useful tool in dentistry and in terms of diagnostics and treatment planning. Between-group PCA (bgPCA) is a method that has been used to analyse shapes in biological morphometrics, although various “pathologies” of bgPCA have recently been proposed. Monte Carlo (MC) simulated datasets were created here in order to explore “pathologies” of multilevel PCA (mPCA), where mPCA with two levels is equivalent to bgPCA. The first set of MC experiments involved 300 uncorrelated normally distributed variables, whereas the second set of MC experiments used correlated multivariate MC data describing 3D facial shape. We confirmed results of numerical experiments from other researchers that indicated that bgPCA (and so also mPCA) can give a false impression of strong differences in component scores between groups when there is none in reality. These spurious differences in component scores via mPCA decreased significantly as the sample sizes per group were increased. Eigenvalues via mPCA were also found to be strongly affected by imbalances in sample sizes per group, although this problem was removed by using weighted forms of covariance matrices suggested by the maximum likelihood solution of the two-level model. However, this did not solve problems of spurious differences between groups in these simulations, which was driven by very small sample sizes in one group. As a “rule of thumb” only, all of our experiments indicate that reasonable results are obtained when sample sizes per group in all groups are at least equal to the number of variables. Interestingly, the sum of all eigenvalues over both levels via mPCA scaled approximately linearly with the inverse of the sample size per group in all experiments. Finally, between-group variation was added explicitly to the MC data generation model in two experiments considered here. Results for the sum of all eigenvalues via mPCA predicted the asymptotic amount for the total amount of variance correctly in this case, whereas standard “single-level” PCA underestimated this quantity.
Citation: Journal of Imaging
PubDate: 2022-03-04
DOI: 10.3390/jimaging8030063
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 64: Rethinking Weight Decay for Efficient Neural
Network Pruning
Authors: Hugo Tessier, Vincent Gripon, Mathieu Léonardon, Matthieu Arzel, Thomas Hannagan, David Bertrand
First page: 64
Abstract: Introduced in the late 1980s for generalization purposes, pruning has now become a staple for compressing deep neural networks. Despite many innovations in recent decades, pruning approaches still face core issues that hinder their performance or scalability. Drawing inspiration from early work in the field, and especially the use of weight decay to achieve sparsity, we introduce Selective Weight Decay (SWD), which carries out efficient, continuous pruning throughout training. Our approach, theoretically grounded on Lagrangian smoothing, is versatile and can be applied to multiple tasks, networks, and pruning structures. We show that SWD compares favorably to state-of-the-art approaches, in terms of performance-to-parameters ratio, on the CIFAR-10, Cora, and ImageNet ILSVRC2012 datasets.
Citation: Journal of Imaging
PubDate: 2022-03-04
DOI: 10.3390/jimaging8030064
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 65: Review of Machine Learning in Lung
Ultrasound in COVID-19 Pandemic
Authors: Jing Wang, Xiaofeng Yang, Boran Zhou, James J. Sohn, Jun Zhou, Jesse T. Jacob, Kristin A. Higgins, Jeffrey D. Bradley, Tian Liu
First page: 65
Abstract: Ultrasound imaging of the lung has played an important role in managing patients with COVID-19–associated pneumonia and acute respiratory distress syndrome (ARDS). During the COVID-19 pandemic, lung ultrasound (LUS) or point-of-care ultrasound (POCUS) has been a popular diagnostic tool due to its unique imaging capability and logistical advantages over chest X-ray and CT. Pneumonia/ARDS is associated with the sonographic appearances of pleural line irregularities and B-line artefacts, which are caused by interstitial thickening and inflammation, and increase in number with severity. Artificial intelligence (AI), particularly machine learning, is increasingly used as a critical tool that assists clinicians in LUS image reading and COVID-19 decision making. We conducted a systematic review from academic databases (PubMed and Google Scholar) and preprints on arXiv or TechRxiv of the state-of-the-art machine learning technologies for LUS images in COVID-19 diagnosis. Openly accessible LUS datasets are listed. Various machine learning architectures have been employed to evaluate LUS and showed high performance. This paper will summarize the current development of AI for COVID-19 management and the outlook for emerging trends of combining AI-based LUS with robotics, telehealth, and other techniques.
Citation: Journal of Imaging
PubDate: 2022-03-05
DOI: 10.3390/jimaging8030065
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 66: An Empirical Evaluation of Convolutional
Networks for Malaria Diagnosis
Authors: Andrea Loddo, Corrado Fadda, Cecilia Di Ruberto
First page: 66
Abstract: Malaria is a globally widespread disease caused by parasitic protozoa transmitted to humans by infected female mosquitoes of Anopheles. It is caused in humans only by the parasite Plasmodium, further classified into four different species. Identifying malaria parasites is possible by analysing digital microscopic blood smears, which is tedious, time-consuming and error prone. So, automation of the process has assumed great importance as it helps the laborious manual process of review and diagnosis. This work focuses on deep learning-based models, by comparing off-the-shelf architectures for classifying healthy and parasite-affected cells, by investigating the four-class classification on the Plasmodium falciparum stages of life and, finally, by evaluating the robustness of the models with cross-dataset experiments on two different datasets. The main contributions to the research in this field can be resumed as follows: (i) comparing off-the-shelf architectures in the task of classifying healthy and parasite-affected cells, (ii) investigating the four-class classification on the P. falciparum stages of life and (iii) evaluating the robustness of the models with cross-dataset experiments. Eleven well-known convolutional neural networks on two public datasets have been exploited. The results show that the networks have great accuracy in binary classification, even though they lack few samples per class. Moreover, the cross-dataset experiments exhibit the need for some further regulations. In particular, ResNet-18 achieved up to 97.68% accuracy in the binary classification, while DenseNet-201 reached 99.40% accuracy on the multiclass classification. The cross-dataset experiments exhibit the limitations of deep learning approaches in such a scenario, even though combining the two datasets permitted DenseNet-201 to reach 97.45% accuracy. Naturally, this needs further investigation to improve the robustness. In general, DenseNet-201 seems to offer the most stable and robust performance, offering as a crucial candidate to further developments and modifications. Moreover, the mobile-oriented architectures showed promising and satisfactory performance in the classification of malaria parasites. The obtained results enable extensive improvements, specifically oriented to the application of object detectors for type and stage of life recognition, even in mobile environments.
Citation: Journal of Imaging
PubDate: 2022-03-07
DOI: 10.3390/jimaging8030066
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 67: Perceptually Optimal Color Representation of
Fully Polarimetric SAR Imagery
Authors: Georgia Koukiou
First page: 67
Abstract: The four bands of fully polarimetric SAR data convey scattering characteristics of the Earth’s background, but perceptually are not very easy for an observer to use. In this work, the four different channels of fully polarimetric SAR images, namely HH, HV, VH, and VV, are combined so that a color image of the Earth’s background is derived that is perceptually excellent for the human eye and at the same time provides accurate information regarding the scattering mechanisms in each pixel. Most of the elementary scattering mechanisms are related to specific color and land cover types. The innovative nature of the proposed approach is due to the two different consecutive coloring procedures. The first one is a fusion procedure that moves all the information contained in the four polarimetric channels into three derived RGB bands. This is achieved by means of Cholesky decomposition and brings to the RGB output the correlation properties of a natural color image. The second procedure moves the color information of the RGB image to the CIELab color space, which is perceptually uniform. The color information is then evenly distributed by means of color equalization in the CIELab color space. After that, the inverse procedure to obtain the final RGB image is performed. These two procedures bring the PolSAR information regarding the scattering mechanisms on the Earth’s surface onto a meaningful color image, the appearance of which is close to Google Earth maps. Simultaneously, they give better color correspondence to various land cover types compared with existing SAR color representation methods.
Citation: Journal of Imaging
PubDate: 2022-03-07
DOI: 10.3390/jimaging8030067
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 68: Photo2Video: Semantic-Aware Deep
Learning-Based Video Generation from Still Content
Authors: Paula Viana, Maria Teresa Andrade, Pedro Carvalho, Luis Vilaça, Inês N. Teixeira, Tiago Costa, Pieter Jonker
First page: 68
Abstract: Applying machine learning (ML), and especially deep learning, to understand visual content is becoming common practice in many application areas. However, little attention has been given to its use within the multimedia creative domain. It is true that ML is already popular for content creation, but the progress achieved so far addresses essentially textual content or the identification and selection of specific types of content. A wealth of possibilities are yet to be explored by bringing the use of ML into the multimedia creative process, allowing the knowledge inferred by the former to influence automatically how new multimedia content is created. The work presented in this article provides contributions in three distinct ways towards this goal: firstly, it proposes a methodology to re-train popular neural network models in identifying new thematic concepts in static visual content and attaching meaningful annotations to the detected regions of interest; secondly, it presents varied visual digital effects and corresponding tools that can be automatically called upon to apply such effects in a previously analyzed photo; thirdly, it defines a complete automated creative workflow, from the acquisition of a photograph and corresponding contextual data, through the ML region-based annotation, to the automatic application of digital effects and generation of a semantically aware multimedia story driven by the previously derived situational and visual contextual data. Additionally, it presents a variant of this automated workflow by offering to the user the possibility of manipulating the automatic annotations in an assisted manner. The final aim is to transform a static digital photo into a short video clip, taking into account the information acquired. The final result strongly contrasts with current standard approaches of creating random movements, by implementing an intelligent content- and context-aware video.
Citation: Journal of Imaging
PubDate: 2022-03-10
DOI: 10.3390/jimaging8030068
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 69: Seamless Copy–Move Replication in
Digital Images
Authors: Tanzeela Qazi, Mushtaq Ali, Khizar Hayat, Baptiste Magnier
First page: 69
Abstract: The importance and relevance of digital-image forensics has attracted researchers to establish different techniques for creating and detecting forgeries. The core category in passive image forgery is copy–move image forgery that affects the originality of image by applying a different transformation. In this paper, a frequency-domain image-manipulation method is presented. The method exploits the localized nature of discrete wavelet transform (DWT) to attain the region of the host image to be manipulated. Both patch and host image are subjected to DWT at the same level l to obtain 3l+1 sub-bands, and each sub-band of the patch is pasted to the identified region in the corresponding sub-band of the host image. Resulting manipulated host sub-bands are then subjected to inverse DWT to obtain the final manipulated host image. The proposed method shows good resistance against detection by two frequency-domain forgery detection methods from the literature. The purpose of this research work is to create a forgery and highlight the need to produce forgery detection methods that are robust against malicious copy–move forgery.
Citation: Journal of Imaging
PubDate: 2022-03-10
DOI: 10.3390/jimaging8030069
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 70: A Novel Deep-Learning-Based Framework for
the Classification of Cardiac Arrhythmia
Authors: Sonain Jamil, MuhibUr Rahman
First page: 70
Abstract: Cardiovascular diseases (CVDs) are the primary cause of death. Every year, many people die due to heart attacks. The electrocardiogram (ECG) signal plays a vital role in diagnosing CVDs. ECG signals provide us with information about the heartbeat. ECGs can detect cardiac arrhythmia. In this article, a novel deep-learning-based approach is proposed to classify ECG signals as normal and into sixteen arrhythmia classes. The ECG signal is preprocessed and converted into a 2D signal using continuous wavelet transform (CWT). The time–frequency domain representation of the CWT is given to the deep convolutional neural network (D-CNN) with an attention block to extract the spatial features vector (SFV). The attention block is proposed to capture global features. For dimensionality reduction in SFV, a novel clump of features (CoF) framework is proposed. The k-fold cross-validation is applied to obtain the reduced feature vector (RFV), and the RFV is given to the classifier to classify the arrhythmia class. The proposed framework achieves 99.84% accuracy with 100% sensitivity and 99.6% specificity. The proposed algorithm outperforms the state-of-the-art accuracy, F1-score, and sensitivity techniques.
Citation: Journal of Imaging
PubDate: 2022-03-10
DOI: 10.3390/jimaging8030070
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 71: Multi-Modality Microscopy Image Style
Augmentation for Nuclei Segmentation
Authors: Ye Liu, Sophia J. Wagner, Tingying Peng
First page: 71
Abstract: Annotating microscopy images for nuclei segmentation by medical experts is laborious and time-consuming. To leverage the few existing annotations, also across multiple modalities, we propose a novel microscopy-style augmentation technique based on a generative adversarial network (GAN). Unlike other style transfer methods, it can not only deal with different cell assay types and lighting conditions, but also with different imaging modalities, such as bright-field and fluorescence microscopy. Using disentangled representations for content and style, we can preserve the structure of the original image while altering its style during augmentation. We evaluate our data augmentation on the 2018 Data Science Bowl dataset consisting of various cell assays, lighting conditions, and imaging modalities. With our style augmentation, the segmentation accuracy of the two top-ranked Mask R-CNN-based nuclei segmentation algorithms in the competition increases significantly. Thus, our augmentation technique renders the downstream task more robust to the test data heterogeneity and helps counteract class imbalance without resampling of minority classes.
Citation: Journal of Imaging
PubDate: 2022-03-11
DOI: 10.3390/jimaging8030071
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 72: Comparison of 2D Optical Imaging and 3D
Microtomography Shape Measurements of a Coastal Bioclastic Calcareous Sand
Authors: Ryan D. Beemer, Linzhu Li, Antonio Leonti, Jeremy Shaw, Joana Fonseca, Iren Valova, Magued Iskander, Cynthia H. Pilskaln
First page: 72
Abstract: This article compares measurements of particle shape parameters from three-dimensional (3D) X-ray micro-computed tomography (μCT) and two-dimensional (2D) dynamic image analysis (DIA) from the optical microscopy of a coastal bioclastic calcareous sand from Western Australia. This biogenic sand from a high energy environment consists largely of the shells and tests of marine organisms and their clasts. A significant difference was observed between the two imaging techniques for measurements of aspect ratio, convexity, and sphericity. Measured values of aspect ratio, sphericity, and convexity are larger in 2D than in 3D. Correlation analysis indicates that sphericity is correlated with convexity in both 2D and 3D. These results are attributed to inherent limitations of DIA when applied to platy sand grains and to the shape being, in part, dependent on the biology of the grain rather than a purely random clastic process, like typical siliceous sands. The statistical data has also been fitted to Johnson Bounded Distribution for the ease of future use. Overall, this research demonstrates the need for high-quality 3D microscopy when conducting a micromechanical analysis of biogenic calcareous sands.
Citation: Journal of Imaging
PubDate: 2022-03-14
DOI: 10.3390/jimaging8030072
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 73: Fabrication of a Human Skin Mockup with a
Multilayered Concentration Map of Pigment Components Using a UV Printer
Authors: Kazuki Nagasawa, Shoji Yamamoto, Wataru Arai, Kunio Hakkaku, Chawan Koopipat, Keita Hirai, Norimichi Tsumura
First page: 73
Abstract: In this paper, we propose a pipeline that reproduces human skin mockups using a UV printer by obtaining the spatial concentration map of pigments from an RGB image of human skin. The pigment concentration distributions were obtained by a separating method of skin pigment components with independent component analysis from the skin image. This method can extract the concentration of melanin and hemoglobin components, which are the main pigments that make up skin tone. Based on this concentration, we developed a procedure to reproduce a skin mockup with a multi-layered structure that is determined by mapping the absorbance of melanin and hemoglobin to CMYK (Cyan, Magenta, Yellow, Black) subtractive color mixing. In our proposed method, the multi-layered structure with different pigments in each layer contributes greatly to the accurate reproduction of skin tones. We use a UV printer because the printer is capable of layered fabrication by using UV-curable inks. As the result, subjective evaluation showed that the artificial skin reproduced by our method has a more skin-like appearance than that produced using conventional printing.
Citation: Journal of Imaging
PubDate: 2022-03-15
DOI: 10.3390/jimaging8030073
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 74: Investigation of Nonlinear Optical
Properties of Quantum Dots Deposited onto a Sample Glass Using
Time-Resolved Inline Digital Holography
Authors: Andrey Belashov, Igor Shevkunov, Ekaterina Kolesova, Anna Orlova, Sergei Putilin, Andrei Veniaminov, Chau-Jern Cheng, Nikolay Petrov
First page: 74
Abstract: We report on the application of time-resolved inline digital holography in the study of the nonlinear optical properties of quantum dots deposited onto sample glass. The Fresnel diffraction patterns of the probe pulse due to noncollinear degenerate phase modulation induced by a femtosecond pump pulse were extracted from the set of inline digital holograms and analyzed. The absolute values of the nonlinear refractive index of both the sample glass substrate and the deposited layer of quantum dots were evaluated using the proposed technique. To characterize the inhomogeneous distribution of the samples’ nonlinear optical properties, we proposed plotting an optical nonlinearity map calculated as a local standard deviation of the diffraction pattern intensities induced by noncollinear degenerate phase modulation.
Citation: Journal of Imaging
PubDate: 2022-03-16
DOI: 10.3390/jimaging8030074
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 75: Visualization of Inferior Alveolar and
Lingual Nerve Pathology by 3D Double-Echo Steady-State MRI: Two Case
Reports with Literature Review
Authors: Adib Al-Haj Husain, Daphne Schönegg, Silvio Valdec, Bernd Stadlinger, Thomas Gander, Harald Essig, Marco Piccirelli, Sebastian Winklhofer
First page: 75
Abstract: Injury to the peripheral branches of the trigeminal nerve, particularly the lingual nerve (LN) and the inferior alveolar nerve (IAN), is a rare but serious complication that can occur during oral and maxillofacial surgery. Mandibular third molar surgery, one of the most common surgical procedures in dentistry, is most often associated with such a nerve injury. Proper preoperative radiologic assessment is hence key to avoiding neurosensory dysfunction. In addition to the well-established conventional X-ray-based imaging modalities, such as panoramic radiography and cone-beam computed tomography, radiation-free magnetic resonance imaging (MRI) with the recently introduced black-bone MRI sequences offers the possibility to simultaneously visualize osseous structures and neural tissue in the oral cavity with high spatial resolution and excellent soft-tissue contrast. Fortunately, most LN and IAN injuries recover spontaneously within six months. However, permanent damage may cause significant loss of quality of life for affected patients. Therefore, therapy should be initiated early in indicated cases, despite the inconsistency in the literature regarding the therapeutic time window. In this report, we present the visualization of two cases of nerve pathology using 3D double-echo steady-state MRI and evaluate evidence-based decision-making for iatrogenic nerve injury regarding a wait-and-see strategy, conservative drug treatment, or surgical re-intervention.
Citation: Journal of Imaging
PubDate: 2022-03-17
DOI: 10.3390/jimaging8030075
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 76: Microsaccades, Drifts, Hopf Bundle and
Neurogeometry
Authors: Dmitri Alekseevsky
First page: 76
Abstract: The first part of the paper contains a short review of the image processing in early vision is static, when the eyes and the stimulus are stable, and in dynamics, when the eyes participate in fixation eye movements. In the second part, we give an interpretation of Donders’ and Listing’s law in terms of the Hopf fibration of the 3-sphere over the 2-sphere. In particular, it is shown that the configuration space of the eye ball (when the head is fixed) is the 2-dimensional hemisphere SL+, called Listing hemisphere, and saccades are described as geodesic segments of SL+ with respect to the standard round metric. We study fixation eye movements (drift and microsaccades) in terms of this model and discuss the role of fixation eye movements in vision. A model of fixation eye movements is proposed that gives an explanation of presaccadic shift of receptive fields.
Citation: Journal of Imaging
PubDate: 2022-03-17
DOI: 10.3390/jimaging8030076
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 77: Metal Artifact Reduction in Spectral X-ray
CT Using Spectral Deep Learning
Authors: Matteo Busi, Christian Kehl, Jeppe R. Frisvad, Ulrik L. Olsen
First page: 77
Abstract: Spectral X-ray computed tomography (SCT) is an emerging method for non-destructive imaging of the inner structure of materials. Compared with the conventional X-ray CT, this technique provides spectral photon energy resolution in a finite number of energy channels, adding a new dimension to the reconstructed volumes and images. While this mitigates energy-dependent distortions such as beam hardening, metal artifacts due to photon starvation effects are still present, especially for low-energy channels where the attenuation coefficients are higher. We present a correction method for metal artifact reduction in SCT that is based on spectral deep learning. The correction efficiently reduces streaking artifacts in all the energy channels measured. We show that the additional information in the energy domain provides relevance for restoring the quality of low-energy reconstruction affected by metal artifacts. The correction method is parameter free and only takes around 15 ms per energy channel, satisfying near-real time requirement of industrial scanners.
Citation: Journal of Imaging
PubDate: 2022-03-17
DOI: 10.3390/jimaging8030077
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 78: The Capabilities of Dedicated Small
Satellite Infrared Missions for the Quantitative Characterization of
Wildfires
Authors: Winfried Halle, Christian Fischer, Dieter Oertel, Boris Zhukov
First page: 78
Abstract: The main objective of this paper was to demonstrate the capability of dedicated small satellite infrared sensors with cooled quantum detectors, such as those successfully utilized three times in Germany’s pioneering BIRD and FireBIRD small satellite infrared missions, in the quantitative characterization of high-temperature events such as wildfires. The Bi-spectral Infrared Detection (BIRD) mission was launched in October 2001. The space segment of FireBIRD consists of the small satellites Technologie Erprobungs-Träger (TET-1), launched in July 2012, and Bi-spectral InfraRed Optical System (BIROS), launched in June 2016. These missions also significantly improved the scientific understanding of space-borne fire monitoring with regard to climate change. The selected examples compare the evaluation of quantitative characteristics using data from BIRD or FireBIRD and from the operational polar orbiting IR sensor systems MODIS, SLSTR and VIIRS. Data from the geostationary satellite “Himawari-8” were compared with FireBIRD data, obtained simultaneously. The geostationary Meteosat Third Generation-Imager (MTG-I) is foreseen to be launched at the end of 2022. In its application to fire, the MTG-I’s Flexible Combined Imager (FCI) will provide related spectral bands at ground sampling distances (GSD) of 3.8 µm and 10.5 µm at the sub-satellite point (SSP) of 1 km or 2 km, depending on the used FCI imaging mode. BIRD wildfire data, obtained over Africa and Portugal, were used to simulate the fire detection and monitoring capability of MTG-I/FCI. A new quality of fire monitoring is predicted, if the 1 km resolution wildfire data from MTG-1/FCI are used together with the co-located fire data acquired by the polar orbiting Visible Infrared Imaging Radiometer Suite (VIIRS), and possibly prospective FireBIRD-type compact IR sensors flying on several small satellites in various low Earth orbits (LEOs).
Citation: Journal of Imaging
PubDate: 2022-03-18
DOI: 10.3390/jimaging8030078
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 79: Comparing Desktop vs. Mobile Interaction for
the Creation of Pervasive Augmented Reality Experiences
Authors: Tiago Madeira, Bernardo Marques, Pedro Neves, Paulo Dias, Beatriz Sousa Santos
First page: 79
Abstract: This paper presents an evaluation and comparison of interaction methods for the configuration and visualization of pervasive Augmented Reality (AR) experiences using two different platforms: desktop and mobile. AR experiences consist of the enhancement of real-world environments by superimposing additional layers of information, real-time interaction, and accurate 3D registration of virtual and real objects. Pervasive AR extends this concept through experiences that are continuous in space, being aware of and responsive to the user’s context and pose. Currently, the time and technical expertise required to create such applications are the main reasons preventing its widespread use. As such, authoring tools which facilitate the development and configuration of pervasive AR experiences have become progressively more relevant. Their operation often involves the navigation of the real-world scene and the use of the AR equipment itself to add the augmented information within the environment. The proposed experimental tool makes use of 3D scans from physical environments to provide a reconstructed digital replica of such spaces for a desktop-based method, and to enable positional tracking for a mobile-based one. While the desktop platform represents a non-immersive setting, the mobile one provides continuous AR in the physical environment. Both versions can be used to place virtual content and ultimately configure an AR experience. The authoring capabilities of the different platforms were compared by conducting a user study focused on evaluating their usability. Although the AR interface was generally considered more intuitive, the desktop platform shows promise in several aspects, such as remote configuration, lower required effort, and overall better scalability.
Citation: Journal of Imaging
PubDate: 2022-03-18
DOI: 10.3390/jimaging8030079
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 80: Neutron Tomography Studies of Two
Lamprophyre Dike Samples: 3D Data Analysis for the Characterization of
Rock Fabric
Authors: Ivan Zel, Bekhzodjon Abdurakhimov, Sergey Kichanov, Olga Lis, Elmira Myrzabekova, Denis Kozlenko, Mannab Tashmetov, Khalbay Ishbaev, Kuatbay Kosbergenov
First page: 80
Abstract: The rock fabric of two lamprophyre dike samples from the Koy-Tash granitoid intrusion (Koy-Tash, Jizzakh region, Uzbekistan) has been studied, using the neutron tomography method. We have performed virtual segmentation of the reconstructed 3D model of the tabular igneous intrusion and the corresponding determination of dike margins orientation. Spatial distributions of inclusions in the dike volume, as well as further analysis of size distributions and shape orientations of inclusions, have been obtained. The observed shape preferred orientations of inclusions as evidence of the magma flow-related fabric. The obtained structural data have been discussed in the frame of the models of rigid particle motion and the straining of vesicles in a moving viscous fluid.
Citation: Journal of Imaging
PubDate: 2022-03-19
DOI: 10.3390/jimaging8030080
Issue No: Vol. 8, No. 3 (2022)
- J. Imaging, Vol. 8, Pages 81: A New Approach in Detectability of
Microcalcifications in the Placenta during Pregnancy Using Textural
Features and K-Nearest Neighbors Algorithm
Authors: Mihaela Miron, Simona Moldovanu, Bogdan Ioan Ștefănescu, Mihai Culea, Sorin Marius Pavel, Anisia Luiza Culea-Florescu
First page: 81
Abstract: (1) Background: Ultrasonography is the main method used during pregnancy to assess the fetal growth, amniotic fluid, umbilical cord and placenta. The placenta’s structure suffers dynamic modifications throughout the whole pregnancy and many of these changes, in which placental microcalcifications are by far the most prominent, are related to the process of aging and maturation and have no effect on fetal wellbeing. However, when placental microcalcifications are noticed earlier during pregnancy, they could suggest a major placental dysfunction with serious consequences for the fetus and mother. For better detectability of microcalcifications, we propose a new approach based on improving the clarity of details and the analysis of the placental structure using first and second order statistics, and fractal dimension. (2) Methods: The methodology is based on four stages: (i) cropping the region of interest and preprocessing steps; (ii) feature extraction, first order—standard deviation (SD), skewness (SK) and kurtosis (KR)—and second order—contrast (C), homogeneity (H), correlation (CR), energy (E) and entropy (EN)—are computed from a gray level co-occurrence matrix (GLCM) and fractal dimension (FD); (iii) statistical analysis (t-test); (iv) classification with the K-Nearest Neighbors algorithm (K-NN algorithm) and performance comparison with results from the support vector machine algorithm (SVM algorithm). (3) Results: Experimental results obtained from real clinical data show an improvement in the detectability and visibility of placental microcalcifications.
Citation: Journal of Imaging
PubDate: 2022-03-19
DOI: 10.3390/jimaging8030081
Issue No: Vol. 8, No. 3 (2022)