| | SLO | ENG | Cookies and privacy

Bigger font | Smaller font

Search the digital library catalog Help

Query: search in
search in
search in
search in
* old and bologna study programme

Options:
  Reset


1 - 10 / 14
First pagePrevious page12Next pageLast page
1.
2.
An end-to-end framework for extracting observable cues of depression from diary recordings
Izidor Mlakar, Umut Arioz, Urška Smrke, Nejc Plohl, Valentino Šafran, Matej Rojc, 2024, original scientific article

Abstract: Because of the prevalence of depression, its often-chronic course, relapse and associated disability, early detection and non-intrusive monitoring is a crucial tool for timely diagnosis and treatment, remission of depression and prevention of relapse. In this way, its impact on quality of life and well-being can be limited. Current attempts to use artificial intelligence for the early classification of depression are mostly data-driven and thus non-transparent and lack effective means to deal with uncertainties. Therefore, in this paper, we propose an end-to-end framework for extracting observable depression cues from diary recordings. Furthermore, we also explore its feasibility for automatic detection of depression symptoms using observable behavioural cues. The proposed end-to-end framework for extracting depression was used to evaluate 28 video recordings from the Symptom Media dataset and 27 recordings from the DAIC-WOZ dataset. We compared the presence of the extracted features between recordings of individuals with and without a depressive disorder. We identified several cues consistent with previous studies in terms of their differentiation between individuals with and without depressive disorder across both datasets among language (i.e., use of negatively valanced words, use of first-person singular pronouns, some features of language complexity, explicit mentions of treatment for depression), speech (i.e., monotonous speech, voiced speech and pauses, speaking rate, low articulation rate), and facial cues (i.e., rotational energy of head movements). The nature/context of the discourse, the impact of other disorders and physical/psychological stress, and the quality and resolution of the recordings all play an important role in matching the digital features to the relevant background. In this way, the work presented in this paper provides a novel approach to extracting a wide range of cues relevant to the classification of depression and opens up new opportunities for further research.
Keywords: digital biomarkers of depression, facial cues, speech cues, language cues, deep learning, end-to-end pipeline, artificial intelligence
Published in DKUM: 17.01.2025; Views: 0; Downloads: 3
.pdf Full text (2,34 MB)

3.
4.
5.
DigiPig : First developments of an automated monitoring system for body, head and tail detection in intensive pig farming
Marko Ocepek, Anja Žnidar, Miha Lavrič, Dejan Škorjanc, Inger Lise Andersen, 2022, original scientific article

Abstract: The goal of this study was to develop an automated monitoring system for the detection of pigs’ bodies, heads and tails. The aim in the first part of the study was to recognize individual pigs (in lying and standing positions) in groups and their body parts (head/ears, and tail) by using machine learning algorithms (feature pyramid network). In the second part of the study, the goal was to improve the detection of tail posture (tail straight and curled) during activity (standing/moving around) by the use of neural network analysis (YOLOv4). Our dataset (n = 583 images, 7579 pig posture) was annotated in Labelbox from 2D video recordings of groups (n = 12–15) of weaned pigs. The model recognized each individual pig’s body with a precision of 96% related to threshold intersection over union (IoU), whilst the precision for tails was 77% and for heads this was 66%, thereby already achieving human-level precision. The precision of pig detection in groups was the highest, while head and tail detection precision were lower. As the first study was relatively time-consuming, in the second part of the study, we performed a YOLOv4 neural network analysis using 30 annotated images of our dataset for detecting straight and curled tails. With this model, we were able to recognize tail postures with a high level of precision (90%).
Keywords: pig, welfare, image processing, object detection, deep learning, smart farming
Published in DKUM: 11.07.2024; Views: 87; Downloads: 9
.pdf Full text (48,11 MB)
This document has many files! More...

6.
Cross-Hole GPR for Soil Moisture Estimation Using Deep Learning
Blaž Pongrac, Dušan Gleich, Marko Malajner, Andrej Sarjaš, 2023, original scientific article

Abstract: This paper presents the design of a high-voltage pulse-based radar and a supervised data processing method for soil moisture estimation. The goal of this research was to design a pulse-based radar to detect changes in soil moisture using a cross-hole approach. The pulse-based radar with three transmitting antennas was placed into a 12 m deep hole, and a receiver with three receive antennas was placed into a different hole separated by 100 m from the transmitter. The pulse generator was based on a Marx generator with an LC filter, and for the receiver, the high-frequency data acquisition card was used, which can acquire signals using 3 Gigabytes per second. Used borehole antennas were designed to operate in the wide frequency band to ensure signal propagation through the soil. A deep regression convolutional network is proposed in this paper to estimate volumetric soil moisture using time-sampled signals. A regression convolutional network is extended to three dimensions to model changes in wave propagation between the transmitted and received signals. The training dataset was acquired during the period of 73 days of acquisition between two boreholes separated by 100 m. The soil moisture measurements were acquired at three points 25 m apart to provide ground truth data. Additionally, water was poured into several specially prepared boreholes between transmitter and receiver antennas to acquire additional dataset for training, validation, and testing of convolutional neural networks. Experimental results showed that the proposed system is able to detect changes in the volumetric soil moisture using Tx and Rx antennas.
Keywords: ground penetrating radar, cross-hole, L-band, deep learning, convolutional neural network, soil moisture estimation
Published in DKUM: 03.04.2024; Views: 448; Downloads: 26
.pdf Full text (3,22 MB)
This document has many files! More...

7.
A graph pointer network-based multi-objective deep reinforcement learning algorithm for solving the traveling salesman problem
Jeewaka Perera, Shih-Hsi Liu, Marjan Mernik, Matej Črepinšek, Miha Ravber, 2023, original scientific article

Abstract: Traveling Salesman Problems (TSPs) have been a long-lasting interesting challenge to researchers in different areas. The difficulty of such problems scales up further when multiple objectives are considered concurrently. Plenty of work in evolutionary algorithms has been introduced to solve multi-objective TSPs with promising results, and the work in deep learning and reinforcement learning has been surging. This paper introduces a multi-objective deep graph pointer network-based reinforcement learning (MODGRL) algorithm for multi-objective TSPs. The MODGRL improves an earlier multi-objective deep reinforcement learning algorithm, called DRL-MOA, by utilizing a graph pointer network to learn the graphical structures of TSPs. Such improvements allow MODGRL to be trained on a small-scale TSP, but can find optimal solutions for large scale TSPs. NSGA-II, MOEA/D and SPEA2 are selected to compare with MODGRL and DRL-MOA. Hypervolume, spread and coverage over Pareto front (CPF) quality indicators were selected to assess the algorithms’ performance. In terms of the hypervolume indicator that represents the convergence and diversity of Pareto-frontiers, MODGRL outperformed all the competitors on the three well-known benchmark problems. Such findings proved that MODGRL, with the improved graph pointer network, indeed performed better, measured by the hypervolume indicator, than DRL-MOA and the three other evolutionary algorithms. MODGRL and DRL-MOA were comparable in the leading group, measured by the spread indicator. Although MODGRL performed better than DRL-MOA, both of them were just average regarding the evenness and diversity measured by the CPF indicator. Such findings remind that different performance indicators measure Pareto-frontiers from different perspectives. Choosing a well-accepted and suitable performance indicator to one’s experimental design is very critical, and may affect the conclusions. Three evolutionary algorithms were also experimented on with extra iterations, to validate whether extra iterations affected the performance. The results show that NSGA-II and SPEA2 were greatly improved measured by the Spread and CPF indicators. Such findings raise fairness concerns on algorithm comparisons using different fixed stopping criteria for different algorithms, which appeared in the DRL-MOA work and many others. Through these lessons, we concluded that MODGRL indeed performed better than DRL-MOA in terms of hypervolumne, and we also urge researchers on fair experimental designs and comparisons, in order to derive scientifically sound conclusions.
Keywords: multi-objective optimization, traveling salesman problems, deep reinforcement learning
Published in DKUM: 28.03.2024; Views: 184; Downloads: 28
.pdf Full text (7,89 MB)
This document has many files! More...

8.
A VAN-Based Multi-Scale Cross-Attention Mechanism for Skin Lesion Segmentation Network
Shuang Liu, Zeng Zhuang, Yanfeng Zheng, Simon Kolmanič, 2023, original scientific article

Abstract: With the rise of deep learning technology, the field of medical image segmentation has undergone rapid development. In recent years, convolutional neural networks (CNNs) have brought many achievements and become the consensus in medical image segmentation tasks. Although many neural networks based on U-shaped structures and methods, such as skip connections have achieved excellent results in medical image segmentation tasks, the properties of convolutional operations limit their ability to effectively learn local and global features. To address this problem, the Transformer from the field of natural language processing (NLP) was introduced to the image segmentation field. Various Transformer-based networks have shown significant performance advantages over mainstream neural networks in different visual tasks, demonstrating the huge potential of Transformers in the field of image segmentation. However, Transformers were originally designed for NLP and ignore the multidimensional nature of images. In the process of operation, they may destroy the 2D structure of the image and cannot effectively capture low-level features. Therefore, we propose a new multi-scale cross-attention method called M-VAN Unet, which is designed based on the Visual Attention Network (VAN) and can effectively learn local and global features. We propose two attention mechanisms, namely MSC-Attention and LKA-Cross-Attention, for capturing low-level features and promoting global information interaction. MSC-Attention is designed for multi-scale channel attention, while LKA-Cross-Attention is a cross-attention mechanism based on the large kernel attention (LKA). Extensive experiments show that our method outperforms current mainstream methods in evaluation metrics such as Dice coefficient and Hausdorff 95 coefficient.
Keywords: CNNs, deep learning, medical image processing, NLP, semantic segmentation
Published in DKUM: 14.03.2024; Views: 496; Downloads: 311
.pdf Full text (1,46 MB)
This document has many files! More...

9.
Despeckling of SAR Images Using Residual Twin CNN and Multi-Resolution Attention Mechanism
Blaž Pongrac, Dušan Gleich, 2023, original scientific article

Abstract: The despeckling of synthetic aperture radar images using two different convolutional neural network architectures is presented in this paper. The first method presents a novel Siamese convolutional neural network with a dilated convolutional network in each branch. Recently, attention mechanisms have been introduced to convolutional networks to better model and recognize features. Therefore, we propose a novel design for a convolutional neural network using an attention mechanism for an encoder–decoder-type network. The framework consists of a multiscale spatial attention network to improve the modeling of semantic information at different spatial levels and an additional attention mechanism to optimize feature propagation. Both proposed methods are different in design but they provide comparable despeckling results in subjective and objective measurements in terms of correlated speckle noise. The experimental results are evaluated on both synthetically generated speckled images and real SAR images. The methods proposed in this paper are able to despeckle SAR images and preserve SAR features.
Keywords: synthetic aperture radar, speckle, speckle suppression, despeckling, deep learning, convolutional neural network
Published in DKUM: 21.02.2024; Views: 280; Downloads: 27
.pdf Full text (13,70 MB)
This document has many files! More...

10.
UAV Thermal Imaging for Unexploded Ordnance Detection by Using Deep Learning
Milan Bajić, Jr., Božidar Potočnik, 2023, original scientific article

Abstract: A few promising solutions for thermal imaging Unexploded Ordnance (UXO) detection were proposed after the start of the military conflict in Ukraine in 2014. At the same time, most of the landmine clearance protocols and practices are based on old, 20th-century technologies. More than 60 countries worldwide are still affected by explosive remnants of war, and new areas are contaminated almost every day. To date, no automated solutions exist for surface UXO detection by using thermal imaging. One of the reasons is also that there are no publicly available data. This research bridges both gaps by introducing an automated UXO detection method, and by publishing thermal imaging data. During a project in Bosnia and Herzegovina in 2019, an organisation, Norwegian People's Aid, collected data about unexploded ordnances and made them available for this research. Thermal images with a size of 720 x 480 pixels were collected by using an Unmanned Aerial Vehicle at a height of 3 m, thus achieving a very small Ground Sampling Distance (GSD). One of the goals of our research was also to verify if the explosive war remnants' detection accuracy could be improved further by using Convolutional Neural Networks (CNN). We have experimented with various existing modern CNN architectures for object identification, whereat the YOLOv5 model was selected as the most promising for retraining. An eleven-class object detection problem was solved primarily in this study. Our data were annotated semi-manually. Five versions of the YOLOv5 model, fine-tuned with a grid-search, were trained end-to-end on randomly selected 640 training and 80 validation images from our dataset. The trained models were verified on the remaining 88 images from our dataset. Objects from each of the eleven classes were identified with more than 90% probability, whereat the Mean Average Precision (mAP) at a 0.5 threshold was 99.5%, and the mAP at thresholds from 0.5 to 0.95 was 87.0% up to 90.5%, depending on the model's complexity. Our results are comparable to the state-of-the-art, whereat these object detection methods have been tested on other similar small datasets with thermal images. Our study is one of the few in the field of Automated UXO detection by using thermal images, and the first that solves the problem of identifying more than one class of objects. On the other hand, publicly available thermal images with a relatively small GSD will enable and stimulate the development of new detection algorithms, where our method and results can serve as a baseline. Only really accurate automatic UXO detection solutions will help to solve one of the least explored worldwide life-threatening problems.
Keywords: unmanned aerial vehicle, unexploded ordnance, thermal imaging, UXOTi_NPA dataset, convolutional neural networks, deep learning
Published in DKUM: 12.02.2024; Views: 389; Downloads: 26
.pdf Full text (16,94 MB)
This document has many files! More...

Search done in 0.26 sec.
Back to top
Logos of partners University of Maribor University of Ljubljana University of Primorska University of Nova Gorica