Publications

Explore our research publications: papers, articles, and conference proceedings from AImageLab.

Tip: type @ to pick an author and # to pick a keyword.

RI-PIENO - Revised and Improved Petrol-Filling Itinerary Estimation aNd Optimization

Authors: Savarese, Marco; De Blasi, Antonio; Zaccagnino, Carmine; Salici, Giacomo; Cascianelli, Silvia; Vezzani, Roberto; Grazia, Carlo Augusto

Efficient energy provisioning is a fundamental requirement for modern transportation systems, making refueling path optimization a critical challenge. Existing solutions … (Read full abstract)

Efficient energy provisioning is a fundamental requirement for modern transportation systems, making refueling path optimization a critical challenge. Existing solutions often focus either on inter-vehicle communication or intravehicle monitoring, leveraging Intelligent Transportation Systems, Digital Twins, and Software-Defined Internet of Vehicles with Cloud/Fog/Edge infrastructures. However, integrated frameworks that adapt dynamically to driver mobility patterns are still underdeveloped. Building on our previous PIENO framework, we present RI-PIENO (Revised and Improved Petrolfilling Itinerary Estimation aNd Optimization), a system that combines intra-vehicle sensor data with external geospatial and fuel price information, processed via IoT-enabled Cloud/Fog services. RI-PIENO models refueling as a dynamic, time-evolving directed acyclic graph that reflects both habitual daily trips and real-time vehicular inputs, transforming the system from a static recommendation tool into a continuously adaptive decision engine. We validate RI-PIENO in a daily-commute use case through realistic multi-driver, multi-week simulations, showing that it achieves significant cost savings and more efficient routing compared to previous approaches. The framework is designed to leverage emerging roadside infrastructure and V2X communication, supporting scalable deployment within next-generation IoT and vehicular networking ecosystems.

2025 Relazione in Atti di Convegno

Root colonization pattern of Pseudomonas spp. strains: a key step in the biocontrol of soilborne pathogens in hops.

Authors: Bellameche, F.; Modica, F.; Cortiello, M.; Costi, E.; Riccioni, C.; De Marchis, F.; Rubini, A.; Belfiori, B.; Bellucci, M.; Brilli, L.; Sberveglieri, V.; Lovino, M.; Núñez-Carmona, E.; Giovanardi, D.

Published in: JOURNAL OF PLANT PATHOLOGY

The control of soil-borne diseases in hops, such as Verticillium wilt remains challenging due to the limited effectiveness of fungicides, … (Read full abstract)

The control of soil-borne diseases in hops, such as Verticillium wilt remains challenging due to the limited effectiveness of fungicides, the perennial nature of hop cultivation, and the long-term persistence of the pathogens in the soil. Microbial biocontrol agents (mBCAs) with plant growth-promoting (PGP) and antagonistic effects offer a sustainable ecofriendly alternative for hops protection. Two Pseudomonas spp. strains from the UniMORE microbial collection were selected for this study based on their strong antagonistic activity against Verticillium spp. and multiple plant growth-promoting (PGP) traits. Rhizospheric and endophytic colonization capacities of the strains DLS1929 and DLS2318 were evaluated in hop plants (cv. Cascade) under controlled conditions at seven- and fourteen-days post-inoculation (DPI). Both bacterial strains were rhizosphere and endorhiza competent, with slight differences in their abundances. The highest cell density was observed at 7 DPI for the strain DLS2318, reaching log10 6.39 CFU g−1 root fresh weight in the rhizosphere and log10 4.75 CFU g−1 root fresh weight in the endorhiza; at 14 DPI, colonization results were in line with the previous assessment. Confocal laser scanning microscopy visualization of both eGFP-tagged Pseudomonas spp. strains confirmed their rhizosphere competence in hop. Additionally, root colonization by these bacteria enhanced the photosynthetic capacity in hop leaves, supporting their potential as a PGP agents observed in vitro. Successful root colonization and PGP effects are key prerequisites for an effective biocontrol of soilborne pathogens. Further studies are required to assess the consistent efficacy in the field of these beneficial mBCA candidates. This research was funded by the Italian Ministry of University and Research (MUR), under the European Union funding – Next Generation EU - PRIN- 2022, (prot. 2022M3HR45) project: “IoHOP: Quality valorization of the Italian hop based on a multi-approach strategy”.

2025 Abstract in Rivista

San Vitale Challenge: Automatic Reconstruction of Ancient Colored Glass Windows

Authors: Di Domenico, N.; Borghi, G.; Franco, A.; Boschetti, M.; Giacomini, F.; Barzaghi, S.; Ferucci, S.; Zambruno, S.; Mularoni, L.; Gao, Q.; Che, C.; Li, G.; Zu, Y.; Hao, J.; Zhang, J.; Ducz, A.; Gego, L.; Imeri, K.; Nemkin, V.; Rakhmatillaev, A.; Szatmari, S.; Rowan, W.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

The sixth-century Basilica of San Vitale in Ravenna, Italy, once featured intricate circular colored glass windows that illuminated its interior. … (Read full abstract)

The sixth-century Basilica of San Vitale in Ravenna, Italy, once featured intricate circular colored glass windows that illuminated its interior. Although these windows are now lost, several fragments were recovered during recent restorations. Unfortunately, reconstructing the original glass windows from these fragments is extremely complex and time-consuming, requiring the use of specialized expertise. Therefore, the development of automatic reconstruction techniques based on Artificial Intelligence is particularly important and challenging, due to, for instance, the presence of uniform color, damaged glass edges, and many fragment outliers. In this direction, the San Vitale Challenge was organized to gather the best methods and algorithms, as described and summarized in this paper. The challenge, split into several sub-tracks of increasing difficulty and realism, received the submission of several solutions, ranging from more classical computer vision algorithms to purely deep learning-based approaches, whose results are quantitatively evaluated and compared. In the last part of the paper, directions for future developments of such systems are discussed.

2025 Relazione in Atti di Convegno

Sanctuaria-Gaze: A Multimodal Egocentric Dataset for Human Attention Analysis in Religious Sites

Authors: Cartella, Giuseppe; Cuculo, Vittorio; Cornia, Marcella; Papasidero, Marco; Ruozzi, Federico; Cucchiara, Rita

Published in: ACM JOURNAL ON COMPUTING AND CULTURAL HERITAGE

We introduce Sanctuaria-Gaze, a multimodal dataset featuring egocentric recordings from 40 visits to four architecturally and culturally significant sanctuaries in … (Read full abstract)

We introduce Sanctuaria-Gaze, a multimodal dataset featuring egocentric recordings from 40 visits to four architecturally and culturally significant sanctuaries in Northern Italy. Collected using wearable devices with integrated eye trackers, the dataset offers RGB videos synchronized with streams of gaze coordinates, head motion, and environmental point cloud, resulting in over four hours of recordings. Along with the dataset, we provide a framework for automatic detection and analysis of Areas of Interest (AOIs). This framework fills a critical gap by offering an open-source, flexible tool for gaze-based research that adapts to dynamic settings without requiring manual intervention. Our study analyzes human visual attention to sacred, architectural, and cultural objects, providing insights into how visitors engage with these elements and how their background influences their interactions. By releasing both the dataset and the analysis framework, Sanctuaria-Gaze aims to advance interdisciplinary research on gaze behavior, human-computer interaction, and visual attention in real-world environments. Code and dataset are available at https://github.com/aimagelab/Sanctuaria-Gaze.

2025 Articolo su rivista

Segmenting Maxillofacial Structures in CBCT Volumes

Authors: Bolelli, Federico; Marchesini, Kevin; Van Nistelrooij, Niels; Lumetti, Luca; Pipoli, Vittorio; Ficarra, Elisa; Vinayahalingam, Shankeeth; Grana, Costantino

Cone-beam computed tomography (CBCT) is a standard imaging modality in orofacial and dental practices, providing essential 3D volumetric imaging of … (Read full abstract)

Cone-beam computed tomography (CBCT) is a standard imaging modality in orofacial and dental practices, providing essential 3D volumetric imaging of anatomical structures, including jawbones, teeth, sinuses, and neurovascular canals. Accurately segmenting these structures is fundamental to numerous clinical applications, such as surgical planning and implant placement. However, manual segmentation of CBCT scans is time-intensive and requires expert input, creating a demand for automated solutions through deep learning. Effective development of such algorithms relies on access to large, well-annotated datasets, yet current datasets are often privately stored or limited in scope and considered structures, especially concerning 3D annotations. This paper proposes ToothFairy2, a comprehensive, publicly accessible CBCT dataset with voxel-level 3D annotations of 42 distinct classes corresponding to maxillofacial structures. We validate the dataset by benchmarking state-of-the-art neural network models, including convolutional, transformer-based, and hybrid Mamba-based architectures, to evaluate segmentation performance across complex anatomical regions. Our work also explores adaptations to the nnU-Net framework to optimize multi-class segmentation for maxillofacial anatomy. The proposed dataset provides a fundamental resource for advancing maxillofacial segmentation and supports future research in automated 3D image analysis in digital dentistry.

2025 Relazione in Atti di Convegno

Segmenting the Inferior Alveolar Canal in CBCTs Volumes: the ToothFairy Challenge

Authors: Bolelli, Federico; Lumetti, Luca; Vinayahalingam, Shankeeth; Di Bartolomeo, Mattia; Pellacani, Arrigo; Marchesini, Kevin; Van Nistelrooij, Niels; Van Lierop, Pieter; Xi, Tong; Liu, Yusheng; Xin, Rui; Yang, Tao; Wang, Lisheng; Wang, Haoshen; Xu, Chenfan; Cui, Zhiming; Wodzinski, Marek Michal; Müller, Henning; Kirchhoff, Yannick; R., Rokuss Maximilian; Maier-Hein, Klaus; Han, Jaehwan; Kim, Wan; Ahn, Hong-Gi; Szczepański, Tomasz; Grzeszczyk Michal, K.; Korzeniowski, Przemyslaw; Caselles Ballester Vicent amd Paolo Burgos-Artizzu, Xavier; Prados Carrasco, Ferran; Berge’, Stefaan; Van Ginneken, Bram; Anesi, Alexandre; Re, ; Grana, Costantino

Published in: IEEE TRANSACTIONS ON MEDICAL IMAGING

In recent years, several algorithms have been developed for the segmentation of the Inferior Alveolar Canal (IAC) in Cone-Beam Computed … (Read full abstract)

In recent years, several algorithms have been developed for the segmentation of the Inferior Alveolar Canal (IAC) in Cone-Beam Computed Tomography (CBCT) scans. However, the availability of public datasets in this domain is limited, resulting in a lack of comparative evaluation studies on a common benchmark. To address this scientific gap and encourage deep learning research in the field, the ToothFairy challenge was organized within the MICCAI 2023 conference. In this context, a public dataset was released to also serve as a benchmark for future research. The dataset comprises 443 CBCT scans, with voxel-level annotations of the IAC available for 153 of them, making it the largest publicly available dataset of its kind. The participants of the challenge were tasked with developing an algorithm to accurately identify the IAC using the 2D and 3D-annotated scans. This paper presents the details of the challenge and the contributions made by the most promising methods proposed by the participants. It represents the first comprehensive comparative evaluation of IAC segmentation methods on a common benchmark dataset, providing insights into the current state-of-the-art algorithms and outlining future research directions. Furthermore, to ensure reproducibility and promote future developments, an open-source repository that collects the implementations of the best submissions was released.

2025 Articolo su rivista

Semantic Residual Prompts for Continual Learning

Authors: Menabue, M.; Frascaroli, E.; Boschini, M.; Sangineto, E.; Bonicelli, L.; Porrello, A.; Calderara, S.

Published in: LECTURE NOTES IN COMPUTER SCIENCE

Prompt-tuning methods for Continual Learning (CL) freeze a large pre-trained model and train a few parameter vectors termed prompts. Most … (Read full abstract)

Prompt-tuning methods for Continual Learning (CL) freeze a large pre-trained model and train a few parameter vectors termed prompts. Most of these methods organize these vectors in a pool of key-value pairs and use the input image as query to retrieve the prompts (values). However, as keys are learned while tasks progress, the prompting selection strategy is itself subject to catastrophic forgetting, an issue often overlooked by existing approaches. For instance, prompts introduced to accommodate new tasks might end up interfering with previously learned prompts. To make the selection strategy more stable, we leverage a foundation model (CLIP) to select our prompts within a two-level adaptation mechanism. Specifically, the first level leverages a standard textual prompt pool for the CLIP textual encoder, leading to stable class prototypes. The second level, instead, uses these prototypes along with the query image as keys to index a second pool. The retrieved prompts serve to adapt a pre-trained ViT, granting plasticity. In doing so, we also propose a novel residual mechanism to transfer CLIP semantics to the ViT layers. Through extensive analysis on established CL benchmarks, we show that our method significantly outperforms both state-of-the-art CL approaches and the zero-shot CLIP test. Notably, our findings hold true even for datasets with a substantial domain gap w.r.t. the pre-training knowledge of the backbone model, as showcased by experiments on satellite imagery and medical datasets. The codebase is available at https://github.com/aimagelab/mammoth.

2025 Relazione in Atti di Convegno

Semantically Conditioned Prompts for Visual Recognition under Missing Modality Scenarios

Authors: Pipoli, Vittorio; Bolelli, Federico; Sarto, Sara; Cornia, Marcella; Baraldi, Lorenzo; Grana, Costantino; Cucchiara, Rita; Ficarra, Elisa

Published in: IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION

This paper tackles the domain of multimodal prompting for visual recognition, specifically when dealing with missing modalities through multimodal Transformers. … (Read full abstract)

This paper tackles the domain of multimodal prompting for visual recognition, specifically when dealing with missing modalities through multimodal Transformers. It presents two main contributions: (i) we introduce a novel prompt learning module which is designed to produce sample-specific prompts and (ii) we show that modality-agnostic prompts can effectively adjust to diverse missing modality scenarios. Our model, termed SCP, exploits the semantic representation of available modalities to query a learnable memory bank, which allows the generation of prompts based on the semantics of the input. Notably, SCP distinguishes itself from existing methodologies for its capacity of self-adjusting to both the missing modality scenario and the semantic context of the input, without prior knowledge about the specific missing modality and the number of modalities. Through extensive experiments, we show the effectiveness of the proposed prompt learning framework and demonstrate enhanced performance and robustness across a spectrum of missing modality cases.

2025 Relazione in Atti di Convegno

State-of-the-art Review and Benchmarking of Barcode Localization Methods

Authors: Vezzali, Enrico; Bolelli, Federico; Santi, Stefano; Grana, Costantino

Published in: ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE

Barcodes, despite their long history, remain an essential technology in supply chain management. In addition, barcodes have found extensive use … (Read full abstract)

Barcodes, despite their long history, remain an essential technology in supply chain management. In addition, barcodes have found extensive use in industrial engineering, particularly in warehouse automation, component tracking, and robot guidance. To detect a barcode in an image, multiple algorithms have been proposed in the literature, with a significant increase of interest in the topic since the rise of deep learning. However, research in the field suffers from many limitations, including the scarcity of public datasets and code implementations which hinders the reproducibility and reliability of published results. For this reason, we developed ``BarBeR'' (Barcode Benchmark Repository), a benchmark designed for testing and comparing barcode detection algorithms. This benchmark includes the code implementation of various detection algorithms for barcodes, along with a suite of useful metrics. Among the supported localization methods, there are multiple deep-learning detection models, that will be used to assess the recent contributions of Artificial Intelligence to this field. In addition, we provide a large, annotated dataset of 8748 barcode images, combining multiple public barcode datasets with standardized annotation formats for both detection and segmentation tasks. Finally, we provide a thorough summary of the history and literature on barcode localization and share the results obtained from running the benchmark on our dataset, offering valuable insights into the performance of different algorithms when applied to real-world problems.

2025 Articolo su rivista

TakuNet: an Energy-Efficient CNN for Real-Time Inference on Embedded UAV systems in Emergency Response Scenarios

Authors: Rossi, Daniel; Borghi, Guido; Vezzani, Roberto

Designing efficient neural networks for embedded devices is a critical challenge, particularly in applications requiring real-time performance, such as aerial … (Read full abstract)

Designing efficient neural networks for embedded devices is a critical challenge, particularly in applications requiring real-time performance, such as aerial imaging with drones and UAVs for emergency responses. In this work, we introduce TakuNet, a novel light-weight architecture which employs techniques such as depth-wise convolutions and an early downsampling stem to reduce computational complexity while maintaining high accuracy. It leverages dense connections for fast convergence during training and uses 16-bit floating-point precision for optimization on embedded hardware accelerators. Experimental evaluation on two public datasets shows that TakuNet achieves near-state-of-the-art accuracy in classifying aerial images of emergency situations, despite its minimal parameter count. Real-world tests on embedded devices, namely Jetson Orin Nano and Raspberry Pi, confirm TakuNet's efficiency, achieving more than 650 fps on the 15W Jetson board, making it suitable for real-time AI processing on resource-constrained platforms and advancing the applicability of drones in emergency scenarios. The code and implementation details are publicly released.

2025 Relazione in Atti di Convegno

Page 10 of 106 • Total publications: 1059