Mid Sweden University

miun.sePublications
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 200) Show all publications
Lycksam, A., O'Nils, M. & Qureshi, F. Z. (2025). A prognostic framework for rotating machines considering multi-component fault scenarios. IEEE Access, 13, 91682-91692
Open this publication in new window or tab >>A prognostic framework for rotating machines considering multi-component fault scenarios
2025 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 13, p. 91682-91692Article in journal (Refereed) Published
Abstract [en]

The importance of preventing failures in rotating machines has led to extensive research on diagnosis and prognosis methods based on vibration data. To achieve a high generalizability for these solutions, they need to handle high discrepancies between the data used for developing the method and the data from the target machine, a lack of historical events from the target machine and multi-component fault scenarios. Currently, state-of-the-art research has successfully found solutions targeting the first two challenges. However, these almost exclusively focus on single-component fault scenarios, meaning it is unclear how a method should be constructed considering a multi-component system. Therefore, this study constructs a framework for multi-component fault scenarios called Rotating Machinery Prognostic Framework (RoMaP) that leverages the advancements in anomaly detection, fault diagnosis and RUL prediction based on vibration data. To evaluate the potential of RoMaP, an instance based on state-of-the-art research was implemented and compared to other methods based on alternative frameworks. The results show that the instance of RoMaP had the best performance across many scenarios expected in an industrial environment, suggesting that it is a suitable approach for monitoring the health of rotating machines. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
bearing, fault diagnosis, predictive maintenance, remaining useful life prediction, rotating machinery
National Category
Other Civil Engineering
Identifiers
urn:nbn:se:miun:diva-54589 (URN)10.1109/ACCESS.2025.3572582 (DOI)001499574200021 ()2-s2.0-105006649662 (Scopus ID)
Available from: 2025-06-10 Created: 2025-06-10 Last updated: 2025-09-25
Seyed Jalaleddin, M., Gallo, V., Shallari, I. & O'Nils, M. (2025). Addressing Contextual Factors in ArUco Marker-Based Distance Estimation: A Machine Learning Approach. In: 2025 IEEE International Instrumentation and Measurement Technology Conference (I2MTC): . Paper presented at 2025 IEEE International Instrumentation and Measurement Technology Conference (I2MTC). IEEE conference proceedings
Open this publication in new window or tab >>Addressing Contextual Factors in ArUco Marker-Based Distance Estimation: A Machine Learning Approach
2025 (English)In: 2025 IEEE International Instrumentation and Measurement Technology Conference (I2MTC), IEEE conference proceedings, 2025Conference paper, Published paper (Refereed)
Abstract [en]

Accurate distance estimation is crucial for applications such as robotics and autonomous systems, where reliable measurements are needed for navigation and interaction with the environment. ArUco markers are a robust solution for distance estimation, offering precise measurements based on their geometric properties. In this paper, we, first, systematically investigate the impact of various contextual factors in an indoor environment, such as angle of observation and illumination, on the systematic error and uncertainty of distance measurements. Our experiments show that illumination (exposure) significantly influence the performance of distance estimation systems, introducing notable systematic errors in real-world settings. Second, we propose a machine learning-based approach, using a neural network, to address the challenge posed by these factors and improve the systematic error of distance estimation. Our results in dynamic outdoor environment, demonstrate the effectiveness of this approach, significantly enhancing the systematic error of distance under dynamic environmental conditions. 

Place, publisher, year, edition, pages
IEEE conference proceedings, 2025
Keywords
ArUco marker, Contextual factor, Machine learning, Neural networks, Systemic Error
National Category
Computer Systems
Identifiers
urn:nbn:se:miun:diva-55270 (URN)10.1109/I2MTC62753.2025.11079149 (DOI)2-s2.0-105012189961 (Scopus ID)9798331505004 (ISBN)
Conference
2025 IEEE International Instrumentation and Measurement Technology Conference (I2MTC)
Available from: 2025-08-12 Created: 2025-08-12 Last updated: 2025-09-25Bibliographically approved
Zakaryapour Sayyad, F., Shallari, I., Seyed Jalaleddin, M., O'Nils, M. & Qureshi, F. Z. (2025). AdVision: An efficient and effective deep learning based advertisement detector for printed media. MACHINE LEARNING WITH APPLICATIONS, 21, Article ID 100686.
Open this publication in new window or tab >>AdVision: An efficient and effective deep learning based advertisement detector for printed media
Show others...
2025 (English)In: MACHINE LEARNING WITH APPLICATIONS, ISSN 2666-8270, Vol. 21, article id 100686Article in journal (Refereed) Published
Abstract [en]

Automated advertisement detection in newspapers is a challenging task due to the diversity in print layouts, formats, and design styles. This task has critical applications in media monitoring, content analysis, and advertising analytics. To address these challenges, we introduce AdVision, a deep-learning-based solution that treats advertisements as unique visual objects. We provide a comparative study of various detection architectures, including one-stage, two-stage, and transformer-based detectors, to identify the most effective approach for detecting advertisements. Our results are validated through extensive experiments conducted under different conditions and metrics. Newspapers from four different countries - Denmark, Norway, Sweden, and the UK - were selected to demonstrate the variety of languages and print formats. Additionally, we conduct a cross-analysis to show how training on one language can generalize to another. To enhance the explainability of our results, we employ GradCAM++ (Chattopadhay et al., 2018) heatmaps. Our experiments demonstrate that the YOLOv8 model achieves superior performance, balancing high precision and recall with minimal inference latency, making it particularly suitable for high-throughput advertisement detection.

Place, publisher, year, edition, pages
Elsevier BV, 2025
Keywords
Cross-linguistic advertisement detection, Deep learning, Newspaper image analysis, Object detection, YOLO, Model generalization
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:miun:diva-55187 (URN)10.1016/j.mlwa.2025.100686 (DOI)001519891500001 ()
Available from: 2025-07-22 Created: 2025-07-22 Last updated: 2025-09-25
Gatner, O., Shallari, I., O'Nils, M., Imran, M., Ciani, L. & Patrizi, G. (2025). Cross-Section-Based Method for LiDAR Dataset Generation with Multipath-Resilient Ground Truth. IEEE Transactions on Instrumentation and Measurement, 74, 1-11
Open this publication in new window or tab >>Cross-Section-Based Method for LiDAR Dataset Generation with Multipath-Resilient Ground Truth
Show others...
2025 (English)In: IEEE Transactions on Instrumentation and Measurement, ISSN 0018-9456, E-ISSN 1557-9662, Vol. 74, p. 1-11Article in journal (Refereed) Published
Abstract [en]

The demand for high-quality LiDAR datasets is increasing as LiDAR technology is being used in various applications, including autonomous vehicles, robotics, and 3D mapping. However, generating accurate ground truth data for LiDAR datasets remains a challenge due to issues like multipath interference (MPI) and other disturbances. The method for generating high-quality ground truth data for LiDAR applications based on cross-section is introduced is this paper. The key concept is based on precisely registering a digital twin, CAD-based which is later converted to a mesh, with LiDAR depth images. By utilizing selective point usage in its cross-sections, the method demonstrates greater robustness to MPI compared to standard approaches. The technique is evaluated using a dataset of LiDAR-acquired point clouds of a living room scene. The results show that the proposed technique achieves significantly better accuracy in affected regions than standard Iterative Closest Point (ICP) based methods. Additionally, the paper proposes a new set of metrics for evaluating the quality of ground truth data, which is more robust to MPI than standard metrics such as RMSE and Chamfer Distance. The proposed technique is a valuable tool for generating large-scale, high-quality datasets for LiDAR applications. Lastly we compared our dataset with latest available datasets. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Denoising, Ground Truth, Lidar, Metrology, Multipath Propagation, Point Clouds, Three-dimensional Reconstruction, Time Of Flight Measurements, 3d Reconstruction, Computer Aided Design, Digital Twin, Iterative Methods, Optical Radar, Three Dimensional Computer Graphics, De-noising, Ground Truth Data, High Quality, Multi-path Interference, Multipath, Point-clouds, Section-based, Time-of-flight Measurements
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:miun:diva-55659 (URN)10.1109/TIM.2025.3612631 (DOI)2-s2.0-105017402785 (Scopus ID)
Available from: 2025-10-07 Created: 2025-10-07 Last updated: 2025-10-08Bibliographically approved
Berg, O. A., Saqib, E., Jantsch, A., O'Nils, M., Krug, S., Shallari, I. & Sánchez Leal, I. (2025). Efficient Inference of parallel partitioned hybrid-Vision Transformers. In: 2025 CYBER-PHYSICAL SYSTEMS AND INTERNET-OF-THINGS WEEK, CPS-IOT WEEK WORKSHOPS: . Paper presented at 4th International Workshop on Real-time and Intelligent Edge Computing-RAGE, MAY 06-09, 2025, Irvine, CA. Association for Computing Machinery (ACM), Article ID 5.
Open this publication in new window or tab >>Efficient Inference of parallel partitioned hybrid-Vision Transformers
Show others...
2025 (English)In: 2025 CYBER-PHYSICAL SYSTEMS AND INTERNET-OF-THINGS WEEK, CPS-IOT WEEK WORKSHOPS, Association for Computing Machinery (ACM) , 2025, article id 5Conference paper, Published paper (Refereed)
Abstract [en]

Recent advancements have explored parallel partitioning of Transformers and Convolutional Neural Network (CNN) based models across networks of edge devices to accelerate deep neural network (DNN) inference. However, partitioning strategies for hybrid Vision Transformers-models integrating convolutional and attention layers-remain underdeveloped, particularly in scenarios with low communication data rates. This work introduces a novel partitioning scheme tailored for hybrid Vision Transformers, addressing communication latency through efficient compressed communication and model size reduction. The proposed approach incorporates a trainable quantization and JPEG compression pipeline to minimize overhead. We evaluate our scheme on two state-of-the-art architectures, edgeViT and CoatNet. For a communication data rate of 10 MB/s and partitioning across 12 devices, we achieve up to a 1.74x speed-up and a 5.34x model size reduction for edgeViT-XXS. Similarly, on a customized CoatNet-0, our method achieves a 1.40x speed-up and a 2.66x reduction in model size, demonstrating the efficacy of the approach in real-world scenarios.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2025
Keywords
Partitioning, Parallelization, IoT, ViT, hybrid Transformers
National Category
Computer Sciences
Identifiers
urn:nbn:se:miun:diva-54746 (URN)10.1145/3722567.3727849 (DOI)001498351300005 ()9798400716119 (ISBN)
Conference
4th International Workshop on Real-time and Intelligent Edge Computing-RAGE, MAY 06-09, 2025, Irvine, CA
Available from: 2025-06-24 Created: 2025-06-24 Last updated: 2025-09-25Bibliographically approved
Zakaryapour Sayyad, F., Shallari, I., Seyed Jalaleddin, M. & O'Nils, M. (2025). Model Evaluation and Selection for Robust and Efficient Advertisement Detection in Print Media. In: Communications in Computer and Information Science: . Paper presented at Communications in Computer and Information Science (pp. 211-224). Springer Nature
Open this publication in new window or tab >>Model Evaluation and Selection for Robust and Efficient Advertisement Detection in Print Media
2025 (English)In: Communications in Computer and Information Science, Springer Nature , 2025, p. 211-224Conference paper, Published paper (Refereed)
Abstract [en]

The localization and identification of advertisements play a pivotal role in content analysis and information retrieval. Acknowledging this significance, this paper focuses on the critical task of model evaluation and selection for robust and efficient advertisement detection. Employing a comprehensive methodology, we assess various deep learning models based on their accuracy, efficiency, and reliability in detecting and localizing advertisements within diverse print media formats. Our study reveals that certain models significantly outperform others in terms of mean Average Precision (mAP) and F1 scores, while also maintaining low inference latencies. These findings have profound implications for the development of more effective and efficient advertisement detection systems in print media. The conclusions drawn from our research provide valuable insights for future advancements in digital advertising and media analytics. 

Place, publisher, year, edition, pages
Springer Nature, 2025
Keywords
Advertisement Detection, Consumer Economy, Deep Learning, Object Detection
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:miun:diva-53059 (URN)10.1007/978-3-031-70906-7_18 (DOI)2-s2.0-85208025830 (Scopus ID)9783031709050 (ISBN)
Conference
Communications in Computer and Information Science
Available from: 2024-11-12 Created: 2024-11-12 Last updated: 2025-09-25Bibliographically approved
Berg, O. A., Saqib, E., Jantsch, A., O'Nils, M., Shallari, I., Sánchez Leal, I. & Krug, S. (2025). Quantization-Aware Training for Autoencoder-Based Partitioning of CNNs. In: IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops: . Paper presented at IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops (pp. 261-266). IEEE conference proceedings (2025)
Open this publication in new window or tab >>Quantization-Aware Training for Autoencoder-Based Partitioning of CNNs
Show others...
2025 (English)In: IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops, IEEE conference proceedings, 2025, no 2025, p. 261-266Conference paper, Published paper (Refereed)
Abstract [en]

We enhance Convolutional Neural Network (CNN) deployment through optimized model partitioning for efficient implementation on resource-constrained IoT devices. Our strategy involves (a) a novel trainable activation sharing function and (b) combining lossy and lossless compression at the partitioning point. Using an autoencoder, trained via Knowledge Distillation, with activation sharing and LZMA compression, we maximize intermediate feature map compression and accelerate IoT-Server communication. Experiments on ResNet50 and YOLOv10n, implemented on a Raspberry Pi 4 with LTE-C4 and WiFi(ah), demonstrate up to 7× speed-up compared to full in-node processing, with less than 1% accuracy drop and 99% IoT-model size reduction via 3-bit quantization (ResNet50 on Tiny-ImageNet-200). We analyze latency across key processes, highlighting significant improvements in IoT-node latency and performance preservation. Our approach enhances existing methods in model partitioning and quantization, offering solutions for reducing latency on the IoT-node while maintaining performance. 

Place, publisher, year, edition, pages
IEEE conference proceedings, 2025
Keywords
IoT, Knowledge Distillation, Neural Network Partitioning, Quantized Aware Training, YOLOv10
National Category
Computer Engineering
Identifiers
urn:nbn:se:miun:diva-55249 (URN)10.1109/PerComWorkshops65533.2025.00075 (DOI)001540474700057 ()2-s2.0-105011480158 (Scopus ID)9798331535537 (ISBN)
Conference
IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops
Available from: 2025-08-11 Created: 2025-08-11 Last updated: 2025-09-26Bibliographically approved
Berg, O. A., Saqib, E., Jantsch, A., Shallari, I., Krug, S., Sánchez Leal, I. & O'Nils, M. (2025). TCL: Time-dependent Clustering Loss for Optimizing Post-Training Feature Map Quantization for Partitioned DNNs. IEEE Access, 13, 103640-103648
Open this publication in new window or tab >>TCL: Time-dependent Clustering Loss for Optimizing Post-Training Feature Map Quantization for Partitioned DNNs
Show others...
2025 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 13, p. 103640-103648Article in journal (Refereed) Published
Abstract [en]

This paper introduces an enhanced approach for deploying deep learning models on resource-constrained IoT devices by combining model partitioning, autoencoder-based compression, quantization with Time Dependent Clustering Loss (TCL) regularization, and lossless compression, to reduce communication overhead, minimizing latency while maintaining accuracy. The autoencoder compresses feature maps at the partitioning point before quantization, effectively reducing data size and preserving accuracy. TCL regularization clusters activations at the partitioning point to align with quantization levels, minimizing quantization error and ensuring accuracy even with extreme low-bitwidth quantization. Our method is evaluated on classification models (ResNet-50, EfficientNetV2-S) and an object detection model (YOLOv10n) using the TinyImageNet-200 and Pascal VOC datasets. Deployed on Raspberry Pi 4 B and GPU, each model is tested across various partitioning points, quantization bit-widths (1-bit, 2-bit, and 3-bit), communication datarate (1MB/s to 10MB/s), and LZMA lossless compression. For a partitioned ResNet-50 after the convolutional stem block, the speed-up against a server solution is 2.33× and 1.85x compared to the all-in-node solution, with only a minimal accuracy drop of less than one percentage points. The proposed framework offers a scalable solution for deploying high-performance AI models on IoT devices, extending the feasibility of real-time inference in resource-constrained environments. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
CNN, IoT, Partitioning, Quantization
National Category
Computer Systems
Identifiers
urn:nbn:se:miun:diva-54756 (URN)10.1109/ACCESS.2025.3579107 (DOI)001512606800010 ()2-s2.0-105008273568 (Scopus ID)
Available from: 2025-06-24 Created: 2025-06-24 Last updated: 2025-09-25
Seyed Jalaleddin, M., Shallari, I. & O'Nils, M. (2025). Temporal Image Sequence Fusion with PSO-Optimised CNN Transfer Learning for Plant Temporal State Categorisation. In: 2025 IEEE Congress on Evolutionary Computation, CEC 2025: . Paper presented at 2025 IEEE Congress on Evolutionary Computation, CEC 2025. IEEE conference proceedings
Open this publication in new window or tab >>Temporal Image Sequence Fusion with PSO-Optimised CNN Transfer Learning for Plant Temporal State Categorisation
2025 (English)In: 2025 IEEE Congress on Evolutionary Computation, CEC 2025, IEEE conference proceedings, 2025Conference paper, Published paper (Refereed)
Abstract [en]

The accurate classification of plant growth stages from image sequences is a challenging problem due to the gradual and continuous nature of plant development. Significant variability exists within the same category of growth, and temporal boundaries between successive categories often exhibit visual similarities, complicating the classification task. Whereas the existing approaches are limited and rely on the analysis of single frames, they do not take temporal information embedded in sequential data, reducing robustness and accuracy. They also show vulnerability to intra-category variability among the images and also to inter-class overlap at the temporal boundaries. To address these challenges, we propose a novel approach, PSO-SqueezeTempVote, for plant temporal state categorisation that combines PSO for hyperparameter optimisation of the SqueezeNet architecture with Temporal Image Sequence Fusion (TISF) for temporal ensemble learning. The PSO part effectively optimises some critical hyperparameters, hence improving the performance of the model on individual frames. Meanwhile, the component of TISF exploits temporal continuity using a sequence of images' aggregation through majority voting. This enables classifying the segment of an image much more precisely than the previously done approaches. Results on our datasets confirm that the PSOSqueezeTempVot has improved the state-of-the-art techniques by almost 13%, out of which approximately 6% was contributed by TISF itself. 

Place, publisher, year, edition, pages
IEEE conference proceedings, 2025
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:miun:diva-55200 (URN)10.1109/CEC65147.2025.11042918 (DOI)001539410900006 ()2-s2.0-105010419966 (Scopus ID)979-8-3315-3431-8 (ISBN)
Conference
2025 IEEE Congress on Evolutionary Computation, CEC 2025
Available from: 2025-07-22 Created: 2025-07-22 Last updated: 2025-10-17Bibliographically approved
Hussain, M., O'Nils, M., Lundgren, J. & Seyed Jalaleddin, M. (2024). A Comprehensive Review On Deep Learning-Based Data Fusion. IEEE Access, 12, 180093-180124
Open this publication in new window or tab >>A Comprehensive Review On Deep Learning-Based Data Fusion
2024 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 12, p. 180093-180124Article in journal (Refereed) Published
Abstract [en]

The rapid progress in sensor technology and computational capabilities has significantly improved real-time data collection, enabling precise monitoring of various phenomena and industrial processes. However, the volume and complexity of heterogeneous data present substantial processing challenges. Traditional data-processing techniques, such as data aggregation, filtering, and statistical analysis, are increasingly supplemented by data fusion methods. These methods can be broadly categorised into traditional analytics-based approaches, like the Kalman Filter and Particle Filter, and learning-based approaches, utilising machine learning and deep learning techniques such as Artificial Neural Networks (ANN) and Convolutional Neural Networks (CNN). These techniques combine data from multiple sources to provide a comprehensive and accurate representation of information, which is critical in number of fields. Despite this, a comprehensive review of learning-based, particularly deep learning-based, data fusion strategies is lacking. This paper presents a thorough review of deep learning-based data fusion methodologies across various fields, examining their evolution over the past five years. It highlights applications in remote sensing, healthcare, industrial fault diagnosis, intelligent transportation, and other domains. The paper categories fusion strategies into early-level, intermediate-level, late-level, and hybrid fusion, emphasising their synergies, challenges, and suitability. It outlines significant advancements, the comparative advantages of deep learning-based methods over traditional approaches, and emerging trends and future directions. To ensure a comprehensive analysis, the review is structured using the ProKnow-C methodology, a rigorous selection process that focuses on relevant literature from recent years.

Place, publisher, year, edition, pages
IEEE, 2024
Keywords
Data fusion, deep learning, early-level fusion, intermediate-level feature fusion, late level decision fusion, hybrid fusion, review.
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:miun:diva-53257 (URN)10.1109/ACCESS.2024.3508271 (DOI)2-s2.0-85210926295 (Scopus ID)
Available from: 2024-12-05 Created: 2024-12-05 Last updated: 2025-09-25Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-8607-4083

Search in DiVA

Show all publications