Publications tagged with Federated learning

Published:

Publications tagged with "Federated learning"

  1. Campanile, L., de Biase, M. S., & Marulli, F. (2025). Edge-Cloud Distributed Approaches to Text Authorship Analysis: A Feasibility Study [Book chapter]. Lecture Notes on Data Engineering and Communications Technologies, 250, 284–293. https://doi.org/10.1007/978-3-031-87778-0_28
    Abstract
    Automatic authorship analysis, often referred to as stylometry, is a captivating yet contentious field that employs computational techniques to determine the authorship of textual artefacts. In recent years, the importance of author profiling has grown significantly due to the proliferation of automatic text generation systems. These include both early-generation bots and the latest generative AI-based models, which have heightened concerns about misinformation and content authenticity. This study proposes a novel approach to evaluate the feasibility and effectiveness of contemporary distributed learning methods. The approach leverages the computational advantages of distributed systems while preserving the privacy of human contributors, enabling the collection and analysis of extensive datasets of “human-written” texts in contrast to those generated by bots. More specifically, the proposed method adopts a Federated Learning (FL) framework, integrating readability and stylometric metrics to deliver a privacy-preserving solution for Authorship Attribution (AA). The primary objective is to enhance the accuracy of AA processes, thus achieving a more robust “authorial fingerprint”. Experimental results reveal that while FL effectively protects privacy and mitigates data exposure risks, the combined use of readability and stylometric features significantly increases the accuracy of AA. This approach demonstrates promise for secure and scalable AA applications, particularly in privacy-sensitive contexts and real-time edge computing scenarios. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
    DOI Publisher Details
    Details
  2. Marulli, F., Campanile, L., Marrone, S., & Verde, L. (2024). Combining Federated and Ensemble Learning in Distributed and Cloud Environments: An Exploratory Study [Book chapter]. Lecture Notes on Data Engineering and Communications Technologies, 203, 297–306. https://doi.org/10.1007/978-3-031-57931-8_29
    Abstract
    Conventional modern Machine Learning (ML) applications involve training models in the cloud and then transferring them back to the edge, especially in an Internet of Things (IoT) enabled environment. However, privacy-related limitations on data transfer from the edge to the cloud raise challenges: among various solutions, Federated Learning (FL) could satisfy privacy related concerns and accommodate power and energy issues of edge devices. This paper proposes a novel approach that combines FL and Ensemble Learning (EL) to improve both security and privacy challenges. The presented methodology introduces an extra layer, the Federation Layer, to enhance security. It uses Bayesian Networks (BNs) to dynamically filter untrusted/unsecure federation clients. This approach presents a solution for increasing the security and robustness of FL systems, considering also privacy and performance aspects. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
    DOI Publisher Details
    Details
  3. Campanile, L., Marrone, S., Marulli, F., & Verde, L. (2022). Challenges and Trends in Federated Learning for Well-being and Healthcare [Conference paper]. Procedia Computer Science, 207, 1144–1153. https://doi.org/10.1016/j.procs.2022.09.170
    Abstract
    Currently, research in Artificial Intelligence, both in Machine Learning and Deep Learning, paves the way for promising innovations in several areas. In healthcare, especially, where large amounts of quantitative and qualitative data are transferred to support studies and early diagnosis and monitoring of any diseases, potential security and privacy issues cannot be underestimated. Federated learning is an approach where privacy issues related to sensitive data management can be significantly reduced, due to the possibility to train algorithms without exchanging data. The main idea behind this approach is that learning models can be trained in a distributed way, where multiple devices or servers with decentralized data samples can provide their contributions without having to exchange their local data. Recent studies provided evidence that prototypes trained by adopting Federated Learning strategies are able to achieve reliable performance, thus by generating robust models without sharing data and, consequently, limiting the impact on security and privacy. This work propose a literature overview of Federated Learning approaches and systems, focusing on its application for healthcare. The main challenges, implications, issues and potentials of this approach in the healthcare are outlined. © 2022 The Authors. Published by Elsevier B.V.
    DOI Publisher Details
    Details
  4. Marulli, F., Verde, L., Marrore, S., & Campanile, L. (2022). A Federated Consensus-Based Model for Enhancing Fake News and Misleading Information Debunking [Conference paper]. Smart Innovation, Systems and Technologies, 309, 587–596. https://doi.org/10.1007/978-981-19-3444-5_50
    Abstract
    Misinformation and Fake News are hard to dislodge. According to experts on this phenomenon, to fight disinformation a less credulous public is needed; so, current AI techniques can support misleading information debunking, given the human tendency to believe “facts” that confirm biases. Much effort has been recently spent by the research community on this plague: several AI-based approaches for automatic detection and classification of Fake News have been proposed; unfortunately, Fake News producers have refined their ability in eluding automatic ML and DL-based detection systems. So, debunking false news represents an effective weapon to contrast the users’ reliance on false information. In this work, we propose a preliminary study aiming to approach the design of effective fake news debunking systems, harnessing two complementary federated approaches. We propose, firstly, a federation of independent classification systems to accomplish a debunking process, by applying a distributed consensus mechanism. Secondly, a federated learning task, involving several cooperating nodes, is accomplished, to obtain a unique merged model, including features of single participants models, trained on different and independent data fragments. This study is a preliminary work aiming to to point out the feasibility and the comparability of these proposed approaches, thus paving the way to an experimental campaign that will be performed on effective real data, thus providing an evidence for an effective and feasible model for detecting potential heterogeneous fake news. Debunking misleading information is mission critical to increase the awareness of facts on the part of news consumers. © 2022, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
    DOI Publisher Details
    Details
  5. Marulli, F., Balzanella, A., Campanile, L., Iacono, M., & Mastroianni, M. (2021). Exploring a Federated Learning Approach to Enhance Authorship Attribution of Misleading Information from Heterogeneous Sources [Conference paper]. Proceedings of the International Joint Conference on Neural Networks, 2021-July. https://doi.org/10.1109/IJCNN52387.2021.9534377
    Abstract
    Authorship Attribution (AA) is currently applied in several applications, among which fraud detection and anti-plagiarism checks: this task can leverage stylometry and Natural Language Processing techniques. In this work, we explored some strategies to enhance the performance of an AA task for the automatic detection of false and misleading information (e.g., fake news). We set up a text classification model for AA based on stylometry exploiting recurrent deep neural networks and implemented two learning tasks trained on the same collection of fake and real news, comparing their performances: one is based on Federated Learning architecture, the other on a centralized architecture. The goal was to discriminate potential fake information from true ones when the fake news comes from heterogeneous sources, with different styles. Preliminary experiments show that a distributed approach significantly improves recall with respect to the centralized model. As expected, precision was lower in the distributed model. This aspect, coupled with the statistical heterogeneity of data, represents some open issues that will be further investigated in future work. © 2021 IEEE.
    DOI Publisher Details
    Details

← Back to all publications