Topic: Data transfer
Published:
2026
- DetailsConference Toward Privacy-Aware Environmental Monitoring of CO2 and Air Pollutants in Southern ItalyCampanile, L., Di Bonito, L. P., Marulli, F., Balzanella, A., & Verde, R. (2026). Toward Privacy-Aware Environmental Monitoring of CO2 and Air Pollutants in Southern Italy [Conference paper]. Lecture Notes in Computer Science, 15893 LNCS, 317–333. https://doi.org/10.1007/978-3-031-97645-2_21
Abstract
The increasing levels of CO2 and air pollutants represent a major challenge to environmental sustainability and public health, particularly in regions characterized by complex geographic and socio-economic dynamics. This work proposes a study focused on the Southern Italy regions, where environmental vulnerabilities are displayed, along with a limited availability of high-granularity data. The main aim of this work is to build and provide a comprehensive and detailed dataset tailored to the region’s unique needs, by leveraging datasets from EDGAR for greenhouse gases and air pollutants, integrated with demographic and territorial morphology data from ISTAT. The creation of composite indicators to monitor trends in emissions and pollution on a fine spatial scale is supported by the data set. These indicators enable initial insight into spatial disparities in pollutant concentrations, offering valuable data to inform targeted policy interventions. The work provided a foundation for next analytical studies, integrating different datasets and highlighting the potential for complex spatiotemporal analysis. The study provides a robust dataset and preliminary insights, enhancing the understanding of environmental dynamics in Southern Italy. Subsequent efforts will focus on extending this methodology to more extensive geographic contexts and incorporating real-time data for adaptive monitoring. The proposed framework also lays the groundwork for privacy-aware environmental monitoring solutions, enabling future integration with edge and IoT-based architectures while addressing privacy and data protection concerns. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
2025
- DetailsMarulli, F., Campanile, L., Ragucci, G., Carbone, S., & Bifulco, M. (2025). Data Generation and Cybersecurity: A Major Opportunity or the Next Nightmare? [Conference paper]. Proceedings of the 2025 IEEE International Conference on Cyber Security and Resilience, CSR 2025, 969–974. https://doi.org/10.1109/CSR64739.2025.11130069
Abstract
In recent years, the proliferation of synthetic data generation techniques-driven by advances in artificial intelli-gence-has opened new possibilities across a wide range of fields, from healthcare to autonomous systems, by addressing critical data scarcity issues. However, this technological progress also brings with it a growing concern: the dual-use nature of synthetic data. While it offers powerful tools for innovation, it simultaneously introduces significant risks related to information disorder and cybersecurity. As AI systems become increasingly capable of producing highly realistic yet entirely fabricated content, the boundaries between authentic and artificial information blur, making it more difficult to detect manipulation, protect digital infrastructures, and maintain public trust. This work undertakes a preliminary exploration of the evolving nexus between Generative AI, Information Disorder, and Cybersecurity: it aims to investigate the complex interplay among these three and to map their dynamic interactions and reciprocal influences, highlighting both the potential benefits and the looming challenges posed by this evolving landscape. Moreover, it seeks to propose a conceptual framework for assessing these interdependencies through a set of indicative metrics, offering a foundation for future empirical evaluation and strategic response. © 2025 IEEE.
2024
- DetailsBook Chapter Combining Federated and Ensemble Learning in Distributed and Cloud Environments: An Exploratory StudyMarulli, F., Campanile, L., Marrone, S., & Verde, L. (2024). Combining Federated and Ensemble Learning in Distributed and Cloud Environments: An Exploratory Study [Book chapter]. Lecture Notes on Data Engineering and Communications Technologies, 203, 297–306. https://doi.org/10.1007/978-3-031-57931-8_29
Abstract
Conventional modern Machine Learning (ML) applications involve training models in the cloud and then transferring them back to the edge, especially in an Internet of Things (IoT) enabled environment. However, privacy-related limitations on data transfer from the edge to the cloud raise challenges: among various solutions, Federated Learning (FL) could satisfy privacy related concerns and accommodate power and energy issues of edge devices. This paper proposes a novel approach that combines FL and Ensemble Learning (EL) to improve both security and privacy challenges. The presented methodology introduces an extra layer, the Federation Layer, to enhance security. It uses Bayesian Networks (BNs) to dynamically filter untrusted/unsecure federation clients. This approach presents a solution for increasing the security and robustness of FL systems, considering also privacy and performance aspects. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
2022
- DetailsCampanile, L., Forgione, F., Mastroianni, M., Palmiero, G., & Sanghez, C. (2022). Evaluating the Impact of Data Anonymization in a Machine Learning Application [Conference paper]. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 13380 LNCS, 389–400. https://doi.org/10.1007/978-3-031-10542-5_27
Abstract
The data protection impact assessment is used to verify the necessity, proportionality and risks of data processing. Our work is based on the data processed by the technical support of a Wireless Service Provider. The team of WISP tech support uses a machine learning system to predict failures. The goal of our the experiments was to evaluate the DPIA with personal data and without personal data. In fact, in a first scenario, the experiments were conducted using a machine learning application powered by non-anonymous personal data. Instead in the second scenario, the data was anonymized before feeding the machine learning system. In this article we evaluate how much the Data Protection Impact Assessment changes when moving from a scenario with raw data to a scenario with anonymized data. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
2021
- DetailsMarulli, F., Verde, L., & Campanile, L. (2021). Exploring data and model poisoning attacks to deep learning-based NLP systems [Conference paper]. Procedia Computer Science, 192, 3570–3579. https://doi.org/10.1016/j.procs.2021.09.130
Abstract
Natural Language Processing (NLP) is being recently explored also to its application in supporting malicious activities and objects detection. Furthermore, NLP and Deep Learning have become targets of malicious attacks too. Very recent researches evidenced that adversarial attacks are able to affect also NLP tasks, in addition to the more popular adversarial attacks on deep learning systems for image processing tasks. More precisely, while small perturbations applied to the data set adopted for training typical NLP tasks (e.g., Part-of-Speech Tagging, Named Entity Recognition, etc..) could be easily recognized, models poisoning, performed by the means of altered data models, typically provided in the transfer learning phase to a deep neural networks (e.g., poisoning attacks by word embeddings), are harder to be detected. In this work, we preliminary explore the effectiveness of a poisoned word embeddings attack aimed at a deep neural network trained to accomplish a Named Entity Recognition (NER) task. By adopting the NER case study, we aimed to analyze the severity of such a kind of attack to accuracy in recognizing the right classes for the given entities. Finally, this study represents a preliminary step to assess the impact and the vulnerabilities of some NLP systems we adopt in our research activities, and further investigating some potential mitigation strategies, in order to make these systems more resilient to data and models poisoning attacks. © 2021 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (https://creativecommons.org/licenses/by-nc-nd/4.0) Peer-review under responsibility of the scientific committee of KES International.
2026
- DetailsConference Toward Privacy-Aware Environmental Monitoring of CO2 and Air Pollutants in Southern ItalyCampanile, L., Di Bonito, L. P., Marulli, F., Balzanella, A., & Verde, R. (2026). Toward Privacy-Aware Environmental Monitoring of CO2 and Air Pollutants in Southern Italy [Conference paper]. Lecture Notes in Computer Science, 15893 LNCS, 317–333. https://doi.org/10.1007/978-3-031-97645-2_21
Abstract
The increasing levels of CO2 and air pollutants represent a major challenge to environmental sustainability and public health, particularly in regions characterized by complex geographic and socio-economic dynamics. This work proposes a study focused on the Southern Italy regions, where environmental vulnerabilities are displayed, along with a limited availability of high-granularity data. The main aim of this work is to build and provide a comprehensive and detailed dataset tailored to the region’s unique needs, by leveraging datasets from EDGAR for greenhouse gases and air pollutants, integrated with demographic and territorial morphology data from ISTAT. The creation of composite indicators to monitor trends in emissions and pollution on a fine spatial scale is supported by the data set. These indicators enable initial insight into spatial disparities in pollutant concentrations, offering valuable data to inform targeted policy interventions. The work provided a foundation for next analytical studies, integrating different datasets and highlighting the potential for complex spatiotemporal analysis. The study provides a robust dataset and preliminary insights, enhancing the understanding of environmental dynamics in Southern Italy. Subsequent efforts will focus on extending this methodology to more extensive geographic contexts and incorporating real-time data for adaptive monitoring. The proposed framework also lays the groundwork for privacy-aware environmental monitoring solutions, enabling future integration with edge and IoT-based architectures while addressing privacy and data protection concerns. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
2025
- DetailsMarulli, F., Campanile, L., Ragucci, G., Carbone, S., & Bifulco, M. (2025). Data Generation and Cybersecurity: A Major Opportunity or the Next Nightmare? [Conference paper]. Proceedings of the 2025 IEEE International Conference on Cyber Security and Resilience, CSR 2025, 969–974. https://doi.org/10.1109/CSR64739.2025.11130069
Abstract
In recent years, the proliferation of synthetic data generation techniques-driven by advances in artificial intelli-gence-has opened new possibilities across a wide range of fields, from healthcare to autonomous systems, by addressing critical data scarcity issues. However, this technological progress also brings with it a growing concern: the dual-use nature of synthetic data. While it offers powerful tools for innovation, it simultaneously introduces significant risks related to information disorder and cybersecurity. As AI systems become increasingly capable of producing highly realistic yet entirely fabricated content, the boundaries between authentic and artificial information blur, making it more difficult to detect manipulation, protect digital infrastructures, and maintain public trust. This work undertakes a preliminary exploration of the evolving nexus between Generative AI, Information Disorder, and Cybersecurity: it aims to investigate the complex interplay among these three and to map their dynamic interactions and reciprocal influences, highlighting both the potential benefits and the looming challenges posed by this evolving landscape. Moreover, it seeks to propose a conceptual framework for assessing these interdependencies through a set of indicative metrics, offering a foundation for future empirical evaluation and strategic response. © 2025 IEEE.
2024
- DetailsBook Chapter Combining Federated and Ensemble Learning in Distributed and Cloud Environments: An Exploratory StudyMarulli, F., Campanile, L., Marrone, S., & Verde, L. (2024). Combining Federated and Ensemble Learning in Distributed and Cloud Environments: An Exploratory Study [Book chapter]. Lecture Notes on Data Engineering and Communications Technologies, 203, 297–306. https://doi.org/10.1007/978-3-031-57931-8_29
Abstract
Conventional modern Machine Learning (ML) applications involve training models in the cloud and then transferring them back to the edge, especially in an Internet of Things (IoT) enabled environment. However, privacy-related limitations on data transfer from the edge to the cloud raise challenges: among various solutions, Federated Learning (FL) could satisfy privacy related concerns and accommodate power and energy issues of edge devices. This paper proposes a novel approach that combines FL and Ensemble Learning (EL) to improve both security and privacy challenges. The presented methodology introduces an extra layer, the Federation Layer, to enhance security. It uses Bayesian Networks (BNs) to dynamically filter untrusted/unsecure federation clients. This approach presents a solution for increasing the security and robustness of FL systems, considering also privacy and performance aspects. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
2022
- DetailsCampanile, L., Forgione, F., Mastroianni, M., Palmiero, G., & Sanghez, C. (2022). Evaluating the Impact of Data Anonymization in a Machine Learning Application [Conference paper]. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 13380 LNCS, 389–400. https://doi.org/10.1007/978-3-031-10542-5_27
Abstract
The data protection impact assessment is used to verify the necessity, proportionality and risks of data processing. Our work is based on the data processed by the technical support of a Wireless Service Provider. The team of WISP tech support uses a machine learning system to predict failures. The goal of our the experiments was to evaluate the DPIA with personal data and without personal data. In fact, in a first scenario, the experiments were conducted using a machine learning application powered by non-anonymous personal data. Instead in the second scenario, the data was anonymized before feeding the machine learning system. In this article we evaluate how much the Data Protection Impact Assessment changes when moving from a scenario with raw data to a scenario with anonymized data. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
2021
- DetailsMarulli, F., Verde, L., & Campanile, L. (2021). Exploring data and model poisoning attacks to deep learning-based NLP systems [Conference paper]. Procedia Computer Science, 192, 3570–3579. https://doi.org/10.1016/j.procs.2021.09.130
Abstract
Natural Language Processing (NLP) is being recently explored also to its application in supporting malicious activities and objects detection. Furthermore, NLP and Deep Learning have become targets of malicious attacks too. Very recent researches evidenced that adversarial attacks are able to affect also NLP tasks, in addition to the more popular adversarial attacks on deep learning systems for image processing tasks. More precisely, while small perturbations applied to the data set adopted for training typical NLP tasks (e.g., Part-of-Speech Tagging, Named Entity Recognition, etc..) could be easily recognized, models poisoning, performed by the means of altered data models, typically provided in the transfer learning phase to a deep neural networks (e.g., poisoning attacks by word embeddings), are harder to be detected. In this work, we preliminary explore the effectiveness of a poisoned word embeddings attack aimed at a deep neural network trained to accomplish a Named Entity Recognition (NER) task. By adopting the NER case study, we aimed to analyze the severity of such a kind of attack to accuracy in recognizing the right classes for the given entities. Finally, this study represents a preliminary step to assess the impact and the vulnerabilities of some NLP systems we adopt in our research activities, and further investigating some potential mitigation strategies, in order to make these systems more resilient to data and models poisoning attacks. © 2021 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (https://creativecommons.org/licenses/by-nc-nd/4.0) Peer-review under responsibility of the scientific committee of KES International.
