RMIT University
Browse

Next-Generation Collaborative Learning with Integrated Trust, Privacy, Governance and Scalability Solutions

Download (28.89 MB)
thesis
posted on 2025-09-11, 04:15 authored by Aditya Pribadi Kalapaaking
<p dir="ltr">This thesis investigates novel strategies to enhance trust, privacy, auditability, and scalability in federated learning systems. In distributed environments, the collaborative training of machine learning models faces critical challenges such as adversarial manipulation, privacy breaches, and inconsistencies arising from data heterogeneity and diverse client capabilities. Addressing these issues requires a multifaceted approach that secures model aggregation, verifies individual contributions, and optimizes resource allocation without compromising performance. </p><p dir="ltr">The initial segment of this research focuses on reinforcing the trustworthiness of the aggregation process. In federated learning, local models are vulnerable to tampering, which can result in a corrupted global model. To mitigate these risks, a secure aggregation procedure is proposed in which each local model is rigorously validated prior to its incorporation into the global model. A consensus mechanism is employed to ensure that the aggregated model adheres to the highest standards of integrity, with every update recorded in a tamper-resistant manner. Experimental evaluations conducted across various neural network architectures and benchmark datasets demonstrate that this method maintains model accuracy while effectively countering potential adversarial threats. </p><p dir="ltr">Subsequently, the thesis addresses the imperative of privacy preservation in scenarios involving sensitive data, such as those encountered in healthcare and industrial applications. By integrating advanced privacy-preserving techniques, the study facilitates the encrypted processing of local model updates, thereby ensuring that sensitive information remains confidential throughout the training and aggregation processes. Empirical results indicate that these privacy measures introduce only marginal overhead while preserving the efficacy of the learning process, thereby providing a balanced solution that upholds data confidentiality without impeding model performance. </p><p dir="ltr">Furthermore, the research emphasizes the necessity for auditability and verifiability within federated learning systems. Recognizing that decentralized training environments often lack centralized oversight, a robust governance mechanism is introduced to record detailed training logs and aggregation steps. This verifiable approach enables all participants to independently confirm that the prescribed protocols have been followed, thereby enhancing transparency and accountability. The capacity to audit the entire learning process not only strengthens stakeholder confidence but also aligns with evolving regulatory standards in data protection and machine learning ethics. </p><p dir="ltr">The final aspect of the study tackles the challenges of scalability and resource optimization. Real-world federated learning applications are characterized by heterogeneous client devices and variable network conditions, which can impede efficient training. To address these issues, dynamic resource management techniques are developed that adaptively partition and allocate computational tasks based on the capabilities of each participant. Comprehensive simulations and real-world experiments reveal that these methods accelerate model convergence, reduce training time, and optimize energy consumption, all while maintaining high levels of global model accuracy. </p><p dir="ltr">In summary, the contributions of this thesis are demonstrated through a comprehensive approach that fortifies federated learning systems against security threats, ensures privacy preservation, and enhances both auditability and resource efficiency. By integrating secure aggregation methods, privacy-preserving processing, verifiable governance, and dynamic resource management, this research offers a robust solution to many of the pressing challenges in distributed machine learning. The findings provide a solid foundation for future investigations aimed at further refining consensus protocols and extending the approach to accommodate heterogeneous model structures in increasingly dynamic environments.</p>

History

Degree Type

Doctorate by Research

Imprint Date

2025-04-24

School name

Computing Technologies, RMIT University

Copyright

© 2025 Aditya Pribadi Kalapaaking

Usage metrics

    Theses

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC