"Interpretation and Robustness of Black-Box ML Models for Secure Cyber " by Qudrat E. Alahy Ratul

Publication Date

5-2024

Date of Final Oral Examination (Defense)

3-25-2022

Type of Culminating Activity

Dissertation

Degree Title

Doctor of Philosophy in Computing

Department Filter

Computer Science

Department

Computer Science

Supervisory Committee Chair

Edoardo Serra, Ph.D.

Supervisory Committee Member

Francesca Spezzano, Ph.D.

Supervisory Committee Member

Maria Soledad Pera, Ph.D.

Abstract

The advent of sophisticated Machine Learning (ML) models has revolutionized the cybersecurity landscape, offering unprecedented capabilities to address complex challenges. Despite the ML model's capabilities, such models are still perceived as black boxes, the critical aspects of interpretability and robustness within these models often remain underexplored, rendering them susceptible to adversarial attacks and hindering their integration into sensitive applications. This dissertation focuses on interpretability and robustness of black-box ML models, aiming to fortify the cyberinfrastructure. Through a meticulous investigation of advanced techniques and methodologies, this work contributes towards elucidating and enhancing the interpretability and robustness of ML models in cybersecurity contexts, thereby laying the groundwork for the development of ML systems that are both reliable and trustworthy.

This research has four contributions: Firstly, it analyzes the resilience of Automatic Scientific Claim Verification (ASCV) tools against sophisticated adversarial rephrasing, underscoring the imperative for robust model design and introducing an innovative attack model that crafts precise adversarial examples. Secondly, it assesses the impact of attribution methods on improving ML interpretability, highlighting their crucial role in fostering human-AI trust. Thirdly, the dissertation introduces the Generality and Precision Shapley Attributions (GAPS) method, a novel approach that enhances ML model explanations. Lastly, as a possible use case, this thesis explores a few-shot transfer learning approach tailored for rapid user personalization in privacy-sensitive image classification, to show the practical applicability of interpretable ML techniques within the privacy domain. Collectively, these contributions not only advance the understanding of ML model dynamics but also pave the way for their effective application in safeguarding cyberspaces.

DOI

https://doi.org/10.18122/td.2258.boisestate

Available for download on Friday, May 01, 2026

Share

COinS