Repository logo
  • Log In
    Log in via Symplectic to deposit your publication(s).
Repository logo
  • Communities & Collections
  • Research Outputs
  • Statistics
  • Log In
    Log in via Symplectic to deposit your publication(s).
  1. Home
  2. Faculty of Engineering
  3. Computing
  4. Computing PhD theses
  5. Robustness in natural language processing
 
  • Details
Robustness in natural language processing
File(s)
Li-Z-2025-PhD-Thesis.pdf (5.55 MB)
Thesis
Author(s)
Li, Zhenhao
Type
Thesis or dissertation
Abstract
The recent success in Natural Language Processing has been driven by the evolution of pre-trained language models, which achieve state-of-the-art performance across a wide range of tasks, such as machine translation, sentiment analysis, question answering, etc. Despite these successes, ensuring robustness of NLP systems remains a significant challenge, especially when faced with non-standard inputs such as noisy texts and adversarial texts. In this thesis, we focus on three common robustness issues in NLP systems: noisy texts, textual adversarial attacks, and unintended data bias. We address the robustness issue from textual noise by integrating extra visual and conversational contexts.

We first incorporate visual contexts and error correction training in multimodal translation. Our findings reveal that the integration of multimodality and error correction effectively mitigates information loss from noise and improves translation robustness. In addition, we extend the exploration of noise robustness to multimodal conversational modeling by proposing a new task: multimodal conversation derailment detection and curating a corresponding new dataset. We propose a multimodal hierarchical model that combines textual, visual, and conversational contexts to predict conversation derailment, and experimentally demonstrate the efficacy of conversational contexts in noisy settings. To counter adversarial attacks, we propose a flexible method, DiffuseDef, to defend against adversarial attacks. Our method trains a diffusion layer on top of a PLM and performs denoising iteratively on the encoder hidden representations. Benefiting from removing latent noise in adversarial texts, our method achieves state-of-the-art performance against common black-box and white-box adversarial attacks. Finally, we tackle the robustness concern that stems from model training, specifically handling unintended bias. We propose an adversarial multitask model that extracts task-common representations and disentangles task-specific representations that can lead to bias. Our experiments on toxicity detection show that the proposed method can successfully mitigate unintended biases arising from three relevant tasks.
Version
Open Access
Date Issued
2025-07-24
Date Awarded
2025-10-01
URI
https://hdl.handle.net/10044/1/124073
DOI
https://doi.org/10.25560/124073
Copyright Statement
Attribution-Non Commercial-No Derivatives 4.0 International Licence (CC BY-NC-ND)
License URL
https://creativecommons.org/licenses/by-nc-nd/4.0/
Advisor
Specia, Lucia
Rei, Marek
Publisher Department
Department of Computing
Publisher Institution
Imperial College London
Qualification Level
Doctoral
Qualification Name
Doctor of Philosophy (PhD)
About
Spiral Depositing with Spiral Publishing with Spiral Symplectic
Contact us
Open access team Report an issue
Other Services
Scholarly Communications Library Services
logo

Imperial College London

South Kensington Campus

London SW7 2AZ, UK

tel: +44 (0)20 7589 5111

Accessibility Modern slavery statement Cookie Policy

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Cookie settings
  • Privacy policy
  • End User Agreement
  • Send Feedback