FIND: human-in-the-loop debugging deep text classifiers
File(s)2010.04987v1.pdf (3 MB)
Working paper
Author(s)
Lertvittayakumjorn, Piyawat
Specia, Lucia
Toni, Francesca
Type
Working Paper
Abstract
Since obtaining a perfect training dataset (i.e., a dataset which is
considerably large, unbiased, and well-representative of unseen cases) is
hardly possible, many real-world text classifiers are trained on the available,
yet imperfect, datasets. These classifiers are thus likely to have undesirable
properties. For instance, they may have biases against some sub-populations or
may not work effectively in the wild due to overfitting. In this paper, we
propose FIND -- a framework which enables humans to debug deep learning text
classifiers by disabling irrelevant hidden features. Experiments show that by
using FIND, humans can improve CNN text classifiers which were trained under
different types of imperfect datasets (including datasets with biases and
datasets with dissimilar train-test distributions).
considerably large, unbiased, and well-representative of unseen cases) is
hardly possible, many real-world text classifiers are trained on the available,
yet imperfect, datasets. These classifiers are thus likely to have undesirable
properties. For instance, they may have biases against some sub-populations or
may not work effectively in the wild due to overfitting. In this paper, we
propose FIND -- a framework which enables humans to debug deep learning text
classifiers by disabling irrelevant hidden features. Experiments show that by
using FIND, humans can improve CNN text classifiers which were trained under
different types of imperfect datasets (including datasets with biases and
datasets with dissimilar train-test distributions).
Date Issued
2020-10-10
Citation
2020
Copyright Statement
© 2020 The Author(s)
Identifier
http://arxiv.org/abs/2010.04987v1
Subjects
cs.CL
cs.CL
cs.HC
cs.LG
Notes
17 pages including appendices; To appear at EMNLP 2020
Publication Status
Published