Noisy Neural Language Modeling for Typing Prediction in BCI Communication
Proceedings of the Eighth Workshop on Speech and Language Processing for Assistive Technologies,
Jun 2019
Abstract
Language models have broad adoption in predictive typing tasks. When the typing history contains numerous errors, as in open-vocabulary predictive typing with brain-computer interface (BCI) systems, we observe significant performance degradation in both n-gram and recurrent neural network language models trained on clean text. In evaluations of ranking character predictions, training recurrent LMs on noisy text makes them much more robust to noisy histories, even when the error model is misspecified. We also propose an effective strategy for combining evidence from multiple ambiguous histories of BCI electroencephalogram measurements.Add the full text or supplementary notes for the publication here using Markdown formatting.