JOURNAL ARTICLE

Semi-supervised Sequence Learning

Andrew M. DaiQuoc V. Le

Year: 2015 Journal:   arXiv (Cornell University) Vol: 28 Pages: 3079-3087   Publisher: Cornell University

Abstract

We present two approaches that use unlabeled data to improve sequence learning with recurrent networks. The first approach is to predict what comes next in a sequence, which is a conventional language model in natural language processing. The second approach is to use a sequence autoencoder, which reads the input sequence into a vector and predicts the input sequence again. These two algorithms can be used as a "pretraining" step for a later supervised sequence learning algorithm. In other words, the parameters obtained from the unsupervised step can be used as a starting point for other supervised training models. In our experiments, we find that long short term memory recurrent networks after being pretrained with the two approaches are more stable and generalize better. With pretraining, we are able to train long short term memory recurrent networks up to a few hundred timesteps, thereby achieving strong performance in many text classification tasks, such as IMDB, DBpedia and 20 Newsgroups.

Keywords:
Sequence (biology) Computer science Autoencoder Artificial intelligence Sequence learning Recurrent neural network Deep learning Pattern recognition (psychology) Sequence labeling Point (geometry) Machine learning Natural language processing Artificial neural network Mathematics Task (project management)

Metrics

487
Cited By
0.00
FWCI (Field Weighted Citation Impact)
38
Refs
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Text and Document Classification Technologies
Physical Sciences →  Computer Science →  Artificial Intelligence
© 2026 ScienceGate Book Chapters — All rights reserved.