skip to main content
10.1145/1143844.1143891acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
Article

Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks

Authors Info & Claims
Published:25 June 2006Publication History

ABSTRACT

Many real-world sequence learning tasks require the prediction of sequences of labels from noisy, unsegmented input data. In speech recognition, for example, an acoustic signal is transcribed into words or sub-word units. Recurrent neural networks (RNNs) are powerful sequence learners that would seem well suited to such tasks. However, because they require pre-segmented training data, and post-processing to transform their outputs into label sequences, their applicability has so far been limited. This paper presents a novel method for training RNNs to label unsegmented sequences directly, thereby solving both problems. An experiment on the TIMIT speech corpus demonstrates its advantages over both a baseline HMM and a hybrid HMM-RNN.

References

  1. Bengio., Y. (1999). Markovian models for sequential data. Neural Computing Surveys, 2, 129--162.Google ScholarGoogle Scholar
  2. Bishop, C. (1995). Neural Networks for Pattern Recognition, chapter 6. Oxford University Press, Inc. Google ScholarGoogle ScholarCross RefCross Ref
  3. Bourlard, H., & Morgan, N. (1994). Connnectionist speech recognition: A hybrid approach. Kluwer Academic Publishers. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Bridle, J. (1990). Probabilistic interpretation of feed-forward classification network outputs, with relationships to statistical pattern recognition. In F. Soulie and J. Herault (Eds.), Neurocomputing: Algorithms, architectures and applications, 227--236. Springer-Verlag.Google ScholarGoogle Scholar
  5. Gers, F.; Schraudolph, N., & Schmidhuber, J. (2002). Learning precise timing with LSTM recurrent networks. Journal of Machine Learning Research, 3, 115--143. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Graves, A., Fernández, S., & Schmidhuber, J. (2005). Bidirectional LSTM networks for improved phoneme classification and recognition. Proceedings of the 2005 International Conference on Artificial Neural Networks. Warsaw, Poland. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Graves, A., & Schmidhuber, J. (2005). Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Networks, 18, 602--610. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Hochreiter, S., & Schmidhuber, J. (1997). Long Short-Term Memory. Neural Computation, 9, 1735--1780. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Kadous, M. W. (2002). Temporal classification: Extending the classification paradigm to multivariate time series. Doctoral dissertation, School of Computer Science & Engineering, University of New South Wales. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Lafferty, J., McCallum, A., & Pereira, F. (2001). Conditional random fields: Probabilistic models for segmenting and labeling sequence data. Proc. 18th International Conf. on Machine Learning (pp. 282--289). Morgan Kaufmann, San Francisco, CA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. LeCun, Y., Bottou, L., Orr, G., & Muller, K. (1998). Efficient backprop. Neural Networks: Tricks of the trade. Springer. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Rabiner, L. R. (1989). A tutorial on hidden markov models and selected applications in speech recognition. Proc. IEEE (pp. 257--286). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  13. Robinson, A. J. (1991). Several improvements to a recurrent error propagation network phone recognition system (Technical Report CUED/FINFENG/TR82). University of Cambridge.Google ScholarGoogle Scholar
  14. Robinson, A. J. (1994). An application of recurrent nets to phone probability estimation. IEEE Transactions on Neural Networks, 5, 298--305.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Schraudolph, N. N. (2002). Fast Curvature Matrix-Vector Products for Second-Order Gradient Descent. Neural Comp., 14, 1723--1738. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Schuster, M., & Paliwal, K. K. (1997). Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45, 2673--2681. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Werbos, P. (1990). Backpropagation through time: What it does and how to do it. Proceedings of the IEEE, 78, 1550--1560.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          ICML '06: Proceedings of the 23rd international conference on Machine learning
          June 2006
          1154 pages
          ISBN:1595933832
          DOI:10.1145/1143844

          Copyright © 2006 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 25 June 2006

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • Article

          Acceptance Rates

          ICML '06 Paper Acceptance Rate140of548submissions,26%Overall Acceptance Rate140of548submissions,26%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader