Imagen de Google Jackets

Supervised Sequence Labelling with Recurrent Neural Networks [electronic resource] / by Alex Graves.

Por: Tipo de material: TextoTextoSeries Studies in Computational Intelligence ; 385 | Studies in Computational Intelligence ; 385Editor: Berlin, Heidelberg : Springer Berlin Heidelberg, 2012Descripción: XIV, 146 p. online resourceTipo de contenido:
  • text
Tipo de medio:
  • computer
Tipo de soporte:
  • online resource
ISBN:
  • 9783642247972
Trabajos contenidos:
  • SpringerLink (Online service)
Tema(s): Formatos físicos adicionales: Sin títuloClasificación CDD:
  • 006.3 23
Clasificación LoC:
  • Q342
Recursos en línea:
Contenidos:
Springer eBooksResumen: Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning toolsrobust to input noise and distortion, able to exploit long-range contextual informationthat would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary. The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal.Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video. Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.
Etiquetas de esta biblioteca: No hay etiquetas de esta biblioteca para este título. Ingresar para agregar etiquetas.
Valoración
    Valoración media: 0.0 (0 votos)
No hay ítems correspondientes a este registro

Introduction -- Supervised Sequence Labelling -- Neural Networks -- Long Short-Term Memory -- A Comparison of Network Architectures -- Hidden Markov Model Hybrids -- Connectionist Temporal Classification -- Multidimensional Networks -- Hierarchical Subsampling Networks.

Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning toolsrobust to input noise and distortion, able to exploit long-range contextual informationthat would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary. The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal.Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video. Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.

ZDB-2-ENG

No hay comentarios en este titulo.

para colocar un comentario.