Deep Neural Networks in a Mathematical Framework [electronic resource] / by Anthony L. Caterini, Dong Eui Chang.

Por: Caterini, Anthony L [author.]Colaborador(es): Chang, Dong Eui [author.] | SpringerLink (Online service)Tipo de material: TextoTextoSeries SpringerBriefs in Computer ScienceEditor: Cham : Springer International Publishing : Imprint: Springer, 2018Edición: 1st ed. 2018Descripción: XIII, 84 p. online resourceTipo de contenido: text Tipo de medio: computer Tipo de portador: online resourceISBN: 9783319753041Tema(s): Artificial intelligence | Pattern recognition | Artificial Intelligence | Pattern RecognitionFormatos físicos adicionales: Printed edition:: Sin título; Printed edition:: Sin títuloClasificación CDD: 006.3 Clasificación LoC:Q334-342Recursos en línea: Libro electrónicoTexto En: Springer Nature eBookResumen: This SpringerBrief describes how to build a rigorous end-to-end mathematical framework for deep neural networks. The authors provide tools to represent and describe neural networks, casting previous results in the field in a more natural light. In particular, the authors derive gradient descent algorithms in a unified way for several neural network structures, including multilayer perceptrons, convolutional neural networks, deep autoencoders and recurrent neural networks. Furthermore, the authors developed framework is both more concise and mathematically intuitive than previous representations of neural networks. This SpringerBrief is one step towards unlocking the black box of Deep Learning. The authors believe that this framework will help catalyze further discoveries regarding the mathematical properties of neural networks.This SpringerBrief is accessible not only to researchers, professionals and students working and studying in the field of deep learning, but also to those outside of the neutral network community.
Star ratings
    Valoración media: 0.0 (0 votos)
Existencias
Tipo de ítem Biblioteca actual Colección Signatura Copia número Estado Fecha de vencimiento Código de barras
Libro Electrónico Biblioteca Electrónica
Colección de Libros Electrónicos 1 No para préstamo

Acceso multiusuario

This SpringerBrief describes how to build a rigorous end-to-end mathematical framework for deep neural networks. The authors provide tools to represent and describe neural networks, casting previous results in the field in a more natural light. In particular, the authors derive gradient descent algorithms in a unified way for several neural network structures, including multilayer perceptrons, convolutional neural networks, deep autoencoders and recurrent neural networks. Furthermore, the authors developed framework is both more concise and mathematically intuitive than previous representations of neural networks. This SpringerBrief is one step towards unlocking the black box of Deep Learning. The authors believe that this framework will help catalyze further discoveries regarding the mathematical properties of neural networks.This SpringerBrief is accessible not only to researchers, professionals and students working and studying in the field of deep learning, but also to those outside of the neutral network community.

UABC ; Temporal ; 01/01/2021-12/31/2023.

Con tecnología Koha