User profiles for Yoshua Bengio

Yoshua Bengio

Professor of computer science, University of Montreal, Mila, IVADO, CIFAR
Verified email at umontreal.ca
Cited by 785400

Deep learning

Y LeCun, Y Bengio, G Hinton - nature, 2015 - nature.com
… & Bengio, Y. On the number of linear regions of deep neural networks. In Proc. … &
Bengio, Y. Neural machine translation by jointly learning to align and translate. In Proc. …

Learning deep architectures for AI

Y Bengio - Foundations and trends® in Machine Learning, 2009 - nowpublishers.com
Theoretical results suggest that in order to learn the kind of complicated functions that can
represent high-level abstractions (eg, in vision, language, and other AI-level tasks), one may …

Representation learning: A review and new perspectives

Y Bengio, A Courville, P Vincent - IEEE transactions on pattern …, 2013 - ieeexplore.ieee.org
The success of machine learning algorithms generally depends on data representation, and
we hypothesize that this is because different representations can entangle and hide more or …

[BOOK][B] Deep learning

I Goodfellow, Y Bengio, A Courville - 2016 - books.google.com
An introduction to a broad range of topics in deep learning, covering mathematical and
conceptual background, deep learning techniques used in industry, and research perspectives.“…

Gradient-based learning applied to document recognition

Y LeCun, L Bottou, Y Bengio… - Proceedings of the …, 1998 - ieeexplore.ieee.org
Multilayer neural networks trained with the back-propagation algorithm constitute the best
example of a successful gradient based learning technique. Given an appropriate network …

Understanding the difficulty of training deep feedforward neural networks

X Glorot, Y Bengio - Proceedings of the thirteenth …, 2010 - proceedings.mlr.press
Whereas before 2006 it appears that deep multi-layer neural networks were not successfully
trained, since then several algorithms have been shown to successfully train them, with …

Scaling learning algorithms toward AI

Y Bengio, Y LeCun - 2007 - direct.mit.edu
One long-term goal of machine learning research is to produce methods that are applicable
to highly complex tasks, such as perception (vision, audition), reasoning, intelligent control, …

Estimating or propagating gradients through stochastic neurons for conditional computation

Y Bengio, N Léonard, A Courville - arXiv preprint arXiv:1308.3432, 2013 - arxiv.org
Stochastic neurons and hard non-linearities can be useful for a number of reasons in deep
learning models, but in many cases they pose a challenging problem: how to estimate the …

Deep generative stochastic networks trainable by backprop

Y Bengio, E Laufer, G Alain… - … Conference on Machine …, 2014 - proceedings.mlr.press
We introduce a novel training principle for probabilistic models that is an alternative to maximum
likelihood. The proposed Generative Stochastic Networks (GSN) framework is based on …

Generative adversarial nets

…, S Ozair, A Courville, Y Bengio - Advances in neural …, 2014 - proceedings.neurips.cc
We propose a new framework for estimating generative models via adversarial nets, in
which we simultaneously train two models: a generative model G that captures the data …