**This is the where the meetup will be streamed starting at around 7:15PM Paris time. **

The theme of today's meetup will about (mostly) Deep Learning. It will be held in conjunction with the

Deep Learning Paris meetup . We should also have the

Kiev Deep Learning Meetup as a guest audience as well. All the presentation slides should be available below by the time the meetup starts. Our host and sponsor in Paris will be

Criteo. Here is the tentative schedule/program that starts at 19h15 Paris time.

and the presentation slides:

+

Yoshua Bengio, Title:

Deep Learning Theory (remote from the London Machine Learning meetup)

Although neural networks have long been considered lacking in theory and much remains to be done, theoretical evidence is mounting and will be discussed, to support distributed representations, depth of representation, the non-convexity of the training objective, and the probabilistic interpretation of learning algorithms (especially of the auto-encoder type, which were lacking one). The talk will focus on the intuitions behind these theoretical results.

+

Sander Dieleman and

Ira Korshunova, Ghent University

Title:

Classifying plankton with deep neural networks by the Deep Sea team from Reservoir Lab
Deep
learning has become a very popular approach for solving computer vision
problems in recent years. In this talk we'll demonstrate how this
approach can be applied in practice. We'll show how our team of 7 built a
model for the automated classification of plankton based on
convolutional neural networks. Using this model, we placed 1st in the
National Data Science Bowl competition on Kaggle.

+

Gabriella Contardo, LIP6, UPMC

Learning to build representations from partial information: Application to cold-start recommendation

Most
of the successful machine learning algorithms rely on data
representation, i.e a way to disentangle and extract useful information
from data, which will help the model in its objective task. Classical
approaches build representations based on fully observed data. But in
many cases, one wants to build representations ''on the fly'', based on a
partially observed information. As an example, learning representations
over users can be done by progressively gathering information about
their profiles. This paper presents an inductive representation-based
model to tackle the twofold more general problem of (i) selecting the
right information to collect for building relevant representations, (ii)
updating these representations based on new incoming information. It is
developed in this paper to design static interview for the cold-start
collaborative filtering problem but it can also be used to go smoothly
to the warm context where all information has been gathered.

+

Guillaume Wenzek
Sentiment Analysis With Recursive Neural Tensor Network / Analyse de sentiment à l'aide de réseaux de neurones récursifs

Sentiment analysis is one of the hardest NLP (Natural Language Processing) task, due to complex linguistic structures such as negation or double-negation. Socher et al. introduced a method that combines a classic NLP tool, a syntaxic parser, with a special kind of neural networks. We will review this method and introduce a few improvements in order to train on a corpus with fewer annotations than the Stanford Sentiment Treebank used in the paper.

** **
** **
**Join the CompressiveSensing subreddit or the Google+ Community and post there !**

Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.