Emmanuel Bengio

ORCID: 0000-0002-3257-4661
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Reinforcement Learning in Robotics
  • Adversarial Robustness in Machine Learning
  • Domain Adaptation and Few-Shot Learning
  • Machine Learning and Algorithms
  • Neural Networks and Applications
  • Computational Drug Discovery Methods
  • Machine Learning in Materials Science
  • Topic Modeling
  • Stochastic Gradient Optimization Techniques
  • Generative Adversarial Networks and Image Synthesis
  • Chemical Reactions and Isotopes
  • Explainable Artificial Intelligence (XAI)
  • Receptor Mechanisms and Signaling
  • Machine Learning and Data Classification
  • Data Stream Mining Techniques
  • vaccines and immunoinformatics approaches
  • RNA and protein synthesis mechanisms
  • Bacillus and Francisella bacterial research
  • Process Optimization and Integration
  • Human Pose and Action Recognition
  • Cell Image Analysis Techniques
  • Statistics Education and Methodologies
  • Bioinformatics and Genomic Networks
  • Text Readability and Simplification
  • Advanced Graph Neural Networks

Mila - Quebec Artificial Intelligence Institute
2023

McGill University
2015-2021

Université de Montréal
2013

In this paper we present the techniques used for University of Montréal's team submissions to 2013 Emotion Recognition in Wild Challenge. The challenge is classify emotions expressed by primary human subject short video clips extracted from feature length movies. This involves analysis acted scenes lasting approximately one-two seconds, including audio track which may contain voices as well background music. Our approach combines multiple deep neural networks different data modalities,...

10.1145/2522848.2531745 article EN 2013-11-27

We examine the role of memorization in deep learning, drawing connections to capacity, generalization, and adversarial robustness. While networks are capable memorizing noise data, our results suggest that they tend prioritize learning simple patterns first. In experiments, we expose qualitative differences gradient-based optimization neural (DNNs) on vs. real data. also demonstrate for appropriately tuned explicit regularization (e.g., dropout) can degrade DNN training performance datasets...

10.48550/arxiv.1706.05394 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Deep learning has become the state-of-art tool in many applications, but evaluation and training of deep models can be time-consuming computationally expensive. The conditional computation approach been proposed to tackle this problem (Bengio et al., 2013; Davis & Arel, 2013). It operates by selectively activating only parts network at a time. In paper, we use reinforcement as optimize policies. More specifically, cast activation-dependent policies for dropping out blocks units problem. We...

10.48550/arxiv.1511.06297 preprint EN other-oa arXiv (Cornell University) 2015-01-01

This paper is about the problem of learning a stochastic policy for generating an object (like molecular graph) from sequence actions, such that probability proportional to given positive reward object. Whereas standard return maximization tends converge single return-maximizing sequence, there are cases where we would like sample diverse set high-return solutions. These arise, example, in black-box function optimization when few rounds possible, each with large batches queries, should be...

10.48550/arxiv.2106.04399 preprint EN cc-by arXiv (Cornell University) 2021-01-01

It has been postulated that a good representation is one disentangles the underlying explanatory factors of variation. However, it remains an open question what kind training framework could potentially achieve that. Whereas most previous work focuses on static setting (e.g., with images), we postulate some causal be discovered if learner allowed to interact its environment. The agent can experiment different actions and observe their effects. More specifically, hypothesize these correspond...

10.48550/arxiv.1708.01289 preprint EN other-oa arXiv (Cornell University) 2017-01-01

It has been postulated that a good representation is one disentangles the underlying explanatory factors of variation. However, it remains an open question what kind training framework could potentially achieve that. Whereas most previous work focuses on static setting (e.g., with images), we postulate some causal be discovered if learner allowed to interact its environment. The agent can experiment different actions and observe their effects. More specifically, hypothesize these correspond...

10.48550/arxiv.1802.09484 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Finding features that disentangle the different causes of variation in real data is a difficult task, has nonetheless received considerable attention static domains like natural images. Interactive environments, which an agent can deliberately take actions, offer opportunity to tackle this task better, because experiment with actions and observe their effects. We introduce idea interactive latent factors control observed be identified by figuring out what control. propose naive method find...

10.48550/arxiv.1703.07718 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Design of de novo biological sequences with desired properties, like protein and DNA sequences, often involves an active loop several rounds molecule ideation expensive wet-lab evaluations. These experiments can consist multiple stages, increasing levels precision cost evaluation, where candidates are filtered. This makes the diversity proposed a key consideration in phase. In this work, we propose learning algorithm leveraging epistemic uncertainty estimation recently GFlowNets as generator...

10.48550/arxiv.2203.04115 preprint EN cc-by arXiv (Cornell University) 2022-01-01
Johannes Schimunek Philipp Seidl Katarina Elez Tim Hempel Tuan Anh Le and 95 more Frank Noé Simon Olsson Lluı́s Raich Robin Winter Hatice Gökcan Filipp Gusev Evgeny Gutkin Olexandr Isayev Maria Kurnikova Chamali H. Narangoda R.I. Zubatyuk Ivan P. Bosko Konstantin V. Furs Anna D. Karpenko Yury V. Kornoushenko Mikita Shuldau Artsemi Yushkevich Mohammed Benabderrahmane Patrick Bousquet‐Melou Ronan Bureau Beatrice Charton Bertrand C. Cirou Gérard Gil William J. Allen Suman Sirimulla Stanley J. Watowich Nick Antonopoulos Nikolaos Epitropakis Agamemnon Krasoulis Vassilis Pitsikalis Stavros Theodorakis Igor Kozlovskii Anton Maliutin Alexander Medvedev Petr Popov Mark Zaretckii Hamid Eghbal-zadeh Christina Halmich Sepp Hochreiter Andreas Mayr Peter Ruch Michael Widrich Francois Berenger Ashutosh Kumar Yoshihiro Yamanishi Kam Y. J. Zhang Emmanuel Bengio Yoshua Bengio Moksh Jain Maksym Korablyov Chenghao Liu Gilles Marcou Enrico Glaab Kelly K. Barnsley Suhasini M. Iyengar Mary Jo Ondrechen V. Joachim Haupt Florian Kaiser Michael Schroeder Luisa Pugliese Simone Albani Christina Athanasiou Andrea R. Beccari Paolo Carloni Giulia D’Arrigo Eleonora Gianquinto Jonas Goßen Anton Hanke Benjamin P. Joseph Daria B. Kokh Sandra Kovachka Candida Manelfi Goutam Mukherjee Abraham Muñiz‐Chicharro Francesco Musiani Ariane Nunes‐Alves Giulia Paiardi Giulia Rossetti S. Kashif Sadiq Francesca Spyrakis Carmine Talarico Alexandros Tsengenes Rebecca C. Wade Conner Copeland Jeremiah Gaiser Daniel R. Olson Amitava Roy Vishwesh Venkatraman Travis J. Wheeler Haribabu Arthanari Klara Blaschitz Marco Cespugli Vedat Durmaz Konstantin Fackeldey Patrick D. Fischer

The COVID-19 pandemic continues to pose a substantial threat human lives and is likely do so for years come. Despite the availability of vaccines, searching efficient small-molecule drugs that are widely available, including in low- middle-income countries, an ongoing challenge. In this work, we report results open science community effort, "Billion molecules against challenge", identify inhibitors SARS-CoV-2 or relevant receptors. Participating teams used wide variety computational methods...

10.1002/minf.202300262 article EN cc-by Molecular Informatics 2023-10-14

Humans interpret texts with respect to some background information, or world knowledge, and we would like develop automatic reading comprehension systems that can do the same. In this paper, introduce a task several models drive progress towards goal. particular, propose of rare entity prediction: given web document entities removed, are tasked predicting correct missing conditioned on context lexical resources. This is challenging due diversity language styles extremely large number...

10.18653/v1/d17-1086 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2017-01-01

Generative Flow Networks (GFlowNets) have been introduced as a method to sample diverse set of candidates in an active learning context, with training objective that makes them approximately proportion given reward function. In this paper, we show number additional theoretical properties GFlowNets. They can be used estimate joint probability distributions and the corresponding marginal where some variables are unspecified and, particular interest, represent over composite objects like sets...

10.48550/arxiv.2111.09266 preprint EN other-oa arXiv (Cornell University) 2021-01-01

We study the link between generalization and interference in temporal-difference (TD) learning. Interference is defined as inner product of two different gradients, representing their alignment. This quantity emerges being interest from a variety observations about neural networks, parameter sharing dynamics find that TD easily leads to low-interference, under-generalizing parameters, while effect seems reversed supervised hypothesize cause can be traced back interplay bootstrapping....

10.48550/arxiv.2003.06350 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Generative flow networks (GFlowNets) are a method for learning stochastic policy generating compositional objects, such as graphs or strings, from given unnormalized density by sequences of actions, where many possible action may lead to the same object. We find previously proposed objectives GFlowNets, matching and detailed balance, which analogous temporal difference learning, be prone inefficient credit propagation across long sequences. thus propose new objective trajectory more...

10.48550/arxiv.2201.13259 preprint EN cc-by arXiv (Cornell University) 2022-01-01

Generative flow networks (GFlowNets) are a family of algorithms for training sequential sampler discrete objects under an unnormalized target density and have been successfully used various probabilistic modeling tasks. Existing objectives GFlowNets either local to states or transitions, propagate reward signal over entire sampling trajectory. We argue that these alternatives represent opposite ends gradient bias-variance tradeoff propose way exploit this mitigate its harmful effects....

10.48550/arxiv.2209.12782 preprint EN cc-by arXiv (Cornell University) 2022-01-01

Machine-learning algorithms can be fooled by small well-designed adversarial perturbations. This is reminiscent of cellular decision-making where ligands (called antagonists) prevent correct signaling, like in early immune recognition. We draw a formal analogy between neural networks used machine learning and models (adaptive proofreading). apply attacks from to simple show explicitly the correspondence antagonism weakly bound ligands. Such absent more nonlinear models, which inspires us...

10.1103/physrevx.9.031012 article EN cc-by Physical Review X 2019-07-26

We study the problem of generating diverse candidates in context Multi-Objective Optimization. In many applications machine learning such as drug discovery and material design, goal is to generate which simultaneously optimize a set potentially conflicting objectives. Moreover, these objectives are often imperfect evaluations some underlying property interest, making it important have multiple options for expensive downstream evaluations. propose GFlowNets (MOGFNs), novel method Pareto...

10.48550/arxiv.2210.12765 preprint EN cc-by arXiv (Cornell University) 2022-01-01

Generative Flow Networks (GFlowNets; GFNs) are a family of reward/energy-based generative methods for combinatorial objects, capable generating diverse and high-utility samples. However, biasing GFNs towards producing samples is non-trivial. In this work, we leverage connections between reinforcement learning (RL) propose to combine the GFN policy with an action-value estimate, $Q$, create greedier sampling policies which can be controlled by mixing parameter. We show that several variants...

10.48550/arxiv.2402.05234 preprint EN arXiv (Cornell University) 2024-02-07

Generative Flow Networks (GFlowNets, GFNs) are a generative framework for learning unnormalized probability mass functions over discrete spaces. Since their inception, GFlowNets have proven to be useful models in applications where the majority of space is unvisited during training. This has inspired some hypothesize that GFlowNets, when paired with deep neural networks (DNNs), favourable generalization properties. In this work, we empirically verify hypothesized mechanisms GFlowNets....

10.48550/arxiv.2402.05309 preprint EN arXiv (Cornell University) 2024-02-07

Despite substantial progress in machine learning for scientific discovery recent years, truly de novo design of small molecules which exhibit a property interest remains significant challenge. We introduce LambdaZero, generative active approach to search synthesizable molecules. Powered by deep reinforcement learning, LambdaZero learns over the vast space discover candidates with desired property. apply molecular docking novel that inhibit enzyme soluble Epoxide Hydrolase 2 (sEH), while...

10.48550/arxiv.2405.01616 preprint EN arXiv (Cornell University) 2024-05-02

Diffusion models have emerged as effective distribution estimators in vision, language, and reinforcement learning, but their use priors downstream tasks poses an intractable posterior inference problem. This paper studies amortized sampling of the over data, $\mathbf{x}\sim p^{\rm post}(\mathbf{x})\propto p(\mathbf{x})r(\mathbf{x})$, a model that consists diffusion generative prior $p(\mathbf{x})$ black-box constraint or likelihood function $r(\mathbf{x})$. We state prove asymptotic...

10.48550/arxiv.2405.20971 preprint EN arXiv (Cornell University) 2024-05-31

The Generative Flow Network (GFlowNet) is a probabilistic framework in which an agent learns stochastic policy and flow functions to sample objects with probability proportional unnormalized reward function. GFlowNets share strong resemblance reinforcement learning (RL), that typically aims maximize reward, due their sequential decision-making processes. Recent works have studied connections between maximum entropy (MaxEnt) RL, modifies the standard objective of RL agents by...

10.48550/arxiv.2406.02213 preprint EN arXiv (Cornell University) 2024-06-04
Coming Soon ...