Guillaume Desjardins

ORCID: 0000-0002-5669-6671
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Reinforcement Learning in Robotics
  • Generative Adversarial Networks and Image Synthesis
  • Model Reduction and Neural Networks
  • Neural Networks and Applications
  • Online and Blended Learning
  • Domain Adaptation and Few-Shot Learning
  • Advanced Neural Network Applications
  • Machine Learning and Algorithms
  • Online Learning and Analytics
  • Music and Audio Processing
  • Job Satisfaction and Organizational Behavior
  • Machine Learning and Data Classification
  • French Language Learning Methods
  • Evolutionary Algorithms and Applications
  • Aging, Elder Care, and Social Issues
  • Educational Tools and Methods
  • Innovative Teaching and Learning Methods
  • Gaussian Processes and Bayesian Inference
  • Multimodal Machine Learning Applications
  • Neural Networks and Reservoir Computing
  • Computational Physics and Python Applications
  • Advanced Bandit Algorithms Research
  • Management and Organizational Studies
  • Digital Media Forensic Detection
  • Information Technology and Learning

Université du Québec en Outaouais
2021-2024

Université Laval
2014-2024

Université TÉLUQ
2019-2022

Google (United States)
2014-2021

DeepMind (United Kingdom)
2014-2018

Google (United Kingdom)
2015-2018

Massachusetts Institute of Technology
2017

Australian National University
2014-2015

Université de Montréal
2009-2014

Département d'Informatique
2013

The ability to learn tasks in a sequential fashion is crucial the development of artificial intelligence. Until now neural networks have not been capable this and it has widely thought that catastrophic forgetting an inevitable feature connectionist models. We show possible overcome limitation train can maintain expertise on they experienced for long time. Our approach remembers old by selectively slowing down learning weights important those tasks. demonstrate our scalable effective solving...

10.1073/pnas.1611835114 article EN Proceedings of the National Academy of Sciences 2017-03-14
The Theano Development Team Rami Al‐Rfou Guillaume Alain Amjad Almahairi Christof Angermueller and 95 more Dzmitry Bahdanau Nicolas Ballas Frédéric Bastien Justin Bayer Anatoly Belikov Alexander Belopolsky Yoshua Bengio Arnaud Bergeron James Bergstra Valentin Bisson Josh Bleecher Snyder Nicolas Bouchard Nicolas Boulanger-Lewandowski Xavier Bouthillier Alexandre de Brébisson Olivier Breuleux Pierre-Luc Carrier Kyunghyun Cho Jan Chorowski Paul Christiano Tim Cooijmans Marc-Alexandre Côté Myriam Côté Aaron Courville Yann Dauphin Olivier Delalleau Julien Demouth Guillaume Desjardins Sander Dieleman Laurent Dinh Mélanie Ducoffe Vincent Dumoulin Samira Ebrahimi Kahou Dumitru Erhan Ziye Fan Orhan Fırat Mathieu Germain Xavier Glorot Ian Goodfellow M. Graham Çağlar Gülçehre Philippe Hamel Iban Harlouchet Jean-Philippe Heng Balázs Hidasi Sina Honari Arjun Jain Sébastien Jean Kai Jia Mikhail Korobov Vivek Kulkarni Alex Lamb Pascal Lamblin Eric Larsen César Laurent Sean Lee Simon Lefrançois Simon Lemieux Nicholas Léonard Zhouhan Lin Jesse A. Livezey Cory Lorenz Jeremiah Lowin Qianli Ma Pierre-Antoine Manzagol Olivier Mastropietro Robert T. McGibbon Roland Memisevic Bart van Merriënboer Vincent Michalski Mehdi Mirza Alberto Orlandi Christopher Pal Razvan Pascanu Mohammad Pezeshki Colin Raffel Daniel Renshaw Matthew Rocklin Adriana Romero M. Roth Peter Sadowski John Salvatier François Savard Jan Schlüter John Schulman Gabriel Schwartz Iulian Vlad Serban Dmitriy Serdyuk Samira Shabanian Étienne Simon Sigurd Spieckermann Siva Subramanyam Jakub Sygnowski Jérémie Tanguay Gijs van Tulder

Theano is a Python library that allows to define, optimize, and evaluate mathematical expressions involving multi-dimensional arrays efficiently. Since its introduction, it has been one of the most used CPU GPU compilers - especially in machine learning community shown steady performance improvements. being actively continuously developed since 2008, multiple frameworks have built on top produce many state-of-the-art models. The present article structured as follows. Section I provides an...

10.48550/arxiv.1605.02688 preprint EN other-oa arXiv (Cornell University) 2016-01-01

Learning to solve complex sequences of tasks--while both leveraging transfer and avoiding catastrophic forgetting--remains a key obstacle achieving human-level intelligence. The progressive networks approach represents step forward in this direction: they are immune forgetting can leverage prior knowledge via lateral connections previously learned features. We evaluate architecture extensively on wide variety reinforcement learning tasks (Atari 3D maze games), show that it outperforms common...

10.48550/arxiv.1606.04671 preprint EN cc-by arXiv (Cornell University) 2016-01-01

Theano is a compiler for mathematical expressions in Python that combines the convenience of NumPy's syntax with speed optimized native machine language. The user composes high-level description mimics and semantics, while being statically typed functional (as opposed to imperative). These allow provide symbolic differentiation. Before performing computation, optimizes choice expressions, translates them into C++ (or CUDA GPU), compiles dynamically loaded modules, all automatically. Common...

10.25080/majora-92bf1922-003 article EN cc-by Proceedings of the Python in Science Conferences 2010-01-01

In this paper we present the techniques used for University of Montréal's team submissions to 2013 Emotion Recognition in Wild Challenge. The challenge is classify emotions expressed by primary human subject short video clips extracted from feature length movies. This involves analysis acted scenes lasting approximately one-two seconds, including audio track which may contain voices as well background music. Our approach combines multiple deep neural networks different data modalities,...

10.1145/2522848.2531745 article EN 2013-11-27

We present new intuitions and theoretical assessments of the emergence disentangled representation in variational autoencoders. Taking a rate-distortion theory perspective, we show circumstances under which representations aligned with underlying generative factors variation data emerge when optimising modified ELBO bound $β$-VAE, as training progresses. From these insights, propose modification to regime that progressively increases information capacity latent code during training. This...

10.48550/arxiv.1804.03599 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Here we propose a novel model family with the objective of learning to disentangle factors variation in data. Our approach is based on spike-and-slab restricted Boltzmann machine which generalize include higher-order interactions among multiple latent variables. Seen from generative perspective, multiplicative emulates entangling variation. Inference can be seen as disentangling these factors. Unlike previous attempts at factors, proposed trained using no supervised information regarding We...

10.48550/arxiv.1210.5474 preprint EN other-oa arXiv (Cornell University) 2012-01-01

We introduce Natural Neural Networks, a novel family of algorithms that speed up convergence by adapting their internal representation during training to improve conditioning the Fisher matrix. In particular, we show specific example employs simple and efficient reparametrization neural network weights implicitly whitening obtained at each layer, while preserving feed-forward computation network. Such networks can be trained efficiently via proposed Projected Gradient Descent algorithm...

10.48550/arxiv.1507.00210 preprint EN other-oa arXiv (Cornell University) 2015-01-01
Agnieszka Wojtczuk‐Turek Dariusz Turek Fiona Edgar Howard J. Klein Janine Bosak and 95 more Belgin Okay‐Somerville Na Fu Sabine Raeder Paweł Jurek Anna Lupina‐Wegener Zuzana Dvořáková Francisca Gutiérrez Crocco Aleksandra Kekkonen Pedro I. Leiva Lenka Mynaříková Mercedes Sánchez‐Apellániz Imran Shafique Bassam Samir Al‐Romeedy Serena Wee Patrick D. Dunlop Florence Stinglhamber Gaëtane Caesens Adriana Cristina Ferreira Caldana Marina Greghi Sticca Valentin Vasilev Martin Lauzier Guillaume Desjardins Gangfeng Zhang Le Tan Lady Brigitte Galvez‐Sierra Érico Rentería Pérez Šrečko Goić Ivana Tadić D. Charvátová Marek Botek Dorthe Høj Jensen Dayamy Lima Rojas Segundo Gonzalo Pazmay Ramos Piret Masso Maria Järlström Nicolas Gillet Tiphaine Huyghebaert‐Zouaghi Maia Robakidze Khatuna Martskvishvili Angela Rachael Dorrough Marc Jekel Carolin Häffner A. Timur Sevincer Elias Kodjo Kekesi Collins Badu Agyemang Eleni Apospori Jerin Jose Alice Salendu Arum Etikariena Harry Susianto Bertina Sjabadhyni Shera Malayeri Masoumeh Seydi Mary Kinahan Alon Lisak Marco Giovanni Mariani Marco Salvati Silvia Moscatelli Eleonora Crapolicchio Claudia Manzi Akihito Shimazu Hiroshi Ikeda Rita Žukauskienė Goda Kaniušonytė Gottfried Catania Mary Anne Lauri Sergio Madero Denise Fernando Klaske Veth Sandesh Dhakal Nataliya Podgorodnichenko Abiodun Musbau Lawal Marius Duhović Hafstad Ana Inés Reátegui Vela Oswaldo Morales Divina M. Edralin Susana Schmitz Joana Neto Félix Neto Boris Popov Jasna Milošević Đorđević Vladimir Mihić Anna Kalistová Ivana Piterová Claude–Hélène Mayer María José Charlo Molina Ruwan Ranasinghe Tesora Nakchedi‐Ooft Rosita Sobhie Mösli Matteo Jennifer Chavanovanich Narumol Petchthip Serdar Karabatı Gülçin Akbaş Beril Türkoğlu

Abstract Sustainable human resource management is gaining importance in organizations due to its role developing a sustainable work environment and well‐being. This paper discusses the relationship between employee perceptions of job satisfaction 54 countries. We propose that HRM positively associated with but this moderated by employees' identification organization country‐level individualism–collectivism. Thus, we suggest national culture functions as second‐level moderator organizational...

10.1002/csr.2815 article EN Corporate Social Responsibility and Environmental Management 2024-05-10

Policies for complex visual tasks have been successfully learned with deep reinforcement learning, using an approach called Q-networks (DQN), but relatively large (task-specific) networks and extensive training are needed to achieve good performance. In this work, we present a novel method policy distillation that can be used extract the of learning agent train new network performs at expert level while being dramatically smaller more efficient. Furthermore, same consolidate multiple...

10.48550/arxiv.1511.06295 preprint EN other-oa arXiv (Cornell University) 2015-01-01

Purpose Ichak Adizes has developed original and practical conceptions of executive interaction, change management corporate development, collectively referred to as “symbergetic organisational therapy”. Although his name is celebrated in some circles, it not widely known within mainstream academia. Further, Adizes’ insights into what organisations are how they achieve optimal performance routinely dealt with Western business schools. After exposing ideas, this paper aims investigate reasons...

10.1108/jmh-09-2024-0135 article EN Journal of Management History 2025-04-21

The spike-and-slab restricted Boltzmann machine (ssRBM) is defined to have both a real-valued "slab" variable and binary "spike" associated with each unit in the hidden layer. model uses its slab variables conditional covariance of observation-thought be important capturing statistical properties natural images. In this paper, we present canonical ssRBM framework together some extensions. These extensions highlight flexibility RBM as platform for exploring more sophisticated probabilistic...

10.1109/tpami.2013.238 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2014-01-31

We introduce RecurrentGemma, an open language model which uses Google's novel Griffin architecture. combines linear recurrences with local attention to achieve excellent performance on language. It has a fixed-sized state, reduces memory use and enables efficient inference long sequences. provide pre-trained 2B non-embedding parameters, instruction tuned variant. Both models comparable Gemma-2B despite being trained fewer tokens.

10.48550/arxiv.2404.07839 preprint EN arXiv (Cornell University) 2024-04-11
Coming Soon ...