Alexander Meulemans

ORCID: 0000-0002-4858-5031
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Neural dynamics and brain function
  • Advanced Memory and Neural Computing
  • Domain Adaptation and Few-Shot Learning
  • Neural Networks and Applications
  • Reinforcement Learning in Robotics
  • Functional Brain Connectivity Studies
  • Multimodal Machine Learning Applications
  • Stochastic Gradient Optimization Techniques
  • Advanced Neural Network Applications
  • Model Reduction and Neural Networks
  • Adversarial Robustness in Machine Learning
  • Topic Modeling
  • Advanced MIMO Systems Optimization
  • Advanced Neuroimaging Techniques and Applications
  • Neural Networks and Reservoir Computing
  • Service-Oriented Architecture and Web Services
  • Antenna Design and Optimization
  • Machine Learning and Data Classification
  • Glioma Diagnosis and Treatment
  • Human Pose and Action Recognition
  • Natural Language Processing Techniques
  • Generative Adversarial Networks and Image Synthesis
  • Web Data Mining and Analysis
  • Data Quality and Management
  • Complex Systems and Decision Making

ETH Zurich
2020-2024

École Polytechnique Fédérale de Lausanne
2021

University of Zurich
2020

SIB Swiss Institute of Bioinformatics
2020

The success of deep learning, a brain-inspired form AI, has sparked interest in understanding how the brain could similarly learn across multiple layers neurons. However, majority biologically-plausible learning algorithms have not yet reached performance backpropagation (BP), nor are they built on strong theoretical foundations. Here, we analyze target propagation (TP), popular but fully understood alternative to BP, from standpoint mathematical optimization. Our theory shows that TP is...

10.48550/arxiv.2006.14331 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Recent advances in machine learning have significantly impacted the field of information extraction, with Large Language Models (LLMs) playing a pivotal role extracting structured from unstructured text. This paper explores challenges and limitations current methodologies entity extraction introduces novel approach to address these issues. We contribute by first introducing formalizing task Structured Entity Extraction (SEE), followed proposing Approximate Set OverlaP (AESOP) Metric designed...

10.48550/arxiv.2402.04437 preprint EN arXiv (Cornell University) 2024-02-06

Abstract It has been suggested that the brain employs probabilistic generative models to optimally interpret sensory information. This hypothesis formalised in distinct frameworks, focusing on explaining separate phenomena. On one hand, predictive coding theory proposed how can be learned by networks of neurons employing local synaptic plasticity. other neural sampling theories have demonstrated stochastic dynamics enable circuits represent posterior distributions latent states environment....

10.1101/2024.02.29.581455 preprint EN cc-by bioRxiv (Cold Spring Harbor Laboratory) 2024-03-02

It has been suggested that the brain employs probabilistic generative models to optimally interpret sensory information. This hypothesis formalised in distinct frameworks, focusing on explaining separate phenomena. On one hand, classic predictive coding theory proposed how can be learned by networks of neurons employing local synaptic plasticity. other neural sampling theories have demonstrated stochastic dynamics enable circuits represent posterior distributions latent states environment....

10.1371/journal.pcbi.1012532 article EN cc-by PLoS Computational Biology 2024-10-30

While a diverse collection of continual learning (CL) methods has been proposed to prevent catastrophic forgetting, thorough investigation their effectiveness for processing sequential data with recurrent neural networks (RNNs) is lacking. Here, we provide the first comprehensive evaluation established CL on variety benchmarks. Specifically, shed light particularities that arise when applying weight-importance methods, such as elastic weight consolidation, RNNs. In contrast feedforward...

10.48550/arxiv.2006.12109 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Equilibrium systems are a powerful way to express neural computations. As special cases, they include models of great current interest in both neuroscience and machine learning, such as deep networks, equilibrium recurrent models, or meta-learning. Here, we present new principle for learning with temporally- spatially-local rule. Our casts least-control problem, where first introduce an optimal controller lead the system towards solution state, then define reducing amount control needed...

10.48550/arxiv.2207.01332 preprint EN cc-by-sa arXiv (Cornell University) 2022-01-01

The success of deep learning, a brain-inspired form AI, has sparked interest in understanding how the brain could similarly learn across multiple layers neurons. However, majority biologically-plausible learning algorithms have not yet reached performance backpropagation (BP), nor are they built on strong theoretical foundations. Here, we analyze target propagation (TP), popular but fully understood alternative to BP, from standpoint mathematical optimization. Our theory shows that TP is...

10.5167/uzh-198834 article EN Neural Information Processing Systems 2020-12-12

The success of deep learning sparked interest in whether the brain learns by using similar techniques for assigning credit to each synaptic weight its contribution network output. However, majority current attempts at biologically-plausible methods are either non-local time, require highly specific connectivity motives, or have no clear link any known mathematical optimization method. Here, we introduce Deep Feedback Control (DFC), a new method that uses feedback controller drive neural...

10.48550/arxiv.2106.07887 preprint EN other-oa arXiv (Cornell University) 2021-01-01

Self-interested individuals often fail to cooperate, posing a fundamental challenge for multi-agent learning. How can we achieve cooperation among self-interested, independent learning agents? Promising recent work has shown that in certain tasks be established between learning-aware agents who model the dynamics of each other. Here, present first unbiased, higher-derivative-free policy gradient algorithm reinforcement learning, which takes into account other are themselves through trial and...

10.48550/arxiv.2410.18636 preprint EN arXiv (Cornell University) 2024-10-24

10.18653/v1/2024.emnlp-main.388 article EN Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2024-01-01

The largely successful method of training neural networks is to learn their weights using some variant stochastic gradient descent (SGD). Here, we show that the solutions found by SGD can be further improved ensembling a subset in late stages learning. At end learning, obtain back single model taking spatial average weight space. To avoid incurring increased computational costs, investigate family low-dimensional late-phase models which interact multiplicatively with remaining parameters....

10.48550/arxiv.2007.12927 preprint EN other-oa arXiv (Cornell University) 2020-01-01

The success of deep learning ignited interest in whether the brain learns hierarchical representations using gradient-based learning. However, current biologically plausible methods for credit assignment neural networks need infinitesimally small feedback signals, which is problematic realistic noisy environments and at odds with experimental evidence neuroscience showing that top-down can significantly influence activity. Building upon control (DFC), a recently proposed method, we combine...

10.48550/arxiv.2204.07249 preprint EN cc-by-sa arXiv (Cornell University) 2022-01-01

Designing reward functions for reinforcement learning is difficult: besides specifying which behavior rewarded a task, the also has to discourage undesired outcomes. Misspecified can lead unintended negative side effects, and overall unsafe behavior. To overcome this problem, recent work proposed augment specified function with an impact regularizer that discourages big on environment. Although initial results regularizers seem promising in mitigating some types of important challenges...

10.48550/arxiv.2101.12509 preprint EN cc-by-sa arXiv (Cornell University) 2021-01-01

To make reinforcement learning more sample efficient, we need better credit assignment methods that measure an action's influence on future rewards. Building upon Hindsight Credit Assignment (HCA), introduce Counterfactual Contribution Analysis (COCOA), a new family of model-based algorithms. Our algorithms achieve precise by measuring the contribution actions obtaining subsequent rewards, quantifying counterfactual query: 'Would agent still have reached this reward if it had taken another...

10.48550/arxiv.2306.16803 preprint EN cc-by-nc-sa arXiv (Cornell University) 2023-01-01
Coming Soon ...