Arthur Gretton

ORCID: 0000-0003-3169-7624
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Statistical Methods and Inference
  • Gaussian Processes and Bayesian Inference
  • Neural Networks and Applications
  • Bayesian Methods and Mixture Models
  • Domain Adaptation and Few-Shot Learning
  • Bayesian Modeling and Causal Inference
  • Advanced Statistical Methods and Models
  • Statistical Methods and Bayesian Inference
  • Machine Learning and Algorithms
  • Blind Source Separation Techniques
  • Face and Expression Recognition
  • Machine Learning and Data Classification
  • Markov Chains and Monte Carlo Methods
  • Advanced Causal Inference Techniques
  • Generative Adversarial Networks and Image Synthesis
  • Anomaly Detection Techniques and Applications
  • Model Reduction and Neural Networks
  • Control Systems and Identification
  • Sparse and Compressive Sensing Techniques
  • Fault Detection and Control Systems
  • Neural dynamics and brain function
  • Adversarial Robustness in Machine Learning
  • Statistical Distribution Estimation and Applications
  • Spectroscopy and Chemometric Analyses
  • Time Series Analysis and Forecasting

University College London
2014-2023

Oxford Centre for Computational Neuroscience
2014-2023

Artificial Intelligence in Medicine (Canada)
2022

Université de Toulouse
2021

Centre National de la Recherche Scientifique
2021

Yonsei University
2021

Institut de Mathématiques de Toulouse
2021

Gatsby Charitable Foundation
2013-2020

Google (United States)
2020

Sainsbury Wellcome Centre
2016

We propose a framework for analyzing and comparing distributions, which we use to construct statistical tests determine if two samples are drawn from different distributions. Our test statistic is the largest difference in expectations over functions unit ball of reproducing kernel Hilbert space (RKHS), called maximum mean discrepancy (MMD).We present distribution free based on large deviation bounds MMD, third asymptotic this statistic. The MMD can be computed quadratic time, although...

10.5555/2188385.2188410 article EN Journal of Machine Learning Research 2012-03-01

Abstract Motivation: Many problems in data integration bioinformatics can be posed as one common question: Are two sets of observations generated by the same distribution? We propose a kernel-based statistical test for this problem, based on fact that distributions are different if and only there exists at least function having expectation distributions. Consequently we use maximum discrepancy between means basis statistic. The Maximum Mean Discrepancy (MMD) take advantage kernel trick,...

10.1093/bioinformatics/btl242 article EN Bioinformatics 2006-07-15

A Hilbert space embedding for probability measures has recently been proposed, with applications including dimensionality reduction, homogeneity testing, and independence testing. This represents any measure as a mean element in reproducing kernel (RKHS). pseudometric on the of can be defined distance between distribution embeddings: we denote this γk, indexed by function k that defines inner product RKHS. We present three theoretical properties γk. First, consider question determining...

10.5555/1756006.1859901 article EN Journal of Machine Learning Research 2010-03-01

Local field potentials (LFPs) reflect subthreshold integrative processes that complement spike train measures. However, little is yet known about the differences between how LFPs and spikes encode rich naturalistic sensory stimuli. We addressed this question by recording from primary visual cortex of anesthetized macaques while presenting a color movie. then determined power at different frequencies represents features in found most informative LFP frequency ranges were 1–8 60–100 Hz. range...

10.1523/jneurosci.0009-08.2008 article EN cc-by-nc-sa Journal of Neuroscience 2008-05-28

We provide a unifying framework linking two classes of statistics used in two-sample and independence testing: on the one hand, energy distances distance covariances from literature; other, maximum mean discrepancies (MMD), that is, between embeddings distributions to reproducing kernel Hilbert spaces (RKHS), as established machine learning. In case where is computed with semimetric negative type, positive definite kernel, termed may be defined such MMD corresponds exactly distance....

10.1214/13-aos1140 article EN other-oa The Annals of Statistics 2013-10-01

We introduce a framework for filtering features that employs the Hilbert-Schmidt Independence Criterion (HSIC) as measure of dependence between and labels.The key idea is good should maximise such dependence.Feature selection various supervised learning problems (including classification regression) unified under this framework, solutions can be approximated using backward-elimination algorithm.We demonstrate usefulness our method on both artificial real world datasets.

10.1145/1273496.1273600 article EN 2007-06-20

Given two probability measures, $\mathbb{P}$ and $\mathbb{Q}$ defined on a measurable space, $S$, the integral metric (IPM) is as $$\gamma_{\mathcal{F}}(\mathbb{P},\mathbb{Q})=\sup\left\{\left\vert \int_{S}f\,d\mathbb{P}-\int_{S}f\,d\mathbb{Q}\right\vert\,:\,f\in\mathcal{F}\right\},$$ where $\mathcal{F}$ class of real-valued bounded functions $S$. By appropriately choosing $\mathcal{F}$, various popular distances between $\mathbb{Q}$, including Kantorovich metric, Fortet-Mourier dual-bounded...

10.1214/12-ejs722 article EN cc-by Electronic Journal of Statistics 2012-01-01

We propose a framework for analyzing and comparing distributions, allowing us to design statistical tests determine if two samples are drawn from different distributions. Our test statistic is the largest difference in expectations over functions unit ball of reproducing kernel Hilbert space (RKHS). present based on large deviation bounds statistic, while third asymptotic distribution this statistic. The can be computed quadratic time, although efficient linear time approximations available....

10.48550/arxiv.0805.2368 preprint EN other-oa arXiv (Cornell University) 2008-01-01

While kernel canonical correlation analysis (CCA) has been applied in many contexts, the convergence of finite sample estimates associated functions to their population counterparts not yet established. This paper gives a mathematical proof statistical CCA, providing theoretical justification for method. The uses covariance operators defined on reproducing Hilbert spaces, and analyzes empirical rank counterparts, which can have infinite rank. result also sufficient condition regularization...

10.5555/1248659.1248673 article EN Journal of Machine Learning Research 2007-05-01

We investigated whether it is possible to infer spike trains solely on the basis of underlying local field potentials (LFPs). Using support vector machines and linear regression models, we found that in primary visual cortex (V1) monkeys, spikes can indeed be inferred from LFPs, at least with moderate success. Although there a considerable degree variation across electrodes, low-frequency structure (in 100-ms range) reasonable accuracy, whereas exact positions are not reliably predicted. Two...

10.1152/jn.00919.2007 article EN Journal of Neurophysiology 2007-12-26

Many modern applications of signal processing and machine learning, ranging from computer vision to computational biology, require the analysis large volumes high-dimensional continuous-valued measurements. Complex statistical features are commonplace, including multimodality, skewness, rich dependency structures. Such problems call for a flexible robust modeling framework that can take into account these diverse features. Most existing approaches, graphical models, rely heavily on...

10.1109/msp.2013.2252713 article EN IEEE Signal Processing Magazine 2013-06-12
Coming Soon ...