Joan Bruna

ORCID: 0000-0002-2847-1512
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Neural Networks and Applications
  • Stochastic Gradient Optimization Techniques
  • Model Reduction and Neural Networks
  • Machine Learning and Algorithms
  • Advanced Graph Neural Networks
  • Image and Signal Denoising Methods
  • Generative Adversarial Networks and Image Synthesis
  • Sparse and Compressive Sensing Techniques
  • Reinforcement Learning in Robotics
  • Domain Adaptation and Few-Shot Learning
  • 3D Shape Modeling and Analysis
  • Anomaly Detection Techniques and Applications
  • Topological and Geometric Data Analysis
  • Complex Network Analysis Techniques
  • Machine Learning in Materials Science
  • Markov Chains and Monte Carlo Methods
  • Adversarial Robustness in Machine Learning
  • Computer Graphics and Visualization Techniques
  • Image Retrieval and Classification Techniques
  • Computational Physics and Python Applications
  • Speech and Audio Processing
  • Music and Audio Processing
  • Advanced Numerical Analysis Techniques
  • Graph Theory and Algorithms
  • Advanced Vision and Imaging

Courant Institute of Mathematical Sciences
2014-2024

New York University
2014-2024

Massachusetts Institute of Technology
2023

Istituto Tecnico Industriale Alessandro Volta
2021

Weatherford College
2021

Princeton University
2020

University of Pennsylvania
2018-2019

University of California, Berkeley
2014-2017

École Polytechnique
2010-2015

Centre de Mathématiques Appliquées
2015

Many scientific fields study data with an underlying structure that is a non-Euclidean space. Some examples include social networks in computational sciences, sensor communications, functional brain imaging, regulatory genetics, and meshed surfaces computer graphics. In many applications, such geometric are large complex (in the case of networks, on scale billions), natural targets for machine learning techniques. particular, we would like to use deep neural which have recently proven be...

10.1109/msp.2017.2693418 article EN IEEE Signal Processing Magazine 2017-07-01

Convolutional Neural Networks are extremely efficient architectures in image and audio recognition tasks, thanks to their ability exploit the local translational invariance of signal classes over domain. In this paper we consider possible generalizations CNNs signals defined on more general domains without action a translation group. particular, propose two constructions, one based upon hierarchical clustering domain, another spectrum graph Laplacian. We show through experiments that for...

10.48550/arxiv.1312.6203 preprint EN other-oa arXiv (Cornell University) 2013-01-01

A wavelet scattering network computes a translation invariant image representation which is stable to deformations and preserves high-frequency information for classification. It cascades transform convolutions with nonlinear modulus averaging operators. The first layer outputs SIFT-type descriptors, whereas the next layers provide complementary that improves mathematical analysis of networks explains important properties deep convolution stationary processes incorporates higher order...

10.1109/tpami.2012.230 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2013-05-31

Deep Learning's recent successes have mostly relied on Convolutional Networks, which exploit fundamental statistical properties of images, sounds and video data: the local stationarity multi-scale compositional structure, that allows expressing long range interactions in terms shorter, localized interactions. However, there exist other important examples, such as text documents or bioinformatic data, may lack some all these strong regularities. In this paper we consider general question how...

10.48550/arxiv.1506.05163 preprint EN other-oa arXiv (Cornell University) 2015-01-01

We present techniques for speeding up the test-time evaluation of large convolutional networks, designed object recognition tasks. These models deliver impressive accuracy but each image requires millions floating point operations, making their deployment on smartphones and Internet-scale clusters problematic. The computation is dominated by convolution operations in lower layers model. exploit linear structure within filters to derive approximations that significantly reduce required...

10.48550/arxiv.1404.0736 preprint EN other-oa arXiv (Cornell University) 2014-01-01

We propose to study the problem of few-shot learning with prism inference on a partially observed graphical model, constructed from collection input images whose label can be either or not. By assimilating generic message-passing algorithms their neural-network counterparts, we define graph neural network architecture that generalizes several recently proposed models. Besides providing improved numerical performance, our framework is easily extended variants learning, such as semi-supervised...

10.48550/arxiv.1711.04043 preprint EN other-oa arXiv (Cornell University) 2017-01-01

The last decade has witnessed an experimental revolution in data science and machine learning, epitomised by deep learning methods. Indeed, many high-dimensional tasks previously thought to be beyond reach -- such as computer vision, playing Go, or protein folding are fact feasible with appropriate computational scale. Remarkably, the essence of is built from two simple algorithmic principles: first, notion representation feature whereby adapted, often hierarchical, features capture...

10.48550/arxiv.2104.13478 preprint EN other-oa arXiv (Cornell University) 2021-01-01

Inverse problems in image and audio, super-resolution particular, can be seen as high-dimensional structured prediction problems, where the goal is to characterize conditional distribution of a high-resolution output given its low-resolution corrupted observation. When scaling ratio small, point estimates achieve impressive performance, but soon they suffer from regression-to-the-mean problem, result their inability capture multi-modality this distribution. Modeling audio distributions hard...

10.48550/arxiv.1511.05666 preprint EN other-oa arXiv (Cornell University) 2015-01-01

We propose a strong baseline model for unsupervised feature learning using video data. By to predict missing frames or extrapolate future from an input sequence, the discovers both spatial and temporal correlations which are useful represent complex deformations motion patterns. The models we largely borrowed language modeling literature, adapted vision domain by quantizing space of image patches into large dictionary. demonstrate approach on filling generation task. For first time, show...

10.48550/arxiv.1412.6604 preprint EN other-oa arXiv (Cornell University) 2014-01-01

Graph neural networks (GNNs) have emerged as a powerful tool for nonlinear processing of graph signals, exhibiting success in recommender systems, power outage prediction, and motion planning, among others. GNNs consist cascade layers, each which applies convolution, followed by pointwise nonlinearity. In this work, we study the impact that changes underlying topology on output GNN. First, show are permutation equivariant, implies they effectively exploit internal symmetries topology. Then,...

10.1109/tsp.2020.3026980 article EN publisher-specific-oa IEEE Transactions on Signal Processing 2020-01-01

The reconstruction of a discrete surface from point cloud is fundamental geometry processing problem that has been studied for decades, with many methods developed. We propose the use deep neural network as geometric prior reconstruction. Specifically, we overfit representing local chart parameterization to part an input using Wasserstein distance measure approximation. By jointly fitting such networks overlapping parts cloud, while enforcing consistency condition, compute manifold atlas....

10.1109/cvpr.2019.01037 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019-06-01

Deep neural networks are highly expressive models that have recently achieved state of the art performance on speech and visual recognition tasks. While their expressiveness is reason they succeed, it also causes them to learn uninterpretable solutions could counter-intuitive properties. In this paper we report two such First, find there no distinction between individual high level units random linear combinations units, according various methods unit analysis. It suggests space, rather than...

10.48550/arxiv.1312.6199 preprint EN cc-by arXiv (Cornell University) 2013-01-01

Current state-of-the-art classification and detection algorithms train deep convolutional networks using labeled data. In this work we study unsupervised feature learning with in the context of temporally coherent unlabeled We focus on from video data, assumption that adjacent frames contain semantically similar information. This is exploited to a pooling auto-encoder regularized by slowness sparsity priors. establish connection between slow metric learning. Using define "temporal coherence"...

10.1109/iccv.2015.465 article EN 2015-12-01

A scattering vector is a local descriptor including multiscale and multi-direction co-occurrence information. It computed with cascade of wavelet decompositions complex modulus. This representation locally translation invariant linearizes deformations. supervised classification algorithm PCA model selection on vectors. State the art results are obtained for handwritten digit recognition texture classification.

10.1109/cvpr.2011.5995635 article EN 2011-06-01

The availability of large labeled datasets has allowed Convolutional Network models to achieve impressive recognition results. However, in many settings manual annotation the data is impractical; instead our noisy labels, i.e. there some freely available label for each image which may or not be accurate. In this paper, we explore performance discriminatively-trained Convnets when trained on such data. We introduce an extra noise layer into network adapts outputs match distribution....

10.48550/arxiv.1406.2080 preprint EN other-oa arXiv (Cornell University) 2014-01-01

Graph Neural Networks (GNNs) have achieved much success on graph-structured data. In light of this, there been increasing interests in studying their expressive power. One line work studies the capability GNNs to approximate permutation-invariant functions graphs, and another focuses power as tests for graph isomorphism. Our connects these two perspectives proves equivalence. We further develop a framework that incorporates both viewpoints using language sigma-algebra, through which we...

10.48550/arxiv.1905.12560 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Parameter estimation with non-Gaussian stochastic fields is a common challenge in astrophysics and cosmology. In this paper, we advocate performing task using the scattering transform, statistical tool sharing ideas convolutional neural networks (CNNs) but requiring no training nor tuning. It generates compact set of coefficients, which can be used as robust summary statistics for information. especially suited presenting localized structures hierarchical clustering, such cosmological...

10.1093/mnras/staa3165 article EN Monthly Notices of the Royal Astronomical Society 2020-10-13
Coming Soon ...