Mark Bun

ORCID: 0000-0002-5855-8045
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Cryptography and Data Security
  • Privacy-Preserving Technologies in Data
  • Complexity and Algorithms in Graphs
  • Machine Learning and Algorithms
  • Quantum Computing Algorithms and Architecture
  • Privacy, Security, and Data Protection
  • Stochastic Gradient Optimization Techniques
  • Internet Traffic Analysis and Secure E-voting
  • Computability, Logic, AI Algorithms
  • Numerical Methods and Algorithms
  • Mobile Crowdsensing and Crowdsourcing
  • semigroups and automata theory
  • Coding theory and cryptography
  • HIV, Drug Use, Sexual Risk
  • Advanced Optimization Algorithms Research
  • Adversarial Robustness in Machine Learning
  • Advanced Bandit Algorithms Research
  • Low-power high-performance VLSI design
  • Cryptography and Residue Arithmetic
  • Quantum Information and Cryptography
  • Advanced Graph Theory Research
  • Probability and Risk Models
  • Access Control and Trust
  • Auction Theory and Applications
  • Age of Information Optimization

Boston University
2016-2024

Simons Foundation
2018-2021

Princeton University
2017-2020

Siberian Branch of the Russian Academy of Sciences
2019

Limnological Institute
2019

Harvard University
2013-2018

Harvard University Press
2013-2016

We propose truncated concentrated differential privacy (tCDP), a refinement of and privacy. This new definition provides robust efficient composition guarantees, supports powerful algorithmic techniques such as amplification via sub-sampling, enables more accurate statistical analyses. In particular, we show central task for which the exponential accuracy improvement.

10.1145/3188745.3188946 article EN 2018-06-20

Differential privacy is a formal mathematical framework for quantifying and managing risks. It provides provable protection against wide range of potential attacks, including those currently unforeseen. primarily studied in the context collection, analysis, release aggregate statistics. These from simple statistical estimations, such as averages, to machine learning. Tools differentially private analysis are now early stages implementation use across variety academic, industry, government...

10.2139/ssrn.3338027 article EN SSRN Electronic Journal 2018-01-01

We prove new upper and lower bounds on the sample complexity of (ε, δ) differentially private algorithms for releasing approximate answers to threshold functions. A function c over a totally ordered domain X evaluates <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">z</sub> (y) = 1 if y ≤ x, 0 otherwise. give first nontrivial bound thresholds with differential privacy, showing that task is impossible an infinite X, moreover requires n ≥ Ω(log*...

10.1109/focs.2015.45 preprint EN 2015-10-01

We show new lower bounds on the sample complexity of (ε, δ)-differentially private algorithms that accurately answer large sets counting queries. A query a database D ∈ ({0, 1}d)n has form "What fraction individual records in satisfy property q?" order to an arbitrary set Q » nd queries within error ±α it is necessary

10.1145/2591796.2591877 preprint EN 2014-05-31

We present a new locally differentially private algorithm for the heavy hitters problem which achieves optimal worst-case error as function of all standardly considered parameters. Prior work obtained rates depend optimally on number users, size domain, and privacy parameter, but sub-optimally failure probability. strengthen existing lower bounds to incorporate probability, show that our upper bound is tight with respect this parameter well. Our based understanding structure protocols....

10.1145/3196959.3196981 article EN 2018-05-15

We present a new locally differentially private algorithm for the heavy hitters problem that achieves optimal worst-case error as function of all standardly considered parameters. Prior work obtained rates depend optimally on number users, size domain, and privacy parameter but sub-optimally failure probability. strengthen existing lower bounds to incorporate probability show our upper bound is tight with respect this well. Our based understanding structure protocols. further develop these...

10.1145/3344722 article EN ACM Transactions on Algorithms 2019-10-04

We investigate the {\em direct-sum} problem in context of differentially private PAC learning: What is sample complexity solving k learning tasks simultaneously under differential privacy, and how does this cost compare to that without privacy? In our setting, an individual example consists a domain element x labeled by unknown concepts (c1,...,ck). The goal multi-learner output hypotheses (h1,...,hk) generalize input examples.

10.1145/2840728.2840747 preprint EN 2016-01-05

We show new information-theoretic lower bounds on the sample complexity of $(\varepsilon, \delta)$-differentially private algorithms that accurately answer large sets counting queries. A query a database $D \in (\{0,1\}^d)^n$ has form "What fraction individual records in satisfy property $q$?" order to an arbitrary set $\mathcal{Q}$ $\gg d/\alpha^2$ queries $D$ within error $\pm \alpha$ it is necessary $ n \geq \tilde{\Omega}({\sqrt{d} \log |\mathcal{Q}|}/{\alpha^2 \varepsilon}). This bound...

10.1137/15m1033587 article EN SIAM Journal on Computing 2018-01-01

The approximate degree of a Boolean function f is the least real polynomial that approximates pointwise to error at most 1/3. Approximate known be lower bound on quantum query complexity. We resolve or nearly and complexities following basic functions: $\bullet$ $k$-distinctness: For any constant $k$, complexity $k$-distinctness $\Omega(n^{3/4-1/(2k)})$. This tight for large $k$ (Belovs, FOCS 2012). Image size testing: testing image $[n] \to [n]$ $\tilde{\Omega}(n^{1/2})$. proves conjecture...

10.1145/3188745.3188784 preprint EN 2018-06-20

We prove that every concept class with finite Littlestone dimension can be learned by an (approximate) differentially-private algorithm. This answers open question of Alon et al. (STOC 2019) who proved the converse statement (this was also asked Neel (FOCS 2019)). Together these two results yield equivalence between online learnability and private PAC learnability. introduce a new notion algorithmic stability called "global stability" which is essential to our proof may independent interest....

10.1109/focs46700.2020.00044 article EN 2020-11-01

The technical literature about data privacy largely consists of two complementary approaches: formal definitions conditions sufficient for preservation and attacks that demonstrate breaches. Differential is an accepted standard in the former sphere. However, differential privacy's powerful adversarial model worst-case guarantees may make it too stringent some situations, especially when achieving comes at a significant cost to utility. Meanwhile, aim expose real worrying risks associated...

10.48550/arxiv.2502.02709 preprint EN arXiv (Cornell University) 2025-02-04

We provide a differentially private algorithm for hypothesis selection. Given samples from an unknown probability distribution $P$ and set of $m$ distributions $\mathcal{H}$, the goal is to output, in $\varepsilon$-differentially manner, $\mathcal{H}$ whose total variation distance comparable that best such (which we denote by $\alpha$). The sample complexity our basic $O\left(\frac{\log m}{\alpha^2} + \frac{\log m}{\alpha \varepsilon}\right)$, representing minimal cost privacy when compared...

10.48550/arxiv.1905.13229 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Modern machine learning models are complex and frequently encode surprising amounts of information about individual inputs. In extreme cases, appear to memorize entire input examples, including seemingly irrelevant (social security numbers from text, for example). this paper, we aim understand whether sort memorization is necessary accurate learning. We describe natural prediction problems in which every sufficiently training algorithm must encode, the model, essentially all a large subset...

10.1145/3406325.3451131 preprint EN 2021-06-15

10.1016/j.ic.2014.12.003 article EN publisher-specific-oa Information and Computation 2014-12-12

The simplest and most widely applied method for guaranteeing differential privacy is to add instance-independent noise a statistic of interest that scaled its global sensitivity. However, sensitivity worst-case notion often too conservative realized dataset instances. We provide methods scaling in an instance-dependent way demonstrate they greater accuracy under average-case distributional assumptions. Specifically, we consider the basic problem privately estimating mean real distribution...

10.48550/arxiv.1906.02830 preprint EN other-oa arXiv (Cornell University) 2019-01-01

The approximate degree of a Boolean function f : {-1, 1} <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">n</sup> → is the least real polynomial that approximates pointwise to error at most 1/3. We introduce generic method for increasing given function, while preserving its computability by constant-depth circuits. Specifically, we show how transform any with d into F on O(n · polylog(n)) variables D = Ω(n...

10.1109/focs.2017.10 preprint EN 2017-10-01

The sign-rank of a matrix A with entries in {-1, +1} is the least rank real B A_{ij}*B_{ij} > 0 for all i, j. Razborov and Sherstov (2008) gave first exponential lower bounds on function AC^0, answering an old question Babai, Frankl, Simon (1986). Specifically, they exhibited = [F(x,y)]_{x,y} specific F:{-1,1}^n*{-1,1}^n -> {-1,1} such that has exp(Omega(n^{1/3}). We prove generalization Sherstov’s result, yielding non-trivial class functions (that includes used by Sherstov). As corollary...

10.4230/lipics.icalp.2016.37 article EN International Colloquium on Automata, Languages and Programming 2016-01-01
Coming Soon ...