Farzan Farnia

ORCID: 0000-0002-6049-9232
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Adversarial Robustness in Machine Learning
  • Sparse and Compressive Sensing Techniques
  • Neural Networks and Applications
  • Stochastic Gradient Optimization Techniques
  • Advanced MIMO Systems Optimization
  • Statistical Methods and Inference
  • Generative Adversarial Networks and Image Synthesis
  • Bayesian Methods and Mixture Models
  • Anomaly Detection Techniques and Applications
  • Domain Adaptation and Few-Shot Learning
  • Cooperative Communication and Network Coding
  • Machine Learning and Algorithms
  • Energy Harvesting in Wireless Networks
  • Gaussian Processes and Bayesian Inference
  • Distributed Sensor Networks and Detection Algorithms
  • Reinforcement Learning in Robotics
  • Bayesian Modeling and Causal Inference
  • Privacy-Preserving Technologies in Data
  • Explainable Artificial Intelligence (XAI)
  • Advanced Neural Network Applications
  • Model Reduction and Neural Networks
  • 3D Shape Modeling and Analysis
  • Wireless Communication Security Techniques
  • Advanced SAR Imaging Techniques
  • Handwritten Text Recognition Techniques

Chinese University of Hong Kong
2022-2024

Massachusetts Institute of Technology
2020-2021

Moscow Institute of Thermal Technology
2020-2021

Stanford University
2014-2020

Decision Systems (United States)
2020

Sharif University of Technology
2013

We consider an energy-harvesting communication system where a transmitter powered by exogenous energy arrival process and equipped with finite battery of size B <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">max</sub> communicates over discrete-time AWGN channel. first concentrate on simple Bernoulli at each time step, either packet E is harvested probability p, or no all, independent the other steps. provide near optimal control policy...

10.1109/jsac.2015.2391611 article EN IEEE Journal on Selected Areas in Communications 2015-01-14

Federated learning is a distributed paradigm that aims at training models using samples across multiple users in network while keeping the on users' devices with aim of efficiency and protecting privacy. In such settings, data often statistically heterogeneous manifests various distribution shifts users, which degrades performance learnt model. The primary goal this paper to develop robust federated algorithm achieves satisfactory against samples. To achieve goal, we first consider...

10.48550/arxiv.2006.08907 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Diffusion-based Neural Combinatorial Optimization (NCO) has demonstrated effectiveness in solving NP-complete (NPC) problems by learning discrete diffusion models for solution generation, eliminating hand-crafted domain knowledge. Despite their success, existing NCO methods face significant challenges both cross-scale and cross-problem generalization, high training costs compared to traditional solvers. While recent studies have introduced training-free guidance approaches that leverage...

10.48550/arxiv.2502.12188 preprint EN arXiv (Cornell University) 2025-02-15

Deep neural networks (DNNs) have set benchmarks on a wide array of supervised learning tasks. Trained DNNs, however, often lack robustness to minor adversarial perturbations the input, which undermines their true practicality. Recent works increased DNNs by fitting using adversarially-perturbed training samples, but improved performance can still be far below seen in non-adversarial settings. A significant portion this gap attributed decrease generalization due training. In work, we extend...

10.48550/arxiv.1811.07457 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Given a task of predicting $Y$ from $X$, loss function $L$, and set probability distributions $Γ$ on $(X,Y)$, what is the optimal decision rule minimizing worst-case expected over $Γ$? In this paper, we address question by introducing generalization principle maximum entropy. Applying to sets with marginal $X$ constrained be empirical data, develop general minimax approach for supervised learning problems. While some functions such as squared-error log loss, rederives well-knwon regression...

10.48550/arxiv.1606.02206 preprint EN other-oa arXiv (Cornell University) 2016-01-01

.Generative adversarial networks (GANs) learn the distribution of observed samples through a zero-sum game between two machine players, generator and discriminator. While GANs achieve great success in learning complex image, sound, text data, they perform suboptimally multimodal distribution-learning benchmarks such as Gaussian mixture models (GMMs). In this paper, we propose Generative Adversarial Training for Mixture Models (GAT-GMM), minimax GAN framework GMMs. Motivated by optimal...

10.1137/21m1445831 article EN SIAM Journal on Mathematics of Data Science 2023-03-10

Generative Adversarial Networks (GANs) have become a popular method to learn probability model from data. In this paper, we provide an understanding of basic issues surrounding GANs including their formulation, generalization and stability on simple LQG benchmark where the generator is Linear, discriminator Quadratic data has high-dimensional Gaussian distribution. Even in benchmark, GAN problem not been well-understood as observe that existing state-of-the-art architectures may fail proper...

10.1109/jsait.2020.2991375 article EN publisher-specific-oa IEEE Journal on Selected Areas in Information Theory 2020-04-29

Generative adversarial networks (GANs) represent a zero-sum game between two machine players, generator and discriminator, designed to learn the distribution of data. While GANs have achieved state-of-the-art performance in several benchmark learning tasks, GAN minimax optimization still poses great theoretical empirical challenges. trained using first-order methods commonly fail converge stable solution where players cannot improve their objective, i.e., Nash equilibrium underlying game....

10.48550/arxiv.2002.09124 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Federated learning is a distributed machine paradigm, which aims to train model using the local data of many clients. A key challenge in federated that samples across clients may not be identically distributed. To address this challenge, personalized with goal tailoring learned distribution every individual client has been proposed. In paper, we focus on problem and propose novel Learning scheme based Optimal Transport (FedOT) as algorithm learns optimal transport maps for transferring...

10.1109/jsait.2022.3182355 article EN publisher-specific-oa IEEE Journal on Selected Areas in Information Theory 2022-06-01

Generative Adversarial Networks (GANs) have become a popular method to learn probability model from data. In this paper, we aim provide an understanding of some the basic issues surrounding GANs including their formulation, generalization and stability on simple benchmark where data has high-dimensional Gaussian distribution. Even in benchmark, GAN problem not been well-understood as observe that existing state-of-the-art architectures may fail proper generative distribution owing (1) (i.e.,...

10.48550/arxiv.1710.10793 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Generative adversarial network (GAN) is a minimax game between generator mimicking the true model and discriminator distinguishing samples produced by from real training samples. Given an unconstrained able to approximate any function, this reduces finding generative minimizing divergence measure, e.g. Jensen-Shannon (JS) divergence, data distribution. However, in practice constrained be smaller class $\mathcal{F}$ such as neural nets. Then, natural question how minimization interpretation...

10.48550/arxiv.1810.11740 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Interpreting neural network classifiers using gradient-based saliency maps has been extensively studied in the deep learning literature. While existing algorithms manage to achieve satisfactory performance application standard image recognition datasets, recent works demonstrate vulnerability of widely-used interpretation schemes norm-bounded perturbations adversarially designed for every individual input sample. However, such adversarial are commonly knowledge an sample, and hence perform...

10.1109/icassp49357.2023.10097188 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2023-05-05

Deep generative models dominate the existing literature in layout pattern generation. However, leaving guarantee of legality to an inexplicable neural network could be problematic several applications. In this paper, we propose DiffPattern generate reliable patterns. introduces a novel diverse topology generation method via discrete diffusion model with compute-efficiently lossless representation. Then white-box assessment is utilized legal patterns given desired design rules. Our...

10.1109/dac56929.2023.10248009 article EN 2023-07-09

Given low order moment information over the random variables X = (X <inf xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">1</inf> , xmlns:xlink="http://www.w3.org/1999/xlink">2</inf> …, xmlns:xlink="http://www.w3.org/1999/xlink">p</inf> ) and Y, what distribution minimizes Hirschfeld-Gebelein-Rényi (HGR) maximal correlation coefficient between while remains faithful to given moments? The answer this question is important especially in fit models (X,...

10.1109/isit.2015.7282681 article EN 2022 IEEE International Symposium on Information Theory (ISIT) 2015-06-01

We consider an energy-harvesting communication system where a transmitter powered by exogenous energy arrival process and equipped with finite battery of size $B_{max}$ communicates over discrete-time AWGN channel. first concentrate on simple Bernoulli at each time step, either packet $E$ is harvested probability $p$, or no all, independent the other steps. provide near optimal control policy approximation to information-theoretic capacity this Our approximations for both problems are...

10.48550/arxiv.1405.1156 preprint EN other-oa arXiv (Cornell University) 2014-01-01

The success of deep neural networks stems from their ability to generalize well on real data; however, et al. have observed that can easily overfit randomly-generated labels. This observation highlights the following question: why do gradient methods succeed in finding generalizable solutions for while there exist with poor generalization behavior? In this work, we use a Fourier-based approach study properties gradient-based over 2-layer band-limited activation functions. Our results...

10.1109/jsait.2020.2983192 article EN publisher-specific-oa IEEE Journal on Selected Areas in Information Theory 2020-03-26
Coming Soon ...