Deepanshu Vasal

ORCID: 0000-0003-1089-8080
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Game Theory and Applications
  • Economic theories and models
  • Auction Theory and Applications
  • Opinion Dynamics and Social Influence
  • Complex Network Analysis Techniques
  • Game Theory and Voting Systems
  • Advanced Bandit Algorithms Research
  • Experimental Behavioral Economics Studies
  • Economic Policies and Impacts
  • Reinforcement Learning in Robotics
  • Complex Systems and Time Series Analysis
  • Optimization and Search Problems
  • Wireless Communication Security Techniques
  • Markov Chains and Monte Carlo Methods
  • Infrastructure Resilience and Vulnerability Analysis
  • Stochastic processes and financial applications
  • Error Correcting Code Techniques
  • Cooperative Communication and Network Coding
  • Simulation Techniques and Applications
  • Adaptive Dynamic Programming Control
  • Distributed Control Multi-Agent Systems
  • Evolutionary Game Theory and Cooperation
  • Advanced MIMO Systems Optimization
  • Smart Grid Energy Management
  • DNA and Biological Computing

Northwestern University
2020-2024

The University of Texas at Austin
2016-2022

University of Michigan
2012-2021

Technical University of Munich
2019

We consider both finite-horizon and infinite-horizon versions of a dynamic game with <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><tex-math notation="LaTeX">$N$</tex-math></inline-formula> selfish players who observe their types privately take actions that are publicly observed. Players' evolve as conditionally independent Markov processes, conditioned on current actions. Their jointly determine instantaneous rewards. In games...

10.1109/tac.2018.2809863 article EN IEEE Transactions on Automatic Control 2018-02-26

We study the problem of Bayesian learning in a dynamical system involving strategic agents with asymmetric information. In series seminal papers literature, this has been studied under simplifying model where selfish players appear sequentially and act once game, based on private noisy observations state public observation past players' actions. It is shown that there exist information cascades users discard their mimic action predecessor. paper, we provide framework for studying dynamics...

10.1109/allerton.2016.7852239 article EN 2016-09-01

We consider a finite horizon dynamic game with two players who observe their types privately and take actions, which are publicly observed. Players' evolve as independent, controlled linear Gaussian processes incur quadratic instantaneous costs. This forms (LQG) asymmetric information. show that under certain conditions, players' strategies in private types, together beliefs form perfect Bayesian equilibrium (PBE) of the game. Furthermore, it is shown this signaling due to fact future on...

10.1109/cdc.2016.7799332 article EN 2016-12-01

We consider a finite horizon dynamic game with N selfish players who observe their types privately and take actions, which are publicly observed. Players' evolve as conditionally independent Markov processes, conditioned on current actions. Their actions jointly determine instantaneous rewards. Since each player has different information set, this forms asymmetric there is no known methodology to find perfect Bayesian equilibria (PBE) for such games in general. In paper, we provide two-step...

10.1109/acc.2016.7525439 article EN 2022 American Control Conference (ACC) 2016-07-01

This paper studies node cooperation in a wireless network from the MAC layer perspective. A simple relay channel with source, relay, and destination is considered where source can transmit packet directly to or through relay. The tradeoff between average energy delay studied by posing problem as stochastic dynamical optimization problem. following two cases are considered: 1) nodes cooperative information decentralized, 2) strategic centralized. With decentralized nodes, structural result...

10.1109/tcomm.2014.2356578 article EN IEEE Transactions on Communications 2014-09-09

In this article, we have proposed a model-free reinforcement learning (RL) algorithm, based on sequential decomposition, to obtain optimal policies for mean field games (MFGs). We consider finite horizon MFGs with large population of homogeneous players, sequentially making strategic decisions. Each player observes private state and mean-field representing the empirical distribution other players' states. The is <italic xmlns:mml="http://www.w3.org/1998/Math/MathML"...

10.1109/tcns.2023.3264934 article EN IEEE Transactions on Control of Network Systems 2023-04-05

In this paper, we consider a finite horizon, non-stationary, mean field game (MFG) with large population of homogeneous players, sequentially making strategic decisions, where each player is affected by other players through an aggregate state termed as state. Each has private type that only it can observe, and representing the empirical distribution players' types, which shared among all them. Recently, authors in [1] provided sequential decomposition algorithm to compute equilibrium (MFE)...

10.1109/cdc42340.2020.9304340 article EN 2021 60th IEEE Conference on Decision and Control (CDC) 2020-12-14

We consider both finite and infinite horizon discounted mean-field games where there is a large population of homogeneous players sequentially making strategic decisions each player affected by other through an aggregate state. Each has private type that only she observes all commonly observe state which represents the empirical distribution players' types. Mean-field equilibrium (MFE) in such defined as solution coupled Bellman dynamic programming backward equation Mckean-Vlasov forward...

10.23919/acc45564.2020.9147646 article EN 2022 American Control Conference (ACC) 2020-07-01

The optimal coding scheme for Additive White Gaussian noise (AWGN) channels with noisy output feedback has been unknown several decades. best-known linear is by Chance and Love, where the coefficients of are numerically optimized based on unique observations. In this paper, we introduce a new class schemes, called <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">sequential schemes</i> , encoder sequentially updates state process feedback. We...

10.1109/tit.2023.3260807 article EN IEEE Transactions on Information Theory 2023-03-23

We consider both finite-horizon and infinite-horizon versions of a dynamic game with N selfish players who observe their types privately take actions that are publicly observed. Players’ evolve as conditionally independent Markov processes, conditioned on current actions. Their jointly determine instantaneous rewards. In games asymmetric information, widely used concept equilibrium is perfect Bayesian (PBE), which consists strategy belief pair simultaneously satisfy sequential rationality...

10.2139/ssrn.3143176 article EN SSRN Electronic Journal 2016-01-01

In this article, we consider a finite-horizon dynamic game with two players who observe their types privately and take actions that are publicly observed. Players' evolve as conditionally independent, controlled linear Gaussian processes, incur quadratic instantaneous costs. This forms asymmetric information. We show under certain conditions, players' strategies in private types, together beliefs, form perfect Bayesian equilibrium (PBE) of the game. is signaling due to fact future beliefs on...

10.1109/tcns.2021.3059835 article EN publisher-specific-oa IEEE Transactions on Control of Network Systems 2021-02-18

In this paper, we consider a sequential stochastic Stackelberg game with two players, leader, and follower. The follower observes the state of system privately while leader does not. Players play equilibrium where plays best response to leader's strategy. such scenario, has advantage committing policy that maximizes its returns given knowledge is going policy. Such pair strategies both players defined as game. Recently, [1] provided decomposition algorithm compute for games which allow...

10.1109/cdc42340.2020.9303846 article EN 2021 60th IEEE Conference on Decision and Control (CDC) 2020-12-14

In this paper, we consider a discrete time stochastic Stackelberg game where there is defender (also called leader) who has to defend target and an attacker follower). The private type that evolves as controlled Markov process. objective compute Stochastic equilibrium of the commits strategy. attacker's strategy best response defender's optimum given plays response. general computing such involves solving fixed-point equation for whole game. present algorithm computes strategies by smaller...

10.2139/ssrn.3411860 article EN SSRN Electronic Journal 2019-01-01

In this paper, we consider a discrete-time Stackelberg mean field game with leader and an infinite number of followers.The the followers each observe types privately that evolve as conditionally independent controlled Markov processes.The commits to dynamic policy best respond other.Knowing would play based on her policy, chooses maximizes reward.We refer resulting outcome equilibrium (SMFE).In provide master equation allows one compute all SMFE.Based our framework, two numerical...

10.1109/cdc51059.2022.9993218 article EN 2022 IEEE 61st Conference on Decision and Control (CDC) 2022-12-06

In this paper, we consider a discrete memoryless point to channel with noisy feedback, where there is sender private message that she wants communicate receiver by sequentially transmitting symbols over channel. After each transmission, receives feedback of the symbol received receiver. The goal design transmission control strategy minimize average probability error. This an instance decentralized information two controllers, and have no common information. There exist methodology in...

10.48550/arxiv.2002.09553 preprint EN other-oa arXiv (Cornell University) 2020-01-01

In this paper, we consider both finite and infinite horizon discounted dynamic mean-field games where there is a large population of homogeneous players sequentially making strategic decisions each player affected by other through an aggregate state. Each has private type that only she observes.Such have been studied in the literature under simplifying assumption state dynamics are stationary. non-stationary present novel backward recursive algorithm to compute Markov perfect equilibrium...

10.2139/ssrn.3385602 article EN SSRN Electronic Journal 2019-01-01

In this paper, we present a sequential decomposition algorithm to compute graphon mean-field equilibrium (GMFE) of dynamic games (GMFGs). We consider large population players sequentially making strategic decisions where the actions each player affect their neighbors which is captured in graph, generated by known graphon. Each observes private state and also common information as represents empirical networked distribution other players’ types. non-stationary dynamics novel backward...

10.2139/ssrn.3520348 article EN SSRN Electronic Journal 2020-01-01

We consider the problem of how strategic users with asymmetric information can learn an underlying time-varying state in a user-recommendation system. Users who observe private signals about state, sequentially make decision buying product whose value varies time ergodic manner. formulate team as instance decentralized stochastic control and characterize its optimal policies. With users, we design incentives such that reveal their true signals, so gap between objective is small overall...

10.1109/acssc.2015.7421305 article EN 2014 48th Asilomar Conference on Signals, Systems and Computers 2015-11-01

In this paper, we consider a multi-agent system with N cooperative agents where each agent privately observes its own private type and publicly others' actions. We propose novel model-free reinforcement learning algorithm to compute the optimal policies for that maximizes their collective reward. This setting belongs broad class of decentralized control problems partial information. use common approach [1], wherein some fictitious chooses best policy based on belief current states agents....

10.1109/ciss50987.2021.9400275 article EN 2021-03-24

Network utility maximization (NUM) is a general framework for designing distributed optimization algorithms networks. Existing studies proposed (economic) mechanisms to solve the NUM but largely neglected issue of large-scale implementation. In this paper, we present Large-Scale Vickery-Clark-Grove (VCG) Mechanism with simpler payment rule. The VCG maximizes network and achieves individual rationality budget balance. We show that, as number agents approaches infinity, each agent's incentive...

10.1109/ciss50987.2021.9400216 article EN 2021-03-24
Coming Soon ...