Kai Li

ORCID: 0000-0003-3840-3270
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Reinforcement Learning in Robotics
  • Artificial Intelligence in Games
  • Big Data and Digital Economy
  • Neural dynamics and brain function
  • Gaussian Processes and Bayesian Inference
  • Air Quality Monitoring and Forecasting
  • Advanced Bandit Algorithms Research
  • Domain Adaptation and Few-Shot Learning
  • Adaptive Dynamic Programming Control
  • Multimodal Machine Learning Applications
  • Fault Detection and Control Systems
  • Adversarial Robustness in Machine Learning
  • Physical Unclonable Functions (PUFs) and Hardware Security
  • Face and Expression Recognition
  • Text and Document Classification Technologies
  • Sports Analytics and Performance
  • IoT and Edge/Fog Computing
  • Semantic Web and Ontologies
  • Logic, Reasoning, and Knowledge
  • Natural Language Processing Techniques
  • Formal Methods in Verification
  • Advanced Memory and Neural Computing
  • Receptor Mechanisms and Signaling
  • Advanced Neural Network Applications
  • Functional Brain Connectivity Studies

University of Chinese Academy of Sciences
2022-2024

Chinese Academy of Sciences
2020-2024

Computer Network Information Center
2022-2024

Beijing Academy of Artificial Intelligence
2022-2023

NEC (United States)
2023

Shandong Institute of Automation
2021-2023

Fujian Normal University
2010

Recent few-shot video classification (FSVC) works achieve promising performance by capturing similarity across support and query samples with different temporal alignment strategies or learning discriminative features via Transformer block within each episode. However, they ignore two important issues: a) It is difficult to capture rich intrinsic action semantics from a limited number of instances task. b) Redundant irrelevant frames in videos easily weaken the positive influence frames. To...

10.1109/iccv51070.2023.01769 article EN 2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2023-10-01

Opponent modeling is essential to exploit sub-optimal opponents in strategic interactions. Most previous works focus on building explicit models predict the opponents' styles or strategies, which require a large amount of data train model and lack adaptability unknown opponents. In this work, we propose novel Learning Exploit (L2E) framework for implicit opponent modeling. L2E acquires ability through few interactions with different during training neural network can quickly adapt new...

10.1109/ijcnn55064.2022.9892077 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2022-07-18

Experience replay plays a crucial role in Reinforcement Learning (RL), enabling the agent to remember and reuse experience from past. Most previous methods sample transitions using simple heuristics like uniformly sampling or prioritizing those good ones. Since humans can learn both bad experiences, more sophisticated algorithms need be developed. Inspired by potential energy physics, this work introduces artificial field into develops Potentialized Replay (PotER) as new effective algorithm...

10.24963/ijcai.2020/290 article EN 2020-07-01

This paper proposes a novel, data-agnostic, model poisoning attack on Federated Learning (FL), by designing new adversarial graph autoencoder (GAE)-based framework. The requires no knowledge of FL training data and achieves both effectiveness undetectability. By listening to the benign local models global model, attacker extracts structural correlations among features substantiating models. then adversarially regenerates while maximizing loss, subsequently generates malicious using structure...

10.48550/arxiv.2311.18498 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Multi-task reinforcement learning endeavors to accomplish a set of different tasks with single policy. To enhance data efficiency by sharing parameters across multiple tasks, common practice segments the network into distinct modules and trains routing recombine these task-specific policies. However, existing approaches employ fixed number for all neglecting that varying difficulties commonly require amounts knowledge. This work presents Dynamic Depth Routing (D2R) framework, which learns...

10.1609/aaai.v38i11.29129 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2024-03-24

Reinforcement learning (RL) algorithms typically require orders of magnitude more interactions than humans to learn effective policies. Research on memory in neuroscience suggests that humans' efficiency benefits from associating their experiences and reconstructing potential events. Inspired by this finding, we introduce a human brainlike structure for agents build general framework based improve the RL sampling efficiency. Since is similar reconstruction process psychology, name newly...

10.1109/tai.2023.3268612 article EN IEEE Transactions on Artificial Intelligence 2023-04-20

Owing to the unremitting efforts from a few institutes, researchers have recently made significant progress in designing superhuman artificial intelligence (AI) no-limit Texas hold'em (NLTH), primary testbed for large-scale imperfect-information game research. However, it remains challenging new study this problem since there are no standard benchmarks comparing with existing methods, which hinders further developments research area. This work presents OpenHoldem, an integrated benchmark...

10.1109/tnnls.2023.3280186 article EN IEEE Transactions on Neural Networks and Learning Systems 2023-06-14

Solving hard exploration tasks with sparse rewards is notoriously challenging in reinforcement learning (RL), which needs to address two key issues simultaneously: exploiting past successful experiences and exploring the unknown environment. Many prior works take expert demonstrations as learn imitate them directly. However, these are often not available practice. Recently, curiosity-driven RL methods provide intrinsic rewards, encouraging agent explore states high novelty. Nonetheless, they...

10.1109/ijcnn54540.2023.10192041 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2023-06-18

This work considers the problem of deep reinforcement learning (RL) with long time dependencies and sparse rewards, as are found in many hard exploration games. A graph-based representation is proposed to allow an agent perform self-navigation for environmental exploration. The graph not only effectively models environment structure, but also efficiently traces state changes corresponding actions. By encouraging earn a new influence-based curiosity reward game observations, whole task...

10.1109/ijcnn52387.2021.9534251 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2021-07-18

Gaussian processes (GPs), or distributions over arbitrary functions in a continuous domain, can be generalized to the multi-output case: linear model of coregionalization (LMC) is one approach. LMCs estimate and exploit correlations across multiple outputs. While estimation performed efficiently for single-output GPs, these assume stationarity, but case cross-covariance interaction not stationary. We propose Large Linear GP (LLGP), which circumvents need stationarity by inducing structure...

10.48550/arxiv.1705.10813 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Opponent modeling is essential to exploit sub-optimal opponents in strategic interactions. Most previous works focus on building explicit models directly predict the opponents' styles or strategies, which require a large amount of data train model and lack adaptability unknown opponents. In this work, we propose novel Learning Exploit (L2E) framework for implicit opponent modeling. L2E acquires ability by few interactions with different during training, thus can adapt new testing quickly. We...

10.48550/arxiv.2102.09381 preprint EN other-oa arXiv (Cornell University) 2021-01-01
Coming Soon ...