Yifan Zang

ORCID: 0000-0003-4537-384X
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Reinforcement Learning in Robotics
  • Artificial Intelligence in Games
  • Neural dynamics and brain function
  • Advanced Memory and Neural Computing
  • Advanced Bandit Algorithms Research
  • Blockchain Technology Applications and Security
  • Modular Robots and Swarm Intelligence
  • Functional Brain Connectivity Studies
  • Cryptography and Data Security
  • Adversarial Robustness in Machine Learning
  • Mobile Crowdsensing and Crowdsourcing
  • Receptor Mechanisms and Signaling
  • Cloud Data Security Solutions
  • Simulation Techniques and Applications

University of Chinese Academy of Sciences
2020-2024

Shandong Institute of Automation
2022-2023

Chinese Academy of Sciences
2020-2023

Beijing Academy of Artificial Intelligence
2022-2023

Jilin Province Science and Technology Department
2018

Jilin University
2018

Opponent modeling is essential to exploit sub-optimal opponents in strategic interactions. Most previous works focus on building explicit models predict the opponents' styles or strategies, which require a large amount of data train model and lack adaptability unknown opponents. In this work, we propose novel Learning Exploit (L2E) framework for implicit opponent modeling. L2E acquires ability through few interactions with different during training neural network can quickly adapt new...

10.1109/ijcnn55064.2022.9892077 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2022-07-18

Despite the potential of Multi-Agent Reinforcement Learning (MARL) in addressing numerous complex tasks, training a single team MARL agents to handle multiple diverse tasks remains challenge. In this paper, we introduce novel Multi-task method based on Knowledge Transfer cooperative (MKT-MARL). By learning from task-specific teachers, our approach empowers attain expert-level performance tasks. MKT-MARL utilizes knowledge distillation algorithm specifically designed for multi-agent...

10.1109/tg.2023.3316697 article EN IEEE Transactions on Games 2023-09-19

Experience replay plays a crucial role in Reinforcement Learning (RL), enabling the agent to remember and reuse experience from past. Most previous methods sample transitions using simple heuristics like uniformly sampling or prioritizing those good ones. Since humans can learn both bad experiences, more sophisticated algorithms need be developed. Inspired by potential energy physics, this work introduces artificial field into develops Potentialized Replay (PotER) as new effective algorithm...

10.24963/ijcai.2020/290 article EN 2020-07-01

Multi-task reinforcement learning endeavors to accomplish a set of different tasks with single policy. To enhance data efficiency by sharing parameters across multiple tasks, common practice segments the network into distinct modules and trains routing recombine these task-specific policies. However, existing approaches employ fixed number for all neglecting that varying difficulties commonly require amounts knowledge. This work presents Dynamic Depth Routing (D2R) framework, which learns...

10.1609/aaai.v38i11.29129 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2024-03-24

Efficient collaboration in the centralized training with decentralized execution (CTDE) paradigm remains a challenge cooperative multi-agent systems. We identify divergent action tendencies among agents as significant obstacle to CTDE's efficiency, requiring large number of samples achieve unified consensus on agents' policies. This divergence stems from lack adequate team consensus-related guidance signals during credit assignment CTDE. To address this, we propose Intrinsic Action Tendency...

10.1609/aaai.v38i16.29711 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2024-03-24

Efficient collaboration in the centralized training with decentralized execution (CTDE) paradigm remains a challenge cooperative multi-agent systems. We identify divergent action tendencies among agents as significant obstacle to CTDE's efficiency, requiring large number of samples achieve unified consensus on agents' policies. This divergence stems from lack adequate team consensus-related guidance signals during credit assignments CTDE. To address this, we propose Intrinsic Action Tendency...

10.48550/arxiv.2406.18152 preprint EN arXiv (Cornell University) 2024-06-26

Reinforcement learning (RL) algorithms typically require orders of magnitude more interactions than humans to learn effective policies. Research on memory in neuroscience suggests that humans' efficiency benefits from associating their experiences and reconstructing potential events. Inspired by this finding, we introduce a human brainlike structure for agents build general framework based improve the RL sampling efficiency. Since is similar reconstruction process psychology, name newly...

10.1109/tai.2023.3268612 article EN IEEE Transactions on Artificial Intelligence 2023-04-20

Multi-task reinforcement learning endeavors to accomplish a set of different tasks with single policy. To enhance data efficiency by sharing parameters across multiple tasks, common practice segments the network into distinct modules and trains routing recombine these task-specific policies. However, existing approaches employ fixed number for all neglecting that varying difficulties commonly require amounts knowledge. This work presents Dynamic Depth Routing (D2R) framework, which learns...

10.48550/arxiv.2312.14472 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Cloud computing has been developing at a rapid speed, playing an important role in many fields, especially environments like hospitals which produce lot of data every day and have specific users.Because the security information stored cloud cannot be guaranteed, we propose safe storage medical based on attribute encryption.This paper focuses how to apply attribute-based encryption hospitals' environment, design access process different users environment by using encryption.Our goal is build...

10.2991/ammsa-18.2018.29 article EN cc-by-nc 2018-01-01
Coming Soon ...