Shouwei Ruan

ORCID: 0009-0007-0481-5855
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Adversarial Robustness in Machine Learning
  • Anomaly Detection Techniques and Applications
  • Domain Adaptation and Few-Shot Learning
  • Physical Unclonable Functions (PUFs) and Hardware Security
  • Multimodal Machine Learning Applications
  • Advanced Malware Detection Techniques
  • Advanced Neural Network Applications
  • Forensic Fingerprint Detection Methods
  • Boron Compounds in Chemistry
  • Bacillus and Francisella bacterial research
  • Cell Image Analysis Techniques
  • Face recognition and analysis
  • Semantic Web and Ontologies
  • Medical Imaging Techniques and Applications
  • Advanced Optical Sensing Technologies
  • Human Pose and Action Recognition

Beihang University
2023-2025

Adversarial patch is one of the important forms performing adversarial attacks in physical world. To improve naturalness and aggressiveness existing patches, location-aware patches are proposed, where patch's location on target object integrated into optimization process to perform attacks. Although it effective, efficiently finding optimal for placing challenging, especially under black-box attack settings. In this paper, we first empirically find that aggregation regions locations show...

10.1109/tpami.2025.3526188 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2025-01-01

Recent studies have demonstrated that visual recognition models lack robustness to distribution shift. However, current work mainly considers model 2D image transformations, leaving viewpoint changes in the 3D world less explored. In general, are prevalent various real-world applications (e.g., autonomous driving), making it imperative evaluate robustness. this paper, we propose a novel method called ViewFool find adversarial viewpoints mislead models. By encoding objects as neural radiance...

10.48550/arxiv.2210.03895 preprint EN cc-by-nc-sa arXiv (Cornell University) 2022-01-01

10.1109/cvpr52733.2024.02314 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2024-06-16

Visual recognition models are not invariant to viewpoint changes in the 3D world, as different viewing directions can dramatically affect predictions given same object. Compared 2D transformations, exploration of invariance deserves more attention for its greater practical significance. Motivated by success adversarial training promoting model robustness, we propose Viewpoint-Invariant Adversarial Training (VIAT) improve robustness common image classifiers. By regarding transformation an...

10.1109/iccv51070.2023.00434 article EN 2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2023-10-01

Viewpoint invariance remains challenging for visual recognition in the 3D world, as altering viewing directions can significantly impact predictions same object. While substantial efforts have been dedicated to making neural networks invariant 2D image translations and rotations, viewpoint is rarely investigated. Motivated by success of adversarial training enhancing model robustness, we propose Viewpoint-Invariant Adversarial Training (VIAT) improve robustness classifiers. Regarding...

10.48550/arxiv.2307.11528 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Vision-Language Pre-training (VLP) models like CLIP have achieved remarkable success in computer vision and particularly demonstrated superior robustness to distribution shifts of 2D images. However, their under 3D viewpoint variations is still limited, which can hinder the development for real-world applications. This paper successfully addresses this concern while keeping VLPs' original performance by breaking through two primary obstacles: 1) scarcity training data 2) suboptimal...

10.48550/arxiv.2404.12139 preprint EN arXiv (Cornell University) 2024-04-18

Embodied intelligence empowers agents with a profound sense of perception, enabling them to respond in manner closely aligned real-world situations. Large Language Models (LLMs) delve into language instructions depth, serving crucial role generating plans for intricate tasks. Thus, LLM-based embodied models further enhance the agent's capacity comprehend and process information. However, this amalgamation also ushers new challenges pursuit heightened intelligence. Specifically, attackers can...

10.48550/arxiv.2405.19802 preprint EN arXiv (Cornell University) 2024-05-30

Adversarial patches present significant challenges to the robustness of deep learning models, making development effective defenses become critical for real-world applications. This paper introduces DIFFender, a novel DIFfusion-based DeFender framework that leverages power text-guided diffusion model counter adversarial patch attacks. At core our approach is discovery Anomaly Perception (AAP) phenomenon, which enables accurately detect and locate by analyzing distributional anomalies....

10.48550/arxiv.2409.09406 preprint EN arXiv (Cornell University) 2024-09-14

With the rise of deep learning, facial recognition technology has seen extensive research and rapid development. Although is considered a mature technology, we find that existing open-source models commercial algorithms lack robustness in certain real-world Out-of-Distribution (OOD) scenarios, raising concerns about reliability these systems. In this paper, introduce OODFace, which explores OOD challenges faced by from two perspectives: common corruptions appearance variations. We...

10.48550/arxiv.2412.02479 preprint EN arXiv (Cornell University) 2024-12-03

Vision Language Models (VLMs) have exhibited remarkable generalization capabilities, yet their robustness in dynamic real-world scenarios remains largely unexplored. To systematically evaluate VLMs' to 3D variations, we propose AdvDreamer, the first framework that generates physically reproducible adversarial transformation (Adv-3DT) samples from single-view images. AdvDreamer integrates advanced generative techniques with two key innovations and aims characterize worst-case distributions of...

10.48550/arxiv.2412.03002 preprint EN arXiv (Cornell University) 2024-12-03

Adversarial patch is one of the important forms performing adversarial attacks in physical world. To improve naturalness and aggressiveness existing patches, location-aware patches are proposed, where patch's location on target object integrated into optimization process to perform attacks. Although it effective, efficiently finding optimal for placing challenging, especially under black-box attack settings. In this paper, we propose Distribution-Optimized Patch (DOPatch), a novel method...

10.48550/arxiv.2306.16131 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Visual recognition models are not invariant to viewpoint changes in the 3D world, as different viewing directions can dramatically affect predictions given same object. Although many efforts have been devoted making neural networks 2D image translations and rotations, invariance is rarely investigated. As most process images perspective view, it challenging impose based only on inputs. Motivated by success of adversarial training promoting model robustness, we propose Viewpoint-Invariant...

10.48550/arxiv.2307.10235 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Compared with transferable untargeted attacks, targeted adversarial attacks could specify the misclassification categories of samples, posing a greater threat to security-critical tasks. In meanwhile, 3D due their potential multi-view robustness, can more comprehensively identify weaknesses in existing deep learning systems, possessing great application value. However, field remains vacant. The goal this work is develop effective technique that generate examples, filling gap field. To...

10.48550/arxiv.2312.09558 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Adversarial attacks, particularly patch pose significant threats to the robustness and reliability of deep learning models. Developing reliable defenses against attacks is crucial for real-world applications, yet current research in this area unsatisfactory. In paper, we propose DIFFender, a novel defense method that leverages text-guided diffusion model defend adversarial patches. DIFFender includes two main stages: localization restoration. stage, find exploit an intriguing property...

10.48550/arxiv.2306.09124 preprint EN other-oa arXiv (Cornell University) 2023-01-01
Coming Soon ...