Ruizhou Ding

ORCID: 0000-0003-4311-3761
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Neural Network Applications
  • Adversarial Robustness in Machine Learning
  • Machine Learning and Data Classification
  • Domain Adaptation and Few-Shot Learning
  • Advanced Memory and Neural Computing
  • CCD and CMOS Imaging Sensors
  • Human Pose and Action Recognition
  • Video Surveillance and Tracking Methods
  • Advanced Image and Video Retrieval Techniques
  • Hydrological Forecasting Using AI
  • Visual Attention and Saliency Detection
  • Machine Learning in Materials Science
  • Flood Risk Assessment and Management
  • Machine Learning and ELM
  • Climate variability and models
  • Parallel Computing and Optimization Techniques
  • Hydrology and Watershed Management Studies
  • Meteorological Phenomena and Simulations
  • Neural Networks and Applications
  • Water Systems and Optimization
  • Anomaly Detection Techniques and Applications

Carnegie Mellon University
2017-2022

Binarized Neural Networks (BNNs) can significantly reduce the inference latency and energy consumption in resource-constrained devices due to their pure-logical computation fewer memory accesses. However, training BNNs is difficult since activation flow encounters degeneration, saturation, gradient mismatch problems. Prior work alleviates these issues by increasing bits adding floating-point scaling factors, thereby sacrificing BNN's efficiency. In this paper, we propose use distribution...

10.1109/cvpr.2019.01167 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019-06-01

Pruning convolutional filters has demonstrated its effectiveness in compressing ConvNets. Prior art filter pruning requires users to specify a target model complexity (e.g., size or FLOP count) for the resulting architecture. However, determining can be difficult optimizing various embodied AI applications such as autonomous robots, drones, and user-facing applications. First, both accuracy speed of ConvNets affect performance application. Second, application hard assess without evaluating...

10.1109/cvpr42600.2020.00159 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020-06-01

In vision-enabled autonomous systems such as robots and cars, video object detection plays a crucial role, both its speed accuracy are important factors to provide reliable operation. The key insight we show in this paper is that not necessarily trade-off when it comes image scaling. Our results re-scaling the lower resolution will sometimes produce better accuracy. Based on observation, propose novel approach, dubbed AdaScale, which adaptively selects input scale improves for detection. To...

10.48550/arxiv.1902.02910 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Can we reduce the search cost of Neural Architecture Search (NAS) from days down to only few hours? NAS methods automate design Convolutional Networks (ConvNets) under hardware constraints and they have emerged as key components AutoML frameworks. However, problem remains challenging due combinatorially large space significant time (at least 200 GPU-hours). In this work, alleviate less than 3 hours, while achieving state-of-the-art image classification results mobile latency constraints. We...

10.1109/jstsp.2020.2971421 article EN publisher-specific-oa IEEE Journal of Selected Topics in Signal Processing 2020-02-03

Deep Neural Networks (DNNs) have been adopted in many systems because of their higher classification accuracy, with custom hardware implementations great candidates for high-speed, accurate inference. While progress achieving large scale, highly DNNs has made, significant energy and area are required due to massive memory accesses computations. Such demands pose a challenge any DNN implementation, yet it is more natural handle platform. To alleviate the increased demand storage energy,...

10.1109/aspdac.2018.8297274 article EN 2018 23rd Asia and South Pacific Design Automation Conference (ASP-DAC) 2018-01-01

Application-specific integrated circuit (ASIC) implementations for Deep Neural Networks (DNNs) have been adopted in many systems because of their higher classification speed. However, although they may be characterized by better accuracy, larger DNNs require significant energy and area, thereby limiting wide adoption. The consumption is driven both memory accesses computation. Binarized (BNNs), as a trade-off between accuracy consumption, can achieve great reduction, good large due to its...

10.1145/3060403.3060465 article EN Proceedings of the Great Lakes Symposium on VLSI 2022 2017-05-10

Deep Neural Networks (DNNs) have been adopted in many systems because of their higher classification accuracy, with custom hardware implementations great candidates for high-speed, accurate inference. While progress achieving large scale, highly DNNs has made, significant energy and area are required due to massive memory accesses computations. Such demands pose a challenge any DNN implementation, yet it is more natural handle platform. To alleviate the increased demand storage energy,...

10.5555/3201607.3201609 article EN Asia and South Pacific Design Automation Conference 2018-01-22

In recent years, supervised learning using Convolutional Neural Networks (CNNs) has achieved great success in image classification tasks, and large scale labeled datasets have contributed significantly to this achievement. However, the definition of a label is often application dependent. For example, an cat can be as "cat" or perhaps more specifically "Persian cat." We refer granularity. paper, we conduct extensive experiments various demonstrate analyze how why training based on fine-grain...

10.1109/icdmw.2018.00131 article EN 2022 IEEE International Conference on Data Mining Workshops (ICDMW) 2018-11-01

To improve the throughput and energy efficiency of Deep Neural Networks (DNNs) on customized hardware, lightweight neural networks constrain weights DNNs to be a limited combination (denoted as k ϵ {1, 2}) powers 2. In such networks, multiply-accumulate operation can replaced with single shift operation, or two shifts an add operation. provide even more design flexibility, for each convolutional filter optimally chosen instead being fixed every filter. this paper, we formulate selection...

10.1145/3316781.3317828 article EN 2019-05-23

Hardware implementations of deep neural networks (DNNs) have been adopted in many systems because their higher classification speed. However, while they may be characterized by better accuracy, larger DNNs require significant energy and area, thereby limiting wide adoption. The consumption is driven both memory accesses computation. Binarized (BNNs), as a tradeoff between accuracy consumption, can achieve great reduction good for large due to regularization effect. BNNs show poor when...

10.1145/3270689 article EN ACM Transactions on Reconfigurable Technology and Systems 2018-09-30

Can we automatically design a Convolutional Network (ConvNet) with the highest image classification accuracy under latency constraint of mobile device? Neural Architecture Search (NAS) for ConvNet is challenging problem due to combinatorially large space and search time (at least 200 GPU-hours). To alleviate this complexity, propose Single-Path NAS, novel differentiable NAS method designing device-efficient ConvNets in less than 4 hours. 1. Novel formulation: our introduces single-path,...

10.48550/arxiv.1905.04159 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Application-specific integrated circuit (ASIC) implementations for Deep Neural Networks (DNNs) have been adopted in many systems because of their higher classification speed. However, although they may be characterized by better accuracy, larger DNNs require significant energy and area, thereby limiting wide adoption. The consumption is driven both memory accesses computation. Binarized (BNNs), as a trade-off between accuracy consumption, can achieve great reduction, good large due to its...

10.48550/arxiv.1802.02178 preprint EN other-oa arXiv (Cornell University) 2018-01-01

In recent years, Convolutional Neural Networks (CNNs) have shown superior capability in visual learning tasks. While accuracy-wise CNNs provide unprecedented performance, they are also known to be computationally intensive and energy demanding for modern computer systems. this paper, we propose Virtual Pooling (ViP), a model-level approach improve speed consumption of CNN-based image classification object detection tasks, with provable error bound. We show the efficacy ViP through...

10.1109/wacv45572.2020.9093418 article EN 2020-03-01

Binarized Neural Networks (BNNs) can significantly reduce the inference latency and energy consumption in resource-constrained devices due to their pure-logical computation fewer memory accesses. However, training BNNs is difficult since activation flow encounters degeneration, saturation, gradient mismatch problems. Prior work alleviates these issues by increasing bits adding floating-point scaling factors, thereby sacrificing BNN's efficiency. In this paper, we propose use distribution...

10.48550/arxiv.1904.02823 preprint EN other-oa arXiv (Cornell University) 2019-01-01

To improve the throughput and energy efficiency of Deep Neural Networks (DNNs) on customized hardware, lightweight neural networks constrain weights DNNs to be a limited combination (denoted as $k\in\{1,2\}$) powers 2. In such networks, multiply-accumulate operation can replaced with single shift operation, or two shifts an add operation. provide even more design flexibility, $k$ for each convolutional filter optimally chosen instead being fixed every filter. this paper, we formulate...

10.48550/arxiv.1904.02835 preprint EN other-oa arXiv (Cornell University) 2019-01-01

To improve precipitation predictions and enable accurate rainfall models for smart water networks, it is imperative to account multivariate multiscale dynamics with long-memory temporal relationships, long-range spatial dependencies, low-frequency variability. While prior art has motivated the use of complex networks capture these trends, existing work limited specific climate phenomena (such as El Niño) regions. In this paper we employ a comprehensive assessment respect across multiple...

10.1145/3055366.3055374 article EN 2017-04-06

In recent years, supervised learning using Convolutional Neural Networks (CNNs) has achieved great success in image classification tasks, and large scale labeled datasets have contributed significantly to this achievement. However, the definition of a label is often application dependent. For example, an cat can be as "cat" or perhaps more specifically "Persian cat." We refer granularity. paper, we conduct extensive experiments various demonstrate analyze how why training based on fine-grain...

10.48550/arxiv.1901.07012 preprint EN other-oa arXiv (Cornell University) 2019-01-01

As the machine learning and systems community strives to achieve higher energy-efficiency through custom DNN accelerators model compression techniques, there is a need for design space exploration framework that incorporates quantization-aware processing elements into accelerator while having accurate fast power, performance, area models. In this work, we present QAPPA, highly parameterized modeling accelerators. Our can facilitate future research on of various choices such as bit precision,...

10.48550/arxiv.2205.08648 preprint EN other-oa arXiv (Cornell University) 2022-01-01

As the machine learning and systems communities strive to achieve higher energy-efficiency through custom deep neural network (DNN) accelerators, varied bit precision or quantization levels, there is a need for design space exploration frameworks that incorporate quantization-aware processing elements (PE) into accelerator while having accurate fast power, performance, area models. In this work, we present QADAM, highly parameterized modeling framework DNN accelerators. Our can facilitate...

10.48550/arxiv.2205.13045 preprint EN other-oa arXiv (Cornell University) 2022-01-01
Coming Soon ...