Junchao Liao

ORCID: 0000-0003-4282-0843
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Multimodal Machine Learning Applications
  • Advanced Neural Network Applications
  • Anomaly Detection Techniques and Applications
  • Text and Document Classification Technologies
  • Domain Adaptation and Few-Shot Learning
  • Data-Driven Disease Surveillance

Huazhong University of Science and Technology
2022-2024

Recently, the semantics of scene text has been proven to be essential in fine-grained image classification. However, existing methods mainly exploit literal meaning for recognition, which might irrelevant when it is not significantly related objects/scenes. We propose an end-to-end trainable network that mines implicit contextual knowledge behind and enhance correlation fine-tune representation. Unlike methods, our model integrates three modalities: visual feature extraction, correlating...

10.1109/cvpr52688.2022.00458 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022-06-01

<title>Abstract</title> Semantic segmentation is a fundamental problem in computer vision and it requires high-resolution feature maps for dense prediction. Current coordinate-guided low-resolution interpolation methods, e.g., bilinear interpolation, produce coarse features which suffer from misalignment insufficient context information. Moreover, enriching semantics to high computation burden, so that challenging meet the requirement of lowlatency inference. We propose novel Guided...

10.21203/rs.3.rs-4915410/v1 preprint EN Research Square (Research Square) 2024-09-13

Recently, the semantics of scene text has been proven to be essential in fine-grained image classification. However, existing methods mainly exploit literal meaning for recognition, which might irrelevant when it is not significantly related objects/scenes. We propose an end-to-end trainable network that mines implicit contextual knowledge behind and enhance correlation fine-tune representation. Unlike methods, our model integrates three modalities: visual feature extraction, correlating...

10.48550/arxiv.2203.14215 preprint EN other-oa arXiv (Cornell University) 2022-01-01
Coming Soon ...