Yixun Liang

ORCID: 0000-0003-4750-8875
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Video Analysis and Summarization
  • Machine Learning and Data Classification
  • Computer Graphics and Visualization Techniques
  • Image Retrieval and Classification Techniques
  • Human Motion and Animation
  • Human Pose and Action Recognition
  • Software Engineering Research
  • Advanced Vision and Imaging
  • Industrial Vision Systems and Defect Detection
  • Medical Image Segmentation Techniques
  • 3D Shape Modeling and Analysis
  • Advanced Image Fusion Techniques

Hong Kong University of Science and Technology
2023-2025

University of Hong Kong
2023-2025

University of Electronic Science and Technology of China
2022

10.1109/cvpr52733.2024.00623 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2024-06-16

Despite recent advancements in the Large Reconstruction Model (LRM) demonstrating impressive results, when extending its input from single image to multiple images, it exhibits inefficiencies, subpar geometric and texture quality, as well slower convergence speed than expected. It is attributed that, LRM formulates 3D reconstruction a naive images-to-3D translation problem, ignoring strong coherence among images. In this paper, we propose Multi-view (M-LRM) designed reconstruct high-quality...

10.1109/tvcg.2025.3572341 article EN IEEE Transactions on Visualization and Computer Graphics 2025-01-01

Pan-sharpening, a task involving information fusion, entails merging panchromatic (PAN) images with high spatial resolution and low-resolution multispectral (LRMS) in order to obtain high-resolution (HRMS) images. Due deep learning's excellent regression capabilities, it has recently become the dominating technique for this assignment. Meanwhile, development of transformer, novel learning architecture natural language processing, provided researchers new insights. In letter, we seek extend...

10.1109/lgrs.2022.3170904 article EN IEEE Geoscience and Remote Sensing Letters 2022-01-01

Abstract Neural radiance fields (NeRF) have demonstrated a promising research direction for novel view synthesis. However, the existing approaches either require per‐scene optimization that takes significant computation time or condition on local features which overlook global context of images. To tackle this shortcoming, we propose Conditionally Parameterized Radiance Fields (CP‐NeRF), plug‐in module enables NeRF to leverage contextual information from different scales. Instead optimizing...

10.1111/cgf.14940 article EN Computer Graphics Forum 2023-10-01

Score Distillation Sampling (SDS) has emerged as a prevalent technique for text-to-3D generation, enabling 3D content creation by distilling view-dependent information from text-to-2D guidance. However, they frequently exhibit shortcomings such over-saturated color and excess smoothness. In this paper, we conduct thorough analysis of SDS refine its formulation, finding that the core design is to model distribution rendered images. Following insight, introduce novel strategy called...

10.48550/arxiv.2409.05099 preprint EN arXiv (Cornell University) 2024-09-08

Leveraging the visual priors of pre-trained text-to-image diffusion models offers a promising solution to enhance zero-shot generalization in dense prediction tasks. However, existing methods often uncritically use original formulation, which may not be optimal due fundamental differences between and image generation. In this paper, we provide systemic analysis formulation for prediction, focusing on both quality efficiency. And find that parameterization type generation, learns predict...

10.48550/arxiv.2409.18124 preprint EN arXiv (Cornell University) 2024-09-26
Coming Soon ...