Yichi Zhang

ORCID: 0009-0007-4046-1003
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Graph Neural Networks
  • Topic Modeling
  • Multimodal Machine Learning Applications
  • Natural Language Processing Techniques
  • Domain Adaptation and Few-Shot Learning
  • Semantic Web and Ontologies
  • Advanced Algorithms and Applications
  • Wireless Sensor Networks and IoT
  • Machine Learning in Healthcare
  • Rough Sets and Fuzzy Logic
  • Data-Driven Disease Surveillance
  • Misinformation and Its Impacts
  • Vaccine Coverage and Hesitancy
  • Image Processing Techniques and Applications
  • Multi-Criteria Decision Making

Zhejiang University
2023-2024

Zhejiang University of Science and Technology
2024

Chinese Center For Disease Control and Prevention
2023

Multi-modal entity alignment (MMEA) aims to discover identical entities across different knowledge graphs (KGs) whose are associated with relevant images. However, current MMEA algorithms rely on KG-level modality fusion strategies for multi-modal representation, which ignores the variations of preferences entities, thus compromising robustness against noise in modalities such as blurry images and relations. This paper introduces MEAformer, a mlti-modal transformer approach meta hybrid,...

10.1145/3581783.3611786 article EN 2023-10-26

Existing domain-specific Large Language Models (LLMs) are typically developed by fine-tuning general-purposed LLMs with large-scale corpora. However, training on corpora often fails to effectively organize domain knowledge of LLMs, leading fragmented understanding. Inspired how humans connect concepts and through mind maps, we aim emulate this approach using ontology hierarchical conceptual reorganize LLM's knowledge. From perspective, propose an ontology-driven self-training framework...

10.48550/arxiv.2502.05478 preprint EN arXiv (Cornell University) 2025-02-08

Multi-modal knowledge graphs (MMKG) store structured world containing rich multi-modal descriptive information. To overcome their inherent incompleteness, graph completion (MMKGC) aims to discover unobserved from given MMKGs, leveraging both structural information the triples and of entities. Existing MMKGC methods usually extract features with pre-trained models employ a fusion module integrate triple prediction. However, this often results in coarse handling data, overlooking nuanced,...

10.48550/arxiv.2404.09468 preprint EN arXiv (Cornell University) 2024-04-15

Graph neural network (GNN)-based methods have demonstrated remarkable performance in various knowledge graph (KG) tasks. However, most existing approaches rely on observing all entities during training, posing a challenge real-world graphs where new emerge frequently. To address this limitation, we introduce Decentralized Attention Network (DAN). DAN leverages neighbor context as the query vector to score neighbors of an entity, thereby distributing entity semantics only among its...

10.48550/arxiv.2010.08114 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Multi-modal knowledge graph completion (MMKGC) aims to automatically discover the unobserved factual from a given multi-modal by collaboratively modeling triple structure and information entities.However, real-world MMKGs present challenges due their diverse imbalanced nature, which means that modality can span various types (e.g., image, text, numeric, audio, video) but its distribution among entities is uneven, leading missing modalities for certain entities.Existing works usually focus on...

10.36227/techrxiv.171259566.60211714/v1 preprint EN cc-by 2024-04-08

Multi-modal knowledge graph completion (MMKGC) aims to automatically discover new triples in the given multi-modal graphs (MMKGs), which is achieved by collaborative modeling structural information concealed massive and features of entities. Existing methods tend focus on crafting elegant entity-wise fusion strategies, yet they overlook utilization multi-perspective within modalities under diverse relational contexts. To address this issue, we introduce a novel MMKGC framework with Mixture...

10.48550/arxiv.2405.16869 preprint EN arXiv (Cornell University) 2024-05-27

Multi-modal knowledge graph completion (MMKGC) aims to automatically discover the unobserved factual from a given multi-modal by collaboratively modeling triple structure and information entities. However, real-world MMKGs present challenges due their diverse imbalanced nature, which means that modality can span various types (e.g., image, text, numeric, audio, video) but its distribution among entities is uneven, leading missing modalities for certain Existing works usually focus on common...

10.48550/arxiv.2406.17605 preprint EN arXiv (Cornell University) 2024-03-27

10.1145/3626772.3657800 article EN Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval 2024-07-10

The unprecedented developments in segmentation foundational models have become a dominant force the field of computer vision, introducing multitude previously unexplored capabilities wide range natural images and videos. Specifically, Segment Anything Model (SAM) signifies noteworthy expansion prompt-driven paradigm into domain image segmentation. recent introduction SAM2 effectively extends original SAM to streaming fashion demonstrates strong performance video However, due substantial...

10.48550/arxiv.2408.12889 preprint EN arXiv (Cornell University) 2024-08-23
Coming Soon ...