K-ON: Stacking Knowledge On the Head Layer of Large Language Model
FOS: Computer and information sciences
Computer Science - Computation and Language
Artificial Intelligence (cs.AI)
Computer Science - Artificial Intelligence
Computation and Language (cs.CL)
DOI:
10.48550/arxiv.2502.06257
Publication Date:
2025-02-10
AUTHORS (8)
ABSTRACT
Recent advancements in large language models (LLMs) have significantly improved various natural processing (NLP) tasks. Typically, LLMs are trained to predict the next token, aligning well with many NLP However, knowledge graph (KG) scenarios, entities fundamental units and identifying an entity requires at least several tokens. This leads a granularity mismatch between KGs languages. To address this issue, we propose K-ON, which integrates KG into LLM by employing multiple head layers for k-step prediction. K-ON can not only generate entity-level results one step, but also enables contrastive loss against entities, is most powerful tool representation learning. Experimental show that outperforms state-of-the-art methods incorporate text even other modalities.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....