Exploring Concept Depth: How Large Language Models Acquire Knowledge at Different Layers?

FOS: Computer and information sciences Computer Science - Machine Learning Computer Science - Computation and Language Artificial Intelligence (cs.AI) Computer Science - Artificial Intelligence Computation and Language (cs.CL) Machine Learning (cs.LG)
DOI: 10.48550/arxiv.2404.07066 Publication Date: 2024-04-10
ABSTRACT
This paper studies the phenomenon that different concepts are learned in layers of large language models, i.e. more difficult fully acquired with deeper layers. We define difficulty by level abstraction, and here it is crudely categorized factual, emotional, inferential. Each category contains a spectrum tasks, arranged from simple to complex. For example, within factual dimension, tasks range lie detection categorizing mathematical problems. employ probing technique extract representations model apply these classification tasks. Our findings reveal models tend efficiently classify simpler indicating shallower Conversely, complex may only be discernible at layers, if all. explores implications for our understanding learning processes internal representations. implementation available \url{https://github.com/Luckfort/CD}.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....