Self-prompted Chain-of-Thought on Large Language Models for Open-domain Multi-hop Reasoning

FOS: Computer and information sciences Computer Science - Computation and Language Artificial Intelligence (cs.AI) Computer Science - Artificial Intelligence Computation and Language (cs.CL)
DOI: 10.48550/arxiv.2310.13552 Publication Date: 2023-01-01
ABSTRACT
In open-domain question-answering (ODQA), most existing questions require single-hop reasoning on commonsense. To further extend this task, we officially introduce multi-hop (ODMR) by answering with explicit steps in setting. Recently, large language models (LLMs) have found significant utility facilitating ODQA without external corpus. Furthermore, chain-of-thought (CoT) prompting boosts the capability of LLMs to a greater extent manual or automated paradigms. However, methods lack quality assurance, while approaches suffer from limited scalability and poor diversity, hindering capabilities LLMs. paper, propose Self-prompted Chain-of-Thought (SP-CoT), an framework mass-produce high CoTs LLMs, for SP-CoT introduces generation pipeline ODMR datasets, adaptive sampler in-context CoT selection self-prompted inference via learning. Extensive experiments four benchmarks show that our proposed not only significantly surpasses previous SOTA large-scale (175B) but also nearly doubles zero-shot performance small-scale (13B) Further analysis reveals remarkable elicit direct concise intermediate recalling $\sim$50\% answers MuSiQue-Ans dataset.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....