This manuscript was automatically generated on May 24, 2023.
John Doe
XXXX-XXXX-XXXX-XXXX
·
johndoe
·
johndoe
·
@johndoe@mastodon.social
Department of Something, University of Whatever
· Funded by Grant XXXXXXXX
Chris Mungall
✉
0000-0002-6601-2165
·
cmungall
Environmental Genomics and Systems Biology Division, Lawrence Berkeley National Laboratory, Berkeley, CA, 94720
✉ — Correspondence possible via GitHub Issues or email to Chris Mungall <cjmungall@lbl.gov>.
Reasoning is a core component of human intelligence, and a key goal of AI research. Reasoning has traditionally been the domain of symbolic AI, but recent advances in deep learning and in particular Large Language Models (LLMs) such as GPT-3 seem to suggest that LLMs have some latent reasoning ability.
To investigate this, we created a GPT-based reasoning agent that is intended to perform ontological reason using a few-shot learning approach, using instruction prompting and in-context examples. We also created a series of benchmarks to test ontological reasoning ability in LLMs and other systems.
Our results indicate that GPT is a poor reasoner, and is only able to perform ontological reasoning on some of the simplest tasks. Even on these simple tasks, results are highly variable, with performance degrading as the size of the ontology and the complexity of the explanation increases. In the cases where it does successfully perform the task, this seems to essentially be an advanced pattern-based form of lookup.
Our results indicate that a maximalist approach to using LLMs may be limiting, and that to be successful AI should use hybrid strategies.
Citation by DOI [1].
..
blah
blah
blah