Article Alert: Can Large Language Models Reason and Plan

Author: Subbarao Kambhampati (Arizona State University)

Abstract: While humans sometimes do show the capability of correcting their own erroneous guesses with self-critiquing, there seems to be no basis for that assumption in the case of LLMs.

Please click here to read the full article. The article is also available on arXiv.

Image credit: Image by Freepik

Your account