top of page
13 January 2024
Date Added:

Can large language models identify and correct their mistakes?

Gladys Tyen
Author(s):
Department/Study Group:
Research
Organisation:
Google
Country / International Org.:
USA
Short Description:
This article examines the ability of large language models (LLMs) to identify and correct reasoning errors, focusing on self-correction in Chain-of-Thought reasoning and its generalizability to unfamiliar tasks.
Abstract:

This article explores the abilities of large language models (LLMs) in identifying and correcting their reasoning mistakes, particularly in Chain-of-Thought (CoT) reasoning. It introduces the BIG-Bench Mistake dataset for evaluating LLMs' error detection and examines their performance in four key areas: detecting logical mistakes in CoT reasoning, using error detection as a correctness measure, correcting known errors, and generalizing mistake identification to new tasks. The findings highlight the current limitations of LLMs in mistake detection but suggest promise in backtracking methods for corrections and the potential of specialized models for improved error identification across different tasks.

mistake management
Tag(s):
Document Type:
Research
If download link is inactive access the document from below original source link
Download Document:

ReadAI

Comments
Rated 0 out of 5 stars.
Couldn’t Load Comments
It looks like there was a technical problem. Try reconnecting or refreshing the page.
bottom of page