MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) has delved deeply into the realm of generative AI, aiming to harness its potential and improve its fidelity. In a recent paper authored by Yilun Du, in collaboration with CSAIL members Shuang Li SM ’20, Ph.D. ’23; MIT professors Antonio Torralba and Joshua Tenenbaum; and Google DeepMind’s Igor Mordatch, the CSAIL team has pioneered a unique strategy that employs multiple AI systems to debate amongst themselves. By doing so, they aim to pinpoint the most accurate answer to a question. The essence of this methodology is to enable these large language models (LLMs) to bolster their commitment to facts and enhance their decision-making capabilities.
The fundamental challenge with LLMs stems from their occasional inconsistent responses, which can lead to errors and faulty logic. MIT’s novel tactic lets each AI agent critically evaluate the responses of its peers, fine-tuning its own answer based on this collaborative critique. This method involves multiple cycles of generating answers, receiving feedback, and then revising based on the insights. Picture a brainstorming session, where each participant iteratively refines their thoughts to converge on a consensus that’s well-considered.
“Employing a novel approach, we don’t simply rely on a single AI model for answers. Instead, our process enlists a multitude of AI models, each bringing unique insights to tackle a question. Although their initial responses may seem truncated or may contain errors, these models can sharpen and improve their own answers by scrutinizing the responses offered by their counterparts,” said Yilun Du, an MIT PhD student in electrical engineering and computer science, affiliate of MIT CSAIL, and lead author on a new paper about the work. “As these AI models engage in discourse and deliberation, they’re better equipped to recognize and rectify issues, enhance their problem-solving abilities, and better verify the precision of their responses. Essentially, we’re cultivating an environment that compels them to delve deeper into the crux of a problem. This stands in contrast to a single, solitary AI model, which often parrots content found on the internet. Our method, however, actively stimulates the AI models to craft more accurate and comprehensive solutions.”
A salient feature of this method is its adaptability to pre-existing models without delving into their internal mechanics. Given its text-based nature, the approach can be applied across a range of LLMs, enhancing the reliability and precision of their results. Their experiments in mathematical problem-solving, spanning elementary to high school math, exhibited marked improvement when the multi-agent debate technique was employed. Moreover, the LLMs showcased a heightened proficiency in producing correct arithmetic calculations, signalling its applicability across varied sectors.
Furthermore, the method offers a remedy for the occasional “hallucinations” seen in language models. By fostering a culture of inter-agent critiques, the AI systems are dissuaded from producing arbitrary data, emphasizing factual correctness instead.
This approach’s versatility doesn’t end at language models. It holds potential in amalgamating varied models with niche skills. By orchestrating a multi-agent dialogue system, these agents can synergize their specialized skills for tasks across different media, be it speech, video, or textual content.
However, the researchers also acknowledge the room for improvement. Current LLMs may grapple with very lengthy contexts, and their critique faculties might need refining. The current multi-agent format, while mirroring human group dynamics, could further encapsulate the intricacies of our discussions to boost collective decision-making. As the team foresees, the road ahead might involve grasping the deeper computational intricacies behind human dialogues and leveraging them to supplement or refine existing LLMs.
“Not only does this approach offer a pathway to elevate the performance of existing language models, but it also presents an automatic means of self-improvement. By utilizing the debate process as supervised data, language models can enhance their factuality and reasoning autonomously, reducing reliance on human feedback and offering a scalable approach to self-improvement,” said Du. “As researchers continue to refine and explore this approach, we can get closer to a future where language models not only mimic human-like language but also exhibit more systematic and reliable thinking, forging a new era of language understanding and application.”
“It makes so much sense to use a deliberative process to improve the model’s overall output, and it’s a big step forward from chain-of-thought prompting,” added Anca Dragan, associate professor at the University of California at Berkeley’s Department of Electrical Engineering and Computer Sciences, who was not involved in the work. “I’m excited about where this can go next. Can people better judge the answers coming out of LLMs when they see the deliberation, whether or not it converges? Can people arrive at better answers by themselves deliberating with an LLM? Can a similar idea be used to help a user probe a LLM’s answer in order to arrive at a better one?”
Featured image: MSem69, PlaygroundAI