Skip to Content

Inside GPT-01: Transforming Gen AI

Robert Engels
Oct 14, 2024

The latest GPT-01 series is outstanding, just like previous releases.

In my tests on challenging tasks that typically confuse Large Language Models (LLMs), it consistently delivered accurate and expected results. The performance is truly impressive. The following are my key observations: 
 

  1. Multi-agent systems and chain-of-thought (CoT) integration: The GPT-01 series generates results through a multi-step process. Recently developed multi-agent systems with various roles and different setups incorporating chain-of-thought and critics have become integral to the system, significantly reducing errors. 
  • Example: OpenAI’s research on multi-agent reinforcement learning (MARL) demonstrates how multiple agents can collaborate to solve complex tasks. For instance, in a simulated environment, agents can work together to achieve a common goal, such as navigating a maze or playing a team-based game. [1] 
  1. Deterministic prompting: Initial testing indicates that the new models exhibit more deterministic behavior. This means that prompts for the same task, even when phrased differently, deliver more consistent answers. This improvement addresses a major issue seen in previous versions, where the phrasing of a prompt could drastically affect the response quality. 
  • Example: Research on prompt engineering has shown that models like GPT-3 can be fine-tuned to provide more consistent responses. For example, a study found that by using specific prompt templates, the variability in responses to similar questions was significantly reduced. [2] 
  1. Reinforcement learning enhancements: Reinforcement learning continues to play a crucial role. This technique, which was pivotal in earlier versions, now helps implement best practices around chain-of-thought, contributing to the model’s enhanced performance. 
  • Example: DeepMind’s AlphaGo used reinforcement learning to master the game of Go. By playing millions of games against itself and learning from each outcome, AlphaGo improved its performance to a superhuman level. [3] 
  1. Transparency in reasoning steps: While the current reasoning steps in the chain-of-thought process are not fully explained or inspectable, this is acceptable for many use cases. However, it remains a concern for applications where correctness is critical. 
  • Example: The Explainable AI (XAI) initiative by Defense Advanced Research Projects Agency (DARPA) aims to create AI systems whose decisions can be understood and trusted by humans. For example, in medical diagnostics, XAI techniques are used to provide insights into how an AI model arrived at a particular diagnosis, making the process more transparent and trustworthy. [4] 

 The new GPT-01 series demonstrates significant improvements in performance and reliability. However, the underlying model still faces challenges with hallucinations, particularly in scenarios where accuracy is critical. As highlighted by Sourav Banerjee in the referred paper (see below), LLMs will continue to experience errors and hallucinations. It is essential to learn how to manage these limitations effectively. The latest release makes this task somewhat easier, but the journey towards perfect accuracy continues.

References: 

[1] https://openai.com/index/emergent-tool-use/ 

[2] https://www.nature.com/articles/s41746-024-01029-4.pdf 

[3] https://deepmind.google/research/breakthroughs/alphago/  

[4] https://www.darpa.mil/program/explainable-artificial-intelligence  

Meet the author

Robert Engels

Global CTIO and Head of Lab for AI Futures and Insights & Data
Robert is an innovation lead and a thought leader in several sectors and regions, and holds the position of Chief Technology Officer for Northern and Central Europe in our Insights & Data Global Business Line. Based in Norway, he is a known lecturer, public speaker, and panel moderator. Robert holds a PhD in artificial intelligence from the Technical University of Karlsruhe (KIT), Germany.