What have been the public reviews of GPT-5 so far?

Sam Altman hearing a headset microphone on stage at an event

Public reception of GPT-5 has been mixed, noting both improvements and limitations. Reviewers indicate that GPT-5 offers a more user-friendly experience, effectively reasoning through complex questions and providing faster responses than previous models. OpenAI claims it feels like talking to a PhD-level expert, representing a significant step forward, albeit still viewed more as an iterative improvement rather than a revolutionary leap [4].

Concerns have been raised about the potential for misinformation, with some experts emphasizing the need for skepticism regarding performance claims and the challenges of AI hallucinations [6][5].

Follow Up Recommendations

Challenges in Aligning Human and Machine Generalisation

Fundamental Differences in Generalisation

One of the core challenges in aligning human and machine generalisation arises from the fundamental differences in how each system forms and applies general concepts. The text explains that humans tend to rely on sparse abstractions, conceptual representations, and causal models. In contrast, many current AI systems, particularly those based on statistical methods, derive generalisation from extensive data as correlated patterns and probability distributions. For instance, it is noted that "humans tend toward sparse abstractions and conceptual representations that can be composed or transferred to new domains via analogical reasoning, whereas generalisations in statistical AI tend to be statistical patterns and probability distributions"[1]. This misalignment in the nature of what is learnt and how it is applied stands as a primary barrier to effective alignment.

Conceptual and Methodological Misalignment

The text clearly highlights that the methodologies underlying human and machine generalisation differ significantly. While human generalisation is viewed in terms of processes (abstraction, extension, and analogy) and results (categories, concepts, and rules), AI generalisation is often cast primarily as the ability to predict or reproduce statistical patterns over large datasets. One passage states that "if we wish to align machines to human-like generalisation ability (as an operator), we need new methods to achieve machine generalisation"[1]. In effect, while humans can generalise fresh from a few examples and adapt these insights across tasks, machines often require heavy data reliance, leading to products that do not encapsulate the inherent flexibility of human cognition. This discrepancy makes it difficult to seamlessly integrate AI systems into human–machine teaming scenarios.

Challenges in Evaluation and Robustness

Another challenge concerns the evaluation of generalisation capabilities and ensuring robustness. AI evaluation methods typically rely on empirical risk minimisation by testing on data that is assumed to be drawn from the same distribution as training data. However, this approach is limited when it comes to out-of-distribution (OOD) data and subtle distributional shifts. The text reflects that statistical learning methods often require large amounts of data and may hide generalisation failures behind data memorisation or overgeneralisation errors (for example, hallucinations in language models)[1]. Moreover, deriving provable guarantees — such as robustness bounds or measures for distribution shifts — poses a further challenge. This is complicated by difficulties in ensuring that training and test data are truly representative and independent, which is crucial for meaningful evaluation of whether a model generalises in practice.

Human-AI Teaming and Realignment Mechanisms

Effective human–machine teaming requires that the outputs of AI systems align closely with human expectations, particularly in high-stakes or decision-critical contexts. However, the text highlights that when such misalignments occur (for example, when AI predictions diverge significantly from human assessments), developing mechanisms for realignment and error correction becomes critical. The text emphasizes the need for collaborative methods that support not only the final decision but also the reasoning process, stating that "when misalignments occur, designing mechanisms for realignment and error correction becomes critical"[1]. One aspect of the challenge is that human cognition often involves explicit explanations based on causal history, whereas many AI systems, especially deep models, operate as opaque black boxes. This discrepancy necessitates the incorporation of explainable prediction methods and neurosymbolic approaches that can provide insights into underlying decision logic.

Integrating Diverse Generalisation Methods

The text also outlines challenges in harmonising the strengths of different AI methods. It distinguishes among statistical methods, knowledge-informed generalisation methods, and instance-based approaches. Each of these has its own set of advantages and limitations. For example, statistical methods deliver universal approximation and inference efficiency, yet they often fall short in compositionality and explainability. In contrast, knowledge-informed methods excel at explicit compositionality and enabling human insight but might be constrained to simpler scenarios due to their reliance on formalised theories[1]. Integrating these varying methods into a unified framework that resonates with human generalisation processes is a critical but unresolved goal. Approaches like neurosymbolic AI are being explored as potential bridges, but they still face significant hurdles, particularly in establishing formal generalisation properties and managing context dependency.

Conclusion

In summary, aligning human and machine generalisation is multifaceted, involving conceptual, methodological, evaluative, and practical challenges. Humans naturally form abstract, composable, and context-sensitive representations from few examples, while many AI systems depend on extensive data and statistical inference, leading to inherently different forms of generalisation. Furthermore, challenges in measuring robustness, explaining decisions, and ensuring that AI outputs align with human cognitive processes exacerbate these differences. The text underscores the need for interdisciplinary approaches that combine observational data with symbolic reasoning, develop formal guarantees for generalisation, and incorporate mechanisms for continuous realignment in human–machine teaming scenarios[1]. Addressing these challenges will be essential for advancing AI systems that truly support and augment human capabilities.


ControlNet: A Breakthrough in Conditional Control for Image Synthesis

In recent years, the advent of text-to-image diffusion models has revolutionized how we generate images. These models allow users to input a descriptive text, which the model then transforms into a visual representation. However, enhancing control over the image generation process has become an essential focus in the field. This blog post discusses a novel approach named ControlNet, which adds conditional controls to text-to-image diffusion models, enabling more precise and context-aware image generation.

Understanding Text-to-Image Diffusion Models

Text-to-image diffusion models like Stable Diffusion work by gradually adding noise to an image and then reversing this process to generate new images from textual descriptions. These models are trained on vast datasets that help them learn to denoise images iteratively. The goal is to produce images that accurately reflect the input text. As stated in the paper, 'Image diffusion models learn to progressively denoise images and generate samples from the training domain'[1].

Despite their impressive capabilities, these models can struggle with specific instructions. For instance, when users require detailed shapes or context, the model may produce generic outputs. This limitation led to the development of Conditional Control, where the model learns to incorporate additional information, such as edges or poses, into its generation process. ControlNet was designed to leverage various conditions to enhance the specificity and relevance of the generated images.

Introducing ControlNet

ControlNet is a neural network architecture that integrates spatial conditioning controls into large pre-trained text-to-image diffusion models. The primary objective of ControlNet is to allow users to add dimensions of control that were not previously possible. The approach involves using a technique called 'learned conditions,' which allows the model to accept additional inputs, like edge maps or human poses, to influence the resulting image.

The authors describe ControlNet as follows: 'ControlNet allows users to add conditions like Canny edges (top), human pose (bottom), etc., to control the image generation of large pre-trained diffusion models'[1]. This means that rather than solely relying on textual prompts, users can provide additional contextual cues that guide the generation process more effectively.

Applications of ControlNet

 title: 'Figure 7: Controlling Stable Diffusion with various conditions without prompts. The top row is input conditions, while all other rows are outputs. We use the empty string as input prompts. All models are trained with general-domain data. The model has to recognize semantic contents in the input condition images to generate images.'
title: 'Figure 7: Controlling Stable Diffusion with various conditions without prompts. The top row is input conditions, while all other rows are outputs. We use the empty string as input prompts. All models are trained with general-domain data. The...Read More

ControlNet has shown promising results in various applications. It can create images based on input conditions without requiring an accompanying text prompt. For example, a sketch input or a depth map could be used as the sole input, and ControlNet would generate a corresponding image that accurately reflects the details in those inputs.

The paper details numerous experiments demonstrating how ControlNet improves the fidelity of generated images by integrating these additional conditions. For instance, when testing with edge maps, the model could produce images that adhere closely to the specified shapes and orientations dictated by the input, leading to “high-quality, detailed, and professional images”[1].

Methodology Behind ControlNet

The architecture of ControlNet involves adding layers that handle different kinds of inputs. It connects to pre-trained diffusion models while introducing zero-convolution layers, which help prevent the detrimental effects of noise during training. The flexibility of ControlNet allows it to adapt to various types of prompts seamlessly.

By leveraging a foundation of large pre-trained models, ControlNet also benefits from their robust performance while fine-tuning them specifically for new tasks. The authors highlight that “extensive experiments verify that ControlNet facilitates wider applications to control image diffusion models”[1]. This adaptability is crucial for tackling diverse use cases and ensuring that the model can respond accurately to its inputs.

Training and Performance

Table 1: Average User Ranking (AUR) of result quality and condition fidelity. We report the user preference ranking (1 to 5 indicates worst to best) of different methods.
Table 1: Average User Ranking (AUR) of result quality and condition fidelity. We report the user preference ranking (1 to 5 indicates worst to best) of different methods.

To train ControlNet, researchers employed a method that involves optimizing for a range of conditions simultaneously. This multifaceted training process equips the model to recognize and interpret various inputs consistently. The results showed significant improvements, particularly noted through user studies where participants ranked the quality and fidelity of generated images. ControlNet was often rated higher than models that only depended on text prompts, proving the effectiveness of incorporating additional controls[1].

Another compelling aspect discussed in the paper is the impact of training datasets on performance. The researchers illustrated that the model's training does not collapse when it is limited to fewer images, indicating its robustness in learning from varying quantities of data. Users were able to achieve desirable results even when the training set was significantly restricted[1].

Conclusion: The Future of Image Generation

In summary, ControlNet represents a significant advancement in the capabilities of text-to-image generation technologies. By integrating conditional controls, it offers users greater specificity and reliability in image creation. This added flexibility makes it particularly beneficial for artists and designers seeking to generate highly customized images based on various inputs.

As these models continue to evolve, the seamless integration of more complex conditions will likely lead to even more sophisticated image synthesis technologies. With ongoing enhancements and refinements, ControlNet positions itself as a powerful tool in the intersection of artificial intelligence and creative expression, paving the way for innovative applications across multiple domains.


What benchmarks prove TTD-DR's effectiveness?

 title: 'Flowcharts illustrating various research frameworks: Huggingface Open DR, GPT Researcher, Open Deep Research, and Test-Time Diffusion DR'

The effectiveness of the Test-Time Diffusion Deep Researcher (TTD-DR) is substantiated through rigorous evaluation across various benchmarks. Specifically, TTD-DR achieves state-of-the-art results on complex tasks, such as generating long-form research reports and addressing multi-hop reasoning queries. Notably, it significantly outperforms existing deep research agents in these areas, as evidenced by win rates of 69.1% and 74.5% compared to OpenAI Deep Research for two long-form benchmarks[1].

Furthermore, comprehensive evaluations showcase TTD-DR's superior performance in generating coherent and comprehensive reports, alongside its ability to find concise answers to challenging queries. This is demonstrated through various datasets, including 'LongForm Research' and 'DeepConsult'[1].


How did "AlphaGo" defeat human champions?

 title: 'Humans strike back: How Lee Sedol won a game against AlphaGo'

AlphaGo defeated human champions through a combination of advanced machine learning techniques and innovative gameplay strategies. The AI system utilized deep neural networks and reinforcement learning, allowing it to learn from vast amounts of gameplay data and improve over time. Initially, it was trained by playing numerous games against human opponents, after which it played against different versions of itself, continuously refining its algorithms based on successful moves and winning percentages[3].

One significant factor in its victories was AlphaGo's ability to work with an enormous number of potential board configurations—far surpassing human capabilities. Go is considered a significantly more complex game than chess, with an estimated 10 to the power of 170 possible board positions, requiring an AI like AlphaGo to assess an immense search space quickly[3][4].

During its matches against the world champion Lee Sedol, AlphaGo showcased unexpected and highly creative moves that disrupted conventional strategies. For example, in one match, AlphaGo executed a 'shoulder hit' move that had never been seen in professional play, displaying a level of cunning that surprised even seasoned players[2][3]. In contrast, Lee Sedol, despite being a top player, struggled to adapt to AlphaGo's aggressive and unconventional playing style, leading to his defeat in several games[5].

However, Lee managed to win one game in the series by employing a clever move known as the 'Hand of God,' exploiting AlphaGo's mistake during a critical phase of the game. This victory highlighted that while AlphaGo was incredibly powerful, it still had vulnerabilities that could be exploited by skilled human players. Nonetheless, AlphaGo's overall performance established it as one of the strongest Go players in history, defeating Lee Sedol 4-1 in their five-game match series[1][3].

Follow Up Recommendations

Test your knowledge about AI interactions.

What are the names of the two open-weight reasoning models introduced by OpenAI? 🤖
Difficulty: Easy
What technique is used by gpt-oss models to reduce their memory footprint? 💾
Difficulty: Medium
In what year was the gpt-oss model card document published? 📅
Difficulty: Hard
Space: Let’s explore the gpt-oss-120b and gpt-oss-20b Model Card

AlphaGo: Revolutionizing the Game of Go with Artificial Intelligence

Introduction to AlphaGo

The game of Go, known for its deep strategic complexity, has long been a benchmark for artificial intelligence (AI) development. Achieving excellence in Go presents significant challenges due to its vast search space and the difficulty in evaluating board positions. Researchers at DeepMind introduced AlphaGo, a system that combines deep neural networks with tree search techniques, marking a pivotal moment in AI's capability to compete against top human players. In a series of high-stakes games, AlphaGo defeated elite Go players, showcasing the profound implications of AI in cognitive games.

The Architecture of AlphaGo

AlphaGo employs a novel architecture that integrates two primary neural networks: a policy network and a value network. The policy network is designed to predict the next move by using a variety of input features from the board, such as the presence of stones and potential capture opportunities. This network is crucial for narrowing down the vast number of possible moves to those that are most promising. A notable achievement of this architecture is its ability to draw on dozens of human games, learning from the best strategies and developing its own superhuman plays.

The value network complements the policy network by estimating the eventual outcome of the game from any given board position. It evaluates positions on a scale of winning probability, effectively guiding the search process in a more informed manner. The training of these networks involved extensive supervised learning from historical games, enhancing their capabilities to better predict moves and evaluate game states.

Training via Reinforcement Learning

AlphaGo's training process involved a combination of supervised learning and reinforcement learning. Initially, it trained its policy network on over 30 million board positions sourced from human games. This training resulted in a model that could predict moves with remarkable accuracy, achieving a test accuracy of 57.5% against the state-of-the-art[1].

Once the policy network was established, the team implemented reinforcement learning through self-play. In this phase, AlphaGo played numerous games against itself, refining its skills through extensive exploration of strategies. The result was a program that not only mimicked human play but also developed unique strategies that even top players had never considered.

Monte Carlo Tree Search (MCTS)

A key element of AlphaGo's decision-making process is the use of Monte Carlo Tree Search (MCTS). This algorithm enhances the effectiveness of the neural networks by sampling possible future moves and simulating their outcomes. Essentially, MCTS builds a search tree where each node corresponds to a game state, enabling the system to evaluate the ramifications of decisions over numerous simulated games.

During the simulations, AlphaGo uses its policy network to select positions via probability distributions, which allows it to explore the most promising moves while balancing exploration and exploitation. This combination of MCTS with deep learning led to unprecedented efficiency and effectiveness in decision-making, ultimately allowing AlphaGo to outplay traditional Go programs, such as Crazy Stone and Zen, as well as human champions.

Evaluating AlphaGo's Performance

AlphaGo's introduction to competitive settings was marked by its match against European Go champion Fan Hui. In this series, AlphaGo won five out of five matches, one by a margin of 2.5 points, and the others by resignation. The performance metrics and strategies were scrutinized, revealing its superior capability to evaluate and execute moves autonomously[1].

None
None

Moreover, the effectiveness of AlphaGo was also tested against various Go programs in a tournament setting. The results were striking; AlphaGo demonstrated a substantial advantage, winning a vast majority of its games. Its performance against other AI competitors and human players showcased a significant leap in the field of artificial intelligence, highlighting the success of integrating deep learning with strategic game planning.

Implications for Artificial Intelligence

AlphaGo represents a landmark achievement in artificial intelligence, demonstrating that machines can not only learn from human behavior but can also innovate beyond traditional human strategies. The methods employed in developing AlphaGo have far-reaching implications for various fields, including robotics, healthcare, and any domain requiring strategic thinking and decision-making.

The success of AlphaGo has sparked interest in further research into deep reinforcement learning and its applications to other complex decision-making problems, showcasing the potential of AI in tackling tasks previously thought to be uniquely human.

Conclusion

The development of AlphaGo is a testament to the advancements in artificial intelligence, marking a significant milestone in the convergence of machine learning and cognitive strategy. Its ability to defeat top-tier players and traditional Go programs alike emphasizes the transformative power of AI, pushing the boundaries of what machines can achieve in complex domains. As research continues, the lessons learned from AlphaGo’s design and operational strategies will undoubtedly influence future AI systems across various sectors[1].


Quotes on system-2 reasoning in AI agents

Agent frameworks still rely on human-defined workflows to structure their actions.
[1]
System 2 encompasses slow, deliberate, and analytical thinking, which is crucial for solving complex tasks...
[1]
Deliberate, structured, and analytical thinking, enabling agents to handle complex, multi-step tasks that go beyond the reactive behaviors of system 1.
[1]
By incorporating reflective thinking, the model continuously evaluates its past actions, identifies potential mistakes, and adjusts its behavior to improve performance over time.
[1]
Enabling the model to express its decision-making process explicitly, fostering better alignment with task objectives.
[1]
Space: Browser AI Agents

Contributions of Self-Supervised Learning to AI

'a blue sphere in space'
title: 'Self-Supervised Learning and Its Applications' and caption: 'a blue sphere in space'

Self-supervised learning (SSL) has emerged as a transformative approach within the field of artificial intelligence (AI), particularly addressing the challenges associated with labeled data dependencies. This report highlights the essential contributions of SSL and examines its implications for various AI applications.

Reducing Dependency on Labeled Data

One of the primary contributions of self-supervised learning is its ability to significantly reduce the reliance on manual labeling of datasets. Traditional supervised learning methods require vast amounts of labeled data, which can be costly and time-consuming to produce. In contrast, SSL generates implicit labels from unstructured data, leveraging the inherent structures and patterns within the data itself. This innovation has made SSL a game-changer for AI, particularly in sectors where annotated data is scarce or difficult to obtain[2].

Applications in Multiple Domains

Top-1 accuracy for dermatology condition classification for MICLe, SimCLR, and supervised models under different unlabeled pretraining dataset and varied sizes of label fractions
title: 'Top-1 accuracy for dermatology condition classification for MICLe, SimCLR, and supervised models under different unlabeled pretraining dataset and varied sizes of label fractions' and caption: 'a graph of different sizes and colors'

The versatility of self-supervised learning is evident across several domains, including computer vision, natural language processing (NLP), and healthcare. In computer vision, SSL techniques can enable models to learn quality representations from unlabeled images. For instance, tasks such as image reconstruction, colorization, and predicting future video frames exemplify how SSL can achieve meaningful insights without explicit supervision. As a result, SSL algorithms can accelerate the development of applications like image classification and object detection[2][1].

In NLP, self-supervised learning has facilitated advancements in language models like BERT and GPT. These models have utilized self-supervised objectives to understand and generate language. BERT, for instance, employs techniques such as Next Sentence Prediction, allowing the model to understand relationships between sentences, hence improving various language comprehension tasks[1]. This self-supervised training has led to significant improvements in tasks such as sentiment analysis, translation, and text generation[2].

Cost-Effective and Time-Efficient Solutions

Self-supervised learning addresses several persistent issues in other learning procedures, most notably the high costs associated with labeled data. By mitigating the need for extensive manual annotation, SSL reduces the financial and time burdens normally imposed by model training, thus enabling faster and more cost-effective development of AI systems[1][2]. This is especially relevant in fields like healthcare, where annotating medical images can be prohibitively expensive. SSL can analyze medical imaging data, facilitating the rapid development of diagnostic tools without the need for extensive labeled datasets.

Bridging Supervised and Unsupervised Learning

Semi-supervised learning
title: 'Semi-supervised learning' and caption: 'a diagram of a machine learning model'

SSL serves as a vital link between supervised and unsupervised learning techniques, capturing essential features and relationships within data through cleverly designed pretext tasks. In self-supervised learning, models tackle objectives generated from the data itself, transforming unsupervised tasks into supervised learning problems through the generation of pseudo-labels. These tasks can be creative assignments, predictive tasks, or distinctive learning experiences derived from data augmentations, which teach models to recognize patterns without the need for external labels[2][1].

For example, SSL models can learn to reconstruct images or predict elements of sequences, creating robust embeddings that can later be fine-tuned for specific supervised tasks with small amounts of labeled data. This blend of SSL with supervised learning enhances the efficacy and robustness of models, revealing its potential to boost performance in various applications[2][1].

Enhancements in Model Training and Generalization

BERT
title: 'BERT' and caption: 'a diagram of a mask'

Self-supervised learning has been pivotal in enhancing model training and generalization. By pre-training models on large unlabeled datasets, SSL allows for robust feature extraction, which is crucial for subsequent fine-tuning on specific tasks. This two-step training process—first generating strong feature representations and then adapting them for particular uses—results in greater model performance and generalization capabilities across different tasks and domains[1][2].

Scalability and Future Potential

The scalability of self-supervised learning presents significant opportunities for future research and application. As SSL models are trained on vast amounts of unlabeled data, the ambition is to continue pushing the boundaries of what AI systems can learn using fewer resources. Future trends may involve integrating SSL techniques with other methodologies, including reinforcement learning and transfer learning, to create adaptable models capable of learning continuously and responding to dynamic environments with minimal supervision[2][1].

Conclusion

Self-supervised learning has undoubtedly reshaped the landscape of artificial intelligence by providing solutions that alleviate the challenges posed by the necessity of labeled data. Its application across various fields highlights the approach's versatility and efficiency. As research and development continue, SSL is set to play a crucial role in the ongoing evolution and sophistication of AI technologies, promising to unlock new capabilities and improve accessibility in a data-driven world.


How well do you know GPT-5's multilingual abilities?

What is one of the languages that GPT-5 can perform well in? 🌍
Difficulty: Easy
How does GPT-5's performance in multilingual contexts compare to existing models? 🌐
Difficulty: Medium
What evaluation did GPT-5 models undergo to measure multilingual performance? 📊
Difficulty: Hard
Space: Let’s explore the GPT-5 Model Card