- Verses AI (CBOE:VERS) announced a major achievement in its latest “Mastermind” challenge, where its flagship product, Genius, outperformed OpenAI’s o1-preview model and China’s leading AI model, DeepSeek’s R1
- This demonstration showcased Genius’s superior performance in the code-breaking game Mastermind, highlighting its speed, efficiency, and cost-effectiveness
- Genius solved the code 100 per cent of the time in a consistent number of steps, while DeepSeek solved the game only 45 per cent of the time
- Verses AI stock (CBOE:VERS) last traded at C$1.40
Verses AI (CBOE:VERS) announced a major achievement in its latest “Mastermind” challenge, where its flagship product, Genius, outperformed OpenAI’s o1-preview model and China’s leading AI model, DeepSeek’s R1.
This demonstration showcased Genius’s superior performance in the code-breaking game Mastermind, highlighting its speed, efficiency, and cost-effectiveness.
In a side-by-side comparison, Genius demonstrated its prowess by winning the Mastermind challenge against DeepSeek’s R1 model, which has been positioned as a disruptive and more cost-efficient large language model. Genius not only outperformed DeepSeek’s R1 model but did so 245 times faster and at a cost 779 times cheaper.
During the challenge, both models attempted to crack the Mastermind code in 100 games, each within up to 10 guesses. Each guess provided a hint, requiring the models to reason about the missing parts of the correct answer, with all six code colors needing to be correct to crack the code.
You can play the game at mastermindgame.org.
Highlights
- Genius solved the code 100 per cent of the time in a consistent number of steps, while DeepSeek solved the game only 45 per cent of the time
- Genius consistently solved games in 1.1–4.5 seconds, whereas DeepSeek’s solve times averaged 934 seconds (approximately 15.5 minutes)
- Genius’s total compute time for 100 games was just over 5 minutes, compared to DeepSeek’s 26 hours
- Genius’s compute cost was estimated at $0.05 for all 100 games, compared to DeepSeek’s R1 model at US$38.94
Leadership insights
“This is a good showcase of Genius’ domain-specific model’s advanced performance in a multi-step reasoning problem,” Hari Thiruvengada, Verses’ chief technology officer said in a news release. “Mastermind was the perfect choice for this test because it requires multi-step logical reasoning, predictive cause-and-effect understanding, and dynamic adaptation to crack the code. This exercise highlights how Genius excels in step-by-step reasoning by leveraging domain agents using a Bayesian approach and Active Inference.”
“Many AI models struggle to efficiently deliver results and as demonstrated in this challenge, Genius’ ability to perform multi-step reasoning and dynamically adjust to feedback is crucial for enabling agents that are not only more efficient but, more importantly, accurate and reliable enough to operate in dynamic real-world scenarios,”, Verses’ CEO and founder, Gabriel René added. “While this test highlights the competitive advantages of our technology, we believe Genius is also highly complementary to large language models—enhancing their capabilities by providing the additional ‘brainpower’ needed to make AI agents smarter and more trustworthy within their domains.”
About Verses AI
Vancouiver-based Verses is a cognitive computing company building next-generation intelligent software systems modeled after the wisdom and genius of nature.
Verses AI stock (CBOE:VERS) last traded at C$1.40.
Join the discussion: Find out what everybody’s saying about this AI stock’s performance in the Atari Challenge on the Verses AI Inc. Bullboard and check out the rest of Stockhouse’s stock forums and message boards.
The material provided in this article is for information only and should not be treated as investment advice. For full disclaimer information, please click here.
(Top image generated by AI.)