Best Wallet - No KYC Crypto Wallet with Exclusive Airdrops and Hottest New Tokens - Download Now!

New Open-Source AI Model Challenges DeepSeek, Proves Less Can Be More

Key Takeaways

  • OpenThinker-32B rivals DeepSeek’s performance despite using far fewer training examples;
  • The model’s open-source nature allows for ongoing improvements from the AI research community;
  • While it trails DeepSeek in coding, OpenThinker-32B's efficiency and adaptability make it a strong competitor.
New Open-Source AI Model Challenges DeepSeek, Proves Less Can Be More

A group of researchers from universities and tech firms has introduced a new artificial intelligence (AI) model that competes with DeepSeek, one of China’s most advanced systems.

The open-source model, OpenThinker-32B, achieved similar or better results in key performance tests while requiring far less training data, according to the February 12 blog post.

OpenThinker-32B was trained using just 114,000 examples, a fraction of DeepSeek’s 800,000. The dataset, called OpenThoughts-114k, included detailed solutions, coding test cases, starter code, and subject-specific information.

What is BNB? The Truth Behind Binance Smart Chain (Animated)

Did you know?

Want to get smarter & wealthier with crypto?

Subscribe - We publish new crypto explainer videos every week!

Training took about 90 hours on four nodes, each equipped with eight H100 GPUs. Another dataset, containing 137,000 unverified samples, was processed separately using Italy’s Leonardo Supercomputer, consuming 11,520 A100 GPU hours in just 30 hours.

When tested, OpenThinker-32B delivered a 90.6% accuracy rate on the MATH500 benchmark, surpassing DeepSeek’s 89.4%. It also scored 61.6 on GPQA-Diamond, compared to DeepSeek’s 57.6, which shows strength in general reasoning tasks.

For coding tasks, OpenThinker-32B lagged slightly behind, scoring 68.9 against DeepSeek’s 71.2. Since OpenThinker-32B is open source, these numbers could improve as developers contribute refinements.

Built on Alibaba’s Qwen2.5-32B-Instruct language model, OpenThinker-32B supports a 16,000-token context window. While this is smaller than other AI models, it is still enough to handle complex equations and long programming tasks.

On February 13, Elon Musk revealed the latest version of xAI's chatbot, Grok 3. What can it do? Read the full story.

Aaron S. Editor-In-Chief
Having completed a Master’s degree in Economics, Politics, and Cultures of the East Asia region, Aaron has written scientific papers analyzing the differences between Western and Collective forms of capitalism in the post-World War II era.
With close to a decade of experience in the FinTech industry, Aaron understands all of the biggest issues and struggles that crypto enthusiasts face. He’s a passionate analyst who is concerned with data-driven and fact-based content, as well as that which speaks to both Web3 natives and industry newcomers.
Aaron is the go-to person for everything and anything related to digital currencies. With a huge passion for blockchain & Web3 education, Aaron strives to transform the space as we know it, and make it more approachable to complete beginners.
Aaron has been quoted by multiple established outlets, and is a published author himself. Even during his free time, he enjoys researching the market trends, and looking for the next supernova.

Loading...
binance
×
Verified

$600 WELCOME BONUS

Earn Huge Exclusive Binance Learners Rewards
5.0 Rating