
Artificial intelligence (AI) has made significant strides in recent years, yet challenges persist in achieving efficient, cost-effective, and high-performance models. Developing large language models (LLMs) often requires substantial computational resources and financial investment, which can be prohibitive for many organizations. Additionally, ensuring that these models possess strong reasoning capabilities and can be deployed effectively on consumer-grade hardware remains a hurdle.
DeepSeek AI has addressed these challenges head-on with the release of DeepSeek-V3-0324, a significant upgrade to its V3 large language model. This new model not only enhances performance but also operates at an impressive speed of 20 tokens per second on a Mac Studio, a consumer-grade device. This advancement intensifies the competition with industry leaders like OpenAI, showcasing DeepSeek’s commitment to making high-quality AI models more accessible and efficient.
DeepSeek-V3-0324 introduces several technical improvements over its predecessor. Notably, it demonstrates significant enhancements in reasoning capabilities, with benchmark scores showing substantial increases:
MMLU-Pro: 75.9 → 81.2 (+5.3)
GPQA: 59.1 → 68.4 (+9.3)
AIME: 39.6 → 59.4 (+19.8)
LiveCodeBench: 39.2 → 49.2 (+10.0)
These improvements indicate a more robust understanding and processing of complex tasks. Additionally, the model has enhanced front-end web development skills, producing more executable code and aesthetically pleasing web pages and game interfaces. Its Chinese writing proficiency has also seen advancements, aligning with the R1 writing style and improving the quality of medium-to-long-form content. Furthermore, function calling accuracy has been increased, addressing issues present in previous versions.
The release of DeepSeek-V3-0324 under the MIT License underscores DeepSeek AI’s dedication to open-source collaboration, allowing developers worldwide to utilize and build upon this technology without restrictive licensing constraints. The model’s ability to run efficiently on devices like the Mac Studio, achieving 20 tokens per second, exemplifies its practical applicability and efficiency. This performance level not only makes advanced AI more accessible but also reduces the dependency on expensive, specialized hardware, thereby lowering the barrier to entry for many users and organizations.
In conclusion, DeepSeek AI’s release of DeepSeek-V3-0324 marks a significant milestone in the AI landscape. By addressing key challenges related to performance, cost, and accessibility, DeepSeek has positioned itself as a formidable competitor to established entities like OpenAI. The model’s technical advancements and open-source availability promise to democratize AI technology further, fostering innovation and broader adoption across various sectors.
Check out the Model on Hugging Face. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 85k+ ML SubReddit.

Asif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.
Be the first to comment