Tested: DeepSeek-V3 matches top AI models at lower cost

A detailed analysis published by Sunil Kumar Dash reveals that DeepSeek’s latest AI model achieves performance comparable to leading closed-source models while offering significant cost advantages. The model outperforms existing open-source alternatives in mathematics and reasoning tasks, according to extensive benchmark testing. The analysis demonstrates that DeepSeek-V3 surpasses GPT-4 and Claude 3.5 Sonnet in mathematical …

Read more

Developer shares guide for running AI models locally

A detailed guide for running large language models (LLMs) on personal computers has been published by software developer Abishek Muthian on his blog. The article provides a thorough overview of hardware requirements, essential tools, and recommended models for local LLM deployment. Muthian emphasizes that while he uses high-end hardware including a Core i9 CPU and …

Read more

Engineer details DIY setup for training AI language models

A detailed guide for building a powerful AI training system has been published by machine learning engineer Sabareesh Subramani on his personal website. The setup, costing approximately $12,000, uses four NVIDIA 4090 graphics cards to train large language models (LLMs) similar to but much smaller than ChatGPT. The system can effectively train AI models with …

Read more

Open model DeepSeek-V3 performs similar to closed competition

Chinese AI startup DeepSeek has launched DeepSeek-V3, a powerful new AI model that outperforms existing open-source alternatives. According to reporting by Shubham Sharma at VentureBeat, the model features 671 billion parameters but activates only 37 billion for each task through its mixture-of-experts architecture. The model was trained on 14.8 trillion diverse tokens and demonstrates superior …

Read more

New AI evaluation model Glider matches GPT-4’s performance with fewer resources

Startup Patronus AI has developed a breakthrough AI evaluation model that achieves comparable results to much larger systems while using significantly fewer computational resources. As reported by Michael Nuñez for VentureBeat, the new open-source model named Glider uses only 3.8 billion parameters yet matches or exceeds the performance of GPT-4 on key benchmarks. The model …

Read more

Nvidia and DataStax launch storage-efficient AI retrieval system

Nvidia and DataStax have introduced a new AI technology that reduces data storage requirements by 35 times for generative AI systems. As reported by Michael Nuñez for VentureBeat, the Nvidia NeMo Retriever microservices, integrated with DataStax’s AI platform, enables faster and more accurate information retrieval across multiple languages. The technology has already shown impressive results …

Read more

Cohere launches new compact AI language model Command R7B

AI company Cohere has introduced Command R7B, a new compact language model designed for enterprise applications. According to VentureBeat reporter Taryn Plumb, the model supports 23 languages and specializes in retrieval-augmented generation (RAG). Command R7B outperforms similar-sized models from competitors like Google, Meta, and Mistral in mathematics and coding tasks. The model features a 128K …

Read more

Sakana AI develops new memory optimization for language models

Tokyo-based startup Sakana AI has created a breakthrough technique that reduces memory usage in large language models by up to 75%. As reported by Ben Dickson, the system called “universal transformer memory” uses neural attention memory modules (NAMMs) to efficiently manage information processing. These modules analyze the model’s attention layers to determine which information to …

Read more

Microsoft’s Phi-4 AI model achieves high performance with fewer resources

Microsoft has introduced a new AI model that delivers superior mathematical reasoning capabilities while using significantly less computing power than larger competitors. According to Michael Nuñez’s report in VentureBeat, the 14-billion-parameter Phi-4 model outperforms larger systems like Google’s Gemini Pro 1.5. The model excels particularly in mathematical problem-solving, achieving top scores on standardized math competition …

Read more

NitroFusion creates instant images on basic hardware

The University of Surrey’s Institute for People-Centred Artificial Intelligence (PAI) has unveiled NitroFusion, a revolutionary AI model that generates images in real-time as users type. The groundbreaking technology, developed by the university’s SketchX laboratory, operates on consumer-grade graphics cards, making it accessible to individual creators and small studios. Unlike existing image generation platforms that require …

Read more