DeepSeek releases new reasoning models and introduces distilled versions

Chinese AI company DeepSeek has announced the release of its new reasoning-focused language models DeepSeek-R1-Zero and DeepSeek-R1, along with six smaller distilled versions. The main models, built on DeepSeek’s V3 architecture, feature 671 billion total parameters with 37 billion activated parameters and a context length of 128,000 tokens. According to company statements, DeepSeek-R1 achieves performance … Read more

Diffbot launches new AI model with real-time fact checking

Diffbot, a Silicon Valley company, has introduced a new AI model that combines AI capabilities with real-time fact verification. As reported by Michael Nuñez for VentureBeat, the system uses graph retrieval-augmented generation (GraphRAG) technology based on Meta’s Llama 3.3. The model connects to Diffbot’s Knowledge Graph, a database containing over one trillion facts that updates … Read more

Nvidia announces new AI models and technologies at CES 2025

Nvidia has unveiled multiple new AI initiatives at CES 2025, centered around their Nemotron model families and Cosmos World Foundation Models. The company’s CEO Jensen Huang presented these developments during his opening keynote, introducing AI models designed to advance both enterprise and consumer applications. The Nemotron family includes language and vision models available as NIM … Read more

Nvidia introduces “desktop AI supercomputer” Project Digits for $3,000

At CES 2025 in Las Vegas, Nvidia announced Project Digits, a compact desktop AI supercomputer aimed at researchers, data scientists, and students. The device, scheduled for release in May 2025 at a price point of $3,000, represents the company’s effort to bring powerful AI computing capabilities to individual desks. At the core of Project Digits … Read more

Tested: DeepSeek-V3 matches top AI models at lower cost

A detailed analysis published by Sunil Kumar Dash reveals that DeepSeek’s latest AI model achieves performance comparable to leading closed-source models while offering significant cost advantages. The model outperforms existing open-source alternatives in mathematics and reasoning tasks, according to extensive benchmark testing. The analysis demonstrates that DeepSeek-V3 surpasses GPT-4 and Claude 3.5 Sonnet in mathematical … Read more

Developer shares guide for running AI models locally

A detailed guide for running large language models (LLMs) on personal computers has been published by software developer Abishek Muthian on his blog. The article provides a thorough overview of hardware requirements, essential tools, and recommended models for local LLM deployment. Muthian emphasizes that while he uses high-end hardware including a Core i9 CPU and … Read more

Engineer details DIY setup for training AI language models

A detailed guide for building a powerful AI training system has been published by machine learning engineer Sabareesh Subramani on his personal website. The setup, costing approximately $12,000, uses four NVIDIA 4090 graphics cards to train large language models (LLMs) similar to but much smaller than ChatGPT. The system can effectively train AI models with … Read more

Open model DeepSeek-V3 performs similar to closed competition

Chinese AI startup DeepSeek has launched DeepSeek-V3, a powerful new AI model that outperforms existing open-source alternatives. According to reporting by Shubham Sharma at VentureBeat, the model features 671 billion parameters but activates only 37 billion for each task through its mixture-of-experts architecture. The model was trained on 14.8 trillion diverse tokens and demonstrates superior … Read more

New AI evaluation model Glider matches GPT-4’s performance with fewer resources

Startup Patronus AI has developed a breakthrough AI evaluation model that achieves comparable results to much larger systems while using significantly fewer computational resources. As reported by Michael Nuñez for VentureBeat, the new open-source model named Glider uses only 3.8 billion parameters yet matches or exceeds the performance of GPT-4 on key benchmarks. The model … Read more

Nvidia and DataStax launch storage-efficient AI retrieval system

Nvidia and DataStax have introduced a new AI technology that reduces data storage requirements by 35 times for generative AI systems. As reported by Michael Nuñez for VentureBeat, the Nvidia NeMo Retriever microservices, integrated with DataStax’s AI platform, enables faster and more accurate information retrieval across multiple languages. The technology has already shown impressive results … Read more