Claude Sonnet 4.5 can operate autonomously longer than ever

Anthropic announced Claude Sonnet 4.5, an AI model that can operate autonomously for up to 30 hours on complex tasks. The company demonstrated this capability by having the model create a chat application similar to Slack, producing 11,000 lines of code before stopping upon task completion. The new model represents a significant improvement over Anthropic’s …

Read more

Qwen3-Omni is an open-source model for text, image, audio, and video

The Chinese technology company Alibaba has released Qwen3-Omni, a new generative AI model that can process a combination of text, images, audio, and video. The model is notable for its “omni-modal” capabilities and its open-source license, positioning it as a direct competitor to proprietary models from U.S. tech companies like OpenAI and Google. According to …

Read more

Hugging Face releases small language model with full training recipe

Hugging Face has launched SmolLM3, a new 3-billion-parameter language model designed for high performance and efficiency. In their official post, the company states that the model outperforms others in its size class and is competitive with some larger alternatives. A key feature is its dual-mode capability, which allows it to provide direct answers or show …

Read more

MiniMax launches powerful open-source AI model with big context window

Chinese AI company MiniMax has released MiniMax-M1, a new open-source language model that can process up to one million tokens of context. This makes it capable of handling entire book collections in a single conversation, reports Carl Franzen for VentureBeat. The model is available for free commercial use under an Apache 2.0 license on platforms …

Read more

Chinese startup DeepSeek releases major update

Chinese AI startup DeepSeek has released a significant update to its open-source reasoning model, bringing it closer to competing with paid services from OpenAI and Google. The new DeepSeek-R1-0528 model shows substantial improvements in complex reasoning tasks across mathematics, science, and programming. VentureBeat’s Carl Franzen reports that the updated model achieved 87.5% accuracy on the …

Read more

Google introduces fast new AI model using diffusion technology

Google unveiled Gemini Diffusion at its I/O developer conference, marking a significant shift in how AI models generate text. The experimental model uses diffusion technology instead of the traditional transformer approach that powers ChatGPT and similar systems. The key advantage is speed. Gemini Diffusion generates text at 857 to 2,000 tokens per second, which is …

Read more

Prime Intellect releases model trained with decentralized reinforcement learning

Prime Intellect has introduced INTELLECT-2, a 32B parameter AI model that represents the first of its kind trained using globally distributed reinforcement learning. The model employs a decentralized approach, utilizing compute resources from contributors around the world rather than centralized GPU clusters. In a technical report, Prime Intellect details their custom-built infrastructure components, including PRIME-RL, …

Read more

Anthropic introduces “Max” plan for increased Claude AI usage

Anthropic has launched a new “Max” subscription tier for its Claude AI assistant, offering up to 20 times higher usage limits than its Pro plan. According to the announcement from Anthropic today, the new plan is designed for users who collaborate extensively with Claude and need expanded access for demanding projects. The Max plan comes …

Read more

Google makes Gemini 2.5 Pro widely available at competitive pricing

Google has announced that its Gemini 2.5 Pro model is now available in public preview through the Gemini API in Google AI Studio, with Vertex AI rollout expected shortly. According to Google, this model is their “most intelligent” to date and has been priced competitively at $1.24 per million input tokens and $10 per million …

Read more

Midjourney research aims to make LLMs write more creatively

Midjourney, primarily known for AI image generation, has released new research in collaboration with New York University on training large language models to produce more creative text. Carl Franzen reports for VentureBeat that the research introduces two new techniques: Diversified Direct Preference Optimization (DDPO) and Diversified Odds Ratio Preference Optimization (DORPO). These methods encourage LLMs …

Read more