Meta’s Llama AI models reach 1 billion downloads

Meta’s Llama AI models have reached 1 billion downloads, according to CEO Mark Zuckerberg’s announcement on Threads. Kyle Wiggers reports that this represents a 53% increase from the 650 million downloads recorded in December 2024. The models, available under a proprietary license, power Meta’s AI assistant across Facebook, Instagram, and WhatsApp. Despite commercial restrictions in … Read more

Nvidia unveils Llama Nemotron models to advance AI agents and reasoning capabilities

At the GPU Technology Conference (GTC) 2025, Nvidia announced a new family of AI models called Llama Nemotron designed to enhance reasoning capabilities for autonomous AI agents. These models are based on Meta’s open-source Llama models but have been refined through post-training optimization techniques to improve their performance in complex tasks such as multistep math, … Read more

Nous Research launches AI model with optional reasoning mode

Nous Research has released DeepHermes-3, a new AI language model that allows users to switch between detailed reasoning and quick responses. As reported by Carl Franzen for VentureBeat, this 8-billion parameter model builds on Meta’s Llama technology. Users can activate a special reasoning mode that makes the AI show its thought process before providing answers. … Read more

DeepSeek releases new reasoning models and introduces distilled versions

Chinese AI company DeepSeek has announced the release of its new reasoning-focused language models DeepSeek-R1-Zero and DeepSeek-R1, along with six smaller distilled versions. The main models, built on DeepSeek’s V3 architecture, feature 671 billion total parameters with 37 billion activated parameters and a context length of 128,000 tokens. According to company statements, DeepSeek-R1 achieves performance … Read more

New AI model LlamaV-o1 explains its reasoning process

Researchers at the Mohamed bin Zayed University of Artificial Intelligence have developed a new AI model that shows how it arrives at its conclusions. As reported by Michael Nuñez for VentureBeat, LlamaV-o1 combines visual and textual analysis while providing step-by-step explanations of its reasoning process. The model excels at complex tasks like interpreting financial charts … Read more

Meta’s Llama 3.3 70B model runs GPT-4 level AI on high-end laptops

Meta has released Llama 3.3 70B, a new large language model that achieves GPT-4 level performance while running on high-end consumer laptops. The breakthrough was documented by developer Simon Willison testing the model on a 64 GB MacBook Pro M2, demonstrating capabilities comparable to much larger models like Meta’s own Llama 3.1 405B. The model … Read more

Meta uses OpenAI’s GPT-4 alongside Llama in internal coding tool

Meta’s internal coding assistant Metamate combines both OpenAI’s GPT-4 and Meta’s own Llama AI model to support developers, according to a report by Kali Hays in Fortune. The tool has been using GPT-4 since early 2024, despite CEO Mark Zuckerberg’s public promotion of Llama as a leading AI model. Current and former Meta employees, speaking … Read more

Meta rebuilds company strategy around open-source AI model Llama

Meta has fundamentally transformed its business strategy by focusing on Llama, its open-source artificial intelligence model. According to Sharon Goldman’s detailed report in Fortune, CEO Mark Zuckerberg made the pivotal decision to release Llama 2 as open-source in July 2023, despite internal concerns about monetization and security risks. The model has since been downloaded over … Read more

AnyChat unifies access to multiple AI language models

AnyChat, a new development tool, enables seamless integration of multiple large language models (LLMs) through a single interface. Developer Ahsen Khaliq, machine learning growth lead at Gradio, created the platform to allow users to switch between models like ChatGPT, Google’s Gemini, Perplexity, Claude, and Meta’s LLaMA without being restricted to one provider, as reported by … Read more

Cerebras Inference achieves breakthrough performance for Llama 3.1-70B

Cerebras has announced a major update to its Cerebras Inference platform, which now runs the Llama 3.1-70B language model at an impressive 2,100 tokens per second – a threefold performance increase compared to the previous release. According to James Wang from the official Cerebras blog, this performance is 16 times faster than the fastest GPU … Read more