Open model DeepSeek-V3 performs similar to closed competition

Chinese AI startup DeepSeek has launched DeepSeek-V3, a powerful new AI model that outperforms existing open-source alternatives. According to reporting by Shubham Sharma at VentureBeat, the model features 671 billion parameters but activates only 37 billion for each task through its mixture-of-experts architecture. The model was trained on 14.8 trillion diverse tokens and demonstrates superior …

Read more

Report: Microsoft diversifies AI models for 365 Copilot

Microsoft is expanding beyond OpenAI’s technology for its Microsoft 365 Copilot service, according to a Reuters report by Anna Tong and Krystal Hu. The company is working to integrate both internal and third-party AI models to reduce costs and improve performance. While OpenAI remains a partner for advanced AI models, Microsoft aims to make the …

Read more

Perplexity acquires data integration startup Carbon

AI search company Perplexity has acquired Carbon, a startup specializing in data retrieval frameworks. According to Shubham Sharma’s report in VentureBeat, this acquisition aims to help enterprises better connect their internal data sources to AI search capabilities. Carbon’s technology enables integration with over 20 data connectors and supports various file formats including text, audio, and …

Read more

Stable Diffusion 3.5 launches on Amazon’s enterprise AI platform

Stability AI has released its latest text-to-image generation model, Stable Diffusion 3.5 Large, on Amazon’s Bedrock service. As reported by Sean Michael Kerner for VentureBeat, this marks AWS as the only public cloud service offering Stability AI’s flagship models. The integration allows enterprises to access multiple AI models through a single unified API, meeting the …

Read more

New AI evaluation model Glider matches GPT-4’s performance with fewer resources

Startup Patronus AI has developed a breakthrough AI evaluation model that achieves comparable results to much larger systems while using significantly fewer computational resources. As reported by Michael Nuñez for VentureBeat, the new open-source model named Glider uses only 3.8 billion parameters yet matches or exceeds the performance of GPT-4 on key benchmarks. The model …

Read more

IBM launches improved Granite 3.1 language models

IBM has released a new version of its open-source large language models, Granite 3.1, featuring significant improvements in performance and capabilities. According to reporting by Sean Michael Kerner for VentureBeat, the new models offer extended context length and integrated hallucination detection. The Granite 8B Instruct model reportedly outperforms similar-sized competitors including Meta Llama 3.1 and …

Read more

Nvidia and DataStax launch storage-efficient AI retrieval system

Nvidia and DataStax have introduced a new AI technology that reduces data storage requirements by 35 times for generative AI systems. As reported by Michael Nuñez for VentureBeat, the Nvidia NeMo Retriever microservices, integrated with DataStax’s AI platform, enables faster and more accurate information retrieval across multiple languages. The technology has already shown impressive results …

Read more

Slack integrates Salesforce AI agents for workplace automation

Slack is implementing Salesforce’s Agentforce AI agents into its collaboration platform to enhance workplace productivity. According to Michael Nuñez at VentureBeat, the integration will give AI agents access to organizational conversations and data within Slack channels. Rob Seaman, Slack’s chief product officer, emphasized that the system will provide AI agents with contextual knowledge, reasoning capabilities, …

Read more

Cohere launches new compact AI language model Command R7B

AI company Cohere has introduced Command R7B, a new compact language model designed for enterprise applications. According to VentureBeat reporter Taryn Plumb, the model supports 23 languages and specializes in retrieval-augmented generation (RAG). Command R7B outperforms similar-sized models from competitors like Google, Meta, and Mistral in mathematics and coding tasks. The model features a 128K …

Read more

Sakana AI develops new memory optimization for language models

Tokyo-based startup Sakana AI has created a breakthrough technique that reduces memory usage in large language models by up to 75%. As reported by Ben Dickson, the system called “universal transformer memory” uses neural attention memory modules (NAMMs) to efficiently manage information processing. These modules analyze the model’s attention layers to determine which information to …

Read more