Tests show strong performance of Google’s Gemini 2.0 Flash model

Independent developer Simon Willison has conducted extensive testing of Google’s newly announced Gemini 2.0 Flash model, documenting the results on his blog. The tests reveal significant capabilities in multimodal processing, spatial reasoning, and code execution. The model demonstrated exceptional accuracy in analyzing complex images, as shown in a detailed assessment of a crowded pelican photograph … Read more

Google launches Gemini 2.0 AI model with expanded capabilities and agent features

Google has announced Gemini 2.0, its latest artificial intelligence model that introduces significant advances in multimodal capabilities and autonomous agent features. The experimental version, Gemini 2.0 Flash, is being released first to developers and trusted testers through Google’s AI platforms. According to Google, the new model can generate text, images, and multilingual audio while operating … Read more

Amazon launches Nova family of AI models for text, image and video generation

Amazon Web Services has introduced Nova, a new family of artificial intelligence models designed for text, image and video generation. The announcement was made by CEO Andy Jassy at the AWS re:Invent conference. The Nova family consists of four text-generating models: Micro, Lite, Pro, and Premier. Micro, Lite, and Pro are immediately available to AWS … Read more

AnyChat unifies access to multiple AI language models

AnyChat, a new development tool, enables seamless integration of multiple large language models (LLMs) through a single interface. Developer Ahsen Khaliq, machine learning growth lead at Gradio, created the platform to allow users to switch between models like ChatGPT, Google’s Gemini, Perplexity, Claude, and Meta’s LLaMA without being restricted to one provider, as reported by … Read more

Mistral AI launches enhanced language model and ChatGPT competitor

French AI startup Mistral has unveiled Pixtral Large, a new 124-billion-parameter language model, alongside major updates to its Le Chat platform, reports Carl Franzen. The new model features advanced multimodal capabilities, including image processing and optical character recognition, while maintaining a significant context window of 128,000 tokens. The model is available for research purposes through … Read more

Moondream raises $4.5M for compact yet powerful AI vision-language model

Moondream, a startup backed by Felicis Ventures, Microsoft’s M12 GitHub Fund, and Ascend, has emerged from stealth with $4.5 million in pre-seed funding. According to VentureBeat’s Michael Nuñez, the company has developed an open-source vision-language model that boasts 1.6 billion parameters but matches the performance of models four times its size. The model, which can … Read more

Spirit LM is Meta’s first freely available multimodal model

Meta has launched Spirit LM, its first Spirit LM is Meta’s first freely available multimodal language model that integrates text and speech inputs and outputs, positioning it as a competitor to models like OpenAI’s GPT-4o. Developed by Meta’s Fundamental AI Research (FAIR) team, Spirit LM aims to enhance AI voice experiences by improving speech generation’s … Read more

ARIA is open and natively multimodal

ARIA is an open multimodal native mixture-of-experts model designed to integrate diverse forms of information for comprehensive understanding, outperforming existing proprietary models in various tasks. With 24.9 billion total parameters, it activates 3.9 billion and 3.5 billion parameters for visual and text tokens, respectively. The model is pre-trained on a substantial dataset comprising 6.4 trillion … Read more

Nvidia surprises with powerful, open AI models

Nvidia has released a powerful open-source AI model that rivals proprietary systems from industry leaders like OpenAI and Google. The model, called NVLM 1.0, demonstrates exceptional performance in vision and language tasks while also enhancing text-only capabilities. Michael Nuñez reports on this development for VentureBeat. The main model, NVLM-D-72B, with 72 billion parameters, can process … Read more

Meta Llama 3.2 is here

Meta has today released the new version of its AI model series: Llama 3.2, which for the first time includes vision models that can process both images and text. The larger versions with 11 and 90 billion parameters should be able to compete with closed systems like Claude 3 Haiku for image processing. Also new … Read more