Sakana AI introduces Continuous Thought Machines, a novel neural network that mimics brain processes

Sakana AI, co-founded by former Google AI scientists, has unveiled a new neural network architecture called Continuous Thought Machines (CTM). Unlike traditional transformer-based models that process information in parallel, CTMs incorporate a time-based dimension that mimics how biological brains operate, allowing for more flexible and adaptive reasoning. The key innovation in CTMs is their treatment … Read more

New benchmark reveals leading AI models confidently produce false information

A new benchmark called Phare has revealed that leading large language models (LLMs) frequently generate false information with high confidence, particularly when handling misinformation. The research, conducted by Giskard with partners including Google DeepMind, evaluated top models from eight AI labs across multiple languages. The Phare benchmark focuses on four critical domains: hallucination, bias and … Read more

Scientists struggle to understand how LLMs work

Researchers building large language models (LLMs) face a major challenge in understanding how these AI systems actually function, according to a recent article in Quanta Magazine by James O’Brien. The development process resembles gardening more than traditional engineering, with scientists having limited control over how models develop. Martin Wattenberg, a language model researcher at Harvard … Read more

Study finds LM Arena may favor major AI labs in its benchmarking

A new study by researchers from Cohere, Stanford, MIT, and Ai2 alleges that LM Arena, the organization behind the Chatbot Arena AI benchmark, provided preferential treatment to major AI companies. According to Maxwell Zeff’s TechCrunch report, companies like Meta, OpenAI, Google, and Amazon were allowed to privately test multiple model variants and only publish scores … Read more

AI helps scientists develop new experiments and discoveries

AI systems are increasingly being used to design experiments and drive scientific discoveries, according to research highlighted in Quanta Magazine. Mario Krenn, a quantum physicist who now leads the Artificial Scientist Lab, developed an AI program called Melvin that successfully designed quantum physics experiments when humans were stuck. Gregory Barber, writing for Quanta Magazine, describes … Read more

Google DeepMind researchers predict “Era of Experience” in AI

Google DeepMind’s David Silver and Richard S. Sutton predict a major shift in artificial intelligence development, which they call the “Era of Experience.” In a preprint paper for MIT Press, the researchers argue that AI will increasingly learn from its own experiences rather than human-generated data. The authors suggest that current AI systems, particularly large … Read more

Anthropic develops method to analyze AI’s values in real conversations

Anthropic, the company behind the AI assistant Claude, has developed a new technique to observe and analyze how its AI expresses values during real-world conversations with users. The research, conducted by Anthropic’s Societal Impacts team, examines whether Claude adheres to the company’s goal of making it “helpful, honest, and harmless” when interacting with users. The … Read more

Midjourney research aims to make LLMs write more creatively

Midjourney, primarily known for AI image generation, has released new research in collaboration with New York University on training large language models to produce more creative text. Carl Franzen reports for VentureBeat that the research introduces two new techniques: Diversified Direct Preference Optimization (DDPO) and Diversified Odds Ratio Preference Optimization (DORPO). These methods encourage LLMs … Read more

Anthropic reveals insights into Claude’s internal thought processes

Anthropic has published new research that sheds light on how its AI assistant Claude “thinks” internally. Two recent papers explore the model’s internal mechanisms through a novel interpretability approach the company compares to an “AI microscope.” This research reveals several surprising findings about Claude’s cognitive processes, including how it handles multiple languages, plans ahead when … Read more

New AI techniques promise huge cost savings and improved performance for enterprises

Recent research has unveiled two promising approaches that could dramatically reduce the costs of running large language models (LLMs) while simultaneously improving their performance on complex reasoning tasks. These innovations come at a critical time as enterprises increasingly deploy AI solutions but struggle with computational expenses. Chain of draft: Less is more Researchers at Zoom … Read more