Google researchers develop AI model that can learn continuously

Researchers at Google have introduced a new AI paradigm called “Nested Learning” to address a major weakness in current large language models (LLMs). Ben Dickson reports for VentureBeat that this approach could enable AI systems to learn and update their knowledge continuously after their initial training. Today’s LLMs are largely static. Their knowledge is limited …

Read more

This simple sentence can make AI models more creative

Researchers have developed a method called Verbalized Sampling that uses a single sentence to make generative AI models produce more diverse and creative responses. The technique works on large language models like GPT-4 and Claude without requiring any retraining. Carl Franzen reports for VentureBeat that this method addresses the common problem of AI models giving …

Read more

MIT researchers create AI models that can teach themselves

Researchers at the Massachusetts Institute of Technology (MIT) have developed a technique that allows large language models to improve on their own. Carl Franzen reports for VentureBeat that the method, called SEAL, enables AI to autonomously generate its own training data. Instead of relying on fixed external datasets, models using the SEAL framework create instructions …

Read more

University study suggests ChatGPT’s vocabulary is entering human speech

Researchers at Florida State University have found that buzzwords commonly used by AI are appearing more frequently in unscripted human conversations, McKenzie Harris reports for Florida State University News. The study analyzed 22.1 million words of spoken language, revealing a measurable increase in the use of words such as “delve,” “intricate,” and “underscore” after the …

Read more

Researchers develop human-like memory for AI

Chinese researchers have created a system named MemOS, designed to provide artificial intelligence with a persistent, human-like memory. According to a report by Michael Nuñez in VentureBeat, the technology addresses a fundamental limitation that causes AI models to forget information between user interactions. Current AI assistants often cannot recall past conversations, a problem the researchers …

Read more

Anthropic reveals how its multi-agent research system achieves 90% better performance

Anthropic has published detailed insights into how it built Claude’s research capabilities, revealing that its multi-agent system outperforms single-agent approaches by 90.2%. The post was written by Jeremy Hadfield, Barry Zhang, Kenneth Lien, Florian Scholz, Jeremy Fox, and Daniel Ford from Anthropic. The research feature allows Claude to search across the web, Google Workspace, and …

Read more

Stanford researchers develop test to measure AI chatbot flattery

Stanford University researchers have created a new benchmark to measure excessive flattery in AI chatbots after OpenAI rolled back updates to GPT-4o due to complaints about overly polite responses. The research, conducted with Carnegie Mellon University and University of Oxford, was reported by Emilia David. The team developed “Elephant,” a test that evaluates how much …

Read more

Google introduces fast new AI model using diffusion technology

Google unveiled Gemini Diffusion at its I/O developer conference, marking a significant shift in how AI models generate text. The experimental model uses diffusion technology instead of the traditional transformer approach that powers ChatGPT and similar systems. The key advantage is speed. Gemini Diffusion generates text at 857 to 2,000 tokens per second, which is …

Read more

DarkBench framework identifies manipulative behaviors in AI chatbots

AI safety researchers have created the first benchmark specifically designed to detect manipulative behaviors in large language models, following a concerning incident with ChatGPT-4o’s excessive flattery toward users. Leon Yen reported on the development for VentureBeat. The DarkBench framework, developed by Apart Research founder Esben Kran and collaborators, identifies six categories of problematic AI behaviors. …

Read more

Sakana AI introduces Continuous Thought Machines, a novel neural network that mimics brain processes

Sakana AI, co-founded by former Google AI scientists, has unveiled a new neural network architecture called Continuous Thought Machines (CTM). Unlike traditional transformer-based models that process information in parallel, CTMs incorporate a time-based dimension that mimics how biological brains operate, allowing for more flexible and adaptive reasoning. The key innovation in CTMs is their treatment …

Read more