1. Who Evaluates AI’s Social Impacts? Mapping Coverage and Gaps in First- and Third-Party Evaluations
Key points:
As AI models become more embedded in daily life, there is an increasing reliance on evaluations to measure their societal effects, ranging from fairness and environmental impact to labour and privacy risks. This research provides a first comprehensive overview of how these evaluations are conducted, comparing reports from developers at the time of model release with independent evaluations carried out afterwards. The authors examined almost 200 documents from AI companies and a similar number from external organisations.
Their analysis reveals a consistent trend: companies tend to publish limited information about social harms, while external researchers cover a broader range of issues, yet still lack the necessary data for a comprehensive overview. Crucial indicators, such as the origin of training data, the human labour involved in moderation and the computational resources used, are rarely disclosed. The paper argues that this fragmented landscape weakens accountability and makes it harder for policymakers to understand real-world risks.
Authors: Anka Reuel, Avijit Ghosh, Jenny Chim, Andrew Tran, Yanan Long, Jennifer Mickel, Usman Gohar, Srishti Yadav et al.
2. Introducing Nested Learning: A new ML paradigm for continual learning
Key points:
Google Research has presented Nested Learning, a new framework designed to help machine-learning models update themselves over time without forgetting what they have previously learnt. Unlike conventional approaches, where the architecture is fixed first and then optimised afterwards, this method treats a model as a collection of multiple learning processes operating at different speeds and levels of detail.
The team introduced a prototype system called HOPE, which demonstrates that this structure can mitigate catastrophic forgetting and process long, information-rich sequences more efficiently than traditional designs. These results suggest a promising path towards AI systems that can adapt continuously and at multiple levels, more closely resembling the way humans learn over different timescales.
Authors: Ali Behrouz, Vahab Mirrokni
3. From Memorization to Reasoning in the Spectrum of Loss Curvature
Key points:
As LLMs increase in size and capability, a key question arises: how do they transition from memorising training data to reasoning about new inputs?
This study introduces the concept of loss curvature, a mathematical measure of how a model’s loss landscape behaves, as a means of exploring this transition. By analysing how loss curvature varies across tasks and layers, the authors demonstrate that memorisation and reasoning correspond to distinct curvature regimes. Their experiments reveal that deeper layers with sharper curvature are associated with memorisation, whereas flatter curvature is associated with more generalisable reasoning. These findings suggest that training dynamics and architectural choices significantly impact where and how reasoning emerges in LLMs, providing an insight for designing more robust and trustworthy AI systems.
Authors: Jack Merullo, Srihita Vatsavaya, Lucius Bushnaq, Owen Lewis
4. How Do AI Agents Do Human Work? Comparing AI and Human Workflows Across Diverse Occupations
Key points:
This paper provides a comprehensive analysis of the similarities and differences between the work processes of AI agents and humans across a wide range of occupations. The researchers collected detailed data from 48 individuals and four AI systems spanning 16 job categories. Their analysis shows that, while AI agents can perform highly efficiently in tasks with a clear structure, they still fall short in areas that demand flexibility, troubleshooting or situational judgement. The authors argue that integrating AI agents into workplaces requires more than merely assessing task performance; it also necessitates careful consideration of how these systems align with existing workflows, how responsibilities are distributed and where potential breakdowns may arise.
Authors: Zora Zhiruo Wang, Yijia Shao, Omar Shaikh, Daniel Fried, Graham Neubig, Diyi Yang
5. GTIG AI Threat Tracker: Advances in Threat Actor Usage of AI Tools
Key points:
Threat actors’ use of artificial intelligence is undergoing a notable transformation, according to the Google Cloud analysis. Instead of relying on AI simply to enhance performance, attackers are now integrating AI-driven malware directly into their activities. Google’s Threat Intelligence Group (GTIG) reports that families like PROMPTFLUX and PROMPTSTEAL draw on large language models in real time to craft malicious scripts and disguise their activity to avoid detection. The update also explains that adversaries are increasingly turning to tailored social-engineering strategies to circumvent the protective safeguards built into generative-AI systems such as Gemini.
Authors: Google Threat Intelligence Group