Skip to content
View samholt's full-sized avatar
đź’­
Building the future
đź’­
Building the future

Highlights

  • Pro

Block or report samholt

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
samholt/README.md

I am a Research Scientist at Google DeepMind working on LLM agents and their learning dynamics, with a particular focus on reinforcement learning (RL), and a broader goal of helping build universal AI assistants—systems that can understand and remember context, use tools, act on users’ behalf across devices, and, in the longer term, help accelerate scientific discovery.

Before joining Google DeepMind full-time, I completed a PhD in machine learning at the University of Cambridge, supervised by Prof. Mihaela van der Schaar FRS in the Machine Learning and Artificial Intelligence group. My research focused on reinforcement learning for and with large language models: memory-augmented agents, tool use, scalable long-context generation, and automating parts of the scientific pipeline. My first-author work has appeared at NeurIPS, ICLR, ICML, AISTATS, and RSS, including multiple spotlights and a long oral. Earlier, I also worked with Google DeepMind on hierarchical RL and the MuJoCo Playground project.

Some recurring themes in my work are:

  • LLM agents with memory, tools, and lookahead
  • Unbounded code and content generation (e.g., large codebases, long-form text)
  • Agents optimising algorithms and simulators for science
  • RL at inference time for adapting language model behaviour

Here on GitHub you’ll mostly find research code and prototypes for these directions, along with open-source tools and experiments developed in my PhD focused on scalable, self-improving agentic systems.

Pinned Loading

  1. L2MAC L2MAC Public

    🚀 The LLM Automatic Computer Framework: L2MAC

    Python 144 41

  2. NeuralLaplace NeuralLaplace Public

    Neural Laplace: Differentiable Laplace Reconstructions for modelling any time observation with O(1) complexity.

    Python 81 12

  3. PacktPublishing/Practical-Machine-Learning-with-TensorFlow-2.0-and-Scikit-Learn PacktPublishing/Practical-Machine-Learning-with-TensorFlow-2.0-and-Scikit-Learn Public

    Practical Machine Learning with TensorFlow 2.0 and Scikit-Learn [Video], published by Packt

    Jupyter Notebook 58 26