Alexander F. Spies
London, UK | [email protected] | linkedin.com/in/afspies | afspies.com | +44 7854 494 600
Summary
AI Safety Researcher (PhD, Imperial College London) specializing in mechanistic interpretability and causal world models, with hands-on experience shipping robust LLM systems and building eval/finetuning/tool-use pipelines. Published on maze-task world models & transformer internals; led research teams and developed safety-relevant tooling. Seeking to work on evaluations and interpretability techniques for frontier models.
Education
Imperial College London
Oct 2020 - Sep 2025Thesis: Interpretable Representations in Artificial Neural Networks
- Improving representations in Object-Centric Learning and reasoning
- Mechanistic analysis of vision & language transformers
- Advisors – Prof. Alessandra Russo & Prof. Michael Shanahan
Imperial College London
Sep 2019 - Sep 2020- Thesis: Learning World Models in the Animal-AI Environment
- Independent project: Neurosymbolic Learning & Neurally Weighted dSILP
University of California, Berkeley
Aug 2017 - May 2018- Completed graduate-level courses as an undergrad, alongside research
University of Manchester
Sep 2015 - Jun 2019- Thesis: AI for the Automated Diagnosis of Atrial Fibrillation
Professional Experience
Epic Games
Jan 2025 - PresentResearch engineer working on production LLM systems and agentic pipelines (promoted from intern to full-time in June 2025).
- Shipped production LLM services (10k+ weekly queries) for low-resource language support; hardened prompts, data paths, and fallback behaviors for robust performance across diverse use cases.
- Built finetuning/eval infra: UnSloth & SageMaker training, W&B sweeps, vLLM serving; standardized experiment tracking → faster ablations + reproducible benchmarks.
- Developing embedded agentic pipelines for multi-turn code generation, repo understanding, and retrieval + tool-use.
UnSearch (AI Safety Camp)
Mar 2023 - Oct 2024Led independent research groups on mechanistic interpretability as "model organisms of misalignment."
- Defined agenda on mechanistic interpretability for maze-solving LMs; trained transformers and Sparse Autoencoders, managed 9 researchers across 2 projects → 2 workshop papers + best-poster award.
- Built SAE-based analysis pipelines to study internal circuits & world-model structure; results demonstrated causal world models emerging in maze-solving transformers.
- Research artifacts, figures, and code published in workshops and available on GitHub/personal site.
National Institute of Informatics
Aug 2023 - Jun 2024Mechanistic analysis of Transformers trained on maze‑solving tasks.
Lawrence Berkeley National Laboratory
Feb 2018 - Jul 2018Investigated non‑local thresholds in pixel detectors; co‑authored JINST publication.
German Electron Synchrotron (DESY)
Jul 2018 - Sep 2018Exclusion analysis of Higgs decay channels in MSSM.
Publications
Detailed List →(*indicates equal contributions)
Selected Publications
[1]
Transformers Use Causal World Models in Maze‑Solving Tasks
World Models Workshop (ICLR 2025), Oct 2024
[2]
Structured World Representations in Maze‑Solving Transformers
Unifying Representations in Neural Models Workshop (NeurIPS 2023), Dec 2023
[3]
Sparse Relational Reasoning with Object‑Centric Representations
Dynamic Neural Networks Workshop (ICML 2022) — spotlight, Jul 2022
Skills
Alignment & Interpretability
LLM Training/Serving
MLOps & Engineering
Languages
Awards & Grants
Long‑Term Future Fund Grant — Safe AI Research
Jul 2024FAR Labs Residency
Jun 2024Best Poster — Technical AI Safety Conference
Apr 2024JSPS Postdoctoral Fellowship
May 2023Google Cloud Research Grant
Aug 20221st Place — AIHack 2022
Mar 2022Full PhD Scholarship (UKRI)
Sep 2020Leadership & Service
Pivotal Fellowship
Jan 2025 - Apr 2025- Provided technical guidance on AI Safety Research to 8+ Research Fellows
Imperial College London
Jan 2021 - Dec 2024Reviewer
2022 - Present- NeurIPS, ICLR, ICML, AAAI, UAI, Artificial Intelligence (Journal); MATS program applications
Imperial College London & Manchester
Sep 2021 - Feb 2025- Led technical coursework for Deep Learning, ML Math, Data Structures & Algorithms, and Python
- Engineered GPU-backed autograding pipeline for 120+ students using Otter Grader and Paperspace