Alexander F. Spies

London, UK | [email protected] | linkedin.com/in/afspies | afspies.com | +44 7854 494 600

Summary

AI Safety Researcher (PhD, Imperial College London) specializing in mechanistic interpretability and causal world models, with hands-on experience shipping robust LLM systems and building eval/finetuning/tool-use pipelines. Published on maze-task world models & transformer internals; led research teams and developed safety-relevant tooling. Seeking to work on evaluations and interpretability techniques for frontier models.

Education

Imperial College London

Oct 2020 - Sep 2025
PhD in Computer Science, AI London, UK

Thesis: Interpretable Representations in Artificial Neural Networks

  • Improving representations in Object-Centric Learning and reasoning
  • Mechanistic analysis of vision & language transformers
  • Advisors – Prof. Alessandra Russo & Prof. Michael Shanahan

Imperial College London

Sep 2019 - Sep 2020
MSc in Computing (AI & ML) London, UK
  • Thesis: Learning World Models in the Animal-AI Environment
  • Independent project: Neurosymbolic Learning & Neurally Weighted dSILP

University of California, Berkeley

Aug 2017 - May 2018
Study Abroad Year, Major: Physics Berkeley, CA, USA
  • Completed graduate-level courses as an undergrad, alongside research

University of Manchester

Sep 2015 - Jun 2019
MPhys in Theoretical Physics Manchester, UK
  • Thesis: AI for the Automated Diagnosis of Atrial Fibrillation

Professional Experience

Epic Games

Jan 2025 - Present
Research Engineer London, UK

Research engineer working on production LLM systems and agentic pipelines (promoted from intern to full-time in June 2025).

  • Shipped production LLM services (10k+ weekly queries) for low-resource language support; hardened prompts, data paths, and fallback behaviors for robust performance across diverse use cases.
  • Built finetuning/eval infra: UnSloth & SageMaker training, W&B sweeps, vLLM serving; standardized experiment tracking → faster ablations + reproducible benchmarks.
  • Developing embedded agentic pipelines for multi-turn code generation, repo understanding, and retrieval + tool-use.

UnSearch (AI Safety Camp)

Mar 2023 - Oct 2024
Research Team Lead Remote

Led independent research groups on mechanistic interpretability as "model organisms of misalignment."

  • Defined agenda on mechanistic interpretability for maze-solving LMs; trained transformers and Sparse Autoencoders, managed 9 researchers across 2 projects → 2 workshop papers + best-poster award.
  • Built SAE-based analysis pipelines to study internal circuits & world-model structure; results demonstrated causal world models emerging in maze-solving transformers.
  • Research artifacts, figures, and code published in workshops and available on GitHub/personal site.

National Institute of Informatics

Aug 2023 - Jun 2024
JSPS Doctoral Fellow Tokyo, Japan

Mechanistic analysis of Transformers trained on maze‑solving tasks.

Lawrence Berkeley National Laboratory

Feb 2018 - Jul 2018
Undergraduate Researcher Berkeley, CA, USA

Investigated non‑local thresholds in pixel detectors; co‑authored JINST publication.

German Electron Synchrotron (DESY)

Jul 2018 - Sep 2018
Research Intern Hamburg, Germany

Exclusion analysis of Higgs decay channels in MSSM.

Publications

Detailed List

(*indicates equal contributions)

Selected Publications

[1]

Transformers Use Causal World Models in Maze‑Solving Tasks

A.F. Spies, W. Edwards, M.I. Ivanitskiy, et al.

World Models Workshop (ICLR 2025), Oct 2024

[2]

Structured World Representations in Maze‑Solving Transformers

M.I. Ivanitskiy*, A.F. Spies*, T. Räuker*, et al.

Unifying Representations in Neural Models Workshop (NeurIPS 2023), Dec 2023

[3]

Sparse Relational Reasoning with Object‑Centric Representations

A.F. Spies, A. Russo, M. Shanahan

Dynamic Neural Networks Workshop (ICML 2022) — spotlight, Jul 2022

Skills

Alignment & Interpretability

Mechanistic interpretability, Sparse Autoencoders, Circuit Analysis, Evals, Causal World Models

LLM Training/Serving

PyTorch, vLLM, UnSloth, SageMaker, W&B, JAX, Agentic Pipelines

MLOps & Engineering

Experiment tracking, reproducible sweeps, dataset curation; Python (expert), C++/Java (working), Git

Languages

English (native), German (native), Japanese (beginner)

Awards & Grants

Long‑Term Future Fund Grant — Safe AI Research

Jul 2024

FAR Labs Residency

Jun 2024

Best Poster — Technical AI Safety Conference

Apr 2024

JSPS Postdoctoral Fellowship

May 2023

Google Cloud Research Grant

Aug 2022

1st Place — AIHack 2022

Mar 2022

Full PhD Scholarship (UKRI)

Sep 2020

Leadership & Service

Pivotal Fellowship

Jan 2025 - Apr 2025
Technical Research Advisor London, UK
  • Provided technical guidance on AI Safety Research to 8+ Research Fellows

Imperial College London

Jan 2021 - Dec 2024
Co‑founder — ICARL Seminar Series London, UK

Reviewer

2022 - Present
Journals, ML Conferences & MATS Program
  • NeurIPS, ICLR, ICML, AAAI, UAI, Artificial Intelligence (Journal); MATS program applications

Imperial College London & Manchester

Sep 2021 - Feb 2025
Teaching Assistant
  • Led technical coursework for Deep Learning, ML Math, Data Structures & Algorithms, and Python
  • Engineered GPU-backed autograding pipeline for 120+ students using Otter Grader and Paperspace