Taylor W. Killian
Taylor Killian

Taylor W. Killian

Principal Scientist, Lila Sciences

Reinforcement Learning | Machine Learning | Decision Making Under Uncertainty

Download CV

Currently, I am a Principal Scientist at Lila Sciences within the AI Research organization. We are enthusiastically focused on developing Scientific Superintelligence, bringing advanced AI into automated laboratory settings to accelerate scientific discovery, there's a lot of exciting things on the horizon!

I work in the fields of reinforcement learning, machine learning, and causal inference. I have long been interested in decision making and the mechanisms by which humans summarize and reason about the world. In my work, I aim to develop models and algorithms that enable actors (whether human or not) to efficiently make decisions in the face of various forms of uncertainty. Ultimately, my goal is to develop algorithmic techniques that extend beyond the domain in which they are trained, adapting to their end uses and any unique aspects/preferences therein.

I'm always keen on hearing about interesting ideas and love collaborating with others on a variety of problems, applied and foundational, as far as there is alignment with my areas of focus. Don't hesitate to reach out!

News

  • 17 Jul 25 Presented the work, at ICML, I contributed to during my postdoc at Apple!
  • 24 Feb 25 My first day at MBZUAI! I'm excited for the journey ahead and all that we'll be able to accomplish together!
  • 5 Aug 24 I'm honored to be recognized as an Expert Reviewer for TMLR
  • 24 May 24 "Risk Sensitive Dead-end Identification in Safety-Critical Offline Reinforcement Learning" was accepted to be presented as part of the "Journal-to-Conference" track at RLC this coming August!
  • 16 Mar 24 I successfully defended my Dissertation today and am now officially Dr. Killian!

Research

Current research projects and interests.

Ongoing

Reinforcement Learning for Foundation Models

Overview Leading research efforts at MBZUAI’s Institute of Foundation Models to develop practical RL techniques for improving language model reasoning and alignment. Our work spans multiple reasoning domains through the...

Learn More
2023-2024

Safety-Critical Offline Reinforcement Learning

Overview Developing risk-sensitive methods for identifying dangerous states and treatments in healthcare settings. Focus on dead-end identification using distributional RL and conservative value estimation for improved patient safety. Motivation In...

Learn More
View All Research

Teaching

Courses, mentorship, and educational activities.

Planned

Introduction to Machine Learning

Course Overview A comprehensive introduction to machine learning for undergraduate students covering supervised learning, unsupervised learning, and practical ML skills. The course emphasizes both theoretical understanding and hands-on implementation of...

Learn More
Planned

Introduction to Reinforcement Learning

Course Overview An introductory course designed for advanced undergraduates and beginning graduate students covering the fundamentals of reinforcement learning. Students will learn core concepts, implement basic algorithms, and understand when...

Learn More
Planned

Advanced Topics in Reinforcement Learning

Course Overview Graduate-level course covering modern RL techniques including offline RL, safe RL, and applications to real-world problems. Emphasis on bridging theory and practice with hands-on projects. Course Description This...

Learn More
Ongoing

PhD Student Mentorship

Overview Actively mentoring PhD students at MBZUAI on projects spanning RL for LLMs, safe decision-making, and practical applications of machine learning. Focus on developing both technical skills and research independence....

Learn More
View All Teaching

Selected Publications

2026

Coming Soon

New publications and preprints will appear here.

2025

Revisiting Reinforcement Learning for LLM Reasoning from A Cross-Domain Perspective
Zhoujun Cheng, Shibo Hao, Tianyang Liu, Fan Zhou, Yutao Xie, Feng Yao, Yuexin Bian, Yonghao Zhuang, Nilabjo Dey, Yuheng Zha, Yi Gu, Kun Zhou, Yuqi Wang, Yuan Li, Richard Fan, Jianshu She, Chengqian Gao, Abulhair Saparov, Haonan Li, Taylor W. Killian, Mikhail Yurochkin, Zhengzhong Liu, Eric P. Xing, Zhiting Hu
arXiv Preprint
Reinforcement learning has emerged as a promising approach to improve large language model reasoning, yet most open efforts focus narrowly on math and code, limiting our understanding of its broader applicability to general reasoning. We introduce Guru, a curated RL reasoning corpus spanning six reasoning domains.
Robust Autonomy Emerges from Self-Play
Marco Cusumano-Towner, David Hafner, Alex Hertzberg, Brody Huval, Aleksei Petrenko, Eugene Vinitsky, Erik Wijmans, Taylor W. Killian, Stuart Bowers, Ozan Sener, Philipp Krahenbuhl, Vladlen Koltun
ICML 2025
We developed a robust autonomous driving agent, in simulation, via self-play at massive scale. This simulator was designed to run in extensively parallel settings where we could aggressively randomize each agent's physical and behavior characteristics and generate substantial amounts of experience.

2024

Clinically Motivated Sequential Decision Making Under Uncertainty in Offline Settings
Taylor W. Killian
PhD Thesis, University of Toronto, Department of Computer Science
In order to develop practical machine learning aided technology for the benefit of human users, it is critical to anchor scientific research and development by the intended real-world use cases. In this thesis, I propose specific modeling decisions that can be made to develop actionable insights from sequentially observed healthcare data.

2023

Risk Sensitive Dead-end Identification in Safety-Critical Offline Reinforcement Learning
Taylor W. Killian, Sonali Parbhoo, Marzyeh Ghassemi
Transactions on Machine Learning Research (TMLR)
We improve upon our prior dead-ends work by taking a risk-sensitive approach to dead-end discovery, leveraging distributional RL for value estimation. This allows for earlier indication of dead-ends in a manner that is tunable based on the risk tolerance of the designed task.
View All Publications arXiv

Musings and Reflections

I'm working on sharing insights from my research and experiences in reinforcement learning, machine learning, and more.

View All Posts