Elliot J. Crowley

I am a Lecturer (Assistant Professor) in Machine Learning and Computer Vision at the School of Engineering in the University of Edinburgh. I am part of the Bayesian and Neural Systems research group.

My research interests include:

  • simplifying machine learning
  • neural architecture search (and AutoML more generally)
  • efficient network training
  • low-resource deep learning
  • engineering applications of machine learning

I have an MEng in Engineering Science and a DPhil (PhD), both from the University of Oxford. My DPhil was on "Visual recognition in Art using Machine Learning" with Andrew Zisserman in the VGG group. After my DPhil, I was a postdoc at the School of Informatics in Edinburgh with Amos Storkey.

I hold an EPSRC New Investigator Award and I am an investigator on the dAIEdge Horizon Network.

I have also written an undergraduate machine learning course.


  • October 2023. [Job] Amos Storkey and I are recruiting a postdoc to work on ML/AI at the edge as part of the dAIEdge Horizon Network. Click here for more details. Please email amos+daiedgera *at* inf.ed.ac.uk you are interested! Deadline: 28th November

  • July 2023. [Job] I am recruiting a postdoc to work on NAS/AutoML. Click here for more details. Please email me if you are interested! Deadline: 31st July

  • June 2023. Would you like to do a PhD exploring how to make LLM training affordable? Click here. This is fully funded for UK students. Do email me if you are interested!

  • May 2023. I am co-organising the The First Edinburgh Workshop on Affordable Machine Learning to be held in June.
  • January 2023. I am co-organising the Fourth Workshop on Neural Architecture Search, Third lightweight NAS challenge at CVPR 2023.

Selected Publications

GPViT: A High Resolution Non-Hierarchical Vision Transformer with Group Propagation

ICLR 2023 (Accepted as a notable paper)

Chenhongyi Yang*, Jiarui Xu*, Shalini De Mello, Elliot J. Crowley, Xiaolong Wang

A new vision transformer architecture that serves as an excellent backbone across different fine-grained vision tasks.

Prediction-Guided Distillation for Dense Object Detection

ECCV 2022

Chenhongyi Yang, Mateusz Ochal, Amos Storkey, Elliot J. Crowley

A knowledge distillation framework for single stage detectors that uses a few key predictive regions to obtain high performance.

Neural Architecture Search without Training

ICML 2021 (Long talk)

Joseph Mellor, Jack Turner, Amos Storkey, Elliot J. Crowley

A low-cost measure for scoring networks at initialisation that can be used to perform neural architecture search in seconds.

Neural Architecture Search as Program Transformation Exploration

ASPLOS 2021 (Distinguished Paper)

Jack Turner, Elliot J. Crowley, Michael O'Boyle

A compiler-oriented approach to neural architecture search which can generate new convolution operations.

Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels

NeurIPS 2020 (Spotlight)

Massimiliano Patacchiola, Jack Turner, Elliot J. Crowley, Michael O'Boyle, Amos Storkey

A simple Bayesian alternative to standard meta-learning.

BlockSwap: Fisher-guided Block Substitution for Network Compression on a Budget

ICLR 2020

Jack Turner*, Elliot J. Crowley*, Michael O'Boyle, Amos Storkey, Gavia Gray

A fast algorithm for obtaining a compressed network architecture using Fisher information.

* equal contribution. A full list of publication is on Scholar.

Thanks to Jack Turner and Chenhongyi Yang for the website template.