2023

10/05/2023

Gave a talk on NLP Applications and Large Language Models to the Capital Enterprise startup network.

13/04/2023

Honoured to have been nominated by my students for the UCL Inspiring Teaching Delivery award 🙏

13/03/2023

Gave an invited talk on Dynamic Advsersarial Data Collection for Large Language Models at the UCL AI Centre seminar on The Present and Future of Large Language Models in Theory and Practice.

13/03/2023

That’s a wrap! Another year of the MSIN0221 Natural Language Processing lectures comes to an end. Exciting to see the growing interest in NLP and its application!

2022

24/11/2022

Presented recent work on DADC and GAAs at the King’s College London Distributed Artificial Intelligence group. Thanks for the insightful discussions!

17/10/2022

Super excited to announce that I have joined Cohere and will be working on making large language models more useful and robust.

10/07/2022

I’m in Seattle for NAACL 2022! I’ll be presenting Models in the Loop: Aiding Crowdworkers with Generative Annotation Assistants on Wednesday, 13th July at 10:45 PST. And don’t forget to join us at the DADC workshop on Thursday, 14th July for same amazing keynote talks, a diverse panel, presentations from our Shared Task participants and best paper winners, posters, prizes & much more!

18/05/2022

Our work Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity has been selected as an outstanding paper at ACL 2022!

13/05/2022

Our work Models in the Loop: Aiding Crowdworkers with Generative Annotation Assistants has been accepted as an oral presentation at NAACL 2022!

09/05/2022

Excited to announce that I have joined DeepMind as a Research Scientist Intern.

06/04/2022

Gave an invited talk on Dynamic Adversarial Data Collection for Question Answering at the Oracle Labs ML Seminar Series. This was a particularly fun and interactive one, thanks for the invite!

26/03/2022

The call for participation for the Shared Task at the DADC Workshop co-located with NAACL ‘22 in Seattle is now live! We have three fantastic tracks for you to participate in. Sign up here!

25/03/2022

Presented our work on Dynamic Adversarial Data Collection for QA at the University of Oxford.

19/03/2022

Additional resources from our work on Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation at EMNLP 2021 are now available! We are releasing a collection of synthetically-generated adversarial QA pairs and related resources as well as the models used to generate the questions.

14/03/2022

Just gave the last lecture of the MSIN0221 Natural Language Processing module for this year. Fantastic cohort as always and it was great to be back to in-person teaching!

20/01/2022

AdversarialQA is currently the 3rd most downloaded QA dataset on Huggingface 🤗 Datasets right after the benchmark SQuADv1.1 and SQuADv2!

04/01/2022

Our proposal for the First Workshop on Dynamic Adversarial Data Collection has been accepted! See you at NAACL ‘22 in Seattle!

2021

09/11/2021

Presented our work on Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation at EMNLP 2021. The recording is available here.

24/09/2021

Dynabench is 1 year old! To celebrate, we’ve released Dynatask to help researchers host their own tasks.

10/09/2021

Presented a live demonstration of Dynamic Benchmarking at the UCL AI Centre 2nd Anniversary Showcase.

27/08/2021

Our work Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation has been accepted to the EMNLP 2021 Main Conference!

26/08/2021

Our work Contrasting Human-and Machine-Generated Word-Level Adversarial Examples for Text Classification has been accepted to the EMNLP 2021 Main Conference!

24/08/2021

ldbd.ly helps you make sense of ever-changing dynamic leaderboards.

18/04/2021

Our work Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation is now available on arXiv!

17/04/2021

Our work Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity is now available on arXiv!

07/04/2021

The Dynabench paper introducing our unified research platform for dynamic benchmarking has been accepted to NAACL 2021!

06/04/2021

Excited to announce that I have joined Facebook AI Research as an external research collaborator working on generation-assisted human adversarial annotation.

13/01/2021

The AdversarialQA dataset is now available in Huggingface 🤗 Datasets! Usage is as simple as from datasets import load_dataset; adversarial_qa = load_dataset('adversarial_qa', 'adversarialQA')

2020

12/12/2020

The HAMLETS NeurIPS 2020 workshop kicks off today. Join us to learn more about Human And Model in the Loop Evaluation and Training Strategies.

20/11/2020

Presented Humans-and-Machines in the Loop for Dynamic Benchmarking and Evaluation at the Annual MURI Review Meeting.

11/11/2020

Presented Adversarial Human Annotation for Dynamic Benchmarking and Evaluation at the UCL AI Centre session on AI in science, industry and society at TheAlgo2020.

25/10/2020

Our work Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension will be presented at EMNLP 2020.

24/09/2020

Dynabench, in collaboration with Stanford University, the University of North Carolina at Chapel Hill, and Facebook AI, is now live! Can you fool the QA model?

22/09/2020

Call for papers for our NeurIPS2020 workshop HAMLETS: Human And Model in the Loop Evaluation & Training Strategies is now live!

15/09/2020

Our work Undersensitivity in Neural Reading Comprehension has been accepted in Findings of EMNLP 2020!

01/09/2020

Excited to announce that I have joined Facebook AI Research as a research intern working on adversarial benchmarking and robustness.

17/06/2020

Our work Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension has been accepted to Transactions of the Association for Computational Linguistics (TACL)!

26/04/2020

Delivered the MSIN0221: Natural Language Processing module to this year’s UCL MSc Business Analytics cohort.

14/02/2020

Presented Adversarial Human Annotation for Reading Comprehension at the University of Cambridge NLIP Seminar Series.

06/02/2020

Accepted onto Cohort III of the Conception X programme!

2019

17/06/2019

Presented Asking Harder Questions at the UCL NLP Inaugural Event followed by a poster session on ShARC.

15/05/2019

Delivered a two-part workshop titled Overview of NLP to this year’s UCL MSc Business Analytics cohort.

15/04/2019

Co-presented Interpretation of Natural Language Rules in Conversational Machine Reading at the South England Natural Language Processing (SENLP) meetup.

12/04/2019

Led a workshop titled Introduction to Python and Machine Learning at the Peking University HSBC Business School (PHBS) in Oxford.

14/01/2019

I have started a PhD at UCL under the guidance of Pontus Stenetorp and Sebastian Riedel.

2018

03/11/2018

Presented Interpretation of Natural Language Rules in Conversational Machine Reading at EMNLP together with Patrick Lewis and other co-authors.

25/10/2018

The ShARC dataset from our EMNLP ‘18 paper is now live!

31/08/2018

Bloomsbury AI has joined Facebook to strengthen its efforts in natural language processing research.

22/08/2018

Cape (open source) is the new state-of-the-art for open-domain question answering on TriviaQA.

17/08/2018

Our large-scale question answering system, Cape, is now available open source!

10/08/2018

Interpretation of Natural Language Rules in Conversational Machine Reading has been accepted at EMNLP.

25/04/2018

We’ve been accepted into the Allen & Overy Fuse accelerator programme.

2017

22/11/2017

Invited presentation of the work we’re doing at Bloomsbury AI at the A Common Language for Intelligence meet-up hosted by Grakn AI.

12/05/2017

I have joined NLP-focused startup Bloomsbury AI, working on open-domain question answering.