Profile avatar
girving.bsky.social
Chief Scientist at the UK AI Safety Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
214 posts 3,102 followers 106 following
Regular Contributor
Active Commenter

FYI if you haven’t looked it up recently: Asteroid 2024 YR4 is down to only a 0.0027% chance of hitting Earth on 22 December 2032 (if it hits, it hits on that day), per Wikipedia and cneos.jpl.nasa.gov/sentry/detai....

This is a much wilder result than the recent tree evaluation proof that it builds on. I feel like it’s only 50-80% that it holds up though.

We're starting two new mitigation teams at AISI, Alignment and Control, which together with Safeguards will form a solutions unit working on direct research, collaboration, and external funding for frontier AI mitigations. Here is a thread on why you should join! 🧵

Especially excited about work like this because it means I can both eat my cake* and have it too**. *: Do a boring TCS PhD that isn't alignment relevant because it plays to my intellectual strengths and is self actualisation. **: Contribute usefully to technical AI safety.

Another neat alignment-motivated complexity theory conjecture from ARC! I am excited for more theory folk to work on alignment, and crisply defined conjectures are a great starting point. Some thoughts on how this conjecture relates to the overall problem. 🧵 www.alignment.org/blog/a-compu...

It’s fun to imagine an Infuriating Truth Fairy that occasionally pops up to give you perfect certainty whether a conjecture is true or false, but only where ~all of the value comes from the details of the (not given) proof.

ALLFED's paper, “Resilient foods for preventing global famine: a review of food supply interventions for global catastrophic food shocks including nuclear winter and infrastructure collapse”, is now published in Critical Reviews in Food Science and Nutrition tandfonline.com/doi/full/10....

This week I typed the author list of an upcoming paper into a code comment, and Cursor helpfully suggested adding 1. First, one of my previous coauthors from DeepMind days 2. Next, the brother of one of my AISI colleagues Both plausible coauthors!

Two new AISI papers: one on safety cases the other on safety frameworks! So far AISI has published a safety case template for inability arguments and a safety case sketch for AI control arguments; the new paper discusses why we are excited about them at a high-level. 🧵 x.com/MarieBassBuh...