Profile avatar
peterhenderson.bsky.social
Assistant Professor leading the AI, Law, & Society Lab @ Princeton 📚JD/PhD @ Stanford
168 posts 3,415 followers 547 following
Regular Contributor
Active Commenter

I wrote up my thoughts on the Library of Congress debacle. I even got to refer to one of the earliest cases in my career--defending the giant pandas at the National Zoo. Although it originally posted before the awesome! CADC IAF decision, I added a few mentions. www.lawfaremedia.org/article/bure...

🚨Reddit sues Anthropic!🚨 This is going to be a really interesting case. Some quick thoughts... 🧵👇 1️⃣ Notice: no copyright claim. Reddit doesn't really own the copyright to user-uploaded content, so nothing to do here. Reddit also doesn't make any federal claims to keep it in state court.

Hallucinated-Citation Watch (May 27 – Jun 2 2025) 📊🧵 7️⃣ Seven separate decisions flagged fake authorities for the week, bringing the total in our tracker to 172 cases. ⚖️ 4 U.S. federal district courts, 1 U.S. state-level tax court, 1 Canadian superior court, 1 Australian tribunal

Seems like despite the change in name and mission statement, lots of the core work that AISI was focusing on (cyber and CBRN) will continue on. TBD how things change though…

FWIW we’re now at 167 cases of nonexistent law/cases being cited across the world. It’s a mix of pro se litigants, attorneys, and even adjudicators. AI is here to stay. Even if attorneys stop using it, pro se litigants won’t.

Evaluations of offensive cyber abilities of frontier models often assume a fixed, 1-shot system. But this is just the tip of the iceberg, the risk "bubble" is bigger (and cheaper 💰) than you think. Evaluating frontier models for cyber risks? Check out our latest work!👇 arxiv.org/abs/2505.18384

Our tracker of “hallucinated” or nonexistent citations in real-world legal contexts has reached over 140 cases across the world. There’s a notable spike in the last 6 months. 📈📈📈 www.polarislab.org/ai-law-track...

The next ~1-4 years will be taking the 2017-2020 years of Deep RL and scaling up: exploration, generalization, long-horizon tasks, credit assignment, continual learning, multi-agent interaction! Lots of cool work to be done! 🎮🤖 But we shouldn't forget big lessons from back then.

Still searching, but was able to find two cases pre-2023 where non-existent citations were used! Modified the tracker to account for these.

Honestly it's happening so much that I can't keep up with adding all of them to our tracker. I have ~30 cases that people have sent my way that I'm still ingesting, so will be over 100 soon. www.polarislab.org/ai-law-track...

Tempted to use AI to help interpret statutes or draft opinions? 📜🤖 Take pause. As we explained in @lawfaremedia.org, closed models can smuggle in the hidden value judgments of everyone who touched the deployment/creation pipeline.

Another hallucinated citation in court. At this point, our tracker is up to ~70 cases worldwide of hallucinated citations in court, including hallucinations from 2 adjudicators. New Case: storage.courtlistener.com/recap/gov.us... Tracker: www.polarislab.org/ai-law-track...

SAG-AFTRA files an unfair labor practices complaint with the NLRB for Epic Games "utilizing AI-generated voices to replace bargaining unit work" on Fortnite. www.sagaftra.org/sites/defaul...

For what it’s worth, there’s a recent paper that benchmarks Bluebook citation formatting on frontier models and finds compliant/accurate citations only ~70% of the time. arxiv.org/abs/2505.02763

Update: so it was actually an attorney who relied on Claude for bluebook citation formatting that introduced the hallucinated the citation! www.reuters.com/legal/legali...

Up to 26 tracked cases so far across the United States, Canada, the UK, and Israel! Thanks for the pointers from folks.

There are so many hallucinated citations in court nowadays, that I'm starting to put together a tracker. Check it out and feel free to send ones that I've missed along. New tabs coming for more categories of AI+Law cases! www.polarislab.org/ai-law-track...

So apparently someone from Anthropic submitted a report to the court... with a hallucinated citation by Claude. chatgptiseatingtheworld.com/2025/05/13/a...

ADeLe, a new evaluation method, explains what AI systems are good at—and where they’re likely to fail. By breaking tasks into ability-based requirements, it has the potential to provide a clearer way to evaluate and predict AI model performance: msft.it/6014SkVGC

Thrilled to share this accessible MSR blogpost that summarizes our latest work on building a Science of AI Evaluation, where we manage to both reliably explain and predict success/failure of general-purpose AI models on new, unforeseen tasks and environments!

Insightful thread by @ajosephoconnell.bsky.social on removal of head of Copyright Office and Librarian of Congress!

House Energy and Commerce reconciliation text has language preempting all state AI regulations. To be seen whether this big de-regulatory push will remain in the final version. docs.house.gov/meetings/IF/...

Was very honored to receive a teaching award for my AI Law & Policy course! Christian Chung was also an amazing preceptor and received an award for his work too! Check out the syllabus below. Syllabus: princeton-polaris-lab.github.io/ailaw/ Announcement: www.cs.princeton.edu/news/compute...

Very excited that our work, "Safety Alignment Should be Made More Than Just a Few Tokens Deep" was recognized for an Outstanding Paper Award at #ICLR2025! It was great working with such wonderful co-authors and Xiangyu is also the first alumni from my group! You can check out more from us below!

From @peterhenderson.bsky.social's POLARIS Lab: 🚨Public Defenders - We need your insights! We’re conducting a research study to explore how #AI can better serve you & your clients. Your experiences can shape the future of AI legal tools designed specifically for public defenders. Sign-up 🔗🔻