Profile avatar
kanishka.bsky.social
Research assistant professor at TTIC. Works on language, concepts, and generalization. "minicons guy" acc. to NJK ; 🐧 he/him 🕸️👁️: https://kanishka.website
117 posts 2,412 followers 227 following
Prolific Poster
Active Commenter

exciting to see more language acquisition work shifting the lens to peer communication! @clairebergey.bsky.social is organizing a CogSci symposium highlighting related lines of work this summer!

Fun new study by Ellie Breitfeld!

Made a new assignment for a class on Computational Psycholinguistics: - I trained a Transformer language model on sentences sampled from a PCFG - The students' task: Given the Transformer, try to infer the PCFG (w/ a leaderboard for who got closest) Would recommend! 1/n

not at naacl (nanaacl)

Six months ago, I accepted a postdoc offer in a fabulous lab. Three days ago, the Trump administration terminated the NSF grant supporting the position. Today, my offer had to be rescinded, and my PhD defense is a week away. (1/3)

If you're at #NAACL2025 and want to hear about similarity effects for property inheritance in LMs, please stop by! I will be presenting this work on Wednesday at the 11-12:30 poster session on Interpretability & analysis for language models (Hall 3). aclanthology.org/2025.naacl-l...

If you are at #NAACL2025 @naaclmeeting.bsky.social catch @juand-r.bsky.social presenting our poster on the interplay between similarity and category membership in the property inferences of LMs @ Poster Session 1 on Wednesday! Or if you're at home like me, read our paper: arxiv.org/abs/2410.22590

Our final South by Semantics lecture at UT Austin is happening on Wednesday April 23!

Always wanted to test this, and because of Sriram's awesome effort we finally have some results! Check out how LMs respond to suspicious coincidences:

Had nothing but the best time postdoc-ing with Kyle -- consider applying to work with him!

One of the ways that LLMs can be inconsistent is the "generator-validator gap," where LLMs deem their own answers incorrect. 🎯 We demonstrate that ranking-based discriminator training can significantly reduce this gap, and improvements on one task often generalize to others! 🧵👇

Yay - so excited Nick is joining us at TTIC! 🥳 🎉

Super excited to submit a big sabbatical project this year: "Continuous developmental changes in word recognition support language learning across early childhood": osf.io/preprints/ps...

The Department of Linguistics at Yale University is seeking to hire a one-year lecturer with expertise in Computational Linguistics and/or Syntax for the 2025-2026 school year. See the job ad for more info: linguistlist.org/issues/36/11... Please send to any who might be interested!

Does anyone know how to get image representations from the Llama-3.2-11B-Vision models? Specifically a pooled representation of the input image

✨New paper ✨ Introducing 🌍MultiBLiMP 1.0: A Massively Multilingual Benchmark of Minimal Pairs for Subject-Verb Agreement, covering 101 languages! We present over 125,000 minimal pairs and evaluate 17 LLMs, finding that support is still lacking for many languages. 🧵⬇️

Cognitive scientists who work across levels of analysis should be called Marr's Rovers

@kmahowald.bsky.social with a beautiful high-tech illustration 🎨 while describing @qyao.bsky.social's latest paper at the HSP online seminar series! Paper: arxiv.org/abs/2503.20850

another day another minicons update (potentially a significant one for psycholinguists?) "Word" scoring is now a thing! You just have to supply your own splitting function! pip install -U minicons for merriment

What might LMs learn about datives if no one gives anything to anyone and no one gives anyone anything? Check out this amazing tour de force work by @qyao.bsky.social!

LMs learn argument-based preferences for dative constructions (preferring recipient first when it’s shorter), consistent with humans. Is this from memorizing preferences in training? New paper w/ @kanishka.bsky.social , @weissweiler.bsky.social , @kmahowald.bsky.social arxiv.org/abs/2503.20850

[1/10] Is scene understanding solved? Models today can label pixels and detect objects with high accuracy. But does that mean they truly understand scenes? Super excited to share our new paper and a new task in computer vision: Visual Jenga! 📄 arxiv.org/abs/2503.21770 🔗 visualjenga.github.io

What a win for BU!!

Life update: I'm starting as faculty at Boston University @bucds.bsky.social in 2026! BU has SCHEMES for LM interpretability & analysis, I couldn't be more pumped to join a burgeoning supergroup w/ @najoung.bsky.social @amuuueller.bsky.social. Looking for my first students, so apply and reach out!

Looking for emergency reviewers for two ARR papers on linguistic coordination in LLM agents. Reviews needed within 2 days 🙏

Require a bunch of emergency reviewers for a paper on personality induction in LLMs -- please DM me if you'd like to review it (turnaround maybe in 2-4 days)

AANN usage in the wild: “an unbelievable 34000 people”

Ah yes USA - where two months feel like two years!

I’ve long used FiveThirtyEight’s interactive “Hack Your Way To Scientific Glory” to illustrate the idea of p-hacking when I teach statistics. But ABC/Disney killed the site earlier this month :( So I made my own with #rstats and Observable and #QuartoPub ! stats.andrewheiss.com/hack-your-way/

I just realized reml is the reverse of lmer (context: R)