Profile avatar
velezbeltran.bsky.social
Machine Learning PhD Student @ Blei Lab & Columbia University. Working on probabilistic ML | uncertainty quantification | LLM interpretability. Excited about everything ML, AI and engineering!
84 posts 1,689 followers 1,000 following
Prolific Poster
Conversation Starter

First 11 chapters of RLHF Book have v0 draft done. Should be useful now. Next: * Crafting more blog content into future topics, * DPO+ chapter, * Meeting with publishers to get wheels turning on physical copies, * Cleaning & cohesiveness rlhfbook.com

🔥 Benchmark Alert! MotifBench sets a new standard for evaluating protein design methods in motif scaffolding. Why does this matter? Reproducibility & fair comparison have been lacking—until now. Paper: arxiv.org/abs/2502.12479 | Repo: github.com/blt2114/Moti... A thread ⬇️

The HuggingFace/Nanotron team just shipped an entire pretraining textbook in interactive format. huggingface.co/spaces/nanot... It’s not just a great pedagogic support, but many unprecedented data and experiments presented for the first time in a systematic way.

I just wanted to see what it looked like 😭

Good God, please. I just want some gradients that don't vanish 😭

I was hoping that recent events would lead to a mass exodus from X. Many have left, but most of the ML and LLM people have not. I have lost a lot of respect for the ML community.

Now that bluesky has gifs (it didn't work?), I can share (again) my educational notebook on discrete flow matching (by Itai Gat et al.). Also please check the original article and official implementation by Meta! 🐍 github.com/gle-bellier/... 🐍 github.com/facebookrese... 📄 arxiv.org/abs/2407.15595

Really excited about this! We note a connection between diffusion/flow models and neural/latent SDEs. We show how to use this for simulation-free learning of fully flexible SDEs. We refer to this as SDE Matching and show speed improvements of several orders of magnitude. arxiv.org/abs/2502.02472

I have a sinking feeling that by 2029 I'm going to be faking a British accent so no one will think I was one of the *Americans* working on AI during the regime.

NGL, it's kind of surprising that more people haven't migrated here, especially given what Musk has been doing these days. I don't get it.

Since everyone wants to learn RL for language models now post DeepSeek, reminder that I've been working on this book quietly in the background for months. Policy gradient chapter is coming together. Plugging away at the book every day now. rlhfbook.com/c/11-policy-...

Please stop anthropomorphizing language models, it makes them feel really bad

This comments section is the first time I've felt even a shred of hope in eight days.

Nazi salutes and speaking at neo-Nazi rallies seems bad. There's history that we should learn from.

Something I really like about NLP research is that it makes everything super intuitive. This week I have been thinking about variational inference in NLP and a lot of the things that seemed to require mathematical intuition just become trivial when thinking about language. So cool:)

One thing about me is that I will fight Nazis until I’m six feet in the ground.

Every day things are going to become a bit stupider and more evil than the previous day, and there is no foreseeable end to that trend. Using the taxpayer funds to pump your crypto? Go ahead. The next day will bring something worse.

New randomized, controlled trial by the World Bank of students using GPT-4 as a tutor in Nigeria. Six weeks of after-school AI tutoring = 2 years of typical learning gains, outperforming 80% of other educational interventions. And it helped all students, especially girls who were initially behind.

Does anyone have any good resources to learn about quantization? Any essential papers to read and resources about how to use/quantize models in practice are greatly appreciated!

1-> 2 -> 3 -> 3.5 -> 4 -> 4o -> o1 -> o3 I guess we need AGI just to figure out how to name things

If you are into ML theory (RL or not) with a proven track record, and you are interested in an industry research position, PM me. Feel free to spread the word.

🧵 Excited to share #Echidna, a Bayesian framework for quantifying the impact of gene dosage on phenotypic plasticity: tinyurl.com/296kf7hf! With @elhamazizi.bsky.social and @mingxz.bsky.social, we integrate scRNA-seq & WGS to uncover how CNAs drive tumor evolution and transcriptional variability.

Proud of this work spearheaded by the phenomenal @jlfan.bsky.social and @mingxz.bsky.social in collaboration w/ Ben Izar! The past 3 years we've worked hard to unravel how #CNVs shape #tumor phenotypic plasticity seen in #singlecell #RNAseq data ➡️ #Echidna 🦔

Hello! We will be presenting Estimating the Hallucination Rate of Generative AI at NeurIPS. Come if you'd like to chat about epistemic uncertainty for In-Context Learning, or uncertainty more generally. :) Location: East Exhibit Hall A-C #2703 Time: Friday @ 4:30 Paper: arxiv.org/abs/2406.07457

I'm on my way to #NeurIPS2024. On Friday I'm going to present my latest paper with Yuval Benjamini. The gist is in the comments, and come chat with me to hear more!

The circuit hypothesis proposes that LLM capabilities emerge from small subnetworks within the model. But how can we actually test this? 🤔 joint work with @velezbeltran.bsky.social @maggiemakar.bsky.social @anndvision.bsky.social @bleilab.bsky.social Adria @far.ai Achille and Caro

I have been a little bit scarce on social media in the last few months. Some of that has just been from being busy at work, but some of it has had to do with my father's passing. He hated the very idea of social media, but he religiously followed my Twitter, and then Bluesky, feeds. /n

We are thrilled to share #Decipher 🔍! Extremely proud of Achille and Joy for leading the development of this creative #ML tool combining VAEs with deep exponential models for comparing disease & healthy trajectories ➕ applying it to study leukemic onset! biorxiv.org/conte…

We're hiring! We have a staff scientist position open for graduating PhDs and postdocs. If you're excited about single-cell omics in cancer (scRNA, scATAC, scWGS, CITE, spatial—we've got it all), consider applying! We have great people, data, and resources, and welcome visa sponsorships.

I wanted to make my first post about a project close to my heart. Linear algebra is an underappreciated foundation for machine learning. Our new framework CoLA (Compositional Linear Algebra) exploits algebraic structure arising from modelling assumptions for significant computational savings! 1/4

Thank you Nature and @anilananth.bsky.social for this great feature on LLMs and AGI (and for highlighting our work arxiv.org/abs/2406.03689)

From our lab finally I convinced Yuli Slavutsky (generalization/uncertainty/robustness) @yulislavutsky.bsky.social and Claudia Shi (science of LLMs and AI safety) @claudiashi.bsky.social to use bluesky. Go give them a follow! 😁