Profile avatar
iscienceluvr.bsky.social
PhD at 19 | Founder and CEO at @MedARC_AI | Research Director at @StabilityAI | @kaggle Notebooks GM | Biomed. engineer @ 14 | TEDx talk➡https://bit.ly/3tpAuan
80 posts 2,580 followers 150 following
Prolific Poster
Conversation Starter

I restarted my blog a few weeks ago. The 1st post was: Debunking DeepSeek Delusions I discussed 5 main myths that I saw spreading online back during the DeepSeek hype. It may be a little less relevant now, but hopefully still interesting to folks. Check it out → www.tanishq.ai/blog/posts/d...

Are folks still here? 😅

Okay so this is so far the most important paper in AI of the year

Anthropic, please add a higher tier plan for unlimited messages 😭🙏

Decentralized Diffusion Models UC Berkeley and Luma AI introduce Decentralized Diffusion Models, a way to train diffusion models on decentralized compute with no communication between nodes. abs: arxiv.org/abs/2501.05450 project page: decentralizeddiffusion.github.io

The GAN is dead; long live the GAN! A Modern Baseline GAN This is a very interesting paper, exploring making GANs simpler and more performant. abs: arxiv.org/abs/2501.05441 code: github.com/brownvc/R3GAN

Happy birthday to my incredible and awesome Mamma! 🥳🎉🎂 To many more years of health and happiness. Tiara (my sister) and I love you very much ❤️❤️❤️

Happy 19th birthday to my amazing sister Tiara Abraham! 🥳🎉 🎂 Proud of you graduating with your Master's degree at 18 and starting your doctorate in music degree this past year! Excited to see what this final teen year holds for you!

Inventors of flow matching have released a comprehensive guide going over the math & code of flow matching! Also covers variants like non-Euclidean & discrete flow matching. A PyTorch library is also released with this guide! This looks like a very good read! 🔥 arxiv: arxiv.org/abs/2412.06264

Normalizing Flows are Capable Generative Models Apple introduces TarFlow, a new Transformer-based variant of Masked Autoregressive Flows. SOTA on likelihood estimation for images, quality and diversity comparable to diffusion models. arxiv.org/abs/2412.06329

Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models "We introduce a simple strategy that makes refusal behavior controllable at test-time without retraining: the refusal token." arxiv.org/abs/2412.06748

Can foundation models actively gather information in interactive environments to test hypotheses? "Our experiments with Gemini 1.5 reveal significant exploratory capabilities" arxiv.org/abs/2412.06438

Training Large Language Models to Reason in a Continuous Latent Space Introduces a new paradigm for LLM reasoning called Chain of Continuous Thought (COCONUT) Directly feed the last hidden state (a continuous thought) as the input embedding for the next token. arxiv.org/abs/2412.06769

[MASK] is All You Need New paper from CompVis group, introduces a new method called Discrete Interpolants that builds on top of discrete flow matching. Achieves SOTA performance on MS-COCO, competitive results on ImageNet 256. arxiv.org/abs/2412.06787

A new tutorial on RL by Kevin Patrick Murphy, a Research Scientist at Google DeepMind who also wrote several comprehensive, well-regarded textbooks on ML/DL. This ought to be a good read 👀 arxiv.org/abs/2412.05265

Birth and Death of a Rose abs: arxiv.org/abs/2412.05278 Generating temporal object intrinsics - temporally evolving sequences of object geometry, reflectance, and texture, such as blooming of a rose - from pre-trained 2D foundation models.

Frontier Models are Capable of In-context Scheming abs: arxiv.org/abs/2412.04984 "Our results show that o1, Claude 3.5 Sonnet, Claude 3 Opus, Gemini 1.5 Pro, and Llama 3.1 405B all demonstrate in-context scheming capabilities"

BigDocs: An Open and Permissively-Licensed Dataset for Training Multimodal Models on Document and Code Tasks abs: arxiv.org/abs/2412.04626 project page: bigdocs.github.io BigDocs-7.5M is a high-quality, open-access dataset comprising 7.5 million multimodal documents across 30 tasks.

Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling abs: arxiv.org/abs/2412.05271 model: huggingface.co/OpenGVLab/In... Introduces new InternVL-2.5 model, the first open-source MLLMs to surpass 70% on the MMMU benchmark

NVILA: Efficient Frontier Visual Language Models abs: arxiv.org/abs/2412.04468 NVIDIA introduces NVILA, a family of open VLMs designed to optimize both efficiency and accuracy.

Infinity: Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis abs: arxiv.org/abs/2412.04431 New visual autoregression framework that performs bitwise token prediction w/ an infinite-vocabulary tokenizer & classifier, a new record for autoregressive text-to-image models.

🤔 Why do we extract diffusion features from noisy images? Isn’t that destroying information? Yes, it is - but we found a way to do better. 🚀 Here’s how we unlock better features, no noise, no hassle. 📝 Project Page: compvis.github.io/cleandift 💻 Code: github.com/CompVis/clea... 🧵👇

Leading computer vision researchers Lucas Beyer (@giffmana.ai), Alexander Kolesnikov (@kolesnikov.ch), Xiaohua Zhai have left Google DeepMind to join OpenAI! They were behind recent SOTA vision approaches and open-source models like ViT, SigLIP, PaliGemma

The AI winter has started 😔

the restrictions on post and video length is gonna make it harder to paper-post here ngl

Reverse Thinking Makes LLMs Stronger Reasoners abs: arxiv.org/abs/2411.19865 Train an LLM to be able to generate forward reasoning from question, backward question, and backward reaoning from backward question Shows an average 13.53% improvement over the student model’s zero-shot performance

GaussianSpeech: Audio-Driven Gaussian Avatars abs: arxiv.org/abs/2411.18675 project page: shivangi-aneja.github.io/projects/gau...

some people managed to find some AoC-solving code from qianxyz in a github repo that has now been deleted seems like an automated pipeline using gpt-4o-mini with a pretty basic prompt

how does someone solve Advent of Code problem in 9 seconds??!!

how does someone solve Advent of Code problem in 9 seconds??!!

At #XPANSE in Abu Dhabi last week: - Met @anilseth.bsky.social backstage between our talks, discussed studying the nature of consciousness w/ neuroimaging. Appreciated him gifting me a signed copy of his book! - Met @seanmcarroll.bsky.social who gave a great talk about entropy vs. complexity

Many SOTA image generation models use an adversarial loss (VAE for latent diffusion for example), which counts I would say...

My Bluesky follower count (1.6k followers) has now surpassed my Threads follower count (1.1k). I still see a few AI folks on Threads but it seems so much more dead compared to BlueSky.

Every time conference reviews and rebuttals come in we hear complaints about how bad the process is. Which ML conference has the best review process and what's stopping other conferences from improving their processes?

Here is a list of ML OSS & Open Source / Science enthusiasts I found on Bluesky 🦋 go.bsky.app/8MFcfXd Let me know if you find such people here! I'm still new here and probably the list misses many must-add people, so let's built it together💪