Profile avatar
conradlz.bsky.social
Leader | 30K+ Hours coding | Computer Scientist | Innovator @ Enterprise | Founder-Mode
31 posts 30 followers 215 following
Regular Contributor
Conversation Starter

9 of the top 10 github trending repos currently on Github are from deepseek-ai Wow

Feel really good to enjoy bluesky with all the other platforms changing with political climate.

CUDA and HIP both aren't good programming languages. Triton exists because the smartest kernal devs were wasting days manually optimizing code that ended up buggy anyways. My money is on one of the ai asic startup companies making a programming language that sticks in the near future.

Happy Holidays! Thought provoking idea: the most studied pure functions are ones which operate on numeric input and produce a numeric output. This is an incredibly small portion of all possible functions and operations that could be programmed.

Let's treat this one with enthusiastic participation. People will certainly be affected, but only those made aware will voice their opinions.

The ARC-Prize competition concluded with mindsai scoring 55.5% on ARC-AGI o3 would have won by a huge margin. arcprize.org

OpenAI skips o2, previews o3 scores, and they're truly crazy. Huge progress on the few benchmarks we think are truly hard today. Including ARC AGI. Rip to people who say any of "progress is done," "scale is done," or "llms cant reason" 2024 was awesome. I love my job.

Hyper Connections Can nueral networks learn the optimal strength of connections in order to improve performance? Excellent question by ByteDance and even better answer in the paper: arxiv.org/abs/2409.19606

What does this say about our economy and society?

HunyuanVideo from Tencent A new large video foundation model with open source weights, training, and model code github.com/Tencent/Huny...

Interesting idea for simultaneous training in multiple environments with a shared replay buffer. github.com/Albiemc1303/...

According to Ilya, the era of pre-training is beginning to sunset. What era do you want next?

Pre-training as we know it will end - Dr. Ilya Sutskever at NeurIPS 2024

Don't worry everyone, I have figured out a way to delay any potential AI takeover indefinitely. If you know you know. (Turn sound on)

Impressive and elegant approach to causal discovery in this competition-winning work by github user "thetourney". thetourney.github.io/adia-report/

I want to train AI to build Gundam, and have them compete in a tournament to see which is the ultimate fighter.

Shoutouts to Meta for open sourcing SPDL(Scalable and Performant Data Loading). I can easily imagine a future where software 2.0 (model-based software) has its own frameworks, languages, and tooling where these pieces will be foundational. github.com/facebookrese...

The open-source reasoning model QwQ (Qwen with Questions) is actually making the cost of reasoning cheap.

Announcing 🥂 FineWeb2: A sparkling update with 1000s of 🗣️languages. We applied the same data-driven approach that led to SOTA English performance in🍷 FineWeb to thousands of languages. 🥂 FineWeb2 has 8TB of compressed text data and outperforms other datasets.

Will research engineering become more delegated to agentic systems? In this paper we have a benchmark to hill-climb with agentic systems. arxiv.org/abs/2411.15114