Profile avatar
cschroeder.bsky.social
PhD Candidate @ Leipzig University. Active Learning, Text Classification and LLMs. Check out my active learning library: small-text. #NLP #NLProc #ActiveLearning #LLM #ML #AI
54 posts 930 followers 2,670 following
Regular Contributor
Active Commenter

Oh no, what happened to Argilla? @hf.co Could you explain what's going on? It has barely been a year since you bought it. #nlproc #nlp #ml

Big fan of @ai2.bsky.social's semantic scholar feeds. Usually great for paper recommendations. Yesterday it recommended... a paper that blatantly plagiarized from a former student's thesis that I co-supervised. So, I guess the algorithm really knows my interests ๐Ÿ˜….

Our recent paper on the impact of register (genre) on LLM performance. Key points: news do poor in evaluation, while opinionated texts are among the best. We hope this work can be used to understand the impact of register on LLMs and improve training data mixes! arxiv.org/abs/2504.01542

Ever wonder how LLM developers choose their pretraining data? Itโ€™s not guessworkโ€” all AI labs create small-scale models as experiments, but the models and their data are rarely shared. DataDecide opens up the process: 1,050 models, 30k checkpoints, 25 datasets & 10 benchmarks ๐Ÿงต

A bit of a mess around the conflict of COLM with the ARR (and to lesser degree ICML) reviews release. We feel this is creating a lot of pressure and uncertainty. So, we are pushing our deadlines: Abstracts due March 22 AoE (+48hr) Full papers due March 28 AoE (+24hr) Plz RT ๐Ÿ™

Can a Large Language Model (LLM) with zero Pokรฉmon-specific training achieve expert-level performance in competitive Pokรฉmon battles? Introducing PokรฉChamp, our minimax LLM agent that reaches top 30%-10% human-level Elo on Pokรฉmon Showdown! New paper on arXiv and code on github!

(1/8) Excited to share some new work: TESS 2! TESS 2 is an instruction-tuned diffusion LM that can perform close to AR counterparts for general QA tasks, trained by adapting from an existing pretrained AR model. ๐Ÿ“œ Paper: arxiv.org/abs/2502.13917 ๐Ÿค– Demo: huggingface.co/spaces/hamis... More below โฌ‡๏ธ

After 6+ months in the making and over a year of GPU compute, we're excited to release the "Ultra-Scale Playbook": hf.co/spaces/nanot... A book to learn all about 5D parallelism, ZeRO, CUDA kernels, how/why overlap compute & coms with theory, motivation, interactive plots and 4000+ experiments!

More than 8500 submissions to ACL 2025 (ARR February 2025 cycle)! That is an increase of 3000 submissions compared to ACL 2024. It will be a fun reviewing period. ๐Ÿ˜…๐Ÿ’ฏ @aclmeeting.bsky.social #ACL2025 #ACL2025nlp #NLP

๐Ÿ”ฅ ๐…๐ข๐ง๐š๐ฅ ๐‚๐š๐ฅ๐ฅ ๐š๐ง๐ ๐ƒ๐ž๐š๐๐ฅ๐ข๐ง๐ž ๐„๐ฑ๐ญ๐ž๐ง๐ฌ๐ข๐จ๐ง: Survey on Data Annotation and Active Learning We need your support in web survey in which we investigate how recent advancements in NLP, particularly LLMs, have influenced the need for labeled data in supervised machine learning. #NLP #NLProc #ML #AI

Hallo and happy New Year #NLProc :) Julia Romberg, a postdoc in my group in Cologne, together with other collaborators, is conducting a survey on the use of Active Learning in NLP. Find the link in the thread below!

Hereโ€™s just one of the many exciting questions from our survey. If these topics resonate with you and you have experience working on supervised learning with text (i.e., supervised learning in Natural Language Processing), we warmly invite you to participate!

๐Ÿ’™ ๐——๐—ฎ๐˜๐—ฎ ๐—”๐—ป๐—ป๐—ผ๐˜๐—ฎ๐˜๐—ถ๐—ผ๐—ป ๐—•๐—ผ๐˜๐˜๐—น๐—ฒ๐—ป๐—ฒ๐—ฐ๐—ธ ๐—ฎ๐—ป๐—ฑ ๐—”๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐—ณ๐—ผ๐—ฟ ๐—ก๐—Ÿ๐—ฃ ๐—ถ๐—ป ๐˜๐—ต๐—ฒ ๐—˜๐—ฟ๐—ฎ ๐—ผ๐—ณ ๐—Ÿ๐—Ÿ๐— ๐˜€ ๐Ÿ’ก Have you ever had to overcome a lack of labeled data to deal with an NLP task? We are conducting a survey to explore the strategies used to overcome this bottleneck. #NLP #ML

Hello bluesky #NLProc world! Happy to announce the 12th Argument Mining workshop will be colocated with #ACL2025 in Vienna!

A librarian that previously worked at the British Library created a relatively small dataset of bsky posts, hundreds of times smaller than previous researchers, to help folks create toxicity filters and stuff. So people bullied him & posted death threats. He took it down. Nice one, folks.