Profile avatar
dvilasuero.hf.co
Everything datasets and human feedback for AI at Hugging Face. Prev: co-founder and CEO of Argilla (acquired by Hugging Face)
56 posts 3,388 followers 573 following
Prolific Poster
Conversation Starter

πŸš€ The open source community is unstoppable: 4M total downloads for DeepSeek models on @hf.co , with 3.2M coming from the +600 models created by the community. That's 30% more than yesterday!

πŸ’« Generate RAG data with the Synthetic Data Generator to improve your RAG system! 1️⃣ Generate from your documents, dataset, or dataset description. 2️⃣ Configure it. 3️⃣ Generate the synthetic dataset. 4️⃣ Fine-tune the retrieval and reranking models. 5️⃣ Build a RAG pipeline.

New chapter in the Hugging Face NLP course! πŸ€— πŸš€ We've added a new chapter about the very basics of Argilla to the Hugging Face NLP course. Learn how to set up an Argilla instance, load & annotate datasets, and export them to the Hub.Β  Any feedback for improvements welcome!

πŸŽ‰ 50,000+ annotations reached! The FineWeb2-C community is helping build better language models on annotation at a time. πŸ“Š Current stats: - 115 languages represented - 419 amazing contributors - 24 languages with complete datasets But we're not done yet! 🧡

High-quality data for fine-tuning language models for free and at the click of a button! Prompt and wait for your dataset to push to Argilla or the Hub Evaluate, review and fine-tune a model. Blog:

Was 2024 the year of datasets? Is 2025 the year for community-built datasets? It's exciting to see the progress of many languages in FineWeb-C: - Total annotations submitted: 41,577 - Languages with annotations: 106 - Total contributors: 363

The finish line is near! We're building FineWeb-Edu for many languages and need your help πŸ€— Many FineWeb-C languages are close to 1,000 annotations! Assamese is 99.4% done, French needs 64 more annotations, Tamil: 216. Please help us reach the goal: huggingface.co/spaces/data-...

πŸ’₯ Ending 2024: A full data annotation journey on the Hugging Face Hubβ€”from raw data to training-ready datasets! With Argilla 2.6.0, push your data to the Hub from the UI Let’s make 2025 the year anyone can build more transparent and accountable AIβ€”no coding or model skills needed.

πŸš€ Argilla v2.6.0 is here! πŸŽ‰ Let me show you how EASY it is to export your annotated datasets from Argilla to the Hugging Face Hub. 🀩 Take a look to this quick demo πŸ‘‡ πŸ’β€β™‚οΈ More info about the release at github.com/argilla-io/a... #AI #MachineLearning #OpenSource #DataScience #HuggingFace #Argilla

πŸ”₯ We got great feedback on this: "Synthetic Data Generator" A no-code tool to create datasets with LLMs, making it a breeze, allowing ANYONE to create datasets and models in minutes and without any code. Blog: https://buff.ly/4gybyoT GitHub: https://buff.ly/49IDSmd Space: https://buff.ly/3Y1S99z

Well, around 10 percent of the initial goal is complete, and so far, it's been quite a one-man army effort. We're still in the hunt for more people to join and contribute to this open-source initiative. @hf.co data-is-better-together-fineweb-c.hf.space/share-your-p...

The sprint for crowd sourced annotations with argilla is in full swing over at data-is-better-together-fineweb-c.hf.space I've just contributed 100 examples to this dataset: data-is-better-together-fineweb-c.hf.space/share-your-p... Big thanks to @dvilasuero.hf.co, @nataliaelv.hf.co and team πŸ™Œ

I've been building a small library for working with prompt templates on the @huggingface.bsky.social Hub: `pip install prompt-templates`. Motivation: The community currently shares prompt templates in a wide variety of formats: in datasets, in model cards, as strings in .py files, as .txt/... 🧡

Desperate to contribute to the development of Scots language AI. I've just contributed 16 examples to this dataset: data-is-better-together-fineweb-c.hf.space/share-your-p...

I've just contributed 156 examples to the FineWeb 2 Spanish dataset: data-is-better-together-fineweb-c.hf.space/share-your-p... If you want to contribute, sign in with @hf.co and find your language

Help shape the future of multilingual Open Source AI! Join the FineWeb 2 Community Annotation Sprint to create an open training dataset with full transparency and human validation in many languages. Review datasets in your language and help identify the best sources for training.

✨ Argilla 2.5.0 is live and it comes with webhook listener support to supercharge your workflows! πŸš€ #AI #MachineLearning #Webhooks #TechUpdate

πŸ‘ Open Image Preferences is an Apache 2.0 licensed dataset for text-to-image generation by the @hf.co community. This dataset contains 10K text-to-image preference pairs across image generation categories, using different model families and prompt complexities. Blog: huggingface.co/blog/image-p...

Open Image Preferences released! πŸš€ - Open-source dataset for text2image - 10K samples manually evaluated by the HF community. - Binarized format for SFT, DPO, or ORPO. It comes with a nice blog post explaining the steps to pre-process and generate the data, along with the results.

Announcing Global-MMLU - an improved MMLU Open dataset with evaluation coverage across 42 languages. The result of months of work with the goal of advancing Multilingual LLM evaluation. Built together with the community and amazing collaborators at Cohere4AI, MILA, MIT, and many more.

We're about to launch the biggest collaboration effort since the Open Assistant. Let's get the highest quality data for open foundation models with all the nuances & diversity of each language, all with data provenance and transparency Join us as language lead: docs.google.com/forms/d/10XI...

Next week we're launching a collaborative annotation effort to build a big multilingual dataset, so you can have high-quality data in your language. We are really close to getting leads for 100 languages! Can you help us cover the remaining 200?

For anyone interested in fine-tuning or aligning LLMs, I’m running this free and open course called smol course. It’s not a big deal, it’s just smol. 🧡>>

πŸ™Œ I just wanted to share a few thoughts about the latest Argilla release, 2.5.0, as it's a pretty big one! Argilla now has full support for webhooks, which means you can do some pretty cool stuff, like model training on the fly as annotations are created. 🀯 #MachineLearning #NLP #DataLabeling

[SATURDAY THREAD] β˜•οΈ πŸ§‘β€πŸŽ“ In case you spent the week reading GDPR legislation and missed everything. It’s all about vision language models and image preference datasets. >> 🧡 Here are the models and datasets you can use in your projects.

Recently, I added a feature to #Argilla to optimize plugin loading πŸŽ‰. It removes unnecessary code, improves readability, and lets future plugins load automatically. πŸš€ Check out the PR πŸ‘‡ and make your first contribution to our repo. github.com/argilla-io/a... #dev_experience #clean_code

πŸš€ We’re excited to announce Argilla v2.5.0, which includes: * Argilla webhooks, * A new design for the datasets home page. * Python 3.13 and Pydantic v2 support. πŸ“™ Read here πŸ‘‡ the full release notes github.com/argilla-io/a...

A dataset of 1 million or 2 million Bluesky posts is completely irrelevant to training large language models. The primary usecase for the datasets that people are losing their shit over isn't ChatGPT, it's social science research and developing systems that improve Bluesky.

The best path forward in AI requires technologists to be reflective/self-critical about how their work impacts society. Transparency helps this. Appreciate Bsky for flagging AI ethics &my colleague’s response. Let’s make informed consent a real thing. More later; Recommend: bsky.app/profile/cfie...

I've removed the Bluesky data from the repo. While I wanted to support tool development for the platform, I recognize this approach violated principles of transparency and consent in data collection. I apologize for this mistake.

The community has labelled over 3000 image preferences in a few hours. One open source image preferences dataset coming right up!

At @huggingface.bsky.social πŸ€— we're preparing a collaborative annotation effort to build an open-source multilingual dataset. If you'd like to get high-quality open data for your language, check if yours is listed in this form and sign up! forms.gle/DHJdtvoSNxAA...

🎨 Want better open-source AI art models? We need your help! Most top image generators are trained on human preferencesβ€”but those datasets are closed. Let's build our own! Rate images in pairs and help make AI art accessible to everyone πŸ”“ πŸ‘‰ huggingface.co/blog/burtens...

First dataset for the new @huggingface.bsky.social @bsky.app community organisation: one-million-bluesky-posts πŸ¦‹ πŸ“Š 1M public posts from Bluesky's firehose API πŸ” Includes text, metadata, and language predictions πŸ”¬ Perfect to experiment with using ML for Bluesky πŸ€— huggingface.co/datasets/blu...

Did you know that on Argilla, we’re adding a new feature to export labeled datasets directly to the Hugging Face Hub? πŸ€” We’re leveraging the Hugging Face datasets library for seamless integration, including defining span labeling Stay tuned for the release!🧠✨ #MachineLearning #NLP #DataLabeling

πŸ‘€ Who said the Argilla tool was only for text? I am proud of my brilliant teammates for setting up this significant initiative πŸ€— @benburtenshaw.bsky.social @davidberenstein.bsky.social @danielvanstrien.bsky.social @dvilasuero.hf.co