Gabriel Martín Blázquez's picture

Gabriel Martín Blázquez

gabrielmbmb

AI & ML interests

ML Engineer

Recent Activity

liked a model about 3 hours ago
sesame/csm-1b
liked a model 1 day ago
MBZUAI/LLMVoX
upvoted a collection 4 days ago
EuroBERT
View all activity

Organizations

Spaces-explorers's profile picture SomosNLP's profile picture Blog-explorers's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture LLHF's profile picture SLLHF's profile picture Data Is Better Together Contributor's profile picture

Posts 4

view post
Post
1890
Yesterday   @mattshumer released mattshumer/Reflection-Llama-3.1-70B, an impressive model that achieved incredible results in benchmarks like MMLU. The model was fine-tuned using Reflection-Tuning and the dataset used wasn't released, but I created a small recipe with distilabel that allows generating a dataset with a similar output format:

1. We use MagPie 🐦 in combination with https://huggingface.co./meta-llama/Meta-Llama-3.1-70B-Instruct to generate reasoning instructions.
2. We generate a response again using https://huggingface.co./meta-llama/Meta-Llama-3.1-70B-Instruct, but we steer the LLM to generate an specific output format using a custom system prompt. In the system prompt, we instruct the LLM that it will have first to think 💭 and have reflections that will help resolving ambiguities. After that, we instruct the LLM to generate an output based on the previous thinking

In this dataset gabrielmbmb/distilabel-reflection-tuning you can found 5 rows that I generated with this recipe. You can also found the code of the pipeline in the file called reflection.py.

Articles 2

Article
203

Open R1: Update #2