Skip to content

Commit 8890ec1

Browse files
committed
conflict
2 parents d915404 + 9266281 commit 8890ec1

File tree

3 files changed

+39
-15
lines changed

3 files changed

+39
-15
lines changed

_blog.yml

+11-1
Original file line numberDiff line numberDiff line change
@@ -5921,6 +5921,17 @@
59215921
- hub
59225922
- inference
59235923

5924+
- local: llama-guard-4
5925+
title: "Welcoming Llama Guard 4 on Hugging Face Hub"
5926+
author: merve
5927+
thumbnail: /blog/assets/llama-guard-4/thumbnail.png
5928+
date: April 29, 2025
5929+
tags:
5930+
- llama
5931+
- llm
5932+
- vision
5933+
- vlm
5934+
59245935
- local: gradio-mcp
59255936
title: "How to Build an MCP Server with Gradio"
59265937
author: abidlabs
@@ -5930,4 +5941,3 @@
59305941
- gradio
59315942
- tool
59325943
- llm
5933-

lerobot-goes-to-driving-school.md

+12-2
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,16 @@ authors:
99

1010
---
1111

12+
# TL;DR
13+
14+
A snapshot of [L2D](https://huggingface.co/datasets/yaak-ai/L2D), the world's largest self-driving dataset!
15+
- 90+ TeraBytes of multimodal data (5000+ hours of driving) from 30 cities in Germany.
16+
- 6 surrounding HD cameras + vehicle state (Speed/Heading/GPS/IMU)
17+
- Continuous (Gas/Brake/Steering) and discrete actions (Gear/Turn Signals)
18+
- OpenStreetMap [matched waypoints](#OpenStreetMap) from birds-eye-view.
19+
- Natural language instructions. F.ex ["When the light turns green, drive over the tram tracks and then through the roundabout"](https://huggingface.co/spaces/lerobot/visualize_dataset?dataset=yaak-ai%2FL2D&episode=82))
20+
- Expert (driving instructors) and student (learner drivers) policies
21+
1222
# LeRobot goes to driving school
1323

1424
State-of-the art [Vision Language Models](https://huggingface.co/blog/vlms) and Large Language Models are trained on open-source
@@ -249,8 +259,8 @@ information about the episodes. Each release **R1+** is a superset of the previo
249259

250260
| HF | Nutron | Date | Episodes | Duration | Size | instructions | task\_id | observation.state.route | suboptimal |
251261
| :---: | :---: | :---: | :---: | :---: | :---: | :---: | :---: | :---: | :---: |
252-
| [R0](https://huggingface.co/datasets/yaak-ai/L2D) | [R0](https://nutron-sandbox.yaak.ai/collections/fcbb0dfd-40ae-4fd2-b023-7f300f35c5c7/300b7174-b6aa-4598-83e8-fc28cc5fcbe3/list) | March 2025 | 100 | 0.5+ hr | 9,5 GB | ☑️ | | | |
253-
| R1 | R1 | April 2025 | 1K | 5+ hr | 95 GB | ☑️ | | | |
262+
| [R0](https://huggingface.co/datasets/yaak-ai/L2D/tree/R0) | [R0](https://nutron-sandbox.yaak.ai/collections/fcbb0dfd-40ae-4fd2-b023-7f300f35c5c7/300b7174-b6aa-4598-83e8-fc28cc5fcbe3/search/list/session-logs?context=5s) | March 2025 | 100 | 0.5+ hr | 9,5 GB | ☑️ | | | |
263+
| [R1](https://huggingface.co/datasets/yaak-ai/L2D) | [R1](https://nutron-sandbox.yaak.ai/collections/fcbb0dfd-40ae-4fd2-b023-7f300f35c5c7/1cb18573-f731-47b1-ae89-7ea2f026b8d0/search/list/session-logs?context=5s) | April 2025 | 1K | 5+ hr | 95 GB | ☑️ | | | |
254264
| R2 | R2 | May 2025 | 10K | 50+ hr | 1 TB | ☑️ | ☑️ | ☑️ | ☑️ |
255265
| R3 | R3 | June 2025 | 100K | 500+ hr | 10 TB | ☑️ | ☑️ | ☑️ | ☑️ |
256266
| R4 | R4 | July 2025 | 1M | 5000+ hr | 90 TB | ☑️ | ☑️ | ☑️ | ☑️ |

open-deep-research.md

+16-12
Original file line numberDiff line numberDiff line change
@@ -28,14 +28,17 @@ The clock is ticking, let’s go! ⏱️
2828

2929
## Table of Contents
3030

31-
- [What are Agent frameworks and why they matter?](#what-are-agent-frameworks-and-why-they-matter)
32-
- [The GAIA benchmark](#the-gaia-benchmark)
33-
- [Building an open Deep Research](#building-an-open-deep-research)
34-
- [Using a CodeAgent](#using-a-codeagent)
35-
- [Making the right tools 🛠️](#making-the-right-tools-🛠️)
36-
- [Results 🏅](#results-🏅)
37-
- [Community reproductions](#community-reproductions)
38-
- [Most important next steps](#most-important-next-steps)
31+
- [Open-source DeepResearch – Freeing our search agents](#open-source-deepresearch--freeing-our-search-agents)
32+
- [TLDR](#tldr)
33+
- [Table of Contents](#table-of-contents)
34+
- [What are Agent frameworks and why they matter?](#what-are-agent-frameworks-and-why-they-matter)
35+
- [The GAIA benchmark](#the-gaia-benchmark)
36+
- [Building an open Deep Research](#building-an-open-deep-research)
37+
- [Using a CodeAgent](#using-a-codeagent)
38+
- [Making the right tools 🛠️](#making-the-right-tools-️)
39+
- [Results 🏅](#results-)
40+
- [Community Reproductions](#community-reproductions)
41+
- [Most important next steps](#most-important-next-steps)
3942

4043

4144
## What are Agent frameworks and why they matter?
@@ -112,17 +115,18 @@ From building `smolagents` we can also cite a notable additional advantage, whic
112115

113116
Now we need to provide the agent with the right set of tools.
114117

115-
**1.** A web browser. While a fully fledged web browser interaction like [Operator](https://openai.com/index/introducing-operator/) will be needed to reach full performance, we started with an extremely simple text-based web browser for now for our first proof-of-concept. You can find the code [here](https://github.com/huggingface/smolagents/blob/gaia-submission-r1/examples/open_deep_research/scripts/text_web_browser.py)
118+
**1.** A web browser. While a fully fledged web browser interaction like [Operator](https://openai.com/index/introducing-operator/) will be needed to reach full performance, we started with an extremely simple text-based web browser for now for our first proof-of-concept. You can find the code [here](https://github.com/huggingface/smolagents/tree/main/examples/open_deep_research/scripts/text_web_browser.py)
116119

117-
**2.** A simple text inspector, to be able to **read a bunch of text file format**, find it [here](https://github.com/huggingface/smolagents/blob/gaia-submission-r1/examples/open_deep_research/scripts/text_inspector_tool.py).
120+
121+
**2.** A simple text inspector, to be able to **read a bunch of text file format**, find it [here](https://github.com/huggingface/smolagents/tree/main/examples/open_deep_research/scripts/text_inspector_tool.py).
118122

119123
These tools were taken from the excellent [Magentic-One](https://www.microsoft.com/en-us/research/articles/magentic-one-a-generalist-multi-agent-system-for-solving-complex-tasks/) agent by Microsoft Research, kudos to them! We didn’t change them much, as our goal was to get as high a performance as we can with the lowest complexity possible.
120124

121125
Here is a short roadmap of improvements which we feel would really improve these tools’ performance (feel free to open a PR and contribute!):
122126

123127
- extending the number of file formats which can be read.
124128
- proposing a more fine-grained handling of files.
125-
- replacing the web browser with a vision-based one, which we’ve started doing [here](https://github.com/huggingface/smolagents/blob/gaia-submission-r1/src/smolagents/vision_web_browser.py).
129+
- replacing the web browser with a vision-based one, which we’ve started doing [here](https://github.com/huggingface/smolagents/tree/main/src/smolagents/vision_web_browser.py).
126130

127131
## Results 🏅
128132

@@ -169,6 +173,6 @@ So we’re tackling that next! In a more general problem: we’re going to build
169173

170174
We’re also [hiring a full time engineer](https://apply.workable.com/huggingface/j/AF1D4E3FEB/) to help us work on this and more, apply if you’re interested 🙂
171175

172-
- To get started with Open Deep Research, try the examples [here](https://github.com/huggingface/smolagents/tree/gaia-submission-r1/examples/open_deep_research).
176+
- To get started with Open Deep Research, try the examples [here](https://github.com/huggingface/smolagents/tree/main/examples/open_deep_research).
173177
- Check the [smolagents](https://github.com/huggingface/smolagents) repo.
174178
- Read more about smolagents [docs](https://huggingface.co/docs/smolagents/index), [introduction blog post](https://huggingface.co/blog/smolagents).

0 commit comments

Comments
 (0)