We just released RUKA, a $1300 humanoid hand that is 3D-printable, strong, precise, and fully open sourced!
The key technical breakthrough here is that we can control joints and fingertips of the robot **without joint encoders**. All we need here is self-supervised data collection and learning.
18.04.2025 18:53
π 29
π 7
π¬ 1
π 0
When life gives you lemons, you pick them up.
(trained with robotutilitymodels.com)
28.03.2025 04:02
π 15
π 4
π¬ 1
π 0
The robot behaviors shown below are trained without any teleop, sim2real, genai, or motion planning. Simply show the robot a few examples of doing the task yourself, and our new method, called Point Policy, spits out a robot-compatible policy!
28.02.2025 19:09
π 21
π 5
π¬ 1
π 1
We just released AnySense, an iPhone app for effortless data acquisition and streaming for robotics. We leverage Appleβs development frameworks to record and stream:
1. RGBD + Pose data
2. Audio from the mic or custom contact microphones
3. Seamless Bluetooth integration for external sensors
26.02.2025 15:14
π 34
π 10
π¬ 2
π 0
Stretch Community News - February 2025 β Hello Robot
Gazebo Harmonic, Stretch AI, Dynamic memory, and more!
What's new in the Stretch community this month?
βοΈ Gazebo Harmonic
βοΈ Dynamic semantic maps for open-vocabulary tasks
βοΈ Natural-language narration of robot experiences
βοΈ Implicit human-robot communication
And more! Follow the link below for more details:
hello-robot.com/community-up...
06.02.2025 00:15
π 4
π 1
π¬ 0
π 0
We all want a home robot that can actually help us out. Why can't I ask my robot "where did I leave my water bottle" and get a good answer?
In Graph-EQA, we build a 3d memory as the robot explores, using that memory to make decisions.
saumyasaxena.github.io/grapheqa/
30.12.2024 16:20
π 45
π 8
π¬ 3
π 1
A look at the future: chatting with my robot via Discord to ask it to find something in my house.
This uses an LLM to understand what the human wants and generate a task plan, then builds an open-vocab 3d scene representation to find and pick up objects
31.12.2024 16:38
π 19
π 4
π¬ 2
π 1
I'd like to introduce what I've been working at @hellorobot.bsky.social: Stretch AI, a set of open-source tools for language-guided autonomy, exploration, navigation, and learning from demonstration.
Check it out: github.com/hello-robot/...
Thread ->
03.12.2024 16:51
π 131
π 23
π¬ 5
π 4
New paper! We show that by using keypoint-based image representation, robot policies become robust to different object types and background changes.
We call this method Prescriptive Point Priors for robot Policies or P3-PO in short. Full project is here: point-priors.github.io
10.12.2024 20:32
π 37
π 7
π¬ 1
π 2
Modern policy architectures are unnecessarily complex. In our #NeurIPS2024 project called BAKU, we focus on what really matters for good policy learning.
BAKU is modular, language-conditioned, compatible with multiple sensor streams & action multi-modality, and importantly fully open-source!
09.12.2024 23:33
π 30
π 9
π¬ 1
π 2
Since we are nearing the end of the year, I'll revisit some of our work I'm most excited about from the last year and maybe a sneak peek of what we are up to next.
To start of, Robot Utility Models, which enables zero-shot deployment. In the video below, the robot hasnt seen these doors before.
08.12.2024 02:32
π 36
π 8
π¬ 2
π 3