Shoutout to the authors of the wonderful papers i.e. CtRNet-X, DUSt3R, Segment Anything, CLIP and Pytorch3D and for open-sourcing their codebase to advance science and make this effort happen!
Please check these works out if you havenβt already!
24.04.2025 00:33
π 1
π 0
π¬ 0
π 0
We have released our improved extrinsics. Try it out now at droid-dataset.github.io and read more details about it in the updated DROID paper at arxiv.org/abs/2403.12945
This was a fun collaboration with
@vitorguizilini, @SashaKhazatsky and @KarlPertsch!
23.04.2025 23:50
π 1
π 0
π¬ 1
π 0
Thereβs room to improve. Future work could explore:
β’ Extending to in-the-wild scenes via foundation models for robot segmentation & keypoints.
β’ Ensembling predictions over time for better temporal consistency.
β’ Fine-tuning pointmap models on real robot data to handle cluttered tabletops.
8/n
23.04.2025 23:50
π 1
π 0
π¬ 1
π 0
Large-scale auto calibration in robotics is challenging, and our pipeline has some limits:
β’ CtRNet-X is trained on Panda; generalization to other robots is untested.
β’ DUSt3R struggles with clutter or minimal view overlap.
β’ Steps 2οΈβ£ & 3οΈβ£ may yield false positives in tough lighting or geometry.
7/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
Similarly, we plot the distribution of number of matched points and cumulative curve after 3οΈβ£, helping to identify the top quantile of well-calibrated camera pairs within each lab.
6/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
Automatically calibrating a large-scale dataset is challenging. We provide quality assessment metrics across all three stages, with flexibility to narrow bounds for downstream tasks as needed.
1οΈβ£ and 2οΈβ£ quality metrics show IOU and Reprojection-error distributions post-calibration.
5/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
Below we show the Camera-to-Camera transformations, post-calibration improves the alignment of obtained pointclouds!
4/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
We provide:
π€ ~36k calibrated episodes with good-quality extrinsic calibration
π¦Ύ ~24k calibrated multi-view episodes with good-quality multi-view camera calibration
β
Quality assessment metrics for all provided camera poses
3/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
To achieve this, we utilize:
1οΈβ£ Auto Segment Anything (SAM) based filtering (Camera-to-Base Calibration)
2οΈβ£ Tuned CtRNet-X for bringing in additional cams (Camera-to-Base Calibration)
3οΈβ£ Pretrained DUST3R with depth-based pose optimization (Camera-to-Camera Calibration)
2/n
23.04.2025 23:50
π 0
π 0
π¬ 1
π 0
Introducing β¨Posed DROIDβ¨, results of our efforts at automatic post-hoc calibration of a large-scale robotics manipulation dataset.
Try it out at: droid-dataset.github.io
Learn more at:
π arXiv: arxiv.org/pdf/2403.12945
π Blog: medium.com/p/4ddfc45361d3
π§΅ 1/n
23.04.2025 23:50
π 5
π 1
π¬ 1
π 0
3D Vision Language Models (VLMs) for Robotic Manipulation: Opportunities and Challenges
π Learn more & submit your work: robo-3dvlms.github.io
Join us in shaping the future of robotics, 3D vision, and language models! π€π #CVPR2025
10.02.2025 17:00
π 3
π 0
π¬ 0
π 0
π€ Weβre honored to host top experts in the field:
β Angel Chang (Simon Fraser University)
β Chelsea Finn (Stanford University)
β Hao Su (UC San Diego)
β Katerina Fragkiadaki (CMU)
β Yunzhu Li (Columbia University)
β Ranjay Krishna (University of Washington)
5/N
10.02.2025 17:00
π 3
π 0
π¬ 1
π 0
π― Key Topics:
β
3D Vision-Language Policy Learning
β
Pretraining for 3D VLMs
β
3D Representations for Policy Learning
β
3D Benchmarks & Simulation Frameworks
β
3D Vision-Language Action Models
β
3D Instruction-Tuning & Pretraining Datasets for Robotics
4/N
10.02.2025 17:00
π 2
π 0
π¬ 1
π 0
π’ Call for Papers: Submission opens today!
π
Deadline: April 15, 2024 (11:59 PM PST)
π Format: Up to 4 pages (excluding references/appendices), CVPR template, anonymized submissions
π Accepted papers: Poster presentations, with selected papers receiving spotlight talks!
3/N
10.02.2025 17:00
π 0
π 0
π¬ 1
π 0
π Explore how 3D perception and language models can enhance robotic manipulation in the era of foundation models. Engage with leading experts and be part of this new frontier in 3D-based VLMs/VLAs for robotics.
2/N
10.02.2025 17:00
π 0
π 0
π¬ 1
π 0
πExciting News! Join us at the inaugural #CVPR2025 Workshop on 3D Vision Language Models (VLMs) for Robotics Manipulation on June 11, 2025, in Nashville, TN! π¦Ύ
robo-3dvlms.github.io
1/N
@cvprconference.bsky.social
10.02.2025 17:00
π 13
π 0
π¬ 1
π 1
Welcome onboard!
18.01.2025 02:08
π 3
π 0
π¬ 0
π 0
Done, welcome aboard!
17.01.2025 18:55
π 4
π 0
π¬ 0
π 0
Welcome on board!
20.12.2024 09:16
π 5
π 0
π¬ 0
π 0
Congrats on the release, demos look cool and it's open source π
04.12.2024 08:51
π 9
π 2
π¬ 1
π 0
In a world full of AI, authenticity will be the most valuable thing in the universe.
30.11.2024 02:39
π 75
π 6
π¬ 7
π 1
Hello π
28.11.2024 14:53
π 1
π 0
π¬ 1
π 0
1 Andrew Davison, Imperial College London - BMVA Symposium: Robotics Foundation & World Models
YouTube video by BMVA: British Machine Vision Association
For my first post on Bluesky, this recent talk I did at the recent BMVA one day meeting on World Models is a good summary of my work on Computer Vision, Robotics and SLAM, and my thoughts on a bigger picture of #SpatialAI.
youtu.be/NLnPG95vNhQ?...
28.11.2024 14:22
π 90
π 23
π¬ 5
π 2
Just included :) Welcome @ajdavison.bsky.social!
go.bsky.app/HcQYMj
28.11.2024 14:51
π 5
π 0
π¬ 0
π 0
Check out this BEAUTIFUL interactive blog about cameras and lenses
ciechanow.ski/cameras-and-...
27.11.2024 02:54
π 75
π 16
π¬ 6
π 1
Hello π Would love to join!
26.11.2024 02:39
π 1
π 0
π¬ 0
π 0
GitHub - SLAM-Handbook-contributors/slam-handbook-public-release: Release repo for our SLAM Handbook
Release repo for our SLAM Handbook. Contribute to SLAM-Handbook-contributors/slam-handbook-public-release development by creating an account on GitHub.
We are in the process of editing a SLAM handbook, to be published by Cambridge University Press, with many *stellar* contributors. Part 1 is available as an online draft for public comments. Help us find bugs/problems!
Link to release repo is here: lnkd.in/gZhTkaxb
16.11.2024 15:45
π 83
π 26
π¬ 5
π 2
Added you!
25.11.2024 10:12
π 0
π 0
π¬ 0
π 0
Welcome onbaord!
25.11.2024 09:59
π 1
π 0
π¬ 0
π 0