Transcrib3D: 3D Referring Expression Resolution through Large Language Models
arxiv(2024)
摘要
If robots are to work effectively alongside people, they must be able to
interpret natural language references to objects in their 3D environment.
Understanding 3D referring expressions is challenging – it requires the
ability to both parse the 3D structure of the scene and correctly ground
free-form language in the presence of distraction and clutter. We introduce
Transcrib3D, an approach that brings together 3D detection methods and the
emergent reasoning capabilities of large language models (LLMs). Transcrib3D
uses text as the unifying medium, which allows us to sidestep the need to learn
shared representations connecting multi-modal inputs, which would require
massive amounts of annotated 3D data. As a demonstration of its effectiveness,
Transcrib3D achieves state-of-the-art results on 3D reference resolution
benchmarks, with a great leap in performance from previous multi-modality
baselines. To improve upon zero-shot performance and facilitate local
deployment on edge computers and robots, we propose self-correction for
fine-tuning that trains smaller models, resulting in performance close to that
of large models. We show that our method enables a real robot to perform
pick-and-place tasks given queries that contain challenging referring
expressions. Project site is at https://ripl.github.io/Transcrib3D.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要