Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1806.03831
Cited By
Interactive Visual Grounding of Referring Expressions for Human-Robot Interaction
11 June 2018
Mohit Shridhar
David Hsu
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Interactive Visual Grounding of Referring Expressions for Human-Robot Interaction"
9 / 9 papers shown
Title
Deformable Attentive Visual Enhancement for Referring Segmentation Using Vision-Language Model
Alaa Dalaq
Muzammil Behzad
VLM
41
0
0
25 May 2025
RAIDER: Tool-Equipped Large Language Model Agent for Robotic Action Issue Detection, Explanation and Recovery
Silvia Izquierdo-Badiola
Carlos Rizzo
Guillem Alenyà
LLMAG
LM&Ro
113
0
0
22 Mar 2025
HiFi-CS: Towards Open Vocabulary Visual Grounding For Robotic Grasping Using Vision-Language Models
V. Bhat
Prashanth Krishnamurthy
Ramesh Karri
Farshad Khorrami
75
4
0
16 Sep 2024
General-purpose Clothes Manipulation with Semantic Keypoints
Yuhong Deng
David Hsu
82
2
0
15 Aug 2024
A Joint Speaker-Listener-Reinforcer Model for Referring Expressions
Licheng Yu
Hao Tan
Joey Tianyi Zhou
Tamara L. Berg
ObjD
73
275
0
30 Dec 2016
CLEVR: A Diagnostic Dataset for Compositional Language and Elementary Visual Reasoning
Justin Johnson
B. Hariharan
Laurens van der Maaten
Li Fei-Fei
C. L. Zitnick
Ross B. Girshick
CoGe
271
2,346
0
20 Dec 2016
DenseCap: Fully Convolutional Localization Networks for Dense Captioning
Justin Johnson
A. Karpathy
Li Fei-Fei
VLM
106
1,165
0
24 Nov 2015
Natural Language Object Retrieval
Ronghang Hu
Huazhe Xu
Marcus Rohrbach
Jiashi Feng
Kate Saenko
Trevor Darrell
ObjD
65
552
0
13 Nov 2015
A Joint Model of Language and Perception for Grounded Attribute Learning
Cynthia Matuszek
Nicholas FitzGerald
Luke Zettlemoyer
Liefeng Bo
Dieter Fox
LM&Ro
46
316
0
27 Jun 2012
1