ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2109.13116
  4. Cited By
VQA-MHUG: A Gaze Dataset to Study Multimodal Neural Attention in Visual
  Question Answering

VQA-MHUG: A Gaze Dataset to Study Multimodal Neural Attention in Visual Question Answering

27 September 2021
Ekta Sood
Fabian Kögel
Florian Strohm
Prajit Dhar
Andreas Bulling
ArXivPDFHTML

Papers citing "VQA-MHUG: A Gaze Dataset to Study Multimodal Neural Attention in Visual Question Answering"

10 / 10 papers shown
Title
GazeLLM: Multimodal LLMs incorporating Human Visual Attention
GazeLLM: Multimodal LLMs incorporating Human Visual Attention
Jun Rekimoto
38
0
0
31 Mar 2025
VISTA: A Visual and Textual Attention Dataset for Interpreting
  Multimodal Models
VISTA: A Visual and Textual Attention Dataset for Interpreting Multimodal Models
Harshit
Tolga Tasdizen
CoGe
VLM
28
1
0
06 Oct 2024
A Gaze-grounded Visual Question Answering Dataset for Clarifying
  Ambiguous Japanese Questions
A Gaze-grounded Visual Question Answering Dataset for Clarifying Ambiguous Japanese Questions
Shun Inadumi
Seiya Kawano
Akishige Yuguchi
Yasutomo Kawanishi
Koichiro Yoshino
30
1
0
26 Mar 2024
Learning User Embeddings from Human Gaze for Personalised Saliency
  Prediction
Learning User Embeddings from Human Gaze for Personalised Saliency Prediction
Florian Strohm
Mihai Bâce
Andreas Bulling
26
1
0
20 Mar 2024
Trends, Applications, and Challenges in Human Attention Modelling
Trends, Applications, and Challenges in Human Attention Modelling
Giuseppe Cartella
Marcella Cornia
Vittorio Cuculo
Alessandro D’Amelio
Dario Zanca
Giuseppe Boccignone
Rita Cucchiara
38
6
0
28 Feb 2024
Describing Images $\textit{Fast and Slow}$: Quantifying and Predicting
  the Variation in Human Signals during Visuo-Linguistic Processes
Describing Images Fast and Slow\textit{Fast and Slow}Fast and Slow: Quantifying and Predicting the Variation in Human Signals during Visuo-Linguistic Processes
Ece Takmaz
Sandro Pezzelle
Raquel Fernández
22
1
0
02 Feb 2024
Enhancing Human-like Multi-Modal Reasoning: A New Challenging Dataset
  and Comprehensive Framework
Enhancing Human-like Multi-Modal Reasoning: A New Challenging Dataset and Comprehensive Framework
Jingxuan Wei
Cheng Tan
Zhangyang Gao
Linzhuang Sun
Siyuan Li
Bihui Yu
R. Guo
Stan Z. Li
LRM
31
8
0
24 Jul 2023
The Copenhagen Corpus of Eye Tracking Recordings from Natural Reading of
  Danish Texts
The Copenhagen Corpus of Eye Tracking Recordings from Natural Reading of Danish Texts
Nora Hollenstein
Maria Barrett
Marina Bjornsdóttir
20
14
0
28 Apr 2022
Multimodal Integration of Human-Like Attention in Visual Question
  Answering
Multimodal Integration of Human-Like Attention in Visual Question Answering
Ekta Sood
Fabian Kögel
Philippe Muller
Dominike Thomas
Mihai Bâce
Andreas Bulling
35
16
0
27 Sep 2021
Aggregated Residual Transformations for Deep Neural Networks
Aggregated Residual Transformations for Deep Neural Networks
Saining Xie
Ross B. Girshick
Piotr Dollár
Z. Tu
Kaiming He
297
10,216
0
16 Nov 2016
1