ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1702.01991
  4. Cited By
Representations of language in a model of visually grounded speech
  signal

Representations of language in a model of visually grounded speech signal

7 February 2017
Grzegorz Chrupała
Lieke Gelderloos
A. Alishahi
ArXivPDFHTML

Papers citing "Representations of language in a model of visually grounded speech signal"

34 / 84 papers shown
Title
Word Recognition, Competition, and Activation in a Model of Visually
  Grounded Speech
Word Recognition, Competition, and Activation in a Model of Visually Grounded Speech
William N. Havard
Jean-Pierre Chevrot
Laurent Besacier
17
21
0
18 Sep 2019
Language learning using Speech to Image retrieval
Language learning using Speech to Image retrieval
Danny Merkx
S. Frank
M. Ernestus
14
43
0
09 Sep 2019
Do Cross Modal Systems Leverage Semantic Relationships?
Do Cross Modal Systems Leverage Semantic Relationships?
Shah Nawaz
Muhammad Kamran Janjua
I. Gallo
Arif Mahmood
Alessandro Calefati
Faisal Shafait
19
8
0
03 Sep 2019
Higher-order Comparisons of Sentence Encoder Representations
Higher-order Comparisons of Sentence Encoder Representations
Mostafa Abdou
Artur Kulmizev
Felix Hill
D. Low
Anders Søgaard
20
16
0
01 Sep 2019
MaSS: A Large and Clean Multilingual Corpus of Sentence-aligned Spoken
  Utterances Extracted from the Bible
MaSS: A Large and Clean Multilingual Corpus of Sentence-aligned Spoken Utterances Extracted from the Bible
Marcely Zanon Boito
William N. Havard
Mahault Garnerin
Éric Le Ferrand
Laurent Besacier
22
47
0
30 Jul 2019
Trends in Integration of Vision and Language Research: A Survey of
  Tasks, Datasets, and Methods
Trends in Integration of Vision and Language Research: A Survey of Tasks, Datasets, and Methods
Aditya Mogadala
M. Kalimuthu
Dietrich Klakow
VLM
20
132
0
22 Jul 2019
Transfer Learning from Audio-Visual Grounding to Speech Recognition
Transfer Learning from Audio-Visual Grounding to Speech Recognition
Wei-Ning Hsu
David F. Harwath
James R. Glass
SSL
18
32
0
09 Jul 2019
Analyzing Phonetic and Graphemic Representations in End-to-End Automatic
  Speech Recognition
Analyzing Phonetic and Graphemic Representations in End-to-End Automatic Speech Recognition
Yonatan Belinkov
Ahmed M. Ali
James R. Glass
17
32
0
09 Jul 2019
On the Contributions of Visual and Textual Supervision in Low-Resource
  Semantic Speech Retrieval
On the Contributions of Visual and Textual Supervision in Low-Resource Semantic Speech Retrieval
Ankita Pasad
Bowen Shi
Herman Kamper
Karen Livescu
11
12
0
24 Apr 2019
Self-Supervised Audio-Visual Co-Segmentation
Self-Supervised Audio-Visual Co-Segmentation
Andrew Rouditchenko
Hang Zhao
Chuang Gan
Josh H. McDermott
Antonio Torralba
VLM
SSL
9
104
0
18 Apr 2019
Semantic query-by-example speech search using visual grounding
Semantic query-by-example speech search using visual grounding
Herman Kamper
Aristotelis Anastassiou
Karen Livescu
11
29
0
15 Apr 2019
Learning semantic sentence representations from visually grounded
  language without lexical knowledge
Learning semantic sentence representations from visually grounded language without lexical knowledge
Danny Merkx
S. Frank
SSL
11
13
0
27 Mar 2019
Towards Visually Grounded Sub-Word Speech Unit Discovery
Towards Visually Grounded Sub-Word Speech Unit Discovery
David F. Harwath
James R. Glass
16
35
0
21 Feb 2019
Models of Visually Grounded Speech Signal Pay Attention To Nouns: a
  Bilingual Experiment on English and Japanese
Models of Visually Grounded Speech Signal Pay Attention To Nouns: a Bilingual Experiment on English and Japanese
William N. Havard
Jean-Pierre Chevrot
Laurent Besacier
15
24
0
08 Feb 2019
Symbolic inductive bias for visually grounded learning of spoken
  language
Symbolic inductive bias for visually grounded learning of spoken language
Grzegorz Chrupała
14
28
0
21 Dec 2018
Analysis Methods in Neural Language Processing: A Survey
Analysis Methods in Neural Language Processing: A Survey
Yonatan Belinkov
James R. Glass
30
547
0
21 Dec 2018
Interpretable Textual Neuron Representations for NLP
Interpretable Textual Neuron Representations for NLP
Nina Poerner
Benjamin Roth
Hinrich Schütze
FAtt
AI4CE
MILM
8
26
0
19 Sep 2018
Explaining Character-Aware Neural Networks for Word-Level Prediction: Do
  They Discover Linguistic Rules?
Explaining Character-Aware Neural Networks for Word-Level Prediction: Do They Discover Linguistic Rules?
Fréderic Godin
Kris Demuynck
J. Dambre
W. D. Neve
T. Demeester
AI4CE
16
17
0
28 Aug 2018
Revisiting Cross Modal Retrieval
Revisiting Cross Modal Retrieval
Shah Nawaz
Muhammad Kamran Janjua
Alessandro Calefati
I. Gallo
6
6
0
19 Jul 2018
Revisiting the Hierarchical Multiscale LSTM
Revisiting the Hierarchical Multiscale LSTM
Ákos Kádár
Marc-Alexandre Côté
Grzegorz Chrupała
A. Alishahi
11
13
0
10 Jul 2018
Visually grounded cross-lingual keyword spotting in speech
Visually grounded cross-lingual keyword spotting in speech
Herman Kamper
Michael Roth
11
34
0
13 Jun 2018
Vision as an Interlingua: Learning Multilingual Semantic Embeddings of
  Untranscribed Speech
Vision as an Interlingua: Learning Multilingual Semantic Embeddings of Untranscribed Speech
David F. Harwath
Galen Chuang
James R. Glass
12
58
0
09 Apr 2018
Jointly Discovering Visual Objects and Spoken Words from Raw Sensory
  Input
Jointly Discovering Visual Objects and Spoken Words from Raw Sensory Input
David F. Harwath
Adrià Recasens
Dídac Surís
Galen Chuang
Antonio Torralba
James R. Glass
24
201
0
04 Apr 2018
On the difficulty of a distributional semantics of spoken language
On the difficulty of a distributional semantics of spoken language
Grzegorz Chrupała
Lieke Gelderloos
Ákos Kádár
A. Alishahi
17
6
0
23 Mar 2018
Linguistic unit discovery from multi-modal inputs in unwritten
  languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop
Linguistic unit discovery from multi-modal inputs in unwritten languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop
O. Scharenborg
Laurent Besacier
A. Black
M. Hasegawa-Johnson
Florian Metze
...
Elin Larsen
Danny Merkx
Rachid Riad
Liming Wang
Emmanuel Dupoux
20
33
0
14 Feb 2018
Object Referring in Visual Scene with Spoken Language
Object Referring in Visual Scene with Spoken Language
A. Vasudevan
Dengxin Dai
Luc Van Gool
29
18
0
10 Nov 2017
Semantic speech retrieval with a visually grounded model of
  untranscribed speech
Semantic speech retrieval with a visually grounded model of untranscribed speech
Herman Kamper
Gregory Shakhnarovich
Karen Livescu
21
53
0
05 Oct 2017
Analyzing Hidden Representations in End-to-End Automatic Speech
  Recognition Systems
Analyzing Hidden Representations in End-to-End Automatic Speech Recognition Systems
Yonatan Belinkov
James R. Glass
16
83
0
13 Sep 2017
Acoustic Feature Learning via Deep Variational Canonical Correlation
  Analysis
Acoustic Feature Learning via Deep Variational Canonical Correlation Analysis
Qingming Tang
Weiran Wang
Karen Livescu
DRL
13
20
0
11 Aug 2017
SPEECH-COCO: 600k Visually Grounded Spoken Captions Aligned to MSCOCO
  Data Set
SPEECH-COCO: 600k Visually Grounded Spoken Captions Aligned to MSCOCO Data Set
William N. Havard
Laurent Besacier
O. Rosec
15
28
0
26 Jul 2017
Encoding of phonology in a recurrent neural model of grounded speech
Encoding of phonology in a recurrent neural model of grounded speech
A. Alishahi
Marie Barking
Grzegorz Chrupała
10
58
0
12 Jun 2017
Imagination improves Multimodal Translation
Imagination improves Multimodal Translation
Desmond Elliott
Ákos Kádár
23
136
0
11 May 2017
Visually grounded learning of keyword prediction from untranscribed
  speech
Visually grounded learning of keyword prediction from untranscribed speech
Herman Kamper
Shane Settle
Gregory Shakhnarovich
Karen Livescu
11
63
0
23 Mar 2017
Pixel Recurrent Neural Networks
Pixel Recurrent Neural Networks
Aaron van den Oord
Nal Kalchbrenner
Koray Kavukcuoglu
SSeg
GAN
233
2,547
0
25 Jan 2016
Previous
12