ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1606.01933
  4. Cited By
A Decomposable Attention Model for Natural Language Inference

A Decomposable Attention Model for Natural Language Inference

6 June 2016
Ankur P. Parikh
Oscar Täckström
Dipanjan Das
Jakob Uszkoreit
ArXivPDFHTML

Papers citing "A Decomposable Attention Model for Natural Language Inference"

50 / 186 papers shown
Title
Fastformer: Additive Attention Can Be All You Need
Fastformer: Additive Attention Can Be All You Need
Chuhan Wu
Fangzhao Wu
Tao Qi
Yongfeng Huang
Xing Xie
46
117
0
20 Aug 2021
Relation-aware Compositional Zero-shot Learning for Attribute-Object
  Pair Recognition
Relation-aware Compositional Zero-shot Learning for Attribute-Object Pair Recognition
Ziwei Xu
Guangzhi Wang
Yongkang Wong
Mohan S. Kankanhalli
49
26
0
10 Aug 2021
RAIN: Reinforced Hybrid Attention Inference Network for Motion
  Forecasting
RAIN: Reinforced Hybrid Attention Inference Network for Motion Forecasting
Jiachen Li
Fan Yang
Hengbo Ma
Srikanth Malla
M. Tomizuka
Chiho Choi
24
42
0
03 Aug 2021
Towards Robustness Against Natural Language Word Substitutions
Towards Robustness Against Natural Language Word Substitutions
Xinshuai Dong
A. Luu
Rongrong Ji
Hong Liu
SILM
AAML
32
113
0
28 Jul 2021
A Survey on Data Augmentation for Text Classification
A Survey on Data Augmentation for Text Classification
Markus Bayer
M. Kaufhold
Christian A. Reuter
36
334
0
07 Jul 2021
Productivity, Portability, Performance: Data-Centric Python
Productivity, Portability, Performance: Data-Centric Python
Yiheng Wang
Yao Zhang
Yanzhang Wang
Yan Wan
Jiao Wang
Zhongyuan Wu
Yuhao Yang
Bowen She
54
94
0
01 Jul 2021
DocNLI: A Large-scale Dataset for Document-level Natural Language
  Inference
DocNLI: A Large-scale Dataset for Document-level Natural Language Inference
Wenpeng Yin
Dragomir R. Radev
Caiming Xiong
HILM
26
97
0
17 Jun 2021
GraphiT: Encoding Graph Structure in Transformers
GraphiT: Encoding Graph Structure in Transformers
Grégoire Mialon
Dexiong Chen
Margot Selosse
Julien Mairal
20
164
0
10 Jun 2021
Towards More Equitable Question Answering Systems: How Much More Data Do
  You Need?
Towards More Equitable Question Answering Systems: How Much More Data Do You Need?
Arnab Debnath
Navid Rajabi
F. Alam
Antonios Anastasopoulos
22
11
0
28 May 2021
Recent Advances in Deep Learning Based Dialogue Systems: A Systematic
  Survey
Recent Advances in Deep Learning Based Dialogue Systems: A Systematic Survey
Jinjie Ni
Tom Young
Vlad Pandelea
Fuzhao Xue
Erik Cambria
54
268
0
10 May 2021
Poolingformer: Long Document Modeling with Pooling Attention
Poolingformer: Long Document Modeling with Pooling Attention
Hang Zhang
Yeyun Gong
Yelong Shen
Weisheng Li
Jiancheng Lv
Nan Duan
Weizhu Chen
35
98
0
10 May 2021
Rethinking Search: Making Domain Experts out of Dilettantes
Rethinking Search: Making Domain Experts out of Dilettantes
Donald Metzler
Yi Tay
Dara Bahri
Marc Najork
LRM
30
46
0
05 May 2021
Attention-based Stylisation for Exemplar Image Colourisation
Attention-based Stylisation for Exemplar Image Colourisation
Marc Górriz Blanch
Issa Khalifeh
Alan F. Smeaton
Noel E. O'Connor
M. Mrak
28
4
0
04 May 2021
RoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position Embedding
Jianlin Su
Yu Lu
Shengfeng Pan
Ahmed Murtadha
Bo Wen
Yunfeng Liu
38
2,176
0
20 Apr 2021
Making Attention Mechanisms More Robust and Interpretable with Virtual
  Adversarial Training
Making Attention Mechanisms More Robust and Interpretable with Virtual Adversarial Training
Shunsuke Kitada
Hitoshi Iyatomi
AAML
20
8
0
18 Apr 2021
Geometry-Free View Synthesis: Transformers and no 3D Priors
Geometry-Free View Synthesis: Transformers and no 3D Priors
Robin Rombach
Patrick Esser
Bjorn Ommer
ViT
16
93
0
15 Apr 2021
Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Ashish Vaswani
Prajit Ramachandran
A. Srinivas
Niki Parmar
Blake A. Hechtman
Jonathon Shlens
16
395
0
23 Mar 2021
SILT: Efficient transformer training for inter-lingual inference
SILT: Efficient transformer training for inter-lingual inference
Javier Huertas-Tato
Alejandro Martín
David Camacho
24
11
0
17 Mar 2021
Attention is Not All You Need: Pure Attention Loses Rank Doubly
  Exponentially with Depth
Attention is Not All You Need: Pure Attention Loses Rank Doubly Exponentially with Depth
Yihe Dong
Jean-Baptiste Cordonnier
Andreas Loukas
37
373
0
05 Mar 2021
Linear Transformers Are Secretly Fast Weight Programmers
Linear Transformers Are Secretly Fast Weight Programmers
Imanol Schlag
Kazuki Irie
Jürgen Schmidhuber
34
224
0
22 Feb 2021
UPDeT: Universal Multi-agent Reinforcement Learning via Policy
  Decoupling with Transformers
UPDeT: Universal Multi-agent Reinforcement Learning via Policy Decoupling with Transformers
Siyi Hu
Fengda Zhu
Xiaojun Chang
Xiaodan Liang
OffRL
31
71
0
20 Jan 2021
Taming Transformers for High-Resolution Image Synthesis
Taming Transformers for High-Resolution Image Synthesis
Patrick Esser
Robin Rombach
Bjorn Ommer
ViT
64
2,819
0
17 Dec 2020
Transformer Interpretability Beyond Attention Visualization
Transformer Interpretability Beyond Attention Visualization
Hila Chefer
Shir Gur
Lior Wolf
31
644
0
17 Dec 2020
End-to-End Human Pose and Mesh Reconstruction with Transformers
End-to-End Human Pose and Mesh Reconstruction with Transformers
Kevin Qinghong Lin
Lijuan Wang
Zicheng Liu
ViT
34
613
0
17 Dec 2020
R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
  Matching
R2^22-Net: Relation of Relation Learning Network for Sentence Semantic Matching
Anton van den Hengel
Le Wu
Guangyi Lv
Meng Wang
Enhong Chen
Shulan Ruan
30
20
0
16 Dec 2020
Rewriter-Evaluator Architecture for Neural Machine Translation
Rewriter-Evaluator Architecture for Neural Machine Translation
Yangming Li
Kaisheng Yao
8
2
0
10 Dec 2020
Infusing Finetuning with Semantic Dependencies
Infusing Finetuning with Semantic Dependencies
Zhaofeng Wu
Hao Peng
Noah A. Smith
22
36
0
10 Dec 2020
Learning to Model and Ignore Dataset Bias with Mixed Capacity Ensembles
Learning to Model and Ignore Dataset Bias with Mixed Capacity Ensembles
Christopher Clark
Mark Yatskar
Luke Zettlemoyer
18
61
0
07 Nov 2020
Improving Audio Anomalies Recognition Using Temporal Convolutional
  Attention Network
Improving Audio Anomalies Recognition Using Temporal Convolutional Attention Network
Qiang Huang
Thomas Hain
16
10
0
21 Oct 2020
Neural Networks for Entity Matching: A Survey
Neural Networks for Entity Matching: A Survey
Nils Barlaug
J. Gulla
35
91
0
21 Oct 2020
Explainable Automated Fact-Checking for Public Health Claims
Explainable Automated Fact-Checking for Public Health Claims
Neema Kotonya
Francesca Toni
218
248
0
19 Oct 2020
The elephant in the interpretability room: Why use attention as
  explanation when we have saliency methods?
The elephant in the interpretability room: Why use attention as explanation when we have saliency methods?
Jasmijn Bastings
Katja Filippova
XAI
LRM
30
172
0
12 Oct 2020
Explaining Deep Neural Networks
Explaining Deep Neural Networks
Oana-Maria Camburu
XAI
FAtt
28
26
0
04 Oct 2020
Overcoming Data Sparsity in Group Recommendation
Overcoming Data Sparsity in Group Recommendation
Hongzhi Yin
Qinyong Wang
Kai Zheng
Zhixu Li
Xiaofang Zhou
6
61
0
02 Oct 2020
XTE: Explainable Text Entailment
XTE: Explainable Text Entailment
V. S. Silva
André Freitas
Siegfried Handschuh
20
6
0
25 Sep 2020
FarsTail: A Persian Natural Language Inference Dataset
FarsTail: A Persian Natural Language Inference Dataset
Hossein Amirkhani
Mohammad AzariJafari
Zohreh Pourjafari
Soroush Faridan-Jahromi
Zeinab Kouhkan
Azadeh Amirak
19
30
0
18 Sep 2020
DeepRemaster: Temporal Source-Reference Attention Networks for
  Comprehensive Video Enhancement
DeepRemaster: Temporal Source-Reference Attention Networks for Comprehensive Video Enhancement
S. Iizuka
E. Simo-Serra
21
39
0
18 Sep 2020
Explainable CNN-attention Networks (C-Attention Network) for Automated
  Detection of Alzheimer's Disease
Explainable CNN-attention Networks (C-Attention Network) for Automated Detection of Alzheimer's Disease
Ning Wang
Mingxuan Chen
K. P. Subbalakshmi
20
22
0
25 Jun 2020
Attention-based Quantum Tomography
Attention-based Quantum Tomography
Peter Cha
P. Ginsparg
Felix Wu
Juan Carrasquilla
Peter L. McMahon
Eun-Ah Kim
26
72
0
22 Jun 2020
The Depth-to-Width Interplay in Self-Attention
The Depth-to-Width Interplay in Self-Attention
Yoav Levine
Noam Wies
Or Sharir
Hofit Bata
Amnon Shashua
30
45
0
22 Jun 2020
Defense against Adversarial Attacks in NLP via Dirichlet Neighborhood
  Ensemble
Defense against Adversarial Attacks in NLP via Dirichlet Neighborhood Ensemble
Yi Zhou
Xiaoqing Zheng
Cho-Jui Hsieh
Kai-Wei Chang
Xuanjing Huang
SILM
39
48
0
20 Jun 2020
Guided Transformer: Leveraging Multiple External Sources for
  Representation Learning in Conversational Search
Guided Transformer: Leveraging Multiple External Sources for Representation Learning in Conversational Search
Helia Hashemi
Hamed Zamani
W. Bruce Croft
11
62
0
13 Jun 2020
Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason
  Over Implicit Knowledge
Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason Over Implicit Knowledge
Alon Talmor
Oyvind Tafjord
Peter Clark
Yoav Goldberg
Jonathan Berant
ReLM
LRM
28
39
0
11 Jun 2020
Rationalizing Text Matching: Learning Sparse Alignments via Optimal
  Transport
Rationalizing Text Matching: Learning Sparse Alignments via Optimal Transport
Kyle Swanson
L. Yu
Tao Lei
OT
29
37
0
27 May 2020
Logical Inferences with Comparatives and Generalized Quantifiers
Logical Inferences with Comparatives and Generalized Quantifiers
Izumi Haruta
K. Mineshima
D. Bekki
ELM
19
11
0
16 May 2020
Elastic weight consolidation for better bias inoculation
Elastic weight consolidation for better bias inoculation
James Thorne
Andreas Vlachos
17
11
0
29 Apr 2020
A Spatio-temporal Transformer for 3D Human Motion Prediction
A Spatio-temporal Transformer for 3D Human Motion Prediction
Emre Aksan
Manuel Kaufmann
Peng Cao
Otmar Hilliges
ViT
23
221
0
18 Apr 2020
DialBERT: A Hierarchical Pre-Trained Model for Conversation
  Disentanglement
DialBERT: A Hierarchical Pre-Trained Model for Conversation Disentanglement
Tianda Li
Jia-Chen Gu
Xiao-Dan Zhu
Quan Liu
Zhenhua Ling
Zhiming Su
Si Wei
21
27
0
08 Apr 2020
HypoNLI: Exploring the Artificial Patterns of Hypothesis-only Bias in
  Natural Language Inference
HypoNLI: Exploring the Artificial Patterns of Hypothesis-only Bias in Natural Language Inference
Tianyu Liu
Xin Zheng
Baobao Chang
Zhifang Sui
43
23
0
05 Mar 2020
A$^3$: Accelerating Attention Mechanisms in Neural Networks with
  Approximation
A3^33: Accelerating Attention Mechanisms in Neural Networks with Approximation
Tae Jun Ham
Sungjun Jung
Seonghak Kim
Young H. Oh
Yeonhong Park
...
Jung-Hun Park
Sanghee Lee
Kyoung Park
Jae W. Lee
D. Jeong
22
211
0
22 Feb 2020
Previous
1234
Next