ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.08237
  4. Cited By
XLNet: Generalized Autoregressive Pretraining for Language Understanding

XLNet: Generalized Autoregressive Pretraining for Language Understanding

19 June 2019
Zhilin Yang
Zihang Dai
Yiming Yang
J. Carbonell
Ruslan Salakhutdinov
Quoc V. Le
    AI4CE
ArXivPDFHTML

Papers citing "XLNet: Generalized Autoregressive Pretraining for Language Understanding"

50 / 1,487 papers shown
Title
Recommending Metamodel Concepts during Modeling Activities with
  Pre-Trained Language Models
Recommending Metamodel Concepts during Modeling Activities with Pre-Trained Language Models
Martin Weyssow
H. Sahraoui
Eugene Syriani
18
50
0
04 Apr 2021
IITK@Detox at SemEval-2021 Task 5: Semi-Supervised Learning and Dice
  Loss for Toxic Spans Detection
IITK@Detox at SemEval-2021 Task 5: Semi-Supervised Learning and Dice Loss for Toxic Spans Detection
Archit Bansal
Abhay Kaushik
Ashutosh Modi
21
3
0
04 Apr 2021
SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory
  Prediction
SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory Prediction
Liushuai Shi
Le Wang
Chengjiang Long
Sanping Zhou
Mo Zhou
Zhenxing Niu
G. Hua
38
218
0
04 Apr 2021
Normal vs. Adversarial: Salience-based Analysis of Adversarial Samples
  for Relation Extraction
Normal vs. Adversarial: Salience-based Analysis of Adversarial Samples for Relation Extraction
Luoqiu Li
Xiang Chen
Zhen Bi
Xin Xie
Shumin Deng
Ningyu Zhang
Chuanqi Tan
Mosha Chen
Huajun Chen
AAML
39
7
0
01 Apr 2021
Dual Contrastive Loss and Attention for GANs
Dual Contrastive Loss and Attention for GANs
Ning Yu
Guilin Liu
Aysegül Dündar
Andrew Tao
Bryan Catanzaro
Larry S. Davis
Mario Fritz
GAN
34
60
0
31 Mar 2021
Self-supervised Image-text Pre-training With Mixed Data In Chest X-rays
Self-supervised Image-text Pre-training With Mixed Data In Chest X-rays
Xiaosong Wang
Ziyue Xu
Leo K. Tam
Dong Yang
Daguang Xu
ViT
MedIm
25
23
0
30 Mar 2021
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal
  Dependencies
Retraining DistilBERT for a Voice Shopping Assistant by Using Universal Dependencies
P. Jayarao
Arpit Sharma
21
2
0
29 Mar 2021
On the Adversarial Robustness of Vision Transformers
On the Adversarial Robustness of Vision Transformers
Rulin Shao
Zhouxing Shi
Jinfeng Yi
Pin-Yu Chen
Cho-Jui Hsieh
ViT
35
138
0
29 Mar 2021
Efficient Explanations from Empirical Explainers
Efficient Explanations from Empirical Explainers
Robert Schwarzenberg
Nils Feldhus
Sebastian Möller
FAtt
37
9
0
29 Mar 2021
Machine Learning Meets Natural Language Processing -- The story so far
Machine Learning Meets Natural Language Processing -- The story so far
N. Galanis
P. Vafiadis
K.-G. Mirzaev
G. Papakostas
40
7
0
27 Mar 2021
AgentFormer: Agent-Aware Transformers for Socio-Temporal Multi-Agent
  Forecasting
AgentFormer: Agent-Aware Transformers for Socio-Temporal Multi-Agent Forecasting
Ye Yuan
Xinshuo Weng
Yanglan Ou
Kris Kitani
AI4TS
45
442
0
25 Mar 2021
FastMoE: A Fast Mixture-of-Expert Training System
FastMoE: A Fast Mixture-of-Expert Training System
Jiaao He
J. Qiu
Aohan Zeng
Zhilin Yang
Jidong Zhai
Jie Tang
ALM
MoE
45
94
0
24 Mar 2021
Czert -- Czech BERT-like Model for Language Representation
Czert -- Czech BERT-like Model for Language Representation
Jakub Sido
O. Pražák
P. Pribán
Jan Pasek
Michal Seják
Miloslav Konopík
31
43
0
24 Mar 2021
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New
  Multitask Benchmark
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
Nicholas Lourie
Ronan Le Bras
Chandra Bhagavatula
Yejin Choi
LRM
30
137
0
24 Mar 2021
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning
  Architectures
The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures
Sushant Singh
A. Mahmood
AI4TS
60
94
0
23 Mar 2021
Tiny Transformers for Environmental Sound Classification at the Edge
Tiny Transformers for Environmental Sound Classification at the Edge
David Elliott
Carlos E. Otero
Steven Wyatt
Evan Martino
26
15
0
22 Mar 2021
BERT: A Review of Applications in Natural Language Processing and
  Understanding
BERT: A Review of Applications in Natural Language Processing and Understanding
M. V. Koroteev
VLM
25
197
0
22 Mar 2021
Identifying Machine-Paraphrased Plagiarism
Identifying Machine-Paraphrased Plagiarism
Jan Philip Wahle
Terry Ruas
Tomávs Foltýnek
Norman Meuschke
Bela Gipp
11
30
0
22 Mar 2021
Grey-box Adversarial Attack And Defence For Sentiment Classification
Grey-box Adversarial Attack And Defence For Sentiment Classification
Ying Xu
Xu Zhong
Antonio Jimeno Yepes
Jey Han Lau
VLM
AAML
16
53
0
22 Mar 2021
Exploiting Method Names to Improve Code Summarization: A Deliberation
  Multi-Task Learning Approach
Exploiting Method Names to Improve Code Summarization: A Deliberation Multi-Task Learning Approach
Rui Xie
Wei Ye
Jinan Sun
Shikun Zhang
28
26
0
21 Mar 2021
ROSITA: Refined BERT cOmpreSsion with InTegrAted techniques
ROSITA: Refined BERT cOmpreSsion with InTegrAted techniques
Yuanxin Liu
Zheng Lin
Fengcheng Yuan
VLM
MQ
10
18
0
21 Mar 2021
Attribute Alignment: Controlling Text Generation from Pre-trained
  Language Models
Attribute Alignment: Controlling Text Generation from Pre-trained Language Models
Dian Yu
Zhou Yu
Kenji Sagae
21
37
0
20 Mar 2021
Extractive Summarization of Call Transcripts
Extractive Summarization of Call Transcripts
Pratik K. Biswas
Aleksandr Iakubovich
18
10
0
19 Mar 2021
GPT Understands, Too
GPT Understands, Too
Xiao Liu
Yanan Zheng
Zhengxiao Du
Ming Ding
Yujie Qian
Zhilin Yang
Jie Tang
VLM
87
1,147
0
18 Mar 2021
GLM: General Language Model Pretraining with Autoregressive Blank
  Infilling
GLM: General Language Model Pretraining with Autoregressive Blank Infilling
Zhengxiao Du
Yujie Qian
Xiao Liu
Ming Ding
J. Qiu
Zhilin Yang
Jie Tang
BDL
AI4CE
53
1,496
0
18 Mar 2021
Model Extraction and Adversarial Transferability, Your BERT is
  Vulnerable!
Model Extraction and Adversarial Transferability, Your BERT is Vulnerable!
Xuanli He
Lingjuan Lyu
Qiongkai Xu
Lichao Sun
MIACV
SILM
36
91
0
18 Mar 2021
Towards Few-Shot Fact-Checking via Perplexity
Towards Few-Shot Fact-Checking via Perplexity
Nayeon Lee
Yejin Bang
Andrea Madotto
Madian Khabsa
Pascale Fung
AAML
13
90
0
17 Mar 2021
LightningDOT: Pre-training Visual-Semantic Embeddings for Real-Time
  Image-Text Retrieval
LightningDOT: Pre-training Visual-Semantic Embeddings for Real-Time Image-Text Retrieval
Siqi Sun
Yen-Chun Chen
Linjie Li
Shuohang Wang
Yuwei Fang
Jingjing Liu
VLM
41
82
0
16 Mar 2021
TransFG: A Transformer Architecture for Fine-grained Recognition
TransFG: A Transformer Architecture for Fine-grained Recognition
Ju He
Jieneng Chen
Shuai Liu
Adam Kortylewski
Cheng Yang
Yutong Bai
Changhu Wang
ViT
39
376
0
14 Mar 2021
Involution: Inverting the Inherence of Convolution for Visual
  Recognition
Involution: Inverting the Inherence of Convolution for Visual Recognition
Duo Li
Jie Hu
Changhu Wang
Xiangtai Li
Qi She
Lei Zhu
Tong Zhang
Qifeng Chen
BDL
19
304
0
10 Mar 2021
When is it permissible for artificial intelligence to lie? A trust-based
  approach
When is it permissible for artificial intelligence to lie? A trust-based approach
Tae Wan Kim
Tong Lu
Lu
Kyusong Lee
Zhaoqi Cheng
Yanhan Tang
J. N. Hooker
24
4
0
09 Mar 2021
Large Pre-trained Language Models Contain Human-like Biases of What is
  Right and Wrong to Do
Large Pre-trained Language Models Contain Human-like Biases of What is Right and Wrong to Do
P. Schramowski
Cigdem Turan
Nico Andersen
Constantin Rothkopf
Kristian Kersting
33
281
0
08 Mar 2021
MalBERT: Using Transformers for Cybersecurity and Malicious Software
  Detection
MalBERT: Using Transformers for Cybersecurity and Malicious Software Detection
Abir Rahali
M. Akhloufi
32
30
0
05 Mar 2021
Attention is Not All You Need: Pure Attention Loses Rank Doubly
  Exponentially with Depth
Attention is Not All You Need: Pure Attention Loses Rank Doubly Exponentially with Depth
Yihe Dong
Jean-Baptiste Cordonnier
Andreas Loukas
52
373
0
05 Mar 2021
Natural Language Understanding for Argumentative Dialogue Systems in the
  Opinion Building Domain
Natural Language Understanding for Argumentative Dialogue Systems in the Opinion Building Domain
W. A. Abro
Annalena Aicher
Niklas Rach
Stefan Ultes
Wolfgang Minker
Guilin Qi
33
32
0
03 Mar 2021
OAG-BERT: Towards A Unified Backbone Language Model For Academic
  Knowledge Services
OAG-BERT: Towards A Unified Backbone Language Model For Academic Knowledge Services
Xiao Liu
Da Yin
Jingnan Zheng
Xingjian Zhang
Peng Zhang
Hongxia Yang
Yuxiao Dong
Jie Tang
VLM
45
31
0
03 Mar 2021
M6: A Chinese Multimodal Pretrainer
M6: A Chinese Multimodal Pretrainer
Junyang Lin
Rui Men
An Yang
Chan Zhou
Ming Ding
...
Yong Li
Wei Lin
Jingren Zhou
J. Tang
Hongxia Yang
VLM
MoE
37
133
0
01 Mar 2021
PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using
  Transfer Learning
PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using Transfer Learning
Nasi Jofche
Kostadin Mishev
Riste Stojanov
Milos Jovanovik
D. Trajanov
22
17
0
25 Feb 2021
ZJUKLAB at SemEval-2021 Task 4: Negative Augmentation with Language
  Model for Reading Comprehension of Abstract Meaning
ZJUKLAB at SemEval-2021 Task 4: Negative Augmentation with Language Model for Reading Comprehension of Abstract Meaning
Xin Xie
Xiangnan Chen
Xiang Chen
Yong Wang
Ningyu Zhang
Shumin Deng
Huajun Chen
42
2
0
25 Feb 2021
Robust and Transferable Anomaly Detection in Log Data using Pre-Trained
  Language Models
Robust and Transferable Anomaly Detection in Log Data using Pre-Trained Language Models
Harold Ott
Jasmin Bogatinovski
Alexander Acker
S. Nedelkoski
O. Kao
19
29
0
23 Feb 2021
LogME: Practical Assessment of Pre-trained Models for Transfer Learning
LogME: Practical Assessment of Pre-trained Models for Transfer Learning
Kaichao You
Yong Liu
Jianmin Wang
Mingsheng Long
32
178
0
22 Feb 2021
UniT: Multimodal Multitask Learning with a Unified Transformer
UniT: Multimodal Multitask Learning with a Unified Transformer
Ronghang Hu
Amanpreet Singh
ViT
25
296
0
22 Feb 2021
Multilingual Answer Sentence Reranking via Automatically Translated Data
Multilingual Answer Sentence Reranking via Automatically Translated Data
Thuy Vu
Alessandro Moschitti
30
5
0
20 Feb 2021
MUDES: Multilingual Detection of Offensive Spans
MUDES: Multilingual Detection of Offensive Spans
Tharindu Ranasinghe
Marcos Zampieri
27
41
0
18 Feb 2021
Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize
  Long-Tail Visual Concepts
Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Soravit Changpinyo
P. Sharma
Nan Ding
Radu Soricut
VLM
314
1,086
0
17 Feb 2021
COCO-LM: Correcting and Contrasting Text Sequences for Language Model
  Pretraining
COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Yu Meng
Chenyan Xiong
Payal Bajaj
Saurabh Tiwary
Paul N. Bennett
Jiawei Han
Xia Song
127
203
0
16 Feb 2021
Boosting Low-Resource Biomedical QA via Entity-Aware Masking Strategies
Boosting Low-Resource Biomedical QA via Entity-Aware Masking Strategies
Gabriele Pergola
E. Kochkina
Lin Gui
Maria Liakata
Yulan He
88
31
0
16 Feb 2021
Exploring Transformers in Natural Language Generation: GPT, BERT, and
  XLNet
Exploring Transformers in Natural Language Generation: GPT, BERT, and XLNet
M. O. Topal
Anil Bas
Imke van Heerden
LLMAG
AI4CE
26
88
0
16 Feb 2021
Improving speech recognition models with small samples for air traffic
  control systems
Improving speech recognition models with small samples for air traffic control systems
Yi Lin
Qin Li
Bo Yang
Zhen Yan
Huachun Tan
Zhengmao Chen
42
32
0
16 Feb 2021
Overview of the TREC 2020 deep learning track
Overview of the TREC 2020 deep learning track
Nick Craswell
Bhaskar Mitra
Emine Yilmaz
Daniel Fernando Campos
54
371
0
15 Feb 2021
Previous
123...192021...282930
Next