ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2101.08133
13
53

Active Learning for Sequence Tagging with Deep Pre-trained Models and Bayesian Uncertainty Estimates

20 January 2021
Artem Shelmanov
Dmitri Puzyrev
L. Kupriyanova
D. Belyakov
Daniil Larionov
Nikita Khromov
Olga Kozlova
Ekaterina Artemova
Dmitry V. Dylov
Alexander Panchenko
    BDL
    UQLM
    UQCV
ArXivPDFHTML
Abstract

Annotating training data for sequence tagging of texts is usually very time-consuming. Recent advances in transfer learning for natural language processing in conjunction with active learning open the possibility to significantly reduce the necessary annotation budget. We are the first to thoroughly investigate this powerful combination for the sequence tagging task. We conduct an extensive empirical study of various Bayesian uncertainty estimation methods and Monte Carlo dropout options for deep pre-trained models in the active learning framework and find the best combinations for different types of models. Besides, we also demonstrate that to acquire instances during active learning, a full-size Transformer can be substituted with a distilled version, which yields better computational performance and reduces obstacles for applying deep active learning in practice.

View on arXiv
Comments on this paper