ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1708.05604
  4. Cited By
Accelerating recurrent neural network training using sequence bucketing
  and multi-GPU data parallelization

Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization

18 August 2017
Viacheslav Khomenko
Oleg Shyshkov
Olga Radyvonenko
Kostiantyn Bokhan
ArXiv (abs)PDFHTML

Papers citing "Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization"

8 / 8 papers shown
Title
Training and Inference Efficiency of Encoder-Decoder Speech Models
Training and Inference Efficiency of Encoder-Decoder Speech Models
Piotr .Zelasko
Kunal Dhawan
Daniel Galvez
Krishna Puvvada
Ankita Pasad
Nithin Rao Koluguri
Ke Hu
Vitaly Lavrukhin
Jagadeesh Balam
Boris Ginsburg
96
0
0
07 Mar 2025
Theano: A Python framework for fast computation of mathematical
  expressions
Theano: A Python framework for fast computation of mathematical expressions
The Theano Development Team
Rami Al-Rfou
Guillaume Alain
Amjad Almahairi
Christof Angermüller
...
Kelvin Xu
Lijun Xue
Li Yao
Saizheng Zhang
Ying Zhang
201
2,340
0
09 May 2016
Recurrent Memory Networks for Language Modeling
Recurrent Memory Networks for Language Modeling
Ke M. Tran
Arianna Bisazza
Christof Monz
42
21
0
06 Jan 2016
BlackOut: Speeding up Recurrent Neural Network Language Models With Very
  Large Vocabularies
BlackOut: Speeding up Recurrent Neural Network Language Models With Very Large Vocabularies
Shihao Ji
S.V.N. Vishwanathan
N. Satish
Michael J. Anderson
Pradeep Dubey
71
77
0
21 Nov 2015
Empirical Evaluation of Gated Recurrent Neural Networks on Sequence
  Modeling
Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling
Junyoung Chung
Çağlar Gülçehre
Kyunghyun Cho
Yoshua Bengio
598
12,734
0
11 Dec 2014
Learning to Execute
Learning to Execute
Wojciech Zaremba
Ilya Sutskever
ODL
87
560
0
17 Oct 2014
Dropout improves Recurrent Neural Networks for Handwriting Recognition
Dropout improves Recurrent Neural Networks for Handwriting Recognition
Vu Pham
Théodore Bluche
Christopher Kermorvant
J. Louradour
111
567
0
05 Nov 2013
ADADELTA: An Adaptive Learning Rate Method
ADADELTA: An Adaptive Learning Rate Method
Matthew D. Zeiler
ODL
161
6,630
0
22 Dec 2012
1