ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1611.05402
  4. Cited By
The ZipML Framework for Training Models with End-to-End Low Precision:
  The Cans, the Cannots, and a Little Bit of Deep Learning

The ZipML Framework for Training Models with End-to-End Low Precision: The Cans, the Cannots, and a Little Bit of Deep Learning

16 November 2016
Hantian Zhang
Jerry Li
Kaan Kara
Dan Alistarh
Ji Liu
Ce Zhang
    MQ
ArXivPDFHTML

Papers citing "The ZipML Framework for Training Models with End-to-End Low Precision: The Cans, the Cannots, and a Little Bit of Deep Learning"

4 / 4 papers shown
Title
How Low Can We Go: Trading Memory for Error in Low-Precision Training
How Low Can We Go: Trading Memory for Error in Low-Precision Training
Chengrun Yang
Ziyang Wu
Jerry Chee
Christopher De Sa
Madeleine Udell
23
2
0
17 Jun 2021
On-Device Machine Learning: An Algorithms and Learning Theory
  Perspective
On-Device Machine Learning: An Algorithms and Learning Theory Perspective
Sauptik Dhar
Junyao Guo
Jiayi Liu
S. Tripathi
Unmesh Kurup
Mohak Shah
33
141
0
02 Nov 2019
CodedPrivateML: A Fast and Privacy-Preserving Framework for Distributed
  Machine Learning
CodedPrivateML: A Fast and Privacy-Preserving Framework for Distributed Machine Learning
Jinhyun So
Başak Güler
A. Avestimehr
FedML
30
114
0
02 Feb 2019
Gradient Sparsification for Communication-Efficient Distributed
  Optimization
Gradient Sparsification for Communication-Efficient Distributed Optimization
Jianqiao Wangni
Jialei Wang
Ji Liu
Tong Zhang
15
522
0
26 Oct 2017
1