ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.06084
  4. Cited By
Optimizing Large Language Models through Quantization: A Comparative
  Analysis of PTQ and QAT Techniques

Optimizing Large Language Models through Quantization: A Comparative Analysis of PTQ and QAT Techniques

9 November 2024
Jahid Hasan
    MQ
ArXivPDFHTML

Papers citing "Optimizing Large Language Models through Quantization: A Comparative Analysis of PTQ and QAT Techniques"

1 / 1 papers shown
Title
Sustainable LLM Inference for Edge AI: Evaluating Quantized LLMs for Energy Efficiency, Output Accuracy, and Inference Latency
Sustainable LLM Inference for Edge AI: Evaluating Quantized LLMs for Energy Efficiency, Output Accuracy, and Inference Latency
E. J. Husom
Arda Goknil
Merve Astekin
Lwin Khin Shar
Andre Kåsen
S. Sen
Benedikt Andreas Mithassel
Ahmet Soylu
MQ
43
0
0
04 Apr 2025
1