Skip to content
February 6, 2026
Mochiai.blog
Mochiai.blog
Random Article
  • Home
  • LLM inference optimization

Tag: LLM inference optimization

Accelerating LLM inference with post-training weight and activation using AWQ and GPTQ on Amazon SageMaker AI | Amazon Web Services
Categories Artificial intelligence

Accelerating LLM inference with post-training weight and activation using AWQ and GPTQ on Amazon SageMaker AI | Amazon Web Services

  • By
  • January 9, 2026

Foundation models (FMs) and large language models (LLMs) have been rapidly scaling, often doubling in parameter count within months, leading to significant improvements in…

Read More
Architecture
Categories Artificial intelligence

Optimizing LLM inference on Amazon SageMaker AI with BentoMLs LLM- Optimizer | Amazon Web Services

  • By
  • December 24, 2025

The rise of powerful large language models (LLMs) that can be consumed via API calls has made it remarkably straightforward to integrate artificial intelligence…

Read More

Loading...

Categories

  • AI Medical
  • AI Reasoning Model
  • Artificial intelligence
  • Best Exam for AI
  • Cybersecurity
  • Machine Learning
  • Programming & Tech
  • Technology
  • Uncategorized
  • VM

Archives

  • February 2026
  • January 2026
  • December 2025
  • November 2025
  • October 2025
  • September 2025
  • August 2025
  • July 2025
  • May 2025
  • April 2025
  • March 2025
  • February 2025
  • January 2025
  • December 2024
  • November 2024
  • October 2024
  • September 2024
  • August 2024
  • July 2024
  • June 2024
  • May 2024
  • April 2024
  • April 2016

Copyright © 2026
 - Powered by Magze.