Table of Contents Introduction Why Edge Inference Is Hard: Constraints & Opportunities Small Language Models (SLMs): The Right Fit for Edge Quantization Fundamentals 4.1 Post‑Training Quantization (PTQ) 4.2 Quantization‑Aware Training (QAT) Quantization Strategies Tailored for Real‑Time Edge 5.1 Uniform vs. Non‑Uniform Quantization 5.2 Per‑Tensor vs. Per‑Channel Scaling 5.3 Weight‑Only Quantization 5.4 Activation Quantization & Mixed‑Precision 5.5 Group‑Wise and Block‑Wise Quantization (GPTQ, AWQ, SmoothQuant) Toolchains & Libraries You Can Use Today Step‑by‑Step Practical Workflow 7.1 Selecting an SLM 7.2 Preparing Calibration Data 7.3 Applying Quantization (Code Example) 7.4 Benchmarking Latency & Accuracy Real‑World Case Studies 8.1 Smart Camera Captioning on Raspberry Pi 4 8.2 Voice Assistant on NVIDIA Jetson Nano 8.3 Industrial IoT Summarizer on Coral Dev Board Optimizing for Real‑Time: Beyond Quantization 9.1 Token‑Level Streaming & KV‑Cache Management 9.2 Batch‑Size‑One & Pipeline Parallelism 9.3 Hardware‑Accelerator Specific Tricks Trade‑offs, Pitfalls, and Best Practices Future Directions in Edge LLM Quantization Conclusion Resources Introduction Large language models (LLMs) have transformed everything from code generation to conversational AI. Yet the majority of breakthroughs still happen in the cloud, where GPUs, high‑speed interconnects, and terabytes of RAM are taken for granted. For many applications—autonomous drones, on‑device assistants, industrial control panels, or privacy‑sensitive healthcare devices—sending data to a remote server is simply not an option. The challenge is clear: run LLM inference locally, in real time, on hardware that is orders of magnitude less capable than a data‑center GPU.
...