Comparing Efficiency Strategies for LLM Deployment and Summarizing PowerInfer‑2’s Impact Post date November 3, 2025 Post author By Writings, Papers and Blogs on Text Models Post categories In edge-computing, mobile-ai, model-optimization, neural-efficiency, on-device-llm, power-infer, quantization, speculative-decoding
A Quick Guide to Quantization for LLMs Post date September 11, 2025 Post author By James Post categories In ai, ai-quantization-explained, artificial-intelligence, large-language-models, llm, quantization, quantization-for-llms, technology
Accelerating Neural Networks: The Power of Quantization Post date May 28, 2025 Post author By Vladislav Ag Post categories In accelerating-neural-networks, ai, deep-learning, neural-networks, pytorch, quantization, symmetric-quantization, what-is-quantization
Quantization-Aware Training With PyTorch Post date May 5, 2025 Post author By Sahib Dhanjal Post categories In deep-learning, Optimization, python, pytorch, quantization
Increased LLM Vulnerabilities from Fine-tuning and Quantization: Appendix Post date October 17, 2024 Post author By Quantization Post categories In adversarial-attacks, alignment-training, fine-tuning, guardrails, jailbreaking, large-language-models-(llms), quantization, vulnerabilities
Increased LLM Vulnerabilities from Fine-tuning and Quantization: Conclusion and References Post date October 17, 2024 Post author By Quantization Post categories In adversarial-attacks, alignment-training, fine-tuning, guardrails, jailbreaking, large-language-models-(llms), quantization, vulnerabilities
Increased LLM Vulnerabilities from Fine-tuning and Quantization: Experiment Set-up & Results Post date October 17, 2024 Post author By Quantization Post categories In adversarial-attacks, alignment-training, fine-tuning, guardrails, jailbreaking, large-language-models-(llms), quantization, vulnerabilities