Unlock Peak Mobile Performance: A Deep Dive into PowerInfer-2’s Neuron-Aware Runtime Post date August 26, 2025 Post author By Writings, Papers and Blogs on Text Models Post categories In Edge AI, heterogeneous-computing, llm-inference-optimization, mobile-computing, neuron-cluster, on-device-ai, power-infer-2, system-for-ml
The Conductor in Your Pocket: How PowerInfer-2 Orchestrates Smartphone Hardware for LLM Inference Post date August 26, 2025 Post author By Writings, Papers and Blogs on Text Models Post categories In Edge AI, heterogeneous-computing, llm-inference, mobile-computing, neuron-cluster, on-device-ai, power-infer-2, system-for-ml
Why Your Phone’s AI is Slow: A Story of Sparse Neurons and Finicky Flash Storage Post date August 26, 2025 Post author By Writings, Papers and Blogs on Text Models Post categories In edge-computing, llm-inference, mobile-system, on-device-ai, performance-analysis, sparse-activation, system-for-ml, ufs4
PowerInfer-2 Achieves 29x Speedup, Running 47-Billion Parameter LLMs on Smartphones Post date August 26, 2025 Post author By Writings, Papers and Blogs on Text Models Post categories In Edge AI, efficient-ai, heterogeneous-computing, mobile-ai, on-device-language-models, power-infer-2, system-for-ml