Energy-Efficient Ai Model Design for Edge Devices Using Neural Network Pruning and Optimization Techniques
  • Author(s): Rishabh Agrawal ; Himanshu Kumar
  • Paper ID: 1711192
  • Page: 1141-1155
  • Published Date: 30-04-2023
  • Published In: Iconic Research And Engineering Journals
  • Publisher: IRE Journals
  • e-ISSN: 2456-8880
  • Volume/Issue: Volume 6 Issue 10 April-2023
Abstract

AI is progressively being implemented at the edge computing platforms of smartphones, wearables, industrial sensors, and autonomous systems. Although such deployments can support real-time processing, preserve privacy, and decrease reliance on the network, they tend to be restricted by limited computational power, limited memory, and strict power requirements. Without much optimization, the traditional deep neural networks with their large number of parameters and high computational needs are ill-suited to such environments. The present article discusses the neural network pruning and complementary optimization methods as possible solutions to these issues by suggesting the energy-efficient design of AI models. Pruning is used to remove redundant parameters to reduce model size and operations, and quantization is used to encode high-precision weights into low-bit representations to reduce memory and energy usage. Efficiency is additionally improved with knowledge distillation and lightweight architectures with no performance costs, and compiler-level optimizations are applied to guarantee that compressed models can produce real-world runtime gains on a variety of hardware platforms. The discussion combines theoretical knowledge and practical processes, such as step-by-step design processes and example codes, and latency, memory footprint, and energy consumption measuring guidelines on actual models. Issues like accuracy loss, heterogeneity of hardware, and use of standardized benchmarks are critically discussed, and future research directions, including ultra-low-bit networks, hardware-aware neural architecture search, and energy-centric training objectives, are discussed. Through a combination of cutting-edge approaches and deployment-focused ideas, this piece of work highlights that AI minimal energy usage is not only a technical one but an important action towards sustainable, scaled, and accessible edge computing.

Keywords

Energy-Efficient AI; Edge Computing; Neural Network Pruning; Model Optimization; Quantization; Knowledge Distillation; Lightweight Architectures; Compiler Optimizations

Citations

IRE Journals:
Rishabh Agrawal , Himanshu Kumar "Energy-Efficient Ai Model Design for Edge Devices Using Neural Network Pruning and Optimization Techniques" Iconic Research And Engineering Journals Volume 6 Issue 10 2023 Page 1141-1155

IEEE:
Rishabh Agrawal , Himanshu Kumar "Energy-Efficient Ai Model Design for Edge Devices Using Neural Network Pruning and Optimization Techniques" Iconic Research And Engineering Journals, 6(10)