Abstract
Integrating AI inference into wireless sensing edge networks presents notable challenges due to limited resources, changing environments, and diverse devices. In this study, we proposed a novel resource allocation framework that enhances energy efficiency, reduces latency, and ensures fairness across distributed edge nodes for AI inference. The framework models a multi-objective optimization problem that reflects the interdependence of computation, communication, and energy at each device. We also develop a decentralized algorithm based on dual decomposition and projected gradient ascent, by using local data. The extensive simulations demonstrate that our proposed method reduces the average inference latency by 31.4% and energy consumption by 27.8% compared to the greedy and round-robin techniques. The system utility is improved by up to 59.2%, and fairness, measured using Jain’s index, remains within 8% of the ideal. Additionally, throughput analysis further confirms that our approach gains up to 49 tasks/sec, outperforming existing strategies by more than 40%. These findings show that the resource-aware AI inference approach is scalable, energy-efficient, and appropriate for real-time use in multi-user wireless edge networks.
| Original language | English |
|---|---|
| Article number | 108363 |
| Pages (from-to) | 1-11 |
| Number of pages | 11 |
| Journal | Computer Communications |
| Volume | 245 |
| Early online date | 13 Nov 2025 |
| DOIs | |
| Publication status | Published online - 13 Nov 2025 |
Bibliographical note
Publisher Copyright:© 2025 Elsevier B.V.
Data Access Statement
No data was used for the research described in the article.Funding
The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University, Saudi Arabia for funding this work through the Large Research Project under grant number RGP2/544/46.
| Funders | Funder number |
|---|---|
| King Khalid University | RGP2/544/46 |
Keywords
- Resource allocation
- Edge AI
- Wireless sensing networks
- Decentralized optimization
- Energy efficiency
- Inference latency
- Fairness
- Throughput