What is it about?

An optimal preemption framework is proposed to minimize Age of Information (AoI) in single-link systems with stochastic packet arrivals and known packet lengths. The problem is formulated as a Markov Decision Process (MDP) solved through relative value iteration (RVI), demonstrating that packet-length-dependent threshold policies achieve minimal average AoI. For scenarios with unknown traffic statistics, a Deep Q-Network (DQN) algorithm learns these thresholds adaptively through real-time interactions without prior distribution knowledge. Numerical evaluations under exponential and bounded Pareto packet length distributions reveal 15.5 percent average AoI reduction compared to non-preemptive baselines using the RVI method, while the DQN method achieves 14.4 percent improvement with less than 1.1 percent performance gap from the optimal policy. Both methods exhibit threshold-driven decision structures that balance immediate AoI reduction against long-term scheduling efficiency. Comparative analysis reveals the framework’s robustness across stationary and dynamic environments, with the DQN maintaining near-optimal performance through dimension-reduced state representation. These results establish a unified solution for AoI minimization that transitions seamlessly between model-aware and model-agnostic configurations, addressing critical challenges in real-time Internet of Things (IoT) networks and status update systems requiring freshness guarantees.

Featured Image

Why is it important?

This paper is significantly important because it pioneers an optimal preemption framework for Age of Information (AoI) minimization by leveraging known packet lengths, a practical yet often overlooked feature in modern networks. Its uniqueness lies in the discovery and proof of a packet-length-dependent threshold policy, which elegantly balances the trade-off between finishing a current transmission and switching to a fresher arrival. This research is particularly timely as it addresses the ultra-low latency and high freshness requirements of 6G and industrial IoT, offering a dual-approach solution—combining rigorous mathematical optimization (RVI) with adaptive Deep Reinforcement Learning (DQN)—that reduces average AoI by up to 15.5%, providing a robust and scalable blueprint for real-time status update systems.

Perspectives

From my perspective, this publication is particularly compelling because it bridges the gap between high-level communication theory and practical network implementation. While much of the existing research on Age of Information (AoI) relies on idealized, memoryless models where we "hope" for the best, this work acknowledges a simple but powerful reality: in modern digital systems, we often know exactly how large a data packet is before we start sending it. We move away from a "one-size-fits-all" preemption rule. By proving that the optimal strategy is a dynamic threshold—one that accounts for both the age of the current data and the specific size of the new arrival—we provide a much more nuanced tool for network engineers. Furthermore, the inclusion of a DQN-based learning approach demonstrates a forward-thinking attitude toward "self-optimizing" networks. It suggests a future where systems don't need to be perfectly pre-configured by humans but can instead learn the most efficient way to stay "fresh" simply by observing their own traffic patterns. This blend of rigorous mathematical proof and adaptive machine learning makes the paper a significant step toward truly autonomous, low-latency communication systems.

xiyue Li
Sun Yat-Sen University

Read the Original

This page is a summary of: Optimal Preemption Policy for Age of Information Minimization with Random Arrival and Known Packet Length, ACM Transactions on Sensor Networks, February 2026, ACM (Association for Computing Machinery),
DOI: 10.1145/3794849.
You can read the full text:

Read

Contributors

The following have contributed to this page