Rivaling Transformers: Multi-Scale Structured State-Space Mixtures for Agentic 6G O-RAN

dc.contributor.authorRezazadeh, Farhad
dc.contributor.authorChergui, Hatim
dc.contributor.authorDebbah, Merouane
dc.contributor.authorSong, Houbing
dc.contributor.authorNiyato, Dusit
dc.contributor.authorLiu, Lingjia
dc.date.accessioned2025-11-21T00:30:24Z
dc.date.issued2025-10-06
dc.description.abstractIn sixth-generation (6G) Open Radio Access Networks (O-RAN), proactive control is preferable. A key open challenge is delivering control-grade predictions within Near-Real-Time (Near-RT) latency and computational constraints under multi-timescale dynamics. We therefore cast RAN Intelligent Controller (RIC) analytics as an agentic perceive-predict xApp that turns noisy, multivariate RAN telemetry into short-horizon per-User Equipment (UE) key performance indicator (KPI) forecasts to drive anticipatory control. In this regard, Transformers are powerful for sequence learning and time-series forecasting, but they are memory-intensive, which limits Near-RT RIC use. Therefore, we need models that maintain accuracy while reducing latency and data movement. To this end, we propose a lightweight Multi-Scale Structured State-Space Mixtures (MS³M)¹ forecaster that mixes HiPPO-LegS kernels to capture multi-timescale radio dynamics. We develop stable discrete state-space models (SSMs) via bilinear (Tustin) discretization and apply their causal impulse responses as per-feature depthwise convolutions. Squeeze-and-Excitation gating dynamically reweights KPI channels as conditions change, and a compact gated channel-mixing layer models cross-feature nonlinearities without Transformer-level cost. The model is KPI-agnostic -- Reference Signal Received Power (RSRP) serves as a canonical use case -- and is trained on sliding windows to predict the immediate next step. Empirical evaluations conducted using our bespoke O-RAN testbed KPI time-series dataset (59,441 windows across 13 KPIs). Crucially for O-RAN constraints, MS³M achieves a 0.057 s per-inference latency with ∼0.70M parameters, yielding 3-10x lower latency than the Transformer baselines evaluated on the same hardware, while maintaining competitive accuracy.
dc.description.urihttp://arxiv.org/abs/2510.05255
dc.format.extent12 pages
dc.genrejournal articles
dc.genrepreprints
dc.identifierdoi:10.13016/m2ihqy-1xpd
dc.identifier.urihttps://doi.org/10.48550/arXiv.2510.05255
dc.identifier.urihttp://hdl.handle.net/11603/40882
dc.language.isoen
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Information Systems Department
dc.relation.ispartofUMBC Faculty Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.subjectComputer Science - Networking and Internet Architecture
dc.subjectUMBC Security and Optimization for Networked Globe Laboratory (SONG Lab)
dc.titleRivaling Transformers: Multi-Scale Structured State-Space Mixtures for Agentic 6G O-RAN
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0003-2631-9223

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
251005255v1.pdf
Size:
1.91 MB
Format:
Adobe Portable Document Format