AI Agents Tune Next-Gen Wireless Networks

Author: Denis Avetisyan


A new framework uses artificial intelligence to optimize cell-free Open RAN systems, paving the way for more efficient and adaptable wireless infrastructure.

The system models a radio intelligent controller (RIC) architecture where supervisor agents translate operator intent into objectives, while near-real-time agents determine user priority and active radio units, all leveraging a shared large language model customized with distinct QLoRA adapters to facilitate efficient and adaptable network control.
The system models a radio intelligent controller (RIC) architecture where supervisor agents translate operator intent into objectives, while near-real-time agents determine user priority and active radio units, all leveraging a shared large language model customized with distinct QLoRA adapters to facilitate efficient and adaptable network control.

This review details an agentic AI approach leveraging large language models and low-rank adaptation for intent-driven optimization of cell-free Massive MIMO O-RAN networks.

Achieving truly autonomous network management remains a challenge despite advancements in open radio access networks (RAN). This paper, ‘Agentic AI for Intent-driven Optimization in Cell-free O-RAN’, introduces a novel framework leveraging large language models to translate high-level operator intents into optimized resource allocation for cell-free massive MIMO systems. By employing a multi-agent approach and parameter-efficient fine-tuning, the proposed system achieves significant energy savings-reducing active O-RUs by up to 41.93%-while drastically minimizing memory requirements. Could this agentic AI architecture pave the way for more intelligent and sustainable wireless networks?


The Evolving Radio Landscape: Embracing Intelligence Through Disaggregation

Conventional Radio Access Networks (RANs), historically built as monolithic, vertically integrated systems, are increasingly challenged by the exponential growth of mobile data demands and the diverse requirements of modern applications. These legacy architectures, designed for predictable traffic patterns, struggle to adapt to the dynamic and often unpredictable fluctuations in throughput and latency necessitated by 5G and beyond. The rigidity of these systems hinders operators’ ability to rapidly deploy new services, scale capacity efficiently, and optimize performance for specific user experiences. Consequently, networks built on these foundations face limitations in supporting emerging technologies like massive Machine Type Communications and ultra-reliable low latency communications, ultimately impacting the potential of future mobile innovations and requiring a fundamental shift towards more agile and adaptable network designs.

The move towards disaggregated Radio Access Networks (RANs), prominently exemplified by the O-RAN Architecture, while promising greater flexibility and vendor diversity, fundamentally increases system complexity. Traditionally, RAN components were tightly integrated and provided by a single vendor, simplifying management and optimization. However, O-RAN breaks down these components – the radio unit, distributed unit, and centralized unit – allowing for interoperability between different suppliers. This disaggregation introduces a multitude of interfaces and potential points of failure, demanding sophisticated, intelligent control mechanisms to ensure seamless operation. Without automated configuration, orchestration, and optimization, managing a disaggregated RAN becomes exponentially more challenging, potentially negating the benefits of increased flexibility and innovation. Consequently, the success of O-RAN hinges on the development and deployment of artificial intelligence and machine learning driven solutions capable of handling this inherent complexity and delivering optimized performance in dynamic network environments.

The move towards open radio access network (RAN) interfaces, while promising greater flexibility and vendor diversity, fundamentally alters network management requirements. Traditional, monolithic RAN systems relied on tightly coupled hardware and software, simplifying optimization but hindering innovation. Open interfaces, however, introduce a distributed and multi-vendor environment demanding intelligent automation for efficient operation. This necessitates a paradigm shift from manual configuration and optimization to self-organizing networks (SON) and artificial intelligence (AI)-driven solutions capable of dynamically adapting to fluctuating traffic patterns, predicting network congestion, and proactively resolving performance issues. Consequently, innovative approaches-including machine learning algorithms for resource allocation and predictive maintenance-are crucial to unlock the full potential of open RAN and ensure optimal network performance and user experience in increasingly complex mobile networks.

This framework iteratively adjusts user weights and O-RU activations-based on operator objectives, monitoring feedback, and prior knowledge-until minimum rate requirements are met, enabling dynamic resource allocation.
This framework iteratively adjusts user weights and O-RU activations-based on operator objectives, monitoring feedback, and prior knowledge-until minimum rate requirements are met, enabling dynamic resource allocation.

Agentic AI: A Collaborative Approach to Network Control

Agentic AI in Radio Access Network (RAN) control employs a distributed architecture consisting of multiple specialized agents, each responsible for a specific task such as performance monitoring, anomaly detection, or resource allocation. These agents operate collaboratively, sharing data and coordinating actions to optimize network performance. This approach contrasts with monolithic control systems by enabling greater adaptability to dynamic network conditions and complex operational scenarios. The modular design allows for independent scaling and updating of individual agents without disrupting overall system functionality, and facilitates the integration of new capabilities as they become available. Furthermore, the collaborative framework allows agents to validate each other’s findings and mitigate the risk of single-point failures, enhancing the resilience of the RAN control plane.

The agentic AI framework relies on Large Language Models (LLMs) to provide the core reasoning and analytical capabilities for each specialized agent. These LLMs are not utilized for simple task execution, but rather for interpreting network data, assessing potential issues, and formulating responses based on their training and defined roles. The LLM’s ability to process and understand natural language allows for the translation of complex network states into actionable insights, and enables agents to collaborate by exchanging information in a structured, understandable format. This facilitates a level of sophisticated decision-making that surpasses traditional rule-based systems, allowing for dynamic adaptation to changing network conditions and proactive issue resolution.

The system automates complex Radio Access Network (RAN) adjustments by interpreting high-level Operator Intent – expressed as desired network outcomes – and converting these into specific, executable objectives. This translation process involves decomposing the intent into measurable Key Performance Indicators (KPIs) and then formulating a series of targeted actions for RAN control elements. Automation extends to tasks such as power optimization, beamforming adjustments, resource allocation, and mobility management, all driven by the translated objectives. The system continuously monitors performance against these objectives and iteratively refines actions to maintain desired network behavior, reducing the need for manual intervention and optimizing RAN performance.

Realizing Intelligent Control: xApps, rApps and Dynamic Optimization

The Near-Real-Time (Near-RT) Radio Intelligent Controller (RIC) platform is designed to host applications, termed xApps, that directly interface with and control network resources with low latency. Specifically, xApps such as the O-RU Management Agent and Monitoring Agent are deployed within the Near-RT RIC to facilitate near-instantaneous adjustments to radio access network parameters. The O-RU Management Agent manages Open Radio Units (O-RUs), while the Monitoring Agent collects performance data. This architecture enables dynamic control loops operating on timescales of tens of milliseconds, crucial for responding to rapidly changing network conditions and optimizing resource allocation in real-time.

The O-RU Management Agent employs Multi-Agent Deep Reinforcement Learning (MADRL) to optimize the selection of active O-RUs (Open Radio Units) within the network. This approach treats each O-RU as an independent agent, allowing for decentralized decision-making based on local observations and rewards. The MADRL algorithm dynamically adjusts the active set of O-RUs to maximize network performance, considering factors such as traffic load, interference, and energy consumption. Through continuous learning and adaptation, the agent aims to identify the optimal configuration of active O-RUs for varying network conditions, improving resource utilization and overall system capacity. The system leverages deep neural networks to approximate the optimal policy for each agent, enabling efficient exploration of the state space and convergence towards a stable and high-performing solution.

The Non-Real-Time (Non-RT) Radio Intelligent Controller (RIC) hosts Radio Application (rApp) instances, with the Supervisor Agent functioning as a key component for translating high-level operator intents into specific, quantifiable objectives for the near-real-time (Near-RT) xApps. This translation process involves decomposing operator goals – such as maximizing network throughput or minimizing latency for specific user groups – into measurable key performance indicators (KPIs). The Supervisor Agent then formulates these KPIs as optimization targets, defining the desired behavior and constraints for the xApps operating in the Near-RT RIC. This ensures that xApp actions are aligned with overall network objectives and operator policies, facilitating coordinated and intelligent network control.

The User Weighting Agent (UWA) functions by assigning priorities to users based on their individual utility, which is quantified using Utility Functions. These functions translate user preferences and Quality of Service (QoS) requirements into a measurable value. To optimize data rates subject to network constraints, the UWA utilizes Lagrange Multipliers, a mathematical technique for finding the maxima or minima of a function subject to equality constraints. Furthermore, the UWA incorporates a Memory Module to retain information about past network performance and user behavior; this historical data is then used to refine the Utility Function parameters and improve the accuracy of prioritization and data rate allocation over time, allowing for adaptive resource management.

User 3's data rate varies depending on the operator's intended action.
User 3’s data rate varies depending on the operator’s intended action.

Adaptability and Future Directions: Cell-Free RAN and Beyond

This innovative framework seamlessly integrates with Cell-Free Open Radio Access Network (O-RAN) architectures, fostering a collaborative environment where multiple O-Radio Units (O-RUs) can collectively deliver network services. By dynamically coordinating resources and intelligently distributing workloads, the system transcends the limitations of traditional, isolated O-RU deployments. This adaptability allows for enhanced network capacity, improved coverage, and greater resilience against individual component failures. The resulting network operates with heightened efficiency, as demand is intelligently shared and balanced across available O-RUs, paving the way for more flexible and scalable future RAN deployments.

The efficiency of modern radio access networks hinges on intelligent resource allocation, and the O-RU Management Agent within this framework achieves this through a sophisticated combination of techniques. Specifically, it employs Weighted Minimum Mean Square Error (WMMSE) for optimized precoding, shaping the transmitted signals to minimize interference and maximize signal quality for each user. Complementing this, a Proportional Fair Scheduler dynamically allocates radio resources – such as time slots and power – based on each user’s instantaneous channel condition and long-term fairness considerations. This synergistic approach ensures that users experiencing poor signal quality receive preferential treatment, while preventing any single user from monopolizing network resources. The result is a system capable of adapting to varying traffic demands and channel conditions, leading to improved spectral efficiency and a more consistent user experience.

To address the substantial computational demands of large language models (LLMs) within radio access networks, researchers have successfully integrated Quantized Low-Rank Adaptation (QLoRA) techniques. This innovative approach dramatically reduces the number of trainable parameters by freezing the pre-trained LLM and introducing a small number of trainable low-rank matrices. By quantizing these matrices to lower precision, QLoRA minimizes both computational overhead and memory footprint without significantly compromising performance. The result is a highly efficient LLM-based agent capable of complex RAN tasks while operating with substantially fewer resources, paving the way for more scalable and sustainable network deployments.

Recent advancements in radio access network (RAN) architecture have yielded substantial gains in network efficiency through intelligent agent deployment. Specifically, this agentic approach, leveraging collaborative O-RUs, demonstrates a marked 41.93% reduction in the number of active O-RUs when contrasted with traditional baseline schemes. This optimization isn’t simply a matter of conserving power; fewer active units translate directly to reduced capital expenditure and operational costs for network providers. The framework achieves this by dynamically allocating resources and intelligently managing radio access, ensuring that capacity is delivered precisely where and when it’s needed, thus minimizing unnecessary hardware activation and maximizing spectral efficiency. The result is a more responsive, cost-effective, and ultimately, sustainable wireless infrastructure.

The proposed framework demonstrates a substantial advancement in resource efficiency for Radio Access Network (RAN) deployments, achieving a remarkable 92% reduction in memory usage when contrasted with traditional approaches that rely on deploying individual Large Language Model (LLM) agents. This minimization is critical as LLMs are inherently memory-intensive; consolidating agent functionality within a unified framework drastically lowers the overall system footprint. The resulting scalability enables more widespread implementation of intelligent RAN features, even within resource-constrained environments, and facilitates the support of a greater number of connected devices and services without compromising performance or incurring prohibitive infrastructure costs. This efficient memory utilization paves the way for more sustainable and economically viable future RAN architectures.

The proportion of active On-Request Units (O-RUs) decreases as either the number of users or the total number of O-RUs increases.
The proportion of active On-Request Units (O-RUs) decreases as either the number of users or the total number of O-RUs increases.

The pursuit of optimized network performance, as detailed in this work concerning agentic AI for cell-free O-RAN, necessitates a rigorous reduction of complexity. The framework’s reliance on low-rank adaptation (QLoRA) exemplifies this principle; it distills the essence of large language models into a manageable form without sacrificing functionality. This aligns with Donald Knuth’s observation: “Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it.” The agentic AI framework prioritizes a streamlined approach to intent translation and optimization, recognizing that elegance and efficiency are paramount to robust, maintainable systems. It is not about adding layers of intricacy, but about achieving maximum effect with minimal components.

Beyond the Horizon

This work demonstrates a path, not a destination. The agentic framework, while promising, sidesteps a crucial question: true intent is rarely singular. Networks serve diverse, often conflicting, objectives. Current approaches translate intent as a directive; future systems must negotiate it, balancing competing demands with a grace absent here. Abstractions age, principles don’t.

QLoRA offers memory efficiency, yet every complexity needs an alibi. Scaling these models-and the agentic systems built upon them-demands more than parameter reduction. It requires fundamentally new architectures that prioritize knowledge distillation and continuous learning, moving beyond static optimization to adaptive resilience. The current focus remains heavily on the ‘how’, neglecting the ‘why’ of network intelligence.

The true test lies not in achieving incremental gains in energy savings, but in realizing a genuinely autonomous network. One that anticipates needs, proactively resolves conflicts, and learns from its own experience without human intervention. That ambition requires a shift in perspective-from optimization as a calculation, to optimization as an emergent property of a self-aware system.


Original article: https://arxiv.org/pdf/2602.22539.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-01 10:13