Skip to content
transformer-assisted-hierarchical-deep-reinforcement-learning-for-energy-and-spectrum-efficient-mimo-mc-cdma-in-6g-networks-–-scientific-reports

Transformer-assisted hierarchical deep reinforcement learning for energy and spectrum efficient MIMO-MC-CDMA in 6G networks – Scientific Reports

  • Article
  • Open access
  • Published:
  • K. R. Saranya1,
  • Y. Suganya1,
  • L. Josephine Usha1 &
  • P. Valarmathi2 

Scientific Reports , Article number:  (2026) Cite this article

We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.

Subjects

Abstract

TH-DRL is proposed for optimizing spectral and energy efficiency in 6G MIMO-MC-CDMA systems. Unlike existing flat DRL or optimization-based approaches, the proposed TH-DRL framework uniquely integrates a lightweight Transformer encoder with a hierarchical decision-making architecture to jointly optimize subcarrier allocation, power control, and SIC ordering. Architecturally, the structure integrates a Transformer encoder with a two-tier hierarchical DRL model to enhance adaptability in dynamic wireless conditions. The Transformer learns spatiotemporal dependencies from channel state information, interference patterns, and user dynamics to generate context-aware features that is used for making effective decisions. A high-level policy-gradient agent handles subcarrier allocation and user clustering, while a low-level DQN agent manages power control and successive interference cancellation order, jointly improving throughput and energy efficiency. Convergence, scalability, and detection performance are evaluated based on Rayleigh channels at 28/100 GHz with bandwidths of 400 MHz–1 GHz serving a varying number of users (10–100) served by a base station equipped with 64–256 antennas. Training consists of a replay buffer of samples within a range of 104–10⁶ over 5000–10,000 episodes. The results showed that the convergence is stable around episode 600 with consistent gain over the baseline methods achieving 15–18% higher spectral efficiency, up to 22% energy savings, and peak performance at SE = 32.7 bits/s/Hz, EE = 14.8 bits/J, SINR ≈ 34 dB, and BER ≈ 10⁻5. This confirms that Transformer-enhanced hierarchical DRL offers scalable, low-latency, energy-aware resource management for dense 6G networks.

Funding

No funding was acquired for the research.

Author information

Authors and Affiliations

  1. Department of Computer Science and Engineering, School of Computing, SRM Institute of Science and Technology, Tiruchirappalli, Tamil Nadu, India

    K. R. Saranya, Y. Suganya & L. Josephine Usha

  2. Department of Computer Science & Engineering, Mookambigai College of Engineering, Pudukottai, India

    P. Valarmathi

Authors

  1. K. R. Saranya
  2. Y. Suganya
  3. L. Josephine Usha
  4. P. Valarmathi

Corresponding author

Correspondence to K. R. Saranya.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Saranya, K.R., Suganya, Y., Usha, L.J. et al. Transformer-assisted hierarchical deep reinforcement learning for energy and spectrum efficient MIMO-MC-CDMA in 6G networks. Sci Rep (2026). https://doi.org/10.1038/s41598-026-43204-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1038/s41598-026-43204-5

Keywords

colind88

Back To Top