It’s become a common pattern over the past few months: A business leader makes an ominous claim about the impact of AI on the labor market, sparking weeks of discourse and sending U.S. workers reeling. Now, a “godfather” of AI is pushing back on those claims—and warning about the dangers they spur. In an interview
Transformer-assisted hierarchical deep reinforcement learning for energy and spectrum efficient MIMO-MC-CDMA in 6G networks – Scientific Reports
- Article
- Open access
- Published:
- K. R. Saranya1,
- Y. Suganya1,
- L. Josephine Usha1 &
- …
- P. Valarmathi2
Scientific Reports , Article number: (2026) Cite this article
We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.
Subjects
Abstract
TH-DRL is proposed for optimizing spectral and energy efficiency in 6G MIMO-MC-CDMA systems. Unlike existing flat DRL or optimization-based approaches, the proposed TH-DRL framework uniquely integrates a lightweight Transformer encoder with a hierarchical decision-making architecture to jointly optimize subcarrier allocation, power control, and SIC ordering. Architecturally, the structure integrates a Transformer encoder with a two-tier hierarchical DRL model to enhance adaptability in dynamic wireless conditions. The Transformer learns spatiotemporal dependencies from channel state information, interference patterns, and user dynamics to generate context-aware features that is used for making effective decisions. A high-level policy-gradient agent handles subcarrier allocation and user clustering, while a low-level DQN agent manages power control and successive interference cancellation order, jointly improving throughput and energy efficiency. Convergence, scalability, and detection performance are evaluated based on Rayleigh channels at 28/100 GHz with bandwidths of 400 MHz–1 GHz serving a varying number of users (10–100) served by a base station equipped with 64–256 antennas. Training consists of a replay buffer of samples within a range of 104–10⁶ over 5000–10,000 episodes. The results showed that the convergence is stable around episode 600 with consistent gain over the baseline methods achieving 15–18% higher spectral efficiency, up to 22% energy savings, and peak performance at SE = 32.7 bits/s/Hz, EE = 14.8 bits/J, SINR ≈ 34 dB, and BER ≈ 10⁻5. This confirms that Transformer-enhanced hierarchical DRL offers scalable, low-latency, energy-aware resource management for dense 6G networks.
Funding
No funding was acquired for the research.
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
Reprints and permissions
About this article
Cite this article
Saranya, K.R., Suganya, Y., Usha, L.J. et al. Transformer-assisted hierarchical deep reinforcement learning for energy and spectrum efficient MIMO-MC-CDMA in 6G networks. Sci Rep (2026). https://doi.org/10.1038/s41598-026-43204-5
Download citation
-
Received:
-
Accepted:
-
Published:
-
DOI: https://doi.org/10.1038/s41598-026-43204-5
