摘要
In this letter, we propose a low-complexity hypernetwork for channel neural decoding with learning weights in tensor ring (TR) format, called TRHyper. The internal parameters and the number of layers of the TRHyper based channel neural decoding algorithm can be updated without retraining. We design the size of each TRHyper layer according to the size of the factor tensor in tensor ring format. During the training phase, we reuse the storage space for the learning weights of the main decoding network, so the proposed TRHyper no longer require additional storage space for its learning weights. Numerical results show that for low-density parity check (LDPC) codes, the performance of the TRHyper based channel neural decoder is similar to that of the original decoder, while for Bose-Chaudhuri-Hocquenghem (BCH) codes, the performance slightly exceeds the original decoder.
| 原文 | English |
|---|---|
| 頁(從 - 到) | 918-922 |
| 頁數 | 5 |
| 期刊 | IEEE Communications Letters |
| 卷 | 29 |
| 發行號 | 5 |
| DOIs | |
| 出版狀態 | Published - 2025 |
指紋
深入研究「TRHyper: Low-Complexity Hypernetwork for Channel Neural Decoding With Learning Weights in Tensor Ring Format」主題。共同形成了獨特的指紋。引用此
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver