Abstract
The problem of maximum likelihood decoding with a neural decoder for error-correcting code is considered. It is shown that the neural decoder can be improved with two novel loss terms on the node's activations. The first loss term imposes a sparse constraint on the node's activations. Whereas, the second loss term tried to mimic the node's activations from a teacher decoder which has better performance. The proposed method has the same run time complexity and model size as the neural Belief Propagation decoder, while improving the decoding performance by up to 1.1dB on BCH codes.
Original language | English |
---|---|
Pages (from-to) | 2527-2531 |
Number of pages | 5 |
Journal | IEEE Communications Letters |
Volume | 26 |
Issue number | 11 |
DOIs | |
State | Published - 1 Nov 2022 |
Externally published | Yes |
Keywords
- Information theory
- deep learning
- error correcting codes
- neural decoder
ASJC Scopus subject areas
- Modeling and Simulation
- Computer Science Applications
- Electrical and Electronic Engineering