Learned Scheduling of LDPC Decoders Based on Multi-armed Bandits
Document Type
Conference Proceeding
Publication Date
6-1-2020
Abstract
The multi-armed bandit (MAB) problem refers to the dilemma encountered by a gambler when deciding which arm of a multi-armed slot machine to pull in order to maximize the total reward earned in a sequence of pulls. In this paper, we model the scheduling of a node-wise sequential LDPC decoder as a Markov decision process, where the underlying Tanner graph is viewed as a slot machine with multiple arms corresponding to the check nodes. A fictitious gambler decides which check node to pull (schedule) next by observing a reward associated with each pull. This interaction enables the gambler to discover an optimized scheduling policy that aims to reach a codeword output by propagating the fewest possible messages. Based on this policy, we contrive a novel MAB-based node-wise scheduling (MABNS) algorithm to perform sequential decoding of LDPC codes. Simulation results show that the MAB-NS scheme, aided by an appropriate scheduling policy, outperforms traditional scheduling schemes in terms of complexity and bit error probability.
Identifier
85090412598 (Scopus)
ISBN
[9781728164328]
Publication Title
IEEE International Symposium on Information Theory Proceedings
External Full Text Location
https://doi.org/10.1109/ISIT44484.2020.9174337
ISSN
21578095
First Page
2789
Last Page
2794
Volume
2020-June
Grant
ECCS-1711056
Fund Ref
National Science Foundation
Recommended Citation
Habib, Salman; Beemer, Allison; and Kliewer, Jorg, "Learned Scheduling of LDPC Decoders Based on Multi-armed Bandits" (2020). Faculty Publications. 5244.
https://digitalcommons.njit.edu/fac_pubs/5244
