TY - GEN
T1 - Neural Estimation of Multi-User Capacity Regions over Discrete Channels
AU - Huleihel, Bashar
AU - Tsur, Dor
AU - Aharoni, Ziv
AU - Sabag, Oron
AU - Permuter, Haim H.
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/1/1
Y1 - 2024/1/1
N2 - This paper presents a data-driven methodology for estimating capacity regions in multi-user communication scenarios, focusing on channels with discrete alphabets, both with and without feedback. Prior research has successfully utilized neural networks for estimating capacity regions in continuous domains. However, the shift to discrete alphabets introduces a significant challenge due to the lack of end-to-end differentiability of the joint model. To tackle this issue, we first formulate the optimization problem of the causally conditioned directed information rate as a decentralized Markov decision process (MDP). Building on this formulation, we introduce a tractable optimization procedure specifically designed to estimate rate pairs that lie on the boundary of the capacity region. In addressing the inherent complexity of the MDP state space, we employ a reinforcement learning (RL) algorithm to learn optimal policies. We demonstrate the performance of our methodology by applying it to various communication scenarios, including the two-way channel and the multiple access channel (MAC). The results showcase the adaptability and performance of the proposed RL-based framework in estimating capacity regions without explicit knowledge of the underlying channel model, whether there is feedback or not.
AB - This paper presents a data-driven methodology for estimating capacity regions in multi-user communication scenarios, focusing on channels with discrete alphabets, both with and without feedback. Prior research has successfully utilized neural networks for estimating capacity regions in continuous domains. However, the shift to discrete alphabets introduces a significant challenge due to the lack of end-to-end differentiability of the joint model. To tackle this issue, we first formulate the optimization problem of the causally conditioned directed information rate as a decentralized Markov decision process (MDP). Building on this formulation, we introduce a tractable optimization procedure specifically designed to estimate rate pairs that lie on the boundary of the capacity region. In addressing the inherent complexity of the MDP state space, we employ a reinforcement learning (RL) algorithm to learn optimal policies. We demonstrate the performance of our methodology by applying it to various communication scenarios, including the two-way channel and the multiple access channel (MAC). The results showcase the adaptability and performance of the proposed RL-based framework in estimating capacity regions without explicit knowledge of the underlying channel model, whether there is feedback or not.
UR - http://www.scopus.com/inward/record.url?scp=85202842782&partnerID=8YFLogxK
U2 - 10.1109/ISIT57864.2024.10619290
DO - 10.1109/ISIT57864.2024.10619290
M3 - Conference contribution
AN - SCOPUS:85202842782
T3 - IEEE International Symposium on Information Theory - Proceedings
SP - 1191
EP - 1196
BT - 2024 IEEE International Symposium on Information Theory, ISIT 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers
T2 - 2024 IEEE International Symposium on Information Theory, ISIT 2024
Y2 - 7 July 2024 through 12 July 2024
ER -