Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access
This paper addresses the challenge of multiple users concurrently sharing a single channel in a wireless network, a problem typically managed by Carrier Sense Multiple Access (CSMA) protocols. Traditional CSMA methods, however, often lack robustness to environmental changes due to their reliance on...
Saved in:
| Main Authors: | , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2024-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10769449/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1846128580588732416 |
|---|---|
| author | Chanho Lee Seungkeun Park Taesu Cheong |
| author_facet | Chanho Lee Seungkeun Park Taesu Cheong |
| author_sort | Chanho Lee |
| collection | DOAJ |
| description | This paper addresses the challenge of multiple users concurrently sharing a single channel in a wireless network, a problem typically managed by Carrier Sense Multiple Access (CSMA) protocols. Traditional CSMA methods, however, often lack robustness to environmental changes due to their reliance on static parameters. To overcome this limitation, we propose the Dynamic-Persistent Carrier Sense Multiple Access (DP-CSMA) method, a dynamic and flexible solution inspired by both non-persistent and p-persistent CSMA protocols. Our method incorporates a deep reinforcement learning (DRL) model that dynamically adjusts the waiting period based on the current state and the decision-making process when the channel is sensed idle. This strategy transcends the limitations of static hyperparameters, such as the probability factor in p-persistent CSMA or the contention window in CSMA/CA, which demand careful tuning relative to the number of users. The DRL model in our system captures the dynamic history of previous states and actions using a Long Short-Term Memory (LSTM) model. It efficiently compresses repetitively taken actions into a skill, thereby ensuring a sufficient amount of information is encoded in the action history. Furthermore, our method generates skill-based policies that can induce variable lengths of waiting time for the agents, efficiently handling action sequences of varying lengths and to optimize channel access. We compare the performance of our method with conventional techniques in terms of throughput and evaluate the effectiveness of each method in utilizing the shared medium. |
| format | Article |
| id | doaj-art-fbb5a4cdcf204b46831f543ab4c828f4 |
| institution | Kabale University |
| issn | 2169-3536 |
| language | English |
| publishDate | 2024-01-01 |
| publisher | IEEE |
| record_format | Article |
| series | IEEE Access |
| spelling | doaj-art-fbb5a4cdcf204b46831f543ab4c828f42024-12-11T00:05:31ZengIEEEIEEE Access2169-35362024-01-011217870517871610.1109/ACCESS.2024.350697210769449Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel AccessChanho Lee0https://orcid.org/0009-0008-6621-231XSeungkeun Park1https://orcid.org/0000-0003-4956-8775Taesu Cheong2https://orcid.org/0000-0002-8340-825XSchool of Industrial and Management Engineering, Korea University, Seoul, South KoreaRadio Research Division, Electronics and Telecommunications Research Institute (ETRI), Daejeon, South KoreaSchool of Industrial and Management Engineering, Korea University, Seoul, South KoreaThis paper addresses the challenge of multiple users concurrently sharing a single channel in a wireless network, a problem typically managed by Carrier Sense Multiple Access (CSMA) protocols. Traditional CSMA methods, however, often lack robustness to environmental changes due to their reliance on static parameters. To overcome this limitation, we propose the Dynamic-Persistent Carrier Sense Multiple Access (DP-CSMA) method, a dynamic and flexible solution inspired by both non-persistent and p-persistent CSMA protocols. Our method incorporates a deep reinforcement learning (DRL) model that dynamically adjusts the waiting period based on the current state and the decision-making process when the channel is sensed idle. This strategy transcends the limitations of static hyperparameters, such as the probability factor in p-persistent CSMA or the contention window in CSMA/CA, which demand careful tuning relative to the number of users. The DRL model in our system captures the dynamic history of previous states and actions using a Long Short-Term Memory (LSTM) model. It efficiently compresses repetitively taken actions into a skill, thereby ensuring a sufficient amount of information is encoded in the action history. Furthermore, our method generates skill-based policies that can induce variable lengths of waiting time for the agents, efficiently handling action sequences of varying lengths and to optimize channel access. We compare the performance of our method with conventional techniques in terms of throughput and evaluate the effectiveness of each method in utilizing the shared medium.https://ieeexplore.ieee.org/document/10769449/Channel sharingdynamic channel accessreinforcement learning |
| spellingShingle | Chanho Lee Seungkeun Park Taesu Cheong Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access IEEE Access Channel sharing dynamic channel access reinforcement learning |
| title | Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access |
| title_full | Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access |
| title_fullStr | Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access |
| title_full_unstemmed | Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access |
| title_short | Dynamic-Persistent CSMA: A Reinforcement Learning Approach for Multi-User Channel Access |
| title_sort | dynamic persistent csma a reinforcement learning approach for multi user channel access |
| topic | Channel sharing dynamic channel access reinforcement learning |
| url | https://ieeexplore.ieee.org/document/10769449/ |
| work_keys_str_mv | AT chanholee dynamicpersistentcsmaareinforcementlearningapproachformultiuserchannelaccess AT seungkeunpark dynamicpersistentcsmaareinforcementlearningapproachformultiuserchannelaccess AT taesucheong dynamicpersistentcsmaareinforcementlearningapproachformultiuserchannelaccess |