TY - JOUR
T1 - Algorithms for dynamic spectrum access with learning for cognitive radio
AU - Unnikrishnan, Jayakrishnan
AU - Veeravalli, Venugopal V.
N1 - Manuscript received June 10, 2008; accepted July 12, 2009. First published August 04, 2009; current version published January 13, 2010. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Qing Zhao. This work was supported by a Vodafone Foundation Graduate Fellowship and by a NSF Grant CCF 07-29031, through the University of Illinois. This paper was presented in part at the 47th IEEE Conference on Decision and Control, Cancun, Mexico, 2008 and at theAsilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, 2008.
PY - 2010/2
Y1 - 2010/2
N2 - We study the problem of dynamic spectrum sensing and access in cognitive radio systems as a partially observed Markov decision process (POMDP). A group of cognitive users cooperatively tries to exploit vacancies in primary (licensed) channels whose occupancies follow a Markovian evolution. We first consider the scenario where the cognitive users have perfect knowledge of the distribution of the signals they receive from the primary users. For this problem, we obtain a greedy channel selection and access policy that maximizes the instantaneous reward, while satisfying a constraint on the probability of interfering with licensed transmissions. We also derive an analytical universal upper bound on the performance of the optimal policy. Through simulation, we show that our scheme achieves good performance relative to the upper bound and improved performance relative to an existing scheme. We then consider the more practical scenario where the exact distribution of the signal from the primary is unknown. We assume a parametric model for the distribution and develop an algorithm that can learn the true distribution, still guaranteeing the constraint on the interference probability. We show that this algorithm outperforms the naive design that assumes a worst case value for the parameter. We also provide a proof for the convergence of the learning algorithm.
AB - We study the problem of dynamic spectrum sensing and access in cognitive radio systems as a partially observed Markov decision process (POMDP). A group of cognitive users cooperatively tries to exploit vacancies in primary (licensed) channels whose occupancies follow a Markovian evolution. We first consider the scenario where the cognitive users have perfect knowledge of the distribution of the signals they receive from the primary users. For this problem, we obtain a greedy channel selection and access policy that maximizes the instantaneous reward, while satisfying a constraint on the probability of interfering with licensed transmissions. We also derive an analytical universal upper bound on the performance of the optimal policy. Through simulation, we show that our scheme achieves good performance relative to the upper bound and improved performance relative to an existing scheme. We then consider the more practical scenario where the exact distribution of the signal from the primary is unknown. We assume a parametric model for the distribution and develop an algorithm that can learn the true distribution, still guaranteeing the constraint on the interference probability. We show that this algorithm outperforms the naive design that assumes a worst case value for the parameter. We also provide a proof for the convergence of the learning algorithm.
UR - https://www.scopus.com/pages/publications/74949140186
UR - https://www.scopus.com/pages/publications/74949140186#tab=citedBy
U2 - 10.1109/TSP.2009.2028970
DO - 10.1109/TSP.2009.2028970
M3 - Article
AN - SCOPUS:74949140186
SN - 1053-587X
VL - 58
SP - 750
EP - 760
JO - IEEE Transactions on Signal Processing
JF - IEEE Transactions on Signal Processing
IS - 2
M1 - 5191108
ER -