Search

Article

x

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

Markov decision model based handoff selection algorithm for heterogeneous wireless networks

Liang Xiao Qian Zhi-Hong Tian Hong-Liang Wang Xue

Citation:

Markov decision model based handoff selection algorithm for heterogeneous wireless networks

Liang Xiao, Qian Zhi-Hong, Tian Hong-Liang, Wang Xue
PDF
Get Citation

(PLEASE TRANSLATE TO ENGLISH

BY GOOGLE TRANSLATE IF NEEDED.)

  • Coexistence of multiple wireless access technologies will be an indicator of next-generation wireless network, and the integration of heterogeneous wireless networks will meet the needs of high-performance services for mobile users. According to unique quality of service (QoS) requirements of different service type users in heterogeneous environment, the Markov decision model based handoff selection algorithm is proposed in this paper. A heterogeneous wireless network architecture based on the software defined network (SDN) is established to realize the transparency control of heterogeneous networks. Network state information of heterogeneous wireless networks is mastered by SDN controller. It is responsible for scheduling network resources dynamically according to the performance characteristics of each network. If the network state information in equal interval is sampled, the next moment state of network is only related to the current network state and action, but it is not related to the historical state. The problem of handoff selection for heterogeneous wireless networks is modeled as a Markov process with discrete time and continuous state. To predict the next moment state of network by Markov process to obtain a reward, when the reward is positive, it represents the income; when it is negative, it represents the cost. An immediate reward function is constructed for real-time service and non real-time service users respectively according to their different state attributes of the network. Considering five state attributes of wireless network as follows:delay, delay jitter, bandwidth, error rate and network load, the immediate reward function is constructed with weighted summation. Due to the difference in attribute weight distribution among different service type users, the attribute weights are determined by the analytic hierarchy process. In the long term, the objective function which consists of immediate reward function sequence is used to measure future long-term rewards. Then expected reward function based on the state action pair is constructed to obtain the handoff strategy of the maximum expected return by the iterative method of successive approximation. The proposed Markov decision model based handoff selection algorithm is used in simulation of the Matlab platform. The simulation results show that the proposed method can select the optimal handoff strategy for different service type users and reduce the blocking rate, thereby improving the QoS of users and resource utilization of wireless networks.
      Corresponding author: Qian Zhi-Hong, dr.qzh@163.com
    • Funds: Project supported by the National Natural Science Foundation of China (Grant No. 61371092), the Young Scientists Fund of the National Natural Science Foundation of China (Grant No. 61401175), the Key Science and Technology Program of Jilin Province, China (Grant No. 20140204019GX), and the Key Science and Technology Program of Changchun City, China (Grant No. 2014026/14KG021).
    [1]

    Falowo O E, Chan H A 2012 Eurasip J. Wirel. Comm. 221

    [2]

    Zhu K, Niyato D, Wang P 2010 Proceedings of IEEE Wireless Communications and Networking Conference Sydney, Australia, April 18-21, 2010 p1

    [3]

    Yan X, Šekerciğlu lu Y A, Narayanan S 2010 Comput. Networking 54 1848

    [4]

    Liu J, Xiong Q Y, Shi X, Wang K, Shi W R 2015 Chin. Phys. B 24 076401

    [5]

    Ahmed A, Boulahia L M, Gaiti D 2014 IEEE Commun. Surv. Tutorials 16 776

    [6]

    Hasib A, Fapojuwo A 2008 IEEE Trans. Veh. Technol. 57 2426

    [7]

    Kunarak S, Sulessathira R, Dutkiewicz E 2013 Proceedings of IEEE International Conference of Region 10 Xi'an, China, October 22-25, 2013 p1

    [8]

    Salem M, Ismail M, Misran N 2011 J. Appl. Sci. 11 336

    [9]

    Niyato D, Hossain E 2009 IEEE Trans. Veh. Technol. 58 2008

    [10]

    Naghavi P, Rastegar S H, Shah-Mansouri V, Kebriaei H 2016 IEEE Wirel. Commun. Lett. 5 52

    [11]

    Stevens-Navarro E, Martinez-Morales J D, Pineda-Rico U 2012 J. Appl. Res. Technol. 10 534

    [12]

    Wang N, Shi W X, Fan S S, Liu S X 2011 Proceedings of 2nd International Conference on Challenges in Environmental Science and Computer Engineering Haikou, China, December 14-15, 2011 p55

    [13]

    Liu K M 2014 J. Inf. Comput. Secor. 11 3373

    [14]

    Zhu S F, Liu F, Chai Z Y, Qi Y T, Wu J S 2012 Acta Phys. Sin. 61 096401 (in Chinese)[朱思峰, 刘芳, 柴争义, 戚玉涛, 吴建设2012物理学报61 096401]

    [15]

    Ning Z L, Song Q Y, Liu Y J, Wang F Z, Wu X Y 2014 Comput. Electr. Eng. 40 456

    [16]

    Ma B, Deng H, Xie X Z, Liao X F 2015 China Commun. 12 106

    [17]

    Ma B, Xie X Z, Liao X F 2015 J. Electron. Inform. Technol. 37 874 (in Chinese)[马彬, 谢显中, 廖晓峰2015电子与信息学报37 874]

    [18]

    Chen T, Matinmikko M, Chen X F, Zhou X, Ahokangas P 2015 IEEE Commun. Mag. 53 126

    [19]

    Wang H C, Chen S Z, Xu H, Ai M, Shi Y 2015 IEEE Network 29 16

    [20]

    Shen Y 2013 Chin. Phys. B 22 058902

    [21]

    Yang X L, Tan X Z, Guan K 2015 Acta Phys. Sin. 64 108403 (in Chinese)[杨小龙, 谭学治, 关凯2015物理学报64 108403]

    [22]

    Tsai C, Yang F N 2013 J. Hydraul. Eng. 139 1265

    [23]

    Fei R, Cui D W 2009 Acta Phys. Sin. 58 5133 (in Chinese)[费蓉, 崔杜武2009物理学报58 5133]

    [24]

    Alavipoor F S, Karimi S, Balist J, Khakian A H 2016 Global. J. Environ. Sci. Manage. 2 197

    [25]

    Marco W, Martijn V O 2012 Reinforcement Learning:State of the Art (Berlin:Springer) pp223-229

  • [1]

    Falowo O E, Chan H A 2012 Eurasip J. Wirel. Comm. 221

    [2]

    Zhu K, Niyato D, Wang P 2010 Proceedings of IEEE Wireless Communications and Networking Conference Sydney, Australia, April 18-21, 2010 p1

    [3]

    Yan X, Šekerciğlu lu Y A, Narayanan S 2010 Comput. Networking 54 1848

    [4]

    Liu J, Xiong Q Y, Shi X, Wang K, Shi W R 2015 Chin. Phys. B 24 076401

    [5]

    Ahmed A, Boulahia L M, Gaiti D 2014 IEEE Commun. Surv. Tutorials 16 776

    [6]

    Hasib A, Fapojuwo A 2008 IEEE Trans. Veh. Technol. 57 2426

    [7]

    Kunarak S, Sulessathira R, Dutkiewicz E 2013 Proceedings of IEEE International Conference of Region 10 Xi'an, China, October 22-25, 2013 p1

    [8]

    Salem M, Ismail M, Misran N 2011 J. Appl. Sci. 11 336

    [9]

    Niyato D, Hossain E 2009 IEEE Trans. Veh. Technol. 58 2008

    [10]

    Naghavi P, Rastegar S H, Shah-Mansouri V, Kebriaei H 2016 IEEE Wirel. Commun. Lett. 5 52

    [11]

    Stevens-Navarro E, Martinez-Morales J D, Pineda-Rico U 2012 J. Appl. Res. Technol. 10 534

    [12]

    Wang N, Shi W X, Fan S S, Liu S X 2011 Proceedings of 2nd International Conference on Challenges in Environmental Science and Computer Engineering Haikou, China, December 14-15, 2011 p55

    [13]

    Liu K M 2014 J. Inf. Comput. Secor. 11 3373

    [14]

    Zhu S F, Liu F, Chai Z Y, Qi Y T, Wu J S 2012 Acta Phys. Sin. 61 096401 (in Chinese)[朱思峰, 刘芳, 柴争义, 戚玉涛, 吴建设2012物理学报61 096401]

    [15]

    Ning Z L, Song Q Y, Liu Y J, Wang F Z, Wu X Y 2014 Comput. Electr. Eng. 40 456

    [16]

    Ma B, Deng H, Xie X Z, Liao X F 2015 China Commun. 12 106

    [17]

    Ma B, Xie X Z, Liao X F 2015 J. Electron. Inform. Technol. 37 874 (in Chinese)[马彬, 谢显中, 廖晓峰2015电子与信息学报37 874]

    [18]

    Chen T, Matinmikko M, Chen X F, Zhou X, Ahokangas P 2015 IEEE Commun. Mag. 53 126

    [19]

    Wang H C, Chen S Z, Xu H, Ai M, Shi Y 2015 IEEE Network 29 16

    [20]

    Shen Y 2013 Chin. Phys. B 22 058902

    [21]

    Yang X L, Tan X Z, Guan K 2015 Acta Phys. Sin. 64 108403 (in Chinese)[杨小龙, 谭学治, 关凯2015物理学报64 108403]

    [22]

    Tsai C, Yang F N 2013 J. Hydraul. Eng. 139 1265

    [23]

    Fei R, Cui D W 2009 Acta Phys. Sin. 58 5133 (in Chinese)[费蓉, 崔杜武2009物理学报58 5133]

    [24]

    Alavipoor F S, Karimi S, Balist J, Khakian A H 2016 Global. J. Environ. Sci. Manage. 2 197

    [25]

    Marco W, Martijn V O 2012 Reinforcement Learning:State of the Art (Berlin:Springer) pp223-229

Metrics
  • Abstract views:  5883
  • PDF Downloads:  249
  • Cited By: 0
Publishing process
  • Received Date:  26 July 2016
  • Accepted Date:  07 September 2016
  • Published Online:  05 December 2016

/

返回文章
返回