Skip to main content
Log in

MADRL-based bitrate allocation for QoE fairness in 360° video streaming with viewport prediction

  • Regular Paper
  • Published:
Multimedia Systems Aims and scope Submit manuscript

Abstract

Compared to traditional video streaming, 360° video streaming consumes significantly more bandwidth. To address this, tile-based methods have been proposed to reduce bandwidth consumption and enhance the quality of experience (QoE) for users. These methods employ viewport prediction algorithm to select tiles containing the user's region of interest and utilize bitrate allocation algorithm to determine appropriate bitrates for different tiles. However, current viewport prediction algorithms insufficiently utilize viewport spatial position information, affecting prediction accuracy. Additional tile transmissions caused by prediction errors result in bandwidth wastage. Furthermore, the spatial offset of the viewport directly influences tile transmission throughput, making accurate viewport prediction essential for enhancing the performance of bitrate allocation algorithms. Existing bitrate allocation algorithms primarily focus on maximizing QoE for a single user. In the practical environment, factors such as varying viewports among users and time-varying network features can lead to unbalanced QoE distribution among users, making it challenging to ensure QoE fairness among multiple users. To tackle these issues, we propose an algorithm combining Spatio-Temporal GCN-based viewport prediction and Multi-agent deep Reinforcement learning-based bitrate allocation for users’ QoE-fairness (STMRQ). Firstly, a viewport prediction algorithm based on spatio-temporal graph convolutional network is introduced, which extracts the spatio-temporal features of video tiles through a novel viewport position matrix, thereby improving the accuracy and computational efficiency of viewport prediction. Subsequently, a 360° video streaming bitrate allocation algorithm based on multi-agent reinforcement learning is proposed, modeling the environment of multi-user dynamic bandwidth streaming on demand. By leveraging state parameters such as user viewport variability, client status, and global network features, we employ a multi-agent reinforcement learning algorithm to train bitrate allocation strategies in a multi-user environment. Extensive experimental evaluations demonstrate that, compared to existing viewport prediction methods and bitrate allocation algorithms, the proposed STMRQ achieves a higher average QoE across common datasets and effectively ensures QoE fairness. Additionally, the accuracy of viewport prediction is improved by up to double, with relatively low computational overhead during training.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+
from €37.37 /Month
  • Starting from 10 chapters or articles per month
  • Access and download chapters and articles from more than 300k books and 2,500 journals
  • Cancel anytime
View plans

Buy Now

Price includes VAT (Netherlands)

Instant access to the full article PDF.

Fig. 1
The alternative text for this image may have been generated using AI.
Fig. 2
The alternative text for this image may have been generated using AI.
Fig. 3
The alternative text for this image may have been generated using AI.
Fig. 4
The alternative text for this image may have been generated using AI.
Fig. 5
The alternative text for this image may have been generated using AI.
Fig. 6
The alternative text for this image may have been generated using AI.
Fig. 7
The alternative text for this image may have been generated using AI.
Fig. 8
The alternative text for this image may have been generated using AI.
Fig. 9
The alternative text for this image may have been generated using AI.
Fig. 10
The alternative text for this image may have been generated using AI.
Fig. 11
The alternative text for this image may have been generated using AI.
Fig. 12
The alternative text for this image may have been generated using AI.
Fig. 13
The alternative text for this image may have been generated using AI.
Fig. 14
The alternative text for this image may have been generated using AI.
Fig. 15
The alternative text for this image may have been generated using AI.
Fig. 16
The alternative text for this image may have been generated using AI.
Fig. 17
The alternative text for this image may have been generated using AI.
Fig. 18
The alternative text for this image may have been generated using AI.

Similar content being viewed by others

Data availability

No datasets were generated or analysed during the current study.

References

  1. Wang, H., Ning, H., Lin, Y., Wang, W., Dhelim, S., et al.: A survey on the metaverse: the state-of-the-art, technologies, applications, and challenges. IEEE Internet Things J. 10(16), 14671–14688 (2023). https://doi.org/10.1109/JIOT.2023.3278329

    Article  Google Scholar 

  2. Thomas Alsop: Extended reality (XR) market size worldwide from 2021 to 2026. Retrieved Jan. 20, 2024 from https://www.statista.com/statistics/591181/global-augmented-virtual-reality-market-size/ (2023)

  3. CISCO: Cisco visual networking index: Global mobile data traffic forecast update, 2017–2022. Retrieved Nov. 20, 2023 from https://s3.amazonaws.com/media.mediapost.com/uploads/CiscoForecast.pdf (2019)

  4. Alt B., Ballard T., Steinmetz R., Koeppl H., Rizk A.: CBA: Contextual quality adaptation for adaptive bitrate video streaming. In: Proceedings of the IEEE International Conference on Computer Communications (INFOCOM’19), pp. 1000–1008 (2019) https://doi.org/10.1109/INFOCOM.2019.8737418

  5. Sodagar, I.: The MPEG-DASH standard for multimedia streaming over the internet. IEEE Multimed. 18(4), 62–67 (2011). https://doi.org/10.1109/MMUL.2011.71

    Article  Google Scholar 

  6. Guan Y., Zheng C., Zhang X., Guo Z., Jiang J.: Pano: optimizing 360° video streaming with a better understanding of quality perception. In: Proceedings of the ACM Special Interest Group on Data Communication (SIGCOMM’19), pp. 394–407 (2019) https://doi.org/10.1145/3341302.3342063

  7. Wang, H., Martinez-Velazquez, R., Dong, H., El Saddik, A.: Experimental studies of metaverse streaming. IEEE Consumer Electron. Mag. 14(1), 26–36 (2025). https://doi.org/10.1109/MCE.2024.3364118

    Article  Google Scholar 

  8. Qi, L., Xu, X., Wu, X., Ni, Q., Yuan, Y., Zhang, X.: Digital-twin-enabled 6G mobile network video streaming using mobile crowdsourcing. IEEE J. Sel. Areas Commun. Commun. 41(10), 3161–3174 (2023). https://doi.org/10.1109/JSAC.2023.3310077

    Article  Google Scholar 

  9. Xie L., Xu Z., Ban Y., Zhang X., Guo Z.: 360ProbDASH: improving QoE of 360 video streaming using tile-based HTTP adaptive streaming. In: Proceedings of the 25th ACM International Conference on Multimedia (MM’17), pp. 315–323 (2017) https://doi.org/10.1145/3123266.3123291

  10. Zou, J., Li, C., Liu, C., Yang, Q., Xiong, H., Steinbach, E.: Probabilistic tile visibility-based server-side rate adaptation for adaptive 360-degree video streaming. IEEE J. Sel. Top. Signal Process. 14(1), 161–176 (2020). https://doi.org/10.1109/JSTSP.2019.2956716

    Article  Google Scholar 

  11. Qian F., Ji L., Han B., Gopalakrishnan V.: Optimizing 360 video delivery over cellular networks. In: Proceedings of the 5th Workshop on All Things Cellular: Operations, Applications and Challenges (ATC’16), pp. 1–6 (2016) https://doi.org/10.1145/2980055.2980056

  12. Hou X., Dey S., Zhang J., Budagavi M.: Predictive view generation to enable mobile 360-degree and VR experiences. In: Proceedings of the 2018 Morning Workshop on Virtual Reality and Augmented Reality Network (VR/AR Network’18), pp. 20–26 (2018)https://doi.org/10.1145/3229625.3229629

  13. Nguyen, H.C., Dao, T.N., Pham, N.S., Dang, T.L., Nguyen, T.D., Truong, T.H.: An accurate viewport estimation method for 360 video streaming using deep learning. EAI Endorsed Trans. Indus. Netw. Intell. Syst. 9(4), e2 (2022). https://doi.org/10.4108/eetinis.v9i4.2218

    Article  Google Scholar 

  14. Zhang L., Xu W., Lu D., Cui L., Liu J.: MFVP: Mobile-friendly viewport prediction for live 360-degree video streaming. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME’22), pp. 1–6 (2022) https://doi.org/10.1109/ICME52920.2022.9859789

  15. Wu C., Zhang R., Wang Z., Sun L.: A spherical convolution approach for learning long term viewport prediction in 360 immersive video. In: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI’20), pp. 14003–14010 (2020)

  16. Wang M., Peng S., Chen X., Zhao Y., Xu M., Xu C.: CoLive: an edge-assisted online learning framework for viewport prediction in 360° live streaming. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME’22), pp. 1–6 (2022) https://doi.org/10.1109/ICME52920.2022.9859963

  17. Li, J., Han, L., Zhang, C., Li, Q., Liu, Z.: Spherical convolution empowered viewport prediction in 360 video multicast with limited fov feedback. ACM Trans. Multimed. Comput. Commun. Appl.Multimed. Comput. Commun. Appl. 19(1), 1–23 (2023). https://doi.org/10.1145/3511603

    Article  Google Scholar 

  18. Kan, N., Zou, J., Li, C., Dai, W., Xiong, H.: RAPT360: reinforcement learning-based rate adaptation for 360-degree video streaming with adaptive prediction and tiling. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1607–1623 (2022). https://doi.org/10.1109/TCSVT.2021.3076585

    Article  Google Scholar 

  19. Zhang Y., Zhao P., Bian K., Liu Y., Song L., Li X.: DRL360: 360-degree Video streaming with deep reinforcement learning. In: Proceedings of the IEEE International Conference on Computer Communications (INFOCOM’19), pp. 1252–1260 (2019) https://doi.org/10.1109/INFOCOM.2019.8737361

  20. Jiang, Z., Zhang, X., Xu, Y., Ma, Z., Sun, J., Zhang, Y.: Reinforcement learning based rate adaptation for 360-degree video streaming. IEEE Trans. Broadcast. 67(2), 409–423 (2021). https://doi.org/10.1109/TBC.2020.3028286

    Article  Google Scholar 

  21. Nathan V., Sivaraman V., Addanki R., et al.: End-to-end transport for video QoE fairness. In: Proceedings of the ACM Special Interest Group on Data Communication (SIGCOMM’19), pp. 408–423(2019) https://doi.org/10.1145/3341302.3342077

  22. Chiariotti, F.: A survey on 360-degree video: Coding, quality of experience and streaming. Comput. Commun. Commun. 177(1), 133–155 (2021). https://doi.org/10.1016/j.comcom.2021.06.029

    Article  Google Scholar 

  23. Wang, H., Dong, H., El Saddik, A.: Immersive multimedia communication: state-of-the-art on eXtended reality streaming. ACM Trans. Multimed. Comput. Commun. Appl.Multimed. Comput. Commun. Appl. (2025). https://doi.org/10.1145/3721292

    Article  Google Scholar 

  24. Lu Y., Zhu Y., Wang Z.: Personalized 360-degree video streaming: a meta-learning approach. In: Proceedings of the 30th ACM International Conference on Multimedia (MM’22), pp. 3143–3151 (2022) https://doi.org/10.1145/3503161.3548047

  25. Ban Y., Xie L., Xu Z., Zhang X., Guo Z., Wang Y.: CUB360: exploiting cross-users behaviors for viewport prediction in 360 video adaptive streaming. In: Proceedings of the IEEE international conference on multimedia and expo (ICME’18), pp. 1–6 (2018) https://doi.org/10.1109/ICME.2018.8486606

  26. Li, Z., Zhong, P., Huang, J., Gao, F., Wang, J.: Achieving QoE fairness in bitrate allocation of 360° video streaming. IEEE Trans. Multimedia 26, 1169–1178 (2024). https://doi.org/10.1109/TMM.2023.3277286

    Article  Google Scholar 

  27. Sreedhar K. K., Aminlou A., Hannuksela M. M., Gabbouj M.: Viewport-adaptive encoding and streaming of 360-degree video for virtual reality applications. In: Proceedings of the IEEE International Symposium on Multimedia (ISM’16), pp. 583–586 (2016) https://doi.org/10.1109/ISM.2016.0126

  28. Sánchez, D.L.F.Y., Bhullar, G.S., Skupin, R., Hellge, C., Schierl, T.: Delay impact on MPEG OMAF’s tile-based viewport-dependent 360° video streaming. IEEE J. Emerging Sel. Top. Circuits Syst. 9(1), 18–28 (2019). https://doi.org/10.1109/JETCAS.2019.2899516

    Article  Google Scholar 

  29. Bao Y., Wu H., Zhang T., Ramli A. A., Liu X.: Shooting a moving target: motion-prediction-based transmission for 360-degree videos. In: Proceedings of the IEEE International Conference on Big Data (Big Data’16), pp. 1161–1170 (2016) https://doi.org/10.1109/BigData.2016.7840720

  30. Yang S., He Y., Zheng X.: FoVR: Attention-based VR streaming through bandwidth-limited wireless networks. In: Proceedings of the Annual IEEE International Conference on Sensing, Communication, and Networking (SECON’16), pp. 1–9 (2016) https://doi.org/10.1109/SAHCN.2019.8824804

  31. Qian F., Han B., Xiao Q., Gopalakrishnan V.: Flare: Practical viewport-adaptive 360-degree video streaming for mobile devices. In: Proceedings of the 24th Annual International Conference on Mobile Computing and Networking (MobiCom’18), pp. 99–114 (2018) https://doi.org/10.1145/3241539.3241565

  32. Yadav P. K., Ooi W. T.: Tile rate allocation for 360-degree tiled adaptive video streaming. In: Proceedings of the 28th ACM International Conference on Multimedia (MM’20), pp. 3724–3733 (2020) https://doi.org/10.1145/3394171.3413550

  33. Wang, H., Long, Z., Dong, H., El Saddik, A.: MADRL-based rate adaptation for 360° video streaming with multi-viewpoint prediction. IEEE Internet Things J. 11(15), 26503–26517 (2024). https://doi.org/10.1109/JIOT.2024.3398548

    Article  Google Scholar 

  34. Liu C., Kan N., Zou J., Yang Q., Xiong H.: Server-side rate adaptation for multi-user 360-degree video streaming. In: Proceedings of the 25th IEEE International Conference on Image Processing (ICIP’18), pp. 3264–3268 (2018) https://doi.org/10.1109/ICIP.2018.8451447

  35. Ban Y., Zhang Y., Zhang H., Zhang X., Guo Z.: MA360: Multi-agent deep reinforcement learning based live 360-degree video streaming on edge. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME’20), pp. 1–6 (2020) https://doi.org/10.1109/ICME46284.2020.9102836

  36. Wu C., Tan Z., Wang Z., Yang S.: A dataset for exploring user behaviors in VR spherical video streaming. In: Proceedings of the 8th ACM on Multimedia Systems Conference (MMSys’17), pp. 193–198 (2017)https://doi.org/10.1145/3083187.3083210

  37. GPAC. Ultramedia open source infrastructure for video streaming and next-gen multimedia transcoding, packaging and delivery. Retrieved Mar. 28, 2024 from https://gpac.io/

  38. Roy A., Roy K. K., Ali A. A., Amin M. A., Rahman A. K. M. M.: Spatio-temporal modeling for traffic forecasting using graph neural network. In: Proceedings of the International Joint Conference on Neural Networks (IJCNN’21), pp. 1–8 (2021) https://doi.org/10.1109/IJCNN52387.2021.9533319

  39. Ackermann J., Gabler V., Osa T., Sugiyama M.: Reducing overestimation bias in multi-agent domains using double centralized critics. arXiv, arXiv:1910.01465 [cs.LG] (2019). https://doi.org/10.48550/arXiv.1910.01465

  40. Viitanen M., Koivula A., Lemmetti A., Ylä-Outinen A., Vanne J., Hämäläinen T. D.: Kvazaar: Open-source HEVC/H.265 encoder. In: Proceedings of the 24th ACM international conference on Multimedia (MM’16), pp. 1179–1182 (2016) https://doi.org/10.1145/2964284.2973796

  41. Yaqoob, A., Muntean, G.: Advanced predictive tile selection using dynamic tiling for prioritized 360° video VR streaming. ACM Trans. Multimed. Comput. Commun. Appl.Multimed. Comput. Commun. Appl. 20(1), 1–28 (2024). https://doi.org/10.1145/3603146

    Article  Google Scholar 

  42. Park, S.K., Bhattacharya, A., Yang, Z., Das, S.R., Samaras, D.: Mosaic: advancing user quality of experience in 360-degree video streaming with machine learning. IEEE Trans. Netw. Serv. Manage.Netw. Serv. Manage. 18(1), 1000–1015 (2021). https://doi.org/10.1109/TNSM.2021.3053183

    Article  Google Scholar 

  43. Hooft, J.V.D., Petrangeli, S., Wauters, T., Huysegems, R., Alface, P.R., Bostoen, T., Turck, F.D.: HTTP/2-based adaptive streaming of HEVC video over 4G/LTE networks. IEEE Commun. Lett.Commun. Lett. 20(11), 2177–2180 (2016). https://doi.org/10.1109/LCOMM.2016.2601087

    Article  Google Scholar 

  44. Ye, Y., Tang, Y., Wang, H., Zhang, X., Strbac, G.: A scalable privacy-preserving multi-agent deep reinforcement learning approach for large-scale peer-to-peer transactive energy trading. IEEE Trans. Smart Grid 12(6), 5185–5200 (2021). https://doi.org/10.1109/TSG.2021.3103917

    Article  Google Scholar 

  45. Lowe R., Wu Y., Tamar A., Harb J., Abbeel P., Mordatch I.: Multi-agent actor-critic for mixed cooperative-competitive environments. In: Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), pp. 6379–6390 (2017) https://doi.org/10.48550/arXiv.1706.02275

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No: 61861013), Open Project Program of Guangxi Key Laboratory of Digital Infrastructure (Grant No: GXDIOP2024008), and Science and Technology Major Project of Guangxi, China (Grant No: AA18118031).

Author information

Authors and Affiliations

Authors

Contributions

Methodology-L.S.J.;Wrote the main manuscript text-L.S.J.,L.Y.M.; Reviewed the manuscript-L.S.J.,W.Y.,L.Y.M.,L.S.

Corresponding author

Correspondence to Yong Wang.

Ethics declarations

Conflict of interest

The authors declare no competing interests.

Additional information

Communicated by Qiu Shen.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, S., Wang, Y., Li, S. et al. MADRL-based bitrate allocation for QoE fairness in 360° video streaming with viewport prediction. Multimedia Systems 31, 343 (2025). https://doi.org/10.1007/s00530-025-01920-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Version of record:

  • DOI: https://doi.org/10.1007/s00530-025-01920-4

Keywords

Profiles

  1. Yong Wang
  2. Yuming Liu