{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T03:50:46Z","timestamp":1776743446426,"version":"3.51.2"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:00:00Z","timestamp":1750291200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:00:00Z","timestamp":1750291200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62271344"],"award-info":[{"award-number":["62271344"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,9]]},"DOI":"10.1007\/s11760-025-04330-1","type":"journal-article","created":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T16:54:40Z","timestamp":1750352080000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Phase aware speech enhancement based on depthwise deep convolutional network"],"prefix":"10.1007","volume":"19","author":[{"given":"Yasir","family":"Iqbal","sequence":"first","affiliation":[]},{"given":"Tao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Anjum","family":"Iqbal","sequence":"additional","affiliation":[]},{"given":"Ikram","family":"Azaz","sequence":"additional","affiliation":[]},{"given":"Umar","family":"Sadique","sequence":"additional","affiliation":[]},{"given":"Muhammad","family":"Fahad","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Yanzhang","family":"Geng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,19]]},"reference":[{"key":"4330_CR1","doi-asserted-by":"crossref","unstructured":"Kim, E., Seo, H.: SE-Conformer: Time-Domain Speech Enhancement Using Conformer Interspeech, pp. 2736\u20132740. (2021)","DOI":"10.21437\/Interspeech.2021-2207"},{"key":"4330_CR2","doi-asserted-by":"publisher","first-page":"380","DOI":"10.1109\/TASLP.2019.2955276","volume":"28","author":"K Tan","year":"2019","unstructured":"Tan, K., Wang, D.: Learning complex spectral mapping with gated convolutional recurrent networks for monaural speech enhancement. IEEE\/ACM Trans. Audio Speech Lang. Process. 28, 380\u2013390 (2019)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"4330_CR3","doi-asserted-by":"crossref","unstructured":"Boll, S.: Suppression of acoustic noise in speech using spectral subtraction, IEEE Transactions on acoustics, speech, and signal processing, vol. 27, no. 2, pp. 113\u2013120, (1979)","DOI":"10.1109\/TASSP.1979.1163209"},{"issue":"3","key":"4330_CR4","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1109\/TASSP.1978.1163086","volume":"26","author":"J Lim","year":"1978","unstructured":"Lim, J., Oppenheim, A.: All-pole modeling of degraded speech. IEEE Trans. Acoust. Speech Signal Process. 26(3), 197\u2013210 (1978)","journal-title":"IEEE Trans. Acoust. Speech Signal Process."},{"key":"4330_CR5","doi-asserted-by":"crossref","unstructured":"Ephraim, Y., Malah, D.: Speech enhancement using a minimum mean-square error log-spectral amplitude estimator, IEEE transactions on acoustics, speech, and signal processing, vol. 33, no. 2, pp. 443\u2013445, (1985)","DOI":"10.1109\/TASSP.1985.1164550"},{"key":"4330_CR6","doi-asserted-by":"crossref","unstructured":"Hao, X., et al.: Masking and inpainting: A two-stage speech enhancement approach for low SNR and non-stationary noise, in ICASSP 2020\u20132020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),: IEEE, pp. 6959\u20136963. (2020)","DOI":"10.1109\/ICASSP40776.2020.9053188"},{"issue":"01","key":"4330_CR7","doi-asserted-by":"publisher","first-page":"2550001","DOI":"10.1142\/s0219467825500019","volume":"25","author":"C Jannu","year":"2025","unstructured":"Jannu, C., Vanambathina, S.D.: An overview of speech enhancement based on deep learning techniques. Int. J. Image Graphics. 25(01), 2550001 (2025). https:\/\/doi.org\/10.1142\/s0219467825500019","journal-title":"Int. J. Image Graphics"},{"key":"4330_CR8","doi-asserted-by":"crossref","unstructured":"Su, J., Jin, Z., Finkelstein, A.: HiFi-GAN: High-fidelity denoising and dereverberation based on speech deep features in adversarial networks, arXiv preprint arXiv:.05694, 2020. (2006)","DOI":"10.21437\/Interspeech.2020-2143"},{"key":"4330_CR9","doi-asserted-by":"publisher","first-page":"123","DOI":"10.3233\/IDT-230211","volume":"18","author":"RR Rai","year":"2024","unstructured":"Rai, R.R., Mathivanan, M.: Recalling-Enhanced recurrent neural network optimized with chimp optimization algorithm based speech enhancement for hearing aids. Intell. Decis. Technol. 18, 123\u2013134 (2024). https:\/\/doi.org\/10.3233\/IDT-230211","journal-title":"Intell. Decis. Technol."},{"key":"4330_CR10","unstructured":"Oostermeijer, K., Wang, Q., Du, J.: Frequency gating: Improved convolutional neural networks for speech enhancement in the time-frequency domain, in 2020 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC),: IEEE, pp. 465\u2013470. (2020)"},{"key":"4330_CR11","doi-asserted-by":"crossref","unstructured":"Girirajan, S., Pandian, A.: Real-Time speech enhancement based on convolutional recurrent neural network. Intell. Autom. Soft Comput., 35, 2, (2023)","DOI":"10.32604\/iasc.2023.028090"},{"key":"4330_CR12","doi-asserted-by":"publisher","first-page":"104408","DOI":"10.1016\/j.dsp.2024.104408","volume":"147","author":"N Saleem","year":"2024","unstructured":"Saleem, N., Gunawan, T.S., Dhahbi, S., Bourouis, S.: Time domain speech enhancement with CNN and time-attention transformer. Digit. Signal Proc. 147, 104408 (2024)","journal-title":"Digit. Signal Proc."},{"issue":"3","key":"4330_CR13","doi-asserted-by":"publisher","first-page":"1959","DOI":"10.1007\/s11277-021-08313-6","volume":"119","author":"A Karthik","year":"2021","unstructured":"Karthik, A., MazherIqbal, J.: Efficient speech enhancement using recurrent Convolution encoder and decoder. Wireless Pers. Commun. 119(3), 1959\u20131973 (2021)","journal-title":"Wireless Pers. Commun."},{"key":"4330_CR14","doi-asserted-by":"publisher","first-page":"109592","DOI":"10.1016\/j.apacoust.2023.109592","volume":"212","author":"S Sivapatham","year":"2023","unstructured":"Sivapatham, S., Kar, A., Bodile, R., Mladenovic, V., Sooraksa, P.: A deep neural network-correlation phase sensitive mask based Estimation to improve speech intelligibility. Appl. Acoust. 212, 109592 (2023)","journal-title":"Appl. Acoust."},{"key":"4330_CR15","doi-asserted-by":"crossref","unstructured":"Alameri, B.M., Kadhim, I.J., Hadi, S.Q., Hassoon, A.F., Abd, M.M., Premaratne, P.: Convolutional deep neural network and full connectivity for speech enhancement. Int. J. Online Biomedical Eng., 19, 4, (2023)","DOI":"10.3991\/ijoe.v19i04.37577"},{"key":"4330_CR16","doi-asserted-by":"publisher","unstructured":"Paliwal, K., W\u00f3jcicki, K., Shannon, B.: The importance of phase in speech enhancement, Speech Communication, vol. 53, no. 4, pp. 465\u2013494, \/04\/01\/ 2011, (2011). https:\/\/doi.org\/10.1016\/j.specom.2010.12.003","DOI":"10.1016\/j.specom.2010.12.003"},{"key":"4330_CR17","doi-asserted-by":"crossref","unstructured":"Hu, Y., et al.: DCCRN: Deep complex convolution recurrent network for phase-aware speech enhancement, arXiv preprint arXiv:2008.00264, (2020)","DOI":"10.21437\/Interspeech.2020-2537"},{"key":"4330_CR18","doi-asserted-by":"crossref","unstructured":"Iqbal, Y., Zhang, T., Fahad, M., Iqbal, A., Geng, Y., Zhao, X.: Speech enhancement using deep complex convolutional neural network (DCCNN) model, Signal, Image and Video Processing, pp. 1\u201318, (2024)","DOI":"10.1007\/s11760-024-03500-x"},{"key":"4330_CR19","doi-asserted-by":"crossref","unstructured":"Yin, D., Luo, C., Xiong, Z., Zeng, W.: Phasen: A phase-and-harmonics-aware speech enhancement network, in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 05, pp. 9458\u20139465. (2020)","DOI":"10.1609\/aaai.v34i05.6489"},{"key":"4330_CR20","doi-asserted-by":"crossref","unstructured":"Nustede, E.J., Anem\u00fcller, J.: Single-channel speech enhancement with deep complex u-networks and probabilistic latent space models, in ICASSP 2023\u20132023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),: IEEE, pp. 1\u20135. (2023)","DOI":"10.1109\/ICASSP49357.2023.10096208"},{"key":"4330_CR21","unstructured":"Hasannezhad, M., Ouyang, Z., Zhu, W.-P., Champagne, B.: An integrated CNN-GRU framework for complex ratio mask estimation in speech enhancement, in 2020 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC),: IEEE, pp. 764\u2013768. (2020)"},{"key":"4330_CR22","unstructured":"Zhou, L., Gao, Y., Wang, Z., Li, J., Zhang, W.: Complex spectral mapping with attention based convolution recurrent neural network for speech enhancement, arXiv preprint arXiv:2104.05267, (2021)"},{"key":"4330_CR23","doi-asserted-by":"crossref","unstructured":"Tan, K., Wang, D.: Complex spectral mapping with a convolutional recurrent network for monaural speech enhancement, in ICASSP 2019\u20132019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),: IEEE, pp. 6865\u20136869. (2019)","DOI":"10.1109\/ICASSP.2019.8682834"},{"key":"4330_CR24","doi-asserted-by":"publisher","first-page":"1829","DOI":"10.1109\/TASLP.2021.3079813","volume":"29","author":"A Li","year":"2021","unstructured":"Li, A., Liu, W., Zheng, C., Fan, C., Li, X.: Two heads are better than one: A Two-Stage complex spectral mapping approach for monaural speech enhancement. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 1829\u20131843 (2021). https:\/\/doi.org\/10.1109\/TASLP.2021.3079813","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"4330_CR25","doi-asserted-by":"publisher","first-page":"103008","DOI":"10.1016\/j.specom.2023.103008","volume":"156","author":"FE Wahab","year":"2024","unstructured":"Wahab, F.E., Ye, Z., Saleem, N., Ullah, R.: Compact deep neural networks for real-time speech enhancement on resource-limited devices. Speech Commun. 156, 103008 (2024)","journal-title":"Speech Commun."},{"key":"4330_CR26","doi-asserted-by":"crossref","unstructured":"Jannu, C., Vanambathina, S.D.: An attention based densely connected u-net with convolutional gru for speech enhancement, in 3rd international conference on artificial intelligence and signal processing (AISP), 2023: IEEE, pp. 1\u20135. (2023)","DOI":"10.1109\/AISP57993.2023.10134933"},{"issue":"12","key":"4330_CR27","doi-asserted-by":"publisher","first-page":"7467","DOI":"10.1007\/s00034-023-02455-7","volume":"42","author":"C Jannu","year":"2023","unstructured":"Jannu, C., Vanambathina, S.D.: Multi-stage progressive learning-based speech enhancement using time\u2013frequency attentive squeezed Temporal convolutional networks. Circuits Syst. Signal. Process. 42(12), 7467\u20137493 (2023)","journal-title":"Circuits Syst. Signal. Process."},{"issue":"1","key":"4330_CR28","first-page":"1195","volume":"45","author":"C Jannu","year":"2023","unstructured":"Jannu, C., Vanambathina, S.D.: DCT based densely connected convolutional GRU for real-time speech enhancement. J. Intell. Fuzzy Syst. 45(1), 1195\u20131208 (2023)","journal-title":"J. Intell. Fuzzy Syst."},{"issue":"4","key":"4330_CR29","first-page":"10907","volume":"46","author":"V Parisae","year":"2024","unstructured":"Parisae, V., Nagakishore Bhavanam, S.: Multi scale encoder-decoder network with time frequency attention and s-tcn for single channel speech enhancement. J. Intell. Fuzzy Syst. 46(4), 10907\u201310907 (2024)","journal-title":"J. Intell. Fuzzy Syst."},{"key":"4330_CR30","doi-asserted-by":"crossref","unstructured":"Parisae, V., Nagakishore Bhavanam, S.: Stacked u-net with time\u2013frequency attention and deep connection net for single channel speech enhancement. Int. J. Image Graphics, p. 2550067, (2024)","DOI":"10.1142\/S0219467825500676"},{"key":"4330_CR31","doi-asserted-by":"publisher","unstructured":"Jang, J.-G., Quan, C., Lee, H.D., Kang, U.: Falcon: Lightweight and accurate Convolution based on depthwise separable Convolution. Knowl. Inf. Syst., 65, 5, pp. 2225\u20132249, 2023\/05\/01 2023, https:\/\/doi.org\/10.1007\/s10115-022-01818-x","DOI":"10.1007\/s10115-022-01818-x"},{"key":"4330_CR32","doi-asserted-by":"crossref","unstructured":"Xu, X., Hao, J.: U-former: Improving monaural speech enhancement with multi-head self and cross attention, in 26th International Conference on Pattern Recognition (ICPR), 2022: IEEE, pp. 663\u2013669. (2022)","DOI":"10.1109\/ICPR56361.2022.9956638"},{"key":"4330_CR33","doi-asserted-by":"crossref","unstructured":"Valentini-Botinhao, C., Wang, X., Takaki, S., Yamagishi, J.: Investigating RNN-based speech enhancement methods for noise-robust Text-to-Speech, in SSW, pp. 146\u2013152. (2016)","DOI":"10.21437\/SSW.2016-24"},{"key":"4330_CR34","doi-asserted-by":"crossref","unstructured":"Veaux, C., Yamagishi, J., King, S.: The voice bank corpus: Design, collection and data analysis of a large regional accent speech database, in international conference oriental COCOSDA held jointly with 2013 conference on Asian spoken language research and evaluation (O-COCOSDA\/CASLRE), 2013: IEEE, pp. 1\u20134. (2013)","DOI":"10.1109\/ICSDA.2013.6709856"},{"key":"4330_CR35","doi-asserted-by":"crossref","unstructured":"Thiemann, J., Ito, N., Vincent, E.: The diverse environments multi-channel acoustic noise database (demand): A database of multichannel environmental noise recordings, in Proceedings of Meetings on Acoustics, vol. 19, no. 1: AIP Publishing. (2013)","DOI":"10.1121\/1.4799597"},{"key":"4330_CR36","doi-asserted-by":"crossref","unstructured":"Paul, D.B., Baker, J.: The design for the Wall Street Journal-based CSR corpus, in Speech and Natural Language: Proceedings of a Workshop Held at Harriman, New York, February 23\u201326, 1992. (1992)","DOI":"10.3115\/1075527.1075614"},{"key":"4330_CR37","doi-asserted-by":"crossref","unstructured":"Reddy, C.K., et al.: Interspeech 2021 deep noise suppression challenge, arXiv preprint arXiv:2101.01902, (2021)","DOI":"10.21437\/Interspeech.2021-1609"},{"key":"4330_CR38","doi-asserted-by":"crossref","unstructured":"Varga, A., Steeneken, H.J.: Assessment for automatic speech recognition: II. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition systems, Speech communication, vol. 12, no. 3, pp. 247\u2013251, (1993)","DOI":"10.1016\/0167-6393(93)90095-3"},{"key":"4330_CR39","doi-asserted-by":"publisher","first-page":"109385","DOI":"10.1016\/j.apacoust.2023.109385","volume":"209","author":"H Guo","year":"2023","unstructured":"Guo, H., et al.: MAMGAN: Multiscale attention metric GAN for monaural speech enhancement in the time domain. Appl. Acoust. 209, 109385 (2023)","journal-title":"Appl. Acoust."},{"key":"4330_CR40","doi-asserted-by":"crossref","unstructured":"Park, H.J., Kang, B.H., Shin, W., Kim, J.S., Han, S.W.: Manner: Multi-view attention network for noise erasure, in ICASSP 2022\u2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022: IEEE, pp. 7842\u20137846. (2022)","DOI":"10.1109\/ICASSP43922.2022.9747120"},{"key":"4330_CR41","doi-asserted-by":"crossref","unstructured":"Lin, Z., Wang, J., Li, R., Shen, F., Xuan, X.: PrimeK-Net: Multi-scale Spectral Learning via Group Prime-Kernel Convolutional Neural Networks for Single Channel Speech Enhancement, in ICASSP\u20132025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2025: IEEE, pp. 1\u20135. (2025)","DOI":"10.1109\/ICASSP49660.2025.10890034"},{"key":"4330_CR42","doi-asserted-by":"crossref","unstructured":"Zhou, H., Zhou, Y., Cheng, Z., Zhao, Y., Liu, Y.: Improved Encoder-Decoder architecture with Human-like perception attention for monaural speech enhancement. IEEE. Signal. Process. Lett., (2025)","DOI":"10.1109\/LSP.2025.3558690"},{"key":"4330_CR43","doi-asserted-by":"publisher","first-page":"233121652312099","DOI":"10.1177\/23312165231209913","volume":"27","author":"C Zheng","year":"2023","unstructured":"Zheng, C., et al.: Sixty years of frequency-domain monaural speech enhancement: From traditional to deep learning methods. Trends Hear. 27, 23312165231209913 (2023)","journal-title":"Trends Hear."},{"key":"4330_CR44","doi-asserted-by":"crossref","unstructured":"Hao, X., Su, X., Horaud, R., Li, X.: Fullsubnet: A full-band and sub-band fusion model for real-time single-channel speech enhancement, in ICASSP\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021: IEEE, pp. 6633\u20136637. (2021)","DOI":"10.1109\/ICASSP39728.2021.9414177"},{"key":"4330_CR45","doi-asserted-by":"crossref","unstructured":"Li, A., Liu, W., Luo, X., Zheng, C., Li, X.: ICASSP 2021 deep noise suppression challenge: Decoupling magnitude and phase optimization with a two-stage deep network, in ICASSP\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021: IEEE, pp. 6628\u20136632. (2021)","DOI":"10.1109\/ICASSP39728.2021.9414062"},{"key":"4330_CR46","doi-asserted-by":"crossref","unstructured":"Li, A., You, S., Yu, G., Zheng, C., Li, X.: Taylor, can you hear me now? a Taylor-unfolding framework for monaural speech enhancement, arXiv preprint arXiv:2205.00206, (2022)","DOI":"10.24963\/ijcai.2022\/582"},{"key":"4330_CR47","doi-asserted-by":"crossref","unstructured":"Reddy, C.K., Gopal, V., Cutler, R.: DNSMOS: A non-intrusive perceptual objective speech quality metric to evaluate noise suppressors, in ICASSP 2021\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP),: IEEE, pp. 6493\u20136497. (2021)","DOI":"10.1109\/ICASSP39728.2021.9414878"},{"issue":"3","key":"4330_CR48","doi-asserted-by":"publisher","first-page":"1627","DOI":"10.1121\/1.5055562","volume":"144","author":"Y Zhao","year":"2018","unstructured":"Zhao, Y., Wang, D., Johnson, E.M., Healy, E.W.: A deep learning based segregation algorithm to increase speech intelligibility for hearing-impaired listeners in reverberant-noisy conditions. J. Acoust. Soc. Am. 144(3), 1627\u20131637 (2018)","journal-title":"J. Acoust. Soc. Am."},{"key":"4330_CR49","doi-asserted-by":"crossref","unstructured":"Wang, H., Tian, B.: ZipEnhancer: Dual-Path Down-Up Sampling-based Zipformer for Monaural Speech Enhancement, arXiv preprint arXiv:2501.05183, (2025)","DOI":"10.1109\/ICASSP49660.2025.10888703"},{"key":"4330_CR50","doi-asserted-by":"crossref","unstructured":"Lu, Y.-X., Ai, Y., Ling, Z.-H.: Explicit Estimation of magnitude and phase spectra in parallel for high-quality speech enhancement. Neural Netw., p. 107562, (2025)","DOI":"10.1016\/j.neunet.2025.107562"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04330-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04330-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04330-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T14:45:58Z","timestamp":1751553958000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04330-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,19]]},"references-count":50,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2025,9]]}},"alternative-id":["4330"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04330-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,19]]},"assertion":[{"value":"20 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 May 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 June 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 June 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"757"}}