{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,15]],"date-time":"2025-03-15T05:10:29Z","timestamp":1742015429860,"version":"3.38.0"},"reference-count":50,"publisher":"Association for Natural Language Processing","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Journal of Natural Language Processing"],"published-print":{"date-parts":[[2025]]},"DOI":"10.5715\/jnlp.32.252","type":"journal-article","created":{"date-parts":[[2025,3,14]],"date-time":"2025-03-14T22:19:27Z","timestamp":1741990767000},"page":"252-282","source":"Crossref","is-referenced-by-count":0,"title":["DiLM: Distilling Dataset into Language Model for Text-level Dataset Distillation"],"prefix":"10.5715","volume":"32","author":[{"given":"Aru","family":"Maekawa","sequence":"first","affiliation":[{"name":"Institute of Science Tokyo"}]},{"given":"Satoshi","family":"Kosugi","sequence":"additional","affiliation":[{"name":"Institute of Science Tokyo"}]},{"given":"Kotaro","family":"Funakoshi","sequence":"additional","affiliation":[{"name":"Institute of Science Tokyo"}]},{"given":"Manabu","family":"Okumura","sequence":"additional","affiliation":[{"name":"Institute of Science Tokyo"}]}],"member":"3685","reference":[{"key":"1","unstructured":"Aljundi, R., Lin, M., Goujaud, B., and Bengio, Y. (2019). \u201cGradient Based Sample Selection for Online Continual Learning.\u201d In Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019Alch\u00e9-Buc, F., Fox, E., and Garnett, R. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 32, pp. 11816\u201311825. Curran Associates, Inc."},{"key":"2","unstructured":"Bohdal, O., Yang, Y., and Hospedales, T. M. (2020). \u201cFlexible Dataset Distillation: Learn Labels Instead of Images.\u201d <i>CoRR<\/i>, abs\/2006.08572."},{"key":"3","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., McCandlish, S., Radford, A., Sutskever, I., and Amodei, D. (2020). \u201cLanguage Models are Few-Shot Learners.\u201d In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 33, pp. 1877\u20131901. Curran Associates, Inc."},{"key":"4","doi-asserted-by":"crossref","unstructured":"Cazenavette, G., Wang, T., Torralba, A., Efros, A. A., and Zhu, J.-Y. (2022). \u201cDataset Distillation by Matching Training Trajectories.\u201d In <i>IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2022, New Orleans, LA, USA, June 19\u201320, 2022<\/i>, pp. 4749\u20134758. IEEE.","DOI":"10.1109\/CVPRW56347.2022.00521"},{"key":"5","doi-asserted-by":"crossref","unstructured":"Cazenavette, G., Wang, T., Torralba, A., Efros, A. A., and Zhu, J.-Y. (2023). \u201cGeneralizing Dataset Distillation via Deep Generative Prior.\u201d In <i>IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17\u201324, 2023<\/i>, pp. 3739\u20133748. IEEE.","DOI":"10.1109\/CVPR52729.2023.00364"},{"key":"6","unstructured":"Chen, D., Kerkouche, R., and Fritz, M. (2022). \u201cPrivate Set Generation with Discriminative Information.\u201d In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 35, pp. 14678\u201314690. Curran Associates, Inc."},{"key":"7","unstructured":"Chen, Y., Welling, M., and Smola, A. (2010). \u201cSuper-Samples from Kernel Herding.\u201d In <i>Proceedings of the 36th Conference on Uncertainty in Artificial Intelligence<\/i>, UAI\u201910, pp. 109\u2013116, Arlington, Virginia, USA. AUAI Press."},{"key":"8","unstructured":"Cui, J., Wang, R., Si, S., and Hsieh, C.-J. (2023). \u201cScaling Up Dataset Distillation to ImageNet-1K with Constant Memory.\u201d In Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., and Scarlett, J. (Eds.), <i>Proceedings of the 40th International Conference on Machine Learning<\/i>, Vol. 202 of <i>Proceedings of Machine Learning Research<\/i>, pp. 6565\u20136590. PMLR."},{"key":"9","unstructured":"Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). \u201cBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\u201d In <i>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)<\/i>, pp. 4171\u20134186, Minneapolis, Minnesota. Association for Computational Linguistics."},{"key":"10","unstructured":"Dong, T., Zhao, B., and Lyu, L. (2022). \u201cPrivacy for Free: How does Dataset Condensation Help Privacy?\u201d In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (Eds.), <i>Proceedings of the 39th International Conference on Machine Learning<\/i>, Vol. 162 of <i>Proceedings of Machine Learning Research<\/i>, pp. 5378\u20135396. PMLR."},{"key":"11","doi-asserted-by":"crossref","unstructured":"Geng, J., Chen, Z., Wang, Y., Woisetschlaeger, H., Schimmler, S., Mayer, R., Zhao, Z., and Rong, C. (2023). \u201cA Survey on Dataset Distillation: Approaches, Applications and Future Directions.\u201d In Elkind, E. (Ed.), <i>Proceedings of the 32nd International Joint Conference on Artificial Intelligence, IJCAI-23<\/i>, pp. 6610\u20136618. International Joint Conferences on Artificial Intelligence Organization. Survey Track.","DOI":"10.24963\/ijcai.2023\/741"},{"key":"12","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. (2014). \u201cGenerative Adversarial Nets.\u201d In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 27. Curran Associates, Inc."},{"key":"13","doi-asserted-by":"crossref","unstructured":"Hiraoka, T., Takase, S., Uchiumi, K., Keyaki, A., and Okazaki, N. (2020). \u201cOptimizing Word Segmentation for Downstream Task.\u201d In <i>Findings of the Association for Computational Linguistics: EMNLP 2020<\/i>, pp. 1341\u20131351, Online. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2020.findings-emnlp.120"},{"key":"14","unstructured":"Li, Y. and Li, W. (2021). \u201cData Distillation for Text Classification.\u201d <i>CoRR<\/i>, abs\/2104.08448."},{"key":"15","doi-asserted-by":"crossref","unstructured":"Liu, Y., Gu, J., Wang, K., Zhu, Z., Jiang, W., and You, Y. (2023). \u201cDREAM: Efficient Dataset Distillation by Representative Matching.\u201d In <i>Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)<\/i>, pp. 17314\u201317324.","DOI":"10.1109\/ICCV51070.2023.01588"},{"key":"16","unstructured":"Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov, V. (2019). \u201cRoBERTa: A Robustly Optimized BERT Pretraining Approach.\u201d <i>CoRR<\/i>, abs\/1907.11692."},{"key":"17","unstructured":"Loshchilov, I. and Hutter, F. (2019). \u201cDecoupled Weight Decay Regularization.\u201d In <i>7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6\u20139, 2019<\/i>. OpenReview.net."},{"key":"18","unstructured":"Maclaurin, D., Duvenaud, D., and Adams, R. P. (2015). \u201cGradient-based Hyperparameter Optimization through Reversible Learning.\u201d In Bach, F. R. and Blei, D. M. (Eds.), <i>Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6\u201311 July 2015<\/i>, Vol. 37 of <i>JMLR Workshop and Conference Proceedings<\/i>, pp. 2113\u20132122. JMLR.org."},{"key":"19","doi-asserted-by":"crossref","unstructured":"Maekawa, A., Kobayashi, N., Funakoshi, K., and Okumura, M. (2023). \u201cDataset Distillation with Attention Labels for Fine-tuning BERT.\u201d In <i>Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)<\/i>, pp. 119\u2013127, Toronto, Canada. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2023.acl-short.12"},{"key":"20","doi-asserted-by":"crossref","unstructured":"Maekawa, A., Kosugi, S., Funakoshi, K., and Okumura, M. (2024). \u201cDiLM: Distilling Dataset into Language Model for Text-level Dataset Distillation.\u201d In Duh, K., Gomez, H., and Bethard, S. (Eds.), <i>Findings of the Association for Computational Linguistics: NAACL 2024<\/i>, pp. 3138\u20133153, Mexico City, Mexico. Association for Computational Linguistics.","DOI":"10.18653\/v1\/2024.findings-naacl.199"},{"key":"21","unstructured":"Medvedev, D. and D\u2019yakonov, A. (2021). \u201cLearning to Generate Synthetic Training Data Using Gradient Matching and Implicit Differentiation.\u201d In Burnaev, E., Ignatov, D. I., Ivanov, S., Khachay, M. Y., Koltsova, O., Kutuzov, A., Kuznetsov, S. O., Loukachevitch, N. V., Napoli, A., Panchenko, A., Pardalos, P. M., Saram\u00e4ki, J., Savchenko, A. V., Tsymbalov, E., and Tutubalina, E. (Eds.), <i>Recent Trends in Analysis of Images, Social Networks and Texts - 10th International Conference, AIST 2021, Tbilisi, Georgia, December 16\u201318, 2021, Revised Supplementary Proceedings<\/i>, Vol. 1573 of <i>Communications in Computer and Information Science<\/i>, pp. 138\u2013150. Springer."},{"key":"22","unstructured":"Mirzasoleiman, B., Bilmes, J., and Leskovec, J. (2020). \u201cCoresets for Data-efficient Training of Machine Learning Models.\u201d In Daum\u00e9 III, H. and Singh, A. (Eds.), <i>Proceedings of the 37th International Conference on Machine Learning<\/i>, Vol. 119 of <i>Proceedings of Machine Learning Research<\/i>, pp. 6950\u20136960. PMLR."},{"key":"23","unstructured":"Moore, R. C. and Lewis, W. (2010). \u201cIntelligent Selection of Language Model Training Data.\u201d In <i>Proceedings of the ACL 2010 Conference Short Papers<\/i>, pp. 220\u2013224, Uppsala, Sweden. Association for Computational Linguistics."},{"key":"24","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., and Manning, C. (2014). \u201cGloVe: Global Vectors for Word Representation.\u201d In <i>Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)<\/i>, pp. 1532\u20131543, Doha, Qatar. Association for Computational Linguistics.","DOI":"10.3115\/v1\/D14-1162"},{"key":"25","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., and Sutskever, I. (2019). \u201cLanguage Models are Unsupervised Multitask Learners.\u201d <i>OpenAI blog<\/i>, 1 (8), p. 9."},{"key":"26","unstructured":"Sahni, S. and Patel, H. M. (2023). \u201cExploring Multilingual Text Data Distillation.\u201d <i>CoRR<\/i>, abs\/2308.04982."},{"key":"27","doi-asserted-by":"crossref","unstructured":"Sangermano, M., Carta, A., Cossu, A., and Bacciu, D. (2022). \u201cSample Condensation in Online Continual Learning.\u201d In <i>International Joint Conference on Neural Networks, IJCNN 2022, Padua, Italy, July 18\u201323, 2022<\/i>, pp. 1\u20138. IEEE.","DOI":"10.1109\/IJCNN55064.2022.9892299"},{"key":"28","unstructured":"Sener, O. and Savarese, S. (2018). \u201cActive Learning for Convolutional Neural Networks: A Core-Set Approach.\u201d In <i>6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30\u2013May 3, 2018, Conference Track Proceedings<\/i>. OpenReview.net."},{"key":"29","unstructured":"Such, F. P., Rawal, A., Lehman, J., Stanley, K., and Clune, J. (2020). \u201cGenerative Teaching Networks: Accelerating Neural Architecture Search by Learning to Generate Synthetic Training Data.\u201d In Daum\u00e9 III, H. and Singh, A. (Eds.), <i>Proceedings of the 37th International Conference on Machine Learning<\/i>, Vol. 119 of <i>Proceedings of Machine Learning Research<\/i>, pp. 9206\u20139216. PMLR."},{"key":"30","doi-asserted-by":"crossref","unstructured":"Sucholutsky, I. and Schonlau, M. (2021). \u201cSoft-Label Dataset Distillation and Text Dataset Distillation.\u201d In <i>2021 International Joint Conference on Neural Networks (IJCNN)<\/i>, pp. 1\u20138.","DOI":"10.1109\/IJCNN52387.2021.9533769"},{"key":"31","unstructured":"Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., Bikel, D., Blecher, L., Canton-Ferrer, C., Chen, M., Cucurull, G., Esiobu, D., Fernandes, J., Fu, J., Fu, W., Fuller, B., Gao, C., Goswami, V., Goyal, N., Hartshorn, A., Hosseini, S., Hou, R., Inan, H., Kardas, M., Kerkez, V., Khabsa, M., Kloumann, I., Korenev, A., Koura, P. S., Lachaux, M.-A., Lavril, T., Lee, J., Liskovich, D., Lu, Y., Mao, Y., Martinet, X., Mihaylov, T., Mishra, P., Molybog, I., Nie, Y., Poulton, A., Reizenstein, J., Rungta, R., Saladi, K., Schelten, A., Silva, R., Smith, E. M., Subramanian, R., Tan, X. E., Tang, B., Taylor, R., et al. (2023). \u201cLlama 2: Open Foundation and Fine-Tuned Chat Models.\u201d <i>CoRR<\/i>, abs\/2307.09288."},{"key":"32","doi-asserted-by":"crossref","unstructured":"Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., and Bowman, S. (2018). \u201cGLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding.\u201d In <i>Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP<\/i>, pp. 353\u2013355, Brussels, Belgium. Association for Computational Linguistics.","DOI":"10.18653\/v1\/W18-5446"},{"key":"33","unstructured":"Wang, K., Gu, J., Zhou, D., Zhu, Z., Jiang, W., and You, Y. (2023). \u201cDiM: Distilling Dataset into Generative Model.\u201d <i>CoRR<\/i>, abs\/2303.04707."},{"key":"34","doi-asserted-by":"crossref","unstructured":"Wang, K., Zhao, B., Peng, X., Zhu, Z., Yang, S., Wang, S., Huang, G., Bilen, H., Wang, X., and You, Y. (2022). \u201cCAFE: Learning to Condense Dataset by Aligning Features.\u201d In <i>2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)<\/i>, pp. 12186\u201312195.","DOI":"10.1109\/CVPR52688.2022.01188"},{"key":"35","unstructured":"Wang, T., Zhu, J.-Y., Torralba, A., and Efros, A. A. (2018). \u201cDataset Distillation.\u201d <i>CoRR<\/i>, abs\/1811.10959."},{"key":"36","doi-asserted-by":"crossref","unstructured":"Welling, M. (2009). \u201cHerding Dynamical Weights to Learn.\u201d In <i>Proceedings of the 26th Annual International Conference on Machine Learning<\/i>, ICML \u201909, pp. 1121\u20131128, New York, NY, USA. Association for Computing Machinery.","DOI":"10.1145\/1553374.1553517"},{"key":"37","doi-asserted-by":"crossref","unstructured":"Wiewel, F. and Yang, B. (2021). \u201cCondensed Composite Memory Continual Learning.\u201d In <i>International Joint Conference on Neural Networks, IJCNN 2021, Shenzhen, China, July 18\u201322, 2021<\/i>, pp. 1\u20138. IEEE.","DOI":"10.1109\/IJCNN52387.2021.9533491"},{"key":"38","doi-asserted-by":"crossref","unstructured":"Wolf, G. W. (2011). \u201cFacility Location: Concepts, Models, Algorithms and Case Studies. Series: Contributions to Management Science.\u201d <i>International Journal of Geographical Information Science<\/i>, 25 (2), pp. 331\u2013333.","DOI":"10.1080\/13658816.2010.528422"},{"key":"39","doi-asserted-by":"crossref","unstructured":"Xiong, Y., Wang, R., Cheng, M., Yu, F., and Hsieh, C.-J. (2023). \u201cFedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning.\u201d In <i>IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17\u201324, 2023<\/i>, pp. 16323\u201316332. IEEE.","DOI":"10.1109\/CVPR52729.2023.01566"},{"key":"40","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J. G., Salakhutdinov, R., and Le, Q. V. (2019). \u201cXLNet: Generalized Autoregressive Pretraining for Language Understanding.\u201d In Wallach, H. M., Larochelle, H., Beygelzimer, A., d\u2019Alch\u00e9-Buc, F., Fox, E. B., and Garnett, R. (Eds.), <i>Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8\u201314, 2019, Vancouver, BC, Canada<\/i>, pp. 5754\u20135764."},{"key":"41","doi-asserted-by":"crossref","unstructured":"Yu, L., Zhang, W., Wang, J., and Yu, Y. (2017). \u201cSeqGAN: Sequence Generative Adversarial Nets with Policy Gradient.\u201d In Singh, S. and Markovitch, S. (Eds.), <i>Proceedings of the 31st AAAI Conference on Artificial Intelligence, February 4\u20139, 2017, San Francisco, California, USA<\/i>, pp. 2852\u20132858. AAAI Press.","DOI":"10.1609\/aaai.v31i1.10804"},{"key":"42","unstructured":"Yu, R., Liu, S., and Wang, X. (2023). \u201cDataset Distillation: A Comprehensive Review.\u201d <i>CoRR<\/i>, abs\/2301.07014."},{"key":"43","unstructured":"Zhang, J., Chen, C., Li, B., Lyu, L., Wu, S., Ding, S., Shen, C., and Wu, C. (2022a). \u201cDENSE: Data-Free One-Shot Federated Learning.\u201d In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (Eds.), <i>Advances in Neural Information Processing Systems<\/i>, Vol. 35, pp. 21414\u201321428. Curran Associates, Inc."},{"key":"44","unstructured":"Zhang, S., Roller, S., Goyal, N., Artetxe, M., Chen, M., Chen, S., Dewan, C., Diab, M. T., Li, X., Lin, X. V., Mihaylov, T., Ott, M., Shleifer, S., Shuster, K., Simig, D., Koura, P. S., Sridhar, A., Wang, T., and Zettlemoyer, L. (2022b). \u201cOPT: Open Pre-trained Transformer Language Models.\u201d <i>CoRR<\/i>, abs\/2205.01068."},{"key":"45","unstructured":"Zhang, Y., Gan, Z., and Carin, L. (2016). \u201cGenerating Text via Adversarial Training.\u201d In <i>NIPS Workshop on Adversarial Training<\/i>. academia.edu."},{"key":"46","unstructured":"Zhang, Y., Gan, Z., Fan, K., Chen, Z., Henao, R., Shen, D., and Carin, L. (2017). \u201cAdversarial Feature Matching for Text Generation.\u201d In Precup, D. and Teh, Y. W. (Eds.), <i>Proceedings of the 34th International Conference on Machine Learning<\/i>, Vol. 70 of <i>Proceedings of Machine Learning Research<\/i>, pp. 4006\u20134015. PMLR."},{"key":"47","unstructured":"Zhao, B. and Bilen, H. (2021). \u201cDataset Condensation with Differentiable Siamese Augmentation.\u201d In Meila, M. and Zhang, T. (Eds.), <i>Proceedings of the 38th International Conference on Machine Learning<\/i>, Vol. 139 of <i>Proceedings of Machine Learning Research<\/i>, pp. 12674\u201312685. PMLR."},{"key":"48","unstructured":"Zhao, B. and Bilen, H. (2022). \u201cSynthesizing Informative Training Samples with GAN.\u201d <i>CoRR<\/i>, abs\/2204.07513."},{"key":"49","doi-asserted-by":"crossref","unstructured":"Zhao, B. and Bilen, H. (2023). \u201cDataset Condensation with Distribution Matching.\u201d In <i>IEEE\/CVF Winter Conference on Applications of Computer Vision, WACV 2023, Waikoloa, HI, USA, January 2\u20137, 2023<\/i>, pp. 6503\u20136512. IEEE.","DOI":"10.1109\/WACV56688.2023.00645"},{"key":"50","unstructured":"Zhao, B., Mopuri, K. R., and Bilen, H. (2021). \u201cDataset Condensation with Gradient Matching.\u201d In <i>9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3\u20137, 2021<\/i>. OpenReview.net."}],"container-title":["Journal of Natural Language Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/32\/1\/32_252\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,15]],"date-time":"2025-03-15T04:37:01Z","timestamp":1742013421000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/jnlp\/32\/1\/32_252\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":50,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025]]}},"URL":"https:\/\/doi.org\/10.5715\/jnlp.32.252","relation":{},"ISSN":["1340-7619","2185-8314"],"issn-type":[{"value":"1340-7619","type":"print"},{"value":"2185-8314","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}