{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T03:36:47Z","timestamp":1769744207688,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":49,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556953","type":"print"},{"value":"9789819556960","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5696-0_19","type":"book-chapter","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T14:04:14Z","timestamp":1769695454000},"page":"266-280","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Modular MeanFlow: Towards Stable and\u00a0Scalable One-Step Generative Modeling"],"prefix":"10.1007","author":[{"given":"Haochen","family":"You","sequence":"first","affiliation":[]},{"given":"Baojing","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Hongyang","family":"He","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,30]]},"reference":[{"key":"19_CR1","unstructured":"Bartosh, G., Vetrov, D., Naesseth, C.A.: Neural diffusion models. arXiv preprint arXiv:2310.08337 (2023)"},{"key":"19_CR2","doi-asserted-by":"crossref","unstructured":"Chen, M., Mei, S., Fan, J., Wang, M.: Opportunities and challenges of diffusion models for generative AI. National Sci. Rev. 11(12) (2024)","DOI":"10.1093\/nsr\/nwae348"},{"key":"19_CR3","first-page":"68552","volume":"36","author":"S Chen","year":"2023","unstructured":"Chen, S., Chewi, S., Lee, H., Li, Y., Lu, J., Salim, A.: The probability flow ode is provably fast. Adv. Neural. Inf. Process. Syst. 36, 68552\u201368575 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR4","unstructured":"Davtyan, A., Dadi, L.T., Cevher, V., Favaro, P.: Faster inference of flow-based generative models via improved data-noise coupling. In: The Thirteenth International Conference on Learning Representations (2025)"},{"key":"19_CR5","unstructured":"Dinh, L., Krueger, D., Bengio, Y.: Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516 (2014)"},{"key":"19_CR6","unstructured":"Dinh, L., Sohl-Dickstein, J., Bengio, S.: Density estimation using real nvp. arXiv preprint arXiv:1605.08803 (2016)"},{"key":"19_CR7","doi-asserted-by":"crossref","unstructured":"Dong, L., et al.: Tsd-sr: One-step diffusion with target score distillation for real-world image super-resolution. In: Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 23174\u201323184 (2025)","DOI":"10.1109\/CVPR52734.2025.02158"},{"key":"19_CR8","unstructured":"Esser, P., et\u00a0al.: Scaling rectified flow transformers for high-resolution image synthesis. In: Forty-first International Conference on Machine Learning (2024)"},{"key":"19_CR9","unstructured":"Geng, Z., Deng, M., Bai, X., Kolter, J.Z., He, K.: Mean flows for one-step generative modeling. arXiv preprint arXiv:2505.13447 (2025)"},{"key":"19_CR10","first-page":"41914","volume":"36","author":"Z Geng","year":"2023","unstructured":"Geng, Z., Pokle, A., Kolter, J.Z.: One-step diffusion distillation via deep equilibrium models. Adv. Neural. Inf. Process. Syst. 36, 41914\u201341931 (2023)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR11","doi-asserted-by":"crossref","unstructured":"He, H., Xie, H., Shen, G., Fu, B., You, H., Sanchez\u00a0Silva, V.: 4s-classifier: empowering conservation through semi-supervised learning for rare and endangered species (2025)","DOI":"10.1109\/IJCNN64981.2025.11227447"},{"key":"19_CR12","unstructured":"He, H., Xie, H., You, H., Sanchez\u00a0Silva, V.: Semi-vim: bidirectional state space model for mitigating label imbalance in semi-supervised learning (2025)"},{"key":"19_CR13","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"19_CR14","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR15","doi-asserted-by":"publisher","first-page":"110464","DOI":"10.52202\/079017-3507","volume":"37","author":"P Holderrieth","year":"2024","unstructured":"Holderrieth, P., Xu, Y., Jaakkola, T.: Hamiltonian score matching and generative flows. Adv. Neural. Inf. Process. Syst. 37, 110464\u2013110493 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR16","unstructured":"Jeha, P., Grathwohl, W., Andersen, M.R., Ek, C.H., Frellsen, J.: Variance reduction of diffusion model\u2019s gradients with Taylor approximation-based control variate. arXiv preprint arXiv:2408.12270 (2024)"},{"key":"19_CR17","unstructured":"Kim, D., et al.: Consistency trajectory models: Learning probability flow ode trajectory of diffusion. arXiv preprint arXiv:2310.02279 (2023)"},{"key":"19_CR18","unstructured":"Kim, D., et al.: Pagoda: Progressive growing of a one-step generator from a low-resolution diffusion teacher. arXiv preprint arXiv:2405.14822 (2024)"},{"key":"19_CR19","unstructured":"Kingma, D.P., Dhariwal, P.: Glow: Generative flow with invertible 1x1 convolutions. Adv. Neural Inf. Process. Syst. 31 (2018)"},{"key":"19_CR20","unstructured":"Krizhevsky, A., Hinton, G., et\u00a0al.: Learning multiple layers of features from tiny images (2009)"},{"key":"19_CR21","first-page":"63082","volume":"37","author":"S Lee","year":"2024","unstructured":"Lee, S., Lin, Z., Fanti, G.: Improving the training of rectified flows. Adv. Neural. Inf. Process. Syst. 37, 63082\u201363109 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR22","unstructured":"Li, Y., Dong, J., Dong, Z., Yang, C., An, Z., Xu, Y.: Srkd: Towards efficient 3D point cloud segmentation via structure-and relation-aware knowledge distillation. arXiv preprint arXiv:2506.17290 (2025)"},{"key":"19_CR23","unstructured":"Li, Y., et al.: Sepprune: Structured pruning for efficient deep speech separation. arXiv preprint arXiv:2505.12079 (2025)"},{"key":"19_CR24","doi-asserted-by":"crossref","unstructured":"Li, Y., et al.: Comae: Comprehensive attribute exploration for zero-shot hashing. ICMR (2025)","DOI":"10.1145\/3731715.3733294"},{"key":"19_CR25","unstructured":"Li, Y., Lu, Y., Dong, Z., Yang, C., Chen, Y., Gou, J.: Sglp: A similarity guided fast layer partition pruning for compressing large deep models. arXiv preprint arXiv:2410.14720 (2024)"},{"key":"19_CR26","unstructured":"Li, Y., et al.: Frequency-aligned knowledge distillation for lightweight spatiotemporal forecasting. arXiv:2507.02939 (2025)"},{"key":"19_CR27","unstructured":"Lipman, Y., Chen, R.T., Ben-Hamu, H., Nickel, M., Le, M.: Flow matching for generative modeling. arXiv preprint arXiv:2210.02747 (2022)"},{"key":"19_CR28","first-page":"5775","volume":"35","author":"C Lu","year":"2022","unstructured":"Lu, C., Zhou, Y., Bao, F., Chen, J., Li, C., Zhu, J.: Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. Adv. Neural. Inf. Process. Syst. 35, 5775\u20135787 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR29","doi-asserted-by":"crossref","unstructured":"One-step diffusion distillation through score implicit matching: Luo, W., Huang, Z., Geng, Z., Kolter, J.Z., Qi, G.j. Adv. Neural. Inf. Process. Syst. 37, 115377\u2013115408 (2024)","DOI":"10.52202\/079017-3664"},{"key":"19_CR30","doi-asserted-by":"crossref","unstructured":"Ma, Z., et\u00a0al.: Efficient diffusion models: A comprehensive survey from principles to practices. IEEE Trans. Pattern Anal. Mach. Intell. (2025)","DOI":"10.1109\/TPAMI.2025.3569700"},{"key":"19_CR31","unstructured":"Pascanu, R., Mikolov, T., Bengio, Y.: On the difficulty of training recurrent neural networks. In: International Conference on Machine Learning, pp. 1310\u20131318. Pmlr (2013)"},{"issue":"2","key":"19_CR32","doi-asserted-by":"publisher","DOI":"10.1111\/avsc.70022","volume":"28","author":"D R\u00e1tiva Gaona","year":"2025","unstructured":"R\u00e1tiva Gaona, D., Wehncke, E.V., Mariano, N.A., Osorio Beristain, M.: Hydrochory, a key ecological function of a tropical dry forest river threatened by a dam and open-pit coal mining in colombia. Appl. Veg. Sci. 28(2), e70022 (2025)","journal-title":"Appl. Veg. Sci."},{"key":"19_CR33","unstructured":"Salimans, T., Ho, J.: Progressive distillation for fast sampling of diffusion models. arXiv preprint arXiv:2202.00512 (2022)"},{"key":"19_CR34","unstructured":"Song, Y., Dhariwal, P.: Improved techniques for training consistency models. arXiv preprint arXiv:2310.14189 (2023)"},{"key":"19_CR35","unstructured":"Song, Y., Dhariwal, P., Chen, M., Sutskever, I.: Consistency models (2023)"},{"key":"19_CR36","unstructured":"Song, Y., Sohl-Dickstein, J., Kingma, D.P., Kumar, A., Ermon, S., Poole, B.: Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456 (2020)"},{"key":"19_CR37","unstructured":"Wang, F.Y., Geng, Z., Li, H.: Stable consistency tuning: Understanding and improving consistency models. arXiv preprint arXiv:2410.18958 (2024)"},{"key":"19_CR38","unstructured":"Wang, J., Zhang, D., Luo, J., Yang, Y., Luo, F.: Integration flow models. arXiv preprint arXiv:2504.20179 (2025)"},{"key":"19_CR39","unstructured":"Wang, Y., Bai, W., Zhang, C., Zhang, D., Luo, W., Sun, H.: Uni-instruct: One-step diffusion model through unified diffusion divergence instruction. arXiv preprint arXiv:2505.20755 (2025)"},{"key":"19_CR40","first-page":"45073","volume":"37","author":"S Xie","year":"2024","unstructured":"Xie, S., et al.: Em distillation for one-step diffusion models. Adv. Neural. Inf. Process. Syst. 37, 45073\u201345104 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR41","unstructured":"Xie, Y., Cheng, X.: Flow-based generative models as iterative algorithms in probability space. arXiv preprint arXiv:2502.13394 (2025)"},{"key":"19_CR42","doi-asserted-by":"crossref","unstructured":"Xu, S., Ye, Y., Li, M., You, H., Wang, K., Zhang, W.: Drco: a toolkit for intelligently curbing illegal wildlife trade (2025)","DOI":"10.21203\/rs.3.rs-6602662\/v1"},{"key":"19_CR43","doi-asserted-by":"crossref","unstructured":"Yin, T., et al.: One-step diffusion with distribution matching distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6613\u20136623 (2024)","DOI":"10.1109\/CVPR52733.2024.00632"},{"key":"19_CR44","doi-asserted-by":"crossref","unstructured":"You, H., Liu, B.: Application of Pseudometric functions in clustering and a novel similarity measure based on path information discrepancy. In: International Conference on Neural Information Processing, pp. 59\u201373. Springer (2024)","DOI":"10.1007\/978-981-96-6579-2_5"},{"key":"19_CR45","doi-asserted-by":"crossref","unstructured":"You, H., Liu, B.: Mover: Multimodal optimal transport with volume-based embedding regularization. arXiv preprint arXiv:2508.12149 (2025)","DOI":"10.1145\/3746252.3760958"},{"key":"19_CR46","doi-asserted-by":"crossref","unstructured":"Zeng, H., Li, Y., Niu, R., Yang, C., Wen, S.: Enhancing spatiotemporal prediction through the integration of mamba state space models and diffusion transformers. Knowledge-Based Systems (2025)","DOI":"10.1016\/j.knosys.2025.113347"},{"key":"19_CR47","first-page":"15511","volume":"33","author":"B Zhang","year":"2020","unstructured":"Zhang, B., Jin, J., Fang, C., Wang, L.: Improved analysis of clipping algorithms for non-convex optimization. Adv. Neural. Inf. Process. Syst. 33, 15511\u201315521 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"19_CR48","unstructured":"Zheng, H., Nie, W., Vahdat, A., Azizzadenesheli, K., Anandkumar, A.: Fast sampling of diffusion models via operator learning. In: International Conference on Machine Learning, pp. 42390\u201342402. PMLR (2023)"},{"key":"19_CR49","unstructured":"Zhengyu\u00a0Huang, D., Huang, J., Lin, Z.: Convergence analysis of probability flow ode for score-based generative models. arXiv e-prints pp. arXiv\u20132404 (2024)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5696-0_19","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T14:04:36Z","timestamp":1769695476000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5696-0_19"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556953","9789819556960"],"references-count":49,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5696-0_19","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"30 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}