{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T01:06:31Z","timestamp":1742951191183,"version":"3.40.3"},"publisher-location":"Cham","reference-count":97,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031726422"},{"type":"electronic","value":"9783031726439"}],"license":[{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T00:00:00Z","timestamp":1732233600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72643-9_10","type":"book-chapter","created":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T20:48:00Z","timestamp":1732222080000},"page":"158-175","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["How Video Meetings Change Your Expression"],"prefix":"10.1007","author":[{"given":"Sumit","family":"Sarin","sequence":"first","affiliation":[]},{"given":"Utkarsh","family":"Mall","sequence":"additional","affiliation":[]},{"given":"Purva","family":"Tendulkar","sequence":"additional","affiliation":[]},{"given":"Carl","family":"Vondrick","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,22]]},"reference":[{"key":"10_CR1","doi-asserted-by":"crossref","unstructured":"Zhao, N., Zhang, X., Noah, J.A., Tiede, M., Hirsch, J.: Separable processes for live \u201cin-person\u201d and live \u201czoom-like\u201d faces. Imaging Neurosci. (2023)","DOI":"10.1162\/imag_a_00027"},{"key":"10_CR2","doi-asserted-by":"crossref","unstructured":"Balters, S., Miller, J.G., Li, R., Hawthorne, G., Reiss, A.L.: Virtual (Zoom) Interactions Alter Conversational Behavior and Inter-Brain Coherence. bioRxiv (2023)","DOI":"10.1523\/JNEUROSCI.1401-22.2023"},{"key":"10_CR3","doi-asserted-by":"crossref","unstructured":"Matz, S., Harari, G.: Personality\u2013place transactions: mapping the relationships between big five personality traits, states, and daily places. J. Personal. Soc. Psychol. (2020)","DOI":"10.1037\/pspp0000297"},{"key":"10_CR4","unstructured":"Khan, M.R.: A review of the effects of virtual communication on performance and satisfaction across the last ten years of research. J. Appl. Behav. Anal. (2021)"},{"key":"10_CR5","doi-asserted-by":"crossref","unstructured":"Archibald, M., Ambagtsheer, R., Casey, M., Lawless, M.: Using zoom videoconferencing for qualitative data collection: perceptions and experiences of researchers and participants. Int. J. Qualit. Methods (2019)","DOI":"10.1177\/1609406919874596"},{"key":"10_CR6","doi-asserted-by":"crossref","unstructured":"Nesher\u00a0Shoshan, H., Wehrt, W.: Understanding \u201czoom fatigue\u201d: a mixed-method approach. Appl. Psychol. (2022)","DOI":"10.1111\/apps.12360"},{"key":"10_CR7","doi-asserted-by":"crossref","unstructured":"Fauville, G., Luo, M., Queiroz, A.C.M., Bailenson, J.N., Hancock, J.: Zoom Exhaustion and Fatigue Scale. Comput. Human Behav. Rep. (2021)","DOI":"10.2139\/ssrn.3786329"},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Bailenson, J.N.: Nonverbal Overload: A Theoretical Argument for the Causes of Zoom Fatigue, Mind, and Behavior, Technology (2021)","DOI":"10.1037\/tmb0000030"},{"key":"10_CR9","doi-asserted-by":"crossref","unstructured":"Boland, J., Fonseca, P., Mermelstein, I., Williamson, M.: Zoom disrupts the rhythm of conversation. J. Exp. Psychol. Gen. (2021)","DOI":"10.1037\/xge0001150"},{"key":"10_CR10","doi-asserted-by":"crossref","unstructured":"Fauville, G., Luo, M., Queiroz, A.C., Bailenson, J., Hancock, J.: Zoom exhaustion and fatigue scale. SSRN Electron. J. (2021)","DOI":"10.2139\/ssrn.3786329"},{"key":"10_CR11","doi-asserted-by":"crossref","unstructured":"Hoehe, M., Thibaut, F.: Going digital: how technology use may influence human brains and behavior. Dialog. Clin. Neurosci. (2020)","DOI":"10.31887\/DCNS.2020.22.2\/mhoehe"},{"key":"10_CR12","doi-asserted-by":"crossref","unstructured":"Numata, T., et al.: Achieving affective human\u2013virtual agent communication by enabling virtual agents to imitate positive expressions. Sci. Rep. (2020)","DOI":"10.1038\/s41598-020-62870-7"},{"key":"10_CR13","doi-asserted-by":"crossref","unstructured":"Smith, H.J., Neff, M.: Communication behavior in embodied virtual reality. In: ACM CHI (2018)","DOI":"10.1145\/3173574.3173863"},{"key":"10_CR14","unstructured":"Geng, S., Teotia, R., Tendulkar, P., Menon, S., Vondrick, C.: Affective faces for goal-driven dyadic communication. CoRR (2023)"},{"key":"10_CR15","doi-asserted-by":"crossref","unstructured":"Fong, R., Patrick, M., Vedaldi, A.: Understanding deep networks via extremal perturbations and smooth masks. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00304"},{"key":"10_CR16","unstructured":"Petsiuk, V., Das, A., Saenko, K.: Rise: randomized input sampling for explanation of black-box models. CoRR (2018)"},{"key":"10_CR17","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"10_CR18","unstructured":"Shitole, V., Li, F., Kahng, M., Tadepalli, P., Fern, A.: One explanation is not enough: structured attention graphs for image classification. In: NeurIPS (2021)"},{"key":"10_CR19","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: ECCV (2014)","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"10_CR20","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.319"},{"key":"10_CR21","doi-asserted-by":"crossref","unstructured":"Gunning, D., Aha, D.: Darpa\u2019s explainable artificial intelligence (XAI) Program. AI Magazine (2019)","DOI":"10.1145\/3301275.3308446"},{"key":"10_CR22","unstructured":"Goyal, Y., Wu, Z., Ernst, J., Batra, D., Parikh, D., Lee, S.: Counterfactual visual explanations. In: ICML (2019)"},{"key":"10_CR23","doi-asserted-by":"crossref","unstructured":"Vandenhende, S., Mahajan, D., Radenovic, F., Ghadiyaram, D.: Making heads or tails: towards semantically consistent visual counterfactuals. In: ECCV (2022)","DOI":"10.1007\/978-3-031-19775-8_16"},{"key":"10_CR24","doi-asserted-by":"crossref","unstructured":"Wang, P., Vasconcelos, N.: Scout: self-aware discriminant counterfactual explanations. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00900"},{"key":"10_CR25","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: \u201cWhy should i trust you?\u201d explaining the predictions of any classifier. In: SIGKDD (2016)","DOI":"10.18653\/v1\/N16-3020"},{"key":"10_CR26","unstructured":"Koh, P.W., Liang, P.: Understanding black-box predictions via influence functions. In: ICML (2017)"},{"key":"10_CR27","unstructured":"Yeh, C.-K., Kim, J., Yen, I.\u00a0E.-H., Ravikumar, P.K.: Representer point selection for explaining deep neural networks. In: NeurIPS (2018)"},{"key":"10_CR28","unstructured":"Tsai, C.-P., Yeh, C.-K., Ravikumar, P.: Sample based explanations via generalized representers. In: CoRR (2023)"},{"key":"10_CR29","unstructured":"Sui, Y., Wu, G., Sanner, S.: Representer point selection via local Jacobian expansion for post-hoc classifier explanation of deep neural networks and ensemble models. In: NeurIPS (2021)"},{"key":"10_CR30","unstructured":"Pruthi, G., Liu, F., Sundararajan, M., Kale, S.: Estimating training data influence by tracking gradient descent. CoRR (2020)"},{"key":"10_CR31","unstructured":"Silva, A., Chopra, R., Gombolay, M.C.: Cross-loss influence functions to explain deep network representations. In: AISTATS (2020)"},{"key":"10_CR32","doi-asserted-by":"crossref","unstructured":"Guo, H., Rajani, N., Hase, P., Bansal, M., Xiong, C.: \u201cFastif: scalable influence functions for efficient model interpretation and debugging. CoRR (2020)","DOI":"10.18653\/v1\/2021.emnlp-main.808"},{"key":"10_CR33","doi-asserted-by":"crossref","unstructured":"Pan, W., Cui, S., Bian, J., Zhang, C., Wang, F.: Explaining algorithmic fairness through fairness-aware causal path decomposition. In: SIGKDD (2021)","DOI":"10.1145\/3447548.3467258"},{"key":"10_CR34","doi-asserted-by":"crossref","unstructured":"Pradhan, R., Zhu, J., Glavic, B., Salimi, B.: Interpretable data-based explanations for fairness debugging. In: SIGMOD (2022)","DOI":"10.1145\/3514221.3517886"},{"key":"10_CR35","doi-asserted-by":"crossref","unstructured":"Meng, C., Trinh, L., Xu, N., Enouen, J., Liu, Y.: Interpretability and fairness evaluation of deep learning models on mimic-iv dataset. Sci. Rep. (2022)","DOI":"10.21203\/rs.3.rs-402058\/v1"},{"key":"10_CR36","doi-asserted-by":"crossref","unstructured":"Alelyani, S.: Detection and evaluation of machine learning bias. Appl. Sci. (2021)","DOI":"10.3390\/app11146271"},{"key":"10_CR37","doi-asserted-by":"crossref","unstructured":"Gilpin, L.H., Bau, D., Yuan, B.Z., Bajwa, A., Specter, M., Kagal, L.: Explaining explanations: an overview of interpretability of machine learning. In: DSAA (2018)","DOI":"10.1109\/DSAA.2018.00018"},{"key":"10_CR38","doi-asserted-by":"crossref","unstructured":"Kim, S.S., Meister, N., Ramaswamy, V.V., Fong, R., Russakovsky, O.: Hive: evaluating the human interpretability of visual explanations. In: ECCV (2022)","DOI":"10.1007\/978-3-031-19775-8_17"},{"key":"10_CR39","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., et al.: Squinting at VGA models: introspecting VGA models with sub-questions. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01002"},{"key":"10_CR40","doi-asserted-by":"crossref","unstructured":"Das, A., Agrawal, H., Zitnick, L., Parikh, D., Batra, D.: Human attention in visual question answering: do humans and deep networks look at the same regions? In: Computer Vision and Image Understanding (2017)","DOI":"10.1016\/j.cviu.2017.10.001"},{"key":"10_CR41","unstructured":"Brendel, W., Bethge, M.: Approximating CNNs with bag-of-local-features models works surprisingly well on imagenet. CoRR (2019)"},{"key":"10_CR42","doi-asserted-by":"crossref","unstructured":"Bohle, M., Fritz, M., Schiele, B.: Convolutional dynamic alignment networks for interpretable classifications. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00990"},{"key":"10_CR43","doi-asserted-by":"crossref","unstructured":"B\u00f6hle, M., Fritz, M., Schiele, B.: B-cos networks: alignment is all we need for interpretability. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01008"},{"key":"10_CR44","unstructured":"Chen, C., Li, O., Tao, D., Barnett, A., Rudin, C., Su, J.K.: This looks like that: deep learning for interpretable image recognition. In: NeurIPS (2019)"},{"key":"10_CR45","doi-asserted-by":"crossref","unstructured":"Donnelly, J., Barnett, A.J., Chen, C.: Deformable protopnet: an interpretable image classifier using deformable prototypes. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01002"},{"key":"10_CR46","unstructured":"Koh, P.W., et al.: Concept bottleneck models. In: ICML (2020)"},{"key":"10_CR47","doi-asserted-by":"crossref","unstructured":"Hastie, T., Tibshirani, R.: Generalized additive models. Statist. Sci. (1986)","DOI":"10.1214\/ss\/1177013604"},{"key":"10_CR48","doi-asserted-by":"crossref","unstructured":"Lou, Y., Caruana, R., Gehrke, J., Hooker, G.: Accurate intelligible models with pairwise interactions. In: SIGKDD (2013)","DOI":"10.1145\/2487575.2487579"},{"key":"10_CR49","unstructured":"Dubey, A., Radenovic, F., Mahajan, D.: Scalable interpretability via polynomials. In: NeurIPS (2022)"},{"key":"10_CR50","unstructured":"Radenovic, F., Dubey, A., Mahajan, D.: Neural basis models for interpretability. In: NeurIPS (2022)"},{"key":"10_CR51","unstructured":"Chang, C.-H., Caruana, R., Goldenberg, A.: Node-gam: neural generalized additive model for interpretable deep learning. In: ICLR (2022)"},{"key":"10_CR52","unstructured":"Agarwal, R., et al.: Neural additive models: interpretable machine learning with neural nets. In: NeurIPS (2021)"},{"key":"10_CR53","unstructured":"Burgess, C.P., et al.: Understanding disentangling in $$\\beta $$-vae. In: CoRR (2018)"},{"key":"10_CR54","unstructured":"Zhu, Z., Luo, P., Wang, X., Tang, X.: Multi-view perceptron: a deep model for learning face identity and view representations. In: NeurIPS (2014)"},{"key":"10_CR55","unstructured":"Reed, S.E., Sohn, K., Zhang, Y., Lee, H.: Learning to disentangle factors of variation with manifold interaction. In: ICML (2014)"},{"key":"10_CR56","unstructured":"Whitney, W.F., Chang, M., Kulkarni, T.D., Tenenbaum, J.B.: Understanding visual concepts with continuation learning. CoRR (2016)"},{"key":"10_CR57","unstructured":"Cheung, B., Livezey, J.A., Bansal, A.K., Olshausen, B.A.: Discovering hidden factors of variation in deep networks. CoRR (2014)"},{"key":"10_CR58","unstructured":"Lin, Z., Thekumparampil, K.K., Fanti, G.C., Oh, S.: Infogan-cr: disentangling generative adversarial networks with contrastive regularizers. CoRR (2019)"},{"key":"10_CR59","doi-asserted-by":"crossref","unstructured":"Jeon, I., Lee, W., Pyeon, M., Kim, G.: IB-GAN: disentangled representation learning with information bottleneck generative adversarial networks. In: AAAI (2021)","DOI":"10.1609\/aaai.v35i9.16967"},{"key":"10_CR60","unstructured":"Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., Abbeel, P.: Infogan: interpretable representation learning by information maximizing generative adversarial nets. In: NeurIPS (2016)"},{"key":"10_CR61","unstructured":"Ramesh, A., Choi, Y., LeCun, Y.: A spectral regularizer for unsupervised disentanglement. CoRR (2018)"},{"key":"10_CR62","doi-asserted-by":"crossref","unstructured":"Dalva, Y., Alt\u0131ndi\u015f, S.\u00a0F., Dundar, A.: Vecgan: image-to-image translation with interpretable latent directions. In: ECCV (2022)","DOI":"10.1007\/978-3-031-19787-1_9"},{"key":"10_CR63","unstructured":"Dalva, Y., Pehlivan, H., Moran, C., Hatipo\u011flu, \u00d6.I., D\u00fcndar, A.: Face attribute editing with disentangled latent vectors. CoRR (2023)"},{"key":"10_CR64","unstructured":"Higgins, I., et al.: beta-VAE: learning basic visual concepts with a constrained variational framework. In: ICLR (2017)"},{"key":"10_CR65","unstructured":"Kim, H., Mnih, A.: Disentangling by factorising. In: ICML (2018)"},{"key":"10_CR66","unstructured":"Chen, T.Q., Li, X., Grosse, R.B., Duvenaud, D.K.: Isolating sources of disentanglement in variational autoencoders. CoRR (2018)"},{"key":"10_CR67","unstructured":"Jeong, Y., Song, H.O.: Learning discrete and continuous factors of data via alternating disentanglement. In: ICML (2019)"},{"key":"10_CR68","unstructured":"Kumar, A., Sattigeri, P., Balakrishnan, A.: Variational inference of disentangled latent concepts from unlabeled observations. CoRR (2017)"},{"key":"10_CR69","doi-asserted-by":"crossref","unstructured":"Yesu, K., Shandilya, S., Rekharaj, N., Ankit, K., Sairam, P.S.: Big five personality traits inference from five facial shapes using CNN. In: International Conference on Computing, Power and Communication Technologies (GUCON) (2021)","DOI":"10.1109\/GUCON50781.2021.9573895"},{"key":"10_CR70","doi-asserted-by":"crossref","unstructured":"Knyazev, G.G., Bocharov, A.V., Slobodskaya, H.R., Ryabichenko, T.I.: Personality-linked biases in perception of emotional facial expressions. In: Personality and Individual Differences (2008)","DOI":"10.1016\/j.paid.2007.11.001"},{"key":"10_CR71","doi-asserted-by":"crossref","unstructured":"Kachur, A., Osin, E., Davydov, D., Shutilov, K., Novokshonov, A.: Assessing the big five personality traits using real-life static facial images. Sci. Rep. (2020)","DOI":"10.31234\/osf.io\/3y98a"},{"key":"10_CR72","doi-asserted-by":"crossref","unstructured":"B\u00fcdenbender, B., H\u00f6fling, T.T.A., Gerdes, A.B.M., Alpers, G.W.: Training machine learning algorithms for automatic facial coding: the role of emotional facial expressions prototypicality. PLOS One (2023)","DOI":"10.1371\/journal.pone.0281309"},{"key":"10_CR73","doi-asserted-by":"crossref","unstructured":"Stahelski, A., Anderson, A., Browitt, N., Radeke, M.: Facial expressions and emotion labels are separate initiators of trait inferences from the face. Front. Psychol. (2021)","DOI":"10.3389\/fpsyg.2021.749933"},{"key":"10_CR74","doi-asserted-by":"crossref","unstructured":"Snoek, L., et al.: Testing, explaining, and exploring models of facial expressions of emotions. Sci. Adv. (2023)","DOI":"10.1126\/sciadv.abq8421"},{"key":"10_CR75","doi-asserted-by":"crossref","unstructured":"Straulino, E., Scarpazza, C., Sartori, L.: What is missing in the study of emotion expression? Front. Psychol. (2023)","DOI":"10.3389\/fpsyg.2023.1158136"},{"key":"10_CR76","doi-asserted-by":"crossref","unstructured":"Du, S., Tao, Y., Martinez, A.M.: Compound facial expressions of emotion. PNAS (2014)","DOI":"10.1073\/pnas.1322355111"},{"key":"10_CR77","doi-asserted-by":"crossref","unstructured":"Minetaki, K.: Facial expression and description of personality. In: ACM MISNC (2023)","DOI":"10.1145\/3624875.3624894"},{"key":"10_CR78","doi-asserted-by":"crossref","unstructured":"Jonell, P., Kucherenko, T., Henter, G.E., Beskow, J.: Let\u2019s face it: probabilistic multi-modal interlocutor-aware generation of facial gestures in dyadic settings. In: ACM IVA (2020)","DOI":"10.1145\/3383652.3423911"},{"key":"10_CR79","doi-asserted-by":"crossref","unstructured":"Ng, E., Subramanian, S., Klein, D., Kanazawa, A., Darrell, T., Ginosar, S.: Can language models learn to listen? In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00925"},{"key":"10_CR80","doi-asserted-by":"crossref","unstructured":"Ng, E., et al.: Learning to listen: modeling non-deterministic dyadic facial motion. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01975"},{"key":"10_CR81","unstructured":"Burgess, C.P., et al.: Understanding disentangling in $$\\beta $$-vae. (2018)"},{"key":"10_CR82","unstructured":"Li, Z., Liu, H.: Beta-VAE has 2 behaviors: PCA or ICA? (2023)"},{"key":"10_CR83","unstructured":"Garc\u00eda\u00a0de Herreros\u00a0Garc\u00eda, P.: Towards latent space disentanglement of variational autoencoders for language (2022)"},{"key":"10_CR84","unstructured":"Pastrana, R.: Disentangling variational autoencoders. CoRR (2022)"},{"key":"10_CR85","doi-asserted-by":"crossref","unstructured":"Higgins, I., et al.: Unsupervised deep learning identifies semantic disentanglement in single inferotemporal face patch neurons. Nat. Commun. (2021)","DOI":"10.1038\/s41467-021-26751-5"},{"key":"10_CR86","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NeurIPS (2014)"},{"key":"10_CR87","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"10_CR88","unstructured":"Chakrabarty, A., Das, S.: On translation and reconstruction guarantees of the cycle-consistent generative adversarial networks. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems, vol.\u00a035, pp.\u00a023607\u201323620. Curran Associates, Inc. (2022)"},{"key":"10_CR89","doi-asserted-by":"crossref","unstructured":"Shen, Z., Zhou, S.K., Chen, Y., Georgescu, B., Liu, X., Huang, T.: One-to-one mapping for unpaired image-to-image translation. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp.\u00a01170\u20131179 (2020)","DOI":"10.1109\/WACV45572.2020.9093622"},{"key":"10_CR90","unstructured":"Wang, T.-C., et al.: Video-to-video synthesis. In: NeurIPS (2018)"},{"key":"10_CR91","doi-asserted-by":"crossref","unstructured":"Kuster, C., Popa, T., Bazin, J.-C., Gotsman, C., Gross, M.: Gaze correction for home video conferencing. In: ACM TOG (2012)","DOI":"10.1145\/2366145.2366193"},{"key":"10_CR92","unstructured":"Hill, F.: The gesture that encapsulates remote-work life. The Atlantic, 20 July (2023)"},{"key":"10_CR93","doi-asserted-by":"crossref","unstructured":"Schwarz, G.: Estimating the dimension of a model. The Annals of Statistics (1978)","DOI":"10.1214\/aos\/1176344136"},{"key":"10_CR94","unstructured":"Denby, D.: The three faces of Trump. The New Yorker, August (2015)"},{"key":"10_CR95","unstructured":"Collett, P.: The seven faces of Donald Trump\u2014a psychologist\u2019s view. The Guardian, January (2017)"},{"key":"10_CR96","unstructured":"Golshan, T.: Donald Trump\u2019s unique speaking style, explained by linguists. Vox, January (2017)"},{"key":"10_CR97","unstructured":"Locatello, F., et al.: Challenging common assumptions in the unsupervised learning of disentangled representations. In: ICML (2019)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72643-9_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T21:25:37Z","timestamp":1732224337000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72643-9_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,22]]},"ISBN":["9783031726422","9783031726439"],"references-count":97,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72643-9_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,22]]},"assertion":[{"value":"22 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}