{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T10:31:19Z","timestamp":1769941879042,"version":"3.49.0"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"28-29","license":[{"start":{"date-parts":[[2021,2,12]],"date-time":"2021-02-12T00:00:00Z","timestamp":1613088000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2021,2,12]],"date-time":"2021-02-12T00:00:00Z","timestamp":1613088000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100008122","name":"ITRC","doi-asserted-by":"crossref","award":["IITP-2020-2017-0-01630"],"award-info":[{"award-number":["IITP-2020-2017-0-01630"]}],"id":[{"id":"10.13039\/501100008122","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100014188","name":"MSIT","doi-asserted-by":"crossref","award":["2019R1A2C1010476"],"award-info":[{"award-number":["2019R1A2C1010476"]}],"id":[{"id":"10.13039\/501100014188","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2021,11]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>This paper proposes a novel, lightweight method to generate animated graphical interchange format images (GIFs) using the computational resources of a client device. The method analyzes an acoustic feature from the climax section of an audio file to estimate the timestamp corresponding to the maximum pitch. Further, it processes a small video segment to generate the GIF instead of processing the entire video. This makes the proposed method computationally efficient, unlike baseline approaches that use entire videos to create GIFs. The proposed method retrieves and uses the audio file and video segment so that communication and storage efficiencies are improved in the GIF generation process. Experiments on a set of 16 videos show that the proposed approach is 3.76 times more computationally efficient than a baseline method on an Nvidia Jetson TX2. Additionally, in a qualitative evaluation, the GIFs generated using the proposed method received higher overall ratings compared to those generated by the baseline method. To the best of our knowledge, this is the first technique that uses an acoustic feature in the GIF generation process.<\/jats:p>","DOI":"10.1007\/s11042-020-10236-6","type":"journal-article","created":{"date-parts":[[2021,2,12]],"date-time":"2021-02-12T15:07:11Z","timestamp":1613142431000},"page":"35923-35940","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["Client-driven animated GIF generation framework using an acoustic feature"],"prefix":"10.1007","volume":"80","author":[{"given":"Ghulam","family":"Mujtaba","sequence":"first","affiliation":[]},{"given":"Sangsoon","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Jaehyoun","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Eun-Seok","family":"Ryu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,2,12]]},"reference":[{"issue":"16","key":"10236_CR1","doi-asserted-by":"publisher","first-page":"4114","DOI":"10.1109\/TSP.2014.2326991","volume":"62","author":"J And\u00e9n","year":"2014","unstructured":"And\u00e9n J, Mallat S (2014) Deep scattering spectrum. IEEE Trans Signal Process 62(16):4114\u20134128","journal-title":"IEEE Trans Signal Process"},{"key":"10236_CR2","doi-asserted-by":"crossref","unstructured":"Bakhshi S, Shamma DA, Kennedy L, Song Y, De Juan P, Kaye J (2016) Fast, cheap, and good: Why animated gifs engage us. In: Proceedings of the 2016 chi conference on human factors in computing systems, pp 575\u2013586","DOI":"10.1145\/2858036.2858532"},{"key":"10236_CR3","doi-asserted-by":"crossref","unstructured":"Belshe M, Peon R, Thomson M (2015) Hypertext transfer protocol version 2 (http\/2)","DOI":"10.17487\/RFC7540"},{"key":"10236_CR4","doi-asserted-by":"crossref","unstructured":"Chen W, Rudovic OO, Picard RW (2017) Gifgif+: Collecting emotional animated gifs with clustered multi-task learning. In: 2017 seventh international conference on affective computing and intelligent interaction (ACII), IEEE, pp 510\u2013517","DOI":"10.1109\/ACII.2017.8273647"},{"key":"10236_CR5","doi-asserted-by":"crossref","unstructured":"Chiliguano P, Fazekas G (2016) Hybrid music recommender using content-based and social information. In: 2016 IEEE international conference on acoustics, speech and signal processing, ICASSP, IEEE, pp 2618\u20132622","DOI":"10.1109\/ICASSP.2016.7472151"},{"key":"10236_CR6","unstructured":"Choi K, Fazekas G, Sandler M (2016) Automatic tagging using deep convolutional neural networks. arXiv:160600298"},{"key":"10236_CR7","doi-asserted-by":"crossref","unstructured":"Choi K, Fazekas G, Sandler M, Cho K (2017) Convolutional recurrent neural networks for music classification. In: 2017 IEEE international conference on acoustics, speech and signal processing, ICASSP, IEEE, pp 2392\u20132396","DOI":"10.1109\/ICASSP.2017.7952585"},{"issue":"11","key":"10236_CR8","doi-asserted-by":"publisher","first-page":"2723","DOI":"10.1016\/j.sigpro.2012.04.023","volume":"92","author":"YM Costa","year":"2012","unstructured":"Costa YM, Oliveira L, Koerich AL, Gouyon F, Martins J (2012) Music genre classification using lbp textural features. Signal Process 92 (11):2723\u20132737","journal-title":"Signal Process"},{"key":"10236_CR9","doi-asserted-by":"crossref","unstructured":"Dai J, Liang S, Xue W, Ni C, Liu W (2016) Long short-term memory recurrent neural network based segment features for music genre classification. In: 2016 10Th international symposium on chinese spoken language processing, ISCSLP, IEEE, pp 1\u20135","DOI":"10.1109\/ISCSLP.2016.7918369"},{"key":"10236_CR10","unstructured":"Dev V (2020) Javascript hls client using media source extension. https:\/\/github.com\/video-dev\/hls.js\/"},{"key":"10236_CR11","unstructured":"Developer N (2018) Jetson tx2 module. https:\/\/developer.nvidia.com\/"},{"key":"10236_CR12","unstructured":"FFmpeg (2020) Ffmpeg github page. https:\/\/github.com\/FFmpeg\/FFmpeg"},{"key":"10236_CR13","doi-asserted-by":"crossref","unstructured":"Gygli M, Soleymani M (2016) Analyzing and predicting gif interestingness. In: Proceedings of the 24th ACM international conference on multimedia, pp 122\u2013126","DOI":"10.1145\/2964284.2967195"},{"key":"10236_CR14","doi-asserted-by":"crossref","unstructured":"Gygli M, Song Y, Cao L (2016) Video2gif: Automatic generation of animated gifs from video. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1001\u20131009","DOI":"10.1109\/CVPR.2016.114"},{"key":"10236_CR15","unstructured":"Hamel P, Eck D (2010) Learning features from music audio with deep belief networks. In: ISMIR, Utrecht, The Netherlands, vol 10, pp 339\u2013344"},{"key":"10236_CR16","unstructured":"Hauge M (2017) The five key turning points of all successful movie scripts. http:\/\/www.movieoutline.com\/"},{"issue":"3","key":"10236_CR17","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1109\/30.320795","volume":"40","author":"R Hopkins","year":"1994","unstructured":"Hopkins R (1994) Digital terrestrial hdtv for north america: the grand alliance hdtv system. IEEE Trans Consum Electron 40(3):185\u2013198","journal-title":"IEEE Trans Consum Electron"},{"key":"10236_CR18","doi-asserted-by":"crossref","unstructured":"Jou B, Bhattacharya S, Chang SF (2014) Predicting viewer perceived emotions in animated gifs. In: Proceedings of the 22nd ACM international conference on multimedia, pp 213\u2013216","DOI":"10.1145\/2647868.2656408"},{"key":"10236_CR19","doi-asserted-by":"crossref","unstructured":"Kim J W, Salamon J, Li P, Bello JP (2018) Crepe: A convolutional representation for pitch estimation. In: 2018 IEEE international conference on acoustics, speech and signal processing, ICASSP, IEEE, pp 161\u2013165","DOI":"10.1109\/ICASSP.2018.8461329"},{"key":"10236_CR20","unstructured":"Li TL, Chan AB, Chun AH (2010) Automatic musical pattern feature extraction using convolutional neural network. Genre 10:1x1"},{"key":"10236_CR21","doi-asserted-by":"crossref","unstructured":"Li Y, Xu G, Ge J, Liu P, Fu X (2020) Energy-efficient resource allocation for application including dependent tasks in mobile edge computing. KSII Transactions on Internet & Information Systems 14(6)","DOI":"10.3837\/tiis.2020.06.006"},{"issue":"4","key":"10236_CR22","doi-asserted-by":"publisher","first-page":"1098","DOI":"10.1109\/TMM.2019.2936805","volume":"22","author":"T Liu","year":"2020","unstructured":"Liu T, Wan J, Dai X, Liu F, You Q, Luo J (2020) Sentiment recognition for short annotated gifs using visual-textual fusion. IEEE Trans Multimedia 22(4):1098\u20131110","journal-title":"IEEE Trans Multimedia"},{"key":"10236_CR23","unstructured":"Loshchilov I, Hutter F (2018) Fixing weight decay regularization in adam. arXiv:171105101"},{"key":"10236_CR24","doi-asserted-by":"crossref","unstructured":"McFee B, Raffel C, Liang D, Ellis DP, McVicar M, Battenberg E, Nieto O (2015) librosa: Audio and music signal analysis in python. In: Proceedings of the 14th python in science conference, vol 8","DOI":"10.25080\/Majora-7b98e3ed-003"},{"key":"10236_CR25","doi-asserted-by":"crossref","unstructured":"Mu R, Zeng X (2020) Auxiliary stacked denoising autoencoder based collaborative filtering recommendation. KSII Transactions on Internet & Information Systems 14(6)","DOI":"10.3837\/tiis.2020.06.001"},{"key":"10236_CR26","doi-asserted-by":"publisher","first-page":"60417","DOI":"10.1109\/ACCESS.2020.2982992","volume":"8","author":"G Mujtaba","year":"2020","unstructured":"Mujtaba G, Ryu E S (2020) Client-driven personalized trailer framework using thumbnail containers. IEEE Access 8:60417\u201360427","journal-title":"IEEE Access"},{"key":"10236_CR27","unstructured":"Mujtaba G, Kim S, Park E, Kim S, Ryu J, Ryu ES (2019) Client-driven animated keyframe generation system using music analysis. In: Proceedings of the Korean society of broadcast engineers conference, The Korean Institute of Broadcast and Media Engineers, pp 173\u2013175"},{"issue":"4","key":"10236_CR28","doi-asserted-by":"publisher","first-page":"2023","DOI":"10.1007\/s11277-018-5920-1","volume":"106","author":"G Mujtaba","year":"2019","unstructured":"Mujtaba G, Tahir M, Soomro MH (2019) Energy efficient data encryption techniques in smartphones. Wireless Personal Communications 106(4):2023\u20132035","journal-title":"Wireless Personal Communications"},{"key":"10236_CR29","doi-asserted-by":"crossref","unstructured":"O Leary M (2019) Iis iis iis and modsecurity. In: Cyber Operations, Springer, pp 789\u2013819","DOI":"10.1007\/978-1-4842-4294-0_15"},{"issue":"2","key":"10236_CR30","first-page":"978","volume":"13","author":"S Qiu","year":"2019","unstructured":"Qiu S, Xu G, Ahmad H, Xu G, Qiu X, Xu H (2019) An improved lightweight two-factor authentication and key agreement protocol with dynamic identity based on elliptic curve cryptography. TIIS 13(2):978\u20131002","journal-title":"TIIS"},{"key":"10236_CR31","unstructured":"Redwood Shores C (2019) Giphy selects oracle data cloud to measure viewability across billions of gifs. https:\/\/www.oracle.com\/"},{"key":"10236_CR32","unstructured":"Robert J (2020) Pydub library. https:\/\/github.com\/jiaaro\/"},{"issue":"4","key":"10236_CR33","doi-asserted-by":"publisher","first-page":"1652","DOI":"10.1109\/TCE.2011.6131138","volume":"57","author":"ES Ryu","year":"2011","unstructured":"Ryu ES, Jayant N (2011) Home gateway for three-screen tv using h. 264 svc and raptor fec. IEEE Trans Consum Electron 57(4):1652\u20131660","journal-title":"IEEE Trans Consum Electron"},{"issue":"3","key":"10236_CR34","doi-asserted-by":"publisher","first-page":"319","DOI":"10.1007\/s11042-007-0166-z","volume":"37","author":"ES Ryu","year":"2008","unstructured":"Ryu ES, Yoo C (2008) Towards building large scale live media streaming framework for a u-city. Multimedia Tools and Applications 37(3):319\u2013338","journal-title":"Multimedia Tools and Applications"},{"key":"10236_CR35","unstructured":"Schindler A, Lidy T, Rauber A (2016) Comparing shallow versus deep neural network architectures for automatic music genre classification. In: FMT, pp 17\u201321"},{"key":"10236_CR36","doi-asserted-by":"crossref","unstructured":"Senac C, Pellegrini T, Mouret F, Pinquier J (2017) Music feature maps with convolutional neural networks for music genre classification. In: Proceedings of the 15th international workshop on content-based multimedia indexing, pp 1\u20135","DOI":"10.1145\/3095713.3095733"},{"key":"10236_CR37","doi-asserted-by":"crossref","unstructured":"Sigtia S, Dixon S 2014 Improved music feature learning with deep neural networks. In: 2014 IEEE International conference on acoustics, speech and signal processing, ICASSP, IEEE, pp 6959\u20136963","DOI":"10.1109\/ICASSP.2014.6854949"},{"issue":"5","key":"10236_CR38","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1109\/TSA.2002.800560","volume":"10","author":"G Tzanetakis","year":"2002","unstructured":"Tzanetakis G, Cook P (2002) Musical genre classification of audio signals. IEEE Trans Speech Audio Process 10(5):293\u2013302","journal-title":"IEEE Trans Speech Audio Process"},{"key":"10236_CR39","unstructured":"Van den Oord A, Dieleman S, Schrauwen B (2013) Deep content-based music recommendation. In: Advances in neural information processing systems, pp 2643\u20132651"},{"key":"10236_CR40","unstructured":"Vernallis C et al (2004) Experiencing music video: Aesthetics and cultural context. Columbia University Press"},{"issue":"1","key":"10236_CR41","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2801127","volume":"12","author":"MJ Wu","year":"2015","unstructured":"Wu MJ, Jang JSR (2015) Combining acoustic and multilevel visual features for music genre classification. ACM Transactions on Multimedia Computing Communications, and Applications (TOMM) 12(1):1\u201317","journal-title":"ACM Transactions on Multimedia Computing Communications, and Applications (TOMM)"},{"key":"10236_CR42","doi-asserted-by":"crossref","unstructured":"Zhang C, Evangelopoulos G, Voinea S, Rosasco L, Poggio T (2014) A deep representation for invariance and music classification. In: 2014 IEEE International conference on acoustics, speech and signal processing, ICASSP, IEEE, pp 6984\u20136988","DOI":"10.1109\/ICASSP.2014.6854954"},{"key":"10236_CR43","doi-asserted-by":"crossref","unstructured":"Zhang C, Wu D, Ao L, Wang M, Cai Y (2020) Social-aware collaborative caching based on user preferences for d2d content sharing. KSII Transactions on Internet & Information Systems 14(3)","DOI":"10.3837\/tiis.2020.03.009"},{"key":"10236_CR44","doi-asserted-by":"crossref","unstructured":"Zhang P, Zheng X, Zhang W, Li S, Qian S, He W, Zhang S, Wang Z (2015) A deep neural network for modeling music. In: Proceedings of the 5th ACM on international conference on multimedia retrieval, pp 379\u2013386","DOI":"10.1145\/2671188.2749367"},{"key":"10236_CR45","doi-asserted-by":"crossref","unstructured":"Zhang W, Lei W, Xu X, Xing X (2016) Improved music genre classification with convolutional neural networks. In: Interspeech, pp 3304\u20133308","DOI":"10.21437\/Interspeech.2016-1236"},{"key":"10236_CR46","doi-asserted-by":"crossref","unstructured":"Zhou Y, Song Y, Berg TL (2018) Image2gif: Generating cinemagraphs using recurrent deep q-networks. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE, pp 170\u2013178","DOI":"10.1109\/WACV.2018.00025"},{"key":"10236_CR47","unstructured":"Zurawski R (2004) The hypertext transfer protocol and uniform resource identifier. In: The industrial information technology handbook, CRC Press, pp 456\u2013478"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-020-10236-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-020-10236-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-020-10236-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,11,30]],"date-time":"2021-11-30T17:33:35Z","timestamp":1638293615000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-020-10236-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,2,12]]},"references-count":47,"journal-issue":{"issue":"28-29","published-print":{"date-parts":[[2021,11]]}},"alternative-id":["10236"],"URL":"https:\/\/doi.org\/10.1007\/s11042-020-10236-6","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,2,12]]},"assertion":[{"value":"4 July 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 September 2020","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 December 2020","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 February 2021","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}