{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,15]],"date-time":"2025-10-15T00:40:48Z","timestamp":1760488848880,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":36,"publisher":"ACM","funder":[{"name":"Swiss National Science Foundation","award":["209465"],"award-info":[{"award-number":["209465"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,8,27]]},"DOI":"10.1145\/3704268.3742704","type":"proceedings-article","created":{"date-parts":[[2025,8,27]],"date-time":"2025-08-27T16:44:47Z","timestamp":1756313087000},"page":"1-10","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Mining a Century of Swiss Trademark Data"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-5799-9015","authenticated-orcid":false,"given":"Daniel","family":"Travaglia","sequence":"first","affiliation":[{"name":"AnswerConsulting SA, Mendrisio, Switzerland and USI, Lugano"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-9682-6489","authenticated-orcid":false,"given":"Jesper","family":"Findahl","sequence":"additional","affiliation":[{"name":"Software Institute - USI, Lugano, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-3765-0871","authenticated-orcid":false,"given":"Marco","family":"D'Ambros","sequence":"additional","affiliation":[{"name":"Software Institute - USI, Lugano, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8426-5676","authenticated-orcid":false,"given":"Andrea","family":"Mocci","sequence":"additional","affiliation":[{"name":"Software Institute - USI, Lugano, Switzerland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7682-0521","authenticated-orcid":false,"given":"Raphael","family":"Parchet","sequence":"additional","affiliation":[{"name":"Institute of Economics - USI, Lugano, Switzerland"}]}],"member":"320","published-online":{"date-parts":[[2025,8,27]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Hangbo Bao Li Dong Songhao Piao and Furu Wei. 2022. BEiT: BERT Pre-Training of Image Transformers. (2022). arXiv: 2106.08254."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00104"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Alexis Conneau et al. 2020. Unsupervised Cross-lingual Representation Learning at Scale. (2020). arXiv: 1911.02116.","DOI":"10.18653\/v1\/2020.acl-main.747"},{"key":"e_1_3_2_1_4_1","volume-title":"Words: Transformers for Image Recognition at Scale.","author":"Alexey Dosovitskiy","year":"2021","unstructured":"Alexey Dosovitskiy et al. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. (2021). arXiv: 2010.11929."},{"key":"e_1_3_2_1_5_1","volume-title":"Christopher KI Williams, John Winn, and Andrew Zisserman.","author":"Everingham Mark","year":"2010","unstructured":"Mark Everingham, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. 2010. The pascal visual object classes (voc) challenge. International journal of computer vision, 88, 303--338."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/JCDL.2019.00057"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_8_1","unstructured":"Historical Statistics of Switzerland (HSSO). 2023. Handelsregister: eintragungen l\u00f6schungen und eingetragene firmen nach rechtsform am jahresende 1883-1992. https:\/\/hsso.ch\/2012\/s\/1."},{"key":"e_1_3_2_1_9_1","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). (July","author":"Huang Gao","year":"2017","unstructured":"Gao Huang, Zhuang Liu, Laurens van der Maaten, and Kilian Q. Weinberger. 2017. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). (July 2017)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3548112"},{"key":"e_1_3_2_1_11_1","volume-title":"Tel Aviv, Israel","author":"Geewook","year":"2022","unstructured":"Geewook Kim et al. 2022. OCR-free document understanding transformer. In Computer Vision-ECCV 2022:17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XXVIII. Springer, 498--517."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3547911"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.324"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/1390749.1390753"},{"key":"e_1_3_2_1_17_1","unstructured":"Microsoft. 2023. Azure custom vision. https:\/\/learn.microsoft.com\/en-us\/azure\/ai-services\/custom-vision-service\/overview."},{"key":"e_1_3_2_1_18_1","unstructured":"Microsoft. 2024. Compare custom vision to alternative Azure services. https:\/\/learn.microsoft.com\/en-us\/azure\/ai-services\/custom-vision-service\/concepts\/compare-alternatives."},{"key":"e_1_3_2_1_19_1","unstructured":"Microsoft. 2023. Project florence (AI). https:\/\/www.microsoft.com\/en-us\/research\/project\/projectflorence\/."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","unstructured":"Matthias Minderer Alexey Gritsenko and Neil Houlsby. 2023. Scaling open-vocabulary object detection. (2023). arXiv: https:\/\/doi.org\/10.48550\/arXiv.2306.09683.","DOI":"10.48550\/arXiv.2306.09683"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.156468"},{"key":"e_1_3_2_1_22_1","unstructured":"OpenAI et al. 2024. GPT-4 technical report. (2024). arXiv: 2303.08774."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.1979.4310076"},{"key":"e_1_3_2_1_24_1","volume-title":"Workshop on Document Intelligence at NeurIPS","author":"Park Seunghyun","year":"2019","unstructured":"Seunghyun Park, Seung Shin, Bado Lee, Junyeop Lee, Jaeheung Surh, Minjoon Seo, and Hwalsuk Lee. 2019. Cord: a consolidated receipt dataset for post-OCR parsing. In Workshop on Document Intelligence at NeurIPS 2019."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.91"},{"key":"e_1_3_2_1_26_1","unstructured":"Swiss Federal Institute of Intellectual Property (IPI). 2023. Trademarks database. https:\/\/database.ipi.ch\/database- client\/search\/query\/trademarks?lang=en."},{"key":"e_1_3_2_1_27_1","unstructured":"Swiss National Library NL. 2023. e-Helvetica. digital collections nl and web archive switzerland. https:\/\/www.e-helvetica.nb.admin.ch."},{"key":"e_1_3_2_1_28_1","volume-title":"International conference on machine learning. PMLR, 10347--10357","author":"Touvron Hugo","year":"2021","unstructured":"Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herv\u00e9 J\u00e9gou. 2021. Training data-efficient image transformers & distillation through attention. In International conference on machine learning. PMLR, 10347--10357."},{"key":"e_1_3_2_1_29_1","unstructured":"Daniel Travaglia. 2023. Donut for trademark registration events. https:\/\/huggingface.co\/spaces\/reflex-project\/donut-finetuned-sogc-trademarks."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","unstructured":"Junfeng Wu Yi Jiang Qihao Liu Zehuan Yuan Xiang Bai and Song Bai. 2023. General object foundation model for images and videos at scale. (2023). eprint: https:\/\/doi.org\/10.48550\/arXiv.2312.09158.","DOI":"10.48550\/arXiv.2312.09158"},{"key":"e_1_3_2_1_31_1","unstructured":"Yang Xu et al. 2022. Layoutlmv2: multi-modal pre-training for visually-rich document understanding. (2022). arXiv: 2012.14740."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3403172"},{"key":"e_1_3_2_1_33_1","unstructured":"Yiheng Xu Tengchao Lv Lei Cui Guoxin Wang Yijuan Lu Dinei Florencio Cha Zhang and Furu Wei. 2021. LayoutXLM: multimodal pre-training for multilingual visually-rich document understanding. arXiv preprint arXiv:2104.08836."},{"key":"e_1_3_2_1_34_1","unstructured":"Vikas Yadav and Steven Bethard. 2019. A survey on recent advances in named entity recognition from deep learning models. arXiv preprint arXiv:1910.11470."},{"volume-title":"Computer Vision-ECCV 2014: 13th European Conference","author":"Zhang Ning","key":"e_1_3_2_1_35_1","unstructured":"Ning Zhang, Jeff Donahue, Ross Girshick, and Trevor Darrell. 2014. Part-based R-CNNs for fine-grained category detection. In Computer Vision-ECCV 2014: 13th European Conference. Springer, 834--849."},{"key":"e_1_3_2_1_36_1","volume-title":"Publaynet: largest dataset ever for document layout analysis. (2019). arXiv","author":"Zhong Xu","year":"1908","unstructured":"Xu Zhong, Jianbin Tang, and Antonio Jimeno Yepes. 2019. Publaynet: largest dataset ever for document layout analysis. (2019). arXiv: 1908.07836."}],"event":{"name":"DocEng '25: ACM Symposium on Document Engineering 2025","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Nottingham United Kingdom","acronym":"DocEng '25"},"container-title":["Proceedings of the 2025 ACM Symposium on Document Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3704268.3742704","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T18:25:44Z","timestamp":1760466344000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3704268.3742704"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,8,27]]},"references-count":36,"alternative-id":["10.1145\/3704268.3742704","10.1145\/3704268"],"URL":"https:\/\/doi.org\/10.1145\/3704268.3742704","relation":{},"subject":[],"published":{"date-parts":[[2025,8,27]]},"assertion":[{"value":"2025-08-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}