{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T10:21:08Z","timestamp":1743070868688,"version":"3.40.3"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031220241"},{"type":"electronic","value":"9783031220258"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-22025-8_8","type":"book-chapter","created":{"date-parts":[[2023,2,10]],"date-time":"2023-02-10T23:15:19Z","timestamp":1676070919000},"page":"99-119","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["PointPillars Backbone Type Selection for Fast and Accurate LiDAR Object Detection"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2034-0590","authenticated-orcid":false,"given":"Konrad","family":"Lis","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6798-4444","authenticated-orcid":false,"given":"Tomasz","family":"Kryjak","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,11]]},"reference":[{"key":"8_CR1","unstructured":"Dpuczdx8g for zynq ultrascale+ mpsocs product guide. Tech. Rep. PG338 (v3.4), Xilinx Inc. (2022). https:\/\/www.xilinx.com\/content\/dam\/xilinx\/support\/documentation\/ip_documentation\/dpu\/v3_4\/pg338-dpu.pdf"},{"key":"8_CR2","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: Optimal speed and accuracy of object detection (2020). https:\/\/arxiv.org\/abs\/2004.10934"},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Caesar, H., Bankiti, V., Lang, A.H., Vora, S., et\u00a0al.: nuscenes: A multimodal dataset for autonomous driving. arXiv preprint arXiv:1903.11027 (2019)","DOI":"10.1109\/CVPR42600.2020.01164"},{"key":"8_CR4","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep Learning with Depthwise Separable Convolutions (2016). https:\/\/arxiv.org\/abs\/1610.02357","DOI":"10.1109\/CVPR.2017.195"},{"key":"8_CR5","unstructured":"Contributors, M.: MMDetection3D: OpenMMLab next-generation platform for general 3D object detection (2020). https:\/\/github.com\/open-mmlab\/mmdetection3d"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: The kitti dataset. Int. J. Robot. Res. (IJRR) (2013)","DOI":"10.1177\/0278364913491297"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Gholami, A., Kwon, K., Wu, B., Tai, Z., et al.: Squeezenext: Hardware-aware Neural Network Design (2018). https:\/\/arxiv.org\/abs\/1803.10615","DOI":"10.1109\/CVPRW.2018.00215"},{"key":"8_CR8","doi-asserted-by":"publisher","unstructured":"Gustafson, J.L.: Amdahl\u2019s Law, pp. 53\u201360. Springer US, Boston, MA (2011). https:\/\/doi.org\/10.1007\/978-0-387-09766-4_77","DOI":"10.1007\/978-0-387-09766-4_77"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep Residual Learning for Image Recognition (2015). https:\/\/arxiv.org\/abs\/1512.03385","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR10","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chen, B., Wang, W., et\u00a0al.: Searching for mobilenetv3. In: 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"8_CR11","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., et al.: Mobilenets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017). https:\/\/arxiv.org\/abs\/1704.04861"},{"key":"8_CR12","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., et al..: Squeezenet: Alexnet-level Accuracy with 50x Fewer Parameters and $$<$$0.5mb Model Size (2016). https:\/\/arxiv.org\/abs\/1602.07360"},{"key":"8_CR13","unstructured":"KITTI: Kitti Database Website. http:\/\/www.cvlibs.net\/datasets\/kitti\/ (2020). Last accessed 15 May 2020"},{"key":"8_CR14","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q. (eds.) Advances in Neural Information Processing Systems, vol.\u00a025. Curran Associates, Inc. (2012). https:\/\/proceedings.neurips.cc\/paper\/2012\/file\/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf"},{"key":"8_CR15","doi-asserted-by":"crossref","unstructured":"Lang, A.H., Vora, S., Caesar, H., Zhou, L., et\u00a0al.: Pointpillars: Fast encoders for object detection from point clouds. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12689\u201312697 (2019)","DOI":"10.1109\/CVPR.2019.01298"},{"key":"8_CR16","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1007\/978-3-319-46448-0_2","volume-title":"Computer Vision-ECCV 2016","author":"W Liu","year":"2016","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., et al.: Ssd: Single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) Computer Vision-ECCV 2016, pp. 21\u201337. Springer International Publishing, Cham (2016)"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.T., Sun, J.: Shufflenet v2: Practical Guidelines for Efficient Cnn Architecture Design (2018). https:\/\/arxiv.org\/abs\/1807.11164","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"8_CR18","doi-asserted-by":"crossref","unstructured":"Meyer, G.P., Laddha, A., Kee, E., Vallespi-Gonzalez, C., et al.: Lasernet: An Efficient Probabilistic 3d Object Detector for Autonomous Driving (2019). https:\/\/arxiv.org\/abs\/1903.08701","DOI":"10.1109\/CVPR.2019.01296"},{"key":"8_CR19","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: Pointnet++: Deep hierarchical feature learning on point sets in a metric space. CoRR abs\/1706.02413 (2017). http:\/\/arxiv.org\/abs\/1706.02413"},{"key":"8_CR20","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You Only Look Once: Unified, Real-time Object Detection (2015). https:\/\/arxiv.org\/abs\/1506.02640","DOI":"10.1109\/CVPR.2016.91"},{"key":"8_CR21","doi-asserted-by":"crossref","unstructured":"Redmon, J., Farhadi, A.: Yolo9000: Better, Faster, Stronger (2016). https:\/\/arxiv.org\/abs\/1612.08242","DOI":"10.1109\/CVPR.2017.690"},{"key":"8_CR22","unstructured":"Redmon, J., Farhadi, A.: Yolov3: An Incremental Improvement (2018). https:\/\/arxiv.org\/abs\/1804.02767"},{"key":"8_CR23","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.C.: Mobilenetv2: Inverted Residuals and Linear Bottlenecks (2018). https:\/\/arxiv.org\/abs\/1801.04381","DOI":"10.1109\/CVPR.2018.00474"},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Shi, S., Guo, C., Jiang, L., Wang, Z., et\u00a0al.: Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. ArXiv abs\/1912.13192 (2019)","DOI":"10.1109\/CVPR42600.2020.01054"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Shi, S., Wang, X., Li, H.: Pointrcnn: 3d Object Proposal Generation and Detection from Point Cloud (2018). https:\/\/arxiv.org\/abs\/1812.04244","DOI":"10.1109\/CVPR.2019.00086"},{"key":"8_CR26","unstructured":"Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-scale Image Recognition (2014). https:\/\/arxiv.org\/abs\/1409.1556"},{"key":"8_CR27","volume-title":"Implementation of the pointpillars network for 3d object detection in reprogrammable heterogeneous devices using finn","author":"J Stanisz","year":"2021","unstructured":"Stanisz, J., Lis, K., Gorgon, M.: Implementation of the pointpillars network for 3d object detection in reprogrammable heterogeneous devices using finn. J. Sig. Process, Syst (2021)"},{"key":"8_CR28","doi-asserted-by":"crossref","unstructured":"Sun, P., Kretzschmar, H., Dotiwalla, X., Chouard, A., et al.: Scalability in Perception for Autonomous Driving: Waymo Open Dataset (2019). https:\/\/arxiv.org\/abs\/1912.04838","DOI":"10.1109\/CVPR42600.2020.00252"},{"key":"8_CR29","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Ioffe, S., Vanhoucke, V., Alemi, A.: Inception-v4, Inception-resnet and the Impact of Residual Connections on Learning (2016). https:\/\/arxiv.org\/abs\/1602.07261","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"8_CR30","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: Efficientdet: Scalable and Efficient Object Detection (2019). https:\/\/arxiv.org\/abs\/1911.09070","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"8_CR31","doi-asserted-by":"crossref","unstructured":"Vemparala, M.R., Singh, A., Mzid, A., Fasfous, N., et\u00a0al.: Pruning cnns for lidar-based perception in resource constrained environments. In: 2021 IEEE Intelligent Vehicles Symposium Workshops (IV Workshops), pp. 228\u2013235 (2021)","DOI":"10.1109\/IVWorkshops54471.2021.9669256"},{"key":"8_CR32","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: Scaled-yolov4: Scaling Cross Stage Partial Network (2020). https:\/\/arxiv.org\/abs\/2011.08036","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"8_CR33","unstructured":"Wang, C.Y., Yeh, I.H., Liao, H.Y.M.: You Only Learn One Representation: Unified Network for Multiple Tasks (2021). https:\/\/arxiv.org\/abs\/2105.04206"},{"key":"8_CR34","doi-asserted-by":"publisher","first-page":"22080","DOI":"10.1109\/ACCESS.2021.3055491","volume":"9","author":"LH Wen","year":"2021","unstructured":"Wen, L.H., Jo, K.H.: Fast and accurate 3d object detection for lidar-camera-based autonomous vehicles using one shared voxel-based backbone. IEEE Access 9, 22080\u201322089 (2021)","journal-title":"IEEE Access"},{"key":"8_CR35","doi-asserted-by":"crossref","unstructured":"Wu, X., Peng, L., Yang, H., Xie, L., et al.: Sparse Fuse Dense: Towards High Quality 3d Detection with Depth Completion (2022). https:\/\/arxiv.org\/abs\/2203.09780","DOI":"10.1109\/CVPR52688.2022.00534"},{"issue":"21","key":"8_CR36","doi-asserted-by":"publisher","first-page":"24505","DOI":"10.1109\/JSEN.2021.3114345","volume":"21","author":"Y Wu","year":"2021","unstructured":"Wu, Y., Zhang, S., Ogai, H., Inujima, H., et al.: Realtime single-shot refinement neural network with adaptive receptive field for 3d object detection from lidar point cloud. IEEE Sens. J. 21(21), 24505\u201324519 (2021)","journal-title":"IEEE Sens. J."},{"key":"8_CR37","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Dollr, P., Tu, Z., et al.: Aggregated Residual Transformations for Deep Neural Networks (2016). https:\/\/arxiv.org\/abs\/1611.05431","DOI":"10.1109\/CVPR.2017.634"},{"issue":"10","key":"8_CR38","doi-asserted-by":"publisher","first-page":"3337","DOI":"10.3390\/s18103337","volume":"18","author":"Y Yan","year":"2018","unstructured":"Yan, Y., Mao, Y., Li, B.: Second: Sparsely embedded convolutional detection. Sensors 18(10), 3337 (2018)","journal-title":"Sensors"},{"key":"8_CR39","doi-asserted-by":"crossref","unstructured":"Yin, T., Zhou, X., Kr\u00e4henb\u00fchl, P.: Center-based 3d Object Detection and Tracking (2020). https:\/\/arxiv.org\/abs\/2006.11275","DOI":"10.1109\/CVPR46437.2021.01161"},{"issue":"4","key":"8_CR40","doi-asserted-by":"publisher","first-page":"3434","DOI":"10.1109\/LRA.2018.2852843","volume":"3","author":"Y Zeng","year":"2018","unstructured":"Zeng, Y., Hu, Y., Liu, S., Ye, J., et al.: Rt3d: Real-time 3-d vehicle detection in lidar point cloud for autonomous driving. IEEE Robot. Autom. Lett. 3(4), 3434\u20133440 (2018)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"8_CR41","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: An Extremely Efficient Convolutional Neural Network for Mobile Devices (2017). https:\/\/arxiv.org\/abs\/1707.01083","DOI":"10.1109\/CVPR.2018.00716"},{"key":"8_CR42","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Tuzel, O.: Voxelnet: End-to-end learning for point cloud based 3d object detection. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4490\u20134499 (2018)","DOI":"10.1109\/CVPR.2018.00472"},{"key":"8_CR43","doi-asserted-by":"crossref","unstructured":"Zoph, B., Vasudevan, V., Shlens, J., Le, Q.V.: Learning Transferable Architectures for Scalable Image Recognition (2017). https:\/\/arxiv.org\/abs\/1707.07012","DOI":"10.1109\/CVPR.2018.00907"}],"container-title":["Lecture Notes in Networks and Systems","Computer Vision and Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-22025-8_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,11]],"date-time":"2023-02-11T00:31:20Z","timestamp":1676075480000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-22025-8_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031220241","9783031220258"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-22025-8_8","relation":{},"ISSN":["2367-3370","2367-3389"],"issn-type":[{"type":"print","value":"2367-3370"},{"type":"electronic","value":"2367-3389"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"11 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICCVG","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computer Vision and Graphics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Warsaw","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Poland","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iccvg2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iccvg.sggw.edu.pl","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}