ÀüÀÚ°øÇÐȸ ³í¹®Áö (Journal of The Institute of Electronics and Information Engineers)
Current Result Document : 1,794 / 1,795
ÇѱÛÁ¦¸ñ(Korean Title) |
ÀÓº£µðµå ½Ã½ºÅÛ¿¡¼ÀÇ °´Ã¼ ŽÁö ³×Æ®¿öÅ© Ãß·Ð °¡¼ÓÀ» À§ÇÑ ÇÊÅÍ °¡ÁöÄ¡±â ±â¹ý ¿¬±¸ |
¿µ¹®Á¦¸ñ(English Title) |
A Study on Filter Pruning Method for Accelerating Object Detection Network Inference in Embedded Systems |
ÀúÀÚ(Author) |
ÀüÁöÈÆ
±èÀç¸í
°Áø±¸
±è¿ë¿ì
Jihun Jeon
Jaemyung Kim
Jin-Ku Kang
Yongwoo Kim
|
¿ø¹®¼ö·Ïó(Citation) |
VOL 59 NO. 03 PP. 0069 ~ 0077 (2022. 03) |
Çѱ۳»¿ë (Korean Abstract) |
ÃÖ±Ù¿¡ ÄÄÇ»ÅÍ ºñÀü ºÐ¾ß¿¡¼ ¿ì¼öÇÑ ¼º´ÉÀ» ³ªÅ¸³»´Â CNN(Convolution neural network)ÀÌ °¢±¤À» ¹Þ°í ÀÖ´Ù. ÀÌ´Â µ¥ÀÌÅÍÀÇ Áõ°¡¿Í GPU¿Í °°Àº Çϵå¿þ¾î ¼º´ÉÀÇ Çâ»óÀ¸·Î °¡´ÉÇÏ°Ô µÇ¾úÁö¸¸ ³ôÀº ¼º´ÉÀ» À§ÇØ ³×Æ®¿öÅ©°¡ ±í°í ³Ð¾îÁö¸é¼ ÆÄ¶ó¹ÌÅÍ¿Í ¿¬»ê·®ÀÌ ±âÇϱ޼öÀûÀ¸·Î Áõ°¡ÇÏ¿´´Ù. µû¶ó¼ ¸Þ¸ð¸®, ¿¬»ê ¼º´É, Àü·Â »ç¿ëÀÌ Á¦ÇÑÀûÀÎ ÀÓº£µðµå ȯ°æ¿¡¼ Å« ³×Æ®¿öÅ©ÀÇ È°¿ëÀº ´õ¿í ¾î·Á¿öÁ³´Ù. ÀÌ·¯ÇÑ ¹®Á¦¸¦ ÇØ°áÇϱâ À§ÇØ CNN ¸ðµ¨ÀÇ Á¤È®µµ´Â À¯ÁöÇÑ Ã¤ Áß¿äÇÏÁö ¾ÊÀº ÆÄ¶ó¹ÌÅ͵éÀ» Á¦°ÅÇÏ´Â °¡ÁöÄ¡±â ±â¹ýÀÌ È°¹ßÈ÷ ¿¬±¸µÇ°í ÀÖ´Ù. ÇÏÁö¸¸ ´ëºÎºÐÀÇ °¡ÁöÄ¡±â ±â¹ýÀ» ´Ù·é ±âÁ¸ ¿¬±¸´Â ÆÄ¶ó¹ÌÅÍ Á¦°Å¿Í ÁÙ¾îµç ÆÄ¶ó¹ÌÅÍ¿¡ ÀÇÇØ °¨¼ÒµÈ Ç÷ӽº(FLOPs)¿¡ ´ëÇÑ °á°ú¸¦ º¸¿©ÁÖ¿´´Ù. º» ³í¹®¿¡¼´Â ÆÄ¶ó¹ÌÅÍ ¼ö¿Í ÇÔ²² Ç÷ӽº±îÁö ¿øÇÏ´Â ºñÀ²¸¸Å ÁÙÀÓÀ¸·Î½á °¡¼ÓÈµÈ Ãß·Ð ¼Óµµ¸¦ °®´Â ³×Æ®¿öÅ©¸¦ »ý¼ºÇÒ ¼ö ÀÖ´Â ÇÊÅÍ °¡ÁöÄ¡±â ±â¹ýÀ» Á¦¾ÈÇÑ´Ù. Á¦¾ÈÇÏ´Â ÇÊÅÍ °¡ÁöÄ¡±â ±â¹ý¿¡ ´ëÇÑ ¼º´ÉÀ» Æò°¡Çϱâ À§ÇØ VisDrone µ¥ÀÌÅÍ ¼¼Æ®¿Í YOLOv5¸¦ ÀÌ¿ëÇÏ¿´À¸¸ç, °¡ÁöÄ¡±â ÈÄ °æ·®ÈµÈ ³×Æ®¿öÅ©ÀÇ Ãß·Ð ¼Óµµ¸¦ NVIDIA Jetson Xavier NX Ç÷§ÆûÀ» ÀÌ¿ëÇÏ¿© ÃøÁ¤ÇÏ¿´´Ù. ÆÄ¶ó¹ÌÅÍ¿Í Ç÷ӽº¸¦ °¢°¢ 30%, 40%, 50%¾¿ °¡ÁöÄ¡±âÇÑ °á°ú mAP(0.5:0.95)´Â ±âÁØ °´Ã¼ ŽÁö ³×Æ®¿öÅ© ´ëºñ 0.6%, 0.9%, 1.2% °¨¼ÒÇÏ´Â ¹Ý¸é¿¡, Ãß·Ð ½Ã°£Àº °¢°¢ 16.60%, 25.79%, 30.72%°¡ Çâ»óµÈ °á°ú¸¦ È®ÀÎÇÏ¿´´Ù. |
¿µ¹®³»¿ë (English Abstract) |
Recently, convolutional neural networks (CNNs), which exhibit excellent performance in the field of computer vision, have been in the spotlight. This was made possible by the increase of data and the improvement of hardware performance such as GPU, but as the network became deeper and wider for high performance, parameters and computational amount increased exponentially. Therefore, it has become more difficult to utilize a large network in an embedded environment where memory, computational performance, and power usage are limited. To solve this problem, a pruning technique that removes insignificant parameters while maintaining the accuracy of the CNN model is being actively studied. However, existing studies dealing with most of the pruning techniques showed results for reduced parameters and FLOPs reduced by parameter removal. In this paper, we propose a filter pruning method that can create a network with accelerated inference speed by reducing the number of parameters together with the FLOPs by a desired ratio. To evaluate the performance of the proposed filter pruning method, VisDrone data set and YOLOv5 were used, and the inference speed of the lightweight network after pruning was measured using the NVIDIA Jetson Xavier NX platform. As a result of pruning the parameters and FLOPs by 30%, 40%, and 50%, respectively, the mAP (0.5:0.95) decreased 0.6%, 0.9%, and 1.2% compared to the reference object detection network, whereas inference time was confirmed to be improved by 16.60%, 25.79%, and 30.72%, respectively. |
Ű¿öµå(Keyword) |
Deep learning
Embedded system
Object detection
Filter pruning
FLOPs
|
¿ø¹® |
PDF ´Ù¿î·Îµå
|