DOI QR코드

DOI QR Code

딥러닝 이미지 인식 기술을 활용한 소고기 등심 세부 부위 분류

Deep Learning based Image Recognition Models for Beef Sirloin Classification

  • 한준희 (동아대학교 산업경영공학과) ;
  • 정성훈 (동아대학교 산업경영공학과) ;
  • 박경수 (부산대학교 경영학과) ;
  • 유태선 (부경대학교 시스템경영공학부)
  • Han, Jun-Hee (Departement of Industrial & Management Systems Engineering, Dong-A University) ;
  • Jung, Sung-Hun (Departement of Industrial & Management Systems Engineering, Dong-A University) ;
  • Park, Kyungsu (Department of Business Administration, Pusan National University) ;
  • Yu, Tae-Sun (Division of Systems Management and Engineering, Pukyong National University)
  • 투고 : 2021.06.17
  • 심사 : 2021.08.09
  • 발행 : 2021.09.30

초록

This research examines deep learning based image recognition models for beef sirloin classification. The sirloin of beef can be classified as the upper sirloin, the lower sirloin, and the ribeye, whereas during the distribution process they are often simply unified into the sirloin region. In this work, for detailed classification of beef sirloin regions we develop a model that can learn image information in a reasonable computation time using the MobileNet algorithm. In addition, to increase the accuracy of the model we introduce data augmentation methods as well, which amplifies the image data collected during the distribution process. This data augmentation enables to consider a larger size of training data set by which the accuracy of the model can be significantly improved. The data generated during the data proliferation process was tested using the MobileNet algorithm, where the test data set was obtained from the distribution processes in the real-world practice. Through the computational experiences we confirm that the accuracy of the suggested model is up to 83%. We expect that the classification model of this study can contribute to providing a more accurate and detailed information exchange between suppliers and consumers during the distribution process of beef sirloin.

키워드

과제정보

This work was supported by the Pukyong National University Research Fund in 2020(C-D-2020-0842). This work was also supported in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education under Grant 2020R1I1A3073672 and 2020R1G1A1099829. This research was also supported by the Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTIE: Ministry of Trade Industry and Energy; Grant No. N0002429).

참고문헌

  1. Bagherinezhad, H., Horton, M., Rastegari, M., and Farhadi, A., Label Refinery: Improving ImageNet Classification through Label Progression, arXiv, 2018, arXiv:1805.02641.
  2. Carney, M., Webster, B., Alvardo, I., Phillips, K., Howell, N., Griffith, J., Jongejan, J., Pitaru, A., and Chen, A., Teachable Machine: Approachable Web-Based Tool for Exploring Machine Learning Classification, Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, 2020, New York, USA, pp. 1-8.
  3. Cho, S.H., Park, B.Y., Byun, J.S., Kim, J.H., Ahn, J.N., and Yun, S.G., Visual Evaluation Factors of Pork Loin and Korean Consumer's Preference Choice, Korean Society of Animal Sciences and Technology, 2004, Vol. 46, No. 3, pp. 415-426.
  4. Choi, S., Hwang, H., Kim, J.H., Han, N.Y., Ko, M.J., and Cho, S.H., Quantization and Calibration of Color Information From Machine Vision System for Beef Color Grading, Journal of biocystems Engineering, 2007, Vol. 32, No. 3, pp. 160-165. https://doi.org/10.5307/JBE.2007.32.3.160
  5. Cubuk, E.D., Zoph, B., Mane, D., Vasudevan, V. and Le, Q.V., AutoAugment: Learning Augmentation Strategies From Data, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, Long Beach, USA, pp. 113-123.
  6. Efros, A.A. and Freeman, T.W., Image quilting for texture synthesis and transfer, Proceedings of the 28th annual conference on Computer graphics and interactive techniques, 2001, New York, United States, pp. 341-346.
  7. Fukushima, K., Miyake, S., and Ito, T., Neocognitron: A Neural Network Model for a Mechanism of Visual Pattern Recognition, IEEE Transactions on Systems, Man, and Cybernetics, 1983, Vol. SMC-13, No.5, pp. 826-834. https://doi.org/10.1109/TSMC.1983.6313076
  8. Gatys, A.L., Ecker, S.A., and Bethge, M. Image Style Transfer Using Convolutional Neural Networks, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, Las Vegas, USA, pp. 2414-2423.
  9. He, K., Zhang, X., Ren, S., and Sun, J., Deep Residual Learning for Image Recognition, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, Las Vegas, USA, pp. 770-778.
  10. Jang, A.R., Kim, H.J., and Kim, M.B., Deep Learning-based Analysis of Meat Freshness Measurement, The Korean Society of Broad Engineers, 2020, Vol. 25, No. 3, pp. 418-427.
  11. Jeong, D.W., Kim, D.K., and Ren C., A Survey of Deep Learning in Agriculture: Techniques and Their Applications, JIPS(Journal of Information Processing Systems), 2020, Vol. 16, No.5, pp. 1015-1033.
  12. Jin, J.H., Dundar, A., and Culurciello, E., Flattened Convolutional Neural Networks for Feedforward Acceleration, arXiv , 2015, arXiv:1412.5474.
  13. Kang, J.O., Choi, D.Y., Oh, H.R., and Kim, G.H., Comparison of Physico - chemical Characteristics of the Meat Quality Grades in Hanwoo Beef and Imported Beef From Several Countries, Journal of Animal Science and Technology, 1999, Vol. 41, pp. 555-562.
  14. Kim, Y.G., Yu, Y.M., Kim, J.H., and Ahn, J.N., Theme Common Sense of Meat, Dairy Science Division, National Institute of Animal Science Munsung, 2007, pp. 178-182.
  15. Kim, J.W., Pyo, H.A., Ha, J.W., Lee, C.G., and Kim, J.H., Deep learning algorithms and applications, Communications of the Korean Institute of Information Scientists and Engineers, 2015, Vol. 33, No.8, pp. 25-31.
  16. Krizhevsky, A., Sutskever, I., and Hinton, G.E., ImageNet Classification with Deep Convolutional Neural Networks, Communications of the ACM, 2017, Vol. 60, No. 6, pp. 84-90. https://doi.org/10.1145/3065386
  17. Lawrence, S., Giles, C.L., Tsoi, A.C., and Back, A.D., Face recognition: A convolutional neural-network approach, IEEE Transactions on Neural Networks, 1997, Vol. 8, no. 1, pp. 98-113. https://doi.org/10.1109/72.554195
  18. LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P., Gradient-based learning applied to document recognition, Proceeding of the IEEE, 1998, Vol. 86, No. 11, pp. 2278-2324. https://doi.org/10.1109/5.726791
  19. Lee, J.H., Lee, B.H., and Sin, Y.G., A study on evaluation of the meat safety, Korean Journal of Agricultural Management and Policy, 2005, Vol. 32, No. 4, pp. 728-745.
  20. Lee, M.G., A Multi-Layer Perceptron for Color Index based Vegetation Segmentation, Journal of Society of Korea Industrial and Systems Engineering, 2020, Vol. 43, No. 1, pp. 16-25. https://doi.org/10.11627/jkise.2020.43.1.016
  21. Ministry of Agriculture, Food and Rural Affairs notice No.2007-82, https://www.law.go.kr/LSW/admRulInfoP.do?admRulSeq=2894.
  22. Ministry of Food and Drug Safety notice No.2014-116, https://www.law.go.kr/admRulLsInfoP.do?chrClsCd=010202&admRulSeq=2100000184120.
  23. Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., and Lee, H., Generative Adversarial Text to Image Synthesis, Proceedings of The 33rd International Conference on Machine Learning, 2016, New York City, USA, pp. 1060-1069.
  24. Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg A.C., and Fei-Fei, L., ImageNet Large Scale Visual Recognition Challenge, International Journal of Computer Vision, 2015, Vol. 115, pp. 211-252. https://doi.org/10.1007/s11263-015-0816-y
  25. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A. and Chen, L.C., MobileNetV2: Inverted Residuals and Linear Bottlenecks, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018, Salt Lake City, USA, pp. 4510-4520.
  26. Shiranita, K., Miyajima, T., and Takiyama, R. Determination of meat quality by texture analysis, Pattern Recognition Letters, 1998, Vol. 19, No.14, pp. 1319-1324. https://doi.org/10.1016/S0167-8655(98)00113-5
  27. Shorten, C. and Khoshgoftaar, T.M., A survey on Image Data Augmentation for Deep Learning, Journal of Big Data, 2019, Vol. 6, No. 60.
  28. Simonyan, K. and Zisserman, A., Very Deep Convolutional Networks for Large-Scale Image Recognition., 2014, arXiv, arXiv:1409.1556.
  29. Timsorn, K., Wongchoosuk, C., Wattuya, P., Promdaen, S., and Sittichat, S., Discrimination of chicken freshness using electronic nose combined with PCA and ANN, 2014 11th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON), 2014, Nakhon Ratchasima, Thailand, pp. 1-4.
  30. Wang, M., Liu, B., and Foroosh, H., Factorized Convolutional Neural Networks., 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), 2017, Venice, Italy, pp. 545-553.
  31. Zhong, Z., Zheng, L., Kang, G., Li, S. and Yang, Y., Random erasing data augmentation., In Proceedings of the AAAI Conference on Artificial Intelligence, 2020, New York, USA, pp. 13001-13008.