Browse > Article
http://dx.doi.org/10.7734/COSEIK.2021.34.5.327

Flip Side of Artificial Intelligence Technologies: New Labor-Intensive Industry of the 21st Century  

Heo, Seokjae (Remodeling Research Center, Dankook University)
Na, Seunguk (Department of Architectural Engineering, Dankook University)
Han, Sehee (Department of Architectural Engineering, Dankook University)
Shin, Yoonsoo (Department of Architectural Engineering, Dankook University)
Lee, Sanghyun (Department of Architectural Engineering, Dankook University)
Publication Information
Journal of the Computational Structural Engineering Institute of Korea / v.34, no.5, 2021 , pp. 327-337 More about this Journal
Abstract
The paper acknowledges that many human resources are needed on the research and development (R&D) process of artificial intelligence (AI), and discusses on factors to consider on the current method of development. Enfin, in order to enhance efficiency of AI development, it seems possible through labour division of a few managers and numerous ordinary workers as a type of light industry. Thus, the research team names the development process of AI, which maximizes production efficiency by handling digital resources named 'data' with mechanical equipment called 'computer', as digital light industry of fourth industrial era. As experienced during the previous Industrial Revolution, if human resources are efficiently distributed and utilized, digital light industry would be able to expect progress no less than the second Industrial Revolution, and human resources development for this is considered urgent.
Keywords
digital light industry; fourth industrial era; AI; human resources development; work index; data waste;
Citations & Related Records
연도 인용수 순위
  • Reference
1 Kim, J.M., Hyeon, S.G., Chae, J.H., Do, M.S. (2019) Road Crack Detection based on Object Detection Algorithm using Unmanned Aerial Vehicle Image, J. Korea Inst. Intell. Trans. Syst., 18, pp.155~163.   DOI
2 Krizhevsky, A., Sutskever, I., Hinton, G.E. (2012) Imagenet Classification with Deep Convolutional Neural Networks, Adv. Neural Inf. Proc. Syst., 25, pp.1097~1105.
3 Lee, B.Y., Yi, S.T., Kim, J.K. (2007) Surface Crack Evaluation Method in Concrete Structures, J. Korean Soc. Nondestruct. Test., 27, pp.173~182.
4 Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollar, P.,Zitnick, C. L. (2014) Microsoft COCO: Common Objects in Context, European Conference on Computer Vision, Springer, pp.740~755.
5 Roh, B.S., Shin, W.H., Kim, I.D., Kim, S.W. (2021) Spatially Consistent Representation Learning, EEE/CVF Conference on Computer Vision and Pattern Recognition, pp.1144~1153.
6 Russell, B.C., Torralba, A., Murphy, K.P., Freeman, W.T. (2008) LabelMe: A Database and Web-Based Tool for Image Annotation, Int. J. Comput. Vis., 77(1-3), pp.157~173.   DOI
7 Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I. (2017) Attention is All You Need, 31st Conference on Neural Information Processing Systems(NIPS 2017), pp.5998~6008.
8 Na, S.G., Heo, S.J., Han, S.H. (2021) Construction Waste Reduction through Application of Different Structural Systems for the Slab in a Commercial Building: A South Korean Case, Appl. Sci., 11(3), p.5870.   DOI
9 Ren, S., He, K., Girshick , R., Sun, J. (2015) Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks, Adv. Neural Inf. Proc. Syst., 28, pp.91~99.
10 NIA (National Information Society Agency) (2020) 'NIA2020-032', https://www.nia.or.kr.
11 Lowe, D.G. (1999) Object Recognition from Local Scale-Invariant Features, Proceedings of the Seventh IEEE International Conference on Computer Vision, pp.1150~1157.
12 Baji, T. (2018) Evolution of the GPU Device widely used in AI and Massive Parallel Processing, IEEE 2nd Electron Devices Technology and Manufacturing Conference (EDTM), pp.7~9.
13 Acuna, D., Ling, H., Kar, A., Fidler, S. (2018) Efficient Interactive Annotation of Segmentation Datasets with Polygon-RNN++, IEEE Conference on Computer Vision and Pattern Recognition, pp.859~868.
14 Association, Telecommunications Technology (2021a) AI Dataset Construction Guidebook (Ministry of Science and ICT).
15 Association, Telecommunications Technology (2021b) AI Learning Data Quality Management Guidelines v1.0.
16 Felzenszwalb, P., McAllester, D., Ramanan, D. (2008) A Discriminatively Trained, Multiscale, Deformable Part Model, IEEE Conference on Computer Vision and Pattern Recognition, pp.1~8.
17 Girshick, R. (2015) Fast R-CNN, Proc. IEEE International Conference on Computer Vision, pp.1440~1448.
18 Girshick, R., Donahue, J., Darrell, T., Malik, J. (2014) Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation, IEEE Conference on Computer Vision and Pattern Recognition, pp.580~587.
19 Heo, S.J., Chunwei, Z., Yu, E.J. (2018) Response Simulation, Data Cleansing and Restoration of Dynamic and Static Measurements Based on Deep Learning Algorithms, Int. J. Concr. Struct. Mater., 12, pp.1~13.   DOI
20 He, K., Zhang, X., Ren, S., Sun, J. (2016) Deep Residual Learning for Image Recognition, IEEE Conference on Computer Vision and Pattern Recognition, pp.770~778.
21 Kim, B.S., Lee, J.H., Kang, J.W., Kim, E.S., Kim, H.J. (2021) HOTR: End-to-End Human-Object Interaction Detection with Transformers, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.74~83.
22 Dutta, A., Gupta, A., Zissermann, A. (2016) VGG Image Annotator (VIA), http://www.robots.ox.ac.uk/~vgg/software/via.
23 Redmon, J., Divvala, S., Girshick, R., Farhadi, A. (2016) You Only Look Once: Unified, Real-Time Object Detection, IEEE Conference on Computer Vision and Pattern Recognition, pp.779~788.
24 Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N. (2020) An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale, Computer Vision and Pattern Recognition (cs.CV), arXiv preprint arXiv:2010.11929.
25 Tan, M., Le, Q. (2019) Efficientnet: Rethinking Model Scaling for Convolutional Neural Networks, International Conference on Machine Learning, pp.6105~6114. PMLR.
26 Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q. (2017) Densely Connected Convolutional Networks, IEEE Conference on Computer Vision and Pattern Recognition, pp.4700~4708.
27 He, K., Gkioxari, G., Dollar, P., Girshick, R. (2017) Mask R-CNN, IEEE International Conference on Computer Vision, pp.2961~2969.
28 King, T.M., Arbon, J., Santiago, D., Adamo, D., Chin, W., Chin, W., Shanmugam, R. (2019) AI for Testing Today and Tomorrow: Industry Perspectives, IEEE International Conference On Artificial Intelligence Testing (AITest), pp.81~88.
29 Revaud, J., Heo, M., Rezende, R.S., You, C., Jeong, S.G. (2019) Did It Change? Learning to Detect Point-of-Interest Changes for Proactive Map Updates, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.4086~4095.
30 Simonyan, K., Zisserman, A. (2014) Very Deep Convolutional Networks for Large-Scale Image Recognition, Computer Vision and Pattern Recognition, arXiv:1409.1556.