• Title/Summary/Keyword: vision training

Search Result 421, Processing Time 0.028 seconds

Professional Mobility as a Factor of Professional Success of a Modern Specialist in the Conditions of Distance Learning

  • Semchuk, Bohdan;Havryliuk, Svitlana;Karnaukh, Lesia;Balakirieva, Viktoriia;Palshkova, Iryna;Leonova, Veronika;Bida, Olena
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.8
    • /
    • pp.260-268
    • /
    • 2022
  • The article considers the training, competitiveness of specialists, professional mobility, professionalism and competence of specialists in the context of distance learning. The advantages of distance learning are shown. The characteristic features of distance learning in the preparation of students and in the implementation of these technologies in the educational process of higher educational institutions are determined. Competitiveness, professional mobility, professionalism and competence of a specialist are qualities that determine a person's life and work success. Professional mobility is interpreted as a systemic quality of a specialist's personality, which includes a whole range of knowledge, skills, abilities, personal qualities, value orientations, and so on. The vision of mobility of specialists by foreign scientists is presented. It is noted that the classification of professional mobility presented in the article makes it possible to organize various movements from a single position, to present them as separate manifestations of the general process of professional and pedagogical mobility, to determine which type of mobility ensures the performance of certain social functions. It was found that mobility can be differentiated into differentiated and intergeneration. According to the subject, individual and group mobility are distinguished; according to the direction - internal and external. The classification of employees according to their attitude to mobility is shown, which can be divided into the following groups: actually mobile; potentially mobile; actually stable; potentially stable.

Deep Learning Methods for Recognition of Orchard Crops' Diseases

  • Sabitov, Baratbek;Biibsunova, Saltanat;Kashkaroeva, Altyn;Biibosunov, Bolotbek
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.10
    • /
    • pp.257-261
    • /
    • 2022
  • Diseases of agricultural plants in recent years have spread greatly across the regions of the Kyrgyz Republic and pose a serious threat to the yield of many crops. The consequences of it can greatly affect the food security for an entire country. Due to force majeure, abnormal cases in climatic conditions, the annual incomes of many farmers and agricultural producers can be destroyed locally. Along with this, the rapid detection of plant diseases also remains difficult in many parts of the regions due to the lack of necessary infrastructure. In this case, it is possible to pave the way for the diagnosis of diseases with the help of the latest achievements due to the possibilities of feedback from the farmer - developer in the formation and updating of the database of sick and healthy plants with the help of advances in computer vision, developing on the basis of machine and deep learning. Currently, model training is increasingly used already on publicly available datasets, i.e. it has become popular to build new models already on trained models. The latter is called as transfer training and is developing very quickly. Using a publicly available data set from PlantVillage, which consists of 54,306 or NewPlantVillage with a data volumed with 87,356 images of sick and healthy plant leaves collected under controlled conditions, it is possible to build a deep convolutional neural network to identify 14 types of crops and 26 diseases. At the same time, the trained model can achieve an accuracy of more than 99% on a specially selected test set.

Actor Doctor: Actors Visiting Hospitals for Care (액터 닥터: 병원으로 간 배우들)

  • Lee, Gang-Im
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.3
    • /
    • pp.229-238
    • /
    • 2022
  • Through the converging concept of 'an actor doctor,' this paper deals with an actor and its double, the emotional care giver. In the first place, actor doctors have appeared in the medical scene as hospice volunteers. But, as a developing project, the concept of 'an actor doctor' leads to a professional emotional specialist who works at hospitals during the day and performing at the stage at night. In our contemporary culture, our emotional life is threatened by various social and environmental stimuli. The role of 'an emotional specialist' in our everyday life and hospitals is very essential to our healthy life. Against the backgrounds of the high emotional intelligence and the systemic training, actors are qualified to embody this social vision. From the perspectives of performance studies and cognitive studies, this paper tries scientific explanations of the role, the function, and the techniques of an actor doctor. By doing so, this paper seeks to build a scientific foundation for an actor doctor as an emotional specialist.

Divide and Conquer Strategy for CNN Model in Facial Emotion Recognition based on Thermal Images (얼굴 열화상 기반 감정인식을 위한 CNN 학습전략)

  • Lee, Donghwan;Yoo, Jang-Hee
    • Journal of Software Assessment and Valuation
    • /
    • v.17 no.2
    • /
    • pp.1-10
    • /
    • 2021
  • The ability to recognize human emotions by computer vision is a very important task, with many potential applications. Therefore the demand for emotion recognition using not only RGB images but also thermal images is increasing. Compared to RGB images, thermal images has the advantage of being less affected by lighting conditions but require a more sophisticated recognition method with low-resolution sources. In this paper, we propose a Divide and Conquer-based CNN training strategy to improve the performance of facial thermal image-based emotion recognition. The proposed method first trains to classify difficult-to-classify similar emotion classes into the same class group by confusion matrix analysis and then divides and solves the problem so that the emotion group classified into the same class group is recognized again as actual emotions. In experiments, the proposed method has improved accuracy in all the tests than when recognizing all the presented emotions with a single CNN model.

Coating defect classification method for steel structures with vision-thermography imaging and zero-shot learning

  • Jun Lee;Kiyoung Kim;Hyeonjin Kim;Hoon Sohn
    • Smart Structures and Systems
    • /
    • v.33 no.1
    • /
    • pp.55-64
    • /
    • 2024
  • This paper proposes a fusion imaging-based coating-defect classification method for steel structures that uses zero-shot learning. In the proposed method, a halogen lamp generates heat energy on the coating surface of a steel structure, and the resulting heat responses are measured by an infrared (IR) camera, while photos of the coating surface are captured by a charge-coupled device (CCD) camera. The measured heat responses and visual images are then analyzed using zero-shot learning to classify the coating defects, and the estimated coating defects are visualized throughout the inspection surface of the steel structure. In contrast to older approaches to coating-defect classification that relied on visual inspection and were limited to surface defects, and older artificial neural network (ANN)-based methods that required large amounts of data for training and validation, the proposed method accurately classifies both internal and external defects and can classify coating defects for unobserved classes that are not included in the training. Additionally, the proposed model easily learns about additional classifying conditions, making it simple to add classes for problems of interest and field application. Based on the results of validation via field testing, the defect-type classification performance is improved 22.7% of accuracy by fusing visual and thermal imaging compared to using only a visual dataset. Furthermore, the classification accuracy of the proposed method on a test dataset with only trained classes is validated to be 100%. With word-embedding vectors for the labels of untrained classes, the classification accuracy of the proposed method is 86.4%.

Design of Vision-based Interaction Tool for 3D Interaction in Desktop Environment (데스크탑 환경에서의 3차원 상호작용을 위한 비전기반 인터랙션 도구의 설계)

  • Choi, Yoo-Joo;Rhee, Seon-Min;You, Hyo-Sun;Roh, Young-Sub
    • The KIPS Transactions:PartB
    • /
    • v.15B no.5
    • /
    • pp.421-434
    • /
    • 2008
  • As computer graphics, virtual reality and augmented reality technologies have been developed, in many application areas based on those techniques, interaction for 3D space is required such as selection and manipulation of an 3D object. In this paper, we propose a framework for a vision-based 3D interaction which enables to simulate functions of an expensive 3D mouse for a desktop environment. The proposed framework includes a specially manufactured interaction device using three-color LEDs. By recognizing position and color of the LED from video sequences, various events of the mouse and 6 DOF interactions are supported. Since the proposed device is more intuitive and easier than an existing 3D mouse which is expensive and requires skilled manipulation, it can be used without additional learning or training. In this paper, we explain methods for making a pointing device using three-color LEDs which is one of the components of the proposed framework, calculating 3D position and orientation of the pointer and analyzing color of the LED from video sequences. We verify accuracy and usefulness of the proposed device by showing a measurement result of an error of the 3D position and orientation.

Computer Vision and Neuro- Net Based Automatic Grading of a Mushroom(Lentinus Edodes L.) (컴퓨터시각과 신경회로망에 의한 표고등급의 자동판정)

  • Hwang, Heon;Lee, Choongho;Han, Joonhyun
    • Journal of Bio-Environment Control
    • /
    • v.3 no.1
    • /
    • pp.42-51
    • /
    • 1994
  • Visual features of a mushromm(Lentinus Edodes L.) are critical in sorting and grading as most agricultural products are. Because of its complex and various visual features, grading and sorting of mushrooms have been done manually by the human expert. Though actions involved in human grading look simple, it decision making underneath the simple action comes from the result of the complex neural processing of visual image. Recently, an artificial neural network has drawn a great attention because of its functional capability as a partial substitute of the human brain. Since most agricultural products are not uniquely defined in its physical properties and do not have a well defined job structure, the neuro -net based computer visual information processing is the promising approach toward the automation in the agricultural field. In this paper, first, the neuro - net based classification of simple geometric primitives were done and the generalization property of the network was tested for degraded primitives. And then the neuro-net based grading system was developed for a mushroom. A computer vision system was utilized for extracting and quantifying the qualitative visual features of sampled mushrooms. The extracted visual features of sampled mushrooms and their corresponding grades were used as input/output pairs for training the neural network. The grading performance of the trained network for the mushrooms graded previously by the expert were also presented.

  • PDF

A Vision Transformer Based Recommender System Using Side Information (부가 정보를 활용한 비전 트랜스포머 기반의 추천시스템)

  • Kwon, Yujin;Choi, Minseok;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.3
    • /
    • pp.119-137
    • /
    • 2022
  • Recent recommendation system studies apply various deep learning models to represent user and item interactions better. One of the noteworthy studies is ONCF(Outer product-based Neural Collaborative Filtering) which builds a two-dimensional interaction map via outer product and employs CNN (Convolutional Neural Networks) to learn high-order correlations from the map. However, ONCF has limitations in recommendation performance due to the problems with CNN and the absence of side information. ONCF using CNN has an inductive bias problem that causes poor performances for data with a distribution that does not appear in the training data. This paper proposes to employ a Vision Transformer (ViT) instead of the vanilla CNN used in ONCF. The reason is that ViT showed better results than state-of-the-art CNN in many image classification cases. In addition, we propose a new architecture to reflect side information that ONCF did not consider. Unlike previous studies that reflect side information in a neural network using simple input combination methods, this study uses an independent auxiliary classifier to reflect side information more effectively in the recommender system. ONCF used a single latent vector for user and item, but in this study, a channel is constructed using multiple vectors to enable the model to learn more diverse expressions and to obtain an ensemble effect. The experiments showed our deep learning model improved performance in recommendation compared to ONCF.

Computer Vision-based Continuous Large-scale Site Monitoring System through Edge Computing and Small-Object Detection

  • Kim, Yeonjoo;Kim, Siyeon;Hwang, Sungjoo;Hong, Seok Hwan
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.1243-1244
    • /
    • 2022
  • In recent years, the growing interest in off-site construction has led to factories scaling up their manufacturing and production processes in the construction sector. Consequently, continuous large-scale site monitoring in low-variability environments, such as prefabricated components production plants (precast concrete production), has gained increasing importance. Although many studies on computer vision-based site monitoring have been conducted, challenges for deploying this technology for large-scale field applications still remain. One of the issues is collecting and transmitting vast amounts of video data. Continuous site monitoring systems are based on real-time video data collection and analysis, which requires excessive computational resources and network traffic. In addition, it is difficult to integrate various object information with different sizes and scales into a single scene. Various sizes and types of objects (e.g., workers, heavy equipment, and materials) exist in a plant production environment, and these objects should be detected simultaneously for effective site monitoring. However, with the existing object detection algorithms, it is difficult to simultaneously detect objects with significant differences in size because collecting and training massive amounts of object image data with various scales is necessary. This study thus developed a large-scale site monitoring system using edge computing and a small-object detection system to solve these problems. Edge computing is a distributed information technology architecture wherein the image or video data is processed near the originating source, not on a centralized server or cloud. By inferring information from the AI computing module equipped with CCTVs and communicating only the processed information with the server, it is possible to reduce excessive network traffic. Small-object detection is an innovative method to detect different-sized objects by cropping the raw image and setting the appropriate number of rows and columns for image splitting based on the target object size. This enables the detection of small objects from cropped and magnified images. The detected small objects can then be expressed in the original image. In the inference process, this study used the YOLO-v5 algorithm, known for its fast processing speed and widely used for real-time object detection. This method could effectively detect large and even small objects that were difficult to detect with the existing object detection algorithms. When the large-scale site monitoring system was tested, it performed well in detecting small objects, such as workers in a large-scale view of construction sites, which were inaccurately detected by the existing algorithms. Our next goal is to incorporate various safety monitoring and risk analysis algorithms into this system, such as collision risk estimation, based on the time-to-collision concept, enabling the optimization of safety routes by accumulating workers' paths and inferring the risky areas based on workers' trajectory patterns. Through such developments, this continuous large-scale site monitoring system can guide a construction plant's safety management system more effectively.

  • PDF

A Comparison of Image Classification System for Building Waste Data based on Deep Learning (딥러닝기반 건축폐기물 이미지 분류 시스템 비교)

  • Jae-Kyung Sung;Mincheol Yang;Kyungnam Moon;Yong-Guk Kim
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.3
    • /
    • pp.199-206
    • /
    • 2023
  • This study utilizes deep learning algorithms to automatically classify construction waste into three categories: wood waste, plastic waste, and concrete waste. Two models, VGG-16 and ViT (Vision Transformer), which are convolutional neural network image classification algorithms and NLP-based models that sequence images, respectively, were compared for their performance in classifying construction waste. Image data for construction waste was collected by crawling images from search engines worldwide, and 3,000 images, with 1,000 images for each category, were obtained by excluding images that were difficult to distinguish with the naked eye or that were duplicated and would interfere with the experiment. In addition, to improve the accuracy of the models, data augmentation was performed during training with a total of 30,000 images. Despite the unstructured nature of the collected image data, the experimental results showed that VGG-16 achieved an accuracy of 91.5%, and ViT achieved an accuracy of 92.7%. This seems to suggest the possibility of practical application in actual construction waste data management work. If object detection techniques or semantic segmentation techniques are utilized based on this study, more precise classification will be possible even within a single image, resulting in more accurate waste classification