• Title/Summary/Keyword: Transfer of learning

Search Result 743, Processing Time 0.025 seconds

Two person Interaction Recognition Based on Effective Hybrid Learning

  • Ahmed, Minhaz Uddin;Kim, Yeong Hyeon;Kim, Jin Woo;Bashar, Md Rezaul;Rhee, Phill Kyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.751-770
    • /
    • 2019
  • Action recognition is an essential task in computer vision due to the variety of prospective applications, such as security surveillance, machine learning, and human-computer interaction. The availability of more video data than ever before and the lofty performance of deep convolutional neural networks also make it essential for action recognition in video. Unfortunately, limited crafted video features and the scarcity of benchmark datasets make it challenging to address the multi-person action recognition task in video data. In this work, we propose a deep convolutional neural network-based Effective Hybrid Learning (EHL) framework for two-person interaction classification in video data. Our approach exploits a pre-trained network model (the VGG16 from the University of Oxford Visual Geometry Group) and extends the Faster R-CNN (region-based convolutional neural network a state-of-the-art detector for image classification). We broaden a semi-supervised learning method combined with an active learning method to improve overall performance. Numerous types of two-person interactions exist in the real world, which makes this a challenging task. In our experiment, we consider a limited number of actions, such as hugging, fighting, linking arms, talking, and kidnapping in two environment such simple and complex. We show that our trained model with an active semi-supervised learning architecture gradually improves the performance. In a simple environment using an Intelligent Technology Laboratory (ITLab) dataset from Inha University, performance increased to 95.6% accuracy, and in a complex environment, performance reached 81% accuracy. Our method reduces data-labeling time, compared to supervised learning methods, for the ITLab dataset. We also conduct extensive experiment on Human Action Recognition benchmarks such as UT-Interaction dataset, HMDB51 dataset and obtain better performance than state-of-the-art approaches.

The Process of the Interjoint and Intersegmental Coordination of Side Kick Motion in Taekwondo (태권도 옆차기 동작의 인체관절과 분절사이의 협응 과정)

  • Yoon, Chang-Jin;Chae, Woen-Sik
    • Korean Journal of Applied Biomechanics
    • /
    • v.18 no.4
    • /
    • pp.179-189
    • /
    • 2008
  • The purpose of this study was to investigate interjoint and intersegmental coordination of lower segments in skill process. For the investigation, we examined the difference of resultant linear velocity of segments and angle vs angle graph. Novice subjects were 9 male middle school students who have never been experienced a taekwondo. We analyzed kinematic variables of Side Kick motion through videographical analysis. The conclusions were as follows. 1. Examining the graph of novice subjects' maximal resultant linear velocity of the thigh, shank, and foot segment, as it gets closer to the end of the training, the maximal resultant linear velocity in each segment increases which can be assumed to be a result of the effective momentum transfer between adjacent segments. 2 This research showed a sequential transfer from trunk, to thigh, and then to shank as it gets closer to the end of learning at intersegment angular velocity, and it also showed pattern of throwlike motion and pushlike motion. 3. In three dimension of flexion-extension, adduction-abduction and internal-external rotation of the thigh and shank segment, the angle-angle diagram of knee joint and of hip joint showed that dynamic change was indicated at the beginning of learning but stable coordination pattern was indicated like skilled subject as novice subjects became skilled.

Deep Learning based Domain Adaptation: A Survey (딥러닝 기반의 도메인 적응 기술: 서베이)

  • Na, Jaemin;Hwang, Wonjun
    • Journal of Broadcast Engineering
    • /
    • v.27 no.4
    • /
    • pp.511-518
    • /
    • 2022
  • Supervised learning based on deep learning has made a leap forward in various application fields. However, many supervised learning methods work under the common assumption that training and test data are extracted from the same distribution. If it deviates from this constraint, the deep learning network trained in the training domain is highly likely to deteriorate rapidly in the test domain due to the distribution difference between domains. Domain adaptation is a methodology of transfer learning that trains a deep learning network to make successful inferences in a label-poor test domain (i.e., target domain) based on learned knowledge of a labeled-rich training domain (i.e., source domain). In particular, the unsupervised domain adaptation technique deals with the domain adaptation problem by assuming that only image data without labels in the target domain can be accessed. In this paper, we explore the unsupervised domain adaptation techniques.

Effects of Psychiatric Nursing Practice Education Using Virtual Simulation for Nursing (가상간호시뮬레이션을 활용한 정신간호실습 교육의 효과)

  • Han, Mi Ra;Lee, Jihye
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.10
    • /
    • pp.333-342
    • /
    • 2021
  • The purpose of this study is to compare the differences in transfer motivation and learning self-efficacy before and after applying virtual simulation for nursing in psychiatric nursing practice, and to provide them as basic data for effective psychiatric practical education. This study was conducted from October to December 2020. The subjects were 41 people who were enrolled in the third year of a located in U city, and who had received psychiatric nursing practice education using virtual simulation for nursing. Data were analyzed by paired t-test and pearson's correlation coefficient. After practice compared to before psychiatric nursing practice with virtual simulation nursing applied, transfer motivation was significantly increased and learner self-efficacy increased, but it was not statistically significant. Therefore, It was confirmed that psychiatric nursing practice education using virtual simulation for nursing is partly an effective practice strategy.

Assessing the Impact of Sampling Intensity on Land Use and Land Cover Estimation Using High-Resolution Aerial Images and Deep Learning Algorithms (고해상도 항공 영상과 딥러닝 알고리즘을 이용한 표본강도에 따른 토지이용 및 토지피복 면적 추정)

  • Yong-Kyu Lee;Woo-Dam Sim;Jung-Soo Lee
    • Journal of Korean Society of Forest Science
    • /
    • v.112 no.3
    • /
    • pp.267-279
    • /
    • 2023
  • This research assessed the feasibility of using high-resolution aerial images and deep learning algorithms for estimating the land-use and land-cover areas at the Approach 3 level, as outlined by the Intergovernmental Panel on Climate Change. The results from different sampling densities of high-resolution (51 cm) aerial images were compared with the land-cover map, provided by the Ministry of Environment, and analyzed to estimate the accuracy of the land-use and land-cover areas. Transfer learning was applied to the VGG16 architecture for the deep learning model, and sampling densities of 4 × 4 km, 2 × 4 km, 2 × 2 km, 1 × 2 km, 1 × 1 km, 500 × 500 m, and 250 × 250 m were used for estimating and evaluating the areas. The overall accuracy and kappa coefficient of the deep learning model were 91.1% and 88.8%, respectively. The F-scores, except for the pasture category, were >90% for all categories, indicating superior accuracy of the model. Chi-square tests of the sampling densities showed no significant difference in the area ratios of the land-cover map provided by the Ministry of Environment among all sampling densities except for 4 × 4 km at a significance level of p = 0.1. As the sampling density increased, the standard error and relative efficiency decreased. The relative standard error decreased to ≤15% for all land-cover categories at 1 × 1 km sampling density. These results indicated that a sampling density more detailed than 1 x 1 km is appropriate for estimating land-cover area at the local level.

Comparison Study of the Performance of CNN Models with Multi-view Image Set on the Classification of Ship Hull Blocks (다시점 영상 집합을 활용한 선체 블록 분류를 위한 CNN 모델 성능 비교 연구)

  • Chon, Haemyung;Noh, Jackyou
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.57 no.3
    • /
    • pp.140-151
    • /
    • 2020
  • It is important to identify the location of ship hull blocks with exact block identification number when scheduling the shipbuilding process. The wrong information on the location and identification number of some hull block can cause low productivity by spending time to find where the exact hull block is. In order to solve this problem, it is necessary to equip the system to track the location of the blocks and to identify the identification numbers of the blocks automatically. There were a lot of researches of location tracking system for the hull blocks on the stockyard. However there has been no research to identify the hull blocks on the stockyard. This study compares the performance of 5 Convolutional Neural Network (CNN) models with multi-view image set on the classification of the hull blocks to identify the blocks on the stockyard. The CNN models are open algorithms of ImageNet Large-Scale Visual Recognition Competition (ILSVRC). Four scaled hull block models are used to acquire the images of ship hull blocks. Learning and transfer learning of the CNN models with original training data and augmented data of the original training data were done. 20 tests and predictions in consideration of five CNN models and four cases of training conditions are performed. In order to compare the classification performance of the CNN models, accuracy and average F1-Score from confusion matrix are adopted as the performance measures. As a result of the comparison, Resnet-152v2 model shows the highest accuracy and average F1-Score with full block prediction image set and with cropped block prediction image set.

Change Detection for High-resolution Satellite Images Using Transfer Learning and Deep Learning Network (전이학습과 딥러닝 네트워크를 활용한 고해상도 위성영상의 변화탐지)

  • Song, Ah Ram;Choi, Jae Wan;Kim, Yong Il
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.37 no.3
    • /
    • pp.199-208
    • /
    • 2019
  • As the number of available satellites increases and technology advances, image information outputs are becoming increasingly diverse and a large amount of data is accumulating. In this study, we propose a change detection method for high-resolution satellite images that uses transfer learning and a deep learning network to overcome the limit caused by insufficient training data via the use of pre-trained information. The deep learning network used in this study comprises convolutional layers to extract the spatial and spectral information and convolutional long-short term memory layers to analyze the time series information. To use the learned information, the two initial convolutional layers of the change detection network are designed to use learned values from 40,000 patches of the ISPRS (International Society for Photogrammertry and Remote Sensing) dataset as initial values. In addition, 2D (2-Dimensional) and 3D (3-dimensional) kernels were used to find the optimized structure for the high-resolution satellite images. The experimental results for the KOMPSAT-3A (KOrean Multi-Purpose SATllite-3A) satellite images show that this change detection method can effectively extract changed/unchanged pixels but is less sensitive to changes due to shadow and relief displacements. In addition, the change detection accuracy of two sites was improved by using 3D kernels. This is because a 3D kernel can consider not only the spatial information but also the spectral information. This study indicates that we can effectively detect changes in high-resolution satellite images using the constructed image information and deep learning network. In future work, a pre-trained change detection network will be applied to newly obtained images to extend the scope of the application.

COVID-19 Diagnosis from CXR images through pre-trained Deep Visual Embeddings

  • Khalid, Shahzaib;Syed, Muhammad Shehram Shah;Saba, Erum;Pirzada, Nasrullah
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.5
    • /
    • pp.175-181
    • /
    • 2022
  • COVID-19 is an acute respiratory syndrome that affects the host's breathing and respiratory system. The novel disease's first case was reported in 2019 and has created a state of emergency in the whole world and declared a global pandemic within months after the first case. The disease created elements of socioeconomic crisis globally. The emergency has made it imperative for professionals to take the necessary measures to make early diagnoses of the disease. The conventional diagnosis for COVID-19 is through Polymerase Chain Reaction (PCR) testing. However, in a lot of rural societies, these tests are not available or take a lot of time to provide results. Hence, we propose a COVID-19 classification system by means of machine learning and transfer learning models. The proposed approach identifies individuals with COVID-19 and distinguishes them from those who are healthy with the help of Deep Visual Embeddings (DVE). Five state-of-the-art models: VGG-19, ResNet50, Inceptionv3, MobileNetv3, and EfficientNetB7, were used in this study along with five different pooling schemes to perform deep feature extraction. In addition, the features are normalized using standard scaling, and 4-fold cross-validation is used to validate the performance over multiple versions of the validation data. The best results of 88.86% UAR, 88.27% Specificity, 89.44% Sensitivity, 88.62% Accuracy, 89.06% Precision, and 87.52% F1-score were obtained using ResNet-50 with Average Pooling and Logistic regression with class weight as the classifier.

SuperDepthTransfer: Depth Extraction from Image Using Instance-Based Learning with Superpixels

  • Zhu, Yuesheng;Jiang, Yifeng;Huang, Zhuandi;Luo, Guibo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.10
    • /
    • pp.4968-4986
    • /
    • 2017
  • In this paper, we primarily address the difficulty of automatic generation of a plausible depth map from a single image in an unstructured environment. The aim is to extrapolate a depth map with a more correct, rich, and distinct depth order, which is both quantitatively accurate as well as visually pleasing. Our technique, which is fundamentally based on a preexisting DepthTransfer algorithm, transfers depth information at the level of superpixels. This occurs within a framework that replaces a pixel basis with one of instance-based learning. A vital superpixels feature enhancing matching precision is posterior incorporation of predictive semantic labels into the depth extraction procedure. Finally, a modified Cross Bilateral Filter is leveraged to augment the final depth field. For training and evaluation, experiments were conducted using the Make3D Range Image Dataset and vividly demonstrate that this depth estimation method outperforms state-of-the-art methods for the correlation coefficient metric, mean log10 error and root mean squared error, and achieves comparable performance for the average relative error metric in both efficacy and computational efficiency. This approach can be utilized to automatically convert 2D images into stereo for 3D visualization, producing anaglyph images that are visually superior in realism and simultaneously more immersive.

Performance Comparison and Error Analysis of Korean Bio-medical Named Entity Recognition (한국어 생의학 개체명 인식 성능 비교와 오류 분석)

  • Jae-Hong Lee
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.19 no.4
    • /
    • pp.701-708
    • /
    • 2024
  • The advent of transformer architectures in deep learning has been a major breakthrough in natural language processing research. Object name recognition is a branch of natural language processing and is an important research area for tasks such as information retrieval. It is also important in the biomedical field, but the lack of Korean biomedical corpora for training has limited the development of Korean clinical research using AI. In this study, we built a new biomedical corpus for Korean biomedical entity name recognition and selected language models pre-trained on a large Korean corpus for transfer learning. We compared the name recognition performance of the selected language models by F1-score and the recognition rate by tag, and analyzed the errors. In terms of recognition performance, KlueRoBERTa showed relatively good performance. The error analysis of the tagging process shows that the recognition performance of Disease is excellent, but Body and Treatment are relatively low. This is due to over-segmentation and under-segmentation that fails to properly categorize entity names based on context, and it will be necessary to build a more precise morphological analyzer and a rich lexicon to compensate for the incorrect tagging.