• Title/Summary/Keyword: deep-learning dataset

Search Result 803, Processing Time 0.026 seconds

Learning Deep Representation by Increasing ConvNets Depth for Few Shot Learning

  • Fabian, H.S. Tan;Kang, Dae-Ki
    • International journal of advanced smart convergence
    • /
    • v.8 no.4
    • /
    • pp.75-81
    • /
    • 2019
  • Though recent advancement of deep learning methods have provided satisfactory results from large data domain, somehow yield poor performance on few-shot classification tasks. In order to train a model with strong performance, i.e. deep convolutional neural network, it depends heavily on huge dataset and the labeled classes of the dataset can be extremely humongous. The cost of human annotation and scarcity of the data among the classes have drastically limited the capability of current image classification model. On the contrary, humans are excellent in terms of learning or recognizing new unseen classes with merely small set of labeled examples. Few-shot learning aims to train a classification model with limited labeled samples to recognize new classes that have neverseen during training process. In this paper, we increase the backbone depth of the embedding network in orderto learn the variation between the intra-class. By increasing the network depth of the embedding module, we are able to achieve competitive performance due to the minimized intra-class variation.

Remote Sensing Image Classification for Land Cover Mapping in Developing Countries: A Novel Deep Learning Approach

  • Lynda, Nzurumike Obianuju;Nnanna, Nwojo Agwu;Boukar, Moussa Mahamat
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.2
    • /
    • pp.214-222
    • /
    • 2022
  • Convolutional Neural networks (CNNs) are a category of deep learning networks that have proven very effective in computer vision tasks such as image classification. Notwithstanding, not much has been seen in its use for remote sensing image classification in developing countries. This is majorly due to the scarcity of training data. Recently, transfer learning technique has successfully been used to develop state-of-the art models for remote sensing (RS) image classification tasks using training and testing data from well-known RS data repositories. However, the ability of such model to classify RS test data from a different dataset has not been sufficiently investigated. In this paper, we propose a deep CNN model that can classify RS test data from a dataset different from the training dataset. To achieve our objective, we first, re-trained a ResNet-50 model using EuroSAT, a large-scale RS dataset to develop a base model then we integrated Augmentation and Ensemble learning to improve its generalization ability. We further experimented on the ability of this model to classify a novel dataset (Nig_Images). The final classification results shows that our model achieves a 96% and 80% accuracy on EuroSAT and Nig_Images test data respectively. Adequate knowledge and usage of this framework is expected to encourage research and the usage of deep CNNs for land cover mapping in cases of lack of training data as obtainable in developing countries.

Development of Virtual Simulator and Database for Deep Learning-based Object Detection (딥러닝 기반 장애물 인식을 위한 가상환경 및 데이터베이스 구축)

  • Lee, JaeIn;Gwak, Gisung;Kim, KyongSu;Kang, WonYul;Shin, DaeYoung;Hwang, Sung-Ho
    • Journal of Drive and Control
    • /
    • v.18 no.4
    • /
    • pp.9-18
    • /
    • 2021
  • This study proposes a method for creating learning datasets to recognize obstacles using deep learning algorithms in automated construction machinery or an autonomous vehicle. Recently, many researchers and engineers have developed various recognition algorithms based on deep learning following an increase in computing power. In particular, the image classification technology and image segmentation technology represent deep learning recognition algorithms. They are used to identify obstacles that interfere with the driving situation of an autonomous vehicle. Therefore, various organizations and companies have started distributing open datasets, but there is a remote possibility that they will perfectly match the user's desired environment. In this study, we created an interface of the virtual simulator such that users can easily create their desired training dataset. In addition, the customized dataset was further advanced by using the RDBMS system, and the recognition rate was improved.

Image-based ship detection using deep learning

  • Lee, Sung-Jun;Roh, Myung-Il;Oh, Min-Jae
    • Ocean Systems Engineering
    • /
    • v.10 no.4
    • /
    • pp.415-434
    • /
    • 2020
  • Detecting objects is important for the safe operation of ships, and enables collision avoidance, risk detection, and autonomous sailing. This study proposes a ship detection method from images and videos taken at sea using one of the state-of-the-art deep neural network-based object detection algorithms. A deep learning model is trained using a public maritime dataset, and results show it can detect all types of floating objects and classify them into ten specific classes that include a ship, speedboat, and buoy. The proposed deep learning model is compared to a universal trained model that detects and classifies objects into general classes, such as a person, dog, car, and boat, and results show that the proposed model outperforms the other in the detection of maritime objects. Different deep neural network structures are then compared to obtain the best detection performance. The proposed model also shows a real-time detection speed of approximately 30 frames per second. Hence, it is expected that the proposed model can be used to detect maritime objects and reduce risks while at sea.

Performance analysis of deep learning-based automatic classification of upper endoscopic images according to data construction (딥러닝 기반 상부위장관 내시경 이미지 자동분류의 데이터 구성별 성능 분석 연구)

  • Seo, Jeong Min;Lim, Sang Heon;Kim, Yung Jae;Chung, Jun Won;Kim, Kwang Gi
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.3
    • /
    • pp.451-460
    • /
    • 2022
  • Recently, several deep learning studies have been reported to automatically identify the location of diagnostic devices using endoscopic data. In previous studies, there was no design to determine whether the configuration of the dataset resulted in differences in the accuracy in which artificial intelligence models perform image classification. Studies that are based on large amounts of data are likely to have different results depending on the composition of the dataset or its proportion. In this study, we intended to determine the existence and extent of accuracy according to the composition of the dataset by compiling it into three main types using larynx, esophagus, gastroscopy, and laryngeal endoscopy images.

Two person Interaction Recognition Based on Effective Hybrid Learning

  • Ahmed, Minhaz Uddin;Kim, Yeong Hyeon;Kim, Jin Woo;Bashar, Md Rezaul;Rhee, Phill Kyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.2
    • /
    • pp.751-770
    • /
    • 2019
  • Action recognition is an essential task in computer vision due to the variety of prospective applications, such as security surveillance, machine learning, and human-computer interaction. The availability of more video data than ever before and the lofty performance of deep convolutional neural networks also make it essential for action recognition in video. Unfortunately, limited crafted video features and the scarcity of benchmark datasets make it challenging to address the multi-person action recognition task in video data. In this work, we propose a deep convolutional neural network-based Effective Hybrid Learning (EHL) framework for two-person interaction classification in video data. Our approach exploits a pre-trained network model (the VGG16 from the University of Oxford Visual Geometry Group) and extends the Faster R-CNN (region-based convolutional neural network a state-of-the-art detector for image classification). We broaden a semi-supervised learning method combined with an active learning method to improve overall performance. Numerous types of two-person interactions exist in the real world, which makes this a challenging task. In our experiment, we consider a limited number of actions, such as hugging, fighting, linking arms, talking, and kidnapping in two environment such simple and complex. We show that our trained model with an active semi-supervised learning architecture gradually improves the performance. In a simple environment using an Intelligent Technology Laboratory (ITLab) dataset from Inha University, performance increased to 95.6% accuracy, and in a complex environment, performance reached 81% accuracy. Our method reduces data-labeling time, compared to supervised learning methods, for the ITLab dataset. We also conduct extensive experiment on Human Action Recognition benchmarks such as UT-Interaction dataset, HMDB51 dataset and obtain better performance than state-of-the-art approaches.

Normal data based rotating machine anomaly detection using CNN with self-labeling

  • Bae, Jaewoong;Jung, Wonho;Park, Yong-Hwa
    • Smart Structures and Systems
    • /
    • v.29 no.6
    • /
    • pp.757-766
    • /
    • 2022
  • To train deep learning algorithms, a sufficient number of data are required. However, in most engineering systems, the acquisition of fault data is difficult or sometimes not feasible, while normal data are secured. The dearth of data is one of the major challenges to developing deep learning models, and fault diagnosis in particular cannot be made in the absence of fault data. With this context, this paper proposes an anomaly detection methodology for rotating machines using only normal data with self-labeling. Since only normal data are used for anomaly detection, a self-labeling method is used to generate a new labeled dataset. The overall procedure includes the following three steps: (1) transformation of normal data to self-labeled data based on a pretext task, (2) training the convolutional neural networks (CNN), and (3) anomaly detection using defined anomaly score based on the softmax output of the trained CNN. The softmax value of the abnormal sample shows different behavior from the normal softmax values. To verify the proposed method, four case studies were conducted, on the Case Western Reserve University (CWRU) bearing dataset, IEEE PHM 2012 data challenge dataset, PHMAP 2021 data challenge dataset, and laboratory bearing testbed; and the results were compared to those of existing machine learning and deep learning methods. The results showed that the proposed algorithm could detect faults in the bearing testbed and compressor with over 99.7% accuracy. In particular, it was possible to detect not only bearing faults but also structural faults such as unbalance and belt looseness with very high accuracy. Compared with the existing GAN, the autoencoder-based anomaly detection algorithm, the proposed method showed high anomaly detection performance.

Residual Learning Based CNN for Gesture Recognition in Robot Interaction

  • Han, Hua
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.385-398
    • /
    • 2021
  • The complexity of deep learning models affects the real-time performance of gesture recognition, thereby limiting the application of gesture recognition algorithms in actual scenarios. Hence, a residual learning neural network based on a deep convolutional neural network is proposed. First, small convolution kernels are used to extract the local details of gesture images. Subsequently, a shallow residual structure is built to share weights, thereby avoiding gradient disappearance or gradient explosion as the network layer deepens; consequently, the difficulty of model optimisation is simplified. Additional convolutional neural networks are used to accelerate the refinement of deep abstract features based on the spatial importance of the gesture feature distribution. Finally, a fully connected cascade softmax classifier is used to complete the gesture recognition. Compared with the dense connection multiplexing feature information network, the proposed algorithm is optimised in feature multiplexing to avoid performance fluctuations caused by feature redundancy. Experimental results from the ISOGD gesture dataset and Gesture dataset prove that the proposed algorithm affords a fast convergence speed and high accuracy.

Fashion Clothing Image Classification Deep Learning (패션 의류 영상 분류 딥러닝)

  • Shin, Seong-Yoon;Wang, Guangxing;Shin, Kwang-Seong;Lee, Hyun-Chang
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.676-677
    • /
    • 2022
  • In this paper, we propose a new method based on a deep learning model with an optimized dynamic decay learning rate and improved model structure to achieve fast and accurate classification of fashion clothing images. Experiments are performed using the model proposed in the Fashion-MNIST dataset and compared with methods of CNN, LeNet, LSTM and BiLSTM.

  • PDF

Recommendation system using Deep Autoencoder for Tensor data

  • Park, Jina;Yong, Hwan-Seung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.8
    • /
    • pp.87-93
    • /
    • 2019
  • These days, as interest in the recommendation system with deep learning is increasing, a number of related studies to develop a performance for collaborative filtering through autoencoder, a state-of-the-art deep learning neural network architecture has advanced considerably. The purpose of this study is to propose autoencoder which is used by the recommendation system to predict ratings, and we added more hidden layers to the original architecture of autoencoder so that we implemented deep autoencoder with 3 to 5 hidden layers for much deeper architecture. In this paper, therefore we make a comparison between the performance of them. In this research, we use 2-dimensional arrays and 3-dimensional tensor as the input dataset. As a result, we found a correlation between matrix entry of the 3-dimensional dataset such as item-time and user-time and also figured out that deep autoencoder with extra hidden layers generalized even better performance than autoencoder.