• Title/Summary/Keyword: Deep Learning AI

Search Result 622, Processing Time 0.022 seconds

Development and Application of Tunnel Design Automation Technology Using 3D Spatial Information : BIM-Based Design for Namhae Seomyeon - Yeosu Shindeok National Highway Construction (3D 공간정보를 활용한 터널 설계 자동화 기술 개발 및 적용 사례 : 남해 서면-여수 신덕 국도 건설공사 BIM기반 설계를 중심으로)

  • Eunji Jo;Woojin Kim;Kwangyeom Kim;Jaeho Jung;Sanghyuk Bang
    • Tunnel and Underground Space
    • /
    • v.33 no.4
    • /
    • pp.209-227
    • /
    • 2023
  • The government continues to announce measures to revitalize smart construction technology based on BIM for productivity innovation in the construction industry. In the design phase, the goal is design automation and optimization by converging BIM Data and other advanced technologies. Accordingly, in the basic design of the Namhae Seomyeon-Yeosu Sindeok National Road Construction Project, a domestic undersea tunnel project, BIM-based design was carried out by developing tunnel design automation technology using 3D spatial information according to the tunnel design process. In order to derive the optimal alignment, more than 10,000 alignment cases were generated in 36hr using the generative design technique and a quantitative evaluation of the objective functions defined by the designer was performed. AI-based ground classification and 3D Geo Model were established to evaluate the economic feasibility and stability of the optimal alignment. AI-based ground classification has improved its precision by performing about 30 types of ground classification per borehole, and in the case of the 3D Geo Model, its utilization can be expected in that it can accumulate ground data added during construction. In the case of 3D blasting design, the optimal charge weight was derived in 5 minutes by reviewing all security objects on the project range on Dynamo, and the design result was visualized in 3D space for intuitive and convenient construction management so that it could be used directly during construction.

Analysis of Research Trends in New Drug Development with Artificial Intelligence Using Text Mining (텍스트 마이닝을 이용한 인공지능 활용 신약 개발 연구 동향 분석)

  • Jae Woo Nam;Young Jun Kim
    • Journal of Life Science
    • /
    • v.33 no.8
    • /
    • pp.663-679
    • /
    • 2023
  • This review analyzes research trends related to new drug development using artificial intelligence from 2010 to 2022. This analysis organized the abstracts of 2,421 studies into a corpus, and words with high frequency and high connection centrality were extracted through preprocessing. The analysis revealed a similar word frequency trend between 2010 and 2019 to that between 2020 and 2022. In terms of the research method, many studies using machine learning were conducted from 2010 to 2020, and since 2021, research using deep learning has been increasing. Through these studies, we investigated the trends in research on artificial intelligence utilization by field and the strengths, problems, and challenges of related research. We found that since 2021, the application of artificial intelligence has been expanding, such as research using artificial intelligence for drug rearrangement, using computers to develop anticancer drugs, and applying artificial intelligence to clinical trials. This article briefly presents the prospects of new drug development research using artificial intelligence. If the reliability and safety of bio and medical data are ensured, and the development of the above artificial intelligence technology continues, it is judged that the direction of new drug development using artificial intelligence will proceed to personalized medicine and precision medicine, so we encourage efforts in that field.

A Study of Development and Application of an Inland Water Body Training Dataset Using Sentinel-1 SAR Images in Korea (Sentinel-1 SAR 영상을 활용한 국내 내륙 수체 학습 데이터셋 구축 및 알고리즘 적용 연구)

  • Eu-Ru Lee;Hyung-Sup Jung
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_1
    • /
    • pp.1371-1388
    • /
    • 2023
  • Floods are becoming more severe and frequent due to global warming-induced climate change. Water disasters are rising in Korea due to severe rainfall and wet seasons. This makes preventive climate change measures and efficient water catastrophe responses crucial, and synthetic aperture radar satellite imagery can help. This research created 1,423 water body learning datasets for individual water body regions along the Han and Nakdong waterways to reflect domestic water body properties discovered by Sentinel-1 satellite radar imagery. We created a document with exact data annotation criteria for many situations. After the dataset was processed, U-Net, a deep learning model, analyzed water body detection results. The results from applying the learned model to water body locations not involved in the learning process were studied to validate soil water body monitoring on a national scale. The analysis showed that the created water body area detected water bodies accurately (F1-Score: 0.987, Intersection over Union [IoU]: 0.955). Other domestic water body regions not used for training and evaluation showed similar accuracy (F1-Score: 0.941, IoU: 0.89). Both outcomes showed that the computer accurately spotted water bodies in most areas, however tiny streams and gloomy areas had problems. This work should improve water resource change and disaster damage surveillance. Future studies will likely include more water body attribute datasets. Such databases could help manage and monitor water bodies nationwide and shed light on misclassified regions.

Lightening of Human Pose Estimation Algorithm Using MobileViT and Transfer Learning

  • Kunwoo Kim;Jonghyun Hong;Jonghyuk Park
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.9
    • /
    • pp.17-25
    • /
    • 2023
  • In this paper, we propose a model that can perform human pose estimation through a MobileViT-based model with fewer parameters and faster estimation. The based model demonstrates lightweight performance through a structure that combines features of convolutional neural networks with features of Vision Transformer. Transformer, which is a major mechanism in this study, has become more influential as its based models perform better than convolutional neural network-based models in the field of computer vision. Similarly, in the field of human pose estimation, Vision Transformer-based ViTPose maintains the best performance in all human pose estimation benchmarks such as COCO, OCHuman, and MPII. However, because Vision Transformer has a heavy model structure with a large number of parameters and requires a relatively large amount of computation, it costs users a lot to train the model. Accordingly, the based model overcame the insufficient Inductive Bias calculation problem, which requires a large amount of computation by Vision Transformer, with Local Representation through a convolutional neural network structure. Finally, the proposed model obtained a mean average precision of 0.694 on the MS COCO benchmark with 3.28 GFLOPs and 9.72 million parameters, which are 1/5 and 1/9 the number compared to ViTPose, respectively.

GEase-K: Linear and Nonlinear Autoencoder-based Recommender System with Side Information (GEase-K: 부가 정보를 활용한 선형 및 비선형 오토인코더 기반의 추천시스템)

  • Taebeom Lee;Seung-hak Lee;Min-jeong Ma;Yoonho Cho
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.167-183
    • /
    • 2023
  • In the recent field of recommendation systems, various studies have been conducted to model sparse data effectively. Among these, GLocal-K(Global and Local Kernels for Recommender Systems) is a research endeavor combining global and local kernels to provide personalized recommendations by considering global data patterns and individual user characteristics. However, due to its utilization of kernel tricks, GLocal-K exhibits diminished performance on highly sparse data and struggles to offer recommendations for new users or items due to the absence of side information. In this paper, to address these limitations of GLocal-K, we propose the GEase-K (Global and EASE kernels for Recommender Systems) model, incorporating the EASE(Embarrassingly Shallow Autoencoders for Sparse Data) model and leveraging side information. Initially, we substitute EASE for the local kernel in GLocal-K to enhance recommendation performance on highly sparse data. EASE, functioning as a simple linear operational structure, is an autoencoder that performs highly on extremely sparse data through regularization and learning item similarity. Additionally, we utilize side information to alleviate the cold-start problem. We enhance the understanding of user-item similarities by employing a conditional autoencoder structure during the training process to incorporate side information. In conclusion, GEase-K demonstrates resilience in highly sparse data and cold-start situations by combining linear and nonlinear structures and utilizing side information. Experimental results show that GEase-K outperforms GLocal-K based on the RMSE and MAE metrics on the highly sparse GoodReads and ModCloth datasets. Furthermore, in cold-start experiments divided into four groups using the GoodReads and ModCloth datasets, GEase-K denotes superior performance compared to GLocal-K.

Battery-loaded power management algorithm of electric propulsion ship based on power load and state learning model (전력 부하와 학습모델 기반의 전기추진선박의 배터리 연동 전력관리 알고리즘)

  • Oh, Ji-hyun;Oh, Jin-seok
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.9
    • /
    • pp.1202-1208
    • /
    • 2020
  • In line with the current era of the 4th Industrial Revolution, it is necessary to prepare for the future by integrating AI elements in the ship sector. In addition, it is necessary to respond to this in the field of power management for the appearance of autonomous ships. In this study, we propose a battery-linked electric propulsion system (BLEPS) algorithm using machine learning's DNN. For the experiment, we learned the pattern of ship power consumption for each operation mode based on the ship data through LabView and derived the battery status through Python to check the flexibility of the generator and battery interlocking. As a result of the experiment, the low load operation of the generator was reduced through charging and discharging of the battery, and economic efficiency and reliability were confirmed by reducing the fuel consumption of 1% of LNG.

Development of Convolutional Neural Network Basic Practice Cases (합성곱 신경망 기초 실습 사례 개발)

  • Hur, Kyeong
    • Journal of Practical Engineering Education
    • /
    • v.14 no.2
    • /
    • pp.279-285
    • /
    • 2022
  • In this paper, as a liberal arts course for non-majors, we developed a basic practice case for convolutional neural networks, which is essential for designing a basic convolutional neural network course curriculum. The developed practice case focuses on understanding the working principle of the convolutional neural network and uses a spreadsheet to check the entire visualized process. The developed practice case consisted of generating supervised learning method image training data, implementing the input layer, convolution layer (convolutional layer), pooling layer, and output layer sequentially, and testing the performance of the convolutional neural network on new data. By extending the practice cases developed in this paper, the number of images to be recognized can be expanded, or basic practice cases can be made to create a convolutional neural network that increases the compression rate for high-quality images. Therefore, it can be said that the utility of this convolutional neural network basic practice case is high.

Design of weighted federated learning framework based on local model validation

  • Kim, Jung-Jun;Kang, Jeon Seong;Chung, Hyun-Joon;Park, Byung-Hoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.11
    • /
    • pp.13-18
    • /
    • 2022
  • In this paper, we proposed VW-FedAVG(Validation based Weighted FedAVG) which updates the global model by weighting according to performance verification from the models of each device participating in the training. The first method is designed to validate each local client model through validation dataset before updating the global model with a server side validation structure. The second is a client-side validation structure, which is designed in such a way that the validation data set is evenly distributed to each client and the global model is after validation. MNIST, CIFAR-10 is used, and the IID, Non-IID distribution for image classification obtained higher accuracy than previous studies.

English Conversation System Using Artificial Intelligent of based on Virtual Reality (가상현실 기반의 인공지능 영어회화 시스템)

  • Cheon, EunYoung
    • Journal of the Korea Convergence Society
    • /
    • v.10 no.11
    • /
    • pp.55-61
    • /
    • 2019
  • In order to realize foreign language education, various existing educational media have been provided, but there are disadvantages in that the cost of the parish and the media program is high and the real-time responsiveness is poor. In this paper, we propose an artificial intelligence English conversation system based on VR and speech recognition. We used Google CardBoard VR and Google Speech API to build the system and developed artificial intelligence algorithms for providing virtual reality environment and talking. In the proposed speech recognition server system, the sentences spoken by the user can be divided into word units and compared with the data words stored in the database to provide the highest probability. Users can communicate with and respond to people in virtual reality. The function provided by the conversation is independent of the contextual conversations and themes, and the conversations with the AI assistant are implemented in real time so that the user system can be checked in real time. It is expected to contribute to the expansion of virtual education contents service related to the Fourth Industrial Revolution through the system combining the virtual reality and the voice recognition function proposed in this paper.

Filter-mBART Based Neural Machine Translation Using Parallel Corpus Filtering (병렬 말뭉치 필터링을 적용한 Filter-mBART기반 기계번역 연구)

  • Moon, Hyeonseok;Park, Chanjun;Eo, Sugyeong;Park, JeongBae;Lim, Heuiseok
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.5
    • /
    • pp.1-7
    • /
    • 2021
  • In the latest trend of machine translation research, the model is pretrained through a large mono lingual corpus and then finetuned with a parallel corpus. Although many studies tend to increase the amount of data used in the pretraining stage, it is hard to say that the amount of data must be increased to improve machine translation performance. In this study, through an experiment based on the mBART model using parallel corpus filtering, we propose that high quality data can yield better machine translation performance, even utilizing smaller amount of data. We propose that it is important to consider the quality of data rather than the amount of data, and it can be used as a guideline for building a training corpus.