• Title/Summary/Keyword: Neural Net

Search Result 763, Processing Time 0.024 seconds

Improvement of internal exposure assessments of the inhalation of fuel-type hot particles during long-term outages

  • Moonhyung Cho;Hyeongjin Kim
    • Nuclear Engineering and Technology
    • /
    • v.56 no.9
    • /
    • pp.3925-3932
    • /
    • 2024
  • During outages at nuclear power plants, much more care for radiation workers against internal exposure should be ensured given that more hot particles exist relative to the amount during normal operation. If fuel-type hot particles (FTHP) are inhaled, they can cause more severe health risks compared to activation-type hot particles (ATHP), which contain 60Co, due to the alpha-emitting nuclides within FTHPs. The activities of difficult-to-measure nuclides within FTHPs inhaled by workers are inferred by the age-dating technique using a141Ce/144Ce ratio as measured by whole-body counters. However, this method may be limited to outages that last for only a few months due to the short half-life (32.5 days) of 141Ce. We studied the feasibility of utilizing 241Am, a nuclide with a long half-life of 432.6 years, as an alternative to 141Ce. Additionally, we improved the performance of a stand-type whole-body counter for low-energy gamma spectroscopy to meet the criterion (RMSE ≤0.25) specified in ANSI/HPS N13.30-2011 by employing an artificial neural network (ANN). This study can contribute to more rapid and accurate internal dose assessments for workers who have inhaled FTHPs during long-term outages at nuclear power plants.

A Study on Optimal Output Neuron Allocation of LVQ Neural Network using Variance Estimation (분산추정에 의한 LVQ 신경회로망의 최적 출력뉴런 분할에 관한 연구)

  • 정준원;조성원
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1996.10a
    • /
    • pp.239-242
    • /
    • 1996
  • 본 논문에서는 BP(Back Propagation)에 비해서 빠른 학습시간과 다른 경쟁학습 신경회로망 알고리즘에 비해서 비교적 우수한 성능으로 패턴인식 등에 많이 이용되고 있는 LVQ(Learning Vector Quantization) 알고리즘의 성능을 향상시키기 위한 방법을 논의하고자 한다. 일반적으로 LVQ는 음(negative)의 학습을 하기 때문에 초기 가중치가 제대로 설정되지 않으면 발산할 수 있다는 단점이 있으며, 경쟁학습 계열의 신경망이기 때문에 출력 층의 뉴런 수에 따라 성능에 큰 영향을 받는다고 알려져 있다.[1]. 지도학습 형태를 지닌 LVQ의 경우에 학습패턴이 n개의 클래스를 가지고, 각 클래스 별로 학습패턴의 수가 같은 경우에 일반적으로 전체 출력뉴런에 대해서 (출력뉴런수/n)개의 뉴런을 각 클래스의 목표(desired) 클러스터로 할당하여 학습을 수행하는데, 본 논문에서는 각 클래스에 동일한 수의 출력뉴런을 할당하지 않고, 학습데이터에서 각 클래스의 분산을 추정하여 각 클래스의 분산을 추정분산에 비례하게 목표 출력뉴런을 할당하고, 초기 가중치도 추정분산에 비례하게 각 클래스의 초기 임의 위치 입력백터를 사용하여 학습을 수행하는 방법을 제안한다. 본 논문에서 제안하는 방법은 분류하고자 하는 데이터에 대해서 필요한 최적의 출력뉴런 수를 찾는 것이 아니라 이미 결정되어 있는 출력뉴런 수에 대해서 각 클래스에 할당할 출력 뉴런 수를 데이터의 추정분산에 의해서 결정하는 것으로, 추정분산이 크면 상대적으로 많은 출력 뉴런을 할당하고 작으면 상대적으로 적은 출력뉴런을 할당하고 초기 가중치도 마찬가지 방법으로 결정하며, 이렇게 하면 정해진 출력뉴런 개수 안에서 각 클래스 별로 분류의 어려움에 따라서 출력뉴런을 할당하기 때문에 미학습 뉴런이 줄어들게 되어 성능의 향상을 기대할 수 있으며, 실험적으로 제안된 방법이 더 나은 성능을 보임을 확인했다.initially they expected a more practical program about planting than programs that teach community design. Many people are active in their own towns to create better environments and communities. The network system "Alpha Green-Net" is functional to support graduates of the course. In the future these educational programs for citizens will becomes very important. Other cities are starting to have their own progrms, but they are still very short term. "Alpha Green-Net" is in the process of growing. Many members are very keen to develop their own abilities. In the future these NPOs should become independent. To help these NPOs become independent and active the educational programs should consider and teach about how to do this more in the future.단하였는데 그 결과, 좌측 촉각엽에서 제4형의 신경연접이 퇴행성 변화를 나타내었다. 그러므로 촉각의 지각신경세포는 뇌의 같은 족 촉각엽에 뻗어와 제4형 신경연접을 형성한다고 결론되었다.$/ 값이 210 $\mu\textrm{g}$/$m\ell$로서 효과적인 저해 활성을 나타내었다 따라서, 본 연구에서 빈

  • PDF

A Quality Prediction Model for Ginseng Sprouts based on CNN (CNN을 활용한 새싹삼의 품질 예측 모델 개발)

  • Lee, Chung-Gu;Jeong, Seok-Bong
    • Journal of the Korea Society for Simulation
    • /
    • v.30 no.2
    • /
    • pp.41-48
    • /
    • 2021
  • As the rural population continues to decline and aging, the improvement of agricultural productivity is becoming more important. Early prediction of crop quality can play an important role in improving agricultural productivity and profitability. Although many researches have been conducted recently to classify diseases and predict crop yield using CNN based deep learning and transfer learning technology, there are few studies which predict postharvest crop quality early in the planting stage. In this study, a early quality prediction model is proposed for sprout ginseng, which is drawing attention as a healthy functional foods. For this end, we took pictures of ginseng seedlings in the planting stage and cultivated them through hydroponic cultivation. After harvest, quality data were labeled by classifying the quality of ginseng sprout. With this data, we build early quality prediction models using several pre-trained CNN models through transfer learning technology. And we compare the prediction performance such as learning period and accuracy between each model. The results show more than 80% prediction accuracy in all proposed models, especially ResNet152V2 based model shows the highest accuracy. Through this study, it is expected that it will be able to contribute to production and profitability by automating the existing seedling screening works, which primarily rely on manpower.

Artificial Intelligence for Assistance of Facial Expression Practice Using Emotion Classification (감정 분류를 이용한 표정 연습 보조 인공지능)

  • Dong-Kyu, Kim;So Hwa, Lee;Jae Hwan, Bong
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.6
    • /
    • pp.1137-1144
    • /
    • 2022
  • In this study, an artificial intelligence(AI) was developed to help with facial expression practice in order to express emotions. The developed AI used multimodal inputs consisting of sentences and facial images for deep neural networks (DNNs). The DNNs calculated similarities between the emotions predicted by the sentences and the emotions predicted by facial images. The user practiced facial expressions based on the situation given by sentences, and the AI provided the user with numerical feedback based on the similarity between the emotion predicted by sentence and the emotion predicted by facial expression. ResNet34 structure was trained on FER2013 public data to predict emotions from facial images. To predict emotions in sentences, KoBERT model was trained in transfer learning manner using the conversational speech dataset for emotion classification opened to the public by AIHub. The DNN that predicts emotions from the facial images demonstrated 65% accuracy, which is comparable to human emotional classification ability. The DNN that predicts emotions from the sentences achieved 90% accuracy. The performance of the developed AI was evaluated through experiments with changing facial expressions in which an ordinary person was participated.

A Multi-speaker Speech Synthesis System Using X-vector (x-vector를 이용한 다화자 음성합성 시스템)

  • Jo, Min Su;Kwon, Chul Hong
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.4
    • /
    • pp.675-681
    • /
    • 2021
  • With the recent growth of the AI speaker market, the demand for speech synthesis technology that enables natural conversation with users is increasing. Therefore, there is a need for a multi-speaker speech synthesis system that can generate voices of various tones. In order to synthesize natural speech, it is required to train with a large-capacity. high-quality speech DB. However, it is very difficult in terms of recording time and cost to collect a high-quality, large-capacity speech database uttered by many speakers. Therefore, it is necessary to train the speech synthesis system using the speech DB of a very large number of speakers with a small amount of training data for each speaker, and a technique for naturally expressing the tone and rhyme of multiple speakers is required. In this paper, we propose a technology for constructing a speaker encoder by applying the deep learning-based x-vector technique used in speaker recognition technology, and synthesizing a new speaker's tone with a small amount of data through the speaker encoder. In the multi-speaker speech synthesis system, the module for synthesizing mel-spectrogram from input text is composed of Tacotron2, and the vocoder generating synthesized speech consists of WaveNet with mixture of logistic distributions applied. The x-vector extracted from the trained speaker embedding neural networks is added to Tacotron2 as an input to express the desired speaker's tone.

Estimation of Displacements Using Artificial Intelligence Considering Spatial Correlation of Structural Shape (구조형상 공간상관을 고려한 인공지능 기반 변위 추정)

  • Seung-Hun Shin;Ji-Young Kim;Jong-Yeol Woo;Dae-Gun Kim;Tae-Seok Jin
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.1
    • /
    • pp.1-7
    • /
    • 2023
  • An artificial intelligence (AI) method based on image deep learning is proposed to predict the entire displacement shape of a structure using the feature of partial displacements. The performance of the method was investigated through a structural test of a steel frame. An image-to-image regression (I2IR) training method was developed based on the U-Net layer for image recognition. In the I2IR method, the U-Net is modified to generate images of entire displacement shapes when images of partial displacement shapes of structures are input to the AI network. Furthermore, the training of displacements combined with the location feature was developed so that nodal displacement values with corresponding nodal coordinates could be used in AI training. The proposed training methods can consider correlations between nodal displacements in 3D space, and the accuracy of displacement predictions is improved compared with artificial neural network training methods. Displacements of the steel frame were predicted during the structural tests using the proposed methods and compared with 3D scanning data of displacement shapes. The results show that the proposed AI prediction properly follows the measured displacements using 3D scanning.

Comparative Analysis of Self-supervised Deephashing Models for Efficient Image Retrieval System (효율적인 이미지 검색 시스템을 위한 자기 감독 딥해싱 모델의 비교 분석)

  • Kim Soo In;Jeon Young Jin;Lee Sang Bum;Kim Won Gyum
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.12
    • /
    • pp.519-524
    • /
    • 2023
  • In hashing-based image retrieval, the hash code of a manipulated image is different from the original image, making it difficult to search for the same image. This paper proposes and evaluates a self-supervised deephashing model that generates perceptual hash codes from feature information such as texture, shape, and color of images. The comparison models are autoencoder-based variational inference models, but the encoder is designed with a fully connected layer, convolutional neural network, and transformer modules. The proposed model is a variational inference model that includes a SimAM module of extracting geometric patterns and positional relationships within images. The SimAM module can learn latent vectors highlighting objects or local regions through an energy function using the activation values of neurons and surrounding neurons. The proposed method is a representation learning model that can generate low-dimensional latent vectors from high-dimensional input images, and the latent vectors are binarized into distinguishable hash code. From the experimental results on public datasets such as CIFAR-10, ImageNet, and NUS-WIDE, the proposed model is superior to the comparative model and analyzed to have equivalent performance to the supervised learning-based deephashing model. The proposed model can be used in application systems that require low-dimensional representation of images, such as image search or copyright image determination.

Incremental Image Noise Reduction in Coronary CT Angiography Using a Deep Learning-Based Technique with Iterative Reconstruction

  • Jung Hee Hong;Eun-Ah Park;Whal Lee;Chulkyun Ahn;Jong-Hyo Kim
    • Korean Journal of Radiology
    • /
    • v.21 no.10
    • /
    • pp.1165-1177
    • /
    • 2020
  • Objective: To assess the feasibility of applying a deep learning-based denoising technique to coronary CT angiography (CCTA) along with iterative reconstruction for additional noise reduction. Materials and Methods: We retrospectively enrolled 82 consecutive patients (male:female = 60:22; mean age, 67.0 ± 10.8 years) who had undergone both CCTA and invasive coronary artery angiography from March 2017 to June 2018. All included patients underwent CCTA with iterative reconstruction (ADMIRE level 3, Siemens Healthineers). We developed a deep learning based denoising technique (ClariCT.AI, ClariPI), which was based on a modified U-net type convolutional neural net model designed to predict the possible occurrence of low-dose noise in the originals. Denoised images were obtained by subtracting the predicted noise from the originals. Image noise, CT attenuation, signal-to-noise ratio (SNR), and contrast-to-noise ratio (CNR) were objectively calculated. The edge rise distance (ERD) was measured as an indicator of image sharpness. Two blinded readers subjectively graded the image quality using a 5-point scale. Diagnostic performance of the CCTA was evaluated based on the presence or absence of significant stenosis (≥ 50% lumen reduction). Results: Objective image qualities (original vs. denoised: image noise, 67.22 ± 25.74 vs. 52.64 ± 27.40; SNR [left main], 21.91 ± 6.38 vs. 30.35 ± 10.46; CNR [left main], 23.24 ± 6.52 vs. 31.93 ± 10.72; all p < 0.001) and subjective image quality (2.45 ± 0.62 vs. 3.65 ± 0.60, p < 0.001) improved significantly in the denoised images. The average ERDs of the denoised images were significantly smaller than those of originals (0.98 ± 0.08 vs. 0.09 ± 0.08, p < 0.001). With regard to diagnostic accuracy, no significant differences were observed among paired comparisons. Conclusion: Application of the deep learning technique along with iterative reconstruction can enhance the noise reduction performance with a significant improvement in objective and subjective image qualities of CCTA images.

A Comparative Study on the Effective Deep Learning for Fingerprint Recognition with Scar and Wrinkle (상처와 주름이 있는 지문 판별에 효율적인 심층 학습 비교연구)

  • Kim, JunSeob;Rim, BeanBonyka;Sung, Nak-Jun;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.17-23
    • /
    • 2020
  • Biometric information indicating measurement items related to human characteristics has attracted great attention as security technology with high reliability since there is no fear of theft or loss. Among these biometric information, fingerprints are mainly used in fields such as identity verification and identification. If there is a problem such as a wound, wrinkle, or moisture that is difficult to authenticate to the fingerprint image when identifying the identity, the fingerprint expert can identify the problem with the fingerprint directly through the preprocessing step, and apply the image processing algorithm appropriate to the problem. Solve the problem. In this case, by implementing artificial intelligence software that distinguishes fingerprint images with cuts and wrinkles on the fingerprint, it is easy to check whether there are cuts or wrinkles, and by selecting an appropriate algorithm, the fingerprint image can be easily improved. In this study, we developed a total of 17,080 fingerprint databases by acquiring all finger prints of 1,010 students from the Royal University of Cambodia, 600 Sokoto open data sets, and 98 Korean students. In order to determine if there are any injuries or wrinkles in the built database, criteria were established, and the data were validated by experts. The training and test datasets consisted of Cambodian data and Sokoto data, and the ratio was set to 8: 2. The data of 98 Korean students were set up as a validation data set. Using the constructed data set, five CNN-based architectures such as Classic CNN, AlexNet, VGG-16, Resnet50, and Yolo v3 were implemented. A study was conducted to find the model that performed best on the readings. Among the five architectures, ResNet50 showed the best performance with 81.51%.

Application of Deep Learning for Classification of Ancient Korean Roof-end Tile Images (딥러닝을 활용한 고대 수막새 이미지 분류 검토)

  • KIM Younghyun
    • Korean Journal of Heritage: History & Science
    • /
    • v.57 no.3
    • /
    • pp.24-35
    • /
    • 2024
  • Recently, research using deep learning technologies such as artificial intelligence, convolutional neural networks, etc. has been actively conducted in various fields including healthcare, manufacturing, autonomous driving, and security, and is having a significant influence on society. In line with this trend, the present study attempted to apply deep learning to the classification of archaeological artifacts, specifically ancient Korean roof-end tiles. Using 100 images of roof-end tiles from each of the Goguryeo, Baekje, and Silla dynasties, for a total of 300 base images, a dataset was formed and expanded to 1,200 images using data augmentation techniques. After building a model using transfer learning from the pre-trained EfficientNetB0 model and conducting five-fold cross-validation, an average training accuracy of 98.06% and validation accuracy of 97.08% were achieved. Furthermore, when model performance was evaluated with a test dataset of 240 images, it could classify the roof-end tile images from the three dynasties with a minimum accuracy of 91%. In particular, with a learning rate of 0.0001, the model exhibited the highest performance, with accuracy of 92.92%, precision of 92.96%, recall of 92.92%, and F1 score of 92.93%. This optimal result was obtained by preventing overfitting and underfitting issues using various learning rate settings and finding the optimal hyperparameters. The study's findings confirm the potential for applying deep learning technologies to the classification of Korean archaeological materials, which is significant. Additionally, it was confirmed that the existing ImageNet dataset and parameters could be positively applied to the analysis of archaeological data. This approach could lead to the creation of various models for future archaeological database accumulation, the use of artifacts in museums, and classification and organization of artifacts.