• 제목/요약/키워드: deep similarity

Search Result 223, Processing Time 0.025 seconds

Improving the effectiveness of document extraction summary based on the amount of sentence information (문장 정보량 기반 문서 추출 요약의 효과성 제고)

  • Kim, Eun Hee;Lim, Myung Jin;Shin, Ju Hyun
    • Smart Media Journal
    • /
    • v.11 no.3
    • /
    • pp.31-38
    • /
    • 2022
  • In the document extraction summary study, various methods for selecting important sentences based on the relationship between sentences were proposed. In the Korean document summary using the summation similarity of sentences, the summation similarity of the sentences was regarded as the amount of sentence information, and the summary sentences were extracted by selecting important sentences based on this. However, the problem is that it does not take into account the various importance that each sentence contributes to the entire document. Therefore, in this study, we propose a document extraction summary method that provides a summary by selecting important sentences based on the amount of quantitative and semantic information in the sentence. As a result, the extracted sentence agreement was 58.56% and the ROUGE-L score was 34, which was superior to the method using only the combined similarity. Compared to the deep learning-based method, the extraction method is lighter, but the performance is similar. Through this, it was confirmed that the method of compressing information based on semantic similarity between sentences is an important approach in document extraction summary. In addition, based on the quickly extracted summary, the document generation summary step can be effectively performed.

A Study on Safety Management Improvement System Using Similarity Inspection Technique (유사도검사 기법을 이용한 안전관리 개선시스템 연구)

  • Park, Koo-Rack
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.4
    • /
    • pp.23-29
    • /
    • 2018
  • To reduce the accident rate caused by the delay of corrective action, which is common in the construction site, in order to shorten the time from correcting the existing system to the corrective action, I used a time similarity check to inform the inspectors of the problem in real time, modeling the system so that corrective action can be performed immediately on site, and studied a system that can actively cope with safety accidents. The research result shows that there is more than 90% opening effect and more than 60% safety accident reduction rate. I will continue to study more effective system combining voice recognition and deep learning based on this system.

Experimental Investigation on the Flow Control of Hub Clutch for Automobile (자동차용 허브 클러치의 유동제어에 관한 실험적 연구)

  • 박종남;김동환;김병민
    • Transactions of Materials Processing
    • /
    • v.11 no.5
    • /
    • pp.430-438
    • /
    • 2002
  • This paper suggests the new technology to control metal flow in orther to change of the cold forging from conventional deep drawing forming. This technology can be summarized the complex forming, which consists of bulk forming and sheet forming, and multi-action forging, which be performed double action press. The proposed technology is applied to hub clutch model which is part of auto-transmission for automobile. The purpose of this study is to investigate the material flow behavior of hub clutch through control the relative velocity ratio and the stroke of mandrel and punch using the flow forming technique. First of all, the finite element simulations are applied to analyse optimal process conditions to prevent flow defect(necking defect etc.) from non-uniform metal flow, then the results are compared with the plasticine model material experiments. The punch load for real material is predict from similarity law. Finally, the model material experiment results are in good agreement with the FE simulation ones.

Effect of Input Data Video Interval and Input Data Image Similarity on Learning Accuracy in 3D-CNN

  • Kim, Heeil;Chung, Yeongjee
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.2
    • /
    • pp.208-217
    • /
    • 2021
  • 3D-CNN is one of the deep learning techniques for learning time series data. However, these three-dimensional learning can generate many parameters, requiring high performance or having a significant impact on learning speed. We will use these 3D-CNNs to learn hand gesture and find the parameters that showed the highest accuracy, and then analyze how the accuracy of 3D-CNN varies through input data changes without any structural changes in 3D-CNN. First, choose the interval of the input data. This adjusts the ratio of the stop interval to the gesture interval. Secondly, the corresponding interframe mean value is obtained by measuring and normalizing the similarity of images through interclass 2D cross correlation analysis. This experiment demonstrates that changes in input data affect learning accuracy without structural changes in 3D-CNN. In this paper, we proposed two methods for changing input data. Experimental results show that input data can affect the accuracy of the model.

Semantic Similarity Calculation based on Siamese TRAT (트랜스포머 인코더와 시암넷 결합한 시맨틱 유사도 알고리즘)

  • Lu, Xing-Cen;Joe, Inwhee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.05a
    • /
    • pp.397-400
    • /
    • 2021
  • To solve the problem that existing computing methods cannot adequately represent the semantic features of sentences, Siamese TRAT, a semantic feature extraction model based on Transformer encoder is proposed. The transformer model is used to fully extract the semantic information within sentences and carry out deep semantic coding for sentences. In addition, the interactive attention mechanism is introduced to extract the similar features of the association between two sentences, which makes the model better at capturing the important semantic information inside the sentence. As a result, it improves the semantic understanding and generalization ability of the model. The experimental results show that the proposed model can improve the accuracy significantly for the semantic similarity calculation task of English and Chinese, and is more effective than the existing methods.

Zero-shot voice conversion with HuBERT

  • Hyelee Chung;Hosung Nam
    • Phonetics and Speech Sciences
    • /
    • v.15 no.3
    • /
    • pp.69-74
    • /
    • 2023
  • This study introduces an innovative model for zero-shot voice conversion that utilizes the capabilities of HuBERT. Zero-shot voice conversion models can transform the speech of one speaker to mimic that of another, even when the model has not been exposed to the target speaker's voice during the training phase. Comprising five main components (HuBERT, feature encoder, flow, speaker encoder, and vocoder), the model offers remarkable performance across a range of scenarios. Notably, it excels in the challenging unseen-to-unseen voice-conversion tasks. The effectiveness of the model was assessed based on the mean opinion scores and similarity scores, reflecting high voice quality and similarity to the target speakers. This model demonstrates considerable promise for a range of real-world applications demanding high-quality voice conversion. This study sets a precedent in the exploration of HuBERT-based models for voice conversion, and presents new directions for future research in this domain. Despite its complexities, the robust performance of this model underscores the viability of HuBERT in advancing voice conversion technology, making it a significant contributor to the field.

A comparative study on keypoint detection for developmental dysplasia of hip diagnosis using deep learning models in X-ray and ultrasound images (X-ray 및 초음파 영상을 활용한 고관절 이형성증 진단을 위한 특징점 검출 딥러닝 모델 비교 연구)

  • Sung-Hyun Kim;Kyungsu Lee;Si-Wook Lee;Jin Ho Chang;Jae Youn Hwang;Jihun Kim
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.5
    • /
    • pp.460-468
    • /
    • 2023
  • Developmental Dysplasia of the Hip (DDH) is a pathological condition commonly occurring during the growth phase of infants. It acts as one of the factors that can disrupt an infant's growth and trigger potential complications. Therefore, it is critically important to detect and treat this condition early. The traditional diagnostic methods for DDH involve palpation techniques and diagnosis methods based on the detection of keypoints in the hip joint using X-ray or ultrasound imaging. However, there exist limitations in objectivity and productivity during keypoint detection in the hip joint. This study proposes a deep learning model-based keypoint detection method using X-ray and ultrasound imaging and analyzes the performance of keypoint detection using various deep learning models. Additionally, the study introduces and evaluates various data augmentation techniques to compensate the lack of medical data. This research demonstrated the highest keypoint detection performance when applying the residual network 152 (ResNet152) model with simple & complex augmentation techniques, with average Object Keypoint Similarity (OKS) of approximately 95.33 % and 81.21 % in X-ray and ultrasound images, respectively. These results demonstrate that the application of deep learning models to ultrasound and X-ray images to detect the keypoints in the hip joint could enhance the objectivity and productivity in DDH diagnosis.

Performance Evaluation of Automatic Segmentation based on Deep Learning and Atlas according to CT Image Acquisition Conditions (CT 영상획득 조건에 따른 딥 러닝과 아틀라스 기반의 자동분할 성능 평가)

  • Jung Hoon Kim
    • Journal of the Korean Society of Radiology
    • /
    • v.18 no.3
    • /
    • pp.213-222
    • /
    • 2024
  • This study analyzed the volumes generated by deep learning and atlas-based automatic segmentation methods, as well as the Dice similarity coefficient and 95% Hausdorff distance, according to the conditions of conduction voltage and conduction current in computed tomography for lung radiotherapy. The first result, the volumes generated by the atlas-based smart segmentation method showed the smallest volume change as a function of the change in tube voltage and tube current, while Aview RT ACS and OncoStudio using deep learning showed smaller volumes at tube currents lower than 100 mA. The second result, the Dice similarity coefficient, showed that Aview RT ACS was 2% higher than OncoStuido, and the 95% Hausdorff distance results also showed that Aview RT ACS analyzed an average of 0.2-0.5% higher than OncoStudio. However, the standard deviation of the respective results for tube current and tube voltage is lower for OncoStudio, which suggests that the results are consistent across volume variations. Therefore, caution should be exercised when using deep learning-based automatic segmentation programs at low perfusion voltages and low perfusion currents in CT imaging conditions for lung radiotherapy, and similar results were obtained with conventional atlas-based automatic segmentation programs at certain perfusion voltages and perfusion currents.

Comparison of CNN and GAN-based Deep Learning Models for Ground Roll Suppression (그라운드-롤 제거를 위한 CNN과 GAN 기반 딥러닝 모델 비교 분석)

  • Sangin Cho;Sukjoon Pyun
    • Geophysics and Geophysical Exploration
    • /
    • v.26 no.2
    • /
    • pp.37-51
    • /
    • 2023
  • The ground roll is the most common coherent noise in land seismic data and has an amplitude much larger than the reflection event we usually want to obtain. Therefore, ground roll suppression is a crucial step in seismic data processing. Several techniques, such as f-k filtering and curvelet transform, have been developed to suppress the ground roll. However, the existing methods still require improvements in suppression performance and efficiency. Various studies on the suppression of ground roll in seismic data have recently been conducted using deep learning methods developed for image processing. In this paper, we introduce three models (DnCNN (De-noiseCNN), pix2pix, and CycleGAN), based on convolutional neural network (CNN) or conditional generative adversarial network (cGAN), for ground roll suppression and explain them in detail through numerical examples. Common shot gathers from the same field were divided into training and test datasets to compare the algorithms. We trained the models using the training data and evaluated their performances using the test data. When training these models with field data, ground roll removed data are required; therefore, the ground roll is suppressed by f-k filtering and used as the ground-truth data. To evaluate the performance of the deep learning models and compare the training results, we utilized quantitative indicators such as the correlation coefficient and structural similarity index measure (SSIM) based on the similarity to the ground-truth data. The DnCNN model exhibited the best performance, and we confirmed that other models could also be applied to suppress the ground roll.

A Deep Learning Application for Automated Feature Extraction in Transaction-based Machine Learning (트랜잭션 기반 머신러닝에서 특성 추출 자동화를 위한 딥러닝 응용)

  • Woo, Deock-Chae;Moon, Hyun Sil;Kwon, Suhnbeom;Cho, Yoonho
    • Journal of Information Technology Services
    • /
    • v.18 no.2
    • /
    • pp.143-159
    • /
    • 2019
  • Machine learning (ML) is a method of fitting given data to a mathematical model to derive insights or to predict. In the age of big data, where the amount of available data increases exponentially due to the development of information technology and smart devices, ML shows high prediction performance due to pattern detection without bias. The feature engineering that generates the features that can explain the problem to be solved in the ML process has a great influence on the performance and its importance is continuously emphasized. Despite this importance, however, it is still considered a difficult task as it requires a thorough understanding of the domain characteristics as well as an understanding of source data and the iterative procedure. Therefore, we propose methods to apply deep learning for solving the complexity and difficulty of feature extraction and improving the performance of ML model. Unlike other techniques, the most common reason for the superior performance of deep learning techniques in complex unstructured data processing is that it is possible to extract features from the source data itself. In order to apply these advantages to the business problems, we propose deep learning based methods that can automatically extract features from transaction data or directly predict and classify target variables. In particular, we applied techniques that show high performance in existing text processing based on the structural similarity between transaction data and text data. And we also verified the suitability of each method according to the characteristics of transaction data. Through our study, it is possible not only to search for the possibility of automated feature extraction but also to obtain a benchmark model that shows a certain level of performance before performing the feature extraction task by a human. In addition, it is expected that it will be able to provide guidelines for choosing a suitable deep learning model based on the business problem and the data characteristics.