• Title/Summary/Keyword: Text Detection

Search Result 400, Processing Time 0.025 seconds

A Study on the Improvement of DTW with Speech Silence Detection (음성의 묵음구간 검출을 통한 DTW의 성능개선에 관한 연구)

  • Kim, Jong-Kuk;Jo, Wang-Rae;Bae, Myung-Jin
    • Speech Sciences
    • /
    • v.10 no.4
    • /
    • pp.117-124
    • /
    • 2003
  • Speaker recognition is the technology that confirms the identification of speaker by using the characteristic of speech. Such technique is classified into speaker identification and speaker verification: The first method discriminates the speaker from the preregistered group and recognize the word, the second verifies the speaker who claims the identification. This method that extracts the information of speaker from the speech and confirms the individual identification becomes one of the most efficient technology as the service via telephone network is popularized. Some problems, however, must be solved for the real application as follows; The first thing is concerning that the safe method is necessary to reject the imposter because the recognition is not performed for the only preregistered customer. The second thing is about the fact that the characteristic of speech is changed as time goes by, So this fact causes the severe degradation of recognition rate and the inconvenience of users as the number of times to utter the text increases. The last thing is relating to the fact that the common characteristic among speakers causes the wrong recognition result. The silence parts being included the center of speech cause that identification rate is decreased. In this paper, to make improvement, We proposed identification rate can be improved by removing silence part before processing identification algorithm. The methods detecting speech area are zero crossing rate, energy of signal detect end point and starting point of the speech and process DTW algorithm by using two methods in this paper. As a result, the proposed method is obtained about 3% of improved recognition rate compare with the conventional methods.

  • PDF

Subnet Selection Scheme based on probability to enhance process speed of Big Data (빅 데이터의 처리속도 향상을 위한 확률기반 서브넷 선택 기법)

  • Jeong, Yoon-Su;Kim, Yong-Tae;Park, Gil-Cheol
    • Journal of Digital Convergence
    • /
    • v.13 no.9
    • /
    • pp.201-208
    • /
    • 2015
  • With services such as SNS and facebook, Big Data popularize the use of small size such as micro blogs are increasing. However, the problem of accuracy and computational cost of the search result of big data of a small size is unresolved. In this paper, we propose a subnet selection techniques based probability to improve the browsing speed of the small size of the text information from big data environments, such as micro-blogs. The proposed method is to configure the subnets to give to the attribute information of the data increased the probability data search speed. In addition, the proposed method improves the accessibility of the data by processing a pair of the connection information between the probability of the data constituting the subnet to easily access the distributed data. Experimental results showed the proposed method is 6.8% higher detection rates than CELF algorithm, the average processing time was reduced by 8.2%.

Linguistic Features Discrimination for Social Issue Risk Classification (사회적 이슈 리스크 유형 분류를 위한 어휘 자질 선별)

  • Oh, Hyo-Jung;Yun, Bo-Hyun;Kim, Chan-Young
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.11
    • /
    • pp.541-548
    • /
    • 2016
  • The use of social media is already essential as a source of information for listening user's various opinions and monitoring. We define social 'risks' that issues effect negative influences for public opinion in social media. This paper aims to discriminate various linguistic features and reveal their effects for building an automatic classification model of social risks. Expecially we adopt a word embedding technique for representation of linguistic clues in risk sentences. As a preliminary experiment to analyze characteristics of individual features, we revise errors in automatic linguistic analysis. At the result, the most important feature is NE (Named Entity) information and the best condition is when combine basic linguistic features. word embedding, and word clusters within core predicates. Experimental results under the real situation in social bigdata - including linguistic analysis errors - show 92.08% and 85.84% in precision respectively for frequent risk categories set and full test set.

The Analysis of Chosun Danasty Poetry Using 3D Data Visualization (3D 시각화를 이용한 조선시대 시문 분석)

  • Min, Kyoung-Ju;Lee, Byoung-Chan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.7
    • /
    • pp.861-868
    • /
    • 2021
  • With the development of technology for visualizing big-data, tasks such as intuitively analyzing a lot of data, detecting errors, and deriving meaning are actively progressing. In this paper, we describe the design and implementation of a 3D analysis that collects and stores the writing data in Chinese characters provided by the Korean Classical Database of the Korean Classics Translation Institute, stores and progress the data, and visualizes the writing information in a 3D network diagram. It solves the problem when a large amount of data is expressed in 2D, intuitive that analysis, error detection, meaningful data extraction such as characteristics, similarity, differences, etc. and user convenience can be provided. In this paper, we improved the problems of analyzing Chosun dynasty poetry in Chinese characters using 2D visualization conducted in previous studies.

Analysis of News Big Data for Deriving Social Issues in Korea (한국의 사회적 이슈 도출을 위한 뉴스 빅데이터 분석 연구)

  • Lee, Hong Joo
    • The Journal of Society for e-Business Studies
    • /
    • v.24 no.3
    • /
    • pp.163-182
    • /
    • 2019
  • Analyzing the frequency and correlation of the news keywords in the modern society that are becoming complicated according to the time flow is a very important research to discuss the response and solution to issues. This paper analyzed the relationship between the flow of social keyword and major issues through the analysis of news big data for 10 years (2009~2018). In this study, political issues, education and social culture, gender conflicts and social problems were presented as major issues. And, to study the change and flow of issues, it analyzed the change of the issue by dividing it into five years. Through this, the changes and countermeasures of social issues were studied. As a result, the keywords (economy, police) that are closely related to the people's life were analyzed as keywords that are very important in our society regardless of the flow of time. In addition, keyword such as 'safety' have decreased in increasing rate compared to frequency in recent years. Through this, it can be inferred that it is necessary to improve the awareness of safety in our society.

Resolution Estimation Technique in Gaze Tracking System for HCI (HCI를 위한 시선추적 시스템에서 분해능의 추정기법)

  • Kim, Ki-Bong;Choi, Hyun-Ho
    • Journal of Convergence for Information Technology
    • /
    • v.11 no.1
    • /
    • pp.20-27
    • /
    • 2021
  • Eye tracking is one of the NUI technologies, and it finds out where the user is gazing. This technology allows users to input text or control GUI, and further analyzes the user's gaze so that it can be applied to commercial advertisements. In the eye tracking system, the allowable range varies depending on the quality of the image and the degree of freedom of movement of the user. Therefore, there is a need for a method of estimating the accuracy of eye tracking in advance. The accuracy of eye tracking is greatly affected by how the eye tracking algorithm is implemented in addition to hardware variables. Accordingly, in this paper, we propose a method to estimate how many degrees of gaze changes when the pupil center moves by one pixel by estimating the maximum possible movement distance of the pupil center in the image.

Detection of Number and Character Area of License Plate Using Deep Learning and Semantic Image Segmentation (딥러닝과 의미론적 영상분할을 이용한 자동차 번호판의 숫자 및 문자영역 검출)

  • Lee, Jeong-Hwan
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.1
    • /
    • pp.29-35
    • /
    • 2021
  • License plate recognition plays a key role in intelligent transportation systems. Therefore, it is a very important process to efficiently detect the number and character areas. In this paper, we propose a method to effectively detect license plate number area by applying deep learning and semantic image segmentation algorithm. The proposed method is an algorithm that detects number and text areas directly from the license plate without preprocessing such as pixel projection. The license plate image was acquired from a fixed camera installed on the road, and was used in various real situations taking into account both weather and lighting changes. The input images was normalized to reduce the color change, and the deep learning neural networks used in the experiment were Vgg16, Vgg19, ResNet18, and ResNet50. To examine the performance of the proposed method, we experimented with 500 license plate images. 300 sheets were used for learning and 200 sheets were used for testing. As a result of computer simulation, it was the best when using ResNet50, and 95.77% accuracy was obtained.

Detection of Complaints of Non-Face-to-Face Work before and during COVID-19 by Using Topic Modeling and Sentiment Analysis (동적 토픽 모델링과 감성 분석을 이용한 COVID-19 구간별 비대면 근무 부정요인 검출에 관한 연구)

  • Lee, Sun Min;Chun, Se Jin;Park, Sang Un;Lee, Tae Wook;Kim, Woo Ju
    • The Journal of Information Systems
    • /
    • v.30 no.4
    • /
    • pp.277-301
    • /
    • 2021
  • Purpose The purpose of this study is to analyze the sentiment responses of the general public to non-face-to-face work using text mining methodology. As the number of non-face-to-face complaints is increasing over time, it is difficult to review and analyze in traditional methods such as surveys, and there is a limit to reflect real-time issues. Approach This study has proposed a method of the research model, first by collecting and cleansing the data related to non-face-to-face work among tweets posted on Twitter. Second, topics and keywords are extracted from tweets using LDA(Latent Dirichlet Allocation), a topic modeling technique, and changes for each section are analyzed through DTM(Dynamic Topic Modeling). Third, the complaints of non-face-to-face work are analyzed through the classification of positive and negative polarity in the COVID-19 section. Findings As a result of analyzing 1.54 million tweets related to non-face-to-face work, the number of IDs using non-face-to-face work-related words increased 7.2 times and the number of tweets increased 4.8 times after COVID-19. The top frequently used words related to non-face-to-face work appeared in the order of remote jobs, cybersecurity, technical jobs, productivity, and software. The words that have increased after the COVID-19 were concerned about lockdown and dismissal, and business transformation and also mentioned as to secure business continuity and virtual workplace. New Normal was newly mentioned as a new standard. Negative opinions found to be increased in the early stages of COVID-19 from 34% to 43%, and then stabilized again to 36% through non-face-to-face work sentiment analysis. The complaints were, policies such as strengthening cybersecurity, activating communication to improve work productivity, and diversifying work spaces.

A Robust Real-Time License Plate Recognition System Using Anchor-Free Method and Convolutional Neural Network

  • Kim, Dae-Hoon;Kim, Do-Hyeon;Lee, Dong-Hoon;Kim, Yoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.4
    • /
    • pp.19-26
    • /
    • 2022
  • With the recent development of intelligent transportation systems, car license plate recognition systems are being used in various fields. Such systems need to guarantee real-time performance to recognize the license plate of a driving car. Also, they should keep a high recognition rate even in problematic situations such as small license plates in low-resolution and unclear image due to distortion. In this paper, we propose a real-time car license plate recognition system that improved processing speed using object detection algorithm based on anchor-free method and text recognition algorithm based on Convolutional Neural Network(CNN). In addition, we used Spatial Transformer Network to increase the recognition rate on the low resolution or distorted images. We confirm that the proposed system is faster than previously existing car license plate recognition systems and maintains a high recognition rate in a variety of environment and quality images because the proposed system's recognition rate is 93.769% and the processing speed per image is about 0.006 seconds.

Diagnostic accuracy of imaging examinations for peri-implant bone defects around titanium and zirconium dioxide implants: A systematic review and meta-analysis

  • Chagas, Mariana Murai;Kobayashi-Velasco, Solange;Gimenez, Thais;Cavalcanti, Marcelo Gusmao Paraiso
    • Imaging Science in Dentistry
    • /
    • v.51 no.4
    • /
    • pp.363-372
    • /
    • 2021
  • Purpose: This systematic review and meta-analysis assessed the diagnostic accuracy of imaging examinations for the detection of peri-implant bone defects and compared the diagnostic accuracy between titanium (Ti) and zirconium dioxide (ZrO2) implants. Materials and Methods: Six online databases were searched, and studies were selected based on eligibility criteria. The studies included in the systematic review underwent bias and applicability assessment using the Quality Assessment of Diagnostic Accuracy Studies 2 (QUADAS-2) tool and a random-effect meta-analysis. Summary receiver operating characteristic (sROC) curves were constructed to compare the effect of methodological differences in relation to the variables of each group. Results: The search strategy yielded 719 articles. Titles and abstracts were read and 61 studies were selected for full-text reading. Among them, 24 studies were included in this systematic review. Most included studies had a low risk of bias (QUADAS-2). Cone-beam computed tomography (CBCT) presented sufficient data for quantitative analysis in ZrO2 and Ti implants. The meta-analysis revealed high levels of inconsistency in the latter group. Regarding sROC curves, the area under the curve (AUC) was larger for the overall Ti group (AUC=0.79) than for the overall ZrO2 group (AUC=0.69), but without a statistically significant difference between them. In Ti implants, the AUCs for dehiscence defects(0.73) and fenestration defects(0.87) showed a statistically significant difference. Conclusion: The diagnostic accuracy of CBCT imaging in the assessment of peri-implant bone defects was similar between Ti and ZrO2 implants, and fenestration was more accurately diagnosed than dehiscence in Ti implants.