• Title/Summary/Keyword: 자동 데이터 구축

Search Result 716, Processing Time 0.023 seconds

A Document Collection Method for More Accurate Search Engine (정확도 높은 검색 엔진을 위한 문서 수집 방법)

  • Ha, Eun-Yong;Gwon, Hui-Yong;Hwang, Ho-Yeong
    • The KIPS Transactions:PartA
    • /
    • v.10A no.5
    • /
    • pp.469-478
    • /
    • 2003
  • Internet information search engines using web robots visit servers conneted to the Internet periodically or non-periodically. They extract and classify data collected according to their own method and construct their database, which are the basis of web information search engines. There procedure are repeated very frequently on the Web. Many search engine sites operate this processing strategically to become popular interneet portal sites which provede users ways how to information on the web. Web search engine contacts to thousands of thousands web servers and maintains its existed databases and navigates to get data about newly connected web servers. But these jobs are decided and conducted by search engines. They run web robots to collect data from web servers without knowledge on the states of web servers. Each search engine issues lots of requests and receives responses from web servers. This is one cause to increase internet traffic on the web. If each web server notify web robots about summary on its public documents and then each web robot runs collecting operations using this summary to the corresponding documents on the web servers, the unnecessary internet traffic is eliminated and also the accuracy of data on search engines will become higher. And the processing overhead concerned with web related jobs on web servers and search engines will become lower. In this paper, a monitoring system on the web server is designed and implemented, which monitors states of documents on the web server and summarizes changes of modified documents and sends the summary information to web robots which want to get documents from the web server. And an efficient web robot on the web search engine is also designed and implemented, which uses the notified summary and gets corresponding documents from the web servers and extracts index and updates its databases.

Study of the UAV for Application Plans and Landscape Analysis (UAV를 이용한 경관분석 및 활용방안에 관한 기초연구)

  • Kim, Seung-Min
    • Journal of the Korean Institute of Traditional Landscape Architecture
    • /
    • v.32 no.3
    • /
    • pp.213-220
    • /
    • 2014
  • This is the study to conduct the topographical analysis using the orthophotographic data from the waypoint flight using the UAV and constructed the system required for the automatic waypoint flight using the multicopter.. The results of the waypoint photographing are as follows. First, result of the waypoint flight over the area of 9.3ha, take time photogrammetry took 40 minutes in total. The multicopter have maintained the certain flight altitude and a constant speed that the accurate photographing was conducted over the waypoint determined by the ground station. Then, the effect of the photogrammetry was checked. Second, attached a digital camera to the multicopter which is lightweight and low in cost compared to the general photogrammetric unmanned airplane and then used it to check its mobility and economy. In addition, the matching of the photo data, and production of DEM and DXF files made it possible to analyze the topography. Third, produced the high resolution orthophoto(2cm) for the inside of the river and found out that the analysis is possible for the changes in vegetation and topography around the river. Fourth, It would be used for the more in-depth research on landscape analysis such as terrain analysis and visibility analysis. This method may be widely used to analyze the various terrains in cities and rivers. It can also be used for the landscape control such as cultural remains and tourist sites as well as the control of the cultural and historical resources such as the visibility analysis for the construction of DSM.

Spatial Distribution of Urban Heat and Pollution Islands using Remote Sensing and Private Automated Meteorological Observation System Data -Focused on Busan Metropolitan City, Korea- (위성영상과 민간자동관측시스템 자료를 활용한 도시열섬과 도시오염섬의 공간 분포 특성 - 부산광역시를 대상으로 -)

  • HWANG, Hee-Soo;KANG, Jung Eun
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.23 no.3
    • /
    • pp.100-119
    • /
    • 2020
  • During recent years, the heat environment and particulate matter (PM10) have become serious environmental problems, as increases in heat waves due to rising global temperature interact with weakening atmospheric wind speeds. There exist urban heat islands and urban pollution islands with higher temperatures and air pollution concentrations than other areas. However, few studies have examined these issues together because of a lack of micro-scale data, which can be constructed from spatial data. Today, with the help of satellite images and big data collected by private telecommunication companies, detailed spatial distribution analyses are possible. Therefore, this study aimed to examine the spatial distribution patterns of urban heat islands and urban pollution islands within Busan Metropolitan City and to compare the distributions of the two phenomena. In this study, the land surface temperature of Landsat 8 satellite images, air temperature and particulate matter concentration data derived from a private automated meteorological observation system were gridded in 30m × 30m units, and spatial analysis was performed. Analysis showed that simultaneous zones of urban heat islands and urban pollution islands included some vulnerable residential areas and industrial areas. The political migration areas such as Seo-dong and Bansong-dong, representative vulnerable residential areas in Busan, were included in the co-occurring areas. The areas have a high density of buildings and poor ventilation, most of whose residents are vulnerable to heat waves and air pollution; thus, these areas must be considered first when establishing related policies. In the industrial areas included in the co-occurring areas, concrete or asphalt concrete-based impervious surfaces accounted for an absolute majority, and not only was the proportion of vegetation insufficient, there was also considerable vehicular traffic. A hot-spot analysis examining the reliability of the analysis confirmed that more than 99.96% of the regions corresponded to hot-spot areas at a 99% confidence level.

Analysis of Building Characteristics and Temporal Changes of Fire Alarms (건물 특성과 시간적 변화가 소방시설관리시스템의 화재알람에 미치는 영향 분석 연구)

  • Lim, Gwanmuk;Ko, Seoltae;Kim, Yoosin;Park, Keon Chul
    • Journal of Internet Computing and Services
    • /
    • v.22 no.4
    • /
    • pp.83-98
    • /
    • 2021
  • The purpose of this study to find the factors influencing the fire alarms using IoT firefighting facility management system data of Seoul Fire & Disaster Headquarters, and to present academic implications for establishing an effective prevention system of fire situation. As the number of high and complex buildings increases and former bulidings are advanced, the fire detection facilities that can quickly respond to emergency situations are also increasing. However, if the accuracy of the fire situation is incorrectly detected and the accuracy is lowered, the inconvenience of the residents increases and the reliability decreases. Therefore, it is necessary to improve accuracy of the system through efficient inspection and the internal environment investigation of buildings. The purpose of this study is to find out that false detection may occur due to building characteristics such as usage or time, and to aim of emphasizing the need for efficient system inspection and controlling the internal environment. As a result, it is found that the size(total area) of the building had the greatest effect on the fire alarms, and the fire alarms increased as private buildings, R-type receivers, and a large number of failure or shutoff days. In addition, factors that influencing fire alarms were different depending on the main usage of the building. In terms of time, it was found to follow people's daily patterns during weekdays(9 am to 6 pm), and each peaked around 10 am and 2 pm. This study was claimed that it is necessary to investigate the building environment that caused the fire alarms, along with the system internal inspection. Also, it propose additional recording of building environment data in real-time for follow-up research and system enhancement.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.

Implementation of Smart Shopping Cart using Object Detection Method based on Deep Learning (딥러닝 객체 탐지 기술을 사용한 스마트 쇼핑카트의 구현)

  • Oh, Jin-Seon;Chun, In-Gook
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.7
    • /
    • pp.262-269
    • /
    • 2020
  • Recently, many attempts have been made to reduce the time required for payment in various shopping environments. In addition, for the Fourth Industrial Revolution era, artificial intelligence is advancing, and Internet of Things (IoT) devices are becoming more compact and cheaper. So, by integrating these two technologies, access to building an unmanned environment to save people time has become easier. In this paper, we propose a smart shopping cart system based on low-cost IoT equipment and deep-learning object-detection technology. The proposed smart cart system consists of a camera for real-time product detection, an ultrasonic sensor that acts as a trigger, a weight sensor to determine whether a product is put into or taken out of the shopping cart, an application for smartphones that provides a user interface for a virtual shopping cart, and a deep learning server where learned product data are stored. Communication between each module is through Transmission Control Protocol/Internet Protocol, a Hypertext Transmission Protocol network, a You Only Look Once darknet library, and an object detection system used by the server to recognize products. The user can check a list of items put into the smart cart via the smartphone app, and can automatically pay for them. The smart cart system proposed in this paper can be applied to unmanned stores with high cost-effectiveness.

Distance Measurement of Small Moving Object using Infrared Stereo Camera (적외선 스테레오 카메라를 이용한 소형 이동체의 거리 측정)

  • Oh, Jun-Ho;Lee, Sang-Hwa;Lee, Boo-Hwan;Park, Jong-Il
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.49 no.3
    • /
    • pp.53-61
    • /
    • 2012
  • This paper proposes a real-time distance measurement system of high temperature and high speed target using infrared stereo camera. We construct an infrared stereo camera system that measure the difference between target and background temperatures for automatic target measurement. First, the proposed method detects target region based on target motion and intensity variation of local region using difference between target and background temperatures. Second, stereo matching by left and right target information is used to estimate disparity about real-time distance of target. In the proposed method using infrared stereo camera system, we compare distances in three dimension trajectory measuring instrument and in infrared stereo camera measurement. In this experiment from three video data, the result shows an average 9.68% distance error rate. The proposed method is suitable for distance and position measurement of varied targets using infrared stereo system.

Bio-Signal Detection Monitoring System Using ZigBee and Wireless Network (거리측정 센서 스캐닝과 퍼지 제어를 이용한 생체신호 모니터링 전동 휠체어 자율주행 시스템)

  • Kim, Kuk-Se;Yang, Sang-Gi;Rasheed, M.Tahir;Ahn, Seong-Soo;Lee, Joon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2008.10a
    • /
    • pp.331-339
    • /
    • 2008
  • Nowadays with advancement in technology and aging society, the number of disabled citizens is increasing. The disabled citizens always need a caretaker for daily life routines especially for mobility. In future, the need is considered to increase more. To reduce the burden from the disabled, various devices for healthcare are introduced using computer technology. The power wheelchair is an important and convenient mobility device. The demand of power wheelchair is increasing for assistance in mobility. In this paper we proposed a robotic wheelchair for mobility aid to reduce the burden from the disabled. The main issue in an autonomous wheelchair is the automatic detection and avoidance of obstacles and going to the pre-designated place. The proposed algorithm detects the obstacles and avoids them to drive the wheelchair to the desired place safely. By this way, the disabled will not always have to worry about paying deep attention to the surroundings and his path. User has a handheld bio-sensor monitoring system for get user's bio-signal. If user detects unusual signal, alarm send to protector.

  • PDF

WordNet-Based Category Utility Approach for Author Name Disambiguation (저자명 모호성 해결을 위한 개념망 기반 카테고리 유틸리티)

  • Kim, Je-Min;Park, Young-Tack
    • The KIPS Transactions:PartB
    • /
    • v.16B no.3
    • /
    • pp.225-232
    • /
    • 2009
  • Author name disambiguation is essential for improving performance of document indexing, retrieval, and web search. Author name disambiguation resolves the conflict when multiple authors share the same name label. This paper introduces a novel approach which exploits ontologies and WordNet-based category utility for author name disambiguation. Our method utilizes author knowledge in the form of populated ontology that uses various types of properties: titles, abstracts and co-authors of papers and authors' affiliation. Author ontology has been constructed in the artificial intelligence and semantic web areas semi-automatically using OWL API and heuristics. Author name disambiguation determines the correct author from various candidate authors in the populated author ontology. Candidate authors are evaluated using proposed WordNet-based category utility to resolve disambiguation. Category utility is a tradeoff between intra-class similarity and inter-class dissimilarity of author instances, where author instances are described in terms of attribute-value pairs. WordNet-based category utility has been proposed to exploit concept information in WordNet for semantic analysis for disambiguation. Experiments using the WordNet-based category utility increase the number of disambiguation by about 10% compared with that of category utility, and increase the overall amount of accuracy by around 98%.

Three-Phase English Syntactic Analysis for Improving the Parsing Efficiency (영어 구문 분석의 효율 개선을 위한 3단계 구문 분석)

  • Kim, Sung-Dong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.1
    • /
    • pp.21-28
    • /
    • 2016
  • The performance of an English-Korean machine translation system depends heavily on its English parser. The parser in this paper is a part of the rule-based English-Korean MT system, which includes many syntactic rules and performs the chart-based parsing. The parser generates too many structures due to many syntactic rules, so much time and memory are required. The rule-based parser has difficulty in analyzing and translating the long sentences including the commas because they cause high parsing complexity. In this paper, we propose the 3-phase parsing method with sentence segmentation to efficiently translate the long sentences appearing in usual. Each phase of the syntactic analysis applies its own independent syntactic rules in order to reduce parsing complexity. For the purpose, we classify the syntactic rules into 3 classes and design the 3-phase parsing algorithm. Especially, the syntactic rules in the 3rd class are for the sentence structures composed with commas. We present the automatic rule acquisition method for 3rd class rules from the syntactic analysis of the corpus, with which we aim to continuously improve the coverage of the parsing. The experimental results shows that the proposed 3-phase parsing method is superior to the prior parsing method using only intra-sentence segmentation in terms of the parsing speed/memory efficiency with keeping the translation quality.