• Title/Summary/Keyword: Web Service Extraction

Search Result 50, Processing Time 0.031 seconds

A Study on Extraction of Facade and Backside Component for Web Sevice Development (웹 서비스 개발을 위한 Facade와 Backside 컴포넌트 추출에 관한 연구)

  • Park, Dong-Sik;Shin, Ho-Jun;Kim, Soung-Won;Kim, Haeng-Kon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2003.11c
    • /
    • pp.1551-1554
    • /
    • 2003
  • 최근 SOA(Service Oriented Architecture)와 웹 서비스에 대한 관심이 높아져가고 있다. 웹 서비스는 재사용 측면에서 현재 사용되고 있는 컴포넌트와 많은 공통점이 있다. 웹 서비스의 개발과 수정 및 화장을 위해서 신뢰성 있는 컴포넌트를 이용한 경우 개발 시간의 단축 등의 많은 이점을 가질 수 있다. 이러한 컴포넌트를 이용하여 웹 서비스를 개발한 경우 웹 서비스 도메인에 대한 이해뿐만 아니라 요구사항의 추출 및 식별은 매우 중요한 부분이다. 본 논문에서는 사용자 요구사항으로부터 컴포넌트를 식별하고 이들 컴포넌트를 서비스 지원 컴포넌트인 Facade 컴포넌트와 Backside 컴포넌트로 식별할 수 있는 방법을 제시하고자 한다. 이를 위해 서비스를 오퍼레이션과 데이터 요소로 정의하고, 정의된 요소를 사용한 모델을 제시한다. 이러한 컴포넌트의 식별은 현재 활성화되어 있는 컴포넌트를 사용할 수 있도록 함으로써 웹 서비스 개발의 생산성을 보장하며, 웹 서비스를 컴포넌트로 구성함으로서 컴포넌트의 기본적인 특징인 재사용과 대체성을 통해 웹 서비스의 신뢰성과 수정의 용이성 등을 얻을 수 있다.

  • PDF

A Study on the System Development for Automatic Webtoon Download and Feature Registration (웹툰 저작물 자동 수집 및 특징점 등록 시스템 개발에 관한 연구)

  • Yoon, Hee-Don;Kim, Tae-Hyun;Kang, Ho-Gab;Cho, Seong-Hwan
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.17 no.2
    • /
    • pp.25-31
    • /
    • 2017
  • This paper aims to develope the technology for monitoring and identifying webtoon works that are being illegally distributed through the web-hard, torrent and overseas websites. The technology is to effectively prevent the illegal distribution and protect the copyrights of the webtoon works, leading to create a healthy market. This paper deals with the technology for (a) automatic collection of webtoon works from the webtoon service sites, (b) extraction of the features from the images of the collected webtoon works, and (c) registers them in the feature database. For these purposes, the required system performance and features were analyzed and the structures of webtoon service sites were examined, through which the system for automatic registration of webtoon works and features was developed. The system can work for the early stage blockage of illegal distributions of webtoon works which occur during the period from webtoon posting to feature registration.

3D Building Model Texture Extraction from Multiple Spatial Imagery for 3D City Modeling (3차원 도시모델 생성을 위한 다중 공간영상 기반 건물 모델 텍스쳐 추출)

  • Oh, Jae-Hong;Shin, Sung-Woong;Park, Jin-Ho;Lee, Hyo-Seong
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.25 no.4
    • /
    • pp.347-354
    • /
    • 2007
  • Since large portal service providers started web services for 3D city models around the world using spatial imagery, the competition has been getting intense to provide the models with the higher quality and accuracy. The building models are the most in number among the 3D city model objects, and it takes much time and money to create realistic model due to various shapes and visual appearances of building object. The aforementioned problem is the most significant limitation for the service and the update of the 3D city model of the large area. This study proposed a method of generating realistic 3D building models with quick and economical texture mapping using multiple spatial imagery such as aerial photos or satellite images after reconstructed geometric models of buildings from building layers in digital maps. Based on the experimental results, the suggested method has effectiveness for the generation of the 3D building models using various air-borne imagery and satellite imagery quickly and economically.

A Study on Disaster Information Support using Big Data (빅 데이터를 이용한 재해 정보 지원에 관한 연구)

  • Shin, Bong-Hi;Jeon, Hye-Kyoung
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.8
    • /
    • pp.25-32
    • /
    • 2018
  • Recently, the size and type of disasters in Korea has been diversified. However, Korea has not been able to build various information support systems to predict these disasters.Many other organizations also provide relevant information. This information is mainly provided on the Web, but most of it is not real time information. In this study, we have paid attention to support information using big data to provide better quality real - time information together with information provided by institutions. Big data has a large amount of information with real-time property, and it can make customized service using it. Among them, SNS such as Twitter and Facebook can be used as a new information collection medium in case of disaster. However, it is very difficult to retrieve necessary information from too much information, and it is difficult to collect intuitive information. For this purpose, this study develops an information support system using Twitter. The system retrieves information using the Twitter hashtag. Also, information mapping is performed on the map so that intuitive information can be grasped. For system evaluation, information extraction, degree of mapping, and recommendation speed are evaluated.

Metadata extraction using AI and advanced metadata research for web services (AI를 활용한 메타데이터 추출 및 웹서비스용 메타데이터 고도화 연구)

  • Sung Hwan Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.2
    • /
    • pp.499-503
    • /
    • 2024
  • Broadcasting programs are provided to various media such as Internet replay, OTT, and IPTV services as well as self-broadcasting. In this case, it is very important to provide keywords for search that represent the characteristics of the content well. Broadcasters mainly use the method of manually entering key keywords in the production process and the archive process. This method is insufficient in terms of quantity to secure core metadata, and also reveals limitations in recommending and using content in other media services. This study supports securing a large number of metadata by utilizing closed caption data pre-archived through the DTV closed captioning server developed in EBS. First, core metadata was automatically extracted by applying Google's natural language AI technology. The next step is to propose a method of finding core metadata by reflecting priorities and content characteristics as core research contents. As a technology to obtain differentiated metadata weights, the importance was classified by applying the TF-IDF calculation method. Successful weight data were obtained as a result of the experiment. The string metadata obtained by this study, when combined with future string similarity measurement studies, becomes the basis for securing sophisticated content recommendation metadata from content services provided to other media.

Water leakage accident analysis of water supply networks using big data analysis technique (R기반 빅데이터 분석기법을 활용한 상수도시스템 누수사고 분석)

  • Hong, Sung-Jin;Yoo, Do-Guen
    • Journal of Korea Water Resources Association
    • /
    • v.55 no.spc1
    • /
    • pp.1261-1270
    • /
    • 2022
  • The purpose of this study is to collect and analyze information related to water leaks that cannot be easily accessed, and utilized by using the news search results that people can easily access. We applied a web crawling technique for extracting big data news on water leakage accidents in the water supply system and presented an algorithm in a procedural way to obtain accurate leak accident news. In addition, a data analysis technique suitable for water leakage accident information analysis was developed so that additional information such as the date and time of occurrence, cause of occurrence, location of occurrence, damaged facilities, damage effect. The primary goal of value extraction through big data-based leak analysis proposed in this study is to extract a meaningful value through comparison with the existing waterworks statistical results. In addition, the proposed method can be used to effectively respond to consumers or determine the service level of water supply networks. In other words, the presentation of such analysis results suggests the need to inform the public of information such as accidents a little more, and can be used in conjunction to prepare a radio wave and response system that can quickly respond in case of an accident.

A Study of Integrating ASP Databases with Customer Databases (ASP 용의 데이터베이스와 고객 데이터베이스 연동에 관한 연구)

  • Kim, Ho-Yoon;Lee, Jae-Won
    • The KIPS Transactions:PartD
    • /
    • v.11D no.5
    • /
    • pp.1063-1072
    • /
    • 2004
  • In the ASP(Application Service Provider) business, applications using database sometimes require some data from clients' databases. These days such data are extracted from client database using manual database operations as an EXCEL file and the ASP, once receiving this file, transfers it into the application's database using manual database operations. This paper describes how to deal with data transmitting between the client database and ASP database on the web without using database manual operations for data extraction and insertion. We propose a framework which enables to transmit client data in a systematical way, to match different attribute names of each database for sharing same attribute values, and to avoid exposing information about the network path of client database to the ASP. This approach consists of two steps of data processing. The first is extracting data from client database as XML format by using a downloaded client program from ASP site, the second is uploading and storing the XML file into the ASP database. The implemented prototype system shows the suggested data integration paradigm is valid and ASP business needing integration of client database can be activated using it.

Twitter Issue Tracking System by Topic Modeling Techniques (토픽 모델링을 이용한 트위터 이슈 트래킹 시스템)

  • Bae, Jung-Hwan;Han, Nam-Gi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.109-122
    • /
    • 2014
  • People are nowadays creating a tremendous amount of data on Social Network Service (SNS). In particular, the incorporation of SNS into mobile devices has resulted in massive amounts of data generation, thereby greatly influencing society. This is an unmatched phenomenon in history, and now we live in the Age of Big Data. SNS Data is defined as a condition of Big Data where the amount of data (volume), data input and output speeds (velocity), and the variety of data types (variety) are satisfied. If someone intends to discover the trend of an issue in SNS Big Data, this information can be used as a new important source for the creation of new values because this information covers the whole of society. In this study, a Twitter Issue Tracking System (TITS) is designed and established to meet the needs of analyzing SNS Big Data. TITS extracts issues from Twitter texts and visualizes them on the web. The proposed system provides the following four functions: (1) Provide the topic keyword set that corresponds to daily ranking; (2) Visualize the daily time series graph of a topic for the duration of a month; (3) Provide the importance of a topic through a treemap based on the score system and frequency; (4) Visualize the daily time-series graph of keywords by searching the keyword; The present study analyzes the Big Data generated by SNS in real time. SNS Big Data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. In addition, such analysis requires the latest big data technology to process rapidly a large amount of real-time data, such as the Hadoop distributed system or NoSQL, which is an alternative to relational database. We built TITS based on Hadoop to optimize the processing of big data because Hadoop is designed to scale up from single node computing to thousands of machines. Furthermore, we use MongoDB, which is classified as a NoSQL database. In addition, MongoDB is an open source platform, document-oriented database that provides high performance, high availability, and automatic scaling. Unlike existing relational database, there are no schema or tables with MongoDB, and its most important goal is that of data accessibility and data processing performance. In the Age of Big Data, the visualization of Big Data is more attractive to the Big Data community because it helps analysts to examine such data easily and clearly. Therefore, TITS uses the d3.js library as a visualization tool. This library is designed for the purpose of creating Data Driven Documents that bind document object model (DOM) and any data; the interaction between data is easy and useful for managing real-time data stream with smooth animation. In addition, TITS uses a bootstrap made of pre-configured plug-in style sheets and JavaScript libraries to build a web system. The TITS Graphical User Interface (GUI) is designed using these libraries, and it is capable of detecting issues on Twitter in an easy and intuitive manner. The proposed work demonstrates the superiority of our issue detection techniques by matching detected issues with corresponding online news articles. The contributions of the present study are threefold. First, we suggest an alternative approach to real-time big data analysis, which has become an extremely important issue. Second, we apply a topic modeling technique that is used in various research areas, including Library and Information Science (LIS). Based on this, we can confirm the utility of storytelling and time series analysis. Third, we develop a web-based system, and make the system available for the real-time discovery of topics. The present study conducted experiments with nearly 150 million tweets in Korea during March 2013.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

An Analysis of IT Trends Using Tweet Data (트윗 데이터를 활용한 IT 트렌드 분석)

  • Yi, Jin Baek;Lee, Choong Kwon;Cha, Kyung Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.143-159
    • /
    • 2015
  • Predicting IT trends has been a long and important subject for information systems research. IT trend prediction makes it possible to acknowledge emerging eras of innovation and allocate budgets to prepare against rapidly changing technological trends. Towards the end of each year, various domestic and global organizations predict and announce IT trends for the following year. For example, Gartner Predicts 10 top IT trend during the next year, and these predictions affect IT and industry leaders and organization's basic assumptions about technology and the future of IT, but the accuracy of these reports are difficult to verify. Social media data can be useful tool to verify the accuracy. As social media services have gained in popularity, it is used in a variety of ways, from posting about personal daily life to keeping up to date with news and trends. In the recent years, rates of social media activity in Korea have reached unprecedented levels. Hundreds of millions of users now participate in online social networks and communicate with colleague and friends their opinions and thoughts. In particular, Twitter is currently the major micro blog service, it has an important function named 'tweets' which is to report their current thoughts and actions, comments on news and engage in discussions. For an analysis on IT trends, we chose Tweet data because not only it produces massive unstructured textual data in real time but also it serves as an influential channel for opinion leading on technology. Previous studies found that the tweet data provides useful information and detects the trend of society effectively, these studies also identifies that Twitter can track the issue faster than the other media, newspapers. Therefore, this study investigates how frequently the predicted IT trends for the following year announced by public organizations are mentioned on social network services like Twitter. IT trend predictions for 2013, announced near the end of 2012 from two domestic organizations, the National IT Industry Promotion Agency (NIPA) and the National Information Society Agency (NIA), were used as a basis for this research. The present study analyzes the Twitter data generated from Seoul (Korea) compared with the predictions of the two organizations to analyze the differences. Thus, Twitter data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. To overcome these challenges, we used SAS IRS (Information Retrieval Studio) developed by SAS to capture the trend in real-time processing big stream datasets of Twitter. The system offers a framework for crawling, normalizing, analyzing, indexing and searching tweet data. As a result, we have crawled the entire Twitter sphere in Seoul area and obtained 21,589 tweets in 2013 to review how frequently the IT trend topics announced by the two organizations were mentioned by the people in Seoul. The results shows that most IT trend predicted by NIPA and NIA were all frequently mentioned in Twitter except some topics such as 'new types of security threat', 'green IT', 'next generation semiconductor' since these topics non generalized compound words so they can be mentioned in Twitter with other words. To answer whether the IT trend tweets from Korea is related to the following year's IT trends in real world, we compared Twitter's trending topics with those in Nara Market, Korea's online e-Procurement system which is a nationwide web-based procurement system, dealing with whole procurement process of all public organizations in Korea. The correlation analysis show that Tweet frequencies on IT trending topics predicted by NIPA and NIA are significantly correlated with frequencies on IT topics mentioned in project announcements by Nara market in 2012 and 2013. The main contribution of our research can be found in the following aspects: i) the IT topic predictions announced by NIPA and NIA can provide an effective guideline to IT professionals and researchers in Korea who are looking for verified IT topic trends in the following topic, ii) researchers can use Twitter to get some useful ideas to detect and predict dynamic trends of technological and social issues.