• Title/Summary/Keyword: Web Search Data

Search Result 647, Processing Time 0.022 seconds

The Construction and Search of Environment GIS Data for WEB (WEB기반의 환경 GIS자료 구축과 검색)

  • 김창제
    • Spatial Information Research
    • /
    • v.5 no.2
    • /
    • pp.195-198
    • /
    • 1997
  • The his paper describes Management System of Environment GIS Data using JVM, JDBC, Oracle DB. Environment GIS data using the system are satellite Image, Thematic Map, GIS Data. The data have spatial information and attribute information. The search supplies spatial search using viewing-map on Web browser, nonspatial search using the attributes of data. The registration of data use the registration system of data on Web browser.

  • PDF

Implementation of Search Engine to Minimize Traffic Using Blockchain-Based Web Usage History Management System

  • Yu, Sunghyun;Yeom, Cheolmin;Won, Yoojae
    • Journal of Information Processing Systems
    • /
    • v.17 no.5
    • /
    • pp.989-1003
    • /
    • 2021
  • With the recent increase in the types of services provided by Internet companies, collection of various types of data has become a necessity. Data collectors corresponding to web services profit by collecting users' data indiscriminately and providing it to the associated services. However, the data provider remains unaware of the manner in which the data are collected and used. Furthermore, the data collector of a web service consumes web resources by generating a large amount of web traffic. This traffic can damage servers by causing service outages. In this study, we propose a website search engine that employs a system that controls user information using blockchains and builds its database based on the recorded information. The system is divided into three parts: a collection section that uses proxy, a management section that uses blockchains, and a search engine that uses a built-in database. This structure allows data sovereigns to manage their data more transparently. Search engines that use blockchains do not use internet bots, and instead use the data generated by user behavior. This avoids generation of traffic from internet bots and can, thereby, contribute to creating a better web ecosystem.

Personalized Search Service in Semantic Web (시멘틱 웹 환경에서의 개인화 검색)

  • Kim, Je-Min;Park, Young-Tack
    • The KIPS Transactions:PartB
    • /
    • v.13B no.5 s.108
    • /
    • pp.533-540
    • /
    • 2006
  • The semantic web environment promise semantic search of heterogeneous data from distributed web page. Semantic search would resuit in an overwhelming number of results for users is increased, therefore elevating the need for appropriate personalized ranking schemes. Culture Finder helps semantic web agents obtain personalized culture information. It extracts meta data for each web page(culture news, culture performance, culture exhibition), perform semantic search and compute result ranking point to base user profile. In order to work efficient, Culture Finder uses five major technique: Machine learning technique for generating user profile from user search behavior and meta data repository, an efficient semantic search system for semantic web agent, query analysis for representing query and query result, personalized ranking method to provide suitable search result to user, upper ontology for generating meta data. In this paper, we also present the structure used in the Culture Finder to support personalized search service.

Design and Implementation of Web Crawler utilizing Unstructured data

  • Tanvir, Ahmed Md.;Chung, Mokdong
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.3
    • /
    • pp.374-385
    • /
    • 2019
  • A Web Crawler is a program, which is commonly used by search engines to find the new brainchild on the internet. The use of crawlers has made the web easier for users. In this paper, we have used unstructured data by structuralization to collect data from the web pages. Our system is able to choose the word near our keyword in more than one document using unstructured way. Neighbor data were collected on the keyword through word2vec. The system goal is filtered at the data acquisition level and for a large taxonomy. The main problem in text taxonomy is how to improve the classification accuracy. In order to improve the accuracy, we propose a new weighting method of TF-IDF. In this paper, we modified TF-algorithm to calculate the accuracy of unstructured data. Finally, our system proposes a competent web pages search crawling algorithm, which is derived from TF-IDF and RL Web search algorithm to enhance the searching efficiency of the relevant information. In this paper, an attempt has been made to research and examine the work nature of crawlers and crawling algorithms in search engines for efficient information retrieval.

The comparative effectiveness and evaluation study of user groups of the various web search tools (다양한 형태의 웹 탐색도구의 이용자집단간 비교효용성 및 평가에 관한 연구)

  • 박일종;윤명순
    • Journal of Korean Library and Information Science Society
    • /
    • v.31 no.1
    • /
    • pp.87-114
    • /
    • 2000
  • The purpose of this study is offering appropriate system and training program to helf the system designer and the trainer in addition to analyze information use behavior about the web search tools and evaluate the estimated system by user groups. The results of the study are as follows $\circledS1$ It is desirable to consider age than other demographic variables in the case of web search tool. $\circledS2$ It is desirable to design Directory Search Tool in the case of web search tool which serves the student user group. $\circledS3$ An Intelligent Search Tool is more appropriate for the students who are using keyword search tool than any other tools. $\circledS4$ A discussion about standard classification of the web information should be accomplished soon because users feel confused in using web search tools due t o absence of standard mode of classification about classified item. $\circledS5$ Librarians need the cognition about data on internet s a source of information and need positive service and user training program about these information because student users hardly get help from librarians or library orientation for learning method to use web search tool. $\circledS6$ Internet use experience and years of computer use had effect on their use ability when using web search tool, whereas computer use experience, library use experience and Online Public Access Catalogs (OPAC) use experience had no effect on it. Especially, OPAC use experience had no effect on use ability of web search tool of student user group because student user groups had no information about internet and web search tool and they did not recognized the difference about search method between web search tool and OPAC. $\circledS7$In the case of web search tool, it si important to index the increasing web resource automatically by a searching robot. But in the case of student users, web search tool is much more needed to index by index expert due to the absence of ability about selecting and combining keyword.

  • PDF

The Development of Travel Demand Nowcasting Model Based on Travelers' Attention: Focusing on Web Search Traffic Information (여행자 관심 기반 스마트 여행 수요 예측 모형 개발: 웹검색 트래픽 정보를 중심으로)

  • Park, Do-Hyung
    • The Journal of Information Systems
    • /
    • v.26 no.3
    • /
    • pp.171-185
    • /
    • 2017
  • Purpose Recently, there has been an increase in attempts to analyze social phenomena, consumption trends, and consumption behavior through a vast amount of customer data such as web search traffic information and social buzz information in various fields such as flu prediction and real estate price prediction. Internet portal service providers such as google and naver are disclosing web search traffic information of online users as services such as google trends and naver trends. Academic and industry are paying attention to research on information search behavior and utilization of online users based on the web search traffic information. Although there are many studies predicting social phenomena, consumption trends, political polls, etc. based on web search traffic information, it is hard to find the research to explain and predict tourism demand and establish tourism policy using it. In this study, we try to use web search traffic information to explain the tourism demand for major cities in Gangwon-do, the representative tourist area in Korea, and to develop a nowcasting model for the demand. Design/methodology/approach In the first step, the literature review on travel demand and web search traffic was conducted in parallel in two directions. In the second stage, we conducted a qualitative research to confirm the information retrieval behavior of the traveler. In the next step, we extracted the representative tourist cities of Gangwon-do and confirmed which keywords were used for the search. In the fourth step, we collected tourist demand data to be used as a dependent variable and collected web search traffic information of each keyword to be used as an independent variable. In the fifth step, we set up a time series benchmark model, and added the web search traffic information to this model to confirm whether the prediction model improved. In the last stage, we analyze the prediction models that are finally selected as optimal and confirm whether the influence of the keywords on the prediction of travel demand. Findings This study has developed a tourism demand forecasting model of Gangwon-do, a representative tourist destination in Korea, by expanding and applying web search traffic information to tourism demand forecasting. We compared the existing time series model with the benchmarking model and confirmed the superiority of the proposed model. In addition, this study also confirms that web search traffic information has a positive correlation with travel demand and precedes it by one or two months, thereby asserting its suitability as a prediction model. Furthermore, by deriving search keywords that have a significant effect on tourism demand forecast for each city, representative characteristics of each region can be selected.

Design and Implementation of a Search Engine based on Apache Spark (아파치 스파크 기반 검색엔진의 설계 및 구현)

  • Park, Ki-Sung;Choi, Jae-Hyun;Kim, Jong-Bae;Park, Jae-Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.1
    • /
    • pp.17-28
    • /
    • 2017
  • Recently, a study on data has been actively conducted because the value of the data has become more useful. Web crawler that is program of data collection recently spotlighted because it can take advantage of the various fields. Web crawler can be defined as a tool to analyze the web pages and collects the URL by traversing the web server in an automated manner. For the treatment of Big-data, distributed Web crawler is widely used which is based on the Hadoop MapReduce. But, it is difficult to use and has constraints on the performance. Apache spark that is the In-memory computing platform is an alternative to MapReduce. The search engine which is one of the main purposes of web crawler displays the information you search by keyword gathered by web crawler. If search engines implement a spark-based web crawler instead of traditional MapReduce-based web crawler, it would be a more rapid data collection.

A Study on the Crawling and Classification Strategy for Local Website (로컬 웹사이트의 탐색전략과 웹사이트 유형분석에 관한 연구)

  • Hwang In-Soo
    • Journal of Information Technology Applications and Management
    • /
    • v.13 no.2
    • /
    • pp.55-65
    • /
    • 2006
  • Since the World-Wide Web (WWW) has become a major channel for information delivery, information overload also has become a serious problem to the Internet users. Therefore, effective information searching is critical to the success of Internet services. We present an integrated search engine for searching relevant web pages on the WWW in a certain Internet domain. It supports a local search on the web sites. The spider obtains all of the web pages from the web sites through web links. It operates autonomously without any human supervision. We developed state transition diagram to control navigation and analyze link structure of each web site. We have implemented an integrated local search engine and it shows that a higher satisfaction is obtained. From the user evaluation, we also find that higher precision is obtained.

  • PDF

A Document Collection Method for More Accurate Search Engine (정확도 높은 검색 엔진을 위한 문서 수집 방법)

  • Ha, Eun-Yong;Gwon, Hui-Yong;Hwang, Ho-Yeong
    • The KIPS Transactions:PartA
    • /
    • v.10A no.5
    • /
    • pp.469-478
    • /
    • 2003
  • Internet information search engines using web robots visit servers conneted to the Internet periodically or non-periodically. They extract and classify data collected according to their own method and construct their database, which are the basis of web information search engines. There procedure are repeated very frequently on the Web. Many search engine sites operate this processing strategically to become popular interneet portal sites which provede users ways how to information on the web. Web search engine contacts to thousands of thousands web servers and maintains its existed databases and navigates to get data about newly connected web servers. But these jobs are decided and conducted by search engines. They run web robots to collect data from web servers without knowledge on the states of web servers. Each search engine issues lots of requests and receives responses from web servers. This is one cause to increase internet traffic on the web. If each web server notify web robots about summary on its public documents and then each web robot runs collecting operations using this summary to the corresponding documents on the web servers, the unnecessary internet traffic is eliminated and also the accuracy of data on search engines will become higher. And the processing overhead concerned with web related jobs on web servers and search engines will become lower. In this paper, a monitoring system on the web server is designed and implemented, which monitors states of documents on the web server and summarizes changes of modified documents and sends the summary information to web robots which want to get documents from the web server. And an efficient web robot on the web search engine is also designed and implemented, which uses the notified summary and gets corresponding documents from the web servers and extracts index and updates its databases.

Classification of Web Search Engines and Necessity of a Hybrid Search Engine (웹 검색엔진 분류 및 하이브리드 검색엔진의 필요성)

  • Paik, Juryon
    • Journal of Digital Contents Society
    • /
    • v.19 no.4
    • /
    • pp.719-729
    • /
    • 2018
  • Abstract In 2017, it has been reported that Google had more than 90% of the market share in search-engines of desktops and mobiles. Most people may consider that Google surely searches the entire web area. However, according to many researches for web data, Google only searches less than 10%, surprisingly. The most region is called the Deep Web, and it is indexable by special search engines, which are different from Google because they focus on a specific segment of interest. Those engines build their own deep-web databases and run particular algorithms to provide accurate and professional search results. There is no search engine that indexes the entire Web, currently. The best way is to use several search engines together for broad and efficient searches as best as possible. This paper defines that kind of search engine as Hybrid Search Engine and provides characteristics and differences compared to conventional search engines, along with a frame of hybrid search engine.