• Title/Summary/Keyword: Web-page

Search Result 675, Processing Time 0.025 seconds

C-rank: A Contribution-Based Approach for Web Page Ranking (C-rank: 웹 페이지 랭킹을 위한 기여도 기반 접근법)

  • Lee, Sang-Chul;Kim, Dong-Jin;Son, Ho-Yong;Kim, Sang-Wook;Lee, Jae-Bum
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.1
    • /
    • pp.100-104
    • /
    • 2010
  • In the past decade, various search engines have been developed to retrieve web pages that web surfers want to find from world wide web. In search engines, one of the most important functions is to evaluate and rank web pages for a given web surfer query. The prior algorithms using hyperlink information like PageRank incur the problem of 'topic drift'. To solve the problem, relevance propagation models have been proposed. However, these models suffer from serious performance degradation, and thus cannot be employed in real search engines. In this paper, we propose a new ranking algorithm that alleviates the topic drift problem and also provides efficient performance. Through a variety of experiments, we verify the superiority of the proposed algorithm over prior ones.

Web Log Analysis Using Support Vector Regression

  • Jun, Sung-Hae;Lim, Min-Taik;Jorn, Hong-Seok;Hwang, Jin-Soo;Park, Seong-Yong;Kim, Jee-Yun;Oh, Kyung-Whan
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.1
    • /
    • pp.61-77
    • /
    • 2003
  • Due to the wide expansion of the internet, people can freely get information what they want with lesser efforts. However without adequate forms or rules to follow, it is getting more and more difficult to get necessary information. Because of seemingly chaotic status of the current web environment, it is sometimes called "Dizzy web" The user should wander from page to page to get necessary information. Therefore we need to construct system which properly recommends appropriate information for general user. The representative research field for this system is called Recommendation System(RS), The collaborative recommendation system is one of the RS. It was known to perform better than the other systems. When we perform the web user modeling or other web-mining tasks, the continuous feedback data is very important and frequently used. In this paper, we propose a collaborative recommendation system which can deal with the continuous feedback data and tried to construct the web page prediction system. We use a sojourn time of a user as continuous feedback data and combine the traditional model-based algorithm framework with the Support Vector Regression technique. In our experiments, we show the accuracy of our system and the computing time of page prediction compared with Pearson's correlation algorithm.algorithm.

WebPR : A Dynamic Web Page Recommendation Algorithm Based on Mining Frequent Traversal Patterns (WebPR :빈발 순회패턴 탐사에 기반한 동적 웹페이지 추천 알고리즘)

  • Yoon, Sun-Hee;Kim, Sam-Keun;Lee, Chang-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.11B no.2
    • /
    • pp.187-198
    • /
    • 2004
  • The World-Wide Web is the largest distributed Information space and has grown to encompass diverse information resources. However, although Web is growing exponentially, the individual's capacity to read and digest contents is essentially fixed. From the view point of Web users, they can be confused by explosion of Web information, by constantly changing Web environments, and by lack of understanding needs of Web users. In these Web environments, mining traversal patterns is an important problem in Web mining with a host of application domains including system design and Information services. Conventional traversal pattern mining systems use the inter-pages association in sessions with only a very restricted mechanism (based on vector or matrix) for generating frequent k-Pagesets. We develop a family of novel algorithms (termed WebPR - Web Page Recommend) for mining frequent traversal patterns and then pageset to recommend. Our algorithms provide Web users with new page views, which Include pagesets to recommend, so that users can effectively traverse its Web site. The main distinguishing factors are both a point consistently spanning schemes applying inter-pages association for mining frequent traversal patterns and a point proposing the most efficient tree model. Our experimentation with two real data sets, including Lady Asiana and KBS media server site, clearly validates that our method outperforms conventional methods.

Design and Implementation of an Interestingness Analysis System for Web Personalizatoion & Customization

  • Jung, Youn-Hong;Kim, I-I;Park, Kyoo-seok
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.4
    • /
    • pp.707-713
    • /
    • 2003
  • Convenience and promptness of the internet have been not only making the electronic commerce grow rapidly in case of website, analyzing a navigation pattern of the users has been also making personalization and customization techniques develop rapidly for providing service accordant to individual interestingness. Web personalization and customization skill has been utilizing various methods, such as web log mining to use web log data and web mining to use the transaction of users etc, especially e-CRM analyzing a navigation pattern of the users. In this paper, We measure exact duration time of the users in web page and web site, compute weight about duration time each page, and propose a way to comprehend e-loyalty through the computed weight.

  • PDF

Optimized Web Design Method by Analyzing the Websites (웹사이트 분석을 통한 최적화 설계 방안)

  • Jang, Hee-Seon
    • Convergence Security Journal
    • /
    • v.15 no.2
    • /
    • pp.19-24
    • /
    • 2015
  • As the Internet usage such as Web3.0, future internet, and internet of things increases, the big data through information exchange between the users and web servers increases. Analyzing those web data, the commercial web sites use the analytic results for marketing and campaign, and non-commercial web sites also use the results to improve the user's services satisfaction. In this paper, the quantitative index is presented to analyze the web sites, and optimized web site design method is also presented through the correlation analysis of index and significance test. From the results for 138 web sites, it is observed that strong plus(+) correlation for visits-unique visitors and page views-average visit duration exists. We also observe the minus(-) correlation between bounce rate and page views per user(or ratio of new visits). In specific, to reduce the bounce rate for users, the strategy to increase the page views and ratio of new visits rather than visits and unique visitors is needed.

Design of Web Robot Engine Using Distributed Collection Model Processing (분산수집 모델을 이용한 웹 로봇의 설계 및 구현)

  • Kim, Dae-Yu;Kim, Jung-Tae
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.1
    • /
    • pp.115-121
    • /
    • 2010
  • As internet becomes widespread, a lot of information is opened to public and users of Internet can access effectively information using web searching service. To construct web searching service, the web searching method for collecting of information is needed to obtain web page view. As a number of web page view increases, it is necessary to collect information of high quality information to be searched, therefore, a variety of web engine for searching mechanism is developed. Method of link extraction with javascript in dynamic web page and design of web searching robot are presented m this paper. To evaluate performance analyzes, we fixed one searching model with the proposed method. The searching time takes 2 minute 67 sec for 299 web pages and 12.33 sec for 10 searching model.

Crawling algorithm design and experiment for automatic deep web document collection (심층 웹 문서 자동 수집을 위한 크롤링 알고리즘 설계 및 실험)

  • Yun-Jeong, Kang;Min-Hye, Lee;Dong-Hyun, Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.27 no.1
    • /
    • pp.1-7
    • /
    • 2023
  • Deep web collection means entering a query in a search form and collecting response results. It is estimated that the information possessed by the deep web has about 450 to 550 times more information than the statically constructed surface web. The static method does not show the changed information until the web page is refreshed, but the dynamic web page method updates the necessary information in real time and provides real-time information without reloading the web page, but crawler has difficulty accessing the updated information. Therefore, there is a need for a way to automatically collect information on these deep webs using a crawler. Therefore, this paper proposes a method of utilizing scripts as general links, and for this purpose, an algorithm that can utilize client scripts like regular URLs is proposed and experimented. The proposed algorithm focused on collecting web information by menu navigation and script execution instead of the usual method of entering data into search forms.

A Study on Design and Development of Web Information Collection System Based Compare and Merge Method (웹 페이지 비교통합 기반의 정보 수집 시스템 설계 및 개발에 대한 연구)

  • Jang, Jin-Wook
    • Journal of Information Technology Services
    • /
    • v.13 no.1
    • /
    • pp.147-159
    • /
    • 2014
  • Recently, the quantity of information that is accessible from the Internet is being dramatically increased. Searching the Web for useful information has therefore become increasingly difficult. Thus, much research has been done on web robots which perform internet information filtering based on user interest. If a web site which users want to visit is found, its content is searched by following the searching list or Web sites links in order. This search process takes a long time according as the number of page or site increases so that its performance need to be improved. In order to minimize unnecessary search with web robots, this paper proposes an efficient information collection system based on compare and merge method. In the proposed system, a web robot initially collects information from web sites which users register. From the next visit to the web sites, the web robot compares what it collected with what the web sites have currently. If they are different, the web robot updates what it collected. Only updated web page information is classified according to subject and provided to users so that users can access the updated information quickly.

Security Check Scheduling for Detecting Malicious Web Sites (악성사이트 검출을 위한 안전진단 스케줄링)

  • Choi, Jae Yeong;Kim, Sung Ki;Min, Byoung Joon
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.2 no.9
    • /
    • pp.405-412
    • /
    • 2013
  • Current web has evolved to a mashed-up format according to the change of the implementation and usage patterns. Web services and user experiences have improved, however, security threats are also increased as the web contents that are not yet verified combine together. To mitigate the threats incurred as an adverse effect of the web development, we need to check security on the combined web contents. In this paper, we propose a scheduling method to detect malicious web pages not only inside but also outside through extended links for secure operation of a web site. The scheduling method considers several aspects of each page including connection popularity, suspiciousness, and check elapse time to make a decision on the order for security check on numerous web pages connected with links. We verified the effectiveness of the security check complying with the scheduling method that uses the priority given to each page.

Design Automation System for Concrete Slab Bridge using Vector Graphic (벡터 그래픽을 이용한 콘크리트 슬래브교의 설계자동화)

  • Lho, Byeong Cheol;Kim, Jeong Hoon;Kim, Jong Woo
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.12 no.2
    • /
    • pp.181-192
    • /
    • 2008
  • A Web-based design automation system for concrete slab bridge has been developed to manage the resultant documents as well as to speed up the repetitive design process. Manipulation of engineering drawings in the Web page is one of the critical functions needed for Web-based design automation. Active Server Pages (ASP) is used to that collects design parameters in the database, and eXtensible Markup Language (XML) and XML-based vector graphics are expected to facilitate the representation of engineering drawings in the Web page. In this paper, we present how we used XML and Scalable Vector Graphics (SVG) to compose engineering drawings and represent them in the Web page. XML Data Island we designed to define drawing components turned out effective in manipulating the engineering drawings in the Web page.