• Title/Summary/Keyword: User-specific system

Search Result 721, Processing Time 0.028 seconds

Design and Implementation of Clipcast Service via Terrestrial DMB (지상파 DMB를 이용한 클립캐스트 서비스 설계 및 구현)

  • Cho, Suk-Hyun;Seo, Jong-Soo
    • Journal of Broadcast Engineering
    • /
    • v.16 no.1
    • /
    • pp.23-32
    • /
    • 2011
  • Design and Implementation of Clipcast Service via Terrestrial DMB This paper outlines the system design and the implementation process of clipcast service that can send clips of video, mp3, text, images, etc. to terrestrial DMB terminals. To provide clipcast service in terrestrial DMB, a separate data channel needs to be allocated and this requires changes in the existing bandwidth allocation. Clipcast contents can be sent after midnight at around 3 to 4 AM, when terrestrial DMB viewship is low. If the video service bit rate is lowered to 352 Kbps and the TPEG service band is fully used, then 320 Kbps bit rate can be allocated to clipcast. To enable clipcast service, the terminals' DMB program must be executed, and this can be done through SMS and EPG. Clipcast service applies MOT protocol to transmit multimedia objects, and transmits twice in carousel format for stable transmission of files. Therefore, 72Mbyte data can be transmitted in one hour, which corresponds to about 20 minutes of full motion video service at 500Kbps data rate. When running the clip transmitted through terrestrial DMB data channel, information regarding the length of each clip is received through communication with the CMS(Content Management Server), then error-free files are displayed. The clips can be provided to the users as preview contents of the complete VOD contents. In order to use the complete content, the user needs to access the URL allocated for that specific content and download the content by completing a billing process. This paper suggests the design and implementation of terrestrial DMB system to provide clipcast service, which enables file download services as provided in MediaFLO, DVB-H, and the other mobile broadcasting systems. Unlike the other mobile broadcasting systems, the proposed system applies more reliable SMS method to activate the DMB terminals for highly stable clipcast service. This allows hybrid, i.e, both SMS and EPG activations of terminals for clipcast services.

제주도 지하수자원의 최적 개발가능량 선정에 관한 수리지질학적 연구

  • 한정상;김창길;김남종;한규상
    • Proceedings of the Korean Society of Soil and Groundwater Environment Conference
    • /
    • 1994.07a
    • /
    • pp.184-215
    • /
    • 1994
  • The Hydrogeologic data of 455 water wells comprising geologic and aquifer test were analyzed to determine hydrogeoloic characteristics of Cheju island. The groundwater of Cheju island is occurred in unconsolidated pyroclastic deposits interbedded in highly jointed basaltic and andesic rocks as high level, basal and parabasal types order unconfined condition. The average transmissivity and specific yield of the aquifer are at about 29,300m$^2$/day and 0.12 respectively. The total storage of groundwater is estimated about 44 billion cubic meters(m$^3$). Average annual precipitation is about 3390 million m$^3$ among which average recharge amount is estimated 1494 million m$^3$ equivalent 44.1% of annual precipitation with 638 million m$^3$ of runoff and 1256 million m$^3$ of evapotranspiration. Based on groundwater budget analysis, the sustainable yield is about 620 million m$^3$(41% of annual recharge)and rest of it is discharging into the sea. The geologic logs of recently drilled thermal water wens indicate that very low-permeable marine sediments(Sehwa-ri formation) composed of loosely cemented sandy sat derived from mainly volcanic ashes, at the 1st stage volcanic activity of the area was situated at the 120$\pm$68m below sea level. And also the other low-permeable sedimentary rock called Segipo-formation which is deemed younger than former marine sediment is occured at the area covering north-west and western part of Cheju at the $\pm$70m below sea level. If these impermeable beds are distributed as a basal formation of fresh water zone of Cheju, most of groundwater in Cheju will be para-basal type. These formations will be one of the most important hydrogeologic boundary and groundwater occurences in the area.

  • PDF

A Study on the Development and usefulness of the x/y Plane and z Axis Resolution Phantom for MDCT Detector (MDCT 검출기의 x/y plane과 z축 분해능 팬텀 개발 및 유용성에 관한 연구)

  • Kim, Yung-Kyoon;Han, Dong-Kyoon
    • Journal of the Korean Society of Radiology
    • /
    • v.16 no.1
    • /
    • pp.67-75
    • /
    • 2022
  • The aim of this study is to establish a new QC method that can simultaneously evaluate the resolution of the x/y plane and the z-axis by producing a phantom that can reflect exposure and reconstruction parameter of MDCT system. It was used with Aquilion ONE(Cannon Medical System, Otawara, Japan), and the examination was scanned using of 120 kV, 260 mA, and the D-FOV of 300 mm2. It produced new SSP phantom modules in which two aluminum plates inclined at 45° to a vertical axis and a transverse axis to evaluate high contrast resolution of x/y plane and z axis. And it changed factors such as the algorithm, distance from gantry iso-center. All images were reconstructed in five steps from 0.6 mm to 10.0 mm slice thickness to measure resolution of x/y plane and z-axis. The image data measured FWHM and FWTM using Profile tool of Aquarius iNtusion Edition ver. 4.4.13 P6 software(Terarecon, California, USA), and analysed SPQI and signal intensity by ImageJ program(v1.53n, National Institutes of Health, USA). It decreased by 4.09~11.99%, 4.12~35.52%, and 4.70~37.64% in slice thickness of 2.5 mm, 5.0 mm, and 10.0 mm for evaluating the high contrast resolution of x/y plane according to distance from gantry iso-center. Therefore, the high contrast resolution of the x/y plane decreased when the distance from the iso-center increased or the slice thickness increased. Additionally, the slice thicknesses of 2.5 mm, 5.0 mm, and 10.0 mm with a high algorithm increased 74.83, 15.18 and 81.25%. The FWHM was almost constant on the measured SSP graph for evaluating the accuracy of slice thickness which represents the resolution of x/y plane and z-axis, but it was measured to be higher than the nominal slice thickness set by user. The FWHM and FWTM of z-axis with axial scan mode tended to increase significantly as the distance increased from gantry iso-center than the helical mode. Particularly, the thinner slice thickness that increased error range compare with the nominal slice thickness. The SPQI increased with thick slice thickness, and that was closer to 90% in the helical scan than the axial scan. In conclusion, by producing a phantom suitable for MDCT detectors and capable of quantitative resolution evaluation, it can be used as a specific method in the management of research quality and management of outdated equipment. Thus, it is expected to contribute greatly to the discrimination of lesions in the field of CT imaging.

A Study on Public Interest-based Technology Valuation Models in Water Resources Field (수자원 분야 공익형 기술가치평가 시스템에 대한 연구)

  • Ryu, Seung-Mi;Sung, Tae-Eung
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.177-198
    • /
    • 2018
  • Recently, as economic property it has become necessary to acquire and utilize the framework for water resource measurement and performance management as the property of water resources changes to hold "public property". To date, the evaluation of water technology has been carried out by feasibility study analysis or technology assessment based on net present value (NPV) or benefit-to-cost (B/C) effect, however it is not yet systemized in terms of valuation models to objectively assess an economic value of technology-based business to receive diffusion and feedback of research outcomes. Therefore, K-water (known as a government-supported public company in Korea) company feels the necessity to establish a technology valuation framework suitable for technical characteristics of water resources fields in charge and verify an exemplified case applied to the technology. The K-water evaluation technology applied to this study, as a public interest goods, can be used as a tool to measure the value and achievement contributed to society and to manage them. Therefore, by calculating the value in which the subject technology contributed to the entire society as a public resource, we make use of it as a basis information for the advertising medium of performance on the influence effect of the benefits or the necessity of cost input, and then secure the legitimacy for large-scale R&D cost input in terms of the characteristics of public technology. Hence, K-water company, one of the public corporation in Korea which deals with public goods of 'water resources', will be able to establish a commercialization strategy for business operation and prepare for a basis for the performance calculation of input R&D cost. In this study, K-water has developed a web-based technology valuation model for public interest type water resources based on the technology evaluation system that is suitable for the characteristics of a technology in water resources fields. In particular, by utilizing the evaluation methodology of the Institute of Advanced Industrial Science and Technology (AIST) in Japan to match the expense items to the expense accounts based on the related benefit items, we proposed the so-called 'K-water's proprietary model' which involves the 'cost-benefit' approach and the FCF (Free Cash Flow), and ultimately led to build a pipeline on the K-water research performance management system and then verify the practical case of a technology related to "desalination". We analyze the embedded design logic and evaluation process of web-based valuation system that reflects characteristics of water resources technology, reference information and database(D/B)-associated logic for each model to calculate public interest-based and profit-based technology values in technology integrated management system. We review the hybrid evaluation module that reflects the quantitative index of the qualitative evaluation indices reflecting the unique characteristics of water resources and the visualized user-interface (UI) of the actual web-based evaluation, which both are appended for calculating the business value based on financial data to the existing web-based technology valuation systems in other fields. K-water's technology valuation model is evaluated by distinguishing between public-interest type and profitable-type water technology. First, evaluation modules in profit-type technology valuation model are designed based on 'profitability of technology'. For example, the technology inventory K-water holds has a number of profit-oriented technologies such as water treatment membranes. On the other hand, the public interest-type technology valuation is designed to evaluate the public-interest oriented technology such as the dam, which reflects the characteristics of public benefits and costs. In order to examine the appropriateness of the cost-benefit based public utility valuation model (i.e. K-water specific technology valuation model) presented in this study, we applied to practical cases from calculation of benefit-to-cost analysis on water resource technology with 20 years of lifetime. In future we will additionally conduct verifying the K-water public utility-based valuation model by each business model which reflects various business environmental characteristics.

A Study for Strategy of On-line Shopping Mall: Based on Customer Purchasing and Re-purchasing Pattern (시스템 다이내믹스 기법을 활용한 온라인 쇼핑몰의 전략에 관한 연구 : 소비자의 구매 및 재구매 행동을 중심으로)

  • Lee, Sang-Gun;Min, Suk-Ki;Kang, Min-Cheol
    • Asia pacific journal of information systems
    • /
    • v.18 no.3
    • /
    • pp.91-121
    • /
    • 2008
  • Electronic commerce, commonly known as e-commerce or eCommerce, has become a major business trend in these days. The amount of trade conducted electronically has grown extraordinarily by developing the Internet technology. Most electronic commerce has being conducted between businesses to customers; therefore, the researches with respect to e-commerce are to find customer's needs, behaviors through statistical methods. However, the statistical researches, mostly based on a questionnaire, are the static researches, They can tell us the dynamic relationships between initial purchasing and repurchasing. Therefore, this study proposes dynamic research model for analyzing the cause of initial purchasing and repurchasing. This paper is based on the System-Dynamic theory, using the powerful simulation model with some restriction, The restrictions are based on the theory TAM(Technology Acceptance Model), PAM, and TPB(Theory of Planned Behavior). This article investigates not only the customer's purchasing and repurchasing behavior by passing of time but also the interactive effects to one another. This research model has six scenarios and three steps for analyzing customer behaviors. The first step is the research of purchasing situations. The second step is the research of repurchasing situations. Finally, the third step is to study the relationship between initial purchasing and repurchasing. The purpose of six scenarios is to find the customer's purchasing patterns according to the environmental changes. We set six variables in these scenarios by (1) changing the number of products; (2) changing the number of contents in on-line shopping malls; (3) having multimedia files or not in the shopping mall web sites; (4) grading on-line communities; (5) changing the qualities of products; (6) changing the customer's degree of confidence on products. First three variables are applied to study customer's purchasing behavior, and the other variables are applied to repurchasing behavior study. Through the simulation study, this paper presents some inter-relational result about customer purchasing behaviors, For example, Active community actions are not the increasing factor of purchasing but the increasing factor of word of mouth effect, Additionally. The higher products' quality, the more word of mouth effects increase. The number of products and contents on the web sites have same influence on people's buying behaviors. All simulation methods in this paper is not only display the result of each scenario but also find how to affect each other. Hence, electronic commerce firm can make more realistic marketing strategy about consumer behavior through this dynamic simulation research. Moreover, dynamic analysis method can predict the results which help the decision of marketing strategy by using the time-line graph. Consequently, this dynamic simulation analysis could be a useful research model to make firm's competitive advantage. However, this simulation model needs more further study. With respect to reality, this simulation model has some limitations. There are some missing factors which affect customer's buying behaviors in this model. The first missing factor is the customer's degree of recognition of brands. The second factor is the degree of customer satisfaction. The third factor is the power of word of mouth in the specific region. Generally, word of mouth affects significantly on a region's culture, even people's buying behaviors. The last missing factor is the user interface environment in the internet or other on-line shopping tools. In order to get more realistic result, these factors might be essential matters to make better research in the future studies.

A Checklist to Improve the Fairness in AI Financial Service: Focused on the AI-based Credit Scoring Service (인공지능 기반 금융서비스의 공정성 확보를 위한 체크리스트 제안: 인공지능 기반 개인신용평가를 중심으로)

  • Kim, HaYeong;Heo, JeongYun;Kwon, Hochang
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.3
    • /
    • pp.259-278
    • /
    • 2022
  • With the spread of Artificial Intelligence (AI), various AI-based services are expanding in the financial sector such as service recommendation, automated customer response, fraud detection system(FDS), credit scoring services, etc. At the same time, problems related to reliability and unexpected social controversy are also occurring due to the nature of data-based machine learning. The need Based on this background, this study aimed to contribute to improving trust in AI-based financial services by proposing a checklist to secure fairness in AI-based credit scoring services which directly affects consumers' financial life. Among the key elements of trustworthy AI like transparency, safety, accountability, and fairness, fairness was selected as the subject of the study so that everyone could enjoy the benefits of automated algorithms from the perspective of inclusive finance without social discrimination. We divided the entire fairness related operation process into three areas like data, algorithms, and user areas through literature research. For each area, we constructed four detailed considerations for evaluation resulting in 12 checklists. The relative importance and priority of the categories were evaluated through the analytic hierarchy process (AHP). We use three different groups: financial field workers, artificial intelligence field workers, and general users which represent entire financial stakeholders. According to the importance of each stakeholder, three groups were classified and analyzed, and from a practical perspective, specific checks such as feasibility verification for using learning data and non-financial information and monitoring new inflow data were identified. Moreover, financial consumers in general were found to be highly considerate of the accuracy of result analysis and bias checks. We expect this result could contribute to the design and operation of fair AI-based financial services.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.

A Study on Strategy for developing LBS Entertainment content based on local tourist information (지역 관광 정보를 활용한 LBS 엔터테인먼트 컨텐츠 개발 방안에 관한 연구)

  • Kim, Hyun-Jeong
    • Archives of design research
    • /
    • v.20 no.3 s.71
    • /
    • pp.151-162
    • /
    • 2007
  • How can new media devices and networks provide an effective response to the world's growing sector of cultural and historically-minded travelers? This study emerged from the question of how mobile handsets can change the nature of cultural and historical tourism in ubiquitous city environments. As wireless network and mobile IT have rapidly developed, it becomes possible to deliver cultural and historical information on the site through mobile handset as a tour guidance system. The paper describes the development of a new type of mobile tourism platform for site-specific cultural and historical information. The central objective of the project was to organize this cultural and historical walking tour around the mobile handset and its unique advantages (i.e. portability, multi-media capacity, access to wireless internet, and location-awareness potential) and then integrate the tour with a historical story and role-playing game that would deepen the mobile user's interest in the sites being visited, and enhance his or her overall experience of the area. The project was based on twelve locations that were culturally and historically significant to Korean War era in Busan. After the mobile tour game prototype was developed for this route, it was evaluated at the 10th PIFF (Pusan International Film Festival). After use test, some new strategies for developing mobile "edutainment content" to deliver cultural historical contents of the location were discussed. Combining 'edutainment' with a cultural and historical mobile walking tour brings a new dimension to existing approaches of the tourism and mobile content industry.

  • PDF

Anomaly Detection for User Action with Generative Adversarial Networks (적대적 생성 모델을 활용한 사용자 행위 이상 탐지 방법)

  • Choi, Nam woong;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.43-62
    • /
    • 2019
  • At one time, the anomaly detection sector dominated the method of determining whether there was an abnormality based on the statistics derived from specific data. This methodology was possible because the dimension of the data was simple in the past, so the classical statistical method could work effectively. However, as the characteristics of data have changed complexly in the era of big data, it has become more difficult to accurately analyze and predict the data that occurs throughout the industry in the conventional way. Therefore, SVM and Decision Tree based supervised learning algorithms were used. However, there is peculiarity that supervised learning based model can only accurately predict the test data, when the number of classes is equal to the number of normal classes and most of the data generated in the industry has unbalanced data class. Therefore, the predicted results are not always valid when supervised learning model is applied. In order to overcome these drawbacks, many studies now use the unsupervised learning-based model that is not influenced by class distribution, such as autoencoder or generative adversarial networks. In this paper, we propose a method to detect anomalies using generative adversarial networks. AnoGAN, introduced in the study of Thomas et al (2017), is a classification model that performs abnormal detection of medical images. It was composed of a Convolution Neural Net and was used in the field of detection. On the other hand, sequencing data abnormality detection using generative adversarial network is a lack of research papers compared to image data. Of course, in Li et al (2018), a study by Li et al (LSTM), a type of recurrent neural network, has proposed a model to classify the abnormities of numerical sequence data, but it has not been used for categorical sequence data, as well as feature matching method applied by salans et al.(2016). So it suggests that there are a number of studies to be tried on in the ideal classification of sequence data through a generative adversarial Network. In order to learn the sequence data, the structure of the generative adversarial networks is composed of LSTM, and the 2 stacked-LSTM of the generator is composed of 32-dim hidden unit layers and 64-dim hidden unit layers. The LSTM of the discriminator consists of 64-dim hidden unit layer were used. In the process of deriving abnormal scores from existing paper of Anomaly Detection for Sequence data, entropy values of probability of actual data are used in the process of deriving abnormal scores. but in this paper, as mentioned earlier, abnormal scores have been derived by using feature matching techniques. In addition, the process of optimizing latent variables was designed with LSTM to improve model performance. The modified form of generative adversarial model was more accurate in all experiments than the autoencoder in terms of precision and was approximately 7% higher in accuracy. In terms of Robustness, Generative adversarial networks also performed better than autoencoder. Because generative adversarial networks can learn data distribution from real categorical sequence data, Unaffected by a single normal data. But autoencoder is not. Result of Robustness test showed that he accuracy of the autocoder was 92%, the accuracy of the hostile neural network was 96%, and in terms of sensitivity, the autocoder was 40% and the hostile neural network was 51%. In this paper, experiments have also been conducted to show how much performance changes due to differences in the optimization structure of potential variables. As a result, the level of 1% was improved in terms of sensitivity. These results suggest that it presented a new perspective on optimizing latent variable that were relatively insignificant.

A Ranking Algorithm for Semantic Web Resources: A Class-oriented Approach (시맨틱 웹 자원의 랭킹을 위한 알고리즘: 클래스중심 접근방법)

  • Rho, Sang-Kyu;Park, Hyun-Jung;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.17 no.4
    • /
    • pp.31-59
    • /
    • 2007
  • We frequently use search engines to find relevant information in the Web but still end up with too much information. In order to solve this problem of information overload, ranking algorithms have been applied to various domains. As more information will be available in the future, effectively and efficiently ranking search results will become more critical. In this paper, we propose a ranking algorithm for the Semantic Web resources, specifically RDF resources. Traditionally, the importance of a particular Web page is estimated based on the number of key words found in the page, which is subject to manipulation. In contrast, link analysis methods such as Google's PageRank capitalize on the information which is inherent in the link structure of the Web graph. PageRank considers a certain page highly important if it is referred to by many other pages. The degree of the importance also increases if the importance of the referring pages is high. Kleinberg's algorithm is another link-structure based ranking algorithm for Web pages. Unlike PageRank, Kleinberg's algorithm utilizes two kinds of scores: the authority score and the hub score. If a page has a high authority score, it is an authority on a given topic and many pages refer to it. A page with a high hub score links to many authoritative pages. As mentioned above, the link-structure based ranking method has been playing an essential role in World Wide Web(WWW), and nowadays, many people recognize the effectiveness and efficiency of it. On the other hand, as Resource Description Framework(RDF) data model forms the foundation of the Semantic Web, any information in the Semantic Web can be expressed with RDF graph, making the ranking algorithm for RDF knowledge bases greatly important. The RDF graph consists of nodes and directional links similar to the Web graph. As a result, the link-structure based ranking method seems to be highly applicable to ranking the Semantic Web resources. However, the information space of the Semantic Web is more complex than that of WWW. For instance, WWW can be considered as one huge class, i.e., a collection of Web pages, which has only a recursive property, i.e., a 'refers to' property corresponding to the hyperlinks. However, the Semantic Web encompasses various kinds of classes and properties, and consequently, ranking methods used in WWW should be modified to reflect the complexity of the information space in the Semantic Web. Previous research addressed the ranking problem of query results retrieved from RDF knowledge bases. Mukherjea and Bamba modified Kleinberg's algorithm in order to apply their algorithm to rank the Semantic Web resources. They defined the objectivity score and the subjectivity score of a resource, which correspond to the authority score and the hub score of Kleinberg's, respectively. They concentrated on the diversity of properties and introduced property weights to control the influence of a resource on another resource depending on the characteristic of the property linking the two resources. A node with a high objectivity score becomes the object of many RDF triples, and a node with a high subjectivity score becomes the subject of many RDF triples. They developed several kinds of Semantic Web systems in order to validate their technique and showed some experimental results verifying the applicability of their method to the Semantic Web. Despite their efforts, however, there remained some limitations which they reported in their paper. First, their algorithm is useful only when a Semantic Web system represents most of the knowledge pertaining to a certain domain. In other words, the ratio of links to nodes should be high, or overall resources should be described in detail, to a certain degree for their algorithm to properly work. Second, a Tightly-Knit Community(TKC) effect, the phenomenon that pages which are less important but yet densely connected have higher scores than the ones that are more important but sparsely connected, remains as problematic. Third, a resource may have a high score, not because it is actually important, but simply because it is very common and as a consequence it has many links pointing to it. In this paper, we examine such ranking problems from a novel perspective and propose a new algorithm which can solve the problems under the previous studies. Our proposed method is based on a class-oriented approach. In contrast to the predicate-oriented approach entertained by the previous research, a user, under our approach, determines the weights of a property by comparing its relative significance to the other properties when evaluating the importance of resources in a specific class. This approach stems from the idea that most queries are supposed to find resources belonging to the same class in the Semantic Web, which consists of many heterogeneous classes in RDF Schema. This approach closely reflects the way that people, in the real world, evaluate something, and will turn out to be superior to the predicate-oriented approach for the Semantic Web. Our proposed algorithm can resolve the TKC(Tightly Knit Community) effect, and further can shed lights on other limitations posed by the previous research. In addition, we propose two ways to incorporate data-type properties which have not been employed even in the case when they have some significance on the resource importance. We designed an experiment to show the effectiveness of our proposed algorithm and the validity of ranking results, which was not tried ever in previous research. We also conducted a comprehensive mathematical analysis, which was overlooked in previous research. The mathematical analysis enabled us to simplify the calculation procedure. Finally, we summarize our experimental results and discuss further research issues.