• Title/Summary/Keyword: 사용자 상황인지 모델

Search Result 566, Processing Time 0.025 seconds

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

Adaptive Lock Escalation in Database Management Systems (데이타베이스 관리 시스템에서의 적응형 로크 상승)

  • Chang, Ji-Woong;Lee, Young-Koo;Whang, Kyu-Young;Yang, Jae-Heon
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.742-757
    • /
    • 2001
  • Since database management systems(DBMSS) have limited lock resources, transactions requesting locks beyond the limit mutt be aborted. In the worst carte, if such transactions are aborted repeatedly, the DBMS can become paralyzed, i.e., transaction execute but cannot commit. Lock escalation is considered a solution to this problem. However, existing lock escalation methods do not provide a complete solution. In this paper, we prognose a new lock escalation method, adaptive lock escalation, that selves most of the problems. First, we propose a general model for lock escalation and present the concept of the unescalatable look, which is the major cause making the transactions to abort. Second, we propose the notions of semi lock escalation, lock blocking, and selective relief as the mechanisms to control the number of unescalatable locks. We then propose the adaptive lock escalation method using these notions. Adaptive lock escalation reduces needless aborts and guarantees that the DBMS is not paralyzed under excessive lock requests. It also allows graceful degradation of performance under those circumstances. Third, through extensive simulation, we show that adaptive lock escalation outperforms existing lock escalation methods. The results show that, compared to the existing methods, adaptive lock escalation reduces the number of aborts and the average response time, and increases the throughput to a great extent. Especially, it is shown that the number of concurrent transactions can be increased more than 16 ~256 fold. The contribution of this paper is significant in that it has formally analysed the role of lock escalation in lock resource management and identified the detailed underlying mechanisms. Existing lock escalation methods rely on users or system administrator to handle the problems of excessive lock requests. In contrast, adaptive lock escalation releases the users of this responsibility by providing graceful degradation and preventing system paralysis through automatic control of unescalatable locks Thus adaptive lock escalation can contribute to developing self-tuning: DBMSS that draw a lot of attention these days.

  • PDF

A study on Indoor Insulating Tent Design for Senior Citizens who Live Alone (독거노인을 위한 융복합 실내 보온용 텐트디자인에 관한 연구)

  • Lee, Dae Hyun;Kim, Sang Sik
    • Korea Science and Art Forum
    • /
    • v.37 no.2
    • /
    • pp.219-230
    • /
    • 2019
  • According to the National Statistical Office, the number of senior citizens aged 65 or over has exceeded 7 million as of 2017, and the number of senior citizens who live alone among them exceeded 1.3 million. Most of the senior citizens who live alone suffer from absolute poverty (68.5% of them has monthly income less than 500,000 won) and they have difficulties for basic living in the blind spot of our society. In particular, the heating is quite a serious issue for the senior citizens who live alone and belong to vulnerable social group in the winter, and to make it worse, they are exposed to frequent fire accidents due to the negligence in handling electric appliances such as electric heaters and electric pads. The main reason the indoor tent products are being used by senior citizens who live alone is that it saves energy, ensures warm sleep, and improves fire safety. Following the expansion of the indoor tent market, this study focused on the idea that there is a need for an in door tent suitable for senior citizens who live alone and belong to low income bracket and intended to improve its efficiency in relation to use, resting, and storage. For this, a basic survey was conducted on the products of existing brands to analyze advantages and disadvantages and it was possible to understand the demands that consumers have for existing products. Accordingly, a survey on consumer preference was conducted using a designed model and the Zabara typp, which demonstrated the best efficiency in terms of installation convenience, space usability, and appearance design was selected. Based on the results of selection, the product design and final prototype were completed. The results and details of the study are as follows; First, factors that were not recognized in product development phase could be identified through usability survey and interview with actual users. Second, for the effective aspect of the prototype, senior citizens could install and fold the tend more easily and quickly than expected. Based on these results of this study, it is expected that not only senior citizens who live alone but also various other users can use the tent to create another comfortable private space indoors.

An Analysis of Soil Pressure Gauge Result from KHC Test Road (시험도로 토압계 계측결과 분석)

  • In Byeong-Eock;Kim Ji-Won;Kim Kyong-Ha;Lee Kwang-Ho
    • International Journal of Highway Engineering
    • /
    • v.8 no.3 s.29
    • /
    • pp.129-141
    • /
    • 2006
  • The vertical soil pressure developed in the granular layer of asphalt pavement system is influenced by various factors, including the wheel load magnitude, the loading speed, and asphalt pavement temperature. This research observed the distribution of vertical soil pressure in pavement supporting layer by investigating measured data from soil pressure gage in the KHC Test Road. The existing specification of subbase and subgrade compaction was also evaluated with measured vertical pressure. The finite element analysis was conducted to verify the accuracy of results with measured data because it can maximize research capacity without significant field test. The test data was collected from A5, A7, A14, and A15 test sections at August, September, and November 2004 and August 2005. Those test sections and test data were selected because they had best quality. The size of influence area was evaluated and the vertical pressure variation was investigated with respect to load level, load speed, and pavement temperature. The lower speed, higher load level, and higher pavement temperature increased the vertical pressure and reduced the area of influence. The finite element result showed the similar trend of vertical pressure variation in comparison with measured data. The specification of compaction quality for subbase and subgrade is higher than the level of vertical pressure measured with truck load so that it should be lurker investigated.

  • PDF

Research for Application of Interactive Data Broadcasting Service in DMB (DMB에서의 양방향 데어터방송 서비스도입에 관한 연구)

  • Kim, Jong-Geun;Choe, Seong-Jin;Lee, Seon-Hui
    • Broadcasting and Media Magazine
    • /
    • v.11 no.4
    • /
    • pp.104-117
    • /
    • 2006
  • In this Paper, we analyze the application of Interactive Data Broadcasting in DMB(Digital Multimedia Broadcasting) in the accordance with convergence of service and technology. With the acceleration of digital convergence in the Ubiquitous period substantial development of digital media technology and convergence of broadcasting and telecommunication industry are being witnessed. Consequently these results gave rise to newly combined-products such as DMB(Digital Multimedia Broadcasting), WCDMA(Wide-band code division multiple access), Wibro(Wireless Broadband Internet), IP-TV (Internet protocol TV) and HSDPA(High speed downlink packet access). The preparatory stage for the implementation of Interactive Data Broadcasting Service will be reached by the end of December, 2006. DMB is the first result of a successful convergence service between Broadcasting and Telecommunication in new media era. Multimedia technology and services are the core elements of DMB. The Data Broadcasting will not only offer various services of interactive information such News, Weather, Broadcasting Program etc, but also be linked with characteristic function of mobile phone such as calling and SMS(Short Message Service) via Return Channel.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.