• Title/Summary/Keyword: Event Management

Search Result 1,529, Processing Time 0.029 seconds

An Analysis of Big Video Data with Cloud Computing in Ubiquitous City (클라우드 컴퓨팅을 이용한 유시티 비디오 빅데이터 분석)

  • Lee, Hak Geon;Yun, Chang Ho;Park, Jong Won;Lee, Yong Woo
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.45-52
    • /
    • 2014
  • The Ubiquitous-City (U-City) is a smart or intelligent city to satisfy human beings' desire to enjoy IT services with any device, anytime, anywhere. It is a future city model based on Internet of everything or things (IoE or IoT). It includes a lot of video cameras which are networked together. The networked video cameras support a lot of U-City services as one of the main input data together with sensors. They generate huge amount of video information, real big data for the U-City all the time. It is usually required that the U-City manipulates the big data in real-time. And it is not easy at all. Also, many times, it is required that the accumulated video data are analyzed to detect an event or find a figure among them. It requires a lot of computational power and usually takes a lot of time. Currently we can find researches which try to reduce the processing time of the big video data. Cloud computing can be a good solution to address this matter. There are many cloud computing methodologies which can be used to address the matter. MapReduce is an interesting and attractive methodology for it. It has many advantages and is getting popularity in many areas. Video cameras evolve day by day so that the resolution improves sharply. It leads to the exponential growth of the produced data by the networked video cameras. We are coping with real big data when we have to deal with video image data which are produced by the good quality video cameras. A video surveillance system was not useful until we find the cloud computing. But it is now being widely spread in U-Cities since we find some useful methodologies. Video data are unstructured data thus it is not easy to find a good research result of analyzing the data with MapReduce. This paper presents an analyzing system for the video surveillance system, which is a cloud-computing based video data management system. It is easy to deploy, flexible and reliable. It consists of the video manager, the video monitors, the storage for the video images, the storage client and streaming IN component. The "video monitor" for the video images consists of "video translater" and "protocol manager". The "storage" contains MapReduce analyzer. All components were designed according to the functional requirement of video surveillance system. The "streaming IN" component receives the video data from the networked video cameras and delivers them to the "storage client". It also manages the bottleneck of the network to smooth the data stream. The "storage client" receives the video data from the "streaming IN" component and stores them to the storage. It also helps other components to access the storage. The "video monitor" component transfers the video data by smoothly streaming and manages the protocol. The "video translator" sub-component enables users to manage the resolution, the codec and the frame rate of the video image. The "protocol" sub-component manages the Real Time Streaming Protocol (RTSP) and Real Time Messaging Protocol (RTMP). We use Hadoop Distributed File System(HDFS) for the storage of cloud computing. Hadoop stores the data in HDFS and provides the platform that can process data with simple MapReduce programming model. We suggest our own methodology to analyze the video images using MapReduce in this paper. That is, the workflow of video analysis is presented and detailed explanation is given in this paper. The performance evaluation was experiment and we found that our proposed system worked well. The performance evaluation results are presented in this paper with analysis. With our cluster system, we used compressed $1920{\times}1080(FHD)$ resolution video data, H.264 codec and HDFS as video storage. We measured the processing time according to the number of frame per mapper. Tracing the optimal splitting size of input data and the processing time according to the number of node, we found the linearity of the system performance.

Aspect of the chief of state guard EMP (Electro Magnetic Pulse) protection system for the consideration (국가원수 경호적 측면에서의 EMP(Electro Magnetic Pulse) 방호 시스템에 대한 고찰)

  • Jung, Joo-Sub
    • Korean Security Journal
    • /
    • no.41
    • /
    • pp.37-66
    • /
    • 2014
  • In recent years, with the development of computers and electronics, electronics and communication technology in a growing and each part is dependent on the cross-referencing makes all electronic equipment is obsolete due to direct or indirect damage EMP. Korea and the impending standoff North Korea has a considerable level of technologies related to the EMP, EMP weapons you already have or in a few years, the development of EMP weapons will complete. North Korea launched a long-range missile and conducted a nuclear test on several occasions immediately after, when I saw the high-altitude nuclear blackmail has been strengthening the outright offensive nuclear EMP attacks at any time and practical significance for the EMP will need offensive skills would improve. At this point you can predict the damage situation of Korea's security reality that satisfy the need, more than anything else to build a protective system of the EMP. The scale of the damage that unforeseen but significant military damage and socio-economic damage and fatalities when I looked into the situation which started out as a satellite communications systems and equipment to attack military and security systems and transportation, finance, national emergency system, such as the damage elsewhere. In General, there is no direct casualties reported, but EMP medical devices that rely on lethal damage to people who can show up. In addition, the State power system failure due to a power supply interruption would not have thought the damage would bring State highly dependent on domestic power generation of nuclear plants is a serious nuclear power plant accident in the event of a blackout phenomenon can lead to the plant's internal problems should see a forecast. First of all, a special expert Committee of the EMP, the demand for protective facilities and equipment and conduct an investigation, he takes fits into your budget is under strict criteria by configuring the contractors should be sifting through. He then created the Agency for verification of performance EMP protection after you have verified the performance of maintenance, maintenance, safety and security management, design and construction company organized and systematic process Guard facilities or secret communications equipment and perfect for the EMP, such as protective equipment maneuver system should take.

  • PDF

Development of a Remote Multi-Task Debugger for Qplus-T RTOS (Qplus-T RTOS를 위한 원격 멀티 태스크 디버거의 개발)

  • 이광용;김흥남
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.9 no.4
    • /
    • pp.393-409
    • /
    • 2003
  • In this paper, we present a multi-task debugging environment for Qplus-T embedded-system such as internet information appliances. We will propose the structure and functions of a remote multi-task debugging environment supporting environment effective ross-development. And, we are going enhance the communication architecture between the host and target system to provide more efficient cross-development environment. The remote development toolset called Q+Esto consists to several independent support tools: an interactive shell, a remote debugger, a resource monitor, a target manager and a debug agent. Excepting a debug agent, all these support tools reside on the host systems. Using the remote multi-task debugger on the host, the developer can spawn and debug tasks on the target run-time system. It can also be attached to already-running tasks spawned from the application or from interactive shell. Application code can be viewed as C/C++ source, or as assembly-level code. It incorporates a variety of display windows for source, registers, local/global variables, stack frame, memory, event traces and so on. The target manager implements common functions that are shared by Q+Esto tools, e.g., the host-target communication, object file loading, and management of target-resident host tool´s memory pool and target system´s symbol-table, and so on. These functions are called OPEn C APIs and they greatly improve the extensibility of the Q+Esto Toolset. The Q+Esto target manager is responsible for communicating between host and target system. Also, there exist a counterpart on the target system communicating with the host target manager, which is called debug agent. Debug agent is a daemon task on real-time operating systems in the target system. It gets debugging requests from the host tools including debugger via target manager, interprets the requests, executes them and sends the results to the host.

A Real-Time Stock Market Prediction Using Knowledge Accumulation (지식 누적을 이용한 실시간 주식시장 예측)

  • Kim, Jin-Hwa;Hong, Kwang-Hun;Min, Jin-Young
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.109-130
    • /
    • 2011
  • One of the major problems in the area of data mining is the size of the data, as most data set has huge volume these days. Streams of data are normally accumulated into data storages or databases. Transactions in internet, mobile devices and ubiquitous environment produce streams of data continuously. Some data set are just buried un-used inside huge data storage due to its huge size. Some data set is quickly lost as soon as it is created as it is not saved due to many reasons. How to use this large size data and to use data on stream efficiently are challenging questions in the study of data mining. Stream data is a data set that is accumulated to the data storage from a data source continuously. The size of this data set, in many cases, becomes increasingly large over time. To mine information from this massive data, it takes too many resources such as storage, money and time. These unique characteristics of the stream data make it difficult and expensive to store all the stream data sets accumulated over time. Otherwise, if one uses only recent or partial of data to mine information or pattern, there can be losses of valuable information, which can be useful. To avoid these problems, this study suggests a method efficiently accumulates information or patterns in the form of rule set over time. A rule set is mined from a data set in stream and this rule set is accumulated into a master rule set storage, which is also a model for real-time decision making. One of the main advantages of this method is that it takes much smaller storage space compared to the traditional method, which saves the whole data set. Another advantage of using this method is that the accumulated rule set is used as a prediction model. Prompt response to the request from users is possible anytime as the rule set is ready anytime to be used to make decisions. This makes real-time decision making possible, which is the greatest advantage of this method. Based on theories of ensemble approaches, combination of many different models can produce better prediction model in performance. The consolidated rule set actually covers all the data set while the traditional sampling approach only covers part of the whole data set. This study uses a stock market data that has a heterogeneous data set as the characteristic of data varies over time. The indexes in stock market data can fluctuate in different situations whenever there is an event influencing the stock market index. Therefore the variance of the values in each variable is large compared to that of the homogeneous data set. Prediction with heterogeneous data set is naturally much more difficult, compared to that of homogeneous data set as it is more difficult to predict in unpredictable situation. This study tests two general mining approaches and compare prediction performances of these two suggested methods with the method we suggest in this study. The first approach is inducing a rule set from the recent data set to predict new data set. The seocnd one is inducing a rule set from all the data which have been accumulated from the beginning every time one has to predict new data set. We found neither of these two is as good as the method of accumulated rule set in its performance. Furthermore, the study shows experiments with different prediction models. The first approach is building a prediction model only with more important rule sets and the second approach is the method using all the rule sets by assigning weights on the rules based on their performance. The second approach shows better performance compared to the first one. The experiments also show that the suggested method in this study can be an efficient approach for mining information and pattern with stream data. This method has a limitation of bounding its application to stock market data. More dynamic real-time steam data set is desirable for the application of this method. There is also another problem in this study. When the number of rules is increasing over time, it has to manage special rules such as redundant rules or conflicting rules efficiently.

The possibility of South Korea to become a member state of APSCO: an analysis from Legal and political perspectives (韓國加入亞太空間合作組織的可能性 : 基于法律与政策的分析)

  • Nie, Mingyan
    • The Korean Journal of Air & Space Law and Policy
    • /
    • v.31 no.2
    • /
    • pp.237-269
    • /
    • 2016
  • Asia-Pacific Space Cooperation Organization (APSCO) is the only intergovernmental space cooperation organization in Asia. Since its establishment to date, eight countries have signed the convention and become member states. South Korea participated actively in the preparatory phase of creating the organization, and one conference organized by AP-MCSTA which is the predecessor of APSCO was held in South Korea. However, after the APSCO Convention was opened for signature in 2005 to date, South Korea does not ratify the Convention and become a member. The rapid development of space commercialization and privatization, as well as the fastest growing commercial space market in Asia, provides opportunities for Asian countries to cooperate with each other in relevant space fields. And to participate in the existing cooperation framework (e.g., the APSCO) by the Asian space countries (e.g., South Korea) could be a proper choice. Even if the essential cooperation in particular space fields is challenging, joint space programs among different Asian countries for dealing with the common events can be initiated at the first steps. Since APSCO has learned the successful legal arrangements from ESA, the legal measures established by its Convention are believed to be qualified to ensure the achievement of benefits of different member states. For example, the regulation of the "fair return" principle confirms that the return of interests from the relevant programs is in proportion to the member's investment in the programs. Moreover, the distinguish of basic and optional activities intends to authorize the freedom of the members to choose programs to participate. And for the voting procedure, the acceptance of the "consensus" by the Council is in favor of protecting the member's interest when making decisions. However, political factors that are potential to block the participation of South Korea in APSCO are difficult to be ignored. A recent event is an announcement of deploying THAAD by South Korea, which causes tension between South Korea and China. The cooperation between these two states in space activities will be influenced. A long-standing barrier is that China acts as a non-member of the main international export control mechanism, i.e., the MTCR. The U.S takes this fact as the main reason to prevent South Korea to cooperate with China in developing space programs. Although the political factors that will block the participation of South Korea in APSCO are not easy to removed shortly, legal measures can be taken to reduce the political influence. More specifically, APSCO is recommended to ensure the achievement of commercial interests of different cooperation programs by regulating precisely the implementation of the "fair return" principle. Furthermore, APSCO is also suggested to contribute to managing the common regional events by sharing satellite data. And it is anticipated that these measures can effectively response the requirements of the rapid development of space commercialization and the increasing common needs of Asia, thereby to provide a platform for the further cooperation. In addition, in order to directly reduce the political influence, two legal measures are necessary to be taken: Firstly, to clarify the rights and responsibilities of the host state (i.e., China) as providing assistance, coordination and services to the management of the Organization to release the worries of the other member states that the host state will control the Organization's activities. And secondly, to illustrate that the cooperation in APSCO is for the non-military purpose (a narrow sense of "peaceful purpose") to reduce the political concerns. Regional cooperation in Asia regarding space affairs is considered to be a general trend in the future, so if the participation of South Korea in APSCO can be finally proved to be feasible, there will be an opportunity to discuss the creation of a comprehensive institutionalized framework for space cooperation in Asia.

Mobile Contents Transformation System Research for Personalization Service (개인화 서비스를 위한 모바일 콘텐츠 변환 시스템 연구)

  • Bae, Jong-Hwan;Cho, Young-Hee;Lee, Jung-Jae;Kim, Nam-Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.2
    • /
    • pp.119-128
    • /
    • 2011
  • The Sensor technology and portable device capability able to collect recent user information and the information about the surrounding environment haven been highly developed. A user can be made use of various contents and the option is also extending with this technology development. In particular, the initial portable device had simply a call function, but now that has evolved into 'the 4th screen' which including movie, television, PC ability. also, in the past, a portable device to provided only the services of a SMS, in recent years, it provided to interactive video service, and it include technology which providing various contents. Also, it is rising as media which leading the consumption of contents, because it can be used anytime, anywhere. However, the contents available for the nature of user's handheld devices are limited. because it is very difficult for making the contents separately according to various device specification. To find a solution to this problem, the study on one contents from several device has been progressing. The contents conversion technology making use of the profile of device out of this study comes to the force and profile study has been progressing for this. Furthermore, Demand for a user is also increased and the study on the technology collecting, analyzing demands has been making active progress. And what is more, Grasping user's demands by making use of this technology and the study on the technology analyzing, providing contents has been making active progress as well. First of all, there is a method making good use of ZigBee, Bluetooth technology about the sensor for gathering user's information. ZigBee uses low-power digital radio for wireless headphone, wireless communication network, and being utilized for smart energy, automatic home system, wireless communication application and wireless sensor application. Bluetooth, as industry standards of PAN(Personal Area Networks), is being made of use of low power wireless device for the technology supporting data transmission such as drawing file, video file among Bluetooth device. With analyzing the collected information making use of this technology, it utilizes personalized service based on network knowledge developed by ETRI to service contents tailor-made for a user. Now that personalized service builds up network knowledge about user's various environments, the technology provides context friendly service constructed dynamically on the basis of this. The contents to service dynamically like this offer the contents that it converses with utilizing device profile to working well. Therefore, this paper suggests the system as follow. It collects the information, for example of user's sensitivity, context and location by using sensor technology, and generates the profile as a means of collected information as sensor. It collects the user's propensity to the information by user's input and event and generates profile in the same way besides the gathered information by sensor. Device transmits a generated profile and the profile about a device specification to proxy server. And proxy server transmits a profile to each profile management server. It analyzes profile in proxy server so that it selects the contents user demand and requests in contents server. Contents server receives a profile of user portable device from device profile server and converses the contents by using this. Original source code of contents convert into XML code using the device profile and XML code convert into source code available in user portable device. Thus, contents conversion process is terminated and user friendly system is completed as the user transmits optimal contents for user portable device.

Which types of the strategies diffused to the public through company's announcement do contribute to the long-term performance? (공시된 경영전략의 유형별 장기실적 기여도 분석)

  • Kang, Won
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.4 no.4
    • /
    • pp.45-70
    • /
    • 2009
  • This article investigates which types of the strategies announced by the listed firms contribute to enhancing the long-term performance of the companies. Since 2002, Korean Exchange adopted the "faire disclosure policy" which mandates that all publicly traded companies must disclose material information to all investors at the same time. Thanks to the policy, Korean investors can, now, easily access the board's decision on management strategies on the same day the decision is made. If the companies trustfully carry out their announced strategies, we can decide which types of strategies actually enhance or deteriorate the long-term performance, simply by comparing the announced strategies and the firm's performance. The sample companies are confined to 60 firms that became listed in the KOSDAQ market through back-door listing from 2003 to 2005. Using only the newly listed companies, we can avoid the interference on the long-term performance of the strategies pursued before the event date. This often holds true, for many companies radically modify their strategies after the listing. Furthermore, the back-door listing companies serve our purpose better than IPO companies do, because the former tend to have a variety of announcement within a given period of time beginning the listing date. Using these sample companies, this article analyzes the effect on one year buy-and-hold returns and abnormal buy-and-hold returns after the listing of the various types of strategies announced during the same period of time. The results show that those evidences of restructuring such as 'reduction of capital' and 'resignation of incumbent board members', actually contribute to the increase in adjusted long-term stock returns. Those strategies which can be view as evidence of new investment such as 'increase in tangible assets', 'acquisition of other companies', do also helps the stockholders better off. On the contrary, 'increase in bank loans', 'changes of CEO' and 'merger' deteriorate the equity value. The last findings let us to presume that the back-door listing companies appear to use the bank loans for value-reducing activities; the change in CEO is not a sign of restructuring, but rather a sign of failure of the restructuring; another merger carried out after back-door listing itself is also value-reducing activity. This article's findings on reduction of capital, merger and bank loans oppose the results of the former empirical studies which analyze only the short-term effect on stock price. Therefore, more long-term performance studies on public disclosures are in order.

  • PDF

Intelligent Optimal Route Planning Based on Context Awareness (상황인식 기반 지능형 최적 경로계획)

  • Lee, Hyun-Jung;Chang, Yong-Sik
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.117-137
    • /
    • 2009
  • Recently, intelligent traffic information systems have enabled people to forecast traffic conditions before hitting the road. These convenient systems operate on the basis of data reflecting current road and traffic conditions as well as distance-based data between locations. Thanks to the rapid development of ubiquitous computing, tremendous context data have become readily available making vehicle route planning easier than ever. Previous research in relation to optimization of vehicle route planning merely focused on finding the optimal distance between locations. Contexts reflecting the road and traffic conditions were then not seriously treated as a way to resolve the optimal routing problems based on distance-based route planning, because this kind of information does not have much significant impact on traffic routing until a a complex traffic situation arises. Further, it was also not easy to take into full account the traffic contexts for resolving optimal routing problems because predicting the dynamic traffic situations was regarded a daunting task. However, with rapid increase in traffic complexity the importance of developing contexts reflecting data related to moving costs has emerged. Hence, this research proposes a framework designed to resolve an optimal route planning problem by taking full account of additional moving cost such as road traffic cost and weather cost, among others. Recent technological development particularly in the ubiquitous computing environment has facilitated the collection of such data. This framework is based on the contexts of time, traffic, and environment, which addresses the following issues. First, we clarify and classify the diverse contexts that affect a vehicle's velocity and estimates the optimization of moving cost based on dynamic programming that accounts for the context cost according to the variance of contexts. Second, the velocity reduction rate is applied to find the optimal route (shortest path) using the context data on the current traffic condition. The velocity reduction rate infers to the degree of possible velocity including moving vehicles' considerable road and traffic contexts, indicating the statistical or experimental data. Knowledge generated in this papercan be referenced by several organizations which deal with road and traffic data. Third, in experimentation, we evaluate the effectiveness of the proposed context-based optimal route (shortest path) between locations by comparing it to the previously used distance-based shortest path. A vehicles' optimal route might change due to its diverse velocity caused by unexpected but potential dynamic situations depending on the road condition. This study includes such context variables as 'road congestion', 'work', 'accident', and 'weather' which can alter the traffic condition. The contexts can affect moving vehicle's velocity on the road. Since these context variables except for 'weather' are related to road conditions, relevant data were provided by the Korea Expressway Corporation. The 'weather'-related data were attained from the Korea Meteorological Administration. The aware contexts are classified contexts causing reduction of vehicles' velocity which determines the velocity reduction rate. To find the optimal route (shortest path), we introduced the velocity reduction rate in the context for calculating a vehicle's velocity reflecting composite contexts when one event synchronizes with another. We then proposed a context-based optimal route (shortest path) algorithm based on the dynamic programming. The algorithm is composed of three steps. In the first initialization step, departure and destination locations are given, and the path step is initialized as 0. In the second step, moving costs including composite contexts into account between locations on path are estimated using the velocity reduction rate by context as increasing path steps. In the third step, the optimal route (shortest path) is retrieved through back-tracking. In the provided research model, we designed a framework to account for context awareness, moving cost estimation (taking both composite and single contexts into account), and optimal route (shortest path) algorithm (based on dynamic programming). Through illustrative experimentation using the Wilcoxon signed rank test, we proved that context-based route planning is much more effective than distance-based route planning., In addition, we found that the optimal solution (shortest paths) through the distance-based route planning might not be optimized in real situation because road condition is very dynamic and unpredictable while affecting most vehicles' moving costs. For further study, while more information is needed for a more accurate estimation of moving vehicles' costs, this study still stands viable in the applications to reduce moving costs by effective route planning. For instance, it could be applied to deliverers' decision making to enhance their decision satisfaction when they meet unpredictable dynamic situations in moving vehicles on the road. Overall, we conclude that taking into account the contexts as a part of costs is a meaningful and sensible approach to in resolving the optimal route problem.

Study on the Fire Risk Prediction Assessment due to Deterioration contact of combustible cables in Underground Common Utility Tunnels (지하공동구내 가연성케이블의 열화접촉으로 인한 화재위험성 예측평가)

  • Ko, Jaesun
    • Journal of the Society of Disaster Information
    • /
    • v.11 no.1
    • /
    • pp.135-147
    • /
    • 2015
  • Recent underground common utility tunnels are underground facilities for jointly accommodating more than 2 kinds of air-conditioning and heating facilities, vacuum dust collector, information processing cables as well as electricity, telecommunications, waterworks, city gas, sewerage system required when citizens live their daily lives and facilities responsible for the central function of the country but it is difficult to cope with fire accidents quickly and hard to enter into common utility tunnels to extinguish a fire due to toxic gases and smoke generated when various cables are burnt. Thus, in the event of a fire, not only the nerve center of the country is paralyzed such as significant property damage and loss of communication etc. but citizen inconveniences are caused. Therefore, noticing that most fires break out by a short circuit due to electrical works and degradation contact due to combustible cables as the main causes of fires in domestic and foreign common utility tunnels fire cases that have occurred so far, the purpose of this paper is to scientifically analyze the behavior of a fire by producing the model of actual common utility tunnels and reproducing the fire. A fire experiment was conducted in a state that line type fixed temperature detector, fire door, connection deluge set and ventilation equipment are installed in underground common utility tunnels and transmission power distribution cables are coated with fire proof paints in a certain section and heating pipes are fire proof covered. As a result, in the case of Type II, the maximum temperature was measured as $932^{\circ}C$ and line type fixed temperature detector displayed the fire location exactly in the receiver at a constant temperature. And transmission power distribution cables painted with fire proof paints in a certain section, the case of Type III, were found not to be fire resistant and fire proof covered heating pipes to be fire resistant for about 30 minutes. Also, fire simulation was carried out by entering fire load during a real fire test and as a result, the maximum temperature is $943^{\circ}C$, almost identical with $932^{\circ}C$ during a real fire test. Therefore, it is considered that fire behaviour can be predicted by conducting fire simulation only with common utility tunnels fire load and result values of heat release rate, height of the smoke layer, concentration of O2, CO, CO2 etc. obtained by simulation are determined to be applied as the values during a real fire experiment. In the future, it is expected that more reliable information on domestic underground common utility tunnels fire accidents can be provided and it will contribute to construction and maintenance repair effectively and systematically by analyzing and accumulating experimental data on domestic underground common utility tunnels fire accidents built in this study and fire cases continuously every year and complementing laws and regulations and administration manuals etc.

Multi-task Learning Based Tropical Cyclone Intensity Monitoring and Forecasting through Fusion of Geostationary Satellite Data and Numerical Forecasting Model Output (정지궤도 기상위성 및 수치예보모델 융합을 통한 Multi-task Learning 기반 태풍 강도 실시간 추정 및 예측)

  • Lee, Juhyun;Yoo, Cheolhee;Im, Jungho;Shin, Yeji;Cho, Dongjin
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1037-1051
    • /
    • 2020
  • The accurate monitoring and forecasting of the intensity of tropical cyclones (TCs) are able to effectively reduce the overall costs of disaster management. In this study, we proposed a multi-task learning (MTL) based deep learning model for real-time TC intensity estimation and forecasting with the lead time of 6-12 hours following the event, based on the fusion of geostationary satellite images and numerical forecast model output. A total of 142 TCs which developed in the Northwest Pacific from 2011 to 2016 were used in this study. The Communications system, the Ocean and Meteorological Satellite (COMS) Meteorological Imager (MI) data were used to extract the images of typhoons, and the Climate Forecast System version 2 (CFSv2) provided by the National Center of Environmental Prediction (NCEP) was employed to extract air and ocean forecasting data. This study suggested two schemes with different input variables to the MTL models. Scheme 1 used only satellite-based input data while scheme 2 used both satellite images and numerical forecast modeling. As a result of real-time TC intensity estimation, Both schemes exhibited similar performance. For TC intensity forecasting with the lead time of 6 and 12 hours, scheme 2 improved the performance by 13% and 16%, respectively, in terms of the root mean squared error (RMSE) when compared to scheme 1. Relative root mean squared errors(rRMSE) for most intensity levels were lessthan 30%. The lower mean absolute error (MAE) and RMSE were found for the lower intensity levels of TCs. In the test results of the typhoon HALONG in 2014, scheme 1 tended to overestimate the intensity by about 20 kts at the early development stage. Scheme 2 slightly reduced the error, resulting in an overestimation by about 5 kts. The MTL models reduced the computational cost about 300% when compared to the single-tasking model, which suggested the feasibility of the rapid production of TC intensity forecasts.