• Title/Summary/Keyword: Generate Data

Search Result 3,066, Processing Time 0.029 seconds

Implementation of Reporting Tool Supporting OLAP and Data Mining Analysis Using XMLA (XMLA를 사용한 OLAP과 데이타 마이닝 분석이 가능한 리포팅 툴의 구현)

  • Choe, Jee-Woong;Kim, Myung-Ho
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.15 no.3
    • /
    • pp.154-166
    • /
    • 2009
  • Database query and reporting tools, OLAP tools and data mining tools are typical front-end tools in Business Intelligence environment which is able to support gathering, consolidating and analyzing data produced from business operation activities and provide access to the result to enterprise's users. Traditional reporting tools have an advantage of creating sophisticated dynamic reports including SQL query result sets, which look like documents produced by word processors, and publishing the reports to the Web environment, but data source for the tools is limited to RDBMS. On the other hand, OLAP tools and data mining tools have an advantage of providing powerful information analysis functions on each own way, but built-in visualization components for analysis results are limited to tables or some charts. Thus, this paper presents a system that integrates three typical front-end tools to complement one another for BI environment. Traditional reporting tools only have a query editor for generating SQL statements to bring data from RDBMS. However, the reporting tool presented by this paper can extract data also from OLAP and data mining servers, because editors for OLAP and data mining query requests are added into this tool. Traditional systems produce all documents in the server side. This structure enables reporting tools to avoid repetitive process to generate documents, when many clients intend to access the same dynamic document. But, because this system targets that a few users generate documents for data analysis, this tool generates documents at the client side. Therefore, the tool has a processing mechanism to deal with a number of data despite the limited memory capacity of the report viewer in the client side. Also, this reporting tool has data structure for integrating data from three kinds of data sources into one document. Finally, most of traditional front-end tools for BI are dependent on data source architecture from specific vendor. To overcome the problem, this system uses XMLA that is a protocol based on web service to access to data sources for OLAP and data mining services from various vendors.

Assessment of Trophic State for Daecheong reservoir Using Landsat TM Imagery Data (Landsat TM 영상자료를 이용한 대청호의 영양상태 평가)

  • Han, E.J.;Kim, K.T.;Jeong, D.H.;Cheon, S.Y.;Kim, S.J.;Yu, S.J.;Hwang, J.Y.;Kim, T.S.;Kim, M.H.
    • Journal of Environmental Impact Assessment
    • /
    • v.7 no.1
    • /
    • pp.81-91
    • /
    • 1998
  • The objective of this study was to use remotely sensed data, combined with in situ data, for the assessment of trophic state for Daecheong reservoir. Three Landsat TM(Thematic Mapper) imagery data were processed to portray trophic state conditions. The remotely sensed data and the measured data were obtained on 20 June 1995. Regression models have been developed between the chlorophyll-a concentration and reflectance which was converted to Landsat TM digital data. The regression model was determined based on the correlation coefficient which was higher than 0.7 and was applied to the entire study area to generate a distribution map of chlorophyll-a and trophic state. The equation, providing estimates of chlorophyll-a concentration, represented the year-to-year spatial variation of trophic zones in the reservoir. Satellite remote sensing data derived from Landsat TM had been successfully used for trophic slate mapping in Daecheong reservoir.

  • PDF

Statistical Characteristics and Stochastic Modeling of Water Quality Data at the Influent of Daejeon Wastewater Treatment Plant (대전시 공공하수처리시설 유입수 수질자료의 통계적 특성 및 추계학적 모의)

  • Pak, Gijung;Jung, Minjae;Lee, Hansaem;Kim, Deokwoo;Yoon, Jaeyong;Paik, Kyungrock
    • Journal of Korean Society on Water Environment
    • /
    • v.28 no.1
    • /
    • pp.38-49
    • /
    • 2012
  • In this study, we analyze statistical characteristics of influent water quality in Daejeon waste water treatment plant and apply a stochastic model for data generation. In the analysis, the influent water quality data from year 2003 to 2008, except for year 2006, are used. Among water quality variables, we find strong correlations between BOD and T-N; T-N and T-P; BOD and T-P; $COD_{Mn}$ and T-P; and BOD and $COD_{Mn}$. We also find that different water quality variables follow different theoretical probability distribution functions, which also depends on whether the seasonal cycle is removed. Finally, we generate the influent water quality data using the multi-season 1st Markov model (Thomas-Fiering model). With model parameters calibrated for the period 2003~2005, the generated data for 2007~2008 are well compared with observed data showing good agreement in general. BOD and T-N are underestimated by the stochastic model. This is mainly due to the statistical difference in observed data itself between two periods of 2003~2005 and 2007~2008. Therefore, we expect the stochastic model can be applied with more confidence in the case that the data follows stationary pattern.

Estimation of Sea Surface Wind Speed and Direction From RADARSAT Data

  • Kim, Duk-Jin;Wooil-M. Moon
    • Proceedings of the KSRS Conference
    • /
    • 1999.11a
    • /
    • pp.485-490
    • /
    • 1999
  • Wind vector information over the ocean is currently obtained using multiple beam scatterometer data. The scatterometers on ERS-1/2 generate wind vector information with a spatial resolution of 50km and accuracies of $\pm$2m/s in wind speed and $\pm$20$^{\circ}$ in wind direction. Synthetic aperture radar (SAR) data over the ocean have the potential of providing wind vector information independent of weather conditions with finer resolution. Finer resolution wind vector information can often be useful particularly in coastal regions where the scatterometer wind information is often corrupted because of the lower resolution system characteristics which is often contaminated by the signal returns from the coastal areas or ice in the case of arctic environments. In this paper we tested CMOD_4 and CMOD_IFR2 algorithms for extracting the wind vector from SAR data. These algorithms require precise estimation of normalized radar cross-section and wind direction from the SAR data and the local incidence angle. The CMOD series algorithms were developed for the C-band, VV-Polarized SAR data, typically for the ERS SAR data. Since RADARSAT operates at the same C-band but with HH-Polarization, the CMOD series algorithms should not be used directly. As a preliminary approach of resolving with this problem, we applied the polarization ratio between the HH and VV polarizations in the wind vectors estimation. Two test areas, one in front of Inchon and several sites around Jeju island were selected and investigated for wind vector estimation. The new results were compared with the wind vectors obtained from CMOD algorithms. The wind vector results agree well with the observed wind speed data. However the estimation of wind direction agree with the observed wind direction only when the wind speed is greater than approximately 3.0m/s.

  • PDF

Targeting Data Service for Web-Based Media Contents (웹 기반 미디어 콘텐츠를 위한 맞춤형 데이터 서비스)

  • Park, Sung-Joo;Chung, Kwang-Sue
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.12
    • /
    • pp.1154-1164
    • /
    • 2010
  • As an useful application in broadcasting services, the targeting service has been mainly studied to improve the service satisfaction and user usage in various media service environments based on user profile, preferences, and usage history. Targeting service is expanding its domain from broadcasting contents to interstitial contents and from fixed TV devices to mobile devices. Service data also include advertisement data, coupon, and information about media contents as well as simple broadcasting data. In this paper, the targeting data service is designed and implemented on articles, advertisement and broadcasting information on the basis of the user information. To adapt this to web-based media contents, information on user profile, preferences, and usage history is newly defined on the basis of the user metadata developed in TV-Anytime Forum and the user information defined in OpenSocial. The targeting data service is implemented to generate user preferences information and usage history pattern based on the similarity among user preference, contents information, and usage history. Based on performance evaluation, we prove that the proposed targeting data service is effectively applicable to web-based media contents as well as broadcasting service.

Automated Test Data Generation for Dynamic Branch Coverage (동적 분기 커버리지를 위한 테스트 데이터 자동 생성)

  • Chung, In Sang
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.7
    • /
    • pp.451-460
    • /
    • 2013
  • In order to achieve high test coverage, it is usual to generate test data using various techniques including symbolic execution, data flow analysis or constraints solving. Recently, a technique for automated test data generation that fulfills high coverage effectively without those sophisticated means has been proposed. However, the technique shows its weakness in the generation of test data that leads to high coverage for programs having branch conditions where different memory locations are binded during execution. For certain programs with flag conditions, in particular, high coverage can not be achieved because specific branches are not executed. To address the problem, this paper presents dynamic branch coverage criteria and a test data generation technique based on the notion of dynamic branch. It is shown that the proposed technique compared to the previous approach is more effective by conducting experiments involving programs with flag conditions.

Accuracy Assessment of DTM Generation Using LIDAR Data (LIDAR 자료를 이용한 DTM 생성 정확도 평가)

  • Yoo Hwan Hee;Kim Seong Sam;Chung Dong Ki;Hong Jae Min
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.23 no.3
    • /
    • pp.261-272
    • /
    • 2005
  • 3D models in urban areas are essential for a variety of applications, such as virtual visualization, GIS, and mobile communications. LIDAR (Light Detection and Ranging) is a relatively new technology for obtaining Digital Terrain Models (DTM) of the earth's surface since manual 3D data reconstruction is very costly and time consuming. In this paper an approach to extract ground and non-ground points data from LIDAR data by using filtering is presented and the accuracy for generating DTM from ground points data is evaluated. Numerous filter algorithms have been developed to date. To determine the performance of filtering, we selected three filters which are based on the concepts for height difference, slope, and morphology, and also were applied two different data acquired from high raised apartments areas and low house areas. From the results it has been found that the accuracy for generating DTM from LIDAR data are 0.16 m and 0.59 m in high raised apartments areas and low house areas respectively. We expect that LIDAR data is used to generate the accurate DTM in urban areas.

Development of CAPSS2SMOKE Program for Standardized Input Data of SMOKE Model (배출 모델 표준입력자료 작성을 위한 CAPSS2SMOKE 프로그램 개발)

  • Lee, Yong-Mi;Lee, Dae-Gyun;Lee, Mi-Hyang;Hong, Sung-Chul;Yoo, Chul;Jang, Kee-Won;Hong, Ji-Hyung;Lee, Suk-Jo
    • Journal of Korean Society for Atmospheric Environment
    • /
    • v.29 no.6
    • /
    • pp.838-848
    • /
    • 2013
  • The Community Multiscale Air Quality (CMAQ) model is capable of providing high quality atmospheric chemistry profiles through the utilization of high-resolution meteorology and emissions data. However, it cannot simulate air quality accurately if input data are not appropriate and reliable. One of the most important inputs required by CMAQ is the air pollutants emissions, which determines air pollutants concentrations during the simulation. For the CMAQ simulation of Korean peninsula, we, in general, use the Korean National Emission Inventory data which are estimated by Clean Air Policy Support System (CAPSS). However, since they are not provided by model-ready emission data, we should convert CAPSS emissions into model-ready data. The SMOKE is the emission model we used in this study to generate CMAQ-ready emissions. Because processing the emissions data is very monotonous and tedious work, we have developed CAPSS2SMOKE program to convert CAPSS emissions into SMOKE-ready data with ease and effective. CAPSS2SMOKE program consists of many codes and routines such as source classification code, $PM_{10}$ to $PM_{2.5}$ ratio code, map projection conversion routine, spatial allocation routine, and so on. To verify the CAPSS2SMOKE program, we have run SMOKE using the CAPSS 2009 emissions and found that the SMOKE results inherits CAPSS emissions quite well.

Implementation of Analyzer of the Alert Data using Data Mining (데이타마이닝 기법을 이용한 경보데이타 분석기 구현)

  • 신문선;김은희;문호성;류근호;김기영
    • Journal of KIISE:Databases
    • /
    • v.31 no.1
    • /
    • pp.1-12
    • /
    • 2004
  • As network systems are developed rapidly and network architectures are more complex than before, it needs to use PBNM(Policy-Based Network Management) in network system. Generally, architecture of the PBNM consists of two hierarchical layers: management layer and enforcement layer. A security policy server in the management layer should be able to generate new policy, delete, update the existing policy and decide the policy when security policy is requested. And the security policy server should be able to analyze and manage the alert messages received from Policy enforcement system in the enforcement layer for the available information. In this paper, we propose an alert analyzer using data mining. First, in the framework of the policy-based network security management, we design and implement an alert analyzes that analyzes alert data stored in DBMS. The alert analyzer is a helpful system to manage the fault users or hosts. Second, we implement a data mining system for analyzing alert data. The implemented mining system can support alert analyzer and the high level analyzer efficiently for the security policy management. Finally, the proposed system is evaluated with performance parameter, and is able to find out new alert sequences and similar alert patterns.

A Deep Learning-Based Face Mesh Data Denoising System (딥 러닝 기반 얼굴 메쉬 데이터 디노이징 시스템)

  • Roh, Jihyun;Im, Hyeonseung;Kim, Jongmin
    • Journal of IKEEE
    • /
    • v.23 no.4
    • /
    • pp.1250-1256
    • /
    • 2019
  • Although one can easily generate real-world 3D mesh data using a 3D printer or a depth camera, the generated data inevitably includes unnecessary noise. Therefore, mesh denoising is essential to obtain intact 3D mesh data. However, conventional mathematical denoising methods require preprocessing and often eliminate some important features of the 3D mesh. To address this problem, this paper proposes a deep learning based 3D mesh denoising method. Specifically, we propose a convolution-based autoencoder model consisting of an encoder and a decoder. The convolution operation applied to the mesh data performs denoising considering the relationship between each vertex constituting the mesh data and the surrounding vertices. When the convolution is completed, a sampling operation is performed to improve the learning speed. Experimental results show that the proposed autoencoder model produces faster and higher quality denoised data than the conventional methods.