• Title/Summary/Keyword: 최적 필터

Search Result 695, Processing Time 0.024 seconds

Increasing Accuracy of Classifying Useful Reviews by Removing Neutral Terms (중립도 기반 선택적 단어 제거를 통한 유용 리뷰 분류 정확도 향상 방안)

  • Lee, Minsik;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.129-142
    • /
    • 2016
  • Customer product reviews have become one of the important factors for purchase decision makings. Customers believe that reviews written by others who have already had an experience with the product offer more reliable information than that provided by sellers. However, there are too many products and reviews, the advantage of e-commerce can be overwhelmed by increasing search costs. Reading all of the reviews to find out the pros and cons of a certain product can be exhausting. To help users find the most useful information about products without much difficulty, e-commerce companies try to provide various ways for customers to write and rate product reviews. To assist potential customers, online stores have devised various ways to provide useful customer reviews. Different methods have been developed to classify and recommend useful reviews to customers, primarily using feedback provided by customers about the helpfulness of reviews. Most shopping websites provide customer reviews and offer the following information: the average preference of a product, the number of customers who have participated in preference voting, and preference distribution. Most information on the helpfulness of product reviews is collected through a voting system. Amazon.com asks customers whether a review on a certain product is helpful, and it places the most helpful favorable and the most helpful critical review at the top of the list of product reviews. Some companies also predict the usefulness of a review based on certain attributes including length, author(s), and the words used, publishing only reviews that are likely to be useful. Text mining approaches have been used for classifying useful reviews in advance. To apply a text mining approach based on all reviews for a product, we need to build a term-document matrix. We have to extract all words from reviews and build a matrix with the number of occurrences of a term in a review. Since there are many reviews, the size of term-document matrix is so large. It caused difficulties to apply text mining algorithms with the large term-document matrix. Thus, researchers need to delete some terms in terms of sparsity since sparse words have little effects on classifications or predictions. The purpose of this study is to suggest a better way of building term-document matrix by deleting useless terms for review classification. In this study, we propose neutrality index to select words to be deleted. Many words still appear in both classifications - useful and not useful - and these words have little or negative effects on classification performances. Thus, we defined these words as neutral terms and deleted neutral terms which are appeared in both classifications similarly. After deleting sparse words, we selected words to be deleted in terms of neutrality. We tested our approach with Amazon.com's review data from five different product categories: Cellphones & Accessories, Movies & TV program, Automotive, CDs & Vinyl, Clothing, Shoes & Jewelry. We used reviews which got greater than four votes by users and 60% of the ratio of useful votes among total votes is the threshold to classify useful and not-useful reviews. We randomly selected 1,500 useful reviews and 1,500 not-useful reviews for each product category. And then we applied Information Gain and Support Vector Machine algorithms to classify the reviews and compared the classification performances in terms of precision, recall, and F-measure. Though the performances vary according to product categories and data sets, deleting terms with sparsity and neutrality showed the best performances in terms of F-measure for the two classification algorithms. However, deleting terms with sparsity only showed the best performances in terms of Recall for Information Gain and using all terms showed the best performances in terms of precision for SVM. Thus, it needs to be careful for selecting term deleting methods and classification algorithms based on data sets.

Application of Biofilter for the Removal of VOCs Produced in the Remediation of Oil-Contaminated Soil (유류오염 토양의 복원과정에서 발생되는 휘발성 유기화합물의 제거를 위한 바이오필터의 적용)

  • Lee Eun Young;Choi Woo-Zin;Choi Jin-Kyu
    • Journal of Soil and Groundwater Environment
    • /
    • v.10 no.1
    • /
    • pp.35-42
    • /
    • 2005
  • This research was investigated the applicability of the biofiltration technology for the removal of volatile organic carbons (VOCs) produced from the bioremediation of oil contaminated soil. Diesel was used as surrogate for oil and, two types of biofilter systems made of ceramic and polymer media were compared for the removal efficiencies of diesel VOCs at different inlet concentrations and space velocity (SV) conditions. During the first 30-d operation, the removal efficiencies of the biofilter packed with polymer and the biofilter packed with ceramic were investigated at constant SV of $153\;h^{-1}$ When inlet concentrations of diesel VOCs were below 10 ppmv, the average removal efficiencies of the polymer biofilter and the ceramic biofilter were average $67\%\;and\;75\%$, respectively. When the inlet concentration increased to 30 ppmv, the VOC removal efficiency in the polymer biofilter was $80\%$, while the average removal efficiency in the ceramic biofilter was $60\%. Effect of the inlet concentration and SV on the removal efficiency of total diesel VOCs was investigated. As SV increased from $153\;h^{-1}$ to $204\;h^{-1}$ and $306\;h^{-1}$, the removal efficiency of total diesel VOCs was decreased gradually. The average removal efficiency of the biofilter packed with polymer carrier was decreased from $82\%\;to\;80\%\;and\;77\%$. The biofilter packed with polymer carrier showed that the removal efficiency of benzene and toluene were maintained within the range of $81\%\~86\%$. In contrast, for the biofilter packed with ceramic carrier, when SV increased from $153\;h^{-1}$ to $204\;h^{-1}$ and $306\;h^{-1}$, the removal efficiency of benzene decreased from $87\%\;to79\%\;and\;74\% . respectively. The removal efficiency of toluene decreased from $80\%\;to\;77\%\;and\;76\%$ at SV of $153\;h^{-1},\;204\;h^{-1}\;and\;306\;h^{-1}$, and $306\;h^{-1}$, respectively.

A Smoothing Data Cleaning based on Adaptive Window Sliding for Intelligent RFID Middleware Systems (지능적인 RFID 미들웨어 시스템을 위한 적응형 윈도우 슬라이딩 기반의 유연한 데이터 정제)

  • Shin, DongCheon;Oh, Dongok;Ryu, SeungWan;Park, Seikwon
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.1-18
    • /
    • 2014
  • Over the past years RFID/SN has been an elementary technology in a diversity of applications for the ubiquitous environments, especially for Internet of Things. However, one of obstacles for widespread deployment of RFID technology is the inherent unreliability of the RFID data streams by tag readers. In particular, the problem of false readings such as lost readings and mistaken readings needs to be treated by RFID middleware systems because false readings ultimately degrade the quality of application services due to the dirty data delivered by middleware systems. As a result, for the higher quality of services, an RFID middleware system is responsible for intelligently dealing with false readings for the delivery of clean data to the applications in accordance with the tag reading environment. One of popular techniques used to compensate false readings is a sliding window filter. In a sliding window scheme, it is evident that determining optimal window size intelligently is a nontrivial important task in RFID middleware systems in order to reduce false readings, especially in mobile environments. In this paper, for the purpose of reducing false readings by intelligent window adaption, we propose a new adaptive RFID data cleaning scheme based on window sliding for a single tag. Unlike previous works based on a binomial sampling model, we introduce the weight averaging. Our insight starts from the need to differentiate the past readings and the current readings, since the more recent readings may indicate the more accurate tag transitions. Owing to weight averaging, our scheme is expected to dynamically adapt the window size in an efficient manner even for non-homogeneous reading patterns in mobile environments. In addition, we analyze reading patterns in the window and effects of decreased window so that a more accurate and efficient decision on window adaption can be made. With our scheme, we can expect to obtain the ultimate goal that RFID middleware systems can provide applications with more clean data so that they can ensure high quality of intended services.

Development of Prediction Model for the Na Content of Leaves of Spring Potatoes Using Hyperspectral Imagery (초분광 영상을 이용한 봄감자의 잎 Na 함량 예측 모델 개발)

  • Park, Jun-Woo;Kang, Ye-Seong;Ryu, Chan-Seok;Jang, Si-Hyeong;Kang, Kyung-Suk;Kim, Tae-Yang;Park, Min-Jun;Baek, Hyeon-Chan;Song, Hye-Young;Jun, Sae-Rom;Lee, Su-Hwan
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.23 no.4
    • /
    • pp.316-328
    • /
    • 2021
  • In this study, the leaf Na content prediction model for spring potato was established using 400-1000 nm hyperspectral sensor to develop the multispectral sensor for the salinity monitoring in reclaimed land. The irrigation conditions were standard, drought, and salinity (2, 4, 8 dS/m), and the irrigation amount was calculated based on the amount of evaporation. The leaves' Na contents were measured 1st and 2nd weeks after starting irrigation in the vegetative, tuber formative, and tuber growing periods, respectively. The reflectance of the leaves was converted from 5 nm to 10 nm, 25 nm, and 50 nm of FWHM (full width at half maximum) based on the 10 nm wavelength intervals. Using the variance importance in projections of partial least square regression(PLSR-VIP), ten band ratios were selected as the variables to predict salinity damage levels with Na content of spring potato leaves. The MLR(Multiple linear regression) models were estimated by removing the band ratios one by one in the order of the lowest weight among the ten band ratios. The performance of models was compared by not only R2, MAPE but also the number of band ratios, optimal FWHM to develop the compact multispectral sensor. It was an advantage to use 25 nm of FWHM to predict the amount of Na in leaves for spring potatoes during the 1st and 2nd weeks vegetative and tuber formative periods and 2 weeks tuber growing periods. The selected bandpass filters were 15 bands and mainly in red and red-edge regions such as 430/440, 490/500, 500/510, 550/560, 570/580, 590/600, 640/650, 650/660, 670/680, 680/690, 690/700, 700/710, 710/720, 720/730, 730/740 nm.

Image Evaluation for Optimization of Radiological Protection in CBCT during Image-Guided Radiation Therapy (영상유도 방사선 치료 시 CBCT에서 방사선 방호최적화를 위한 영상평가)

  • Min-Ho Choi;Kyung-Wan Kim;Dong-Yeon Lee
    • Journal of the Korean Society of Radiology
    • /
    • v.17 no.3
    • /
    • pp.305-314
    • /
    • 2023
  • With the development of medical technology and radiation treatment equipment, the frequency of high-precision radiation therapy such as intensity modulation radiation therapy has increased. Image-guided radiation therapy has become essential for radiation therapy in precise and complex treatment plans. In particular, with the introduction of imaging equipment for diagnosis in a linear accelerator, CBCT scanning became possible, which made it possible to calibrate and correct the patient's posture through 3D images. Although more precise reproduction of the patient's posture has become possible, the exposure dose delivered to the patient during the image acquisition process cannot be ignored. Radiation optimization is necessary in the field of radiation therapy, and efforts to reduce exposure are necessary. However, when acquiring 3D CBCT images by changing the imaging conditions to reduce exposure, there should be no image quality or artefacts that would make it impossible to align the patient's position. In this study, Rando phantom was used to scan and evaluate images for each shooting condition. The highest SNR was obtained at 100 kV 80 mA 25 ms F1 filter 180°. As the tube voltage and tube current increased, the noise decreased, and the bowtie filter showed the optimal effect at high tube current. Based on the actual scanned images, it was confirmed that patient alignment was possible under all imaging conditions, and that image-guided radiation therapy for patient alignment was possible under the condition of 70 kV 10 mA 20 ms F0 filter 180°, which showed the lowest SNR. In this study, image evaluation was conducted according to the imaging conditions, and low tube voltage, tube current, and small rotation angle scan are expected to be effective in reducing radiation exposure. Based on this, the patient's exposure dose should be kept as low as possible during CBCT imaging.