• Title/Summary/Keyword: feature refinement

Search Result 46, Processing Time 0.024 seconds

FEROM: Feature Extraction and Refinement for Opinion Mining

  • Jeong, Ha-Na;Shin, Dong-Wook;Choi, Joong-Min
    • ETRI Journal
    • /
    • v.33 no.5
    • /
    • pp.720-730
    • /
    • 2011
  • Opinion mining involves the analysis of customer opinions using product reviews and provides meaningful information including the polarity of the opinions. In opinion mining, feature extraction is important since the customers do not normally express their product opinions holistically but separately according to its individual features. However, previous research on feature-based opinion mining has not had good results due to drawbacks, such as selecting a feature considering only syntactical grammar information or treating features with similar meanings as different. To solve these problems, this paper proposes an enhanced feature extraction and refinement method called FEROM that effectively extracts correct features from review data by exploiting both grammatical properties and semantic characteristics of feature words and refines the features by recognizing and merging similar ones. A series of experiments performed on actual online review data demonstrated that FEROM is highly effective at extracting and refining features for analyzing customer review data and eventually contributes to accurate and functional opinion mining.

Image retrieval algorithm based on feature vector using color of histogram refinement (칼라 히스토그램 정제를 이용한 특징벡터 기반 영상 검색 알고리즘)

  • Kang, Ji-Young;Park, Jong-An;Beak, Jung-Uk
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.376-379
    • /
    • 2008
  • This paper presents an image retrieval algorithm based on feature vector using color of histogram refinement for a faster and more efficient search in the process of content based image retrieval. First, we segment each of R, G, and B images from RGB color image and extract their respective histograms. Secondly, these histograms of individual R, G and B are divided into sixteen of bins each. Finally, we extract the maximum pixel values in each bins' histogram, which are calculated, compared and analyzed, Now, we can perform image retrieval technique using these maximum pixel value. Hence, the proposed algorithm of this paper effectively extracts features by comparing input and database images, making features from R, G and B into a feature vector table, and prove a batter searching performance than the current algorithm that uses histogram matching and ranks, only.

  • PDF

Application Examples Applying Extended Data Expression Technique to Classification Problems (패턴 분류 문제에 확장된 데이터 표현 기법을 적용한 응용 사례)

  • Lee, Jong Chan
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.12
    • /
    • pp.9-15
    • /
    • 2018
  • The main goal of extended data expression is to develop a data structure suitable for common problems in ubiquitous environments. The greatest feature of this method is that the attribute values can be represented with probability. The next feature is that each event in the training data has a weight value that represents its importance. After this data structure has been developed, an algorithm has been devised that can learn it. In the meantime, this algorithm has been applied to various problems in various fields to obtain good results. This paper first introduces the extended data expression technique, UChoo, and rule refinement method, which are the theoretical basis. Next, this paper introduces some examples of application areas such as rule refinement, missing data processing, BEWS problem, and ensemble system.

Refinement of Disparity Map using the Rule-based Fusion of Area and Feature-based Matching Results

  • Um, Gi-Mun;Ahn, Chung-Hyun;Kim, Kyung-Ok;Lee, Kwae-Hi
    • Proceedings of the KSRS Conference
    • /
    • 1999.11a
    • /
    • pp.304-309
    • /
    • 1999
  • In this paper, we presents a new disparity map refinement algorithm using statistical characteristics of disparity map and edge information. The proposed algorithm generate a refined disparity map using disparity maps which are obtained from area and feature-based Stereo Matching by selecting a disparity value of edge point based on the statistics of both disparity maps. Experimental results on aerial stereo image show the better results than conventional fusion algorithms in the disparity error. This algorithm can be applied to the reconstruction of building image from the high resolution remote sensing data.

  • PDF

ISFRNet: A Deep Three-stage Identity and Structure Feature Refinement Network for Facial Image Inpainting

  • Yan Wang;Jitae Shin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.3
    • /
    • pp.881-895
    • /
    • 2023
  • Modern image inpainting techniques based on deep learning have achieved remarkable performance, and more and more people are working on repairing more complex and larger missing areas, although this is still challenging, especially for facial image inpainting. For a face image with a huge missing area, there are very few valid pixels available; however, people have an ability to imagine the complete picture in their mind according to their subjective will. It is important to simulate this capability while maintaining the identity features of the face as much as possible. To achieve this goal, we propose a three-stage network model, which we refer to as the identity and structure feature refinement network (ISFRNet). ISFRNet is based on 1) a pre-trained pSp-styleGAN model that generates an extremely realistic face image with rich structural features; 2) a shallow structured network with a small receptive field; and 3) a modified U-net with two encoders and a decoder, which has a large receptive field. We choose structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), L1 Loss and learned perceptual image patch similarity (LPIPS) to evaluate our model. When the missing region is 20%-40%, the above four metric scores of our model are 28.12, 0.942, 0.015 and 0.090, respectively. When the lost area is between 40% and 60%, the metric scores are 23.31, 0.840, 0.053 and 0.177, respectively. Our inpainting network not only guarantees excellent face identity feature recovery but also exhibits state-of-the-art performance compared to other multi-stage refinement models.

Dual-stream Co-enhanced Network for Unsupervised Video Object Segmentation

  • Hongliang Zhu;Hui Yin;Yanting Liu;Ning Chen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.938-958
    • /
    • 2024
  • Unsupervised Video Object Segmentation (UVOS) is a highly challenging problem in computer vision as the annotation of the target object in the testing video is unknown at all. The main difficulty is to effectively handle the complicated and changeable motion state of the target object and the confusion of similar background objects in video sequence. In this paper, we propose a novel deep Dual-stream Co-enhanced Network (DC-Net) for UVOS via bidirectional motion cues refinement and multi-level feature aggregation, which can fully take advantage of motion cues and effectively integrate different level features to produce high-quality segmentation mask. DC-Net is a dual-stream architecture where the two streams are co-enhanced by each other. One is a motion stream with a Motion-cues Refine Module (MRM), which learns from bidirectional optical flow images and produces fine-grained and complete distinctive motion saliency map, and the other is an appearance stream with a Multi-level Feature Aggregation Module (MFAM) and a Context Attention Module (CAM) which are designed to integrate the different level features effectively. Specifically, the motion saliency map obtained by the motion stream is fused with each stage of the decoder in the appearance stream to improve the segmentation, and in turn the segmentation loss in the appearance stream feeds back into the motion stream to enhance the motion refinement. Experimental results on three datasets (Davis2016, VideoSD, SegTrack-v2) demonstrate that DC-Net has achieved comparable results with some state-of-the-art methods.

Improved Statistical Grey-Level Models for PCB Inspection (PCB 검사를 위한 개선된 통계적 그레이레벨 모델)

  • Bok, Jin Seop;Cho, Tai-Hoon
    • Journal of the Semiconductor & Display Technology
    • /
    • v.12 no.1
    • /
    • pp.1-7
    • /
    • 2013
  • Grey-level statistical models have been widely used in many applications for object location and identification. However, conventional models yield some problems in model refinement when training images are not properly aligned, and have difficulties for real-time recognition of arbitrarily rotated models. This paper presents improved grey-level statistical models that align training images using image or feature matching to overcome problems in model refinement of conventional models, and that enable real-time recognition of arbitrarily rotated objects using efficient hierarchical search methods. Edges or features extracted from a mean training image are used for accurate alignment of models in the search image. On the aligned position and orientation, fitness measure based on grey-level statistical models is computed for object recognition. It is demonstrated in various experiments in PCB inspection that proposed methods are superior to conventional methods in recognition accuracy and speed.

A Three-scale Pedestrian Detection Method based on Refinement Module (Refinement Module 기반 Three-Scale 보행자 검출 기법)

  • Kyungmin Jung;Sooyong Park;Hyun Lee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.5
    • /
    • pp.259-265
    • /
    • 2023
  • Pedestrian detection is used to effectively detect pedestrians in various situations based on deep learning. Pedestrian detection has difficulty detecting pedestrians due to problems such as camera performance, pedestrian description, height, and occlusion. Even in the same pedestrian, performance in detecting them can differ according to the height of the pedestrian. The height of general pedestrians encompasses various scales, such as those of infants, adolescents, and adults, so when the model is applied to one group, the extraction of data becomes inaccurate. Therefore, this study proposed a pedestrian detection method that fine-tunes the pedestrian area by Refining Layer and Feature Concatenation to consider various heights of pedestrians. Through this, the score and location value for the pedestrian area were finely adjusted. Experiments on four types of test data demonstrate that the proposed model achieves 2-5% higher average precision (AP) compared to Faster R-CNN and DRPN.

Residual Learning Based CNN for Gesture Recognition in Robot Interaction

  • Han, Hua
    • Journal of Information Processing Systems
    • /
    • v.17 no.2
    • /
    • pp.385-398
    • /
    • 2021
  • The complexity of deep learning models affects the real-time performance of gesture recognition, thereby limiting the application of gesture recognition algorithms in actual scenarios. Hence, a residual learning neural network based on a deep convolutional neural network is proposed. First, small convolution kernels are used to extract the local details of gesture images. Subsequently, a shallow residual structure is built to share weights, thereby avoiding gradient disappearance or gradient explosion as the network layer deepens; consequently, the difficulty of model optimisation is simplified. Additional convolutional neural networks are used to accelerate the refinement of deep abstract features based on the spatial importance of the gesture feature distribution. Finally, a fully connected cascade softmax classifier is used to complete the gesture recognition. Compared with the dense connection multiplexing feature information network, the proposed algorithm is optimised in feature multiplexing to avoid performance fluctuations caused by feature redundancy. Experimental results from the ISOGD gesture dataset and Gesture dataset prove that the proposed algorithm affords a fast convergence speed and high accuracy.

Stroke Width-Based Contrast Feature for Document Image Binarization

  • Van, Le Thi Khue;Lee, Gueesang
    • Journal of Information Processing Systems
    • /
    • v.10 no.1
    • /
    • pp.55-68
    • /
    • 2014
  • Automatic segmentation of foreground text from the background in degraded document images is very much essential for the smooth reading of the document content and recognition tasks by machine. In this paper, we present a novel approach to the binarization of degraded document images. The proposed method uses a new local contrast feature extracted based on the stroke width of text. First, a pre-processing method is carried out for noise removal. Text boundary detection is then performed on the image constructed from the contrast feature. Then local estimation follows to extract text from the background. Finally, a refinement procedure is applied to the binarized image as a post-processing step to improve the quality of the final results. Experiments and comparisons of extracting text from degraded handwriting and machine-printed document image against some well-known binarization algorithms demonstrate the effectiveness of the proposed method.