• Title/Summary/Keyword: Computer vision technology

Search Result 666, Processing Time 0.032 seconds

Personalized Face Modeling for Photorealistic Synthesis

  • Kim, Kyungmin;Shim, Hyunjung
    • Journal of International Society for Simulation Surgery
    • /
    • v.2 no.2
    • /
    • pp.47-51
    • /
    • 2015
  • Faces play a key role in revealing the personalized attributes such as the identity, emotion, health condition, etc. Due to the importance of faces, computer-assisted face modeling and reconstruction have been actively studied both in computer vision and graphics community. Especially, face reconstruction and realistic face synthesis are well-grounded research problems and various approaches have been proposed during the last decade. In this paper, we discuss a wide range of existing work in face modeling by introducing their target applications, categorizing them upon their methodology and addressing their strength and weakness on performance. Finally, we introduce remaining research issues and suggest the future research direction in face modeling. We believe that this paper provides a high-level overview on face modeling techniques and helps understand the major research issues and the trends of methodology.

An Improved Multiple Interval Pixel Sampling based Background Subtraction Algorithm (개선된 다중 구간 샘플링 배경제거 알고리즘)

  • Mahmood, Muhammad Tariq;Choi, Young Kyu
    • Journal of the Semiconductor & Display Technology
    • /
    • v.18 no.3
    • /
    • pp.1-6
    • /
    • 2019
  • Foreground/background segmentation in video sequences is often one of the first tasks in machine vision applications, making it a critical part of the system. In this paper, we present an improved sample-based technique that provides robust background image as well as segmentation mask. The conventional multiple interval sampling (MIS) algorithm have suffer from the unbalance of computation time per frame and the rapid change of confidence factor of background pixel. To balance the computation amount, a random-based pixel update scheme is proposed and a spatial and temporal smoothing technique is adopted to increase reliability of the confidence factor. The proposed method allows the sampling queue to have more dispersed data in time and space, and provides more continuous and reliable confidence factor. Experimental results revealed that our method works well to estimate stable background image and the foreground mask.

U2Net-based Single-pixel Imaging Salient Object Detection

  • Zhang, Leihong;Shen, Zimin;Lin, Weihong;Zhang, Dawei
    • Current Optics and Photonics
    • /
    • v.6 no.5
    • /
    • pp.463-472
    • /
    • 2022
  • At certain wavelengths, single-pixel imaging is considered to be a solution that can achieve high quality imaging and also reduce costs. However, achieving imaging of complex scenes is an overhead-intensive process for single-pixel imaging systems, so low efficiency and high consumption are the biggest obstacles to their practical application. Improving efficiency to reduce overhead is the solution to this problem. Salient object detection is usually used as a pre-processing step in computer vision tasks, mimicking human functions in complex natural scenes, to reduce overhead and improve efficiency by focusing on regions with a large amount of information. Therefore, in this paper, we explore the implementation of salient object detection based on single-pixel imaging after a single pixel, and propose a scheme to reconstruct images based on Fourier bases and use U2Net models for salient object detection.

Deep Reference-based Dynamic Scene Deblurring

  • Cunzhe Liu;Zhen Hua;Jinjiang Li
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.3
    • /
    • pp.653-669
    • /
    • 2024
  • Dynamic scene deblurring is a complex computer vision problem owing to its difficulty to model mathematically. In this paper, we present a novel approach for image deblurring with the help of the sharp reference image, which utilizes the reference image for high-quality and high-frequency detail results. To better utilize the clear reference image, we develop an encoder-decoder network and two novel modules are designed to guide the network for better image restoration. The proposed Reference Extraction and Aggregation Module can effectively establish the correspondence between blurry image and reference image and explore the most relevant features for better blur removal and the proposed Spatial Feature Fusion Module enables the encoder to perceive blur information at different spatial scales. In the final, the multi-scale feature maps from the encoder and cascaded Reference Extraction and Aggregation Modules are integrated into the decoder for a global fusion and representation. Extensive quantitative and qualitative experimental results from the different benchmarks show the effectiveness of our proposed method.

A Survey of Face Recognition Techniques

  • Jafri, Rabia;Arabnia, Hamid R.
    • Journal of Information Processing Systems
    • /
    • v.5 no.2
    • /
    • pp.41-68
    • /
    • 2009
  • Face recognition presents a challenging problem in the field of image analysis and computer vision, and as such has received a great deal of attention over the last few years because of its many applications in various domains. Face recognition techniques can be broadly divided into three categories based on the face data acquisition methodology: methods that operate on intensity images; those that deal with video sequences; and those that require other sensory data such as 3D information or infra-red imagery. In this paper, an overview of some of the well-known methods in each of these categories is provided and some of the benefits and drawbacks of the schemes mentioned therein are examined. Furthermore, a discussion outlining the incentive for using face recognition, the applications of this technology, and some of the difficulties plaguing current systems with regard to this task has also been provided. This paper also mentions some of the most recent algorithms developed for this purpose and attempts to give an idea of the state of the art of face recognition technology.

Comparative Analysis of the Performance of SIFT and SURF (SIFT 와 SURF 알고리즘의 성능적 비교 분석)

  • Lee, Yong-Hwan;Park, Je-Ho;Kim, Youngseop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.12 no.3
    • /
    • pp.59-64
    • /
    • 2013
  • Accurate and robust image registration is important task in many applications such as image retrieval and computer vision. To perform the image registration, essential required steps are needed in the process: feature detection, extraction, matching, and reconstruction of image. In the process of these function, feature extraction not only plays a key role, but also have a big effect on its performance. There are two representative algorithms for extracting image features, which are scale invariant feature transform (SIFT) and speeded up robust feature (SURF). In this paper, we present and evaluate two methods, focusing on comparative analysis of the performance. Experiments for accurate and robust feature detection are shown on various environments such like scale changes, rotation and affine transformation. Experimental trials revealed that SURF algorithm exhibited a significant result in both extracting feature points and matching time, compared to SIFT method.

Scaling Up Face Masks Classification Using a Deep Neural Network and Classical Method Inspired Hybrid Technique

  • Kumar, Akhil;Kalia, Arvind;Verma, Kinshuk;Sharma, Akashdeep;Kaushal, Manisha;Kalia, Aayushi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.11
    • /
    • pp.3658-3679
    • /
    • 2022
  • Classification of persons wearing and not wearing face masks in images has emerged as a new computer vision problem during the COVID-19 pandemic. In order to address this problem and scale up the research in this domain, in this paper a hybrid technique by employing ResNet-101 and multi-layer perceptron (MLP) classifier has been proposed. The proposed technique is tested and validated on a self-created face masks classification dataset and a standard dataset. On self-created dataset, the proposed technique achieved a classification accuracy of 97.3%. To embrace the proposed technique, six other state-of-the-art CNN feature extractors with six other classical machine learning classifiers have been tested and compared with the proposed technique. The proposed technique achieved better classification accuracy and 1-6% higher precision, recall, and F1 score as compared to other tested deep feature extractors and machine learning classifiers.

An Efficient Deep Learning Based Image Recognition Service System Using AWS Lambda Serverless Computing Technology (AWS Lambda Serverless Computing 기술을 활용한 효율적인 딥러닝 기반 이미지 인식 서비스 시스템)

  • Lee, Hyunchul;Lee, Sungmin;Kim, Kangseok
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.9 no.6
    • /
    • pp.177-186
    • /
    • 2020
  • Recent advances in deep learning technology have improved image recognition performance in the field of computer vision, and serverless computing is emerging as the next generation cloud computing technology for event-based cloud application development and services. Attempts to use deep learning and serverless computing technology to increase the number of real-world image recognition services are increasing. Therefore, this paper describes how to develop an efficient deep learning based image recognition service system using serverless computing technology. The proposed system suggests a method that can serve large neural network model to users at low cost by using AWS Lambda Server based on serverless computing. We also show that we can effectively build a serverless computing system that uses a large neural network model by addressing the shortcomings of AWS Lambda Server, cold start time and capacity limitation. Through experiments, we confirmed that the proposed system, using AWS Lambda Serverless Computing technology, is efficient for servicing large neural network models by solving processing time and capacity limitations as well as cost reduction.

A Review on Meat Quality Evaluation Methods Based on Non-Destructive Computer Vision and Artificial Intelligence Technologies

  • Shi, Yinyan;Wang, Xiaochan;Borhan, Md Saidul;Young, Jennifer;Newman, David;Berg, Eric;Sun, Xin
    • Food Science of Animal Resources
    • /
    • v.41 no.4
    • /
    • pp.563-588
    • /
    • 2021
  • Increasing meat demand in terms of both quality and quantity in conjunction with feeding a growing population has resulted in regulatory agencies imposing stringent guidelines on meat quality and safety. Objective and accurate rapid non-destructive detection methods and evaluation techniques based on artificial intelligence have become the research hotspot in recent years and have been widely applied in the meat industry. Therefore, this review surveyed the key technologies of non-destructive detection for meat quality, mainly including ultrasonic technology, machine (computer) vision technology, near-infrared spectroscopy technology, hyperspectral technology, Raman spectra technology, and electronic nose/tongue. The technical characteristics and evaluation methods were compared and analyzed; the practical applications of non-destructive detection technologies in meat quality assessment were explored; and the current challenges and future research directions were discussed. The literature presented in this review clearly demonstrate that previous research on non-destructive technologies are of great significance to ensure consumers' urgent demand for high-quality meat by promoting automatic, real-time inspection and quality control in meat production. In the near future, with ever-growing application requirements and research developments, it is a trend to integrate such systems to provide effective solutions for various grain quality evaluation applications.

Automatic identification and analysis of multi-object cattle rumination based on computer vision

  • Yueming Wang;Tiantian Chen;Baoshan Li;Qi Li
    • Journal of Animal Science and Technology
    • /
    • v.65 no.3
    • /
    • pp.519-534
    • /
    • 2023
  • Rumination in cattle is closely related to their health, which makes the automatic monitoring of rumination an important part of smart pasture operations. However, manual monitoring of cattle rumination is laborious and wearable sensors are often harmful to animals. Thus, we propose a computer vision-based method to automatically identify multi-object cattle rumination, and to calculate the rumination time and number of chews for each cow. The heads of the cattle in the video were initially tracked with a multi-object tracking algorithm, which combined the You Only Look Once (YOLO) algorithm with the kernelized correlation filter (KCF). Images of the head of each cow were saved at a fixed size, and numbered. Then, a rumination recognition algorithm was constructed with parameters obtained using the frame difference method, and rumination time and number of chews were calculated. The rumination recognition algorithm was used to analyze the head image of each cow to automatically detect multi-object cattle rumination. To verify the feasibility of this method, the algorithm was tested on multi-object cattle rumination videos, and the results were compared with the results produced by human observation. The experimental results showed that the average error in rumination time was 5.902% and the average error in the number of chews was 8.126%. The rumination identification and calculation of rumination information only need to be performed by computers automatically with no manual intervention. It could provide a new contactless rumination identification method for multi-cattle, which provided technical support for smart pasture.