• Title/Summary/Keyword: Computer vision technology

Search Result 669, Processing Time 0.025 seconds

Arbitration Award via Modern Technical means in Saudi Arabia

  • Mohammed Sulaiman Alnasyan
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.7
    • /
    • pp.32-38
    • /
    • 2023
  • This study deals with arbitration award via modern technical means; because e-Arbitration is deemed to be one of the most important substitute means for the settlement of disputes arising from electronic transactions. This type of arbitration is characterized by fast settlement of disputes, as well as fast enforcement of awards rendered thereon. The researcher seeks to indicate the content of the award, the conditions for rendering it, and to analyze the legal provisions related to its legal basis in the Saudi Law of Arbitration. This study shows that an arbitration award, rendered via modern technical means has a number of advantages, such as fast settlement, less cost, and keeping pace with modern technology, which is an aim of Saudi Arabia Vision 2030. The study also points out certain problems facing arbitration via technical means; however, the most important of which is the insufficiency of some legal rules associated with traditional arbitration, as contained in the Saudi Law of Arbitrator, which are incompatible with or applicable to an arbitration award which is rendered via modern means.

Development of a Real-time Translation Application using Screen Capture and OCR in Android Environment (안드로이드 환경에서 화면 캡쳐와 OCR을 활용한 실시간 번역 애플리케이션 개발)

  • Seung-Woo Lee;Sung Jin Kim;Young Hyun Yoon;Jai Soon Baek
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2023.07a
    • /
    • pp.267-268
    • /
    • 2023
  • 본 논문은 안드로이드에서 화면 캡쳐와 OCR을 통한 실시간 번역 애플리케이션 개발을 주제로 한다. 코틀린으로 개발된 애플리케이션은 사용자가 원하는 화면 영역을 캡쳐하여 해당 텍스트를 OCR로 추출하고, 구글 Cloud Vision API와 Cloud Translation API를 활용해 번역한다. 이를 통해 외국어 애플리케이션 사용의 편의성을 향상시키고, 정보의 이해와 공유를 도울 수 있음을 제시한다. 이 기술은 더욱 다양한 분야에서의 활용 가능성을 열어놓고 있다.

  • PDF

Lightweight Single Image Super-Resolution Convolution Neural Network in Portable Device

  • Wang, Jin;Wu, Yiming;He, Shiming;Sharma, Pradip Kumar;Yu, Xiaofeng;Alfarraj, Osama;Tolba, Amr
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.11
    • /
    • pp.4065-4083
    • /
    • 2021
  • Super-resolution can improve the clarity of low-resolution (LR) images, which can increase the accuracy of high-level compute vision tasks. Portable devices have low computing power and storage performance. Large-scale neural network super-resolution methods are not suitable for portable devices. In order to save the computational cost and the number of parameters, Lightweight image processing method can improve the processing speed of portable devices. Therefore, we propose the Enhanced Information Multiple Distillation Network (EIMDN) to adapt lower delay and cost. The EIMDN takes feedback mechanism as the framework and obtains low level features through high level features. Further, we replace the feature extraction convolution operation in Information Multiple Distillation Block (IMDB), with Ghost module, and propose the Enhanced Information Multiple Distillation Block (EIMDB) to reduce the amount of calculation and the number of parameters. Finally, coordinate attention (CA) is used at the end of IMDB and EIMDB to enhance the important information extraction from Spaces and channels. Experimental results show that our proposed can achieve convergence faster with fewer parameters and computation, compared with other lightweight super-resolution methods. Under the condition of higher peak signal-to-noise ratio (PSNR) and higher structural similarity (SSIM), the performance of network reconstruction image texture and target contour is significantly improved.

A Proposal of the Olfactory Information Presentation Method and Its Application for Scent Generator Using Web Service

  • Kim, Jeong-Do;Byun, Hyung-Gi
    • Journal of Sensor Science and Technology
    • /
    • v.21 no.4
    • /
    • pp.249-255
    • /
    • 2012
  • Among the human senses, olfactory information still does not have a proper data presentation method unlike that regarding vision and auditory information. It makes presenting the sense of smell into multimedia information impossible, which may be an exploratory field in human computer interaction. In this paper, we propose an olfactory information presentation method, which is a way to use smell as multimedia information, and show an application for scent generation and odor display using a web service. The olfactory information can present smell characteristics such as intensity, persistence, hedonic tone, and odor description. The structure of data format based on olfactory information can also be organized according to data types such as integer, float, char, string, and bitmap. Furthermore, it can be used for data transmitting via a web service and for odor display using a scent generator. The scent generator, which can display information of smell, is developed to generate 6 odors using 6 aroma solutions and a diluted solution with 14 micro-valves and a micropump. Throughout the experiment, we confirm that the remote user can grasp information of smell transmitted by messenger service and request odor display to the computer controlled scent generator. It contributes to enlarge existing virtual reality and to be proposed as a standard reference method regarding olfactory information presentation for future multimedia technology.

Improvement of Active Shape Model for Detecting Face Features in iOS Platform (iOS 플랫폼에서 Active Shape Model 개선을 통한 얼굴 특징 검출)

  • Lee, Yong-Hwan;Kim, Heung-Jun
    • Journal of the Semiconductor & Display Technology
    • /
    • v.15 no.2
    • /
    • pp.61-65
    • /
    • 2016
  • Facial feature detection is a fundamental function in the field of computer vision such as security, bio-metrics, 3D modeling, and face recognition. There are many algorithms for the function, active shape model is one of the most popular local texture models. This paper addresses issues related to face detection, and implements an efficient extraction algorithm for extracting the facial feature points to use on iOS platform. In this paper, we extend the original ASM algorithm to improve its performance by four modifications. First, to detect a face and to initialize the shape model, we apply a face detection API provided from iOS CoreImage framework. Second, we construct a weighted local structure model for landmarks to utilize the edge points of the face contour. Third, we build a modified model definition and fitting more landmarks than the classical ASM. And last, we extend and build two-dimensional profile model for detecting faces within input images. The proposed algorithm is evaluated on experimental test set containing over 500 face images, and found to successfully extract facial feature points, clearly outperforming the original ASM.

Panoramic Image Stitching using Feature Extracting and Matching on Mobile Device (모바일 기기에서 특징적 추출과 정합을 활용한 파노라마 이미지 스티칭)

  • Lee, Yong-Hwan;Kim, Heung-Jun
    • Journal of the Semiconductor & Display Technology
    • /
    • v.15 no.4
    • /
    • pp.97-102
    • /
    • 2016
  • Image stitching is a process of combining two or more images with overlapping area to create a panorama of input images, which is considered as an active research area in computer vision, especially in the field of augmented reality with 360 degree images. Image stitching techniques can be categorized into two general approaches: direct and feature based techniques. Direct techniques compare all the pixel intensities of the images with each other, while feature based approaches aim to determine a relationship between the images through distinct features extracted from the images. This paper proposes a novel image stitching method based on feature pixels with approximated clustering filter. When the features are extracted from input images, we calculate a meaning of the minutiae, and apply an effective feature extraction algorithm to improve the processing time. With the evaluation of the results, the proposed method is corresponding accurate and effective, compared to the previous approaches.

Finger-Pointing Gesture Analysis for Slide Presentation

  • Harika, Maisevli;Setijadi P, Ary;Hindersah, Hilwadi;Sin, Bong-Kee
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.8
    • /
    • pp.1225-1235
    • /
    • 2016
  • This paper presents a method for computer-assisted slide presentation using vision-based gesture recognition. The proposed method consists of a sequence of steps, first detecting a hand in the scene of projector beam, then estimating the smooth trajectory of a hand or a pointing finger using Kalman Filter, and finally interfacing to an application system. Additional slide navigation control includes moving back and forth the pages of the presentation. The proposed method is to help speakers for an effective presentation with natural improved interaction with the computer. In particular, the proposed method of using finger pointing is believed to be more effective than using a laser pointer since the hand, the pointing or finger are more visible and thus can better grab the attention of the audience.

A Distributed Real-time 3D Pose Estimation Framework based on Asynchronous Multiviews

  • Taemin, Hwang;Jieun, Kim;Minjoon, Kim
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.2
    • /
    • pp.559-575
    • /
    • 2023
  • 3D human pose estimation is widely applied in various fields, including action recognition, sports analysis, and human-computer interaction. 3D human pose estimation has achieved significant progress with the introduction of convolutional neural network (CNN). Recently, several researches have proposed the use of multiview approaches to avoid occlusions in single-view approaches. However, as the number of cameras increases, a 3D pose estimation system relying on a CNN may lack in computational resources. In addition, when a single host system uses multiple cameras, the data transition speed becomes inadequate owing to bandwidth limitations. To address this problem, we propose a distributed real-time 3D pose estimation framework based on asynchronous multiple cameras. The proposed framework comprises a central server and multiple edge devices. Each multiple-edge device estimates a 2D human pose from its view and sendsit to the central server. Subsequently, the central server synchronizes the received 2D human pose data based on the timestamps. Finally, the central server reconstructs a 3D human pose using geometrical triangulation. We demonstrate that the proposed framework increases the percentage of detected joints and successfully estimates 3D human poses in real-time.

Vision-Based Driver Monitoring Technology Trend for Takeover in Autonomous Vehicles (자율주행자동차에서의 제어권전환을 위한 영상 기반 운전자 모니터링 기술 동향)

  • Lee, Dong-Hwan;Kim, Kyong-Ho;Kim, Do-Hyun
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2020.11a
    • /
    • pp.1090-1093
    • /
    • 2020
  • 운전자가 아닌 자율주행 시스템이 운전을 주도하기 위한 기술의 상용화를 위해 많은 기업이 노력 중이다. 특히 운전자의 안전을 보장하기 위한 운전자와 자율주행 시스템 간의 제어권전환이 중요하다. 운전자의 주행과 관련 없는 행동은 제어권전환 상황에서 운전자를 위험에 빠뜨릴 수 있으므로 제어권전환을 돕기 위한 운전자 모니터링 기술에 관한 많은 연구가 진행되고 있다. 운전자 모니터링 기술은 주로 생체 정보, 차량 정보, 영상을 사용하여 운전자의 상태와 부주의 행동 등을 감지하는 기술이다. 최근 머신 러닝, 딥 러닝을 사용한 영상처리 및 인식 기술 등의 발전으로 영상을 사용한 운전자 모니터링 기술이 활발하게 연구되고 있다. 따라서 본 논문에서는 영상기반 운전자 모니터링 기술 동향에 대해 상세히 기술하였다. 특히 운전자의 부주의 행동 중 졸음은 운전자가 주행 상황을 전혀 인지하지 못하게 할 수 있어 더욱 위험한 행동이다. 따라서 영상기반 운전자 모니터링 기술을 졸음 인식과 그 외의 행동 인식으로 분류하여 동향을 정리하였다.

Recognition of Occupants' Cold Discomfort-Related Actions for Energy-Efficient Buildings

  • Song, Kwonsik;Kang, Kyubyung;Min, Byung-Cheol
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.426-432
    • /
    • 2022
  • HVAC systems play a critical role in reducing energy consumption in buildings. Integrating occupants' thermal comfort evaluation into HVAC control strategies is believed to reduce building energy consumption while minimizing their thermal discomfort. Advanced technologies, such as visual sensors and deep learning, enable the recognition of occupants' discomfort-related actions, thus making it possible to estimate their thermal discomfort. Unfortunately, it remains unclear how accurate a deep learning-based classifier is to recognize occupants' discomfort-related actions in a working environment. Therefore, this research evaluates the classification performance of occupants' discomfort-related actions while sitting at a computer desk. To achieve this objective, this study collected RGB video data on nine college students' cold discomfort-related actions and then trained a deep learning-based classifier using the collected data. The classification results are threefold. First, the trained classifier has an average accuracy of 93.9% for classifying six cold discomfort-related actions. Second, each discomfort-related action is recognized with more than 85% accuracy. Third, classification errors are mostly observed among similar discomfort-related actions. These results indicate that using human action data will enable facility managers to estimate occupants' thermal discomfort and, in turn, adjust the operational settings of HVAC systems to improve the energy efficiency of buildings in conjunction with their thermal comfort levels.

  • PDF