• 제목/요약/키워드: Learning System for the Blind

검색결과 42건 처리시간 0.026초

Intelligent Shoes for Detecting Blind Falls Using the Internet of Things

  • Ahmad Abusukhon
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제17권9호
    • /
    • pp.2377-2398
    • /
    • 2023
  • In our daily lives, we engage in a variety of tasks that rely on our senses, such as seeing. Blindness is the absence of the sense of vision. According to the World Health Organization, 2.2 billion people worldwide suffer from various forms of vision impairment. Unfortunately, blind people face a variety of indoor and outdoor challenges on a daily basis, limiting their mobility and preventing them from engaging in other activities. Blind people are very vulnerable to a variety of hazards, including falls. Various barriers, such as stairs, can cause a fall. The Internet of Things (IoT) is used to track falls and send a warning message to the blind caretakers. One of the gaps in the previous works is that they were unable to differentiate between falls true and false. Treating false falls as true falls results in many false alarms being sent to the blind caretakers and thus, they may reject the IoT system. As a means of bridging this chasm, this paper proposes an intelligent shoe that is able to precisely distinguish between false and true falls based on three sensors, namely, the load scale sensor, the light sensor, and the Flex sensor. The proposed IoT system is tested in an indoor environment for various scenarios of falls using four models of machine learning. The results from our system showed an accuracy of 0.96%. Compared to the state-of-the-art, our system is simpler and more accurate since it avoids sending false alarms to the blind caretakers.

시각장애인을 위한 점자 교육 시스템 (OnDot: Braille Training System for the Blind)

  • 김학진;문준혁;송민욱;이세민;공기석
    • 한국인터넷방송통신학회논문지
    • /
    • 제20권6호
    • /
    • pp.41-50
    • /
    • 2020
  • 본 논문에서는 기존의 점자 학습 제품의 단점들을 보완한 점자 교육 시스템을 다룬다. 시각장애인 전용 어플리케이션은 사용자 편의성을 위해 터치 제스처 및 음성 안내를 통하여 전체 기능을 수행할 수 있도록 구성한다. 점자키트는 아두이노와 3D 프린팅을 통해 교육 목적에 맞게 제작한다. 시스템은 다음과 같은 기능들을 지원한다. 첫 째, 초성·종성·모음·약어 등 기초적인 점자의 학습. 둘 째, 단계별 퀴즈를 풀어 학습한 점자를 확인하는 기능. 셋 째, 모르는 점자가 있을 때 번역하는 기능이다. 실험을 통한 터치 제스처의 인식률과 점자 표현의 정확도를 확인하였고 번역의 경우 의도한대로 번역이 되는 것을 확인하였다. 이 시스템을 통해 시각장애인이 효율적으로 점자를 학습할 수 있다.

시각장애인을 위한 음성안내 네비게이션 시스템의 심층신경망 성능 비교 (Comparison of Deep Learning Networks in Voice-Guided System for The Blind)

  • 안륜희;엄성호;유윤섭
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2022년도 추계학술대회
    • /
    • pp.175-177
    • /
    • 2022
  • 본 논문은 시각장애인이 원하는 목적지까지 버스를 이용하여 이동하는 것을 보조하는 시스템 안에 사용될 수 있는 심층신경망에 따라 성능을 비교하였다. 이 시스템은 GPS, 경로 안내 API를 이용하여 현재 위치에서 목적지까지 이르는 경로를 찾고 안내하는 스마트폰 어플리케이션과 임베디드 보드, 심층신경망, 버스정보 API를 이용해 원하는 버스의 탑승 대기시간을 구하고 도착을 감지하는 정거장 설치형 모듈로 이루어져 있다. 정거장 설치형 모듈에 탑승할 버스 번호를 인식하기 위해서 faster-RCNN, YOLOv4, YOLOv5s 세 가지 심층신경망을 적용했고 최상 정확도와 속도면에서 YOLOv5s 심층신경망이 가장 좋은 성능을 보였다.

  • PDF

사각지대를 고려한 이동로봇의 인공표식기반 위치추정시스템 (Landmark based Localization System of Mobile Robots Considering Blind Spots)

  • 허동혁;박태형
    • 로봇학회논문지
    • /
    • 제6권2호
    • /
    • pp.156-164
    • /
    • 2011
  • This paper propose a localization system of indoor mobile robots. The localization system includes camera and artificial landmarks for global positioning, and encoders and gyro sensors for local positioning. The Kalman filter is applied to take into account the stochastic errors of all sensors. Also we develop a dead reckoning system to estimate the global position when the robot moves the blind spots where it cannot see artificial landmarks, The learning engine using modular networks is designed to improve the performance of the dead reckoning system. Experimental results are then presented to verify the usefulness of the proposed localization system.

시각장애인을 위한 딥러닝 기반 표지판 검출 및 인식 (Deep Learning Based Sign Detection and Recognition for the Blind)

  • 전태재;이상윤
    • 전자공학회논문지
    • /
    • 제54권2호
    • /
    • pp.115-122
    • /
    • 2017
  • 본 논문은 딥러닝 알고리즘을 기반으로 하여 시각장애인을 위한 표지판을 검출하고 인식하는 시스템을 제안한다. 제안된 시스템은 크게 표지판 검출 단계와 표지판 인식 단계로 나눠지는데 표지판 검출 단계에서는 영상에서 응집 채널 특징을 추출한 뒤 아다부스트 분류기를 적용하여 표지판 관심영역을 검출하였고, 표지판 인식 단계에서는 검출한 표지판 관심영역들에 합성곱 신경망을 적용하여 어떤 표지판인지 인식하였다. 본 논문에서는 미검출된 표지판의 개수가 최대한 감소하도록 아다부스트 분류기를 설계하였고, 딥러닝 알고리즘을 사용하여 인식 정확도를 높임으로써 검출 단계에서 발생한 양성 오류들을 제거시켰다. 실험 결과, 제안된 방법의 양성 오류 개수가 다른 방법들의 양성 오류 개수보다 효과적으로 감소했음을 확인하였다.

시각장애인의 학습을 위한 텍스트 추출 및 점자 변환 시스템 (HunMinJeomUm: Text Extraction and Braille Conversion System for the Learning of the Blind)

  • 김채리;김지안;김용민;이예지;공기석
    • 한국인터넷방송통신학회논문지
    • /
    • 제21권5호
    • /
    • pp.53-60
    • /
    • 2021
  • 시각장애인의 수는 증가하고 있지만 시각장애인을 위한 점역 교재는 부족하여 본인의 의지에 관계 없이 교육권을 침해받는 경우가 많다. 본 논문에서는 시각장애인의 교육권을 보장하기 위해 점자책으로 나오지 않는 교재나 문서, 사진 등을 보호자의 도움 없이도 혼자 쉽게 공부할 수 있게끔 도와주는 학습 시스템을 다룬다. 장애인 접근성을 고려하여 어플리케이션과 웹페이지를 설계하고 점자 키트는 아두이노와 점자 모듈을 이용하여 제작한다. 이 시스템은 다음과 같은 기능들을 지원한다. 첫째, 원하는 문서 또는 사진을 선택해 OCR을 이용하여 텍스트를 추출한다. 둘째, 추출한 텍스트를 음성과 점자로 변환한다. 셋째, 회원가입 기능을 제공하여 추출된 텍스트를 다시 볼 수 있도록 한다. 다양한 실험을 통해 점자 출력, 음성 출력이 정상적으로 작동하는 것을 확인하고 높은 OCR 인식률을 제공하는 것을 알 수 있었다. 또한, 시각이 완전히 차단된 상태에서도 어플리케이션이 손쉽게 이용 가능하다는 것을 확인했다.

Interface Modeling for Digital Device Control According to Disability Type in Web

  • Park, Joo Hyun;Lee, Jongwoo;Lim, Soon-Bum
    • Journal of Multimedia Information System
    • /
    • 제7권4호
    • /
    • pp.249-256
    • /
    • 2020
  • Learning methods using various assistive and smart devices have been developed to enable independent learning of the disabled. Pointer control is the most important consideration for the disabled when controlling a device and the contents of an existing graphical user interface (GUI) environment; however, difficulties can be encountered when using a pointer, depending on the disability type; Although there are individual differences depending on the blind, low vision, and upper limb disability, problems arise in the accuracy of object selection and execution in common. A multimodal interface pilot solution is presented that enables people with various disability types to control web interactions more easily. First, we classify web interaction types using digital devices and derive essential web interactions among them. Second, to solve problems that occur when performing web interactions considering the disability type, the necessary technology according to the characteristics of each disability type is presented. Finally, a pilot solution for the multimodal interface for each disability type is proposed. We identified three disability types and developed solutions for each type. We developed a remote-control operation voice interface for blind people and a voice output interface applying the selective focusing technique for low-vision people. Finally, we developed a gaze-tracking and voice-command interface for GUI operations for people with upper-limb disability.

Visual Analysis of Deep Q-network

  • Seng, Dewen;Zhang, Jiaming;Shi, Xiaoying
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제15권3호
    • /
    • pp.853-873
    • /
    • 2021
  • In recent years, deep reinforcement learning (DRL) models are enjoying great interest as their success in a variety of challenging tasks. Deep Q-Network (DQN) is a widely used deep reinforcement learning model, which trains an intelligent agent that executes optimal actions while interacting with an environment. This model is well known for its ability to surpass skilled human players across many Atari 2600 games. Although DQN has achieved excellent performance in practice, there lacks a clear understanding of why the model works. In this paper, we present a visual analytics system for understanding deep Q-network in a non-blind matter. Based on the stored data generated from the training and testing process, four coordinated views are designed to expose the internal execution mechanism of DQN from different perspectives. We report the system performance and demonstrate its effectiveness through two case studies. By using our system, users can learn the relationship between states and Q-values, the function of convolutional layers, the strategies learned by DQN and the rationality of decisions made by the agent.

딥러닝 기반의 운전자의 안전/위험 상태 인지 시스템 개발 (Development of Driver's Safety/Danger Status Cognitive Assistance System Based on Deep Learning)

  • 미아오 쉬;이현순;강보영
    • 로봇학회논문지
    • /
    • 제13권1호
    • /
    • pp.38-44
    • /
    • 2018
  • In this paper, we propose Intelligent Driver Assistance System (I-DAS) for driver safety. The proposed system recognizes safety and danger status by analyzing blind spots that the driver cannot see because of a large angle of head movement from the front. Most studies use image pre-processing such as face detection for collecting information about the driver's head movement. This not only increases the computational complexity of the system, but also decreases the accuracy of the recognition because the image processing system dose not use the entire image of the driver's upper body while seated on the driver's seat and when the head moves at a large angle from the front. The proposed system uses a convolutional neural network to replace the face detection system and uses the entire image of the driver's upper body. Therefore, high accuracy can be maintained even when the driver performs head movement at a large angle from the frontal gaze position without image pre-processing. Experimental result shows that the proposed system can accurately recognize the dangerous conditions in the blind zone during operation and performs with 95% accuracy of recognition for five drivers.

Blind Image Quality Assessment on Gaussian Blur Images

  • Wang, Liping;Wang, Chengyou;Zhou, Xiao
    • Journal of Information Processing Systems
    • /
    • 제13권3호
    • /
    • pp.448-463
    • /
    • 2017
  • Multimedia is a ubiquitous and indispensable part of our daily life and learning such as audio, image, and video. Objective and subjective quality evaluations play an important role in various multimedia applications. Blind image quality assessment (BIQA) is used to indicate the perceptual quality of a distorted image, while its reference image is not considered and used. Blur is one of the common image distortions. In this paper, we propose a novel BIQA index for Gaussian blur distortion based on the fact that images with different blur degree will have different changes through the same blur. We describe this discrimination from three aspects: color, edge, and structure. For color, we adopt color histogram; for edge, we use edge intensity map, and saliency map is used as the weighting function to be consistent with human visual system (HVS); for structure, we use structure tensor and structural similarity (SSIM) index. Numerous experiments based on four benchmark databases show that our proposed index is highly consistent with the subjective quality assessment.