• Title/Summary/Keyword: user gesture areas

Search Result 12, Processing Time 0.028 seconds

Investigating Smart TV Gesture Interaction Based on Gesture Types and Styles

  • Ahn, Junyoung;Kim, Kyungdoh
    • Journal of the Ergonomics Society of Korea
    • /
    • v.36 no.2
    • /
    • pp.109-121
    • /
    • 2017
  • Objective: This study aims to find suitable types and styles for gesture interaction as remote control on smart TVs. Background: Smart TV is being developed rapidly in the world, and gesture interaction has a wide range of research areas, especially based on vision techniques. However, most studies are focused on the gesture recognition technology. Also, not many previous studies of gestures types and styles on smart TVs were carried out. Therefore, it is necessary to check what users prefer in terms of gesture types and styles for each operation command. Method: We conducted an experiment to extract the target user manipulation commands required for smart TVs and select the corresponding gestures. To do this, we looked at gesture styles people use for every operation command, and checked whether there are any gesture styles they prefer over others. Through these results, this study was carried out with a process selecting smart TV operation commands and gestures. Results: Eighteen TV commands have been used in this study. With agreement level as a basis, we compared the six types of gestures and five styles of gestures for each command. As for gesture type, participants generally preferred a gesture of Path-Moving type. In the case of Pan and Scroll commands, the highest agreement level (1.00) of 18 commands was shown. As for gesture styles, the participants preferred a manipulative style in 11 commands (Next, Previous, Volume up, Volume down, Play, Stop, Zoom in, Zoom out, Pan, Rotate, Scroll). Conclusion: By conducting an analysis on user-preferred gestures, nine gesture commands are proposed for gesture control on smart TVs. Most participants preferred Path-Moving type and Manipulative style gestures based on the actual operations. Application: The results can be applied to a more advanced form of the gestures in the 3D environment, such as a study on VR. The method used in this study will be utilized in various domains.

A Study on Implementing Kinect-Based Control for LCD Display Contents (LCD Display 설비 Contents의 Kinect기반 동작제어 기술 구현에 관한 연구)

  • Rho, Jungkyu
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.63 no.4
    • /
    • pp.565-569
    • /
    • 2014
  • Recently, various kinds of new computer controlled devices have been introduced in a wide range of areas, and convenient user interfaces for controlling the devices are strongly needed. To implement natural user interfaces(NUIs) on top of the devices, new technologies like a touch screen, Wii Remote, wearable interfaces, and Microsoft Kinect were presented. This paper presents a natural and intuitive gesture-based model for controlling contents of LCD display. Microsoft Kinect sensor and its SDK are used to recognize human gestures, and the gestures are interpreted into corresponding commands to be executed. A command dispatch model is also proposed in order to handle the commands more naturally. I expect the proposed interface can be used in various fields, including display contents control.

Gesture Recognition Method using Tree Classification and Multiclass SVM (다중 클래스 SVM과 트리 분류를 이용한 제스처 인식 방법)

  • Oh, Juhee;Kim, Taehyub;Hong, Hyunki
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.6
    • /
    • pp.238-245
    • /
    • 2013
  • Gesture recognition has been widely one of the research areas for natural user interface. This paper presents a novel gesture recognition method using tree classification and multiclass SVM(Support Vector Machine). In the learning step, 3D trajectory of human gesture obtained by a Kinect sensor is classified into the tree nodes according to their distributions. The gestures are resampled and we obtain the histogram of the chain code from the normalized data. Then multiclass SVM is applied to the classified gestures in the node. The input gesture classified using the constructed tree is recognized with multiclass SVM.

User Gestures as a Voluntary Action in Products Design - Focused on a Gesture Discovered in User Positive Action to Transform Products (제품디자인에 있어서 자발적 행위로의 유저제스처 -사용자의 긍정적 제품변형행위에 관한 제스처를 중심으로-)

  • 진선태;우흥룡
    • Archives of design research
    • /
    • v.17 no.2
    • /
    • pp.95-104
    • /
    • 2004
  • Creativity is a important keyword for users as well as for main design organization who needs it. But little attention has been given to the aspect of user's creativity, also there has been a few attempt to apply it into design development until now. Nowadays in design areas, user's experiences and actions are changing the passive states receiving meanings into the active states creating meanings voluntarily. It is resonable to suppose that creative stage is important for users and they have the possibility of new ideas of uses and creating new productions. User's experiences of objects includes that of being formed or supported previously and that of voluntary interpretations acquired for himself, which it may be the possibilities predicted in design process or unknown user's action areas. It is likely that creative use process by themselves are the actions applied and deviated from usability and function by main design organization, also creative productions are arranged and made by users. These have a scope of examination and research in probability that is occurs frequently in user. In this research, approaching with a term, 'User gestures', User gestures are the characteristic action areas based on user's voluntary behaviors, where are revealed a unessential and non-operational function as a action itself and various transformation and creation of products as a outcome of action. This fact proves clearly that user gestures have a worth of alive spectrum to observe aspects of user culture and could be a attractive approach to seek easily new design concept for designer and developer. A further direction of this study will be following areas, Ethnography methods research of user gestures, Cultural research to phenomenon of user design and UGSBD(User gesture scenario based design) research. And it seems probable that they are applied in design development as follows, User initiative customization products, User participatory recycling products and creativity-experience design.

  • PDF

Usability Test Guidelines for Speech-Oriented Multimodal User Interface (음성기반 멀티모달 사용자 인터페이스의 사용성 평가 방법론)

  • Hong, Ki-Hyung
    • MALSORI
    • /
    • no.67
    • /
    • pp.103-120
    • /
    • 2008
  • Basic components for multimodal interface, such as speech recognition, speech synthesis, gesture recognition, and multimodal fusion, have their own technological limitations. For example, the accuracy of speech recognition decreases for large vocabulary and in noisy environments. In spite of those technological limitations, there are lots of applications in which speech-oriented multimodal user interfaces are very helpful to users. However, in order to expand application areas for speech-oriented multimodal interfaces, we have to develop the interfaces focused on usability. In this paper, we introduce usability and user-centered design methodology in general. There has been much work for evaluating spoken dialogue systems. We give a summary for PARADISE (PARAdigm for Dialogue System Evaluation) and PROMISE (PROcedure for Multimodal Interactive System Evaluation) that are the generalized evaluation frameworks for voice and multimodal user interfaces. Then, we present usability components for speech-oriented multimodal user interfaces and usability testing guidelines that can be used in a user-centered multimodal interface design process.

  • PDF

A Study on User Interface for Quiz Game Contents using Gesture Recognition (제스처인식을 이용한 퀴즈게임 콘텐츠의 사용자 인터페이스에 대한 연구)

  • Ahn, Jung-Ho
    • Journal of Digital Contents Society
    • /
    • v.13 no.1
    • /
    • pp.91-99
    • /
    • 2012
  • In this paper we introduce a quiz application program that digitizes the analogue quiz game. We digitize the quiz components such as quiz proceeding, participants recognition, problem presentation, volunteer recognition who raises his hand first, answer judgement, score addition, winner decision, etc, which are manually performed in the normal quiz game. For automation, we obtained the depth images from the kinect camera which comes into the spotlight recently, so that we located the quiz participants and recognized the user-friendly defined gestures. Analyzing the depth distribution, we detected and segmented the upper body parts and located the hands' areas. Also, we extracted hand features and designed the decision function that classified the hand pose into palm, fist or else, so that a participant can select the example that he wants among presented examples. The implemented quiz application program was tested in real time and showed very satisfactory gesture recognition results.

Analysis of Users' Gestures by Application in Smartphone Touch Interfaces (스마트폰 터치 인터페이스에서 애플리케이션별 사용자 제스처의 분석)

  • Kim, Jisun;Bahn, Hyokyung
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.15 no.2
    • /
    • pp.9-14
    • /
    • 2015
  • Touch interface is widely used in a smartphone instead of a keyboard or a keypad interface that has been adopted in a PC or a featurephone, respectively. Touch interface can recognize a variety of gestures that clearly represent the distinct features of each application's input. This paper analyzes users' gesture of each application captured by the touch interface of a smartphone. Specifically, we extract touch input traces from various application categories such as game, web browser, youtube, image and e-book viewer, video player, camera, and map applications, and then analyzed them. Through this analysis, we observed a certain unique characteristics of each application's touch input, and this can be utilized in various useful areas such as identification of an application user, prevention of running an application by an illegal user, or design of a new interface convenient to a specific user.

A Deep Learning-based Hand Gesture Recognition Robust to External Environments (외부 환경에 강인한 딥러닝 기반 손 제스처 인식)

  • Oh, Dong-Han;Lee, Byeong-Hee;Kim, Tae-Young
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.14 no.5
    • /
    • pp.31-39
    • /
    • 2018
  • Recently, there has been active studies to provide a user-friendly interface in a virtual reality environment by recognizing user hand gestures based on deep learning. However, most studies use separate sensors to obtain hand information or go through pre-process for efficient learning. It also fails to take into account changes in the external environment, such as changes in lighting or some of its hands being obscured. This paper proposes a hand gesture recognition method based on deep learning that is strong in external environments without the need for pre-process of RGB images obtained from general webcam. In this paper we improve the VGGNet and the GoogLeNet structures and compared the performance of each structure. The VGGNet and the GoogLeNet structures presented in this paper showed a recognition rate of 93.88% and 93.75%, respectively, based on data containing dim, partially obscured, or partially out-of-sight hand images. In terms of memory and speed, the GoogLeNet used about 3 times less memory than the VGGNet, and its processing speed was 10 times better. The results of this paper can be processed in real-time and used as a hand gesture interface in various areas such as games, education, and medical services in a virtual reality environment.

Analyzing Input Patterns of Smartphone Applications in Touch Interfaces

  • Bahn, Hyokyung;Kim, Jisun
    • International journal of advanced smart convergence
    • /
    • v.10 no.4
    • /
    • pp.30-37
    • /
    • 2021
  • Touch sensor interface has become the most useful input device in a smartphone. Unlike keypad/keyboard interfaces used in electronic dictionaries and feature phones, smartphone's touch interfaces allow for the recognition of various gestures that represent distinct features of each application's input. In this paper, we analyze application-specific input patterns that appear in smartphone's touch interfaces. Specifically, we capture touch input patterns from various Android applications, and analyze them. Based on this analysis, we observe a certain unique characteristics of application's touch input patterns. This can be utilized in various useful areas like user authentications, prevention of executing application by illegal users, or digital forensic based on logged touch patterns.

Designing Effective Virtual Training: A Case Study in Maritime Safety

  • Jung, Jinki;Kim, Hongtae
    • Journal of the Ergonomics Society of Korea
    • /
    • v.36 no.5
    • /
    • pp.385-394
    • /
    • 2017
  • Objective: The aim of this study is to investigate how to design effective virtual reality-based training (i.e., virtual training) in maritime safety and to present methods for enhancing interface fidelity by employing immersive interaction and 3D user interface (UI) design. Background: Emerging virtual reality technologies and hardware enable to provide immersive experiences to individuals. There is also a theory that the improvement of fidelity can improve the training efficiency. Such a sense of immersion can be utilized as an element for realizing effective training in the virtual space. Method: As an immersive interaction, we implemented gesture-based interaction using leap motion and Myo armband type sensors. Hand gestures captured from both sensors are used to interact with the virtual appliance in the scenario. The proposed 3D UI design is employed to visualize appropriate information for tasks in training. Results: A usability study to evaluate the effectiveness of the proposed method has been carried out. As a result, the usability test of satisfaction, intuitiveness of UI, ease of procedure learning, and equipment understanding showed that virtual training-based exercise was superior to existing training. These improvements were also independent of the type of input devices for virtual training. Conclusion: We have shown through experiments that the proposed interaction design results are more efficient interactions than the existing training method. The improvement of interface fidelity through intuitive and immediate feedback on the input device and the training information improve user satisfaction with the system, as well as training efficiency. Application: Design methods for an effective virtual training system can be applied to other areas by which trainees are required to do sophisticated job with their hands.