• Title/Summary/Keyword: visual-audio

Search Result 424, Processing Time 0.03 seconds

MPEG2 Realtime Multimedia Communication on Broadband Network

  • Kawamura, Yoshifumi;Kasahara, Takao;Komatsu, Shigeru
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 1996.06b
    • /
    • pp.67-71
    • /
    • 1996
  • We have developed an MPEG2 realtime codec system and ATM interface, and evaluated the performance and characteristics of the audio-visual communication in terms of Interoperability specification of The ATM Forum, ITU-T H.310 draft recommendation [1] and the reference points of DAVIC specification through the experiments using the experimental public ATM network.

  • PDF

A Web-based 3D Virtual Reality Pavilion of Korean Traditional Music (웹 기반의 가상현실 3D 국악 박물관 제작)

  • Choi, Ji Ae;Shim, Jae Sun;Kim, Yoon Sang
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.4 no.1
    • /
    • pp.65-68
    • /
    • 2008
  • In this paper, a web-based 3D virtual reality (VR) pavillion of Korean Traditional Music was implemented. The VR pavillion is used for the virtual demonstration and experience of Korean Traditional Music, which provides the information as well as multimedia experience on eight instruments to users through internet. It provides eight web-pages and one an audio-visual classroom on the instruments.

Interactive information process image with minute hand gestures

  • Lim, Chan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2016.04a
    • /
    • pp.799-802
    • /
    • 2016
  • It is definitely an interesting job to work with V4 to create various contents emphasizing different interfaces like 3D graphics, and multimedia such as video, audio, and camera. Moreover, beyond the other interface, as it could be used in the many aspects of the sensory sign such as visual effects, auditory effects, and touchable effects, it feels free to make a better developed model. We intended the users to feel some kind of pleasure and interactions rather than just using in aspect of Media art.

Some effects of audio-visual speech in perceiving Korean

  • Kim, Jee-Sun;Davis, Chris
    • Annual Conference on Human and Language Technology
    • /
    • 1999.10e
    • /
    • pp.335-342
    • /
    • 1999
  • The experiments reported here investigated whether seeing a speaker's face (visible speech) affects the perception and memory of Korean speech sounds. In order to exclude the possibility of top-down, knowledge-based influences on perception and memory, the experiments tested people with no knowledge of Korean. The first experiment examined whether visible speech (Auditory and Visual - AV) assists English native speakers (with no knowledge of Korean) in the detection of a syllable within a Korean speech phrase. It was found that a syllable was more likely to be detected within a phrase when the participants could see the speaker's face. The second experiment investigated whether English native speakers' judgments about the duration of a Korean phrase would be affected by visible speech. It was found that in the AV condition participant's estimates of phrase duration were highly correlated with the actual durations whereas those in the AO condition were not. The results are discussed with respect to the benefits of communication with multimodal information and future applications.

  • PDF

MultiFormat motion picture storage subsystem using DirectShow Filters for a Mutichannel Visual Monitoring System (다채널 영상 감시 시스템을 위한 다중 포맷 동영상 저장 DirectShow Filter설계 및 구현)

  • 정연권;하상석;정선태
    • Proceedings of the IEEK Conference
    • /
    • 2002.06d
    • /
    • pp.113-116
    • /
    • 2002
  • Windows provides Directshow for efficient multimedia streaming processings such as multimedia capture, storage, display and etc. Presently, many motion picture codecs and audio codecs are made to be used in Directshow framework and Windows also supports many codecs (MPEG4, H,263, WMV, WMA, ASF, etc.) in addition to a lot of useful tools for multimedia streaming processing. Therefore, Directshow can be effectively utilized for developing windows-based multimedia streaming applications such as visual monitoring systems which needs to store real-time video data for later retrieval. In this paper, we present our efforts for developing a Directshow Filter System supporting storage of motion pictures in various motion picture codecs. Our Directshow Filter system also provides an additional functionality of motion detection.

  • PDF

Lip Feature Extraction using Contrast of YCbCr (YCbCr 농도 대비를 이용한 입술특징 추출)

  • Kim, Woo-Sung;Min, Kyung-Won;Ko, Han-Seok
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.259-260
    • /
    • 2006
  • Since audio speech recognition is affected by noise in real environment, visual speech recognition is used to support speech recognition. For the visual speech recognition, this paper suggests the extraction of lip-feature using two types of image segmentation and reduced ASM. Input images are transformed to YCbCr based images and lips are segmented using the contrast of Y/Cb/Cr between lip and face. Subsequently, lip-shape model trained by PCA is placed on segmented lip region and then lip features are extracted using ASM.

  • PDF

Design of the Visual Telephone System Based on H.320 (H.320 기반의 영상전화 시스템 설계연구)

  • 정하재;박배욱
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2002.11a
    • /
    • pp.821-825
    • /
    • 2002
  • In the paper, we described a new visual telephone system to solve the unsatisfactory items of existing systems based on H.320. The system has a differential task interrupt transfer feature for real time video phone service. Owing to the result of interrupt transfer of different speed according to the time critical degree of tasks, the flow of audio and video data stream can be kept as constant speed.

  • PDF

Digital Color Image Watermarking for HVS(Human Visual System) using Daubechies wavelet

  • Park, Jong-Tae;Rhee, Kang-Hyeon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.8 no.7
    • /
    • pp.1488-1492
    • /
    • 2004
  • The digital signal has been replaced the analog signal in most of every field of multimedia including still image, animation, and audio due to the enormous extension of computer supply and the fast development of computer network. The consumers of information are able to enjoy the abundance of information because of one of the digital signal traits that very easy to regenerate the original data. Because of the trait, however, it is very hard for the producers of information to keep the copyright with the merit of original copy in quality excellency. In this paper, the watermarking technology which inserts a RGB color watermark in color image using the visual characteristics of wavelet coefficient was proposed. As a result, the PSNR value of image was varied depending on perceptual parameter, but we can obtain 32dB as a whole.

Design and Implementation of Scent-Supported Educational Content using Arduino

  • Hye-kyung Kwon;Heesun Kim
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.260-267
    • /
    • 2023
  • Due to the development of science and technology in the 4th Industrial Revolution, a variety of content is being developed and utilized through educational courses linked to digital textbooks. Students use smart devices to engage in realistic virtual learning experiences, interacting with the content in digital textbooks. However, while many realistic contents offer visual and auditory effects like 3D VR, AR, and holograms, olfactory content that evokes actual sensations has not yet been introduced. Therefore, in this paper, we designed and implemented 4D educational content by adding the sense of smell to existing content. This implemented content was tested in classrooms through a curriculum-based evaluation. Classes taught with olfactory-enhanced content showed a higher percentage of correct answers compared to those using traditional audio-visual materials, indicating improved understanding.

Data Visualization of Site-Specific Underground Sounds

  • Tae-Eun, Kim
    • International journal of advanced smart convergence
    • /
    • v.13 no.1
    • /
    • pp.77-84
    • /
    • 2024
  • This study delves into the subtle sounds emanating from beneath the earth's surface to unveil hidden messages and the movements of life. It transforms these acoustic phenomena into digital data and reimagines them as visual elements. By employing Sismophone microphones and utilizing the FFT function in p5.js, it analyzes the intricate frequency components of subterranean sounds and translates them into various visual elements, including 3D geometric shapes, flowing lines, and moving particles. This project is grounded in the sounds recorded in diverse 'spaces of death,' ranging from the tombs of Joseon Dynasty officials to abandoned areas in modern cities. We leverage the power of sound to transcend space and time, conveying the concealed narratives and messages of forgotten places .Through the visualization of these sounds, this research blurs the boundaries between 'death' and 'life,' 'past' and 'present,' aiming to explore new forms of artistic expression and broaden perceptions through the sensory connection between sound and vision.