• Title/Summary/Keyword: Caption

Search Result 167, Processing Time 0.033 seconds

Development of Instructional Activity Support Model for Capstone Design to Creative Engineering Education (창의적 공학교육을 위한 캡스톤 디자인(Capstone Design) 교수활동지원모형 개발)

  • Park, Su-Hong;Jung, Ju-Young;Ryu, Young-Ho;KIM, Seong-Ok
    • Journal of Fisheries and Marine Sciences Education
    • /
    • v.20 no.2
    • /
    • pp.184-200
    • /
    • 2008
  • The purpose of this paper is to develop instructional activity support model for capstone design in order for improving creative engineering education. To do this, having extracted the core idea of capstone design, and elicited core learning activity process, and grasped core supportive factors according to each core learning activity process that elicited, an improved instructional design model for capstone design was then developed through formative evaluation with respect to the draft of the instructional system development model for capstone design. As to major research methods, case analysis, requirements analysis through interview, and formative evaluation by experts were employed, and then research studies were undertaken. The formative evaluation by experts was carried out for two hours in 2007, and the experts participated in the evaluation consisted of total 6 persons: two specialists of capstone design contents, two professionals in field works, and two expert instructional designers in education engineering. Interview results had been reflected in this research when developing final instructional design model for capstone design. The core learning activity process of the final instructional design model for caption design, which developed in this research, comprises following stages: (1) Team building $\rightarrow$ (2) Integrated meeting between industry and academy $\rightarrow$ (3) Analysis of tasks $\rightarrow$ (4) Clarification of tasks $\rightarrow$(5) Seeking solutions for issues $\rightarrow$ (6) Eliciting priority of solutions $\rightarrow$ (7) Designing solutions and construction $\rightarrow$ (8) Exhibiting outcomes and presentation $\rightarrow$(9) Gaining comprehensive insights Also, in the core learning activity process, supportive factors that support implementation of each step were presented having been categorized into facilitator (teacher, and professionals in field works), learner and tool, etc.

Video Copy Detection Algorithm Against Online Piracy of DTV Broadcast Program (DTV 방송프로그램의 온라인 불법전송 차단을 위한 비디오 복사본 검출 알고리즘)

  • Kim, Joo-Sub;Nam, Je-Ho
    • Journal of Broadcast Engineering
    • /
    • v.13 no.5
    • /
    • pp.662-676
    • /
    • 2008
  • This paper presents a video copy detection algorithm that blocks online transfer of illegally copied DTV broadcast programs. Particularly, the proposed algorithm establishes a set of keyframes by detecting abrupt changes of luminance, and then exploits the spatio-temporal features of keyframes. Comparing with the preregistered features stored in the database of DTV broadcast programs, the proposed scheme performs a function of video filtering in order to distinguish whether an uploaded video is illegally copied or not. Note that we analyze only a set of keyframes instead of an entire video frame. Thus, it is highly efficient to identify illegal copied video when we deal with a vast size of broadcast programs. Also, we confirm that the proposed technique is robust to a variety of video edit-effects that are often applied by online video redistribution, such as apsect-ratio change, logo insertion, caption insertion, visual quality degradation, and resolution change (downscaling). In addition, we perform a benchmark test in which the proposed scheme outperforms previous techniques.

Using Staganography method for Protection of Caption Information in the Motion Picture (스테가노그래피 기법을 이용한 동영상 자막정보의 저작권 보호)

  • Ham, Jun-Sick;Ryou, Hwang-Bin;Hong, Seong-Sik
    • The KIPS Transactions:PartC
    • /
    • v.10C no.7
    • /
    • pp.835-842
    • /
    • 2003
  • The technical advancement in and the widespread expansion of use of the internet has recently led to an increase in high bandwidth multimedia data usage. This is largely due to an unfettered access to the internet. The intrinsic merit of the internet, however, brings about a side effect:multimedia date can easily be duplicated recklessly. This invokes the problem of intellectual property right violation. Focusing on this matter, multi-faceted studies are being conducted now. This paper aims to provide, by applying steganography and encryption methods into the SMIL files which supply subtitle data, the intellectual property protection of both motion picture and its subtitle data, as well as the integrity of the contents in question. In order to achieve the above goal, the authors show a particular method how to safely embed subtitle data into the given frames, after they separate subtitle data to encrypt them. By doing this, the authors suggests, the intellectual property right to the subtitle data as well as their integrity can be protected and achieved respectively. Besides, an integrity service for motion picture can be made.

The Current Situation of the Digital Interface International Standards and an Analysis of Integration Condition of Ships (선박의 디지털 인터페이스 표준화 현황 및 연동조건 분석)

  • Park, Jong-Won;Lim, Yong-Kon;Yun, Chang-Ho;Kim, Ok-Soo;Lee, Jung-Woo;Chung, Han-Na
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.48 no.6
    • /
    • pp.490-500
    • /
    • 2011
  • The paradigm that ships are navigated by means of analog navigation devices and caption's experiences is shifted to that ships can be navigated by digital and autonomous navigation systems due to the rapid development of electronics and communication technologies. Hence, all systems deployed in a ship are automated and integrated since corresponding ship systems are not individually operated but integrated and managed under common operating system. In this paper, we overview recent research projects targeted for digitally integrated ships (referred to as digital ships), outline IEC 61162 series, specifying digital interface standards for digital ships, and finally present several issues for further improvement of the IEC 61162 series. In addition, we broadly categorize digital ships into three systems consisting of multiple sub-systems, including a navigation system, a power system, and a automation system in the aspect of operation and function. Corresponding interfacing range and method, as well as interface specification and data types of one system in a digital ship are also described according to sub-systems, respectively.

High-Quality Multimodal Dataset Construction Methodology for ChatGPT-Based Korean Vision-Language Pre-training (ChatGPT 기반 한국어 Vision-Language Pre-training을 위한 고품질 멀티모달 데이터셋 구축 방법론)

  • Jin Seong;Seung-heon Han;Jong-hun Shin;Soo-jong Lim;Oh-woog Kwon
    • Annual Conference on Human and Language Technology
    • /
    • 2023.10a
    • /
    • pp.603-608
    • /
    • 2023
  • 본 연구는 한국어 Vision-Language Pre-training 모델 학습을 위한 대규모 시각-언어 멀티모달 데이터셋 구축에 대한 필요성을 연구한다. 현재, 한국어 시각-언어 멀티모달 데이터셋은 부족하며, 양질의 데이터 획득이 어려운 상황이다. 따라서, 본 연구에서는 기계 번역을 활용하여 외국어(영문) 시각-언어 데이터를 한국어로 번역하고 이를 기반으로 생성형 AI를 활용한 데이터셋 구축 방법론을 제안한다. 우리는 다양한 캡션 생성 방법 중, ChatGPT를 활용하여 자연스럽고 고품질의 한국어 캡션을 자동으로 생성하기 위한 새로운 방법을 제안한다. 이를 통해 기존의 기계 번역 방법보다 더 나은 캡션 품질을 보장할 수 있으며, 여러가지 번역 결과를 앙상블하여 멀티모달 데이터셋을 효과적으로 구축하는데 활용한다. 뿐만 아니라, 본 연구에서는 의미론적 유사도 기반 평가 방식인 캡션 투영 일치도(Caption Projection Consistency) 소개하고, 다양한 번역 시스템 간의 영-한 캡션 투영 성능을 비교하며 이를 평가하는 기준을 제시한다. 최종적으로, 본 연구는 ChatGPT를 이용한 한국어 멀티모달 이미지-텍스트 멀티모달 데이터셋 구축을 위한 새로운 방법론을 제시하며, 대표적인 기계 번역기들보다 우수한 영한 캡션 투영 성능을 증명한다. 이를 통해, 우리의 연구는 부족한 High-Quality 한국어 데이터 셋을 자동으로 대량 구축할 수 있는 방향을 보여주며, 이 방법을 통해 딥러닝 기반 한국어 Vision-Language Pre-training 모델의 성능 향상에 기여할 것으로 기대한다.

  • PDF

Deep Learning-based Professional Image Interpretation Using Expertise Transplant (전문성 이식을 통한 딥러닝 기반 전문 이미지 해석 방법론)

  • Kim, Taejin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.79-104
    • /
    • 2020
  • Recently, as deep learning has attracted attention, the use of deep learning is being considered as a method for solving problems in various fields. In particular, deep learning is known to have excellent performance when applied to applying unstructured data such as text, sound and images, and many studies have proven its effectiveness. Owing to the remarkable development of text and image deep learning technology, interests in image captioning technology and its application is rapidly increasing. Image captioning is a technique that automatically generates relevant captions for a given image by handling both image comprehension and text generation simultaneously. In spite of the high entry barrier of image captioning that analysts should be able to process both image and text data, image captioning has established itself as one of the key fields in the A.I. research owing to its various applicability. In addition, many researches have been conducted to improve the performance of image captioning in various aspects. Recent researches attempt to create advanced captions that can not only describe an image accurately, but also convey the information contained in the image more sophisticatedly. Despite many recent efforts to improve the performance of image captioning, it is difficult to find any researches to interpret images from the perspective of domain experts in each field not from the perspective of the general public. Even for the same image, the part of interests may differ according to the professional field of the person who has encountered the image. Moreover, the way of interpreting and expressing the image also differs according to the level of expertise. The public tends to recognize the image from a holistic and general perspective, that is, from the perspective of identifying the image's constituent objects and their relationships. On the contrary, the domain experts tend to recognize the image by focusing on some specific elements necessary to interpret the given image based on their expertise. It implies that meaningful parts of an image are mutually different depending on viewers' perspective even for the same image. So, image captioning needs to implement this phenomenon. Therefore, in this study, we propose a method to generate captions specialized in each domain for the image by utilizing the expertise of experts in the corresponding domain. Specifically, after performing pre-training on a large amount of general data, the expertise in the field is transplanted through transfer-learning with a small amount of expertise data. However, simple adaption of transfer learning using expertise data may invoke another type of problems. Simultaneous learning with captions of various characteristics may invoke so-called 'inter-observation interference' problem, which make it difficult to perform pure learning of each characteristic point of view. For learning with vast amount of data, most of this interference is self-purified and has little impact on learning results. On the contrary, in the case of fine-tuning where learning is performed on a small amount of data, the impact of such interference on learning can be relatively large. To solve this problem, therefore, we propose a novel 'Character-Independent Transfer-learning' that performs transfer learning independently for each character. In order to confirm the feasibility of the proposed methodology, we performed experiments utilizing the results of pre-training on MSCOCO dataset which is comprised of 120,000 images and about 600,000 general captions. Additionally, according to the advice of an art therapist, about 300 pairs of 'image / expertise captions' were created, and the data was used for the experiments of expertise transplantation. As a result of the experiment, it was confirmed that the caption generated according to the proposed methodology generates captions from the perspective of implanted expertise whereas the caption generated through learning on general data contains a number of contents irrelevant to expertise interpretation. In this paper, we propose a novel approach of specialized image interpretation. To achieve this goal, we present a method to use transfer learning and generate captions specialized in the specific domain. In the future, by applying the proposed methodology to expertise transplant in various fields, we expected that many researches will be actively conducted to solve the problem of lack of expertise data and to improve performance of image captioning.

A Method of Generating Table-of-Contents for Educational Video (교육용 비디오의 ToC 자동 생성 방법)

  • Lee Gwang-Gook;Kang Jung-Won;Kim Jae-Gon;Kim Whoi-Yul
    • Journal of Broadcast Engineering
    • /
    • v.11 no.1 s.30
    • /
    • pp.28-41
    • /
    • 2006
  • Due to the rapid development of multimedia appliances, the increasing amount of multimedia data enforces the development of automatic video analysis techniques. In this paper, a method of ToC generation is proposed for educational video contents. The proposed method consists of two parts: scene segmentation followed by scene annotation. First, video sequence is divided into scenes by the proposed scene segmentation algorithm utilizing the characteristics of educational video. Then each shot in the scene is annotated in terms of scene type, existence of enclosed caption and main speaker of the shot. The ToC generated by the proposed method represents the structure of a video by the hierarchy of scenes and shots and gives description of each scene and shot by extracted features. Hence the generated ToC can help users to perceive the content of a video at a glance and. to access a desired position of a video easily. Also, the generated ToC automatically by the system can be further edited manually for the refinement to effectively reduce the required time achieving more detailed description of the video content. The experimental result showed that the proposed method can generate ToC for educational video with high accuracy.

Immunoelectrophoretic analysis of major component proteins In cystic fluid of Taenia solium metacestodes (면역전기영동법에 의한 유구낭미충 낭액의 구성 단백질 분석)

  • Yoon Kong;Seung-Yull Cho;Suk-Il Kim;Shin-Yong Kang
    • Parasites, Hosts and Diseases
    • /
    • v.30 no.3
    • /
    • pp.209-218
    • /
    • 1992
  • When cystic fluid of Taenia solium metacestodes (CF) was filtrated through Sephacryl S-300 Superfine, major proteins were in fractions III add IV Major protein in fraction III was Band C protein of 150 kDa and that in fraction IV was Band N protein (Choi et of., 1990). When CF was electrophoresed in 0.9% agarose gel and reacted with anti-CF rabbit serum (RACF), two main bands, a long outer and a short inner band, were precipitated, together with 8 minor bands. RACF reacted with fraction III forming the long outer band whereas RACF formed the short infer band with fraction IV in immunoelectrophoresis (IEP) The long outer precipitin band of CF fraction III was similar to antigen B in hydatid fluid (HF) of Oriol et at. (1971), while the short inner band of CF fraction IV was similar to HF antigen 5 of Caption et at. (1967) . When HF was reacted with RACF, the short inner band was immunoprecipitated without forming the long outer band. Common antigenicity between CF and HF seemed to exist in fraction IV rather than in fraction III of CF. Patient sera of neurocysticercosis reacted more frequently with fraction III than with fraction IV.

  • PDF

View-oriented and Green Marketing Characteristics of Apartment Advertisements on Newspapers (신문 아파트광고에 나타난 조망지향성과 그린마케팅의 속성)

  • Rho, Jae-Hyun;Kim, Ok-Kyung
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.34 no.6 s.119
    • /
    • pp.87-100
    • /
    • 2007
  • This study analyzed the key words in Korea's leading apartment brands and newspaper advertisements to examine especially their view-oriented advertising characteristics and green marketing properties. Four study topics were established and examined for this purpose. The following are the results; 1. The key words on apartment brands and catch phrases are $\ulcorner$emotion-oriented$\lrcorner$ words, such as taste, pride, and happiness, and $\ulcorner$environment-oriented$\lrcorner$ to appeal to customers with beautiful views of the nature. Also key words on newspaper advertisement headlines stress views for $\ulcorner$environment-oriented$\lrcorner$ characteristics as well as $\ulcorner$emotion-oriented$\lrcorner$ characteristics for elegance and luxury, and $\ulcorner$function-oriented$\lrcorner$ characteristics for the future. Views, nature and park are key words and everything about illustration that are especially emphasized even on the headlines of newspaper advertisements. 2. Unlike brands or headlines, sub-headlines and body copies stress $\ulcorner$modern-oriented$\lrcorner$ characteristics for life, culture, and accessibility, followed by $\ulcorner$emotion-oriented$\lrcorner$ characteristics. Key words on caption were also highly $\ulcorner$modern-oriented$\lrcorner$, followed by $\ulcorner$environment-oriented$\lrcorner$ and $\ulcorner$function-oriented$\lrcorner$ characteristics for practical aspects. 3. In result of $X^2-test$, $\ulcorner$environment-oriented$\lrcorner$ key words that convey good views, naturalism, parks, and nature are the major representation of green marketing strategies of apartment brands and newspaper advertisements. However, brands were strongly $\ulcorner$emotion-oriented$\lrcorner$, whereas captions were $\ulcorner$modern-oriented$\lrcorner$ and body copies were $\ulcorner$investment-oriented$\lrcorner$. Both apartment brands and advertisements were consistently $\ulcorner$environment-oriented$\lrcorner$, but were not consistent in other factors. 4. Different parts of newspaper advertisements are focusing on green marketing strategies in terms of environmental protection, but are actually leaning toward modern-oriented lifestyles and accessibility. Thus, it is more well-being marketing rather than green marketing. To pursue true green marketing despite the limits of newspaper advertisements, it is necessary to present products and pricing strategies that represent sustainable.

Reusable Multi-story 3D Animation (재구성이 가능한 멀티스토리 3D 애니메이션)

  • Kim, Sungrae;Kim, Ho Sung;Tak, Ji-young;Park, Ji-en;Lim, Sun-hyuk;Kim, Soosanna;Lee, Kyu-seon;Lee, Ji-hyun
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2007.11a
    • /
    • pp.238-242
    • /
    • 2007
  • The existent UCC sites display only finalized contents by publisher. It dose not provide any platform for resources of UCC that public users could reorganize. This paper has developed a platform to be able to produce reusable content using the scenes of the contents and produced a 3D animation with multiple story. It is necessary for user to search the provided contents for easy reorganization of the contents. The scene is classified by the description and information of the scene for handy search. It is obscure for a movie clip to be represent with only one word. Therefore, the proposed platform provides the search technique with a overlapping choice for the specific categories that include most of elements for the scene. Then the user can choose a specific range of the selected movie clip, make a new story with reorganizing the order, and put a caption or BGM on the movie clip. The complete movie clip has the search preferences as a category, new clips, and top favorites. With the Multi-Story line concept, we made a 3D animation about episodes of thermal dolls in the Doll World. This attempt will come to the new marketing way for a field of the visual media like as Music Video, Drama, Feature Film, Commercial Film.

  • PDF