• Title/Summary/Keyword: 영상 언어

Search Result 529, Processing Time 0.025 seconds

Development of Web-GIS based Real-time Damage Information System (웹GIS 기반의 실시간 피해정보제공시스템 개발에 관한 연구)

  • Kim, Tae-Hoon;Kim, Kye-Hyun;Lee, Chol-Young
    • Proceedings of the Korean Association of Geographic Inforamtion Studies Conference
    • /
    • 2008.06a
    • /
    • pp.381-386
    • /
    • 2008
  • 본 연구에서는 현장 조사를 통해 얻어지는 피해정보를 광역적 무선통신을 이용해 전송하고 웹GIS의 지도상에 실시간으로 업데이트하여 보다 신속하고 효율적으로 관련 정보를 제공하는 실시간 피해정보제공시스템을 개하였다. 연구 대상지역은 매년 대형 태풍과 집중호우로 인한 피해가 발생하고 있는 부산시 강서구를 선정하였다. 도형 데이터베이스는 대상지역 $179.05km^2$에 대하여 1:5,000 및 1:1,000 수치지도를 기반으로 행정경계 등 7개의 주제도로 구축하였다. 또한, IKONOS위성의 1m급 고해상도 위성영상을 스트리밍 방식으로 구축하여 실제 피해현장과 주변 환경에 관한 정보들을 함께 제공할 수 있는 방안을 연구하였다. 웹GIS 시스템의 웹 프레임은 HTML과 ASP언어를 이용하여 개발되었으며, 맵 서비스 프레임은 ArcIMS를 이용하여 개발되었다. 현장에서 조사된 피해정보는 좌표를 기반으로 하여 웹GIS 시스템에 점으로 표출되며 재해대장 웹 페이지를 통해 피해지역의 영상 및 피해추정액등 세부적인 피해정보를 함께 제공하도록 하였다. 웹GIS 시스템은 피해현황을 GIS 데이터와 함께 실시간으로 제공함으로써 신속한 현장파악, 대응책 마련 및 복구계획 수립을 위한 의사결정에 활용 가능할 것이라 예상되며, 소방방재청에서 구축하고 있는 NDMS와 연계를 통하여 여러 국가기관에서 얻어지는 다양한 피해정보의 제공 또한 가능할 것이라 사료된다.

  • PDF

웹 환경에서의 MPEG-21테스트베드의 구현

  • Son Jeong-Hwa;Son Hyeon-Sik;Gwon Hyeok-Min;Jo Yeong-Ran;Kim Man-Bae
    • Journal of The Institute of Information and Telecommunication Facilities Engineering
    • /
    • v.1 no.2
    • /
    • pp.70-81
    • /
    • 2002
  • 1990년대 후반부터 다양한 디지털 통신망을 이용하여 멀티미디어 컨텐츠 서비스가 가능하게 되었다. 하지만 멀티미디어 컨텐츠의 전달 및 이용을 위한 기반 구조들의 독자적 발전 및 다양한 통합 관리 시스템 때문에 멀티미디어 컨텐츠 표현 방식의 호환성, 네트워크 방식과 단말기 등의 잠재적인 문제점이 발생한다. 이를 위해서 현재 존재하는 기술 및 기반 구조들 사이의 연동을 통한 큰 멀티미디어 프레임워크인 MPEG-21이 진행 중이다. 본 논문에서는 현재 표준화 작업이 진행 중인 MPEG-21을 기반으로 하는 웹 (Web) 기반 테스트베드를 제안한다. 기본적으로 테스트베드는 서버(server), 클라이언트(client), DIA(Digital Item Adaptation)의 세 모듈로 구성된다. 서버의 역할은 멀티미디어 컨텐츠를 Digital Item(DI)의 형태로 생성하고, 클라이언트가 DI를 요구할 경우 DIA 모듈을 통해서 변환된 DI를 클라이언트에게 제공한다. DIA 모듈은 서버에서 동작되며 클라이언트로부터 요청된 DI를 분석하고 클라이언트로부터 전송된 환경 정보를 이용하여 클라이언트 환경에 적합하게 변환된 (adapted) DI를 생성하는 것이 주 기능이다. 클라이언트는 서버에 저장되어 있는 DI를 선택하고 사용자 선호도(user preferences), 터미널 능력(terminal capabilities) 등의 필요한 정보를 서버로 전송한다. 테스트베드에서는 스포츠 경기의 동영상, 정지 영상, 경기 내용, 역사를 기록한 파일 등의 DI를 이용한다. 표현 언어는 XML이며, HTTP 기반의 웹 환경에서 구동되도록 설계된다.

  • PDF

An Implementation of ISP for CMOS Image Sensor (CMOS 카메라 이미지 센서용 ISP 구현)

  • Sonh, Seung-Il;Lee, Dong-Hoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.11 no.3
    • /
    • pp.555-562
    • /
    • 2007
  • In order to display Bayer input stream received from CMOS image sensor to the display device, image signal processing must be performed. That is, the hardware performing the image signal processing for Bayer data is called ISP(Image Signal Processor). We can see real image through ISP processing. ISP executes functionalities for gamma correction, interpolation, color space conversion, image effect, image scale, AWB, AE and AF. In this paper, we obtained the optimum algorithm through software verification of ISP module for CMOS camera image sensor and described using VHDL and verified in ModelSim6.0a simulator. Also we downloaded into Xilinx XCV-1000e for the designed ISP module and completed the board level verification using PCI interface.

A Study on Performance Improvement of GVQA Model Using Transformer (트랜스포머를 이용한 GVQA 모델의 성능 개선에 관한 연구)

  • Park, Sung-Wook;Kim, Jun-Yeong;Park, Jun;Lee, Han-Sung;Jung, Se-Hoon;Sim, Cun-Bo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.749-752
    • /
    • 2021
  • 오늘날 인공지능(Artificial Intelligence, AI) 분야에서 가장 구현하기 어려운 분야 중 하나는 추론이다. 근래 추론 분야에서 영상과 언어가 결합한 다중 모드(Multi-modal) 환경에서 영상 기반의 질의 응답(Visual Question Answering, VQA) 과업에 대한 AI 모델이 발표됐다. 얼마 지나지 않아 VQA 모델의 성능을 개선한 GVQA(Grounded Visual Question Answering) 모델도 발표됐다. 하지만 아직 GVQA 모델도 완벽한 성능을 내진 못한다. 본 논문에서는 GVQA 모델의 성능 개선을 위해 VCC(Visual Concept Classifier) 모델을 ViT-G(Vision Transformer-Giant)/14로 변경하고, ACP(Answer Cluster Predictor) 모델을 GPT(Generative Pretrained Transformer)-3으로 변경한다. 이와 같은 방법들은 성능을 개선하는 데 큰 도움이 될 수 있다고 사료된다.

Visualization and Localization of Fusion Image Using VRML for Three-dimensional Modeling of Epileptic Seizure Focus (VRML을 이용한 융합 영상에서 간질환자 발작 진원지의 3차원적 가시화와 위치 측정 구현)

  • 이상호;김동현;유선국;정해조;윤미진;손혜경;강원석;이종두;김희중
    • Progress in Medical Physics
    • /
    • v.14 no.1
    • /
    • pp.34-42
    • /
    • 2003
  • In medical imaging, three-dimensional (3D) display using Virtual Reality Modeling Language (VRML) as a portable file format can give intuitive information more efficiently on the World Wide Web (WWW). The web-based 3D visualization of functional images combined with anatomical images has not studied much in systematic ways. The goal of this study was to achieve a simultaneous observation of 3D anatomic and functional models with planar images on the WWW, providing their locational information in 3D space with a measuring implement using VRML. MRI and ictal-interictal SPECT images were obtained from one epileptic patient. Subtraction ictal SPECT co-registered to MRI (SISCOM) was performed to improve identification of a seizure focus. SISCOM image volumes were held by thresholds above one standard deviation (1-SD) and two standard deviations (2-SD). SISCOM foci and boundaries of gray matter, white matter, and cerebrospinal fluid (CSF) in the MRI volume were segmented and rendered to VRML polygonal surfaces by marching cube algorithm. Line profiles of x and y-axis that represent real lengths on an image were acquired and their maximum lengths were the same as 211.67 mm. The real size vs. the rendered VRML surface size was approximately the ratio of 1 to 605.9. A VRML measuring tool was made and merged with previous VRML surfaces. User interface tools were embedded with Java Script routines to display MRI planar images as cross sections of 3D surface models and to set transparencies of 3D surface models. When transparencies of 3D surface models were properly controlled, a fused display of the brain geometry with 3D distributions of focal activated regions provided intuitively spatial correlations among three 3D surface models. The epileptic seizure focus was in the right temporal lobe of the brain. The real position of the seizure focus could be verified by the VRML measuring tool and the anatomy corresponding to the seizure focus could be confirmed by MRI planar images crossing 3D surface models. The VRML application developed in this study may have several advantages. Firstly, 3D fused display and control of anatomic and functional image were achieved on the m. Secondly, the vector analysis of a 3D surface model was defined by the VRML measuring tool based on the real size. Finally, the anatomy corresponding to the seizure focus was intuitively detected by correlations with MRI images. Our web based visualization of 3-D fusion image and its localization will be a help to online research and education in diagnostic radiology, therapeutic radiology, and surgery applications.

  • PDF

Susceptibility-Contrast-Enhanced MR Venography of Cat Brain Using Tailored RF Pulse at High Magnetic Field of 4.7 Tesla Superconducting Magnet (4.7T 고자장 초전도 자석에서 Tailored RF를 이용한 고양이 뇌의 자화율 강조영상법에 의한 자기공명혈관 조영술)

  • Moon, Chi-Woong;Kim, Sang-Tae;Lee, Dae-Geun;Im, Tae-Hwan;No, Yong-Man;Cho, Jang-Hee;Lee, Yoon
    • Progress in Medical Physics
    • /
    • v.5 no.1
    • /
    • pp.55-66
    • /
    • 1994
  • After proving home-made imaging pulse sequences including tailored RF pulse by phantom, susceptibility-contrast-enhanced MR venograms of cat brain were obtained using tailored RF gradient-echo(TRGE) method. Sagittal MR imaging of the cat brain obtained by TRGE technique shows several veins, for example, dorsal sagittal sinus, straight sinus, vein of corpus callosum and internal cerebral vein, etc., compared with cats anatomical figure. Tailored RF waveform was generated by PASCAL language in ASPECT 3000 computer(Switzland, Bruker). Rectangular-shaped slice profile with bi-linear ramp function as phase distribution in the slice, at which maximum value was 2$\pi$, was fourier transformed to make tailored RF pulse. Experimental MR imaging parameters were TR/TE=205/10 msec, slice thickness TH=7mm, maxtrix size=256$\times$256, in-plane resolution=0.62$\times$0.31mm$^2$, and field of view(FOV)=8cm for both conventional gradient-echo(GE) imaging and TRGE imaging techniques.

  • PDF

A Quantization-adaptive Watermarking Algorithm to Protect MPEG Moving Picture Contents (MPEG 동영상 컨텐츠 보호를 위한 양자화-적응적 워터마킹 알고리즘)

  • Kim Joo-Hyuk;Choi Hyun-Jun;Seo Young-Ho;Kim Dong-Wook
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.6
    • /
    • pp.149-158
    • /
    • 2005
  • This paper proposed a blind watermarking method for video contents which satisfies both the invisibility and the robustness to attacks to prohibit counterfeiting, modification, illegal usage and illegal re-production of video contents. This watermarking algorithm targets MPEG compression system and was designed to control the amount of watermarking to be inserted according to the adaptive quantization scale code to follow the adaptive quantization of the compression system. The inserting positions of the watermark were chosen by considering the frequency property of an image and horizontal, vertical and diagonal property of a $8{\times}8$ image block. Also the amount of watermarking for each watermark bit was decided by considering the quantization step. This algorithm was implemented by C++ and experimented for invisibility and robustness with MPEG-2 system. The experiment results showed that the method satisfied enough the invisibility of the inserted watermark and robustness against attacks. For the general attacks, the error rate of the extracted watermark was less than $10\%$, which is enough in robustness against the attacks. Therefore, this algorithm is expected to be used effectively as a part in many MPEG systems for real-time watermarking, especially in the sensitive applications to the network environments.

Speech Activity Decision with Lip Movement Image Signals (입술움직임 영상신호를 고려한 음성존재 검출)

  • Park, Jun;Lee, Young-Jik;Kim, Eung-Kyeu;Lee, Soo-Jong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.26 no.1
    • /
    • pp.25-31
    • /
    • 2007
  • This paper describes an attempt to prevent the external acoustic noise from being misrecognized as the speech recognition target. For this, in the speech activity detection process for the speech recognition, it confirmed besides the acoustic energy to the lip movement image signal of a speaker. First of all, the successive images are obtained through the image camera for PC. The lip movement whether or not is discriminated. And the lip movement image signal data is stored in the shared memory and shares with the recognition process. In the meantime, in the speech activity detection Process which is the preprocess phase of the speech recognition. by conforming data stored in the shared memory the acoustic energy whether or not by the speech of a speaker is verified. The speech recognition processor and the image processor were connected and was experimented successfully. Then, it confirmed to be normal progression to the output of the speech recognition result if faced the image camera and spoke. On the other hand. it confirmed not to output of the speech recognition result if did not face the image camera and spoke. That is, if the lip movement image is not identified although the acoustic energy is inputted. it regards as the acoustic noise.

A Convergence Study for the Academic Systematization of Cartoon-animation (만화영상학의 학문적 체계화를 위한 융합적 연구)

  • Lim, Jae-Hwan
    • Cartoon and Animation Studies
    • /
    • s.43
    • /
    • pp.285-320
    • /
    • 2016
  • Cartoons and Animation are convergent arts created with a composite application of language arts described in the form of literary texts and sounds, plastic arts visualized in the form of artistic paintings, and film arts produced in the form of moving pictures. An academic university major in cartoons and animation studies established in late 20th century however, did not satisfactorily meet the needs in academic research and development and the free expression of artistic creation was limited. In order to systematize the major in cartoons and animation studies, an convergent approach to establish and clarify following are in demand : the terms and definitions, the historical developments, the research areas and methods, the major education and related jobs and start-ups. New culture and arts industries including cartoons, animation, moving images, and games contents are not yet listed in the industries listing service jointly provided online by the portal site Naver.com and Hyung-Seol publishing company. Above all, cartoons and animation are inseparably related to each other that even if one uses the term separately and independently, the meaning may not be complete. So a new combined term "Animatoon" can be established for the major in cartoons and animation studies and also used for its degree with concentrations of cartoons, animation, moving images, games, and etc. In the Introduction, a new combined term Animatoon is defined and explained the use of this term as the name of the major and degree in cartoons and animation studies. In the body, first, the Historical Developments classified Animatoon in the ancient times, the medieval times, and the modern times and they are analyzed with the help of esthetics and arts using examples of mural frescos, animal painting, religion cartoons, caricatures, cartoons, satire cartoons, comics, animation, 2 or 3 dimensional webtoons, and K-toons. Second, the Research Areas of Animatoon reviewed the theories, genres, artworks, and artists and the Research Methods of Animatoon presented the curriculum that integrated the courses in humanities, science technologies, culture and arts, and etc. Third, the Major Education considered Animatoon education in children, young adults, students of the major and the Related Jobs and Start-Ups explored various jobs relating to personal creation of artwork and collective production of business-oriented artwork. In the Conclusion, the current challenges of Animatoon considered personalization of the artists, specialization of the contents, diversification of the types, and liberalization of the art creation. And the direction of improvement advocated Animatoon to be an academic field of study, to be an art, to be a culture, and to be an industry. The importance of cartoons and animation along with videos and games rose in the 21st century. In order for cartoons and animation to take a leading role, make efforts in studying Animatoon academically and also in developing Animatoon as good contents in the cultural industries.

Functional Brain Mapping Using $H_2^{15}O$ Positron Emission Tomography ( I ): Statistical Parametric Mapping Method ($H_2^{15}O$ 양전자단층촬영술을 이용한 뇌기능 지도 작성(I): 통계적 파라메터 지도작성법)

  • Lee, Dong-Soo;Lee, Jae-Sung;Kim, Kyeong-Min;Chung, June-Key;Lee, Myung-Chul
    • The Korean Journal of Nuclear Medicine
    • /
    • v.32 no.3
    • /
    • pp.225-237
    • /
    • 1998
  • Purpose: We investigated the statistical methods to compose the functional brain map of human working memory and the principal factors that have an effect on the methods for localization. Materials and Methods: Repeated PET scans with successive four tasks, which consist of one control and three different activation tasks, were performed on six right-handed normal volunteers for 2 minutes after bolus injections of 925 MBq $H_2^{15}O$ at the intervals of 30 minutes. Image data were analyzed using SPM96 (Statistical Parametric Mapping) implemented with Matlab (Mathworks Inc., U.S.A.). Images from the same subject were spatially registered and were normalized using linear and nonlinear transformation methods. Significant difference between control and each activation state was estimated at every voxel based on the general linear model. Differences of global counts were removed using analysis of covariance (ANCOVA) with global activity as covariate. Using the mean and variance for each condition which was adjusted using ANCOVA, t-statistics was performed on every voxel To interpret the results more easily, t-values were transformed to the standard Gaussian distribution (Z-score). Results: All the subjects carried out the activation and control tests successfully. Average rate of correct answers was 95%. The numbers of activated blobs were 4 for verbal memory I, 9 for verbal memory II, 9 for visual memory, and 6 for conjunctive activation of these three tasks. The verbal working memory activates predominantly left-sided structures, and the visual memory activates the right hemisphere. Conclusion: We conclude that rCBF PET imaging and statistical parametric mapping method were useful in the localization of the brain regions for verbal and visual working memory.

  • PDF