• Title/Summary/Keyword: Computational Domain

Search Result 1,267, Processing Time 0.026 seconds

Coarse Grid Wave Hindcasting in the Yellow Sea Considering the Effect of Tide and Tidal Current (조석 및 조류 효과를 고려한 황해역 광역 파랑 수치모의 실험)

  • Chun, Hwusub;Ahn, Kyungmo
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.30 no.6
    • /
    • pp.286-297
    • /
    • 2018
  • In the present study, wave measurements at KOGA-W01 were analyzed and then the numerical wind waves simulations have been conducted to investigate the characteristics of wind waves in the Yellow sea. According to the present analysis, even though the location of the wave stations are close to the coastal region, the deep water waves are prevailed due to the short fetch length. Chun and Ahn's (2017a, b) numerical model has been extended to the Yellow Sea in this study. The effects of tide and tidal currents should be included in the model to accommodate the distinctive effect of large tidal range and tidal current in the Yellow Sea. The wave hindcasting results were compared with the wave measurements collected KOGA-W01 and Kyeockpo. The comparison shows the reasonable agreements between wave hindcastings and measured data, however the model significantly underestimate the wave period of swell waves from the south due to the narrow computational domain. Despite the poorly prediction in the significant wave period of swell waves which usually have small wave heights, the estimation of the extreme wave height and corresponding wave period shows good agreement with the measurement data.

Secure Training Support Vector Machine with Partial Sensitive Part

  • Park, Saerom
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.4
    • /
    • pp.1-9
    • /
    • 2021
  • In this paper, we propose a training algorithm of support vector machine (SVM) with a sensitive variable. Although machine learning models enable automatic decision making in the real world applications, regulations prohibit sensitive information from being used to protect privacy. In particular, the privacy protection of the legally protected attributes such as race, gender, and disability is compulsory. We present an efficient least square SVM (LSSVM) training algorithm using a fully homomorphic encryption (FHE) to protect a partial sensitive attribute. Our framework posits that data owner has both non-sensitive attributes and a sensitive attribute while machine learning service provider (MLSP) can get non-sensitive attributes and an encrypted sensitive attribute. As a result, data owner can obtain the encrypted model parameters without exposing their sensitive information to MLSP. In the inference phase, both non-sensitive attributes and a sensitive attribute are encrypted, and all computations should be conducted on encrypted domain. Through the experiments on real data, we identify that our proposed method enables to implement privacy-preserving sensitive LSSVM with FHE that has comparable performance with the original LSSVM algorithm. In addition, we demonstrate that the efficient sensitive LSSVM with FHE significantly improves the computational cost with a small degradation of performance.

A CFD Study on Aerodynamic Performances by Geometrical Configuration of Guide Vanes in a Denitrification Facility (탈질 설비 내 안내 깃의 기하학적 형상에 따른 공력 성능에 대한 전산 해석적 연구)

  • Chang-Sik, Lee;Min-Kyu, Kim;Byung-Hee, Ahn;Hee-Taeg, Chung
    • Clean Technology
    • /
    • v.28 no.4
    • /
    • pp.316-322
    • /
    • 2022
  • The flow pattern at the inlet of the catalyst layer in a selective catalytic reduction (SCR) system is one of the key parameters influencing the performance of the denitrification process. In the curved diffusing parts between the ammonia injection grids and the catalyst layers, guide vanes are installed to improve flow uniformity. In the present study, a numerical simulation has been performed to investigate the effect of the geometrical configuration of the guide vanes on the aerodynamic characteristics of a denitrification facility. This application has been made to the existing SCR process in a large-scaled coal-fired power plant. The flow domain to be solved covers the whole region of the flow passages from the exit of the ammonia injection gun to the exit of the catalyst layers. ANSYS-Fluent was used to calculate the three-dimensional steady viscous flow fields with the proper turbulence model fitted to the flow characteristics. The root mean square of velocity and the pressure drop inside the flow passages were chosen as the key performance parameters. Four types of guides vanes were proposed to improve the flow quality compared to the current configuration. The numerical results showed that the type 4 configuration was the most effective at improving the aerodynamic performance in terms of flow uniformity and pressure loss.

Development of 3D Reverse Time Migration Software for Ultra-high-resolution Seismic Survey (초고해상 탄성파 탐사를 위한 3차원 역시간 구조보정 프로그램 개발)

  • Kim, Dae-sik;Shin, Jungkyun;Ha, Jiho;Kang, Nyeon Keon;Oh, Ju-Won
    • Geophysics and Geophysical Exploration
    • /
    • v.25 no.3
    • /
    • pp.109-119
    • /
    • 2022
  • The computational efficiency of reverse time migration (RTM) based on numerical modeling is not secured due to the high-frequency band of several hundred Hz or higher for data acquired through a three-dimensional (3D) ultra-high-resolution (UHR) seismic survey. Therefore, this study develops an RTM program to derive high-quality 3D geological structures using UHR seismic data. In the traditional 3D RTM program, an excitation amplitude technique that stores only the maximum amplitude of the source wavefield and a domain-limiting technique that minimizes the modeling area where the source and receivers are located were used to significantly reduce memory usage and calculation time. The program developed through this study successfully derived a 3D migration image with a horizontal grid size of 1 m for the 3D UHR seismic survey data obtained from the Korea Institute of Geoscience and Mineral Resources in 2019, and geological analysis was conducted.

Wearable Computers

  • Cho, Gil-Soo;Barfield, Woodrow;Baird, Kevin
    • Fiber Technology and Industry
    • /
    • v.2 no.4
    • /
    • pp.490-508
    • /
    • 1998
  • One of the latest fields of research in the area of output devices is tactual display devices [13,31]. These tactual or haptic devices allow the user to receive haptic feedback output from a variety of sources. This allows the user to actually feel virtual objects and manipulate them by touch. This is an emerging technology and will be instrumental in enhancing the realism of wearable augmented environments for certain applications. Tactual displays have previously been used for scientific visualization in virtual environments by chemists and engineers to improve perception and understanding of force fields and of world models populated with the impenetrable. In addition to tactual displays, the use of wearable audio displays that allow sound to be spatialized are being developed. With wearable computers, designers will soon be able to pair spatialized sound to virtual representations of objects when appropriate to make the wearable computer experience even more realistic to the user. Furthermore, as the number and complexity of wearable computing applications continues to grow, there will be increasing needs for systems that are faster, lighter, and have higher resolution displays. Better networking technology will also need to be developed to allow all users of wearable computers to have high bandwidth connections for real time information gathering and collaboration. In addition to the technology advances that make users need to wear computers in everyday life, there is also the desire to have users want to wear their computers. In order to do this, wearable computing needs to be unobtrusive and socially acceptable. By making wearables smaller and lighter, or actually embedding them in clothing, users can conceal them easily and wear them comfortably. The military is currently working on the development of the Personal Information Carrier (PIC) or digital dog tag. The PIC is a small electronic storage device containing medical information about the wearer. While old military dog tags contained only 5 lines of information, the digital tags may contain volumes of multi-media information including medical history, X-rays, and cardiograms. Using hand held devices in the field, medics would be able to call this information up in real time for better treatment. A fully functional transmittable device is still years off, but this technology once developed in the military, could be adapted tp civilian users and provide ant information, medical or otherwise, in a portable, not obstructive, and fashionable way. Another future device that could increase safety and well being of its users is the nose on-a-chip developed by the Oak Ridge National Lab in Tennessee. This tiny digital silicon chip about the size of a dime, is capable of 'smelling' natural gas leaks in stoves, heaters, and other appliances. It can also detect dangerous levels of carbon monoxide. This device can also be configured to notify the fire department when a leak is detected. This nose chip should be commercially available within 2 years, and is inexpensive, requires low power, and is very sensitive. Along with gas detection capabilities, this device may someday also be configured to detect smoke and other harmful gases. By embedding this chip into workers uniforms, name tags, etc., this could be a lifesaving computational accessory. In addition to the future safety technology soon to be available as accessories are devices that are for entertainment and security. The LCI computer group is developing a Smartpen, that electronically verifies a user's signature. With the increase in credit card use and the rise in forgeries, is the need for commercial industries to constantly verify signatures. This Smartpen writes like a normal pen but uses sensors to detect the motion of the pen as the user signs their name to authenticate the signature. This computational accessory should be available in 1999, and would bring increased peace of mind to consumers and vendors alike. In the entertainment domain, Panasonic is creating the first portable hand-held DVD player. This device weight less than 3 pounds and has a screen about 6' across. The color LCD has the same 16:9 aspect ratio of a cinema screen and supports a high resolution of 280,000 pixels and stereo sound. The player can play standard DVD movies and has a hour battery life for mobile use. To summarize, in this paper we presented concepts related to the design and use of wearable computers with extensions to smart spaces. For some time, researchers in telerobotics have used computer graphics to enhance remote scenes. Recent advances in augmented reality displays make it possible to enhance the user's local environment with 'information'. As shown in this paper, there are many application areas for this technology such as medicine, manufacturing, training, and recreation. Wearable computers allow a much closer association of information with the user. By embedding sensors in the wearable to allow it to see what the user sees, hear what the user hears, sense the user's physical state, and analyze what the user is typing, an intelligent agent may be able to analyze what the user is doing and try to predict the resources he will need next or in the near future. Using this information, the agent may download files, reserve communications bandwidth, post reminders, or automatically send updates to colleagues to help facilitate the user's daily interactions. This intelligent wearable computer would be able to act as a personal assistant, who is always around, knows the user's personal preferences and tastes, and tries to streamline interactions with the rest of the world.

  • PDF

Study on Effect of Convection Current Aeration System on Mixing Characteristics and Water Quality of Reservoir (대류식 순환장치의 저수지수체 유동특성 및 수질영향)

  • Lee, Yo-Sang;Lee, Kwang-Man;Koh, Deok-Koo;Yum, Kyung-Taek
    • Korean Journal of Ecology and Environment
    • /
    • v.42 no.1
    • /
    • pp.85-94
    • /
    • 2009
  • This study examines the operational effectiveness of a Convection Current Aeration System (CCAS) in reservoir. CCAS was run from June, 2008 when the thermocline begun forming in the reservoir. This paper reviews the influence of stratification, dissolved oxygen dynamics and temperature in the lake's natural state from June to October 2008. The survey was done on a week basis. Upwelling flow effects a radius of $7{\sim}10m$ at a surface directly and was irrelevant to the strength of thermocline. On the other hand, it was affected the number of working days, and strength of thermocline at vertical profiles of the reservoir. Longer CCAS run, the deeper was the vertical direct flow area. However it didn't break the thermocline during summer season of 2008. The operating efficiency of the CCAS in the reservoir depends on hydraulics and meteological conditions. Computational Fluid Dynamics (CFD) is a very useful tool for evaluating the operating efficiency of fluid dynamics. The geometry for CFD simulation consists of a cylindrical vessel 25 m radius and 40 m height. The CCAS is located in center of domain. The non-uniform tetrahedral meshes had a bulk of the geometry. The meshes ranged from the coarse to the very fine. This is attributed to the cold water flowing into the downcomer and rising, creating a horizontal flow to the top of the CCAS. The result of CFD demonstrate a closer agreement with surveyed data for temperature and flow velocity. Theoretical dispersion volume were calculated at 8m depth, 120 m diameter working for 30 days and 10 m depth, 130 m diameter working for 50 days.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.