• Title/Summary/Keyword: long-memory process

Search Result 163, Processing Time 0.026 seconds

Improving the Cyclic Stability of Electrochromic Mirrors Composed of Gel Electrolyte (겔 전해질로 구성된 전기변색 거울의 내구성 향상)

  • Ji-Hyeong Lee;Kwang-Mo Kang;Sang Bum Lee;Yoon-Chae Nah
    • Korean Journal of Materials Research
    • /
    • v.34 no.8
    • /
    • pp.400-407
    • /
    • 2024
  • The reversible metal electrodeposition (RME) process is used to prepare electrochromic mirrors with reflective-transparent optical states, by depositing metal particles on transparent conductive substrates. These RME based devices can be used in smart windows to regulate indoor temperatures and light levels, serving dual purposes as lighting elements. Commercialization efforts are focused on achieving large-scale production, long-term durability, and a memory effect that maintains coloration without applied voltage. Enhancing durability has received particular attention, leading to the development of electrochromic mirrors that employ gel electrolytes, which are expected to reduce electrolyte leakage and improve mechanical stability compared to traditional liquid electrolyte devices. The gel electrolytes offer the additional advantage of various colors, by controlling the metal particle size and enabling smoother, denser formations. In this study, we investigated improving the durability of RME devices by adding polyvinyl butyral (PVB) to the liquid electrolyte and optimizing the concentration of PVB. Incorporating 10 % PVB resulted in excellent interfacial properties and superior electrochromic stability, with 92.6 % retention after 1,000 cycles.

A Qualitative Case Study about Social Support and Participation Experience in Senior Counselling Program (라이프케어를 위한 노인심리상담교육 경험의 사회적 지지와 참여에 관한 질적사례연구)

  • Kim, Yong-Nam
    • Journal of Korea Entertainment Industry Association
    • /
    • v.13 no.3
    • /
    • pp.221-229
    • /
    • 2019
  • This study intended to find out the participants' learning process in senior counselling program for life care and the meanings of senior's social support and participation in society. The study method was conducted with this researcher through a narrative' interview during 15-week learning course of senior counselling program in the life-long learning institute at A university. As the results, it could found that senior trainees were able to enhance their intellectual level through the senior counselling program to enhance their self-esteems and improve their memory through the education. Also, through the license of senior counsellor, they became more sophisticated and developed the abilities of comprehension and empathy about the senior in experiencing the intergenerational interactions in the space of social interactions, and the enjoyment of learning changed the senior's behaviors and increased their life quality. This research identified that a learning process enabled to continue the senior's social support and participation in society and made the senior achieve various interactions and practical learning through the exchange of trainees' various cultures. Based on these results, it is expected that various, systematic educational programs will be able to improve the senior's life care in future.

An Empirical Study on the Relationships among Safeguarding Mechanism, Relationship Learning, and Relationship Performance in Technology Cooperation Network by Applying Resource Based Theory (자원기반이론을 적용한 기술협력 네트워크에서 보호 메커니즘, 관계학습, 관계성과의 관계에 대한 실증연구)

  • Kang, Seok-Min
    • Management & Information Systems Review
    • /
    • v.35 no.2
    • /
    • pp.45-66
    • /
    • 2016
  • Firms can make scale of economy and scope of economy by internalizing and using new advanced technology and knowledge from technology cooperation network, decrease risk and cost with partner firm of technology cooperation network, and increase market advantage of product & strengthen firms' position in the market. Due to the advantages of technology cooperation network, the related studies have focused on the positive effect of technology cooperation network. However, the related studies investigating the relationship between technology cooperation network and firm performance have only examined the role of technology cooperation network. Safeguarding mechanism, relationship learning, and relationship performance are categorized into the process of technology cooperation network, and this categorization is applied as resources, capability, and performance by resource based view. The empirical results are reported as belows. First, relationship specific investment and relationship capital positively affect on relationship learning as capability. Second, information sharing, common information understanding, and relationship specific memory development positively affect on long-term orientation, but information sharing has no impact on efficiency and effectiveness. Third, relationship specific investment positively affects on relationship capital and efficiency and effectiveness have positive effects on long-term orientation. Applying technology cooperation network in asymmetric technology dependency with resource based theory, this study suggested the importance of both safeguarding and relationship learning by investigating the relationship among safeguarding, relationship learning, and relationship performance. And it is worthy that this study investigated how firms' behavior change affects relationship performance in the relationship of technology cooperation partner.

  • PDF

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Meaning and Value Analysis of Records of Laos Renewable Energy Support Activities Collection (라오스 재생가능에너지 지원활동 컬렉션의 의미와 가치 연구)

  • Ju, Hyun Mi;Yim, Jin Hee
    • The Korean Journal of Archival Studies
    • /
    • no.51
    • /
    • pp.45-87
    • /
    • 2017
  • In recent years, there were some who have conducted research on deriving the social and historical meanings of records through the analysis of specific records collections. This trend is an attempt to pay attention to human actions in the society and to read the society again through the records created by such actions. In this study, I derive various meanings and values of these records through the analysis of the "Laos Renewable Energy Support Activities" collection. Moreover, I study how the collection was reconstructed by the Human and Memory Archives. The "Laos Renewable Energy Support Activities" is the personal record of the donor who led the project, and contains the process and results of the project. Through this collection, I was able to look at the life of the donor as a foreign aid activist in Laos and realized his values. Furthermore, through the business process record, I was able to discover the implications of climate change response overseas aid projects. In addition, I was able to look at the culture and environment of Laos through the eyes of the donor who has been residing there for a long time.

Water temperature prediction of Daecheong Reservoir by a process-guided deep learning model (역학적 모델과 딥러닝 모델을 융합한 대청호 수온 예측)

  • Kim, Sung Jin;Park, Hyungseok;Lee, Gun Ho;Chung, Se Woong
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.88-88
    • /
    • 2021
  • 최근 수자원과 수질관리 분야에 자료기반 머신러닝 모델과 딥러닝 모델의 활용이 급증하고 있다. 그러나 딥러닝 모델은 Blackbox 모델의 특성상 고전적인 질량, 운동량, 에너지 보존법칙을 고려하지 않고, 데이터에 내재된 패턴과 관계를 해석하기 때문에 물리적 법칙을 만족하지 않는 예측결과를 가져올 수 있다. 또한, 딥러닝 모델의 예측 성능은 학습데이터의 양과 변수 선정에 크게 영향을 받는 모델이기 때문에 양질의 데이터가 제공되지 않으면 모델의 bias와 variation이 클 수 있으며 정확도 높은 예측이 어렵다. 최근 이러한 자료기반 모델링 방법의 단점을 보완하기 위해 프로세스 기반 수치모델과 딥러닝 모델을 결합하여 두 모델링 방법의 장점을 활용하는 연구가 활발히 진행되고 있다(Read et al., 2019). Process-Guided Deep Learning (PGDL) 방법은 물리적 법칙을 반영하여 딥러닝 모델을 훈련시킴으로써 순수한 딥러닝 모델의 물리적 법칙 결여성 문제를 해결할 수 있는 대안으로 활용되고 있다. PGDL 모델은 딥러닝 모델에 물리적인 법칙을 해석할 수 있는 추가변수를 도입하며, 딥러닝 모델의 매개변수 최적화 과정에서 Cost 함수에 물리적 법칙을 위반하는 경우 Penalty를 추가하는 알고리즘을 도입하여 물리적 보존법칙을 만족하도록 모델을 훈련시킨다. 본 연구의 목적은 대청호의 수심별 수온을 예측하기 위해 역학적 모델과 딥러닝 모델을 융합한 PGDL 모델을 개발하고 적용성을 평가하는데 있다. 역학적 모델은 2차원 횡방향 평균 수리·수질 모델인 CE-QUAL-W2을 사용하였으며, 대청호를 대상으로 2017년부터 2018년까지 총 2년간 수온과 에너지 수지를 모의하였다. 기상(기온, 이슬점온도, 풍향, 풍속, 운량), 수문(저수위, 유입·유출 유량), 수온자료를 수집하여 CE-QUAL-W2 모델을 구축하고 보정하였으며, 모델은 저수위 변화, 수온의 수심별 시계열 변동 특성을 적절하게 재현하였다. 또한, 동일기간 대청호 수심별 수온 예측을 위한 순환 신경망 모델인 LSTM(Long Short-Term Memory)을 개발하였으며, 종속변수는 수온계 체인을 통해 수집한 수심별 고빈도 수온 자료를 사용하고 독립 변수는 기온, 풍속, 상대습도, 강수량, 단파복사에너지, 장파복사에너지를 사용하였다. LSTM 모델의 매개변수 최적화는 지도학습을 통해 예측값과 실측값의 RMSE가 최소화 되로록 훈련하였다. PGDL 모델은 동일 기간 LSTM 모델과 동일 입력 자료를 사용하여 구축하였으며, 역학적 모델에서 얻은 에너지 수지를 만족하지 않는 경우 Cost Function에 Penalty를 추가하여 물리적 보존법칙을 만족하도록 훈련하고 수심별 수온 예측결과를 비교·분석하였다.

  • PDF

What Changed and Unchanged After Science Class: Analyzing High School Student's Conceptual Change on Circular Motion Based on Mental Model Theory (과학수업 후 변하는 것과 변하지 않는 것: 정신모형 이론을 중심으로 한 고등학생의 원운동 개념변화 사례 분석)

  • Park, Ji-Yeon;Lee, Gyoung-Ho;Shin, Jong-Ho;Song, Sang-Ho
    • Journal of The Korean Association For Science Education
    • /
    • v.26 no.4
    • /
    • pp.475-491
    • /
    • 2006
  • In physics education, the research on students' conceptions has developed in the discussion on the nature and the difficulty of conceptual change. Recently, mental models have been a theoretical background in concrete arguments on "how students' conceptions are constructed or created." Mental models that integrate information in the presented problem and individual knowledge in their long-term memory have important information about not only expressed ideas but also in the thinking process behind the expressed ideas. The purpose of this study is to investigate the forming process and the characteristics of high school student's mental models about circular motion, and how they were changed by instruction. We used the think-aloud method based on the instrument for identifying student's mental models about circular motion, pretest of physics concept, mind map and interview for investigating student's characteristics. The results of the study showed that instructions based on the mental model theory facilitated scientific expressed model, but several factors that affected forming mental models like epistemological belief didn't change scientifically after 3 lessons.

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

A Study of the Planning Process, Design Idea and Implementation of the Gwanghwamun Plaza (광화문광장 조성과정 및 설계 연구)

  • Shin, Hyun-Don;Zoh, Kyung-Jin
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.41 no.4
    • /
    • pp.24-41
    • /
    • 2013
  • This study aims to investigate the complex planning processes and design ideas of the Gwanghwamun Plaza which was opened in 2009. The opening of the plaza was significant as it was built in order to restore the symbolic meaning of axis in the historical Joseon Dynasty that was distorted during the Japanese Imperialism. The plaza itself attracts many citizens and tourists by providing the empty ground carrying historical ambiences around. In this paper, the story of the Gwanghwamun plaza will be summarized. Particularly, the background of promoting the project will be discussed and the whole planning process will be dealt with. The plaza was realized through several stages. First, the planning stage will be reviewed. The planning stage had been a quite long process since the initial idea was discussed. Since the early 90s, the political decision of making the plaza was made through the change of people's understanding toward public space. At this stage, the city government worked together with diverse citizens and professionals to share the vision and to realize the right decision in making the plaza. Second, the design stage will be elaborated in detail. This is the second design stage. The former was the idea competition and the latter was the turn-key base. The final design scheme emphasized the restoration of symbolic axis and the forgotten Yukjo Street. The scheme consists of four zones such as history restoration zone, prospect and history representation zone, culture zone, and the urban zone. Through the whole design process, the original idea remained as it was. The design concept was "a place of memory and prospect." It emphasizes the history representation, view corridor, cultural activities platform as well as the emptiness and flexibility of the basic premise of the plaza. Finally, the construction stage will be discussed. There were some additions and omissions in the construction process. The design chances in the construction stage will be reviewed in detail. After the opening of the plaza, there were some changes in the detail design. We will discuss how and why these modifications were made. In the end, the social and cultural implication of the plaza will be discussed. The storytelling of the Gwanghwamun Plaza will contribute to the clear understanding of planning and design process of pubic places. Based on this reflection, we are able to think about some suggestions of public projects for the future.