• Title/Summary/Keyword: Optimization Techniques

Search Result 1,420, Processing Time 0.038 seconds

Optimal Configuration of the Truss Structures by Using Decomposition Method of Three-Phases (3단계(段階) 분할기법(分割技法)에 의한 평면(平面)트러스 구조물(構造物)의 형상(形狀) 최적화(最適化)에 관한 연구(硏究))

  • Lee, Gyu Won;Song, Gi Beom
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.12 no.3
    • /
    • pp.39-55
    • /
    • 1992
  • In this research, a Three Level Decomposition technique has been developed for configuration design optimization of truss structures. In the first level, as design variables, behavior variables are used and the strain energy has been treated as the cost function to be maximized so that the truss structure can absorb maximum energy. For design constraint of the optimal design problem, allowable stress, buckling stress, and displacement under multi-loading conditions are considered. In the second level, design problem is formulated using the cross-sectional area as the design variable and the weight of the truss structure as the cost function. As for the design constraint, the equilibrium equation with the optimal displacement obtained in the first level is used. In the third level, the nodal point coordinates of the truss structure are used as coordinating variable and the weight has been taken as the cost function. An advantage of the Three Level Decomposition technique is that the first and second level design problems are simple because they are linear programming problems. Moreover, the method is efficient because it is not necessary to carry out time consuming structural analysis and techniques for sensitivity analysis during the design optimization process. By treating the nodal point coordinates as design variables, the third level becomes unconstrained optimal design problems which is easier to solve. Moreover, by using different convergence criteria at each level of design problem, improved convergence can be obtained. The proposed technique has been tested using four different truss structures to yield almost identical optimum designs in the literature with efficient convergence rate regardless of constraint types and configuration of truss structures.

  • PDF

N- and P-doping of Transition Metal Dichalcogenide (TMD) using Artificially Designed DNA with Lanthanide and Metal Ions

  • Kang, Dong-Ho;Park, Jin-Hong
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 2016.02a
    • /
    • pp.292-292
    • /
    • 2016
  • Transition metal dichalcogenides (TMDs) with a two-dimensional layered structure have been considered highly promising materials for next-generation flexible, wearable, stretchable and transparent devices due to their unique physical, electrical and optical properties. Recent studies on TMD devices have focused on developing a suitable doping technique because precise control of the threshold voltage ($V_{TH}$) and the number of tightly-bound trions are required to achieve high performance electronic and optoelectronic devices, respectively. In particular, it is critical to develop an ultra-low level doping technique for the proper design and optimization of TMD-based devices because high level doping (about $10^{12}cm^{-2}$) causes TMD to act as a near-metallic layer. However, it is difficult to apply an ion implantation technique to TMD materials due to crystal damage that occurs during the implantation process. Although safe doping techniques have recently been developed, most of the previous TMD doping techniques presented very high doping levels of ${\sim}10^{12}cm^{-2}$. Recently, low-level n- and p-doping of TMD materials was achieved using cesium carbonate ($Cs_2CO_3$), octadecyltrichlorosilane (OTS), and M-DNA, but further studies are needed to reduce the doping level down to an intrinsic level. Here, we propose a novel DNA-based doping method on $MoS_2$ and $WSe_2$ films, which enables ultra-low n- and p-doping control and allows for proper adjustments in device performance. This is achieved by selecting and/or combining different types of divalent metal and trivalent lanthanide (Ln) ions on DNA nanostructures. The available n-doping range (${\Delta}n$) on the $MoS_2$ by Ln-DNA (DNA functionalized by trivalent Ln ions) is between $6{\times}10^9cm^{-2}$ and $2.6{\times}10^{10}cm^{-2}$, which is even lower than that provided by pristine DNA (${\sim}6.4{\times}10^{10}cm^{-2}$). The p-doping change (${\Delta}p$) on $WSe_2$ by Ln-DNA is adjusted between $-1.0{\times}10^{10}cm^{-2}$ and $-2.4{\times}10^{10}cm^{-2}$. In the case of Co-DNA (DNA functionalized by both divalent metal and trivalent Ln ions) doping where $Eu^{3+}$ or $Gd^{3+}$ ions were incorporated, a light p-doping phenomenon is observed on $MoS_2$ and $WSe_2$ (respectively, negative ${\Delta}n$ below $-9{\times}10^9cm^{-2}$ and positive ${\Delta}p$ above $1.4{\times}10^{10}cm^{-2}$) because the added $Cu^{2+}$ ions probably reduce the strength of negative charges in Ln-DNA. However, a light n-doping phenomenon (positive ${\Delta}n$ above $10^{10}cm^{-2}$ and negative ${\Delta}p$ below $-1.1{\times}10^{10}cm^{-2}$) occurs in the TMD devices doped by Co-DNA with $Tb^{3+}$ or $Er^{3+}$ ions. A significant (factor of ~5) increase in field-effect mobility is also observed on the $MoS_2$ and $WSe_2$ devices, which are, respectively, doped by $Tb^{3+}$-based Co-DNA (n-doping) and $Gd^{3+}$-based Co-DNA (p-doping), due to the reduction of effective electron and hole barrier heights after the doping. In terms of optoelectronic device performance (photoresponsivity and detectivity), the $Tb^{3+}$ or $Er^{3+}$-Co-DNA (n-doping) and the $Eu^{3+}$ or $Gd^{3+}$-Co-DNA (p-doping) improve the $MoS_2$ and $WSe_2$ photodetectors, respectively.

  • PDF

Dosimetric Comparison of Three Dimensional Conformal Radiation Radiotherapy and Helical Tomotherapy Partial Breast Cancer (유방암 환자의 3D-CRT, TOMO 방법에 따른 선량 분포 평가)

  • Kim, Dae-Woong;Kim, Jong-Won;Choi, Yun-Kyeong;Kim, Jung-Soo;Hwang, Jae-Woong;Jeong, Kyeong-Sik;Choi, Gye-Suk
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.20 no.1
    • /
    • pp.11-15
    • /
    • 2008
  • Purpose: The goal of radiation treatment is to deliver a prescribed radiation dose to the target volume accurately while minimizing dose to normal tissues. In this paper, we comparing the dose distribution between three dimensional conformal radiation radiotherapy (3D-CRT) and helical tomotherapy (TOMO) plan for partial breast cancer. Materials and Methods: Twenty patients were included in the study, and plans for two techniques were developed for each patient (left breast:10 patients, right breast:10 patients). For each patient 3D-CRT planning was using pinnacle planning system, inverse plan was made using Tomotherapy Hi-Art system and using the same targets and optimization goals. We comparing the Homogeneity index (HI), Conformity index (CI) and sparing of the organs at risk for dose-volume histogram. Results: Whereas the HI, CI of TOMO was significantly better than the other, 3D-CRT was observed to have significantly poorer HI, CI. The percentage ipsilateral non-PTV breast volume that was delivered 50% of the prescribed dose was 3D-CRT (mean: 40.4%), TOMO (mean: 18.3%). The average ipsilateral lung volume percentage receiving 20% of the PD was 3D-CRT (mean: 4.8%), TOMO (mean: 14.2), concerning the average heart volume receiving 20% and 10% of the PD during treatment of left breast cancer 3D-CRT (mean: 1.6%, 3.0%), TOMO (mean: 9.7%, 26.3%) Conclusion: In summary, 3D-CRT and TOMO techniques were found to have acceptable PTV coverage in our study. However, in TOMO, high conformity to the PTV and effective breast tissue sparing was achieved at the expense of considerable dose exposure to the lung and heart.

  • PDF

Usefulness evaluation of Hybrid planning through dosimetric comparision of Three Dimensinal Conformal Radiation Radiotherapy and Hybrid planning for left breast cancer (유방암 환자의 방사선 치료시 Energy와 Wedge를 combine한 Hybrid plan의 유용성 평가)

  • Chae, Moon Ki;Park, Byung Soo;Ahn, Jong Ho;Song, Ki Won
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.26 no.1
    • /
    • pp.91-98
    • /
    • 2014
  • Purpose : To compare the dosimetry for the left breast cancer treatment between three dimensional conformal radiation radiotherapy (3D-CRT) and Hybrid planning and to estimate usefulness of Hybrid planning Materials and Methods : Five patients with left breast cancer were included in the study. They were planned using several different radiotherapy techniques including: 1)open rectangular field, 2)tangential wedge-based field 3)field in field, 4)hybrid planning(energy, wedge combine). For each patient planning was using Light Speed RT-16 CT and PINNACLE planning system-ver.9.2. Hybrid plan was made using same system and using the same targets and optimization goals. We comparing the Homogeneity Index(HI), normal organs at the does-volume histogram(DVH) Results : In all plans, the Homogeneity Index(HI) of Hybrid planning was significantly better than other. Dose comparison of HI= 2D-RT:38.32, TW:38.32, FIF:29.22, HYBRID:30.57. 2D-RT, TW, FIF Hybrid$V_{75_-lung}$=112.33, 125.14, 121.3, 123.78. $V_{50_-lung}$=155.43, 159.62, 157.96, 159.06. $V_{25_-lung}$=199.86, 200.22, 198.65, 200.31. $V_{50_-heart}$=26.07, 27.1, 26.85, 27.17 $V_{30_-heart}$=33.71, 34.37, 34.15, 34.65 Conclusion : In summary, 3D-CRT, Hybrid planning techniques were found to have acceptableCTV coverage in our study. However the Hybrid planning increased radiation dose exposure to normal tissue. If you apply for treatment of inhomogeneity areas like lung, For best results will be achieved.

A Study on the Prediction Model of Stock Price Index Trend based on GA-MSVM that Simultaneously Optimizes Feature and Instance Selection (입력변수 및 학습사례 선정을 동시에 최적화하는 GA-MSVM 기반 주가지수 추세 예측 모형에 관한 연구)

  • Lee, Jong-sik;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.4
    • /
    • pp.147-168
    • /
    • 2017
  • There have been many studies on accurate stock market forecasting in academia for a long time, and now there are also various forecasting models using various techniques. Recently, many attempts have been made to predict the stock index using various machine learning methods including Deep Learning. Although the fundamental analysis and the technical analysis method are used for the analysis of the traditional stock investment transaction, the technical analysis method is more useful for the application of the short-term transaction prediction or statistical and mathematical techniques. Most of the studies that have been conducted using these technical indicators have studied the model of predicting stock prices by binary classification - rising or falling - of stock market fluctuations in the future market (usually next trading day). However, it is also true that this binary classification has many unfavorable aspects in predicting trends, identifying trading signals, or signaling portfolio rebalancing. In this study, we try to predict the stock index by expanding the stock index trend (upward trend, boxed, downward trend) to the multiple classification system in the existing binary index method. In order to solve this multi-classification problem, a technique such as Multinomial Logistic Regression Analysis (MLOGIT), Multiple Discriminant Analysis (MDA) or Artificial Neural Networks (ANN) we propose an optimization model using Genetic Algorithm as a wrapper for improving the performance of this model using Multi-classification Support Vector Machines (MSVM), which has proved to be superior in prediction performance. In particular, the proposed model named GA-MSVM is designed to maximize model performance by optimizing not only the kernel function parameters of MSVM, but also the optimal selection of input variables (feature selection) as well as instance selection. In order to verify the performance of the proposed model, we applied the proposed method to the real data. The results show that the proposed method is more effective than the conventional multivariate SVM, which has been known to show the best prediction performance up to now, as well as existing artificial intelligence / data mining techniques such as MDA, MLOGIT, CBR, and it is confirmed that the prediction performance is better than this. Especially, it has been confirmed that the 'instance selection' plays a very important role in predicting the stock index trend, and it is confirmed that the improvement effect of the model is more important than other factors. To verify the usefulness of GA-MSVM, we applied it to Korea's real KOSPI200 stock index trend forecast. Our research is primarily aimed at predicting trend segments to capture signal acquisition or short-term trend transition points. The experimental data set includes technical indicators such as the price and volatility index (2004 ~ 2017) and macroeconomic data (interest rate, exchange rate, S&P 500, etc.) of KOSPI200 stock index in Korea. Using a variety of statistical methods including one-way ANOVA and stepwise MDA, 15 indicators were selected as candidate independent variables. The dependent variable, trend classification, was classified into three states: 1 (upward trend), 0 (boxed), and -1 (downward trend). 70% of the total data for each class was used for training and the remaining 30% was used for verifying. To verify the performance of the proposed model, several comparative model experiments such as MDA, MLOGIT, CBR, ANN and MSVM were conducted. MSVM has adopted the One-Against-One (OAO) approach, which is known as the most accurate approach among the various MSVM approaches. Although there are some limitations, the final experimental results demonstrate that the proposed model, GA-MSVM, performs at a significantly higher level than all comparative models.

Development of a Stock Trading System Using M & W Wave Patterns and Genetic Algorithms (M&W 파동 패턴과 유전자 알고리즘을 이용한 주식 매매 시스템 개발)

  • Yang, Hoonseok;Kim, Sunwoong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.63-83
    • /
    • 2019
  • Investors prefer to look for trading points based on the graph shown in the chart rather than complex analysis, such as corporate intrinsic value analysis and technical auxiliary index analysis. However, the pattern analysis technique is difficult and computerized less than the needs of users. In recent years, there have been many cases of studying stock price patterns using various machine learning techniques including neural networks in the field of artificial intelligence(AI). In particular, the development of IT technology has made it easier to analyze a huge number of chart data to find patterns that can predict stock prices. Although short-term forecasting power of prices has increased in terms of performance so far, long-term forecasting power is limited and is used in short-term trading rather than long-term investment. Other studies have focused on mechanically and accurately identifying patterns that were not recognized by past technology, but it can be vulnerable in practical areas because it is a separate matter whether the patterns found are suitable for trading. When they find a meaningful pattern, they find a point that matches the pattern. They then measure their performance after n days, assuming that they have bought at that point in time. Since this approach is to calculate virtual revenues, there can be many disparities with reality. The existing research method tries to find a pattern with stock price prediction power, but this study proposes to define the patterns first and to trade when the pattern with high success probability appears. The M & W wave pattern published by Merrill(1980) is simple because we can distinguish it by five turning points. Despite the report that some patterns have price predictability, there were no performance reports used in the actual market. The simplicity of a pattern consisting of five turning points has the advantage of reducing the cost of increasing pattern recognition accuracy. In this study, 16 patterns of up conversion and 16 patterns of down conversion are reclassified into ten groups so that they can be easily implemented by the system. Only one pattern with high success rate per group is selected for trading. Patterns that had a high probability of success in the past are likely to succeed in the future. So we trade when such a pattern occurs. It is a real situation because it is measured assuming that both the buy and sell have been executed. We tested three ways to calculate the turning point. The first method, the minimum change rate zig-zag method, removes price movements below a certain percentage and calculates the vertex. In the second method, high-low line zig-zag, the high price that meets the n-day high price line is calculated at the peak price, and the low price that meets the n-day low price line is calculated at the valley price. In the third method, the swing wave method, the high price in the center higher than n high prices on the left and right is calculated as the peak price. If the central low price is lower than the n low price on the left and right, it is calculated as valley price. The swing wave method was superior to the other methods in the test results. It is interpreted that the transaction after checking the completion of the pattern is more effective than the transaction in the unfinished state of the pattern. Genetic algorithms(GA) were the most suitable solution, although it was virtually impossible to find patterns with high success rates because the number of cases was too large in this simulation. We also performed the simulation using the Walk-forward Analysis(WFA) method, which tests the test section and the application section separately. So we were able to respond appropriately to market changes. In this study, we optimize the stock portfolio because there is a risk of over-optimized if we implement the variable optimality for each individual stock. Therefore, we selected the number of constituent stocks as 20 to increase the effect of diversified investment while avoiding optimization. We tested the KOSPI market by dividing it into six categories. In the results, the portfolio of small cap stock was the most successful and the high vol stock portfolio was the second best. This shows that patterns need to have some price volatility in order for patterns to be shaped, but volatility is not the best.

A Study on Market Size Estimation Method by Product Group Using Word2Vec Algorithm (Word2Vec을 활용한 제품군별 시장규모 추정 방법에 관한 연구)

  • Jung, Ye Lim;Kim, Ji Hui;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.1-21
    • /
    • 2020
  • With the rapid development of artificial intelligence technology, various techniques have been developed to extract meaningful information from unstructured text data which constitutes a large portion of big data. Over the past decades, text mining technologies have been utilized in various industries for practical applications. In the field of business intelligence, it has been employed to discover new market and/or technology opportunities and support rational decision making of business participants. The market information such as market size, market growth rate, and market share is essential for setting companies' business strategies. There has been a continuous demand in various fields for specific product level-market information. However, the information has been generally provided at industry level or broad categories based on classification standards, making it difficult to obtain specific and proper information. In this regard, we propose a new methodology that can estimate the market sizes of product groups at more detailed levels than that of previously offered. We applied Word2Vec algorithm, a neural network based semantic word embedding model, to enable automatic market size estimation from individual companies' product information in a bottom-up manner. The overall process is as follows: First, the data related to product information is collected, refined, and restructured into suitable form for applying Word2Vec model. Next, the preprocessed data is embedded into vector space by Word2Vec and then the product groups are derived by extracting similar products names based on cosine similarity calculation. Finally, the sales data on the extracted products is summated to estimate the market size of the product groups. As an experimental data, text data of product names from Statistics Korea's microdata (345,103 cases) were mapped in multidimensional vector space by Word2Vec training. We performed parameters optimization for training and then applied vector dimension of 300 and window size of 15 as optimized parameters for further experiments. We employed index words of Korean Standard Industry Classification (KSIC) as a product name dataset to more efficiently cluster product groups. The product names which are similar to KSIC indexes were extracted based on cosine similarity. The market size of extracted products as one product category was calculated from individual companies' sales data. The market sizes of 11,654 specific product lines were automatically estimated by the proposed model. For the performance verification, the results were compared with actual market size of some items. The Pearson's correlation coefficient was 0.513. Our approach has several advantages differing from the previous studies. First, text mining and machine learning techniques were applied for the first time on market size estimation, overcoming the limitations of traditional sampling based- or multiple assumption required-methods. In addition, the level of market category can be easily and efficiently adjusted according to the purpose of information use by changing cosine similarity threshold. Furthermore, it has a high potential of practical applications since it can resolve unmet needs for detailed market size information in public and private sectors. Specifically, it can be utilized in technology evaluation and technology commercialization support program conducted by governmental institutions, as well as business strategies consulting and market analysis report publishing by private firms. The limitation of our study is that the presented model needs to be improved in terms of accuracy and reliability. The semantic-based word embedding module can be advanced by giving a proper order in the preprocessed dataset or by combining another algorithm such as Jaccard similarity with Word2Vec. Also, the methods of product group clustering can be changed to other types of unsupervised machine learning algorithm. Our group is currently working on subsequent studies and we expect that it can further improve the performance of the conceptually proposed basic model in this study.

Rapid Detection Method for Human Rotavirus from Vegetables by a Combination of Filtration and Integrated Cell Culture/Real-Time Reverse Transcription PCR (Filtration과 Integrated Cell Culture/Real-Time Reverse Transcription PCR 기법을 이용한 채소류에서 Human Rotavirus 신속 검출)

  • Hyeon, Ji-Yeon;Chon, Jung-Whan;Song, Kwang-Young;Hwang, In-Gyun;Kwak, Hyo-Sun;Lee, Jung-Soo;Kim, Moo-Sang;Lee, Jung-Bok;Seo, Kun-Ho
    • Korean Journal of Microbiology
    • /
    • v.47 no.2
    • /
    • pp.117-123
    • /
    • 2011
  • The purpose of this study was to evaluate and compare different elution and concentration methods for optimization of human rotavirus (HRV) detection method using real-time RT-PCR and cell culture techniques. The leafy vegetable samples (lettuce, Chinese cabbage) were artificially inoculated with HRV. Viruses were extracted from the vegetables by two different elution buffers, buffer A (100 mM Tris-HCl, 50 mM glycine, 3% beef extract, pH 9.5) and buffer B (250 mM Threonine, 300 mM NaCl, pH 9.5), and the extracted viruses were concentrated by filtration and PEG precipitation sequentially. To determine infectivity of the viruses, the viruses recovered from the samples were infected to the MA-104 cells, and integrated cell culture real-time RT-PCR was performed at 1, 48, 72, 96, 120, 144, 168 h post-infection (p.i.). The elution buffer A was more efficient in extracting the virus from the produce samples tested than the buffer B, 29.54% and 18.32% of recoveries, respectively. The sensitivity of real-time RT-PCR method was markedly improved when the virus was concentrated by the filtration method. When the viruses were eluted and concentrated by buffer A and filtration, respectively, the average recovery rate was approximately 51.89%. When the viruses recovered from samples were infected to MA-104 cell, infectious HRV was detected within 48 h p.i. by ICC/real-time RT-PCR, whereas cytopathic effects were not observed until 72 h p.i. The optimized detection method evaluated in this study could be useful for rapid and reliable detection of HRV in fresh produce products and applied for detection of other food-borne viruses.

Bankruptcy prediction using an improved bagging ensemble (개선된 배깅 앙상블을 활용한 기업부도예측)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.121-139
    • /
    • 2014
  • Predicting corporate failure has been an important topic in accounting and finance. The costs associated with bankruptcy are high, so the accuracy of bankruptcy prediction is greatly important for financial institutions. Lots of researchers have dealt with the topic associated with bankruptcy prediction in the past three decades. The current research attempts to use ensemble models for improving the performance of bankruptcy prediction. Ensemble classification is to combine individually trained classifiers in order to gain more accurate prediction than individual models. Ensemble techniques are shown to be very useful for improving the generalization ability of the classifier. Bagging is the most commonly used methods for constructing ensemble classifiers. In bagging, the different training data subsets are randomly drawn with replacement from the original training dataset. Base classifiers are trained on the different bootstrap samples. Instance selection is to select critical instances while deleting and removing irrelevant and harmful instances from the original set. Instance selection and bagging are quite well known in data mining. However, few studies have dealt with the integration of instance selection and bagging. This study proposes an improved bagging ensemble based on instance selection using genetic algorithms (GA) for improving the performance of SVM. GA is an efficient optimization procedure based on the theory of natural selection and evolution. GA uses the idea of survival of the fittest by progressively accepting better solutions to the problems. GA searches by maintaining a population of solutions from which better solutions are created rather than making incremental changes to a single solution to the problem. The initial solution population is generated randomly and evolves into the next generation by genetic operators such as selection, crossover and mutation. The solutions coded by strings are evaluated by the fitness function. The proposed model consists of two phases: GA based Instance Selection and Instance based Bagging. In the first phase, GA is used to select optimal instance subset that is used as input data of bagging model. In this study, the chromosome is encoded as a form of binary string for the instance subset. In this phase, the population size was set to 100 while maximum number of generations was set to 150. We set the crossover rate and mutation rate to 0.7 and 0.1 respectively. We used the prediction accuracy of model as the fitness function of GA. SVM model is trained on training data set using the selected instance subset. The prediction accuracy of SVM model over test data set is used as fitness value in order to avoid overfitting. In the second phase, we used the optimal instance subset selected in the first phase as input data of bagging model. We used SVM model as base classifier for bagging ensemble. The majority voting scheme was used as a combining method in this study. This study applies the proposed model to the bankruptcy prediction problem using a real data set from Korean companies. The research data used in this study contains 1832 externally non-audited firms which filed for bankruptcy (916 cases) and non-bankruptcy (916 cases). Financial ratios categorized as stability, profitability, growth, activity and cash flow were investigated through literature review and basic statistical methods and we selected 8 financial ratios as the final input variables. We separated the whole data into three subsets as training, test and validation data set. In this study, we compared the proposed model with several comparative models including the simple individual SVM model, the simple bagging model and the instance selection based SVM model. The McNemar tests were used to examine whether the proposed model significantly outperforms the other models. The experimental results show that the proposed model outperforms the other models.

Low temperature plasma deposition of microcrystalline silicon thin films for active matrix displays: opportunities and challenges

  • Cabarrocas, Pere Roca I;Abramov, Alexey;Pham, Nans;Djeridane, Yassine;Moustapha, Oumkelthoum;Bonnassieux, Yvan;Girotra, Kunal;Chen, Hong;Park, Seung-Kyu;Park, Kyong-Tae;Huh, Jong-Moo;Choi, Joon-Hoo;Kim, Chi-Woo;Lee, Jin-Seok;Souk, Jun-H.
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2008.10a
    • /
    • pp.107-108
    • /
    • 2008
  • The spectacular development of AMLCDs, been made possible by a-Si:H technology, still faces two major drawbacks due to the intrinsic structure of a-Si:H, namely a low mobility and most important a shift of the transfer characteristics of the TFTs when submitted to bias stress. This has lead to strong research in the crystallization of a-Si:H films by laser and furnace annealing to produce polycrystalline silicon TFTs. While these devices show improved mobility and stability, they suffer from uniformity over large areas and increased cost. In the last decade we have focused on microcrystalline silicon (${\mu}c$-Si:H) for bottom gate TFTs, which can hopefully meet all the requirements for mass production of large area AMOLED displays [1,2]. In this presentation we will focus on the transfer of a deposition process based on the use of $SiF_4$-Ar-$H_2$ mixtures from a small area research laboratory reactor into an industrial gen 1 AKT reactor. We will first discuss on the optimization of the process conditions leading to fully crystallized films without any amorphous incubation layer, suitable for bottom gate TFTS, as well as on the use of plasma diagnostics to increase the deposition rate up to 0.5 nm/s [3]. The use of silicon nanocrystals appears as an elegant way to circumvent the opposite requirements of a high deposition rate and a fully crystallized interface [4]. The optimized process conditions are transferred to large area substrates in an industrial environment, on which some process adjustment was required to reproduce the material properties achieved in the laboratory scale reactor. For optimized process conditions, the homogeneity of the optical and electronic properties of the ${\mu}c$-Si:H films deposited on $300{\times}400\;mm$ substrates was checked by a set of complementary techniques. Spectroscopic ellipsometry, Raman spectroscopy, dark conductivity, time resolved microwave conductivity and hydrogen evolution measurements allowed demonstrating an excellent homogeneity in the structure and transport properties of the films. On the basis of these results, optimized process conditions were applied to TFTs, for which both bottom gate and top gate structures were studied aiming to achieve characteristics suitable for driving AMOLED displays. Results on the homogeneity of the TFT characteristics over the large area substrates and stability will be presented, as well as their application as a backplane for an AMOLED display.

  • PDF