• Title/Summary/Keyword: module performance

Search Result 2,736, Processing Time 0.037 seconds

Selection Model of System Trading Strategies using SVM (SVM을 이용한 시스템트레이딩전략의 선택모형)

  • Park, Sungcheol;Kim, Sun Woong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.59-71
    • /
    • 2014
  • System trading is becoming more popular among Korean traders recently. System traders use automatic order systems based on the system generated buy and sell signals. These signals are generated from the predetermined entry and exit rules that were coded by system traders. Most researches on system trading have focused on designing profitable entry and exit rules using technical indicators. However, market conditions, strategy characteristics, and money management also have influences on the profitability of the system trading. Unexpected price deviations from the predetermined trading rules can incur large losses to system traders. Therefore, most professional traders use strategy portfolios rather than only one strategy. Building a good strategy portfolio is important because trading performance depends on strategy portfolios. Despite of the importance of designing strategy portfolio, rule of thumb methods have been used to select trading strategies. In this study, we propose a SVM-based strategy portfolio management system. SVM were introduced by Vapnik and is known to be effective for data mining area. It can build good portfolios within a very short period of time. Since SVM minimizes structural risks, it is best suitable for the futures trading market in which prices do not move exactly the same as the past. Our system trading strategies include moving-average cross system, MACD cross system, trend-following system, buy dips and sell rallies system, DMI system, Keltner channel system, Bollinger Bands system, and Fibonacci system. These strategies are well known and frequently being used by many professional traders. We program these strategies for generating automated system signals for entry and exit. We propose SVM-based strategies selection system and portfolio construction and order routing system. Strategies selection system is a portfolio training system. It generates training data and makes SVM model using optimal portfolio. We make $m{\times}n$ data matrix by dividing KOSPI 200 index futures data with a same period. Optimal strategy portfolio is derived from analyzing each strategy performance. SVM model is generated based on this data and optimal strategy portfolio. We use 80% of the data for training and the remaining 20% is used for testing the strategy. For training, we select two strategies which show the highest profit in the next day. Selection method 1 selects two strategies and method 2 selects maximum two strategies which show profit more than 0.1 point. We use one-against-all method which has fast processing time. We analyse the daily data of KOSPI 200 index futures contracts from January 1990 to November 2011. Price change rates for 50 days are used as SVM input data. The training period is from January 1990 to March 2007 and the test period is from March 2007 to November 2011. We suggest three benchmark strategies portfolio. BM1 holds two contracts of KOSPI 200 index futures for testing period. BM2 is constructed as two strategies which show the largest cumulative profit during 30 days before testing starts. BM3 has two strategies which show best profits during testing period. Trading cost include brokerage commission cost and slippage cost. The proposed strategy portfolio management system shows profit more than double of the benchmark portfolios. BM1 shows 103.44 point profit, BM2 shows 488.61 point profit, and BM3 shows 502.41 point profit after deducting trading cost. The best benchmark is the portfolio of the two best profit strategies during the test period. The proposed system 1 shows 706.22 point profit and proposed system 2 shows 768.95 point profit after deducting trading cost. The equity curves for the entire period show stable pattern. With higher profit, this suggests a good trading direction for system traders. We can make more stable and more profitable portfolios if we add money management module to the system.

Real data-based active sonar signal synthesis method (실데이터 기반 능동 소나 신호 합성 방법론)

  • Yunsu Kim;Juho Kim;Jongwon Seok;Jungpyo Hong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.43 no.1
    • /
    • pp.9-18
    • /
    • 2024
  • The importance of active sonar systems is emerging due to the quietness of underwater targets and the increase in ambient noise due to the increase in maritime traffic. However, the low signal-to-noise ratio of the echo signal due to multipath propagation of the signal, various clutter, ambient noise and reverberation makes it difficult to identify underwater targets using active sonar. Attempts have been made to apply data-based methods such as machine learning or deep learning to improve the performance of underwater target recognition systems, but it is difficult to collect enough data for training due to the nature of sonar datasets. Methods based on mathematical modeling have been mainly used to compensate for insufficient active sonar data. However, methodologies based on mathematical modeling have limitations in accurately simulating complex underwater phenomena. Therefore, in this paper, we propose a sonar signal synthesis method based on a deep neural network. In order to apply the neural network model to the field of sonar signal synthesis, the proposed method appropriately corrects the attention-based encoder and decoder to the sonar signal, which is the main module of the Tacotron model mainly used in the field of speech synthesis. It is possible to synthesize a signal more similar to the actual signal by training the proposed model using the dataset collected by arranging a simulated target in an actual marine environment. In order to verify the performance of the proposed method, Perceptual evaluation of audio quality test was conducted and within score difference -2.3 was shown compared to actual signal in a total of four different environments. These results prove that the active sonar signal generated by the proposed method approximates the actual signal.

A Study on Market Size Estimation Method by Product Group Using Word2Vec Algorithm (Word2Vec을 활용한 제품군별 시장규모 추정 방법에 관한 연구)

  • Jung, Ye Lim;Kim, Ji Hui;Yoo, Hyoung Sun
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.1-21
    • /
    • 2020
  • With the rapid development of artificial intelligence technology, various techniques have been developed to extract meaningful information from unstructured text data which constitutes a large portion of big data. Over the past decades, text mining technologies have been utilized in various industries for practical applications. In the field of business intelligence, it has been employed to discover new market and/or technology opportunities and support rational decision making of business participants. The market information such as market size, market growth rate, and market share is essential for setting companies' business strategies. There has been a continuous demand in various fields for specific product level-market information. However, the information has been generally provided at industry level or broad categories based on classification standards, making it difficult to obtain specific and proper information. In this regard, we propose a new methodology that can estimate the market sizes of product groups at more detailed levels than that of previously offered. We applied Word2Vec algorithm, a neural network based semantic word embedding model, to enable automatic market size estimation from individual companies' product information in a bottom-up manner. The overall process is as follows: First, the data related to product information is collected, refined, and restructured into suitable form for applying Word2Vec model. Next, the preprocessed data is embedded into vector space by Word2Vec and then the product groups are derived by extracting similar products names based on cosine similarity calculation. Finally, the sales data on the extracted products is summated to estimate the market size of the product groups. As an experimental data, text data of product names from Statistics Korea's microdata (345,103 cases) were mapped in multidimensional vector space by Word2Vec training. We performed parameters optimization for training and then applied vector dimension of 300 and window size of 15 as optimized parameters for further experiments. We employed index words of Korean Standard Industry Classification (KSIC) as a product name dataset to more efficiently cluster product groups. The product names which are similar to KSIC indexes were extracted based on cosine similarity. The market size of extracted products as one product category was calculated from individual companies' sales data. The market sizes of 11,654 specific product lines were automatically estimated by the proposed model. For the performance verification, the results were compared with actual market size of some items. The Pearson's correlation coefficient was 0.513. Our approach has several advantages differing from the previous studies. First, text mining and machine learning techniques were applied for the first time on market size estimation, overcoming the limitations of traditional sampling based- or multiple assumption required-methods. In addition, the level of market category can be easily and efficiently adjusted according to the purpose of information use by changing cosine similarity threshold. Furthermore, it has a high potential of practical applications since it can resolve unmet needs for detailed market size information in public and private sectors. Specifically, it can be utilized in technology evaluation and technology commercialization support program conducted by governmental institutions, as well as business strategies consulting and market analysis report publishing by private firms. The limitation of our study is that the presented model needs to be improved in terms of accuracy and reliability. The semantic-based word embedding module can be advanced by giving a proper order in the preprocessed dataset or by combining another algorithm such as Jaccard similarity with Word2Vec. Also, the methods of product group clustering can be changed to other types of unsupervised machine learning algorithm. Our group is currently working on subsequent studies and we expect that it can further improve the performance of the conceptually proposed basic model in this study.

A Fast Processor Architecture and 2-D Data Scheduling Method to Implement the Lifting Scheme 2-D Discrete Wavelet Transform (리프팅 스킴의 2차원 이산 웨이브릿 변환 하드웨어 구현을 위한 고속 프로세서 구조 및 2차원 데이터 스케줄링 방법)

  • Kim Jong Woog;Chong Jong Wha
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.42 no.4 s.334
    • /
    • pp.19-28
    • /
    • 2005
  • In this paper, we proposed a parallel fast 2-D discrete wavelet transform hardware architecture based on lifting scheme. The proposed architecture improved the 2-D processing speed, and reduced internal memory buffer size. The previous lifting scheme based parallel 2-D wavelet transform architectures were consisted with row direction and column direction modules, which were pair of prediction and update filter module. In 2-D wavelet transform, column direction processing used the row direction results, which were not generated in column direction order but in row direction order, so most hardware architecture need internal buffer memory. The proposed architecture focused on the reducing of the internal memory buffer size and the total calculation time. Reducing the total calculation time, we proposed a 4-way data flow scheduling and memory based parallel hardware architecture. The 4-way data flow scheduling can increase the row direction parallel performance, and reduced the initial latency of starting of the row direction calculation. In this hardware architecture, the internal buffer memory didn't used to store the results of the row direction calculation, while it contained intermediate values of column direction calculation. This method is very effective in column direction processing, because the input data of column direction were not generated in column direction order The proposed architecture was implemented with VHDL and Altera Stratix device. The implementation results showed overall calculation time reduced from $N^2/2+\alpha$ to $N^2/4+\beta$, and internal buffer memory size reduced by around $50\%$ of previous works.

Evaluation and Comparison of Effects of Air and Tomato Leaf Temperatures on the Population Dynamics of Greenhouse Whitefly (Trialeurodes vaporariorum) in Cherry Tomato Grown in Greenhouses (시설내 대기 온도와 방울토마토 잎 온도가 온실가루이(Trialeurodes vaporariorum)개체군 발달에 미치는 영향 비교)

  • Park, Jung-Joon;Park, Kuen-Woo;Shin, Key-Il;Cho, Ki-Jong
    • Horticultural Science & Technology
    • /
    • v.29 no.5
    • /
    • pp.420-432
    • /
    • 2011
  • Population dynamics of greenhouse whitefly, Trialeurodes vaporariorum (Westwood), were modeled and simulated to compare the temperature effects of air and tomato leaf inside greenhouse using DYMEX model simulator (pre-programed module based simulation program developed by CSIRO, Australia). The DYMEX model simulator consisted of temperature dependent development and oviposition modules. The normalized cumulative frequency distributions of the developmental period for immature and oviposition frequency rate and survival rate for adult of greenhouse whitefly were fitted to two-parameter Weibull function. Leaf temperature on reversed side of cherry tomato leafs (Lycopersicon esculentum cv. Koko) was monitored according to three tomato plant positions (top, > 1.6 m above the ground level; middle, 0.9 - 1.2 m; bottom, 0.3 - 0.5 m) using an infrared temperature gun. Air temperature was monitored at same three positions using a Hobo self-contained temperature logger. The leaf temperatures from three plant positions were described as a function of the air temperatures with 3-parameter exponential and sigmoidal models. Data sets of observed air temperature and predicted leaf temperatures were prepared, and incorporated into the DYMEX simulator to compare the effects of air and leaf temperature on population dynamics of greenhouse whitefly. The number of greenhouse whitefly immatures was counted by visual inspection in three tomato plant positions to verify the performance of DYMEX simulation in cherry tomato greenhouse where air and leaf temperatures were monitored. The egg stage of greenhouse whitefly was not counted due to its small size. A significant positive correlation between the observed and the predicted numbers of immature and adults were found when the leaf temperatures were incorporated into DYMEX simulation, but no significant correlation was observed with the air temperatures. This study demonstrated that the population dynamics of greenhouse whitefly was affected greatly by the leaf temperatures, rather than air temperatures, and thus the leaf surface temperature should be considered for management of greenhouse whitefly in cherry tomato grown in greenhouses.

Development of NCS Based Vocational Curriculum Model for the Practical and Creative Human Respirces (실전 창의형 인재 양성을 위한 NCS 기반 직업교육과정의 모형 개발)

  • Kim, Dong-Yeon;Kim, Jinsoo
    • 대한공업교육학회지
    • /
    • v.39 no.2
    • /
    • pp.101-121
    • /
    • 2014
  • The study aims to develop the NCS based vocational curriculum model for the practical and creative human resources. For effectiveness of the study, the study consists of literature studies of both domestic and international, contents analysis, case study, expert(9samples) consultation and review, and in-depth-interview of the three advisory members. The validity of the developed model is analyzed through mean, standard deviation and contents validity ratio(CVR). The main results of the model development in our study are as follow. First, our NCS based vocational curriculum model for the practical and creative human resources is developed with the analyses of NCS development manuals, training standard utilization and training curriculum organization manuals, NCS learning module development manual and case studies, NCS research report, NCS based curriculum pilot development resources directed toward the high schools and vocational school as well as the domestic and international literature study on career training model like NCS. Second, based on the findings of our analysis in combination with the findings from the consultations with the expert and advisory committee, total 19 sub-factors of each step and domain are extracted. The sub-factors of domain in step 1 are the competency unit, definition of competency unit, competency unit element, performance criteria, range of variable, guide of assessment, key competency; in step 2, they are subject title, subject objectives, chapter title, chapter objectives, pedagogical methods, assessment methods and basic job competence; and in step 2, they are NCS based subject matrix table, NCS based subject profile, NCS based job training curriculum table, NCS based subjects organization flowchart, NCS based job training operation plan. Third, the final model including step 3 NCS based subject profile are developed in association with the linked organizational sub-factors of step 1 and step 2. Forth, the validity tests for the final model by the step and domain yield the mean 4.67, CVR value 1.00, indicating the superior validity. Also, the means of each sub-factors are all over 4.33 with the CVR value 1.00, indicating the high validity as well. The means of the associated organizations within the model are also over 4.33 with the CVR value of 1.00. Standard deviations are all .50 or lower which are small. Fifth, based on the validity test results and the in-depth-interview of the expert and advisory committee, the model is adjusted complemented to establish final model of the NCS based vocational curriculum for the practical and creative human resources.

Power Generating Performance of Photovoltaic Power System for Greenhouse Equipment Operation (온실설비 작동용 태양광발전시스템의 발전 성능 분석)

  • Yoon, Yong-Cheol;Bae, Yong-Han;Ryou, Young-Sun;Lee, Sung-Hyoun;Suh, Won-Myung
    • Journal of Bio-Environment Control
    • /
    • v.18 no.3
    • /
    • pp.177-184
    • /
    • 2009
  • For the purpose of reducing the cost for greenhouse operation by acquiring the electric power necessary for it, this study installed a solar photovoltaic system on the roof of the building adjacent to green-houses and experimentally examined the quantity of power generation based on weather conditions. The results of the study are as per the below: The maximum, average and minimum temperature while the experiment was conducted was $0.4{\sim}34.1,\;-6.1{\sim}22.2$, and $-14.1{\sim}16.7^{\circ}C$ respectively, and the solar radiation was $28.8MJ{\cdot}m^{-2}$ (maximum), $14.9MJ{\cdot}m^{-2}$ (average), and $0.6MJ{\cdot}m^{-2}$ (minimum). The quantity of electric power didn't increase in proportion to the quantity of solar radiation and instead, it was almost consistent around 750W. Daily maximum, average and minimum consumption of electric power was 5.2kWh, 2.5kWh and 0kWh respectively. Based on the average electric power consumption of the system used for this experiment, it was sufficient in case the capacity and the working time of a hot blast heater are small, but it was short in case they are big. In case the capacity of the hot blast heater is big, the average electric power quantity will be sufficient for array area $21m^2$, about three times of the present area. In summer when the temperature of the array becomes high, the generation of electric power didn't increase in proportion to the quantity of solar radiation, but this experiment result shows a high correlation between two factors (coefficient of correlation 0.84).

Modeling and Validation of Population Dynamics of the American Serpentine Leafminer (Liriomyza trifolii) Using Leaf Surface Temperatures of Greenhouses Cherry Tomatoes (방울토마토에서 잎 표면온도를 적용한 아메리카잎굴파리(Liriomyza trifolii) 개체군 밀도변동 모형작성 및 평가)

  • Park, Jung-Joon;Mo, Hyoung-Ho;Lee, Doo-Hyung;Shin, Key-Il;Cho, Ki-Jong
    • Korean journal of applied entomology
    • /
    • v.51 no.3
    • /
    • pp.235-243
    • /
    • 2012
  • Population dynamics of the American serpentine leafminer, Liriomyza trifolii (Burgess), were observed and modeled in order to compare the effects of air and tomato leaf temperatures inside a greenhouse using DYMEX model builder and simulator (pre-programed module based simulation programs developed by CSIRO, Australia). The DYMEX model simulator consisted of a series of modules with the parameters of temperature dependent development and oviposition models of L. trifolii were incorporated from pre-published data. Leaf surface temperatures of cherry tomato leaves (cv. 'Koko') were monitored according to three tomato plant positions (top, > 1.8 m above the ground level; middle, 0.9 - 1.2 m; bottom, 0.3 - 0.5 m) using an infrared temperature gun. Air temperature was monitored at the same three positions using a self-contained temperature logger. Data sets for the observed air temperature and average leaf surface temperatures were collected (top and bottom surfaces), and incorporated into the DYMEX simulator in order to compare the effects of air and leaf surface temperature on the population dynamics of L. trifolii. The initial population consisted of 50 eggs, which were laid by five female L. trifolii in early June. The number of L. trifolii larvae was counted by visual inspection of the tomato plants in order to verify the performance of DYMEX simulation. The egg, pupa, and adult stage of L. trifolii could not be counted due to its infeasible of visual inspection. A significant positive correlation between the observed and the predicted numbers of larvae was found when the leaf surface temperatures were incorporated into the DYMEX simulation (r = 0.97, p < 0.01), but no significant positive correlation was observed with air temperatures(r = 0.40, p = 0.18). This study demonstrated that the population dynamics of L. trifolii was affected greatly by the leaf temperatures, though to little discernible degree by the air temperatures, and thus the leaf surface temperature should be for a consideration in the management of L. trifolii within cherry tomato greenhouses.

A Quantification Method for the Cold Pool Effect on Nocturnal Temperature in a Closed Catchment (폐쇄집수역의 냉기호 모의를 통한 일 최저기온 분포 추정)

  • Kim, Soo-Ock;Yun, Jin-I.
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.13 no.4
    • /
    • pp.176-184
    • /
    • 2011
  • Cold air on sloping surfaces flows down to the valley bottom in mountainous terrain at calm and clear nights. Based on the assumption that the cold air flow may be the same as the water flow, current models estimate temperature drop by regarding the cold air accumulation at a given location as the water-like free drainage. At a closed catchment whose outlet is blocked by man-made obstacles such as banks and roads, however, the water-like free drainage assumption is no longer valid because the cold air accumulates from the bottom first. We developed an empirical model to estimate quantitatively the effect of cold pool on nocturnal temperature in a closed catchment. In our model, a closed catchment is treated like a "vessel", and a digital elevation model (DEM) was used to calculate the maximum capacity of the cold pool formed in a closed catchment. We introduce a topographical variable named "shape factor", which is the ratio of the cold air accumulation potential across the whole catchment area to the maximum capacity of the cold pool to describe the relative size of temperature drop at a wider range of catchment shapes. The shape factor is then used to simulate the density profile of cold pool formed in a given catchment based on a hypsometric equation. The cold lake module was incorporated with the existing model (i.e., Chung et al., 2006), generating a new model and predicting distribution of minimum temperature over closed catchments. We applied this model to Akyang valley (i.e., a typical closed catchment of 53 $km^2$ area) in the southern skirt of Mt. Jiri National Park where 12 automated weather stations (AWS) are operational. The performance of the model was evaluated based on the feasibility of delineating the temperature pattern accurately at cold pool forming at night. Overall, the model's ability of simulating the spatial pattern of lower temperature were improved especially at the valley bottom, showing a similar pattern of the estimated temperature with that of thermal images obtained across the valley at dawn (0520 to 0600 local standard time) of 17 May 2011. Error in temperature estimation, calculated with the root mean square error using the 10 low-lying AWSs, was substantially decreased from $1.30^{\circ}C$ with the existing model to $0.71^{\circ}C$ with the new model. These results suggest the feasibility of the new method in predicting the site-specific freeze and frost warning at a closed catchment.

A Methodology for Automatic Multi-Categorization of Single-Categorized Documents (단일 카테고리 문서의 다중 카테고리 자동확장 방법론)

  • Hong, Jin-Sung;Kim, Namgyu;Lee, Sangwon
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.77-92
    • /
    • 2014
  • Recently, numerous documents including unstructured data and text have been created due to the rapid increase in the usage of social media and the Internet. Each document is usually provided with a specific category for the convenience of the users. In the past, the categorization was performed manually. However, in the case of manual categorization, not only can the accuracy of the categorization be not guaranteed but the categorization also requires a large amount of time and huge costs. Many studies have been conducted towards the automatic creation of categories to solve the limitations of manual categorization. Unfortunately, most of these methods cannot be applied to categorizing complex documents with multiple topics because the methods work by assuming that one document can be categorized into one category only. In order to overcome this limitation, some studies have attempted to categorize each document into multiple categories. However, they are also limited in that their learning process involves training using a multi-categorized document set. These methods therefore cannot be applied to multi-categorization of most documents unless multi-categorized training sets are provided. To overcome the limitation of the requirement of a multi-categorized training set by traditional multi-categorization algorithms, we propose a new methodology that can extend a category of a single-categorized document to multiple categorizes by analyzing relationships among categories, topics, and documents. First, we attempt to find the relationship between documents and topics by using the result of topic analysis for single-categorized documents. Second, we construct a correspondence table between topics and categories by investigating the relationship between them. Finally, we calculate the matching scores for each document to multiple categories. The results imply that a document can be classified into a certain category if and only if the matching score is higher than the predefined threshold. For example, we can classify a certain document into three categories that have larger matching scores than the predefined threshold. The main contribution of our study is that our methodology can improve the applicability of traditional multi-category classifiers by generating multi-categorized documents from single-categorized documents. Additionally, we propose a module for verifying the accuracy of the proposed methodology. For performance evaluation, we performed intensive experiments with news articles. News articles are clearly categorized based on the theme, whereas the use of vulgar language and slang is smaller than other usual text document. We collected news articles from July 2012 to June 2013. The articles exhibit large variations in terms of the number of types of categories. This is because readers have different levels of interest in each category. Additionally, the result is also attributed to the differences in the frequency of the events in each category. In order to minimize the distortion of the result from the number of articles in different categories, we extracted 3,000 articles equally from each of the eight categories. Therefore, the total number of articles used in our experiments was 24,000. The eight categories were "IT Science," "Economy," "Society," "Life and Culture," "World," "Sports," "Entertainment," and "Politics." By using the news articles that we collected, we calculated the document/category correspondence scores by utilizing topic/category and document/topics correspondence scores. The document/category correspondence score can be said to indicate the degree of correspondence of each document to a certain category. As a result, we could present two additional categories for each of the 23,089 documents. Precision, recall, and F-score were revealed to be 0.605, 0.629, and 0.617 respectively when only the top 1 predicted category was evaluated, whereas they were revealed to be 0.838, 0.290, and 0.431 when the top 1 - 3 predicted categories were considered. It was very interesting to find a large variation between the scores of the eight categories on precision, recall, and F-score.