• Title/Summary/Keyword: 성능평가

Search Result 22,239, Processing Time 0.048 seconds

Target-Aspect-Sentiment Joint Detection with CNN Auxiliary Loss for Aspect-Based Sentiment Analysis (CNN 보조 손실을 이용한 차원 기반 감성 분석)

  • Jeon, Min Jin;Hwang, Ji Won;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.4
    • /
    • pp.1-22
    • /
    • 2021
  • Aspect Based Sentiment Analysis (ABSA), which analyzes sentiment based on aspects that appear in the text, is drawing attention because it can be used in various business industries. ABSA is a study that analyzes sentiment by aspects for multiple aspects that a text has. It is being studied in various forms depending on the purpose, such as analyzing all targets or just aspects and sentiments. Here, the aspect refers to the property of a target, and the target refers to the text that causes the sentiment. For example, for restaurant reviews, you could set the aspect into food taste, food price, quality of service, mood of the restaurant, etc. Also, if there is a review that says, "The pasta was delicious, but the salad was not," the words "steak" and "salad," which are directly mentioned in the sentence, become the "target." So far, in ABSA, most studies have analyzed sentiment only based on aspects or targets. However, even with the same aspects or targets, sentiment analysis may be inaccurate. Instances would be when aspects or sentiment are divided or when sentiment exists without a target. For example, sentences like, "Pizza and the salad were good, but the steak was disappointing." Although the aspect of this sentence is limited to "food," conflicting sentiments coexist. In addition, in the case of sentences such as "Shrimp was delicious, but the price was extravagant," although the target here is "shrimp," there are opposite sentiments coexisting that are dependent on the aspect. Finally, in sentences like "The food arrived too late and is cold now." there is no target (NULL), but it transmits a negative sentiment toward the aspect "service." Like this, failure to consider both aspects and targets - when sentiment or aspect is divided or when sentiment exists without a target - creates a dual dependency problem. To address this problem, this research analyzes sentiment by considering both aspects and targets (Target-Aspect-Sentiment Detection, hereby TASD). This study detected the limitations of existing research in the field of TASD: local contexts are not fully captured, and the number of epochs and batch size dramatically lowers the F1-score. The current model excels in spotting overall context and relations between each word. However, it struggles with phrases in the local context and is relatively slow when learning. Therefore, this study tries to improve the model's performance. To achieve the objective of this research, we additionally used auxiliary loss in aspect-sentiment classification by constructing CNN(Convolutional Neural Network) layers parallel to existing models. If existing models have analyzed aspect-sentiment through BERT encoding, Pooler, and Linear layers, this research added CNN layer-adaptive average pooling to existing models, and learning was progressed by adding additional loss values for aspect-sentiment to existing loss. In other words, when learning, the auxiliary loss, computed through CNN layers, allowed the local context to be captured more fitted. After learning, the model is designed to do aspect-sentiment analysis through the existing method. To evaluate the performance of this model, two datasets, SemEval-2015 task 12 and SemEval-2016 task 5, were used and the f1-score increased compared to the existing models. When the batch was 8 and epoch was 5, the difference was largest between the F1-score of existing models and this study with 29 and 45, respectively. Even when batch and epoch were adjusted, the F1-scores were higher than the existing models. It can be said that even when the batch and epoch numbers were small, they can be learned effectively compared to the existing models. Therefore, it can be useful in situations where resources are limited. Through this study, aspect-based sentiments can be more accurately analyzed. Through various uses in business, such as development or establishing marketing strategies, both consumers and sellers will be able to make efficient decisions. In addition, it is believed that the model can be fully learned and utilized by small businesses, those that do not have much data, given that they use a pre-training model and recorded a relatively high F1-score even with limited resources.

Tc-99m ECD Brain SPECT in MELAS Syndrome and Mitochondrial Myopathy: Comparison with MR findings (MELAS 증후군과 미토콘드리아 근육병에서의 Tc-99m ECD 뇌단일 광전자방출 전산화단층촬영 소견: 자기공명영상과의 비교)

  • Park, Sang-Joon;Ryu, Young-Hoon;Jeon, Tae-Joo;Kim, Jai-Keun;Nam, Ji-Eun;Yoon, Pyeong-Ho;Yoon, Choon-Sik;Lee, Jong-Doo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.32 no.6
    • /
    • pp.490-496
    • /
    • 1998
  • Purpose: We evaluated brain perfusion SPECT findings of MELAS syndrome and mitochondrial myopathy in correlation with MR imaging in search of specific imaging features. Materials and Methods: Subjects were five patients (four females and one male; age range, 1 to 25 year) who presented with repeated stroke-like episodes, seizures or developmental delay or asymptomatic but had elevated lactic acid in CSF and serum. Conventional non-contrast MR imaging and Tc-99m-ethyl cysteinate dimer (ECD) brain perfusion SPECT were Performed and imaging features were analyzed. Results: MRI demonstrated increased T2 signal intensities in the affected areas of gray and white matters mainly in the parietal (4/5) and occipital lobes (4/5) and in the basal ganglia (1/5), which were not restricted to a specific vascular territory. SPECT demonstrated decreased perfusion in the corresponding regions of MRI lesions. In addition, there were perfusion defects in parietal (1 patient), temporal (2), and frontal (1) lobes and basal ganglia (1) and thalami (2). In a patient with mitochondrial myopathy who had normal MRI, decreased perfusion was noted in left parietal area and bilateral thalami. Conclusion: Tc-99m ECD SPECT imaging in patients with MELAS syndrome and mitochondrial myopathy showed hypoperfusion of parieto-occipital cortex, basal ganglia, thalamus and temporal cortex, which were not restricted to a specific vascular territory. There were no specific imaging features on SPECT. The significance of abnormal perfusion on SPECT without corresponding MR abnormalities needs to be evaluated further in larger number of patients.

  • PDF

Study on PM10, PM2.5 Reduction Effects and Measurement Method of Vegetation Bio-Filters System in Multi-Use Facility (다중이용시설 내 식생바이오필터 시스템의 PM10, PM2.5 저감효과 및 측정방법에 대한 연구)

  • Kim, Tae-Han;Choi, Boo-Hun
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.48 no.5
    • /
    • pp.80-88
    • /
    • 2020
  • With the issuance of one-week fine dust emergency reduction measures in March 2019, the public's anxiety about fine dust is increasingly growing. In order to assess the application of air purifying plant-based bio-filters to public facilities, this study presented a method for measuring pollutant reduction effects by creating an indoor environment for continuous discharge of particle pollutants and conducted basic studies to verify whether indoor air quality has improved through the system. In this study conducted in a lecture room in spring, the background concentration was created by using mosquito repellent incense as a pollutant one hour before monitoring. Then, according to the schedule, the fine dust reduction capacity was monitored by irrigating for two hours and venting air for one hour. PM10, PM2.5, and temperature & humidity sensors were installed two meters front of the bio-filters, and velocity probes were installed at the center of the three air vents to conduct time-series monitoring. The average face velocity of three air vents set up in the bio-filter was 0.38±0.16 m/s. Total air-conditioning air volume was calculated at 776.89±320.16㎥/h by applying an air vent area of 0.29m×0.65m after deducing damper area. With the system in operation, average temperature and average relative humidity were maintained at 21.5-22.3℃, and 63.79-73.6%, respectively, which indicates that it satisfies temperature and humidity range of various conditions of preceding studies. When the effects of raising relatively humidity rapidly by operating system's air-conditioning function are used efficiently, it would be possible to reduce indoor fine dust and maintain appropriate relative humidity seasonally. Concentration of fine dust increased the same in all cycles before operating the bio-filter system. After operating the system, in cycle 1 blast section (C-1, β=-3.83, β=-2.45), particulate matters (PM10) were lowered by up to 28.8% or 560.3㎍/㎥ and fine particulate matters (PM2.5) were reduced by up to 28.0% or 350.0㎍/㎥. Then, the concentration of find dust (PM10, PM2.5) was reduced by up to 32.6% or 647.0㎍/㎥ and 32.4% or 401.3㎍/㎥ respectively through reduction in cycle 2 blast section (C-2, β=-5.50, β=-3.30) and up to 30.8% or 732.7㎍/㎥ and 31.0% or 459.3㎍/㎥ respectively through reduction in cycle 3 blast section (C-3, β=5.48, β=-3.51). By referring to standards and regulations related to the installation of vegetation bio-filters in public facilities, this study provided plans on how to set up objective performance evaluation environment. By doing so, it was possible to create monitoring infrastructure more objective than a regular lecture room environment and secure relatively reliable data.

A Comparative Study of the Standard Uptake Values of the PET Reconstruction Methods; Using Contrast Enhanced CT and Non Contrast Enhanced CT (PET/CT 영상에서 조영제를 사용하지 않은 CT와 조영제를 사용한 CT를 이용한 감쇠보정에 따른 표준화섭취계수의 비교)

  • Lee, Seung-Jae;Park, Hoon-Hee;Ahn, Sha-Ron;Oh, Shin-Hyun;NamKoong, Heuk;Lim, Han-Sang;Kim, Jae-Sam;Lee, Chang-Ho
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.12 no.3
    • /
    • pp.235-240
    • /
    • 2008
  • Purpose: At the beginning of PET/CT, Computed Tomography was mainly used only for Attenuation Correction (AC), but as the performance of the CT have been increase, it could give improved diagnostic information with Contrast Media. But it was controversial that Contrast Media could affect AC on PET/CT scan. Some submitted thesis' show that Contrast Media could overestimate when it is for AC data processing. On the contrary, the opinion that Contrast Media could be possible to affect the alteration of SUV because of the overestimated AC. But it does not have a definite effect on the diagnosis. Thus, the affection of Contrast Media on AC was investigated in this study. Materials and Methods: Patient inclusion criteria required a history of a malignancy and performance of an integrated PET/CT scan and contrast- enhanced CT scan within a 1-day period. Thirty oncologic patients who had PET/CT scan from December 2007 to June 2008 underwent staging evaluation and met these criteria. All patients fasted for at least 6 hr before the IV injection of approximately 5.6 MBq/kg (0.15 mCi/kg) of $^{18}F$-FDG and were scanned about 60 min after injection. All patients had a whole body PET/CT performed without IV contrast media followed by a contrast-enhanced CT on the Discovery STe PET/CT scanner. CT data were used for AC and PET images came out after AC. The ROIs drew and measured SUV. A paired t-test of these results was performed to assess the significance of the difference between the SUV obtained from the two attenuation corrected PET images. Results: The mean and maximum Standardized Uptake Values (SUV) for different regions averaged over all Patients. Comparing before using Contrast Media and after using, Most of ROIs have the increased SUV when it did Contrast Enhanced CT compare to Non-Contrast enhanced CT. All regions have increased SUV and also their p value was under 0.05 except the mean SUV of the Heart region. Conclusion: In this regard, the effect on SUV measurements that occurs when a contrast-enhanced CT is used for attenuation correction could have significant clinical ramifications. But some submitted thesis insisted that the percentage change in SUV that can determine or modify clinical management of oncology patients is small. Because there was not much difference that could be discovered by interpreter. But obviously the numerical change was occurred and on the stage finding primary region, small change would be base line, such as the region of liver which has greater change than the other regions needs more attention.

  • PDF

Scalable Collaborative Filtering Technique based on Adaptive Clustering (적응형 군집화 기반 확장 용이한 협업 필터링 기법)

  • Lee, O-Joun;Hong, Min-Sung;Lee, Won-Jin;Lee, Jae-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.73-92
    • /
    • 2014
  • An Adaptive Clustering-based Collaborative Filtering Technique was proposed to solve the fundamental problems of collaborative filtering, such as cold-start problems, scalability problems and data sparsity problems. Previous collaborative filtering techniques were carried out according to the recommendations based on the predicted preference of the user to a particular item using a similar item subset and a similar user subset composed based on the preference of users to items. For this reason, if the density of the user preference matrix is low, the reliability of the recommendation system will decrease rapidly. Therefore, the difficulty of creating a similar item subset and similar user subset will be increased. In addition, as the scale of service increases, the time needed to create a similar item subset and similar user subset increases geometrically, and the response time of the recommendation system is then increased. To solve these problems, this paper suggests a collaborative filtering technique that adapts a condition actively to the model and adopts the concepts of a context-based filtering technique. This technique consists of four major methodologies. First, items are made, the users are clustered according their feature vectors, and an inter-cluster preference between each item cluster and user cluster is then assumed. According to this method, the run-time for creating a similar item subset or user subset can be economized, the reliability of a recommendation system can be made higher than that using only the user preference information for creating a similar item subset or similar user subset, and the cold start problem can be partially solved. Second, recommendations are made using the prior composed item and user clusters and inter-cluster preference between each item cluster and user cluster. In this phase, a list of items is made for users by examining the item clusters in the order of the size of the inter-cluster preference of the user cluster, in which the user belongs, and selecting and ranking the items according to the predicted or recorded user preference information. Using this method, the creation of a recommendation model phase bears the highest load of the recommendation system, and it minimizes the load of the recommendation system in run-time. Therefore, the scalability problem and large scale recommendation system can be performed with collaborative filtering, which is highly reliable. Third, the missing user preference information is predicted using the item and user clusters. Using this method, the problem caused by the low density of the user preference matrix can be mitigated. Existing studies on this used an item-based prediction or user-based prediction. In this paper, Hao Ji's idea, which uses both an item-based prediction and user-based prediction, was improved. The reliability of the recommendation service can be improved by combining the predictive values of both techniques by applying the condition of the recommendation model. By predicting the user preference based on the item or user clusters, the time required to predict the user preference can be reduced, and missing user preference in run-time can be predicted. Fourth, the item and user feature vector can be made to learn the following input of the user feedback. This phase applied normalized user feedback to the item and user feature vector. This method can mitigate the problems caused by the use of the concepts of context-based filtering, such as the item and user feature vector based on the user profile and item properties. The problems with using the item and user feature vector are due to the limitation of quantifying the qualitative features of the items and users. Therefore, the elements of the user and item feature vectors are made to match one to one, and if user feedback to a particular item is obtained, it will be applied to the feature vector using the opposite one. Verification of this method was accomplished by comparing the performance with existing hybrid filtering techniques. Two methods were used for verification: MAE(Mean Absolute Error) and response time. Using MAE, this technique was confirmed to improve the reliability of the recommendation system. Using the response time, this technique was found to be suitable for a large scaled recommendation system. This paper suggested an Adaptive Clustering-based Collaborative Filtering Technique with high reliability and low time complexity, but it had some limitations. This technique focused on reducing the time complexity. Hence, an improvement in reliability was not expected. The next topic will be to improve this technique by rule-based filtering.

Comparative Uptake of Tc-99m Sestamibi and Tc-99m Tetrofosmin in Cancer Cells and Tissue Expressing P-Glycoprotein or Multidrug Resistance Associated Protein (P-Glycoprotein과 Multidrug Resistance Associated Protein을 발현하는 암세포와 종양에서 Tc-99m Sestamibi와 Tc-99m Tetrofosmin의 섭취율 비교)

  • Cho, Jung-Ah;Lee, Jae-Tae;Yoo, Jung-Ah;Seo, Ji-Hyoung;Bae, Jin-Ho;Jeong, Shin-Young;Ahn, Byeong-Cheol;Sohn, Sang-Gyun;Ha, Jeoung-Hee;Lee, Kyu-Bo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.39 no.1
    • /
    • pp.34-43
    • /
    • 2005
  • Purpose: $^{99m}Tc$-sestamibi(MIBI) and $^{99m}Tc$-tetrofosmin have been used as substrates for P-glycoprotein (Pgp) and multidrug resistance associated protein (MRP), which are closely associated with multidrug resistance of the tumors. To understand different handling of radiotracers in cancer cell lines expressing Pgp and MRP, we compared cellular uptakes of $^{99m}Tc$-MIBI and $^{99m}Tc$-tetrofosmin. The effects of cyclosporin A (CsA), well-known multidrug resistant reversing agent, on the uptake of both tracers were also compared. Materials and Methods: HCT15/CL02 human colorectal cancer cells for Pgp expressing cells, and human non-small cell lung cancer A549 cells for MRP expressing cells, were used for in vitro and in vivo studies. RT-PCR, western blot analysis and immunohistochemistry were used for detection of Pgp and MRP. MDR-reversal effect with CsA was evaluated at different drug concentrations after incubation with MIBI or tetrofosmin. Radioactivities of supernatant and pellet were measured with gamma well counter. Tumoral uptake of the tracers were measured from tumor bearing nude mice treated with or without CsA. Results: RT-PCR, western blot analysis of the cells and irnrnunochemical staining revealed selective expression of Pgp and MRP for HCY15/CL02 and A549 cells, respectively. There were no significant difference in cellular uptakes of both tracers in HCT15/CL02 cells, but MIBI uptake was slightly higher than that of tetrofosmin in A549 cells. Co-incubation with CsA resulted in a increase in cellular uptakes of MIBI and tetrofosmin. Uptake of MIBI or tetrofosmin in HCT15/CL02 cells was increased by 10- and 2.4-fold, and by 7.5 and 6.3-fold in A549 cells, respectively. Percentage increase of MIBI was higher than that of tetrofosmin with CsA for both cells (p<0.05). In vivo biodistribution study showed that MIBI (114% at 10 min, 257% at 60 min, 396% at 240 min) and tetrofosmin uptake (110% at 10 min, 205% at 60 min, 410% at 240 min) were progressively increased by the time, up to 240 min with CsA. But increases in tumoral uptake were not significantly different between MIBI and tetrofosmin for both tumors. Conclusion: MIBI seems to be a better tracer than tetrofosmin for evaluating MDR reversal effect of the modulators in vitro, but these differences were not evident in vivo tumoral uptake. Both MIBI and tetrofosmin seem to be suitable tracers for imaging Pgp- and MRP-mediated drug resistance in tumors.

An Analysis of Big Video Data with Cloud Computing in Ubiquitous City (클라우드 컴퓨팅을 이용한 유시티 비디오 빅데이터 분석)

  • Lee, Hak Geon;Yun, Chang Ho;Park, Jong Won;Lee, Yong Woo
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.45-52
    • /
    • 2014
  • The Ubiquitous-City (U-City) is a smart or intelligent city to satisfy human beings' desire to enjoy IT services with any device, anytime, anywhere. It is a future city model based on Internet of everything or things (IoE or IoT). It includes a lot of video cameras which are networked together. The networked video cameras support a lot of U-City services as one of the main input data together with sensors. They generate huge amount of video information, real big data for the U-City all the time. It is usually required that the U-City manipulates the big data in real-time. And it is not easy at all. Also, many times, it is required that the accumulated video data are analyzed to detect an event or find a figure among them. It requires a lot of computational power and usually takes a lot of time. Currently we can find researches which try to reduce the processing time of the big video data. Cloud computing can be a good solution to address this matter. There are many cloud computing methodologies which can be used to address the matter. MapReduce is an interesting and attractive methodology for it. It has many advantages and is getting popularity in many areas. Video cameras evolve day by day so that the resolution improves sharply. It leads to the exponential growth of the produced data by the networked video cameras. We are coping with real big data when we have to deal with video image data which are produced by the good quality video cameras. A video surveillance system was not useful until we find the cloud computing. But it is now being widely spread in U-Cities since we find some useful methodologies. Video data are unstructured data thus it is not easy to find a good research result of analyzing the data with MapReduce. This paper presents an analyzing system for the video surveillance system, which is a cloud-computing based video data management system. It is easy to deploy, flexible and reliable. It consists of the video manager, the video monitors, the storage for the video images, the storage client and streaming IN component. The "video monitor" for the video images consists of "video translater" and "protocol manager". The "storage" contains MapReduce analyzer. All components were designed according to the functional requirement of video surveillance system. The "streaming IN" component receives the video data from the networked video cameras and delivers them to the "storage client". It also manages the bottleneck of the network to smooth the data stream. The "storage client" receives the video data from the "streaming IN" component and stores them to the storage. It also helps other components to access the storage. The "video monitor" component transfers the video data by smoothly streaming and manages the protocol. The "video translator" sub-component enables users to manage the resolution, the codec and the frame rate of the video image. The "protocol" sub-component manages the Real Time Streaming Protocol (RTSP) and Real Time Messaging Protocol (RTMP). We use Hadoop Distributed File System(HDFS) for the storage of cloud computing. Hadoop stores the data in HDFS and provides the platform that can process data with simple MapReduce programming model. We suggest our own methodology to analyze the video images using MapReduce in this paper. That is, the workflow of video analysis is presented and detailed explanation is given in this paper. The performance evaluation was experiment and we found that our proposed system worked well. The performance evaluation results are presented in this paper with analysis. With our cluster system, we used compressed $1920{\times}1080(FHD)$ resolution video data, H.264 codec and HDFS as video storage. We measured the processing time according to the number of frame per mapper. Tracing the optimal splitting size of input data and the processing time according to the number of node, we found the linearity of the system performance.

The CH3CHO Removal Characteristics of Lightweight Aggregate Concrete with TiO2 Spreaded by Low Temperature Firing using Sol-gel Method (Sol-gel법으로 이산화티탄(TiO2)을 저온소성 도포시킨 경량골재콘크리트의 아세트알데히드(CH3CHO) 제거 특성)

  • Lee, Seung Han;Yeo, In Dong;Jung, Yong Wook;Jang, Suk Soo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.31 no.2A
    • /
    • pp.129-136
    • /
    • 2011
  • Recently studies on functional concrete with a photocatalytic material such as $TiO_2$ have actively been carried out in order to remove air pollutants. The absorbtion of $TiO_2$ from those studies is applied by it being directly mixed into concrete or by suspension coated on the surface. When it comes to the effectiveness, the former process is less than that of the latter compared with the $TiO_2$ use. As a result, the direct coating of $TiO_2$ on materials' surface is more used for effectiveness. The Surface spread of it needs to have a more than $400^{\circ}C$ heat treat done to stimulate the activation and adhesion of photocatalysis. Heat treat consequently leads hydration products in concrete to be dehydrated and shrunk and is the cause of cracking. The study produces $TiO_2$ used Sol-gel method which enables it to be coated with a low temperature treat, applies it to pearlite using Lightweight Aggregate Concrete fixed with a low temperature treat and evaluates the spread performance of it. In addition to this, the size of pearlite is divided into two types: One is 2.5 mm to 5.0 mm and the other is more than 5.0 mm for the benefit of finding out the removal characteristics of $CH_3CHO$ whether they are affected by pearlite size, mixing method and ratio with $TiO_2$ and elapsed time. The result of this experiment shows that although $TiO_2$ produced by Sol-gel method is treated with 120 temperature, it maintains a high spread rate on the XRF(X ray Florescence) quantitative analysis which ranks $TiO_2$ 38 percent, $SiO_2$ 29 percent and CaO 18 percent. In the size of perlite from 2.5 mm to 5.0 mm, the removal characteristic of $CH_3CHO$ from a low temperature heated Lightweight concrete appears 20 percent higher when $TiO_2$ with Sol-gel method is spreaded on the 7 percent of surface. In other words, the removal rate is 94 percent compared with the 72 percent where $TiO_2$ is mixed in 10 percent surface. In more than 5.0 mm sized perlite, the removal rate of $CH_3CHO$, when $TiO_2$ is mixed with 10 percent, is 69 percent, which is similar with that of the previous case. It suggests that the size of pearlite has little effects on the removal rate of $CH_3CHO$. In terms of Elapsed time, the removal characteristic seems apparent at the early stage, where the average removal rate for the first 10 hours takes up 84 percent compared with that of 20 hours.

Natural Language Processing Model for Data Visualization Interaction in Chatbot Environment (챗봇 환경에서 데이터 시각화 인터랙션을 위한 자연어처리 모델)

  • Oh, Sang Heon;Hur, Su Jin;Kim, Sung-Hee
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.9 no.11
    • /
    • pp.281-290
    • /
    • 2020
  • With the spread of smartphones, services that want to use personalized data are increasing. In particular, healthcare-related services deal with a variety of data, and data visualization techniques are used to effectively show this. As data visualization techniques are used, interactions in visualization are also naturally emphasized. In the PC environment, since the interaction for data visualization is performed with a mouse, various filtering for data is provided. On the other hand, in the case of interaction in a mobile environment, the screen size is small and it is difficult to recognize whether or not the interaction is possible, so that only limited visualization provided by the app can be provided through a button touch method. In order to overcome the limitation of interaction in such a mobile environment, we intend to enable data visualization interactions through conversations with chatbots so that users can check individual data through various visualizations. To do this, it is necessary to convert the user's query into a query and retrieve the result data through the converted query in the database that is storing data periodically. There are many studies currently being done to convert natural language into queries, but research on converting user queries into queries based on visualization has not been done yet. Therefore, in this paper, we will focus on query generation in a situation where a data visualization technique has been determined in advance. Supported interactions are filtering on task x-axis values and comparison between two groups. The test scenario utilized data on the number of steps, and filtering for the x-axis period was shown as a bar graph, and a comparison between the two groups was shown as a line graph. In order to develop a natural language processing model that can receive requested information through visualization, about 15,800 training data were collected through a survey of 1,000 people. As a result of algorithm development and performance evaluation, about 89% accuracy in classification model and 99% accuracy in query generation model was obtained.

Development of a Window Program for Searching CpG Island (CpG Island 검색용 윈도우 프로그램 개발)

  • Kim, Ki-Bong
    • Journal of Life Science
    • /
    • v.18 no.8
    • /
    • pp.1132-1139
    • /
    • 2008
  • A CpG island is a short stretch of DNA in which the frequency of the CG dinucleotide is higher than other regions. CpG islands are present in the promoters and exonic regions of approximately $30{\sim}60$% of mammalian genes so they are useful markers for genes in organisms containing 5-methylcytosine in their genomes. Recent evidence supports the notion that the hypermethylation of CpG island, by silencing tumor suppressor genes, plays a major causal role in cancer, which has been described in almost every tumor types. In this respect, CpG island search by computational methods is very helpful for cancer research and computational promoter and gene predictions. I therefore developed a window program (called CpGi) on the basis of CpG island criteria defined by D. Takai and P. A. Jones. The program 'CpGi' was implemented in Visual C++ 6.0 and can determine the locations of CpG islands using diverse parameters (%GC, Obs (CpG)/Exp (CpG), window size, step size, gap value, # of CpG, length) specified by user. The analysis result of CpGi provides a graphical map of CpG islands and G+C% plot, where more detailed information on CpG island can be obtained through pop-up window. Two human contigs, i.e. AP00524 (from chromosome 22) and NT_029490.3 (from chromosome 21), were used to compare the performance of CpGi and two other public programs for the accuracy of search results. The two other programs used in the performance comparison are Emboss-CpGPlot and CpG Island Searcher that are web-based public CpG island search programs. The comparison result showed that CpGi is on a level with or outperforms Emboss-CpGPlot and CpG Island Searcher. Having a simple and easy-to-use user interface, CpGi would be a very useful tool for genome analysis and CpG island research. To obtain a copy of CpGi for academic use only, contact corresponding author.