• 제목/요약/키워드: Diverse data sets

검색결과 79건 처리시간 0.026초

Deformable image registration in radiation therapy

  • Oh, Seungjong;Kim, Siyong
    • Radiation Oncology Journal
    • /
    • 제35권2호
    • /
    • pp.101-111
    • /
    • 2017
  • The number of imaging data sets has significantly increased during radiation treatment after introducing a diverse range of advanced techniques into the field of radiation oncology. As a consequence, there have been many studies proposing meaningful applications of imaging data set use. These applications commonly require a method to align the data sets at a reference. Deformable image registration (DIR) is a process which satisfies this requirement by locally registering image data sets into a reference image set. DIR identifies the spatial correspondence in order to minimize the differences between two or among multiple sets of images. This article describes clinical applications, validation, and algorithms of DIR techniques. Applications of DIR in radiation treatment include dose accumulation, mathematical modeling, automatic segmentation, and functional imaging. Validation methods discussed are based on anatomical landmarks, physical phantoms, digital phantoms, and per application purpose. DIR algorithms are also briefly reviewed with respect to two algorithmic components: similarity index and deformation models.

Video augmentation technique for human action recognition using genetic algorithm

  • Nida, Nudrat;Yousaf, Muhammad Haroon;Irtaza, Aun;Velastin, Sergio A.
    • ETRI Journal
    • /
    • 제44권2호
    • /
    • pp.327-338
    • /
    • 2022
  • Classification models for human action recognition require robust features and large training sets for good generalization. However, data augmentation methods are employed for imbalanced training sets to achieve higher accuracy. These samples generated using data augmentation only reflect existing samples within the training set, their feature representations are less diverse and hence, contribute to less precise classification. This paper presents new data augmentation and action representation approaches to grow training sets. The proposed approach is based on two fundamental concepts: virtual video generation for augmentation and representation of the action videos through robust features. Virtual videos are generated from the motion history templates of action videos, which are convolved using a convolutional neural network, to generate deep features. Furthermore, by observing an objective function of the genetic algorithm, the spatiotemporal features of different samples are combined, to generate the representations of the virtual videos and then classified through an extreme learning machine classifier on MuHAVi-Uncut, iXMAS, and IAVID-1 datasets.

최적 Type-2 퍼지신경회로망 설계와 응용 (The Design of Optimized Type-2 Fuzzy Neural Networks and Its Application)

  • 김길성;안인석;오성권
    • 전기학회논문지
    • /
    • 제58권8호
    • /
    • pp.1615-1623
    • /
    • 2009
  • In order to develop reliable on-site partial discharge (PD) pattern recognition algorithm, we introduce Type-2 Fuzzy Neural Networks (T2FNNs) optimized by means of Particle Swarm Optimization(PSO). T2FNNs exploit Type-2 fuzzy sets which have a characteristic of robustness in the diverse area of intelligence systems. Considering the on-site situation where it is not easy to obtain voltage phases to be used for PRPDA (Phase Resolved Partial Discharge Analysis), the PD data sets measured in the laboratory were artificially changed into data sets with shifted voltage phases and added noise in order to test the proposed algorithm. Also, the results obtained by the proposed algorithm were compared with that of conventional Neural Networks(NNs) as well as the existing Radial Basis Function Neural Networks (RBFNNs). The T2FNNs proposed in this study were appeared to have better performance when compared to conventional NNs and RBFNNs.

인공어초 어장에서 수록한 음향학적 다차원 데이터 해석을 위한 GIS의 응용 (The application of GIS in analyzing acoustical and multidimensional data related to artificial reefs ground)

  • 강명희;타케시 나카무라;아키라 하마노
    • 수산해양기술연구
    • /
    • 제47권3호
    • /
    • pp.222-233
    • /
    • 2011
  • This study is for the multi-dimensional analysis of diverse data sets for artificial reefs off the coast of Shimonoseki, Yamaguchi prefecture, Japan. Various data sets recorded in artificial reefs ground were integrated in new GIS software: to reveal the relationships between water temperature and fish schools; to visualize the quantitative connection between the reefs and the fish schools; and to compare the seabed types derived from two different data sources. The results obtained suggest that the application of GIS in analyzing multi-dimensional data is a better way to understand the characteristics of fish schools and environmental information around artificial reefs and particularly in the evaluation of the effectiveness of artificial reefs.

Significant Gene Selection Using Integrated Microarray Data Set with Batch Effect

  • Kim Ki-Yeol;Chung Hyun-Cheol;Jeung Hei-Cheul;Shin Ji-Hye;Kim Tae-Soo;Rha Sun-Young
    • Genomics & Informatics
    • /
    • 제4권3호
    • /
    • pp.110-117
    • /
    • 2006
  • In microarray technology, many diverse experimental features can cause biases including RNA sources, microarray production or different platforms, diverse sample processing and various experiment protocols. These systematic effects cause a substantial obstacle in the analysis of microarray data. When such data sets derived from different experimental processes were used, the analysis result was almost inconsistent and it is not reliable. Therefore, one of the most pressing challenges in the microarray field is how to combine data that comes from two different groups. As the novel trial to integrate two data sets with batch effect, we simply applied standardization to microarray data before the significant gene selection. In the gene selection step, we used new defined measure that considers the distance between a gene and an ideal gene as well as the between-slide and within-slide variations. Also we discussed the association of biological functions and different expression patterns in selected discriminative gene set. As a result, we could confirm that batch effect was minimized by standardization and the selected genes from the standardized data included various expression pattems and the significant biological functions.

Applications of Diverse Data Combinations in Subsurface Characterization using D-optimality Based Pilot Point Methods (DBM)

  • Jung, Yong;Mahinthakumar, G.
    • 한국지하수토양환경학회지:지하수토양환경
    • /
    • 제18권2호
    • /
    • pp.45-53
    • /
    • 2013
  • Many cases of strategically designed groundwater remediation have lack of information of hydraulic conductivity or permeability, which can render remediation methods inefficient. Many studies have been carried out to minimize this shortcoming by determining detailed hydraulic information either through direct or indirect measurements. One popular method for hydraulic characterization is the pilot point method (PPM), where the hydraulic property is estimated at a small number of strategically selected points using secondary measurements such as hydraulic head or tracer concentration. This paper adopted a D-optimality based pilot point method (DBM) developed previously for hydraulic head measurements and extended it to include both hydraulic head and tracer measurements. Based on different combinations of trials, our analysis showed that DBM performs well when hydraulic head is used for pilot point selection and both hydraulic head and tracer measurements are used for determining the conductivity values.

Vulnerability Threat Classification Based on XLNET AND ST5-XXL model

  • Chae-Rim Hong;Jin-Keun Hong
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제16권3호
    • /
    • pp.262-273
    • /
    • 2024
  • We provide a detailed analysis of the data processing and model training process for vulnerability classification using Transformer-based language models, especially sentence text-to-text transformers (ST5)-XXL and XLNet. The main purpose of this study is to compare the performance of the two models, identify the strengths and weaknesses of each, and determine the optimal learning rate to increase the efficiency and stability of model training. We performed data preprocessing, constructed and trained models, and evaluated performance based on data sets with various characteristics. We confirmed that the XLNet model showed excellent performance at learning rates of 1e-05 and 1e-04 and had a significantly lower loss value than the ST5-XXL model. This indicates that XLNet is more efficient for learning. Additionally, we confirmed in our study that learning rate has a significant impact on model performance. The results of the study highlight the usefulness of ST5-XXL and XLNet models in the task of classifying security vulnerabilities and highlight the importance of setting an appropriate learning rate. Future research should include more comprehensive analyzes using diverse data sets and additional models.

Reinforcement learning multi-agent using unsupervised learning in a distributed cloud environment

  • Gu, Seo-Yeon;Moon, Seok-Jae;Park, Byung-Joon
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제14권2호
    • /
    • pp.192-198
    • /
    • 2022
  • Companies are building and utilizing their own data analysis systems according to business characteristics in the distributed cloud. However, as businesses and data types become more complex and diverse, the demand for more efficient analytics has increased. In response to these demands, in this paper, we propose an unsupervised learning-based data analysis agent to which reinforcement learning is applied for effective data analysis. The proposal agent consists of reinforcement learning processing manager and unsupervised learning manager modules. These two modules configure an agent with k-means clustering on multiple nodes and then perform distributed training on multiple data sets. This enables data analysis in a relatively short time compared to conventional systems that perform analysis of large-scale data in one batch.

콘텐트 노드의 유사성 제어를 통한 그래프 구조 데이터 검색의 다양성 향상 (Improving Diversity of Keyword Search on Graph-structured Data by Controlling Similarity of Content Nodes)

  • 박창섭
    • 한국콘텐츠학회논문지
    • /
    • 제20권3호
    • /
    • pp.18-30
    • /
    • 2020
  • 최근 소셜 네트워크, 시맨틱 웹 등 여러 분야에서 그래프 구조 데이터가 널리 사용됨에 따라 대량의 그래프 데이터에 대한 효과적이고 효율적인 검색 방법의 필요성이 커지고 있다. 기존 키워드 기반 검색 방법들은 대부분 주어진 질의에 대한 연관도만을 고려하여 결과를 구한다. 그러나 이런 방법은 질의 연관도는 높지만 콘텐트 노드들을 공유하는 유사한 결과들이 함께 선택될 가능성이 높다. 이런 문제점을 개선하기 위해 본 논문에서는 키워드 질의에 대한 답 트리에 포함된 콘텐트 노드들의 유사성을 제어하여 콘텐트 노드가 다양한 답 트리들을 구하는 top-k 검색 방법을 제안한다. 다양한 답 트리 집합의 기준을 정의하고, 다양한 top-k 결과 집합을 구하기 위한 두 가지 방법으로 점진적 나열 알고리즘과 A 탐색 기법을 이용한 휴리스틱 탐색 알고리즘을 설계한다. 또 휴리스틱 탐색의 성능을 높이기 위한 개선 방법을 제시한다. 실 데이터를 이용한 성능 실험 결과를 통해, 본 논문에서 제안한 휴리스틱 탐색 방법이 질의 연관성뿐만 아니라 콘텐트 노드들의 상이도가 높은 다양한 답 트리들을 효율적으로 구할 수 있음을 보인다.

프로젝트 일정과 자원 평준화를 포함한 다목적 최적화 문제에서 순차적 자원 감소에 기반한 파레토 집합의 생성 (Generation of Pareto Sets based on Resource Reduction for Multi-Objective Problems Involving Project Scheduling and Resource Leveling)

  • 정우진;박성철;임동순
    • 산업경영시스템학회지
    • /
    • 제43권2호
    • /
    • pp.79-86
    • /
    • 2020
  • To make a satisfactory decision regarding project scheduling, a trade-off between the resource-related cost and project duration must be considered. A beneficial method for decision makers is to provide a number of alternative schedules of diverse project duration with minimum resource cost. In view of optimization, the alternative schedules are Pareto sets under multi-objective of project duration and resource cost. Assuming that resource cost is closely related to resource leveling, a heuristic algorithm for resource capacity reduction (HRCR) is developed in this study in order to generate the Pareto sets efficiently. The heuristic is based on the fact that resource leveling can be improved by systematically reducing the resource capacity. Once the reduced resource capacity is given, a schedule with minimum project duration can be obtained by solving a resource-constrained project scheduling problem. In HRCR, VNS (Variable Neighborhood Search) is implemented to solve the resource-constrained project scheduling problem. Extensive experiments to evaluate the HRCR performance are accomplished with standard benchmarking data sets, PSPLIB. Considering 5 resource leveling objective functions, it is shown that HRCR outperforms well-known multi-objective optimization algorithm, SPEA2 (Strength Pareto Evolutionary Algorithm-2), in generating dominant Pareto sets. The number of approximate Pareto optimal also can be extended by modifying weight parameter to reduce resource capacity in HRCR.