• Title/Summary/Keyword: Automated software

Search Result 525, Processing Time 0.026 seconds

Classification and Retrieval of Object - Oriented Reuse Components with HACM (HACM을 사용한 객체지향 재사용 부품의 분류와 검색)

  • Bae, Je-Min;Kim, Sang-Geun;Lee, Kyung-Whan
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.7
    • /
    • pp.1733-1748
    • /
    • 1997
  • In this paper, we propose the classification scheme and retrieval mechanism which can apply to many application domains in order to construct the software reuse library. Classification scheme which is the core of the accessibility in the reusability, is defined by the hierarchical structure using the agglomerative clusters. Agglomerative cluster means the group of the reuse component by the functional relationships. Functional relationships are measured by the HACM which is the representation method about software components to calculate the similarities among the classes in the particular domain. And clustering informations are added to the library structure which determines the functionality and accuracy of the retrieval system. And the system stores the classification results such as the index information with the weights, the similarity matrix, the hierarchical structure. Therefore users can retrieve the software component using the query which is the natural language. The thesis is studied to focus on the findability of software components in the reuse library. As a result, the part of the construction process of the reuse library was automated, and we can construct the object-oriented reuse library with the extendibility and relationship about the reuse components. Also the our process is visualized through the browse hierarchy of the retrieval environment, and the retrieval system is integrated to the reuse system CARS 2.1.

  • PDF

Regression Analysis of the Relationships between Complexity Metrics and Faults on the Telecommunication Program (통신 소프트웨어의 프로그램 결함과 복잡도의 관련성 분석을 위한 회귀분석 모델)

  • Lee, Gyeong-Hwan;Jeong, Chang-Sin;Hwang, Seon-Myeong;Jo, Byeong-Gyu;Park, Ji-Hun;Kim, Gang-Tae
    • Journal of KIISE:Software and Applications
    • /
    • v.26 no.11
    • /
    • pp.1282-1287
    • /
    • 1999
  • 통신 프로그램은 고도의 신뢰성과 기능성, 확장성, 그리고 유지 보수성이 필요하다. 프로그램 테스트의 결과와 McCabe의 Complexity를 측정한 데이타를 가지고 회귀모델을 만들고 그 신뢰성을 분석함으로서 프로그램의 결함과 복잡도의 관련성을 평가한다.본 연구에서 사용한 통신 프로그램은 500개 블록이 59가지 기능을 수행하는 교환 기능 중에서 복잡도가 너무 많아서 통계 처리의 bias가 될 블록을 제외하고 394 블록을 선정하여 SAS에 의해서 통계 분석을 하고 회귀 분석 모델을 설계하였다. t 분포에 의하여 방정식의 유의성 수준을 검증하고 프로그램의 결함수에 가장 큰 영향을 주고 있는 복잡도가 McCabe의 복잡도와 설계 복잡도 임을 밝혀냈다. 이 연구 결과에 의해서 설계 정보 및 유지 보수 정보를 얻을 수 있다. Abstract Switching software requires high reliability, functionality, extendability and maintainability. For doing, software quality model based on MaCabe's complexity measure is investigated. It is experimentally shown using regression analysis the program fault density depends on the complexity and size of the function unit. The software should be verified and tested if it satisfies its requirements with automated analysis tools. In this paper we propose the regression model with the test data.The sample program for the regression model consists of more than 500 blocks, where each block compose of 10 files, which has 59 functions of switching activity.Among them we choose 394 blocks and analyzed for 59 functions by testing tools and SAS package. We developed Regression Analysis Model and evaluated significant of the equation based on McCabe's cyclomatic complexity, block design complexity, design complexity, and integration complexity.The results of our experimental study are that number of fault are under the influence of McCabe's complexity number and design complexity.

Bayesian Optimization Framework for Improved Cross-Version Defect Prediction (향상된 교차 버전 결함 예측을 위한 베이지안 최적화 프레임워크)

  • Choi, Jeongwhan;Ryu, Duksan
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.9
    • /
    • pp.339-348
    • /
    • 2021
  • In recent software defect prediction research, defect prediction between cross projects and cross-version projects are actively studied. Cross-version defect prediction studies assume WP(Within-Project) so far. However, in the CV(Cross-Version) environment, the previous work does not consider the distribution difference between project versions is important. In this study, we propose an automated Bayesian optimization framework that considers distribution differences between different versions. Through this, it automatically selects whether to perform transfer learning according to the difference in distribution. This framework is a technique that optimizes the distribution difference between versions, transfer learning, and hyper-parameters of the classifier. We confirmed that the method of automatically selecting whether to perform transfer learning based on the distribution difference is effective through experiments. Moreover, we can see that using our optimization framework is effective in improving performance and, as a result, can reduce software inspection effort. This is expected to support practical quality assurance activities for new version projects in a cross-version project environment.

AJFCode: An Approach for Full Aspect-Oriented Code Generation from Reusable Aspect Models

  • Mehmood, Abid;Jawawi, Dayang N.A.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.6
    • /
    • pp.1973-1993
    • /
    • 2022
  • Model-driven engineering (MDE) and aspect-oriented software development (AOSD) contribute to the common goal of development of high-quality code in reduced time. To complement each approach with the benefits of the other, various methods of integration of the two approaches were proposed in the past. Aspect-oriented code generation, which targets obtaining aspect-oriented code directly from aspect models, offers some unique advantages over the other integration approaches. However, the existing aspect-oriented code generation approaches do not comprehensively address all aspects of a model-driven code generation system, such as a textual representation of graphical models, conceptual mapping, and incorporation of behavioral diagrams. These problems limit the worth of generated code, especially in practical use. Here, we propose AJFCode, an approach for aspect-oriented model-driven code generation, which comprehensively addresses the various aspects including the graphical models and their text-based representation, mapping between visual model elements and code, and the behavioral code generation. Experiments are conducted to compare the maintainability and reusability characteristics of the aspect-oriented code generated using the AJFCode with the most comprehensive object-oriented code generation approach. AJFCode performs well in terms of all metrics related to maintainability and reusability of code. However, the most significant improvement is noticed in the separation of concerns, coupling, and cohesion. For instance, AJFCode yields significant improvement in concern diffusion over operations (19 vs 51), coupling between components (0 vs 6), and lack of cohesion in operations (5 vs 9) for one of the experimented concerns.

Implementation of Git's Commit Message Complex Classification Model for Software Maintenance

  • Choi, Ji-Hoon;Kim, Joon-Yong;Park, Seong-Hyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.11
    • /
    • pp.131-138
    • /
    • 2022
  • Git's commit message is closely related to the project life cycle, and by this characteristic, it can greatly contribute to cost reduction and improvement of work efficiency by identifying risk factors and project status of project operation activities. Among these related fields, there are many studies that classify commit messages as types of software maintenance, and the maximum accuracy among the studies is 87%. In this paper, the purpose of using a solution using the commit classification model is to design and implement a complex classification model that combines several models to increase the accuracy of the previously published models and increase the reliability of the model. In this paper, a dataset was constructed by extracting automated labeling and source changes and trained using the DistillBERT model. As a result of verification, reliability was secured by obtaining an F1 score of 95%, which is 8% higher than the maximum of 87% reported in previous studies. Using the results of this study, it is expected that the reliability of the model will be increased and it will be possible to apply it to solutions such as software and project management.

An Improved Technique of Fitness Evaluation for Automated Test Data Generation (테스트 데이터 자동 생성을 위한 적합도 평가 방법의 효율성 향상 기법)

  • Lee, Sun-Yul;Choi, Hyun-Jae;Jeong, Yeon-Ji;Bae, Jung-Ho;Kim, Tae-Ho;Chae, Heung-Suk
    • Journal of KIISE:Software and Applications
    • /
    • v.37 no.12
    • /
    • pp.882-891
    • /
    • 2010
  • Many automated dynamic test data generation technique have been proposed. The techniques evaluate fitness of test data through executing instrumented Software Under Test (SUT) and then generate new test data based on evaluated fitness values and optimization algorithms. Previous researches and experiments have been showed that these techniques generate effective test data. However, optimization algorithms in these techniques incur much time to generate test data, which results in huge test case generation cost. In this paper, we propose a technique for reducing the time of evaluating a fitness of test data among steps of dynamic test data generation methods. We introduce the concept of Fitness Evaluation Program (FEP), derived from a path constraint of SUT. We suggest a test data generation method based on FEP and implement a test generation tool, named ConGA. We also apply ConGA to generate test cases for C programs, and evaluate efficiency of the FEP-based test case generation technique. The experiments show that the proposed technique reduces 20% of test data generation time on average.

Automated Classification of Ground-glass Nodules using GGN-Net based on Intensity, Texture, and Shape-Enhanced Images in Chest CT Images (흉부 CT 영상에서 결절의 밝기값, 재질 및 형상 증강 영상 기반의 GGN-Net을 이용한 간유리음영 결절 자동 분류)

  • Byun, So Hyun;Jung, Julip;Hong, Helen;Song, Yong Sub;Kim, Hyungjin;Park, Chang Min
    • Journal of the Korea Computer Graphics Society
    • /
    • v.24 no.5
    • /
    • pp.31-39
    • /
    • 2018
  • In this paper, we propose an automated method for the ground-glass nodule(GGN) classification using GGN-Net based on intensity, texture, and shape-enhanced images in chest CT images. First, we propose the utilization of image that enhances the intensity, texture, and shape information so that the input image includes the presence and size information of the solid component in GGN. Second, we propose GGN-Net which integrates and trains feature maps obtained from various input images through multiple convolution modules on the internal network. To evaluate the classification accuracy of the proposed method, we used 90 pure GGNs, 38 part-solid GGNs less than 5mm with solid component, and 23 part-solid GGNs larger than 5mm with solid component. To evaluate the effect of input image, various input image set is composed and classification results were compared. The results showed that the proposed method using the composition of intensity, texture and shape-enhanced images showed the best result with 82.75% accuracy.

Compensation of Installation Errors in a Laser Vision System and Dimensional Inspection of Automobile Chassis

  • Barkovski Igor Dunin;Samuel G.L.;Yang Seung-Han
    • Journal of Mechanical Science and Technology
    • /
    • v.20 no.4
    • /
    • pp.437-446
    • /
    • 2006
  • Laser vision inspection systems are becoming popular for automated inspection of manufactured components. The performance of such systems can be enhanced by improving accuracy of the hardware and robustness of the software used in the system. This paper presents a new approach for enhancing the capability of a laser vision system by applying hardware compensation and using efficient analysis software. A 3D geometrical model is developed to study and compensate for possible distortions in installation of gantry robot on which the vision system is mounted. Appropriate compensation is applied to the inspection data obtained from the laser vision system based on the parameters in 3D model. The present laser vision system is used for dimensional inspection of car chassis sub frame and lower arm assembly module. An algorithm based on simplex search techniques is used for analyzing the compensated inspection data. The details of 3D model, parameters used for compensation and the measurement data obtained from the system are presented in this paper. The details of search algorithm used for analyzing the measurement data and the results obtained are also presented in the paper. It is observed from the results that, by applying compensation and using appropriate algorithms for analyzing, the error in evaluation of the inspection data can be significantly minimized, thus reducing the risk of rejecting good parts.

A Study on Validation of OFP for UAV using Auto Code Generation (자동 코드생성을 이용한 무인기용 OFP의 검증에 관한 연구)

  • Cho, Sang-Ook;Choi, Kee-Young
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.37 no.4
    • /
    • pp.359-366
    • /
    • 2009
  • MATLAB Autocode generation is a feature that converts a block diagram model in Simulink to a c program. Utilizing this function makes MATLAB/Simulink an integrated developing environment, from controller design to implementation. It can reduce development cost and time significantly. However, this automated process requires high reliability on the software, especially the original Simulink block diagram model. And thus, the verification of the codes becomes important. In this study, a UAV flight program which is generated with Simulink is validated and modified according to DO-178B. As a result of applying the procedures, the final program not only satisfied the functional requirement but is also verified with structural point of view with Decision Coverage 93%, Condition Coverage 95% and MC/DC 90%.

Data Server Oriented Computing Infrastructure for Process Integration and Multidisciplinary Design Optimization (다분야통합최적설계를 위한 데이터 서버 중심의 컴퓨팅 기반구조)

  • 홍은지;이세정;이재호;김승민
    • Korean Journal of Computational Design and Engineering
    • /
    • v.8 no.4
    • /
    • pp.231-242
    • /
    • 2003
  • Multidisciplinary Design Optimization (MDO) is an optimization technique considering simultaneously multiple disciplines such as dynamics, mechanics, structural analysis, thermal and fluid analysis and electromagnetic analysis. A software system enabling multidisciplinary design optimization is called MDO framework. An MDO framework provides an integrated and automated design environment that increases product quality and reliability, and decreases design cycle time and cost. The MDO framework also works as a common collaborative workspace for design experts on multiple disciplines. In this paper, we present the architecture for an MDO framework along with the requirement analysis for the framework. The requirement analysis has been performed through interviews of design experts in industry and thus we claim that it reflects the real needs in industry. The requirements include integrated design environment, friendly user interface, highly extensible open architecture, distributed design environment, application program interface, and efficient data management to handle massive design data. The resultant MDO framework is datasever-oriented and designed around a centralized data server for extensible and effective data exchange in a distributed design environment among multiple design tools and software.