• Title/Summary/Keyword: source code level

Search Result 157, Processing Time 0.026 seconds

A Study on the Design of Security Metrics for Source Code (소스코드의 보안성 메트릭 설계에 관한 연구)

  • Seo, Dong-Su
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.20 no.6
    • /
    • pp.147-155
    • /
    • 2010
  • It has been widely addressed that static analysis techniques can play important role in identifying potential security vulnerability reside in source code. This paper proposes the design and application of security metrics that use both vulnerability information extracted from the static analysis, and significant factors of information that software handles. The security metrics are useful for both developers and evaluators in that the metrics help them identity source code vulnerability in early stage of development. By effectively utilizing the security metrics, evaluators can check the level of source code security, and confirm the final code depending on the characteristics of the source code and the security level of information required.

Development of Web-based Off-site Consequence Analysis Program and its Application for ILRT Extension (격납건물종합누설률시험 주기연장을 위한 웹기반 소외결말분석 프로그램 개발 및 적용)

  • Na, Jang-Hwan;Hwang, Seok-Won;Oh, Ji-Yong
    • Journal of the Korean Society of Safety
    • /
    • v.27 no.5
    • /
    • pp.219-223
    • /
    • 2012
  • For an off-site consequence analysis at nuclear power plant, MELCOR Accident Consequence Code System(MACCS) II code is widely used as a software tool. In this study, the algorithm of web-based off-site consequence analysis program(OSCAP) using the MACCS II code was developed for an Integrated Leak Rate Test (ILRT) interval extension and Level 3 probabilistic safety assessment(PSA), and verification and validation(V&V) of the program was performed. The main input data for the MACCS II code are meteorological, population distribution and source term information. However, it requires lots of time and efforts to generate the main input data for an off-site consequence analysis using the MACCS II code. For example, the meteorological data are collected from each nuclear power site in real time, but the formats of the raw data collected are different from each site. To reduce the efforts and time for risk assessments, the web-based OSCAP has an automatic processing module which converts the format of the raw data collected from each site to the input data format of the MACCS II code. The program also provides an automatic function of converting the latest population data from Statistics Korea, the National Statistical Office, to the population distribution input data format of the MACCS II code. For the source term data, the program includes the release fraction of each source term category resulting from modular accident analysis program(MAAP) code analysis and the core inventory data from ORIGEN. These analysis results of each plant in Korea are stored in a database module of the web-based OSCAP, so the user can select the defaulted source term data of each plant without handling source term input data.

A Study of LLVM-based Embedded System Performance Analyzer (LLVM 기반의 임베디드 시스템 성능 분석기의 연구)

  • Cho, Doosan
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.1
    • /
    • pp.577-582
    • /
    • 2022
  • For developing a new embedded system, an application program/an emulator and a compiler are developed simultaneously. In order to provide the optimal performance of all system components, local optimization should be carried out for the developing process. For this purpose, if a source-level performance analyzer is developed, it is possible to optimize the application program's source code by the performance evaluation. In general, the performance of an application program is determined in the loop iterations. The Intermediate Representation (IR) code generator generates IR code from the source code, and evaluates the execution time with the instructions in the intermediate representation code. If the source code is improved based on the evaluated result, better results can be obtained in the final application code. This study describes the source-level performance analyzer that can be used during the simultaneous development of the new embedded system and its application programs. The performance analyzer makes it possible to more quickly optimize the performance of the new embedded system.

Vertical System Testing Method For Efficient Error Tracing (효과적인 오류 추적을 위한 수직적 시스템 시험 방법)

  • Seo, Kwang-Ik;Choi, Eun-Man
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.2
    • /
    • pp.19-29
    • /
    • 2008
  • In case of unit testing, White-box test can be used to closely check source code and to analyze logic and statement errors. On the other hand, in case of function testing of system level, Black-box test can be mainly used to compare actual and expected results by inputting test data because the scale of function is large. This Black-Box test in system testing level has problem in tracing errors in source code when we find errors. Moreover applying White-box test is not easy for system testing level because the levels of test target are different. Therefore this paper suggests the vertical test method of a practical and integrated system level which can checks up to source code level using White-box test style although it aims to test the highly abstract level like a system function. In addition, the experiment explains how to apply the vertical test by displaying an example which traces from UML specification model to the source code and also shows efficiency of error trace.

  • PDF

Cross-Language Clone Detection based on Common Token (공통 토큰에 기반한 서로 다른 언어의 유사성 검사)

  • Hong, Sung-Moon;Kim, Hyunha;Lee, Jaehyung;Park, Sungwoo;Mo, Ji-Hwan;Doh, Kyung-Goo
    • Journal of Software Assessment and Valuation
    • /
    • v.14 no.2
    • /
    • pp.35-44
    • /
    • 2018
  • Tools for detecting cross-language clones usually compare abstract-syntax-tree representations of source code, which lacks scalability. In order to compare large source code to a practical level, we need a similarity checking technique that works on a token level basis. In this paper, we define common tokens that represent all tokens commonly used in programming languages of different paradigms. Each source code of different language is then transformed into the list of common tokens that are compared. Experimental results using exEyes show that our proposed method using common tokens is effective in detecting cross-language clones.

A Custom Code Generation Technique for ASIPs from High-level Language (고급 언어에서 ASIP을 위한 전용 부호 생성 기술 연구)

  • Alam, S.M. Shamsul;Choi, Goangseog
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.11 no.3
    • /
    • pp.31-43
    • /
    • 2015
  • In this paper, we discuss a code generation technique for custom transport triggered architecture (TTA) from a high-level language structure. This methodology is implemented by using TTA-based Co-design Environment (TCE) tool. The results show how the scheduler exploits instruction level parallelism in the custom target architecture and source program. Thus, the scheduler generates parallel TTA instructions using lower cycle counts than the sequential scheduling algorithm. Moreover, we take Tensilica tool to make a comparison with TCE. Because of the efficiency of TTA, TCE takes less execution cycles compared to Tensilica configurations. Finally, this paper shows that it requires only 7 cycles to generate the parallel TTA instruction set for implementing Cyclic Redundancy Check (CRC) applications as an input design, and presents the code generation technique to move complexity from the processor software to hardware architecture. This method can be applicable lots of channel Codecs like CRC and source Codecs like High Efficiency Video Coding (HEVC).

Development of Free-surface Decomposition Method and Its ApplicationDevelopment of Free-surface Decomposition Method and Its Application

  • Park, Sunho
    • Journal of Advanced Research in Ocean Engineering
    • /
    • v.3 no.2
    • /
    • pp.75-82
    • /
    • 2017
  • With the development of computational fluid dynamics (CFD), studies on shipbuilding and maritime issues including free-surface wave flow have been conducted. Although the volume of fluid (VOF) and level-set methods are widely used to study the free-surface wave flow, disadvantages exist. In particular, it takes a long time to obtain solutions. In this study, a free-surface capturing code is developed for ship and offshore structures. The developed code focuses on accuracy and computation time. Open source CFD libraries, termed OpenFOAM, are used to develop the code. The results obtained using the developed code are compared with those obtained using interFoam. The results show that the developed code could be used to capture the free-surface wave flow without numerical diffusion; moreover, the accuracy of the developed code is largely the same as that of interFoam.

An Efficient Bit-Level Lossless Grayscale Image Compression Based on Adaptive Source Mapping

  • Al-Dmour, Ayman;Abuhelaleh, Mohammed;Musa, Ahmed;Al-Shalabi, Hasan
    • Journal of Information Processing Systems
    • /
    • v.12 no.2
    • /
    • pp.322-331
    • /
    • 2016
  • Image compression is an essential technique for saving time and storage space for the gigantic amount of data generated by images. This paper introduces an adaptive source-mapping scheme that greatly improves bit-level lossless grayscale image compression. In the proposed mapping scheme, the frequency of occurrence of each symbol in the original image is computed. According to their corresponding frequencies, these symbols are sorted in descending order. Based on this order, each symbol is replaced by an 8-bit weighted fixed-length code. This replacement will generate an equivalent binary source with an increased length of successive identical symbols (0s or 1s). Different experiments using Lempel-Ziv lossless image compression algorithms have been conducted on the generated binary source. Results show that the newly proposed mapping scheme achieves some dramatic improvements in regards to compression ratios.

Proof Transformation for Source-level Optimization (최적화 컴파일러에 맞춘 프로그램 오류 검증의 변환)

  • 양홍석;이광근
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2004.04b
    • /
    • pp.850-852
    • /
    • 2004
  • 고급 언어 (high-level language)로 프로그램을 쓰고 그것이 맞다고 증명 했을때, 프로그램과 증명을 동시에 "컴파일"해주는 방법을 찾으려고 한다. 이러한 방법은 "증명 보내기"(Proof Carrying Code)라는 제안을 실용화할 때 요긴하게 사용할 수 있는 기술이다. 지면관계상 "반복문에서 명령어 끄집어내기"(Loop Invariant Code Motion)라는 최적화 과정에 맞게 Hoare 증명을 변환하는 방법에 대해서만 알아보겠다.

  • PDF

Automatic Extraction of Abstract Components for supporting Model-driven Development of Components (모델기반 컴포넌트 개발방법론의 지원을 위한 추상컴포넌트 자동 추출기법)

  • Yun, Sang Kwon;Park, Min Gyu;Choi, Yunja
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.2 no.8
    • /
    • pp.543-554
    • /
    • 2013
  • Model-Driven Development(MDD) helps developers verify requirements and design issues of a software system in the early stage of development process by taking advantage of a software model which is the most highly abstracted form of a software system. In practice, however, many software systems have been developed through a code-centric method that builds a software system bottom-up rather than top-down. So, without support of appropriate tools, it is not easy to introduce MDD to real development process. Although there are many researches about extracting a model from code to help developers introduce MDD to code-centrically developed system, most of them only extracted base-level models. However, using concept of abstract component one can continuously extract higher level model from base-level model. In this paper we propose a practical method for automatic extraction of base level abstract component from source code, which is the first stage of continuous extraction process of abstract component, and validate the method by implementing an extraction tool based on the method. Target code chosen is the source code of TinyOS, an operating system for wireless sensor networks. The tool is applied to the source code of TinyOS, written in nesC language.