• Title/Summary/Keyword: Scheduling System

Search Result 2,159, Processing Time 0.027 seconds

A Comparison Study of Model Parameter Estimation Methods for Prognostics (건전성 예측을 위한 모델변수 추정방법의 비교)

  • An, Dawn;Kim, Nam Ho;Choi, Joo Ho
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.25 no.4
    • /
    • pp.355-362
    • /
    • 2012
  • Remaining useful life(RUL) prediction of a system is important in the prognostics field since it is directly linked with safety and maintenance scheduling. In the physics-based prognostics, accurately estimated model parameters can predict the remaining useful life exactly. It, however, is not a simple task to estimate the model parameters because most real system have multivariate model parameters, also they are correlated each other. This paper presents representative methods to estimate model parameters in the physics-based prognostics and discusses the difference between three methods; the particle filter method(PF), the overall Bayesian method(OBM), and the sequential Bayesian method(SBM). The three methods are based on the same theoretical background, the Bayesian estimation technique, but the methods are distinguished from each other in the sampling methods or uncertainty analysis process. Therefore, a simple physical model as an easy task and the Paris model for crack growth problem are used to discuss the difference between the three methods, and the performance of each method evaluated by using established prognostics metrics is compared.

Applying TMO-Based Object Group Model to Area of Distributed Real-Time Applications and Its Analysis (분산 실시간 응용 분야에 TMO 기반 객체그룹 모델의 적용 및 분석)

  • 신창선;정창원;주수종
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.8
    • /
    • pp.432-444
    • /
    • 2004
  • In this paper, we construct the TMO-based object group model on distributed environment, and analyze and evaluate the executability for distributed real-time service of our object group model by developing the distributed real-time application simulator applying the model. The Time-triggered Message-triggered Object(TMO) is a real-time server object having real-time property itself. The TMO-based object group is defined as a set of objects which logically reconfigured the physically distributed one or more TMOs on network by a given distributed application. For supporting group management of the server objects, the TMO-based object group we suggested provides the functions which register and withdraw the solver objects as a group member to an arbitrary object group, and also provides the functions which insert and delete the access rights of server objects from clients. Also, our model was designed and implemented to support the appropriate object selection and dynamic binding service for a single TMO as well as the duplicated TMOs, and to support the real-time scheduling service for the clients which are requesting the service. Finally, we developed the Defence System against Invading Enemy Planes(DSIEP) simulator as a practical example of distributed real-time application by applying our model, and evaluated the adaptability of distributed service strategies for the group components and the executability of real-time services that the TMO-based object group model provides.

HW/SW Partitioning Techniques for Multi-Mode Multi-Task Embedded Applications (멀티모드 멀티태스크 임베디드 어플리케이션을 위한 HW/SW 분할 기법)

  • Kim, Young-Jun;Kim, Tae-Whan
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.34 no.8
    • /
    • pp.337-347
    • /
    • 2007
  • An embedded system is called a multi-mode embedded system if it performs multiple applications by dynamically reconfiguring the system functionality. Further, the embedded system is called a multi-mode multi-task embedded system if it additionally supports multiple tasks to be executed in a mode. In this Paper, we address a HW/SW partitioning problem, that is, HW/SW partitioning of multi-mode multi-task embedded applications with timing constraints of tasks. The objective of the optimization problem is to find a minimal total system cost of allocation/mapping of processing resources to functional modules in tasks together with a schedule that satisfies the timing constraints. The key success of solving the problem is closely related to the degree of the amount of utilization of the potential parallelism among the executions of modules. However, due to an inherently excessively large search space of the parallelism, and to make the task of schedulabilty analysis easy, the prior HW/SW partitioning methods have not been able to fully exploit the potential parallel execution of modules. To overcome the limitation, we propose a set of comprehensive HW/SW partitioning techniques which solve the three subproblems of the partitioning problem simultaneously: (1) allocation of processing resources, (2) mapping the processing resources to the modules in tasks, and (3) determining an execution schedule of modules. Specifically, based on a precise measurement on the parallel execution and schedulability of modules, we develop a stepwise refinement partitioning technique for single-mode multi-task applications. The proposed techniques is then extended to solve the HW/SW partitioning problem of multi-mode multi-task applications. From experiments with a set of real-life applications, it is shown that the proposed techniques are able to reduce the implementation cost by 19.0% and 17.0% for single- and multi-mode multi-task applications over that by the conventional method, respectively.

Ovarian and Endometrial $^{18}F$-FDG Uptake During the Menstrual Cycle in Normal Premenopausal Patients: Evaluation by PET/CT (월경주기에 따른 $^{18}F$-FDG PET/CT의 자궁 내 섭취에 관한 연구)

  • Bahn, Young-Kag;Park, Hoon-Hee;NamKoong, Hyuk;Kim, Sang-Kyoo;Lim, Han-Sang;Lee, Chang-Ho
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.13 no.3
    • /
    • pp.43-47
    • /
    • 2009
  • Purpose: The menstrual cycle of normal premenopausal patients was divide into menstrual flow phase, proliferative phase, ovulatory phase, secretory phase. The aim of this study was to ovarian and endometrial $^{18}F$-FDG uptake during the menstrual cycle in normal premenopausal patients. Materials and Methods: We identified 200 incidental $^{18}F$-FDG uptake in the ovary. The patient fasted at least 6 hours before receiving an intravenous injection of 370-592 MBq (10-16mCi) of $^{18}F$-FDG. Scanning from the base of the skull though the mid thigh was performed using the Discovery Ste PET/CT system (GE Healthcare, Milwaukee, WI, USA). Ovarian and endometrial $^{18}F$-FDG uptake (expressed as standardized uptake value) was measured on PET/CT image. Results: Two peaks of increased endometrial $^{18}F$-FDG uptake were identified during the menstrual cycle. The $SUV_{avg}$ and $SUV_{max}$ was $2.89{\pm}1.04$ and $3.17{\pm}1.59$ in menstruating patients, $2.4{\pm}0.88$ and $2.98{\pm}1.14$ in proliferative phase patients, $3.59{\pm}1.76$ and $3.17{\pm}1.67$ in ovulatory phase patients, $2.58{\pm}1.39$ and $3.1{\pm}1.8$ in secretory phase patients. Conclusions: Increased ovarian and endometrial $^{18}F$-FDG uptake could be found the time of menstrual flow and ovulatory phase of menstrual cycle. Increased uptake in endometrial adjacent to a cervical tumor does not necessarily reflect endometrial tumor invasion. Since increased uptake was dependent on the menstrual cycle, it can be avoided by scheduling PET/CT just after menstruation. Non-menstrual-related endometrial uptake may be instrumental in establishing a diagnosis in a premenopaual patient.

  • PDF

Design Information Management System Core Development Using Industry Foundation Classes (IFC를 이용한 설계정보관리시스템 핵심부 구축)

  • Lee Keun-hyung;Chin Sang-yoon;Kim Jae-jun
    • Korean Journal of Construction Engineering and Management
    • /
    • v.1 no.2 s.2
    • /
    • pp.98-107
    • /
    • 2000
  • Increased use of computers in AEC (Architecture, Engineering and Construction) has expanded the amount of information gained from CAD (Computer Aided Design), PMIS (Project Management Information System), Structural Analysis Program, and Scheduling Program as well as making it more complex. And the productivity of AEC industry is largely dependent on well management and efficient reuse of this information. Accordingly, such trend incited much research and development on ITC (Information Technology in Construction) and CIC (Computer Integrated Construction) to be conducted. In exemplifying such effort, many researchers studied and researched on IFC (Industry Foundation Classes) since its development by IAI (International Alliance for Interoperability) for the product based information sharing. However, in spite of some valuable outputs, these researches are yet in the preliminary stage and deal mainly with conceptual ideas and trial implementations. Research on unveiling the process of the IFC application development, the core of the Design Information management system, and its applicable plan still need be done. Thus, the purpose of this paper is to determine the technologies needed for Design Information management system using IFC, and to present the key roles and the process of the IFC application development and its applicable plan. This system play a role to integrate the architectural information and the structural information into the product model and to group many each product items with various levels and aspects. To make the process model, we defined two activities, 'Product Modeling', 'Application Development', at the initial level. Then we decomposed the Application Development activity into five activities, 'IFC Schema Compile', 'Class Compile', 'Make Project Database Schema', 'Development of Product Frameworker', 'Make Project Database'. These activities are carried out by C++ Compiler, CAD, ObjectStore, ST-Developer, and ST-ObjectStore. Finally, we proposed the applicable process with six stages, '3D Modeling', 'Creation of Product Information', 'Creation and Update of Database', 'Reformation of Model's Structure with Multiple Hierarchies', 'Integration of Drawings and Specifications', and 'Creation of Quantity Information'. The IFCs, including the other classes which are going to be updated and developed newly on the construction, civil/structure, and facility management, will be used by the experts through the internet distribution technologies including CORBA and DCOM.

  • PDF

A New Bias Scheduling Method for Improving Both Classification Performance and Precision on the Classification and Regression Problems (분류 및 회귀문제에서의 분류 성능과 정확도를 동시에 향상시키기 위한 새로운 바이어스 스케줄링 방법)

  • Kim Eun-Mi;Park Seong-Mi;Kim Kwang-Hee;Lee Bae-Ho
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.11
    • /
    • pp.1021-1028
    • /
    • 2005
  • The general solution for classification and regression problems can be found by matching and modifying matrices with the information in real world and then these matrices are teaming in neural networks. This paper treats primary space as a real world, and dual space that Primary space matches matrices using kernel. In practical study, there are two kinds of problems, complete system which can get an answer using inverse matrix and ill-posed system or singular system which cannot get an answer directly from inverse of the given matrix. Further more the problems are often given by the latter condition; therefore, it is necessary to find regularization parameter to change ill-posed or singular problems into complete system. This paper compares each performance under both classification and regression problems among GCV, L-Curve, which are well known for getting regularization parameter, and kernel methods. Both GCV and L-Curve have excellent performance to get regularization parameters, and the performances are similar although they show little bit different results from the different condition of problems. However, these methods are two-step solution because both have to calculate the regularization parameters to solve given problems, and then those problems can be applied to other solving methods. Compared with UV and L-Curve, kernel methods are one-step solution which is simultaneously teaming a regularization parameter within the teaming process of pattern weights. This paper also suggests dynamic momentum which is leaning under the limited proportional condition between learning epoch and the performance of given problems to increase performance and precision for regularization. Finally, this paper shows the results that suggested solution can get better or equivalent results compared with GCV and L-Curve through the experiments using Iris data which are used to consider standard data in classification, Gaussian data which are typical data for singular system, and Shaw data which is an one-dimension image restoration problems.

User-Level Threads for the ARX Real-Time Operating System (ARX 실시간 운영체계를 위한 사용자 레벨 쓰레드)

  • Seo, Yang-Min;Park, Jung-Keun;Hong, Seong-Soo
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 1998.10a
    • /
    • pp.65-67
    • /
    • 1998
  • 내장 실시간 시스템이 높은 우선순위의 비동기적 이벤트를 적시에 처리하면서 필수적으로 적은 비용의 선점 다중쓰레드를 지원해야한다. 사용자 레벨 쓰레드는 커널 레벨 쓰레드 보다 적은 비용의 유연한 추상적 기법들을 제공하지만, 기존의 실시간 시스템에서는 스케줄링과 시그날(signal) 처리가 단순하다는 이류로 커널 레벨 쓰레드가 선호 되어왔다. 본 논문에서는 내장 실시간 시스템에 적합한 새로운 사용자 레벨 다중 쓰레드 방식을 제안한다. 이 기법은 가상 쓰레드(virtual threads)와 개선된 스케줄링 이벤트 업콜(scheduling event upcall) 메카니즘을 기반으로 한다. 가상 쓰레드는 사용자 레벨 쓰레드에게 커널 레벨의 실행 환경을 제공할 수 있도록 사용자 레벨 쓰레드를 커널 레벨로 형상화한 것이다. 이 쓰레드는 필요에 의해 잠시동안 사용자 레벨 쓰레드에 묶이는 소동적인 존재이다. 스케줄링 이벤트 업콜 메카니즘은 쓰레드 블록킹과 타이머 만기와 같은 커널 이벤트를 유저 프로세서에게 전달할 수 있게 한다. 본 논문의 개선된 업콜 방식은 scheduler activation과 시그날과 같은 전통적인 업콜 구조에서 예측하기 힘든 요소들을 배제했다. 순간적인 시스템의 과부하 상황에서도 이벤트를 놓치지 않으면서 커널과 유저 프로세서의 비싼 동기화 작업들을 피할 수 있도록 하는 잠금(lock)이 필요 없는 이벤트 큐를 상용한다. 본 기법은 서울대학교 실시간 운영체계 실험실에서 구현한 ARX위에 완벽하게 구현되었다. ARX 사용자 레벨 쓰레드가 사용자 레벨 쓰레드의 장점을 손상하지 않으면서 솔라리스와 윈도즈98과 같은 상용 운영체제의 커널 쓰레드보다 성능이 우수함을 실험적 비교에 의해서 입증한다.분에서 uronic acid를 분리동정하였을 때 점미는 0.90%, 백미는 0.66%, 흑미는 1.8%로서 흑미에서 uronic acid 함량이 두 배 이상으로 나타났다. 흑미의 uronic acid 함량이 가장 많이 용출된 분획은 sodium hydroxide 부분으로서 hemicellulose구조가 polyuronic acid의 형태인 것으로 사료된다. 추출획분의 구성단당은 여러 곡물연구의 보고와 유사하게 glucose, arabinose, xylose 함량이 대체로 높게 나타났다. 점미가 수가용성분에서 goucose대비 용출함량이 고르게 나타나는 경향을 보였고 흑미는 알칼리가용분에서 glucose가 상당량(0.68%) 포함되고 있음을 보여주었고 arabinose(0.68%), xylose(0.05%)도 다른 종류에 비해서 다량 함유한 것으로 나타났다. 흑미는 총식이섬유 함량이 높고 pectic substances, hemicellulose, uronic acid 함량이 높아서 콜레스테롤 저하 등의 효과가 기대되며 고섬유식품으로서 조리 특성 연구가 필요한 것으로 사료된다.리하였다. 얻어진 소견(所見)은 다음과 같았다. 1. 모년령(母年齡), 임신회수(姙娠回數), 임신기간(姙娠其間), 출산시체중등(出産時體重等)의 제요인(諸要因)은 주산기사망(周産基死亡)에 대(對)하여 통계적(統計的)으로 유의(有意)한 영향을 미치고 있어 $25{\sim}29$세(歲)의 연령군에서, 2번째 임신과 2번째의 출산에서 그리고 만삭의 임신 기간에, 출산시체중(出産時體重) $3.50{\sim}3.99kg$사이의 아이에서 그 주산기사망률(周産基死亡率)이 각각 가장 낮았다. 2. 사산(死産)과 초생

  • PDF

Control of pH Neutralization Process using Simulation Based Dynamic Programming in Simulation and Experiment (ICCAS 2004)

  • Kim, Dong-Kyu;Lee, Kwang-Soon;Yang, Dae-Ryook
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.620-626
    • /
    • 2004
  • For general nonlinear processes, it is difficult to control with a linear model-based control method and nonlinear controls are considered. Among the numerous approaches suggested, the most rigorous approach is to use dynamic optimization. Many general engineering problems like control, scheduling, planning etc. are expressed by functional optimization problem and most of them can be changed into dynamic programming (DP) problems. However the DP problems are used in just few cases because as the size of the problem grows, the dynamic programming approach is suffered from the burden of calculation which is called as 'curse of dimensionality'. In order to avoid this problem, the Neuro-Dynamic Programming (NDP) approach is proposed by Bertsekas and Tsitsiklis (1996). To get the solution of seriously nonlinear process control, the interest in NDP approach is enlarged and NDP algorithm is applied to diverse areas such as retailing, finance, inventory management, communication networks, etc. and it has been extended to chemical engineering parts. In the NDP approach, we select the optimal control input policy to minimize the value of cost which is calculated by the sum of current stage cost and future stages cost starting from the next state. The cost value is related with a weight square sum of error and input movement. During the calculation of optimal input policy, if the approximate cost function by using simulation data is utilized with Bellman iteration, the burden of calculation can be relieved and the curse of dimensionality problem of DP can be overcome. It is very important issue how to construct the cost-to-go function which has a good approximate performance. The neural network is one of the eager learning methods and it works as a global approximator to cost-to-go function. In this algorithm, the training of neural network is important and difficult part, and it gives significant effect on the performance of control. To avoid the difficulty in neural network training, the lazy learning method like k-nearest neighbor method can be exploited. The training is unnecessary for this method but requires more computation time and greater data storage. The pH neutralization process has long been taken as a representative benchmark problem of nonlin ar chemical process control due to its nonlinearity and time-varying nature. In this study, the NDP algorithm was applied to pH neutralization process. At first, the pH neutralization process control to use NDP algorithm was performed through simulations with various approximators. The global and local approximators are used for NDP calculation. After that, the verification of NDP in real system was made by pH neutralization experiment. The control results by NDP algorithm was compared with those by the PI controller which is traditionally used, in both simulations and experiments. From the comparison of results, the control by NDP algorithm showed faster and better control performance than PI controller. In addition to that, the control by NDP algorithm showed the good results when it applied to the cases with disturbances and multiple set point changes.

  • PDF

Development of One-to-One Shortest Path Algorithm Based on Link Flow Speeds on Urban Networks (도시부 가로망에서의 링크 통행속도 기반 One-to-One 최단시간 경로탐색 알고리즘 개발)

  • Kim, Taehyeong;Kim, Taehyung;Park, Bum-Jin;Kim, Hyoungsoo
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.11 no.5
    • /
    • pp.38-45
    • /
    • 2012
  • Finding shortest paths on time dependent networks is an important task for scheduling and routing plan and real-time navigation system in ITS. In this research, one-to-one time dependent shortest path algorithms based on link flow speeds on urban networks are proposed. For this work, first we select three general shortest path algorithms such as Graph growth algorithm with two queues, Dijkstra's algorithm with approximate buckets and Dijkstra's algorithm with double buckets. These algorithms were developed to compute shortest distance paths from one node to all nodes in a network and have proven to be fast and efficient algorithms in real networks. These algorithms are extended to compute a time dependent shortest path from an origin node to a destination node in real urban networks. Three extended algorithms are implemented on a data set from real urban networks to test and evaluate three algorithms. A data set consists of 4 urban street networks for Anaheim, CA, Baltimore, MD, Chicago, IL, and Philadelphia, PA. Based on the computational results, among the three algorithms for TDSP, the extended Dijkstra's algorithm with double buckets is recommended to solve one-to-one time dependent shortest path for urban street networks.

A Study on Parallel Performance Optimization Method for Acceleration of High Resolution SAR Image Processing (고해상도 SAR 영상처리 고속화를 위한 병렬 성능 최적화 기법 연구)

  • Lee, Kyu Beom;Kim, Gyu Bin;An, Sol Bo Reum;Cho, Jin Yeon;Lim, Byoung-Gyun;Kim, Dong-Hyun;Kim, Jeong Ho
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.46 no.6
    • /
    • pp.503-512
    • /
    • 2018
  • SAR(Synthetic Aperture Radar) is a technology to acquire images by processing signals obtained from radar, and there is an increasing demand for utilization of high-resolution SAR images. In this paper, for high-speed processing of high-resolution SAR image data, a study for SAR image processing algorithms to achieve optimal performance in multi-core based computer architecture is performed. The performance deterioration due to a large amount of input/output data for high resolution images is reduced by maximizing the memory utilization, and the parallelization ratio of the code is increased by using dynamic scheduling and nested parallelism of OpenMP. As a result, not only the total computation time is reduced, but also the upper bound of parallel performance is increased and the actual parallel performance on a multi-core system with 10 cores is improved by more than 8 times. The result of this study is expected to be used effectively in the development of high-resolution SAR image processing software for multi-core systems with large memory.