• Title/Summary/Keyword: Information processing

Search Result 42,013, Processing Time 0.067 seconds

An Exploratory Study of business support policy by growth phases for Small and medium sized enterprises -Focused on Cheonan and Asan in ChungNam- (중소기업의 성장단계별 지원정책에 관한 탐색적 연구 -충청남도 천안·아산지역을 중심으로-)

  • Lee, Jae-Beom
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.14 no.5
    • /
    • pp.2215-2224
    • /
    • 2013
  • This study performed empirical analysis to estimate SMEs needs in terms of business support policy by growth stages(start-up growth expansion). The subject is the SMEs in Cheonan and Asan, ChungNam and the results are as follows. First, In the initial start-up stage, management plays a key role in dealing with money, labor force, markets and technology while running the organization is a key role of the management in the expansion stage. Major policies to help SMEs grow includes money provision needed in the start-up stage, domestic marketing assistance and the provision of human resources in the growth stage, and assistance in foreign marketing and R&D in the expansion stage. Second, To achieve markets businesses aim at entering the existing and niche markets in the initial phase, and creating new markets in the growth phase. Third, Labor force for technology, sales and management planning in the start-up stage, marketing in the growth stage, and labor force for production in the expansion stage are core man- power needed. Fourth, Money for technology development, securing land for factories, organizing man power, securing markets and running the company is needed in the initial and growth stages while fund for facility investment is needed to grow in the expansion stage. Five, Regarding technology, the initial stage needs technology related to new product development, renewing existing products, improving the existing manufacturing process or developing new manufacturing process, while the growth stage needs processing techniques, and the expansion stage needs technology for developing new manufacturing process. Sixth, Making supply contracts with conglomerates, SMEs and public institutions, and sales to foreign markets are ways for SMEs to grow sales. Seventh, What SMEs wish to get includes business incubating support, R&D assistance, information exchanges, practical use of the R&D results, merchandising support, help with the land to build factories and custom-made support for management in the foundation stage while the support they want to get in the growth stage and in the expansion stage is training assistance and trial production respectively.

An efficient interconnection network topology in dual-link CC-NUMA systems (이중 연결 구조 CC-NUMA 시스템의 효율적인 상호 연결망 구성 기법)

  • Suh, Hyo-Joong
    • The KIPS Transactions:PartA
    • /
    • v.11A no.1
    • /
    • pp.49-56
    • /
    • 2004
  • The performance of the multiprocessor systems is limited by the several factors. The system performance is affected by the processor speed, memory delay, and interconnection network bandwidth/latency. By the evolution of semiconductor technology, off the shelf microprocessor speed breaks beyond GHz, and the processors can be scalable up to multiprocessor system by connecting through the interconnection networks. In this situation, the system performances are bound by the latencies and the bandwidth of the interconnection networks. SCI, Myrinet, and Gigabit Ethernet are widely adopted as a high-speed interconnection network links for the high performance cluster systems. Performance improvement of the interconnection network can be achieved by the bandwidth extension and the latency minimization. Speed up of the operation clock speed is a simple way to accomplish the bandwidth and latency betterment, while its physical distance makes the difficulties to attain the high frequency clock. Hence the system performance and scalability suffered from the interconnection network limitation. Duplicating the link of the interconnection network is one of the solutions to resolve the bottleneck of the scalable systems. Dual-ring SCI link structure is an example of the interconnection network improvement. In this paper, I propose a network topology and a transaction path algorism, which optimize the latency and the efficiency under the duplicated links. By the simulation results, the proposed structure shows 1.05 to 1.11 times better latency, and exhibits 1.42 to 2.1 times faster execution compared to the dual ring systems.

Analysis of the 'Problem Solving and Invention' Units of Technology and Home Economics 1 Textbook (기술.가정 1 교과서 '문제해결과 발명' 단원 분석)

  • Jung, Jin Woo
    • 대한공업교육학회지
    • /
    • v.38 no.1
    • /
    • pp.49-67
    • /
    • 2013
  • The purpose of this study is to analyze the external systems and the units 'problem solving and invention' of the middle school technology and home economics 1 textbooks of the revised 2011 national curriculum in an effort to provide some information on the content system of invention education in technology class, as invention education was provided as part of a regular subject for the first time. The findings of the study were as follows: First, 'Technology and Inventions' chapter of Technology and Home Economics 1 Textbooks occupied 10-18% share, with the subchapter of 'Problem Solving and Invention' unit taking up 6.7-29% of the textbooks. Second, for most textbooks, 'Technological Problem Solving', 'Idea Generation' 'Multi-dimensional Projection Method', 'Expansive Thought-Processing Methodology', 'Converging Thought Methodology' and 'Invention in Everyday Lives' were included as main contents based on the accomplishment criteria presented in education process interpretation documents. Third, the detailed structures were generally made up as follows: Introduction (Broad Chapter Title, Subchapter Table of Contents, Introduction, Subchapter Title, Study Objectives, Open Thinking); Development (Unit Title, Thinking Ahead, Core Terms, Main Text, Study Helper, Activities, Research Exercises, Supplemental Readings, In-depth Study Topics, Technology in Everyday Lives, Reading Topics, Discussion Topics, and Career Helpers); and Summary (Subchapter Summary, Study Summary, Terms Summary, Writing Follow-up, Self Review, Broad Chapter Evaluation). Fourth, based on the analysis of figures included, photographs had the largest share, followed by figures, tables, and graphs. The photos were used to illustrate various inventions, invention methodologies, and exercise activities, while figures were included to depict the contents included in the main text, and the tables to assist to preparation of process diagrams or materials lists. Fifth, based on the analysis of content weights, greater weights were placed on 'Inventions and Thoughts', and 'Invention Experiment Activities,' while 'Understanding Inventions' and 'Invention and Patents' chapters did not have a lot of texts involved. Sixth, based on the analysis of content presentation methods, most textbooks combined figures, tables, illustrations and texts to discuss the topics. Based on the above study results, we suggest the following: First, a consistent education curriculum should be developed over the topic of invention; and second, more precise and systematic analysis of textbooks would need to be performed.

Development of Geometrical Quality Control Real-time Analysis Program using an Electronic Portal Imaging (전자포탈영상을 이용한 기하학적 정도관리 실시간 분석 프로그램의 개발)

  • Lee, Sang-Rok;Jung, Kyung-Yong;Jang, Min-Sun;Lee, Byung-Gu;Kwon, Young-Ho
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.24 no.2
    • /
    • pp.77-84
    • /
    • 2012
  • Purpose: To develop a geometrical quality control real-time analysis program using an electronic portal imaging to replace film evaluation method. Materials and Methods: A geometrical quality control item was established with the Eclipse treatment planning system (Version 8.1, Varian, USA) after the Electronic Portal Imaging Device (EPID) took care of the problems occurring from the fixed substructure of the linear accelerator (CL-iX, Varian, USA). Electronic portal image (single exposure before plan) was created at the treatment room's 4DTC (Version 10.2, Varian, USA) and a beam was irradiated in accordance with each item. The gaining the entire electronic portal imaging at the Off-line review and was evaluated by a self-developed geometrical quality control real-time analysis program. As for evaluation methods, the intra-fraction error was analyzed by executing 5 times in a row under identical conditions and procedures on the same day, and in order to confirm the infer-fraction error, it was executed for 10 days under identical conditions of all procedures and was compared with the film evaluation method using an Iso-align$^{TM}$ quality control device. Measurement and analysis time was measured by sorting the time into from the device setup to data achievement and the time amount after the time until the completion of analysis and the convenience of the users and execution processes were compared. Results: The intra-fraction error values for each average 0.1, 0.2, 0.3, 0.2 mm at light-radiation field coincidence, collimator rotation axis, couch rotation axis and gantry rotation axis. By checking the infer-fraction error through 10 days of continuous quality control, the error values obtained were average 1.7, 1.4, 0.7, 1.1 mm for each item. Also, the measurement times were average 36 minutes, 15 minutes for the film evaluation method and electronic portal imaging system, and the analysis times were average 30 minutes, 22 minutes. Conclusion: When conducting a geometrical quality control using an electronic portal imaging, it was found that it is efficient as a quality control tool. It not only reduces costs through not using films, but also reduces the measurement and analysis time which enhances user convenience and can improve the execution process by leaving out film developing procedures etc. Also, images done with evaluation from the self-developed geometrical quality control real-time analysis program, data processing is capable which supports the storage of information.

  • PDF

Design and Implementation of the SSL Component based on CBD (CBD에 기반한 SSL 컴포넌트의 설계 및 구현)

  • Cho Eun-Ae;Moon Chang-Joo;Baik Doo-Kwon
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.12 no.3
    • /
    • pp.192-207
    • /
    • 2006
  • Today, the SSL protocol has been used as core part in various computing environments or security systems. But, the SSL protocol has several problems, because of the rigidity on operating. First, SSL protocol brings considerable burden to the CPU utilization so that performance of the security service in encryption transaction is lowered because it encrypts all data which is transferred between a server and a client. Second, SSL protocol can be vulnerable for cryptanalysis due to the key in fixed algorithm being used. Third, it is difficult to add and use another new cryptography algorithms. Finally. it is difficult for developers to learn use cryptography API(Application Program Interface) for the SSL protocol. Hence, we need to cover these problems, and, at the same time, we need the secure and comfortable method to operate the SSL protocol and to handle the efficient data. In this paper, we propose the SSL component which is designed and implemented using CBD(Component Based Development) concept to satisfy these requirements. The SSL component provides not only data encryption services like the SSL protocol but also convenient APIs for the developer unfamiliar with security. Further, the SSL component can improve the productivity and give reduce development cost. Because the SSL component can be reused. Also, in case of that new algorithms are added or algorithms are changed, it Is compatible and easy to interlock. SSL Component works the SSL protocol service in application layer. First of all, we take out the requirements, and then, we design and implement the SSL Component, confidentiality and integrity component, which support the SSL component, dependently. These all mentioned components are implemented by EJB, it can provide the efficient data handling when data is encrypted/decrypted by choosing the data. Also, it improves the usability by choosing data and mechanism as user intend. In conclusion, as we test and evaluate these component, SSL component is more usable and efficient than existing SSL protocol, because the increase rate of processing time for SSL component is lower that SSL protocol's.

THE CURRENT STATUS OF BIOMEDICAL ENGINEERING IN THE USA

  • Webster, John G.
    • Proceedings of the KOSOMBE Conference
    • /
    • v.1992 no.05
    • /
    • pp.27-47
    • /
    • 1992
  • Engineers have developed new instruments that aid in diagnosis and therapy Ultrasonic imaging has provided a nondamaging method of imaging internal organs. A complex transducer emits ultrasonic waves at many angles and reconstructs a map of internal anatomy and also velocities of blood in vessels. Fast computed tomography permits reconstruction of the 3-dimensional anatomy and perfusion of the heart at 20-Hz rates. Positron emission tomography uses certain isotopes that produce positrons that react with electrons to simultaneously emit two gamma rays in opposite directions. It locates the region of origin by using a ring of discrete scintillation detectors, each in electronic coincidence with an opposing detector. In magnetic resonance imaging, the patient is placed in a very strong magnetic field. The precessing of the hydrogen atoms is perturbed by an interrogating field to yield two-dimensional images of soft tissue having exceptional clarity. As an alternative to radiology image processing, film archiving, and retrieval, picture archiving and communication systems (PACS) are being implemented. Images from computed radiography, magnetic resonance imaging (MRI), nuclear medicine, and ultrasound are digitized, transmitted, and stored in computers for retrieval at distributed work stations. In electrical impedance tomography, electrodes are placed around the thorax. 50-kHz current is injected between two electrodes and voltages are measured on all other electrodes. A computer processes the data to yield an image of the resistivity of a 2-dimensional slice of the thorax. During fetal monitoring, a corkscrew electrode is screwed into the fetal scalp to measure the fetal electrocardiogram. Correlations with uterine contractions yield information on the status of the fetus during delivery To measure cardiac output by thermodilution, cold saline is injected into the right atrium. A thermistor in the right pulmonary artery yields temperature measurements, from which we can calculate cardiac output. In impedance cardiography, we measure the changes in electrical impedance as the heart ejects blood into the arteries. Motion artifacts are large, so signal averaging is useful during monitoring. An intraarterial blood gas monitoring system permits monitoring in real time. Light is sent down optical fibers inserted into the radial artery, where it is absorbed by dyes, which reemit the light at a different wavelength. The emitted light travels up optical fibers where an external instrument determines O2, CO2, and pH. Therapeutic devices include the electrosurgical unit. A high-frequency electric arc is drawn between the knife and the tissue. The arc cuts and the heat coagulates, thus preventing blood loss. Hyperthermia has demonstrated antitumor effects in patients in whom all conventional modes of therapy have failed. Methods of raising tumor temperature include focused ultrasound, radio-frequency power through needles, or microwaves. When the heart stops pumping, we use the defibrillator to restore normal pumping. A brief, high-current pulse through the heart synchronizes all cardiac fibers to restore normal rhythm. When the cardiac rhythm is too slow, we implant the cardiac pacemaker. An electrode within the heart stimulates the cardiac muscle to contract at the normal rate. When the cardiac valves are narrowed or leak, we implant an artificial valve. Silicone rubber and Teflon are used for biocompatibility. Artificial hearts powered by pneumatic hoses have been implanted in humans. However, the quality of life gradually degrades, and death ensues. When kidney stones develop, lithotripsy is used. A spark creates a pressure wave, which is focused on the stone and fragments it. The pieces pass out normally. When kidneys fail, the blood is cleansed during hemodialysis. Urea passes through a porous membrane to a dialysate bath to lower its concentration in the blood. The blind are able to read by scanning the Optacon with their fingertips. A camera scans letters and converts them to an array of vibrating pins. The deaf are able to hear using a cochlear implant. A microphone detects sound and divides it into frequency bands. 22 electrodes within the cochlea stimulate the acoustic the acoustic nerve to provide sound patterns. For those who have lost muscle function in the limbs, researchers are implanting electrodes to stimulate the muscle. Sensors in the legs and arms feed back signals to a computer that coordinates the stimulators to provide limb motion. For those with high spinal cord injury, a puff and sip switch can control a computer and permit the disabled person operate the computer and communicate with the outside world.

  • PDF

The Efficient Merge Operation in Log Buffer-Based Flash Translation Layer for Enhanced Random Writing (임의쓰기 성능향상을 위한 로그블록 기반 FTL의 효율적인 합병연산)

  • Lee, Jun-Hyuk;Roh, Hong-Chan;Park, Sang-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.19D no.2
    • /
    • pp.161-186
    • /
    • 2012
  • Recently, the flash memory consistently increases the storage capacity while the price of the memory is being cheap. This makes the mass storage SSD(Solid State Drive) popular. The flash memory, however, has a lot of defects. In order that these defects should be complimented, it is needed to use the FTL(Flash Translation Layer) as a special layer. To operate restrictions of the hardware efficiently, the FTL that is essential to work plays a role of transferring from the logical sector number of file systems to the physical sector number of the flash memory. Especially, the poor performance is attributed to Erase-Before-Write among the flash memory's restrictions, and even if there are lots of studies based on the log block, a few problems still exists in order for the mass storage flash memory to be operated. If the FAST based on Log Block-Based Flash often is generated in the wide locality causing the random writing, the merge operation will be occur as the sectors is not used in the data block. In other words, the block thrashing which is not effective occurs and then, the flash memory's performance get worse. If the log-block makes the overwriting caused, the log-block is executed like a cache and this technique contributes to developing the flash memory performance improvement. This study for the improvement of the random writing demonstrates that the log block is operated like not only the cache but also the entire flash memory so that the merge operation and the erase operation are diminished as there are a distinct mapping table called as the offset mapping table for the operation. The new FTL is to be defined as the XAST(extensively-Associative Sector Translation). The XAST manages the offset mapping table with efficiency based on the spatial locality and temporal locality.

IPv6 Migration, OSPFv3 Routing based on IPv6, and IPv4/IPv6 Dual-Stack Networks and IPv6 Network: Modeling, and Simulation (IPv6 이관, IPv6 기반의 OSPFv3 라우팅, IPv4/IPv6 듀얼 스택 네트워크와 IPv6 네트워크: 모델링, 시뮬레이션)

  • Kim, Jeong-Su
    • The KIPS Transactions:PartC
    • /
    • v.18C no.5
    • /
    • pp.343-360
    • /
    • 2011
  • The objective of this paper is to analyze and characterize to simulate routing observations on end-to-end routing circuits and a ping experiment of a virtual network after modeling, such as IPv6 migration, an OSPFv3 routing experiment based on an IPv6 environment, and a ping experiment for IPv4/IPv6 dual-stack networks and IPv6 network for OSPFv3 routing using IPv6 planning and operations in an OPNET Modeler. IPv6 deployment based largely on the integrated wired and wireless network was one of the research tasks at hand. The previous studies' researchers recommended that future research work be done on the explicit features of both OSPFv3 and EIGRP protocols in the IPv4/IPv6 environment, and more research should be done to explore how to improve the end-to-end IPv6 performance. Also, most related work was performed with an IPv4 environment but lacked studies related to the OSPFv3 virtual network based on an end-to-end IPv6 environment. Hence, this research continues work in previous studies in analyzing IPv6 migration, an OSPFv3 routing experiment based on IPv6, and a ping experiment for IPv4/IPv6 dual-stack networks and IPv6 network for OSPFv3 routing. In the not too distant future, before enabling the default IPv6, it would help to understand network design and deployment based on an IPv6 environment through IPv6 planning and operations for the end-user perspective such as success or failure of connection on IPv6 migration, exploration of an OSPFv3 routing circuit based on an end-to-end IPv6 environment, and a ping experiment for IPv4/IPv6 dual-stack networks and IPv6 network for OSPFv3 routing. We were able to observe an optimal route for modeling of an end-to-end virtual network through simulation results as well as find what appeared to be a fast ping response time VC server to ensure Internet quality of service better than an HTTP server.

Inexpensive Visual Motion Data Glove for Human-Computer Interface Via Hand Gesture Recognition (손 동작 인식을 통한 인간 - 컴퓨터 인터페이스용 저가형 비주얼 모션 데이터 글러브)

  • Han, Young-Mo
    • The KIPS Transactions:PartB
    • /
    • v.16B no.5
    • /
    • pp.341-346
    • /
    • 2009
  • The motion data glove is a representative human-computer interaction tool that inputs human hand gestures to computers by measuring their motions. The motion data glove is essential equipment used for new computer technologiesincluding home automation, virtual reality, biometrics, motion capture. For its popular usage, this paper attempts to develop an inexpensive visual.type motion data glove that can be used without any special equipment. The proposed approach has the special feature; it can be developed as a low-cost one becauseof not using high-cost motion-sensing fibers that were used in the conventional approaches. That makes its easy production and popular use possible. This approach adopts a visual method that is obtained by improving conventional optic motion capture technology, instead of mechanical method using motion-sensing fibers. Compared to conventional visual methods, the proposed method has the following advantages and originalities Firstly, conventional visual methods use many cameras and equipments to reconstruct 3D pose with eliminating occlusions But the proposed method adopts a mono vision approachthat makes simple and low cost equipments possible. Secondly, conventional mono vision methods have difficulty in reconstructing 3D pose of occluded parts in images because they have weak points about occlusions. But the proposed approach can reconstruct occluded parts in images by using originally designed thin-bar-shaped optic indicators. Thirdly, many cases of conventional methods use nonlinear numerical computation image analysis algorithm, so they have inconvenience about their initialization and computation times. But the proposed method improves these inconveniences by using a closed-form image analysis algorithm that is obtained from original formulation. Fourthly, many cases of conventional closed-form algorithms use approximations in their formulations processes, so they have disadvantages of low accuracy and confined applications due to singularities. But the proposed method improves these disadvantages by original formulation techniques where a closed-form algorithm is derived by using exponential-form twist coordinates, instead of using approximations or local parameterizations such as Euler angels.

Incremental Ensemble Learning for The Combination of Multiple Models of Locally Weighted Regression Using Genetic Algorithm (유전 알고리즘을 이용한 국소가중회귀의 다중모델 결합을 위한 점진적 앙상블 학습)

  • Kim, Sang Hun;Chung, Byung Hee;Lee, Gun Ho
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.7 no.9
    • /
    • pp.351-360
    • /
    • 2018
  • The LWR (Locally Weighted Regression) model, which is traditionally a lazy learning model, is designed to obtain the solution of the prediction according to the input variable, the query point, and it is a kind of the regression equation in the short interval obtained as a result of the learning that gives a higher weight value closer to the query point. We study on an incremental ensemble learning approach for LWR, a form of lazy learning and memory-based learning. The proposed incremental ensemble learning method of LWR is to sequentially generate and integrate LWR models over time using a genetic algorithm to obtain a solution of a specific query point. The weaknesses of existing LWR models are that multiple LWR models can be generated based on the indicator function and data sample selection, and the quality of the predictions can also vary depending on this model. However, no research has been conducted to solve the problem of selection or combination of multiple LWR models. In this study, after generating the initial LWR model according to the indicator function and the sample data set, we iterate evolution learning process to obtain the proper indicator function and assess the LWR models applied to the other sample data sets to overcome the data set bias. We adopt Eager learning method to generate and store LWR model gradually when data is generated for all sections. In order to obtain a prediction solution at a specific point in time, an LWR model is generated based on newly generated data within a predetermined interval and then combined with existing LWR models in a section using a genetic algorithm. The proposed method shows better results than the method of selecting multiple LWR models using the simple average method. The results of this study are compared with the predicted results using multiple regression analysis by applying the real data such as the amount of traffic per hour in a specific area and hourly sales of a resting place of the highway, etc.